Next Article in Journal
Enhanced Modbus/TCP Security Protocol: Authentication and Authorization Functions Supported
Previous Article in Journal
Multi-Sensor Image Fusion Method for Defect Detection in Powder Bed Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data

1
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou 310023, China
2
Noblelift Intelligent Equipment Co., Ltd., Huzhou 313100, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 8019; https://doi.org/10.3390/s22208019
Submission received: 9 September 2022 / Revised: 4 October 2022 / Accepted: 18 October 2022 / Published: 20 October 2022
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Automated guided vehicles are widely used in warehousing environments for automated pallet handling, which is one of the fundamental parts to construct intelligent logistics systems. Pallet detection is a critical technology for automated guided vehicles, which directly affects production efficiency. A novel pallet detection method for automated guided vehicles based on point cloud data is proposed, which consists of five modules including point cloud preprocessing, key point extraction, feature description, surface matching and point cloud registration. The proposed method combines the color with the geometric features of the pallet point cloud and constructs a new Adaptive Color Fast Point Feature Histogram (ACFPFH) feature descriptor by selecting the optimal neighborhood adaptively. In addition, a new surface matching method called the Bidirectional Nearest Neighbor Distance Ratio-Approximate Congruent Triangle Neighborhood (BNNDR-ACTN) is proposed. The proposed method overcomes the problems of current methods such as low efficiency, poor robustness, random parameter selection, and being time-consuming. To verify the performance, the proposed method is compared with the traditional and modified Iterative Closest Point (ICP) methods in two real-world cases. The results show that the Root Mean Square Error (RMSE) is reduced to 0.009 and the running time is reduced to 0.989 s, which demonstrates that the proposed method has faster registration speed while maintaining higher registration accuracy.

1. Introduction

Under the background of “Industry 4.0”, the logistics industry is facing challenges, including structural adjustment, industrial optimization, cost reduction and efficiency improvement, and also has ushered in development opportunities such as information technology, intelligent logistics and machine vision [1]. As an important part of an intelligent logistics system, automated guided vehicles are widely used in warehousing, production, service, aerial work and other scenarios, which can establish a human–machine friendly interactive environment and reduce the incidence of safety accidents [2]. However, in the actual storage environment, due to the influence of many factors such as excessive obstacles, uneven illumination, accumulated handling errors and manual intervention, automated guided vehicles have problems of low efficiency and repeated handling in the process of pallet handling [3]. With the help of a 3D vision sensor, automated guidance vehicles can detect the scene pallet, which can effectively solve these problems. Pallet detection for automated guided vehicles is widely used in various scenarios including storage shelves, the production workshop, drug transport and blast furnace conditions, which are shown in Figure 1. The applications of automated guided vehicles in these scenarios can establish a human–machine friendly interactive environment, improve production efficiency and reduce the incidence of safety accidents.
The existing vision-based object detection methods are mainly divided into two categories: the image-based method [4] and the point cloud-based method [5,6]. There has been a large amount of research on the object detection method based on images [7,8,9,10,11]. Specific to pallet detection, Li et al. [12] applied the Region Growing algorithm to extract the whole target region, and the pallet was located by the Progressive Probabilistic Hough Transform (PPHT) method, which solved the problem of difficult target detection under unstable light conditions. Syu et al. [13] used the monocular vision system on the forklift and combined the Adaptive Structure Feature (ASF) and Direction Weighted Overlapping (DWO) ratio to detect the pallet, which removes most of the non-stationary background and significantly increases the processing efficiency. Li et al. [14] established the pallet dataset and applied the improved deep learning object detection algorithm to obtain detection results, which improved the efficiency and accuracy of the pallet detection. The above methods of object detection based on 2D images have been intensively investigated, which is currently a relatively mature research area. However, the imaging process of 2D images involves mapping from 3D space to 2D space, which loses a lot of useful information during the mapping process. Therefore, object detection based on 2D images can no longer satisfy the needs of current industrial production.
With the rapid development of low-cost depth sensors, object detection has converted from traditional single point and segment measurement to dense point cloud and full profile measurement [15,16,17]. Compared with 2D images, 3D point cloud data provide more information about color, texture, geometric feature and space distribution [18], which makes pallet detection based on the 3D point cloud an active research topic. Firstly, the methods based on artificial features were attached to the pallets. Seelinger et al. [19] presented a vision-based approach to identify the fiducials which were placed on each pallet, which provides automated guided vehicle systems with the capability of performing pallet detection tasks. Two reflectors were fixed left and right on the short side of the pallet in the study by Lecking [20] to realize pallet detection. Although these artificial features simplify pallet detection, it takes effort to label all of the pallets in advance, thereby identifying the above approaches as unfeasible. Guo et al. [21] summarized the existing local feature detection methods and concluded that the contradiction between descriptiveness and computational efficiency of the local feature descriptor was a major challenge faced in feature extraction. Hence, it was essential to construct a robust and descriptive feature descriptor. The Fast Point Feature Histogram (FPFH) is a commonly used local feature descriptor which can perform well in descriptiveness, robustness and efficiency [22]. Additionally, FPFH employed the geometric feature of the pallet to build a descriptor without adding any artificial marks. Tao et al. [23] combined SVM classification and the FPFH descriptor to achieve object detection, which improved the robot detection ability and perception in three-dimensional space. A new point registration algorithm that combines FPFH and greedy projection triangulation was presented by Liu et al. [24], which improved the accuracy of registration. Li et al. [25] proposed a novel method of point registration called the Four Initial Point Pairs (FIPP) algorithm based on the FPFH feature descriptor, and the accuracy of FIPP could reach a better level, but it had low efficiency in mass data. However, few studies considered the color information and the criteria for the selection of the neighborhood radius in the FPFH descriptor. Most researchers adjusted the neighborhood radius manually based on prior knowledge, with certain randomness, low efficiency and high complexity.
In response to the above problems, a novel pallet detection method for automated guided vehicles based on point cloud data is proposed, including point cloud preprocessing, key point extraction, feature description, surface matching and point cloud registration. The main contributions can be summarized as: (1) the proposed method considers the HSV color feature, which improves the detection accuracy; (2) an ACFPFH feature descriptor is proposed and the criteria for adaptive selection of the optimal neighborhood radius are established; (3) a new surface matching method called the Bidirectional Nearest Neighbor Distance Ratio-Approximate Congruent Triangle Neighborhood (BNNDR-ACTN) is proposed, which increases the efficiency and accuracy. The proposed method not only overcomes the drawback of randomness and inefficiency of neighborhood selection in traditional feature extraction but also improves the accuracy and efficiency of pallet detection. Moreover, the proposed method can be well adapted to a variety of complex scenes such as the ground and the shelf.
The rest of the paper is organized as follows: In Section 2, the proposed pallet detection method based on the ACFPFH feature descriptor is described. Section 3 outlines two specific case studies and further comparison analysis for verifying the proposed method in engineering applications. Finally, Section 4 concludes this paper.

2. The Proposed Method

2.1. Overview of the Proposed Method

This section describes an overview of the proposed method. It consists of five modules: point cloud preprocessing, key point extraction, feature description, surface matching and point cloud registration. The framework of the proposed method is shown in Figure 2. The procedure involves the following steps.
Step 1: Point cloud preprocessing. The Percipio FM851-E2 3D vision sensor is used to acquire the point cloud data which represent the whole scene, including the pallet. Outliers are eliminated and the redundant information, such as walls and grounds, is removed using Random Sample Consensus (RANSAC) algorithm.
Step 2: Key point detection. The key points with rich information are extracted from scene point clouds by the Intrinsic Shape Signatures (ISS) algorithm.
Step 3: Feature description. The optimal neighborhood radius of each point is obtained based on the minimum criterion of the neighborhood feature entropy function. The color components and the geometric information based the optimal neighborhood of the key point are encoded into a representative feature descriptor called the Adaptive Color Fast Point Feature Histogram. The pallet template point cloud and its corresponding library of feature descriptors are obtained by performing the above three steps.
Step 4: Surface matching. The matching method based on the Bidirectional Nearest Neighbor Distance Ratio (BNNDR) is employed to complete feature matching between the pallet template point cloud and the scene point cloud. Considering that there are some incorrect matching point pairs which will reduce the registration accuracy, it is essential to eliminate them by the Approximate Congruent Triangle Neighborhood (ACTN).
Step 5: Point cloud registration. The RANSAC algorithm is applied for performing point cloud coarse registration, which can obtain the relationship between the template point cloud and the scene point cloud and provide an ideal initial position for fine registration. The fine registration works to obtain a final optimal transformation matrix using the Iterative Closest Point (ICP) algorithm.

2.2. Point Cloud Preprocessing

2.2.1. Outliers Elimination

Due to the hardware design of the Percipio FM851-E2 3D vision sensor, external environmental interference and other factors, point cloud outliers are inevitable in the measurement. The pallet detection results will have errors if the outliers in the original scene point cloud Qso are not eliminated.
The distance from arbitrary point P i in the point cloud to its neighborhood point P ik ( k = 1 , 2 , , m ) is approximately subject to Gaussian distribution, and the probability density function of the average neighborhood distance is listed below:
f ( d i ) = 1 2 π σ exp ( ( d i μ ) 2 2 σ 2 )
where i = 1 , 2 , , n , n represents the number of points in the point cloud, d i is the average neighborhood distance of arbitrary point P i , μ and σ are the expectation and standard deviation of the average neighborhood distance d i , respectively. Calculating the average neighborhood distance d i , the point P i is considered as an outlier and removed if μ σ < d i < μ + σ .

2.2.2. Plane Segmentation

In warehousing environments, the scene point cloud acquired by the Percipio FM851-E2 3D vision sensor contains a lot of redundant information, such as the grounds and the walls, which will decrease the calculation efficiency. Therefore, it is necessary to remove the useless planes in the scene point cloud [26]. The specific segmentation procedures are as follows:
Step 1: The plane equation in the three-dimensional point cloud is defined as:
Ax + By + Cz + D = 0
where A, B and C are plane parameters, and D is the distance from the plane to the point Pi. Randomly select three points from the scene point cloud QSE after removing outliers and obtain the parameters of the initial plane PI.
Step 2: Calculate the distance Di from the point Pi to the initial plane PI and the angle β i between the point Pi and the normal vector of the initial plane P I . Set distance threshold D ε and angle threshold β ε ; if both D i < D ε and β i < β ε are satisfied, the point P i belongs to the plane PI.
Step 3: Repeat the above procedures until the number of the points in the plane reaches the threshold t, and remove the final fitted plane model to obtain the preprocessed scene point cloud Qs.

2.3. Key Point Extraction

The preprocessed scene point cloud Qs still contains a large number of points, which leads to low efficiency of feature extraction and matching. Selecting key points to simplify the point clouds can retain the features of the point clouds as much as possible while reducing the number of the points. The Intrinsic Shape Signatures (ISS) is a widely used algorithm with a fast calculation speed and high repeatability to realize key point extraction [27]. The extraction procedures of the key points PFi are summarized as follows:
Step 1: The neighborhood points P ik ( k = 1 , 2 , , m ) of P i in the scene point cloud Q S are searched within a certain radius dp. dp is the average closest point distance of the point cloud collected by the 3D vision sensor, which can be calculated as follows:
d p = 1 N d m
where N is the number of the points, and d m is the distance between each point and its closest point.
Compute a weight parameter ω ik for each point P i inversely related to the distance from P ik to P i as follows:
ω ik = 1 P ik P i
Step 2: The covariance matrix Ci of point Pi is generated as follows:
C i = k = 1 m ω ik ( P i k P i ) ( P i k P i ) T k = 1 m ω ik
where m is the number of the neighborhood points P ik , and P i is the center point of the neighborhood points P ik and P i = 1 m k = 1 m P ik .
Step 3: Calculate the eigenvalues of the covariance matrix Ci and sort them from large to small as { λ 1 i , λ 2 i , λ 3 i } .
Step 4: Set the thresholds k2 and k1; the points with the relation of λ 2 λ 1 < k 1 λ 3 λ 2 < k 2 are considered as the key point PFi.

2.4. Feature Description

As for the traditional feature descriptors, the neighborhood radius for all the points is a fixed value. Most studies select the appropriate radius based on empirical knowledge, which has strong subjectivity and low efficiency. Besides, the color information is ignored, which makes it difficult to fully and accurately characterize the objects. Therefore, a novel feature descriptor called ACFPFH is defined, which adaptively selects the optimal neighborhood radius and considers the color and geometric features. The flowchart of the proposed feature description method is shown in Figure 3, and the detailed procedures are described as follows.

2.4.1. Adaptive Optimal Neighborhood Selection

The choice is preferred where the radius is more flexible and allowed to vary within a dataset. The proper neighborhood radius obtained by adaptive selection can reduce the runtime of feature extraction under the premise of ensuring precision. Therefore, a general method for obtaining the adaptive optimal neighborhood radius ropt is proposed in this paper, without the limit of prior knowledge. The detailed procedures about depriving the adaptive optimal neighborhood radius ropt are described as follows.
Step 1: Set the radius range [rmin, rmax] and change interval Δr of the neighborhood search. Set the value of r min equal to the average closest point distance dp, and r max is the maximal acceptable neighborhood radius for all the points of the scene point cloud Qs, which can usually be set to a fixed value.
Considering the radius of interest is usually closer to r min than to r max , the radius is calculated as follows:
r 1 = r min ,   r j + 1 = { r j + Δ r if   r j + 1 < r mid r j + 2 Δ r else
where r mid = r min + r max 2 , j = 1 , 2 , 3 , until r j > r max and Δ r is the adaptive neighborhood radius value step. It results in more reasonable samples near the radius of interest and less when reaching the maximal values.
Step 2: Calculate the covariance matrix Cj and eigenvalues λ 1 , λ 2 , λ 3 of each neighborhood radius r j , which can determine the dimensionality characteristics of the local neighborhood. Table 1 shows the details about the dimensionality characteristics [28]. Construct the dimensionality features, including the 1D linearity feature L λ , 2D planarity feature P λ and 3D scattering feature S λ . They are represented as:
L λ = λ 1 λ 2 λ 1 ,   P λ = λ 2 λ 3 λ 1 ,   S λ = λ 3 λ 1
where L λ + P λ + S λ = 1 , and each of them can be regarded as the probability of the point P i being labeled as a 1D, 2D or 3D structure. Consequently, the task of searching for an optimal neighborhood size can be converted to finding which radius favors the corresponding dimensionality.
Step 3: The entropy function of local neighborhood Eneiborhood is established as a measure of unpredictability based on information entropy theory, and it is defined as [29]:
E neighborhood = L λ ln ( L λ ) P λ ln ( P λ ) S λ ln ( S λ )
The smaller the value of the information entropy, the smaller the uncertainty of the variable, which is the core of the Shannon entropy theory [30]. Accordingly, it can be concluded that the smaller the information entropy value of the local neighborhood, the less the uncertainty of the dimensional feature of the points. That is, the greater the probability that the point belongs to a certain dimensional feature, and the more similar the spatial distribution characteristics of the local data points under the neighborhood radius, then the neighborhood radius tends to be more optimal. More immediately, it is feasible to obtain the adaptive optimal neighborhood radius r e - opt according to the minimum criterion of neighborhood entropy function.
r e - opt = argmin ( E neighborhood )
However, the optimal neighborhood radius r e - opt obtained according to Equations (7) and (8) is based on the assumption that obvious dimensionality characteristics exist in the observed point cloud. When the dimensionality features of the point P i are indistinguishable, the optimality of the estimated neighborhood cannot be determined.
In order to avoid the limitation of the above assumptions for the scene point cloud and improve the estimation accuracy of the optimal neighborhood, a more general solution for calculating the optimal neighborhood radius r opt is proposed in this paper. The eigenvalues directly reflect the dimensional distribution characteristics of the neighborhood points. Consequently, the three eigenvalues are normalized by their sum λ j for obtaining an eigen entropy E e that is defined as:
E e = e 1 ln ( e 1 ) e 2 ln ( e 2 ) e 3 ln ( e 3 )
where the e j = λ j / λ j for j { 1 , 2 , 3 } represents the normalized eigenvalues summing up to 1. The optimal neighborhood radius r opt is obtained according to the minimum criterion of eigen entropy E e .

2.4.2. ACFPFH Feature Description

The ACFPFH feature descriptor consisting of a 3-dimensional HSV color feature and the 33-dimensional FPFH geometric feature is proposed in this section, which is shown as Equation (11):
ACFPFH ( PF i ) = HSV ( PF i ) + FPFH ( PF i )
where the PFi is the key points of the pallet point cloud. The specific calculation procedures of color feature and geometric feature are as follows:
(1) Color feature calculation
The point cloud data acquired by the Percipio FM851-E2 3D vision sensor contain information such as the color and coordinates of the object. Due to the high correlation between components in RGB color space, color cognitive properties cannot be intuitively expressed. Therefore, RGB color space is not suitable for feature similarity detection. Compared with RGB color space, HSV color space is easier to distinguish and more consistent with human visual characteristics. H represents the hue, S represents the saturation and V represents the value. HSV color space is exploited to form a color feature descriptor of the key point PF i , and it can be converted from the RGB color space [31].
V = max ( R , G , B ) S = { 0 ,   if   V = 0 max ( R , G , B ) min ( R , G , B ) max ( R , G , B ) ,   otherwise H = { 0 ,   if   S = 0 60 × ( G B ) / ( S × V ) ,   if   S 0   and   V = R 60 × ( 2 + ( B R ) / ( S × V ) ) ,   if   S 0   and   V = G 60 × ( 4 + ( R G ) / ( S × V ) ) ,   otherwise if   H < 0 ,   H = H + 360
where the value range of R, G and B is [0,255], and the value ranges of H, S and V are [0,360], [0,1] and [0,255], respectively.
(2) Geometric feature calculation
FPFH is an efficient local feature descriptor which reflects the normal relationship between query points and neighborhood points of point cloud data. The detailed calculation procedures are explained as follows:
Step 1: For each key point PF i (or query point P q ), select all of the neighborhood points P qj of the query point P q that are enclosed in the sphere with an adaptive optimal neighborhood r opt , as shown in Figure 4. The red point Pq in the middle of the figure is the query point, and the colored points Pq1-Pq5 in the black circle are the neighborhood points of Pq, and those blue points Pq6-Pq15 are the neighborhood points of the colored points Pq1–Pq5.
Step 2: The point pairs p s , p t are generated based on the query point P q and the neighborhood points P qj . Estimate their corresponding normal n s and n t . The relative relationship between the point pairs p s , p t is obtained by establishing a local frame, as shown in Figure 5.
Taking the point p s as the coordinate origin, the coordinate frame is set up with u , v and w axes. The axis is defined as:
u = n s v = ( p t p s ) × u w = u × v
Step 3: The angles α , φ and θ are calculated for representing the deviation between the normal vectors n s and n t , which forms the simplified point feature histograms (SPFH).
α = arccos ( v n t | v | | n t | ) φ = arccos ( u p | u | | p | ) θ = arctan ( w n t , u n t )
where p represents ( p t p s ) p t p s 2 .
Step 4: For each neighborhood point P qj of PF i , the r opt is re-determined and the neighboring SPFH value is used to weight the final histograms of PF i , whose results are called FPFH.
FPFH ( PF i ) = SPFH ( PF i ) + 1 k i = 1 k 1 ω SPFH ( P qj )
where k represents the number of the neighborhood point Pqj, and ω represents the weight, which is the reciprocal of the distance between Ps and Pt. Figure 6 shows an example of the ACFPFH of one point.

2.5. Surface Matching

Accurate surface matching is an important prerequisite for point cloud registration, which directly affects the performance of pallet detection. For the traditional surface matching method, the one-way feature matching is performed and the method for eliminating incorrect matching point pairs only considers the relationship between points, leading to too many incorrect matching pairs. Therefore, a new surface matching method called BNNDR-ACTN is proposed, which includes feature matching based on the Bidirectional Nearest Neighbor Distance Ratio (BNNDR) and the incorrect matching point pairs’ elimination based on the Approximate Congruent Triangle Neighborhood (ACTN). The architecture of the proposed surface matching method is shown in Figure 7, and the detailed procedures are described as follows.
Module 1: Feature matching
The purpose of point cloud feature matching is to establish the relationship between the feature descriptors of the template point cloud and the scene point cloud, thereby obtaining the initial matching point pair.
Step 1: Forward matching
Define F M = { f M i } and F S = { f S j } as the sets of ACFPFH descriptors of the pallet template point cloud Q M and the scene point cloud Q S , respectively. For each ACFPFH descriptor f S j belonging to the scene point cloud, obtain the nearest ACFPFH descriptor f M i and the second-nearest ACFPFH descriptor f M i in the template point cloud, and their Euclidean distances are represented by d ( f S j , f M i ) and d ( f S j , f M i ) . If the ratio of their distances satisfies the Equation (16), ( P F S j , P F M i ) can be considered as a candidate matching key point pair. That is, the key point P F S j corresponds to f S j , and the key point P F M i corresponds to f M i .
d ( f S j , f M i ) d ( f S j , f M i ) < t h
where the threshold th is a constant between 0 and 1.
Step 2: Backward matching
For each ACFPFH descriptor f M i belonging to the template point cloud, the nearest ACFPFH descriptor f S j and the second-nearest ACFPFH descriptor f S j are obtained in the scene point cloud, and their Euclidean distances are represented by d ( f M i , f S j ) and d ( f M i , f S j ) . If the ratio of their distances satisfies the Equation (17), ( P F S j , P F M i ) is a matching key point pair; otherwise, it is not a matching key point pair.
d ( f M i , f S j ) d ( f M i , f S j ) < t h
If the key points P F S j and P F S j are the same point, ( P F S j , P F M i ) can be considered as a matching key point pair. The final matching key point pairs set M P = { Q S i , Q M i } is obtained by repeating the above steps.
Module 2: Elimination of wrong matching point pairs
The surface of the object is rough and noisy, which leads to some mismatching point pairs. Therefore, after obtaining initial matching pairs, the next step is to eliminate the wrong matching point pairs.
Step 3: Triangle neighborhood generation
Select a query point pair ( Q S 1 , Q M 1 ) from the matching key point pairs set M P = { Q S j , Q M i } and search for the nearest point pairs ( Q S 2 , Q M 2 ) and the second nearest point pairs ( Q S 3 , Q M 3 ) , which can generate triangle neighborhood T S = ( Q S 1 , Q S 2 , Q S 3 ) and T M = ( Q M 1 , Q M 2 , Q M 3 ) .
Step 4: Obtain the correct matching point pairs
The point pair ( Q S 1 , Q M 1 ) is considered as a correct matching point pair if the two triangles T S = ( Q S 1 , Q S 2 , Q S 3 ) and T M = ( Q M 1 , Q M 2 , Q M 3 ) are approximately congruent; otherwise, it will be regarded as a wrong matching point pair and should be eliminated. The Equation (18) is used to determine whether the two triangles are approximately congruent.
{ t < dist ( Q M 1 , Q M 2 ) dist ( Q S 1 , Q S 2 ) W < t t < dist ( Q M 1 , Q M 3 ) dist ( Q S 1 , Q S 3 ) W < t t < dist ( Q M 2 , Q M 3 ) dist ( Q S 2 , Q S 3 ) W < t
where dist (   ·   ) represents the distance between two points, W = max ( dist ) , and t represents the degree of approximation between the point pairs. Then, the final correct matching point pairs set C P = { Q C S i , Q C M i } is obtained by repeating the above steps.
Considering the stability of the triangle, each point in the point cloud is expanded into a triangular neighborhood. Therefore, the point-to-point matching problem is transformed into the neighborhood matching problem, which can obtain more feature information and improve registration accuracy. In addition, each point in the point cloud is regarded as the vertex of the triangle, which benefits the maintenance of the geometric characteristics of the original point cloud.

2.6. Point Cloud Registration

2.6.1. Coarse Registration

The main task of point cloud coarse registration is to obtain the relationship between the template point cloud Q M and scene point cloud Q S and provide an ideal initial position for fine registration. This computation is based on the correct matching point pairs set C P = { Q C S i , Q C M i } . The main steps are as follows:
Step 1: Three correspondences are randomly selected to estimate the rigid transformation matrix R 0 and T 0 .
Step 2: Calculate the distance D ( R 0 , T 0 ) between the point Q C S i and the transformed point Q T C M i based on the transformation matrix R 0 and T 0 . Take the point in point set Q C S whose corresponding distance D ( R 0 , T 0 ) is less than threshold d0 as the inlier point; otherwise, consider it as an exterior point.
D ( R 0 , T 0 ) = Q C S i ( R 0 × Q C M i + T 0 )
Step 3: Repeat the above steps to obtain a different rigid transformation matrix and count its corresponding number of inliers point until the maximum iteration number I0 is reached.
Step 4: Obtain the final rigid transformation matrix R 0 and T 0 with the most interior points; the template point cloud Q M is transformed into the coordinate system of the scene point cloud Q S to complete the coarse registration. Define the transformed template point cloud Q M as Q MT .

2.6.2. Fine Registration

The Iterative Closest Point (ICP) algorithm is used to achieve point cloud fine registration. It is based on minimizing the error function to calculate the optimal rotation matrix and translation matrix. The specific procedures of fine registration are as follows:
Step 1: For each point Q M T i in the transformed template point cloud Q M T , search for its nearest neighbor point Q S i in the scene point cloud Q S , thereby generating the corresponding points pairs set C F = { Q M T i , Q S i } .
Step 2: Use the least square method to solve the rotation matrix R n and translation matrix T n with the smallest average distance e n between the corresponding points.
e n = 1 k i = 1 k Q S i ( Q M T i × R n + T n ) 2
where n is the number of iterations and k is the number of the corresponding points.
Step 3: Repeat the above steps and obtain the optimal rotation matrix R f and translation matrix T f until e n is smaller than distance threshold e f , or the maximum number of iterations I f is reached. The new template point cloud Q MF is obtained by using the transformation matrix R f and T f , and fine registration is completed.

3. Case Studies

3.1. Evaluation Index

In order to validate the performance of the proposed ACFPFH feature descriptor and the overall registration method, two representative indicators are developed to evaluate the experiment results, and the details of these indicators are briefly described as follows. The experiment was performed using MATLAB Code on a desktop with 3.6 GHz inter® Core™ i7-11700kf CPU and 16 G memory.
(1) Precision-recall curve
The precision–recall curve (PRC) is used to evaluate the descriptiveness of a feature descriptor. The precision is calculated as the number of correct matching point pairs with respect to the total number of matching point pairs:
Precision = N CP N MP
where N CP represents the number of correct matching point pairs, and N MP represents the number of matching point pairs.
The recall is calculated as the number of correct matching point pairs with respect to the number of key points of the template point cloud:
Recall = N CP N PF
where N PF represents the number of key points of the template point cloud. The value of the threshold th used for performing feature matching in Section 2.5 varies from 0 to 1 to calculate the precision and recall under each threshold and obtain the PRC.
(2) Root mean square error
Root mean square error (RMSE) is the error evaluation index commonly used in point cloud registration, which represents the average of the sum of squared distances between the corresponding points of the two point clouds. It is defined as:
RMSE = i = 1 m | | P i Q j | | m
where P i and Q j are the corresponding points, and m is the number of the corresponding point pairs. The smaller the value of RMSE, the better the fine registration result.

3.2. Experiment Preparation

In order to verify the effectiveness and feasibility of the proposed pallet detection method, a widely used industrial camera called the Percipio FM851-E2 3D vision sensor is adopted to acquire point cloud data for comparative analysis of the results. The vision sensor shown in Figure 8 consists of an RGB camera and a depth sensor which is composed of an infrared camera and structured light projector. Its length, width and height are 124.0 mm, 86.8 mm and 28.6 mm, respectively. The RGB camera captures the RGB image with a resolution of 1280 × 960 and the depth sensor captures the depth image with a resolution of 320 × 240. Percipio FM851-E2 3D vision sensor active binocular vision technology is used for measuring the distance, and its operative range is from 0.7 m to 6.0 m.
The Percipio FM851-E2 3D vision sensor is mounted on the top of the carriage of a real automated guided vehicle, which means the camera will move along with the forks, as shown in Figure 9. Given that the length of the fork is 1150 mm, the distance between the top of the fork and the front face of the pallet is set to 500 mm so that the automated guided vehicle is able to adjust its position. Meanwhile, it is necessary to ensure that the fork is perpendicular to the front face of the pallet and that the center of the sensor is in line with the center of the pallet. The specific placement of the pallet is shown in Figure 9 and Figure 10. Considering the effect of illumination on point cloud data acquired by the sensor, all experiments are carried out under normal daytime illumination. In this case, the size of the pallet is 1200 mm × 1000 mm × 150 mm, and it is extracted from the scene point cloud and considered as the template point cloud.

3.3. Case Study I

3.3.1. Implementation Process

For the pallets on the ground, the color image of the scene is acquired by the Percipio FM851-E2 3D vision sensor at the same distance of 500 mm, which is shown in Figure 11. Figure 12 shows the pallet template point cloud and the scene point cloud. Then, the outliers of the scene point cloud are eliminated. The normal of the ground and the wall are [0, 1, 0] and [0, 0, 1], respectively. Set distance threshold D ε = 0.02   m and angle threshold β ε = 5 ° ; the plane segmentation is performed for the scene point cloud after removing outliers, and the result is shown in Figure 13. The ISS algorithm is used to extract key points with the search radius of 0.013 m and thresholds κ 1 = 0.6 , κ 2 = 0.75 , which can guarantee the efficiency and accuracy of the method. The number of points in the pallet template point cloud decreased from 2661 to 492, and the number of points in the scene point cloud decreased from 41,351 to 576, as shown in Figure 14, and the red points in Figure 14 represent the key points. The point cloud image shown below contains the RBG information of the point cloud and therefore has different colors.
It is necessary to determine the adaptive neighborhood radius for each point before calculating the ACFPFH feature descriptor. Given that the interval between two sampling points of point cloud data acquired by Percipio FM851-E2 3D vision sensor is 0.007 m, set the radius range r min = 0.007 m , r max = 0.015 m , r m i d = 0.011 m and r Δ = 0.0005 m . The adaptive optimal neighborhood radius of each point is obtained with the minimum criterion of neighborhood information entropy function. The adaptive optimal neighborhood radius distribution of pallet template point cloud and scene point cloud is shown in Figure 15. The horizontal axis represents the value of the neighborhood radius, and the vertical axis represents the number of points corresponding to each neighborhood radius. Among the 2661 points in the pallet template point cloud, there are 855 points with an optimal neighborhood radius of 0.007 m. Among the 41,351 points in the ground scene point cloud, there are 15,421 points with an optimal neighborhood radius of 0.007 m. It meets where the optimal neighborhood radius of points is concentrated at the given minimum neighborhood radius, which aids in improving the efficiency.
Extract the HSV color components of the key points of the pallet template point cloud and the scene point cloud, and calculate the geometric feature based on the adaptive optimal neighborhood radius. The ACFPFH feature descriptor is obtained by superimposing the color and geometric features. The feature matching is completed with the distance ratio threshold of t h = 0.75 . The initial matching result is shown in Figure 16a. The green line connects the corresponding points between the pallet template point cloud and the scene point cloud. Obviously, there are some wrong matching point pairs. The wrong matching point pairs are eliminated by using the wrong matching point pairs elimination algorithm based on the ACTN, and the result is shown in Figure 16b. The RANSAC algorithm is used for coarse registration to calculate the rough transformation matrix, and the ICP algorithm is used to obtain a final transformation matrix and complete fine registration. The parameters of the final transformation matrix are as follows:
R = [ 0.9889 0.0067 0.1487 0.0099 0.9997 0.0206 0.1485 0.0219 0.9887 ] T = [ 0.4709 0.0312 0.0723 ]

3.3.2. Performance Evaluation

The PRC is used to evaluate the descriptiveness of a feature descriptor. The ACFPFH feature descriptor is compared with the classical feature descriptors including FPFH, CFPFH and Signature of Histogram of Orientation (SHOT) with the fixed neighborhood radius.
The set th = { 0.2 , 0.4 , 0.6 , 0.75 , 0.85 , 0.925 , 0.95 , 9.975 , 1.0 } is considered as the selected distance ratio threshold set of the feature matching stage, and the PRC corresponding to different feature descriptors is obtained, as shown in Figure 17. Take th = 0.75 to compare the accuracy of different feature descriptors, as shown in Table 2. Table 3 lists the time required for feature extraction of the scene point cloud of different feature descriptors, and the bold characters are the experimental results of the proposed method.
Traditional feature descriptors such as SHOT and FPFH only describe the geometric feature of the pallet and ignore the color information, so the precision is lower. The CFPFH feature descriptor considers the HSV color information, which improves the precision. The neighborhood radiuses of the above three feature descriptors are obtained by complex and inefficient manual debugging methods, which are not suitable for all points in the point cloud. A large neighborhood radius leads to too many key points in the neighborhood, which reduces the speed of feature extraction. The ACFPFH feature descriptor not only contains color information but also adaptively selects the optimal neighborhood radius for each key point, so that it performs better in terms of effectiveness and precision.
It is well known that the closer the curve is to the upper right, the better the performance of the feature descriptor in the PRC graph. It can be seen from Figure 17 that comparing with SHOT, FPFH and CFPFH feature descriptors with fixed radiuses, the ACFPFH feature descriptor has the best performance. It can be seen from Table 2 and Table 3 that when th = 0.75, compared with the SHOT feature descriptor with a neighborhood radius of 0.011 m, the precision is improved by 29.40%, and the time required for feature extraction is reduced by 14.57%. Compared with the FPFH feature descriptor with a neighborhood radius of 0.011 m, the precision is improved by 39.10%, and the time required for feature extraction is reduced by 11.03%. Compared with the CFPFH feature descriptor with a neighborhood radius of 0.011 m, the precision is improved by 16.68%, and the feature extraction time is reduced by 18.87%.
The RMSE and runtime are used to evaluate the performance of the registration algorithms. Popular algorithms including ICP, SHOT + ICP, FPFH + ICP and CFPFH + ICP are selected to compare with the proposed method in this paper. The number of iterations, the RMSE and the runtime of the above methods are detailed in Table 4. The initial position relationship of the pallet template point cloud and the scene point cloud and the registration results of different methods are shown in Figure 18, and the red points are the template point cloud.
The following conclusions can be drawn from Table 4 and Figure 18: The traditional ICP algorithm has a large registration error due to the large initial pose difference. It takes 27.256 s to realize registration, which cannot meet the real-time requirements in intelligent manufacturing systems. The modified ICP registration methods such as SHOT + ICP, FPFH + ICP and CFPFH + ICP perform coarse registration, providing a better initial position for the fine registration by the ICP algorithm. Compared with the traditional ICP algorithm, the RMSE and runtime are reduced. However, the feature descriptors used by the above methods lack neighborhood selection criteria, which leads to an increase in the overall registration runtime. The proposed method has minimal registration error and the least runtime, which shows higher efficiency and proves that the proposed method has a more significant improvement than other methods. Furthermore, the precision and efficiency of the proposed method also meet the production requirement in intelligent manufacturing systems.

3.4. Case Study II

Shelves are widely used in intelligent manufacturing systems, which can improve the utilization rate of warehouse space and realize the rational allocation of resources while ensuring the quality of goods. Hence, it is necessary to complete the pallet detection of the shelf scene. The color image and the point cloud of the shelf scene are acquired with the same distance from the ground scene, as shown in Figure 19. The same pallet template point cloud is used to perform pallet detection of the shelf scene, and the parameters are consistent with the ground scene in Case Study I. After extracting the key points, the number of points in the scene point cloud decreased from 30,469 to 658. The adaptive optimal neighborhood radius distribution of the scene point cloud in the shelf scene is shown in Figure 20. Figure 21 and Table 5 show the registration result.
Compared with the traditional ICP algorithm, the RMSE of the proposed method is greatly reduced, and the runtime is reduced from 29.523 s to 0.989 s, which proves that the efficiency and accuracy have been greatly improved. Compared with other modified ICP registration methods, the RMSE of the proposed method is the smallest, and the runtime is the shortest. In summary, the proposed method still achieves optimal performance in the shelf scene, which shows its effectiveness in different scenes. Furthermore, the above case studies demonstrate that the proposed method can be well applied in intelligent manufacturing systems to realize accurate and efficient pallet detection. In addition, feature descriptors can often determine the final performance in the process of point cloud registration. Combining a good feature descriptor with a good matching strategy would improve the efficiency of point cloud registration.

4. Conclusions

A novel pallet detection method for automated guided vehicles based on point cloud data is proposed in this paper. The contributions of this paper can be concluded as follows:
  • A novel pallet detection method for automated guided vehicles based on point cloud data is proposed, which can be used for automated guided vehicles to perform automated and effective pallet handling, thereby promoting the transformation and upgrading of the manufacturing industry.
  • A new Adaptive Color Fast Point Feature Histogram (ACFPFH) feature descriptor has been built for the description of pallet features, which overcomes shortcomings such as low efficiency, time-consumption, poor robustness, and random parameter selection in feature description.
  • A new surface matching method called the Bidirectional Nearest Neighbor Distance Ratio-Approximate Congruent Triangle Neighborhood (BNNDR-ACTN) is proposed, which transforms the point-to-point matching problem into the neighborhood matching problem and can obtain more feature information and improve the detection accuracy.
Due to the measurement accuracy of the 3D vision sensor being easily affected by environmental factors such as illumination and obstacles, a more robust and efficient pallet detection method will be researched, which is suitable for more complex scenarios.

Author Contributions

Conceptualization, J.L.; methodology, Y.S.; software, Z.F.; validation, Z.F.; formal analysis, Z.C.; investigation, Y.S.; resources, B.Z.; data curation, M.Z.; writing—original draft preparation, Y.S.; writing—review and editing, Y.S. and Z.F.; visualization, Z.C.; supervision, J.L.; project administration, B.Z.; funding acquisition, Y.S. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Zhejiang Provincial Natural Science Foundation, grant number LQ22E050017, the Zhejiang Science and Technology Plan Project, grant number 2018C01003, the China Postdoctoral Science Foundation, grant number 2021M702894 and the Zhejiang Provincial Postdoctoral Science Foundation, grant number ZJ2021119. All experiments were performed at Noblelift Intelligent Equipment Co., Ltd., we are grateful to Noblelift engineers for their experimental support.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ambroz, M. Raspberry Pi as a low-cost data acquisition system for human powered vehicles. Measurement 2017, 100, 7–18. [Google Scholar] [CrossRef]
  2. Li, Z.; Barenji, A.V.; Jiang, J.; Zhong, R.Y.; Xu, G. A mechanism for scheduling multi robot intelligent warehouse system face with dynamic demand. J. Intell. Manuf. 2018, 31, 469–480. [Google Scholar] [CrossRef]
  3. Casado, F.; Lapido, Y.L.; Losada, D.P. Pose Estimation and Object Tracking Using 2D Images. Procedia Manuf. 2017, 11, 63–71. [Google Scholar] [CrossRef]
  4. Sriram, K.V.; Havaldar, R.H. Analytical review and study on object detection techniques in the image. Int. J. Modeling Simul. Sci. Comput. 2021, 12, 31–50. [Google Scholar] [CrossRef]
  5. Wang, Q.; Tan, Y.; Mei, Z. Computational Methods of Acquisition and Processing of 3D Point Cloud Data for Construction Applications. Arch. Comput. Methods Eng. 2019, 27, 479–499. [Google Scholar] [CrossRef]
  6. Camurri, M.; Vezzani, R.; Cucchiara, R. 3D Hough transform for sphere recognition on point clouds: A systematic study and a new method proposal. Mach. Vis. Appl. 2014, 25, 1877–1891. [Google Scholar] [CrossRef]
  7. García-Pulido, J.A.; Pajares, G.; Dormido, S.; de la Cruz, J.M. Recognition of a landing platform for unmanned aerial vehicles by using computer vision-based techniques. Expert Syst. Appl. 2017, 76, 152–165. [Google Scholar] [CrossRef]
  8. Seidenari, L.; Serra, G.; Bagdanov, A.D.; Del Bimbo, A. Local Pyramidal Descriptors for Image Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1033–1040. [Google Scholar] [CrossRef] [Green Version]
  9. Chen, J.; Chen, L. Multi-Dimensional Color Image Recognition and Mining Based on Feature Mining Algorithm. Autom. Control. Comput. Sci. 2021, 55, 195–201. [Google Scholar] [CrossRef]
  10. Joshi, K.D.; Chauhan, V.; Surgenor, B. A flexible machine vision system for small part inspection based on a hybrid SVM/ANN approach. J. Intell. Manuf. 2018, 31, 103–125. [Google Scholar] [CrossRef]
  11. Bastian, B.T.; Charangatt Victor, J. Detection and pose estimation of auto-rickshaws from traffic images. Mach. Vis. Appl. 2020, 31, 54. [Google Scholar] [CrossRef]
  12. Li, J.; Kang, J.; Chen, Z.; Cui, F.; Fan, Z. A Workpiece Localization Method for Robotic De-Palletizing Based on Region Growing and PPHT. IEEE Access 2020, 8, 6365–6376. [Google Scholar] [CrossRef]
  13. Syu, J.-L.; Li, H.-T.; Chiang, J.S. A computer vision assisted system for autonomous forklift vehicles in real factory environment. Multimed. Tools Appl. 2016, 76, 387–407. [Google Scholar] [CrossRef]
  14. Li, T.; Huang, B.; Li, C. Application of convolution neural network object detection algorithm in logistics warehouse. J. Eng. 2019, 2019, 9053–9058. [Google Scholar] [CrossRef]
  15. Shao, Y.; Wang, K.; Du, S.; Xi, L. High definition metrology enabled three dimensional discontinuous surface filtering by extended tetrolet transform. J. Manuf. Syst. 2018, 49, 75–92. [Google Scholar] [CrossRef]
  16. Huang, D.; Du, S.; Li, G.; Zhao, C.; Deng, Y. Detection and monitoring of defects on three-dimensional curved surfaces based on high-density point cloud data. Precis. Eng. 2018, 53, 79–95. [Google Scholar] [CrossRef]
  17. Jia, S.; Deng, Y.; Lv, J.; Du, S.; Xie, Z. Joint distribution adaptation with diverse feature aggregation: A new transfer learning framework for bearing diagnosis across different machines. Measurement 2022, 187, 110332. [Google Scholar] [CrossRef]
  18. He, K.; Zhang, M.; Zuo, L.; Alhwiti, T.; Megahed, F.M. Enhancing the monitoring of 3D scanned manufactured parts through projections and spatiotemporal control charts. J. Intell. Manuf. 2014, 28, 899–911. [Google Scholar] [CrossRef]
  19. Seelinger, M.; Yoder, J.D. Automatic pallet engagment by a vision guided forklift. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Barcelona, Spain, 18–22 April 2005; pp. 4068–4073. [Google Scholar]
  20. Lecking, D.; Wulf, O.; Wagner, B. Variable pallet pick-up for automatic guided vehicles in industrial environments. In Proceedings of the 11th IEEE International Conference on Emerging Technologies and Factory Automation, Prague, Czech, 20–22 September 2006; pp. 1169–1174. [Google Scholar]
  21. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. 3D Object Recognition in Cluttered Scenes with Local Surface Features: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2270–2287. [Google Scholar] [CrossRef]
  22. Liu, Y.; Kong, D.; Zhao, D.; Gong, X.; Han, G. A Point Cloud Registration Algorithm Based on Feature Extraction and Matching. Math. Probl. Eng. 2018, 2018, 7352691. [Google Scholar] [CrossRef]
  23. Tao, Y.; Zhou, J. Automatic apple recognition based on the fusion of color and 3D feature for robotic fruit picking. Comput. Electron. Agric. 2017, 142, 388–396. [Google Scholar] [CrossRef]
  24. Liu, J.; Bai, D.; Chen, L. 3-D point cloud registration algorithm based on greedy projection triangulation. Appl. Sci. 2018, 8, 1776. [Google Scholar] [CrossRef] [Green Version]
  25. Li, P.; Wang, J.; Zhao, Y.; Wang, Y.; Yao, Y. Improved algorithm for point cloud registration based on fast point feature histograms. J. Appl. Remote Sens. 2016, 10, 045024. [Google Scholar] [CrossRef]
  26. Kitt, B.; Geiger, A.; Lategahn, H. Visual Odometry based on Stereo Image Sequences with RANSAC-based Outlier Rejection Scheme. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), University of California, San Diego (UCSD), San Diego, CA, USA, 21–24 June 2010; pp. 486–492. [Google Scholar]
  27. Xu, G.; Pang, Y.; Bai, Z.; Wang, Y.; Lu, Z. A Fast Point Clouds Registration Algorithm for Laser Scanners. Appl. Sci. 2021, 11, 3426. [Google Scholar] [CrossRef]
  28. Napoli, A.; Glass, S.; Ward, C.; Tucker, C.; Obeid, I. Performance analysis of a generalized motion capture system using microsoft kinect 2.0. Biomed. Signal Proces. 2017, 38, 265–280. [Google Scholar] [CrossRef]
  29. Demantké, J.; Mallet, C.; David, N.; Vallet, B. Dimensionality based scale selection in 3D lidar point clouds. Int. Arch. Photogramm. 2012, 38, 97–102. [Google Scholar] [CrossRef] [Green Version]
  30. Weinmann, M.; Jutzi, B.; Mallet, C. Semantic 3D scene interpretation: A framework combining optimal neighborhood size selection with relevant features. In Proceedings of the ISPRS annals of the photogrammetry, remote sensing and spatial information sciences, Zurich, Switzerland, 5–7 September 2014; pp. 181–188. [Google Scholar]
  31. Chernov, V.; Alander, J.; Bochko, V. Integer-based accurate conversion between RGB and HSV color spaces. Comput. Electr. Eng. 2015, 46, 328–337. [Google Scholar] [CrossRef]
Figure 1. The specific pallet detection scene for automated guided vehicles. (a) Storage shelves, (b) the production workshop, (c) drug transport and (d) blast furnace conditions.
Figure 1. The specific pallet detection scene for automated guided vehicles. (a) Storage shelves, (b) the production workshop, (c) drug transport and (d) blast furnace conditions.
Sensors 22 08019 g001
Figure 2. The framework of the proposed approach.
Figure 2. The framework of the proposed approach.
Sensors 22 08019 g002
Figure 3. The flowchart of the feature description.
Figure 3. The flowchart of the feature description.
Sensors 22 08019 g003
Figure 4. Neighborhood influence area of the point Pq for ACFPFH.
Figure 4. Neighborhood influence area of the point Pq for ACFPFH.
Sensors 22 08019 g004
Figure 5. Local coordinate system.
Figure 5. Local coordinate system.
Sensors 22 08019 g005
Figure 6. The ACFPFH of one point.
Figure 6. The ACFPFH of one point.
Sensors 22 08019 g006
Figure 7. The flowchart the proposed surface matching method.
Figure 7. The flowchart the proposed surface matching method.
Sensors 22 08019 g007
Figure 8. Structure of Percipio FM851-E2.
Figure 8. Structure of Percipio FM851-E2.
Sensors 22 08019 g008
Figure 9. The installation position of the Percipio FM851-E2.
Figure 9. The installation position of the Percipio FM851-E2.
Sensors 22 08019 g009
Figure 10. The pose relationship between the 3D vision sensor and the pallet.
Figure 10. The pose relationship between the 3D vision sensor and the pallet.
Sensors 22 08019 g010
Figure 11. Color image of the original ground scene.
Figure 11. Color image of the original ground scene.
Sensors 22 08019 g011
Figure 12. Pallet template point cloud and original ground scene point cloud. (a) Pallet template point cloud. (b) Original ground scene point cloud.
Figure 12. Pallet template point cloud and original ground scene point cloud. (a) Pallet template point cloud. (b) Original ground scene point cloud.
Sensors 22 08019 g012
Figure 13. Pallet template point cloud and original ground scene point cloud.
Figure 13. Pallet template point cloud and original ground scene point cloud.
Sensors 22 08019 g013
Figure 14. Key points of pallet template point cloud and ground scene point cloud. (a) Key points of pallet template point cloud. (b) Key points of ground scene point cloud.
Figure 14. Key points of pallet template point cloud and ground scene point cloud. (a) Key points of pallet template point cloud. (b) Key points of ground scene point cloud.
Sensors 22 08019 g014
Figure 15. Adaptive optimal neighborhood radius distribution of point cloud. (a) Adaptive optimal neighborhood radius distribution of template point cloud. (b) Adaptive optimal neighborhood radius distribution of ground scene point cloud.
Figure 15. Adaptive optimal neighborhood radius distribution of point cloud. (a) Adaptive optimal neighborhood radius distribution of template point cloud. (b) Adaptive optimal neighborhood radius distribution of ground scene point cloud.
Sensors 22 08019 g015
Figure 16. The result of feature matching. Schemes follow another format. (a) Feature matching. (b) Elimination of wrong matching point pairs. The green lines are the lines connecting the matching point pairs.
Figure 16. The result of feature matching. Schemes follow another format. (a) Feature matching. (b) Elimination of wrong matching point pairs. The green lines are the lines connecting the matching point pairs.
Sensors 22 08019 g016
Figure 17. PRC of different feature descriptors.
Figure 17. PRC of different feature descriptors.
Sensors 22 08019 g017
Figure 18. Point registration results based on different methods in the ground scene. (a) Initial position. (b) Traditional ICP. (c) The method based on SHOT + ICP. (d) The method based on FPFH + ICP. (e) The method based on CFPFH + ICP. (f) The proposed method.
Figure 18. Point registration results based on different methods in the ground scene. (a) Initial position. (b) Traditional ICP. (c) The method based on SHOT + ICP. (d) The method based on FPFH + ICP. (e) The method based on CFPFH + ICP. (f) The proposed method.
Sensors 22 08019 g018
Figure 19. Shelf scene. (a) Color image. (b) Point cloud.
Figure 19. Shelf scene. (a) Color image. (b) Point cloud.
Sensors 22 08019 g019
Figure 20. Adaptive optimal neighborhood radius distribution of point cloud in the shelf scene.
Figure 20. Adaptive optimal neighborhood radius distribution of point cloud in the shelf scene.
Sensors 22 08019 g020
Figure 21. The result of point registration in the shelf scene. (a) Initial pose relationship. (b) Fine registration. The red points represents the template point cloud.
Figure 21. The result of point registration in the shelf scene. (a) Initial pose relationship. (b) Fine registration. The red points represents the template point cloud.
Sensors 22 08019 g021
Table 1. Dimensional features judgment of local neighborhood of point cloud.
Table 1. Dimensional features judgment of local neighborhood of point cloud.
Eigenvalue RelationDimensionality FeatureEigenvalue Relation
λ 1 λ 2 λ 3 Linearity feature λ 1 λ 2 λ 3
λ 1 λ 2 λ 3 Planarity feature λ 1 λ 2 λ 3
λ 1 λ 2 λ 3 Scattering feature λ 1 λ 2 λ 3
Table 2. Precision and recall comparison of different feature descriptors.
Table 2. Precision and recall comparison of different feature descriptors.
NameFeature DimensionNeighborhood
Radius/m
RecallPrecisionAccuracy Comparison of ACFPFH with Other Feature Descriptors (%)
SHOT3520.0110.01930.248129.40
0.0130.02030.279820.38
FPFH330.0110.01830.2140
0.2712
39.10
0.0130.022422.82
CFPFH360.0110.02190.292816.68
0.0130.02640.32567.34
ACFPFH36Adaptive0.02690.3514/
Table 3. Feature extraction time comparison of different feature descriptors.
Table 3. Feature extraction time comparison of different feature descriptors.
NameFeature DimensionNeighborhood
Radius/m
Feature Extraction Time/sTime Comparison of ACFPFH with Other Feature Descriptors (%)
SHOT3520.0110.15114.57
0.0130.18530.27
FPFH330.0110.14511.03
25.86
0.0130.174
CFPFH360.0110.15918.87
0.0130.19333.16
ACFPFH36Adaptive0.129/
Table 4. RMSE and elapsed time comparison of different feature descriptors.
Table 4. RMSE and elapsed time comparison of different feature descriptors.
MethodThe Number of IterationsRMSEThe Runtime/s
Traditional ICP1130.04034427.256
SHOT + ICP820.0247910.986
FPFH + ICP240.0265890.948
CFPFH + ICP440.0215591.039
ACFPFH260.0092510.853
Table 5. RMSE and elapsed time comparison of different feature descriptors in the shelf scene.
Table 5. RMSE and elapsed time comparison of different feature descriptors in the shelf scene.
MethodThe Number of IterationsRMSEThe Runtime/s
Traditional ICP680.04155329.523
SHOT + ICP490.0259871.174
FPFH + ICP320.0267511.118
CFPFH + ICP360.0189541.326
ACFPFH230.0090320.989
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shao, Y.; Fan, Z.; Zhu, B.; Zhou, M.; Chen, Z.; Lu, J. A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data. Sensors 2022, 22, 8019. https://doi.org/10.3390/s22208019

AMA Style

Shao Y, Fan Z, Zhu B, Zhou M, Chen Z, Lu J. A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data. Sensors. 2022; 22(20):8019. https://doi.org/10.3390/s22208019

Chicago/Turabian Style

Shao, Yiping, Zhengshuai Fan, Baochang Zhu, Minlong Zhou, Zhihui Chen, and Jiansha Lu. 2022. "A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data" Sensors 22, no. 20: 8019. https://doi.org/10.3390/s22208019

APA Style

Shao, Y., Fan, Z., Zhu, B., Zhou, M., Chen, Z., & Lu, J. (2022). A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data. Sensors, 22(20), 8019. https://doi.org/10.3390/s22208019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop