Next Article in Journal
Carbohydrate Content Classification Using Postprandial Heart Rate Responses from Non-Invasive Wearables
Previous Article in Journal
Surface Deformation of Xiamen, China Measured by Time-Series InSAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Precision Positioning and Rotation Angle Estimation for a Target Pallet Based on BeiDou Navigation Satellite System and Vision

Qilu Aerospace Information Research Institute, Jinan 250100, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5330; https://doi.org/10.3390/s24165330
Submission received: 15 July 2024 / Revised: 8 August 2024 / Accepted: 8 August 2024 / Published: 17 August 2024

Abstract

:
In outdoor unmanned forklift unloading scenarios, pallet detection and localization face challenges posed by uncontrollable lighting conditions. Furthermore, the stacking and close arrangement of pallets also increase the difficulty of positioning a target pallet. To solve these problems, a method for high-precision positioning and rotation angle estimation for a target pallet using the BeiDou Navigation Satellite System (BDS) and vision is proposed. Deep dual-resolution networks (DDRNets) are used to segment the pallet from depth images and RGB images. Then, keypoints for calculating the position and rotation angle are extracted and further combined with the 3D point cloud data to achieve accurate pallet positioning. Constraining the pixel coordinates and depth coordinates of the center point of the pallet and setting the priority of the pallet according to the unloading direction allow the target pallet to be identified from multiple pallets. The positioning of the target pallet in the forklift navigation coordinate system is achieved by integrating BDS positioning data through coordinate transformation. This method is robust in response to lighting influences and can accurately locate the target pallet. The experimental results show that the pallet positioning error is less than 20 mm, and the rotation angle error is less than 0.37°, which meets the accuracy requirements for automated forklift operations.

1. Introduction

With the rapid development of modern logistics, automated handling equipment, especially unmanned forklifts, has been widely used, which can improve the efficiency of cargo unloading and reduce operating costs [1,2,3]. As one of the key technologies for autonomous operation of forklifts, pallet detection and localization have garnered increasing attention [4]. In outdoor unmanned unloading scenarios, the position of a pallet is highly uncertain and is influenced by uncontrollable lighting conditions, making accurate detection and localization of the pallet more challenging.
Researchers have carried out a lot of research work for pallet detection and localization. The mainstream methods can be categorized into three types: vision-based, point-cloud-based, or a combination of both. Early vision-based detection methods were based on the natural features of the pallet to design segmentation strategies to segment the pallet from the image background and combined with geometric computation methods to realize pallet localization [5,6,7]. However, these methods are susceptible to background interference, and their robustness in complex logistics scenarios is not ideal [8]. Another approach involves adding reference markers to the surface of the pallet to enhance its recognition features, thereby achieving more precise pallet identification and localization [9,10]. However, this method is susceptible to marker deterioration and is not suitable for logistics scenarios with high pallet circulation. In recent years, object detection algorithms based on deep learning have been used to detect pallets in 2D images. Compared with early detection methods based on traditional machine vision, they have improved the speed and accuracy of pallet detection, providing an efficient solution for pallet detection in indoor environments [11,12,13,14].
Point-cloud-based methods use plan segmentation to process the point cloud data, and then use template matching or point cloud registration to achieve pallet localization [15,16,17,18]. Compared with vision-based methods, the point-cloud-based methods have the advantage of resisting the effect of lighting. However, processing point cloud data usually requires a large amount of computing resources, and the quality of the point cloud data directly affects the accuracy of pallet localization, which puts more stringent requirements on the performance of sensors and computing units.
The methods based on vision and point clouds combine the advantages of both and can not only use deep learning algorithms to accurately identify pallets in 2D images, but also use point cloud processing methods to achieve precise positioning of pallets in 3D space [19,20,21,22]. These methods reduce the need for computational resources and achieve a balance between speed and accuracy. However, these methods rely on extracting the pallet from RGB images. When the camera cannot stably obtain valid RGB images due to the influence of uncontrollable outdoor lighting conditions, these methods become ineffective, making them unsuitable for outdoor scenarios.
Furthermore, most research focuses on the identification and positioning of single pallets [23], while ignoring the situation of stacked and closely arranged pallets in actual unloading scenarios. There is a lack of research on target pallet detection and localization.
To solve these problems, a method for high-precision positioning and rotation angle estimation for a target pallet using the BeiDou Navigation Satellite System (BDS) and vision is proposed. The main contributions of this method are as follows: (1) Deep dual-resolution networks (DDRNets) are used to segment the pallet in 2D images. Then the keypoints for calculating the position and rotation angle are extracted and further combined with the 3D point cloud data to achieve pallet positioning. This method takes into account the speed and accuracy of pallet positioning and meets the real-time requirements of unmanned forklifts. (2) The method achieves accurate positioning of the target pallet in the case of stacked and closely arranged pallets by constraining the pixel coordinates and depth coordinates of the center point of the pallet and setting the priority of the pallet according to the unloading direction. (3) This method uses depth images with strong light resistance as the processing object and can effectively cope with the impact of lighting changes. Moreover, this method can work normally without lighting conditions and is suitable for autonomous forklift operations in outdoor scenes.

2. Method

The flowchart of the proposed method in this article is shown in Figure 1. It consists of six phases: image acquisition and preprocessing, pallet segmentation, keypoint extraction and pallet localization, target pallet identification, BDS data acquisition, and coordinate transformation. Firstly, RGB images and depth images of the pallet are captured using an RGB-D camera at the forklift observation location. The Deep Dual-resolution Network-23-slim (DDRNet-23-slim) [24] is used to segment the pallet from 2D images. Then the pallet boundary straight lines are extracted using Progressive Probabilistic Hough Transform (PPHT) [25], and the pallet endpoints are obtained by clustering the intersections of these lines using the K-means clustering algorithm [26]. The intersection of the two diagonals and the correction point of the boundary center point on both sides are extracted as keypoints. Subsequently, the position and rotation angle of the pallet are calculated based on point cloud data indexing by keypoints. Constraining the pixel coordinates and depth coordinates of the center point of the pallet and setting the priority of the pallet according to the unloading direction allow the target pallet to be identified from many pallets. The position and yaw angle of the forklift are acquired by the dual-antenna BDS mobile receiver, and finally, the positioning of the target pallet in the forklift navigation coordinate system is achieved through coordinate transformation [27,28].

2.1. Image Acquisition and Preprocessing

After the flatbed truck stops along the stop line on the ground, the unmanned forklift receives the unloading task and arrives at the observation position on the side of the flatbed truck. Each pallet corresponds to an approximate observation position, which is set in advance based on the stop line. The observation position is shown in Figure 2, and D denotes the vertical distance between the observation position and the stop line. Affected by the camera’s field of view (FOV), ranging accuracy, and fork arm length, the value of D must meet the following conditions: (1) Even if the flatbed truck is parked deviated, the pallet must appear within the camera’s field of view. (2) The camera ranging accuracy decreases as the observation distance increases, so the observation position should not be too far from the stop line. (3) The observation position should not be too close to the stop line to avoid contact between the fork arm and the flatbed truck, which may lead to safety accidents.
The unmanned forklift is equipped with an RGB-D camera to capture RGB images and depth images of the pallet. The depth channel of RGB-D camera has a field of view similar to that of the color channel and is not affected by variant illumination. Even in dark environments, the RGB-D camera can stably acquire depth images. Figure 3 illustrates the RGB image and depth image provided by an RGB-D camera. In order to reduce the interference pixels of other objects in the depth image, the depth threshold D T is set with D as the reference. By eliminating depth data larger than D T , most non-target-object pixels are filtered out. In addition, for the purpose of reducing the interference caused by the jitter of the depth data at the edge of the pallet, Gaussian filtering on a large scale is used to smooth the depth image and reduce the noise.

2.2. Pallet Segmentation

There are many types of pallets used in the logistics industry. The circulation of pallets among warehouses is very strong, which leads to varying degrees of wear and deformation of pallets. Additionally, the pallets are tilted at small angles or partially obscured by plastic film and are affected by different conditions. It is difficult for traditional image processing methods to cope with these situations. On the contrary, image segmentation methods based on deep learning extract features through multi-layer neural networks, which can achieve automated target classification and are more suitable for the above situations.
For unmanned forklifts, it is not desirable to spend a lot of inference time to obtain segmentation results with high precision. Therefore, we apply the lightweight DDRNet-23-slim semantic segmentation model to the pallet segmentation. DDRNet-23-slim features deep dual-resolution networks as efficient backbones for real-time segmentation, providing a trade-off between accuracy and efficiency by scaling model width and depth [24]. Table 1 shows the architecture of DDRNet-23-slim.
Plastic pallets and wooden pallets are widely used in the logistics industry. Table 2 shows the specific information about the pallets used in this article.
In order to match the real environment, pallets under different conditions were collected in the dataset. In addition to pallets under different lighting conditions, pallets partially covered by plastic film, at different heights, or at small tilt angles were also collected, as shown in Figure 4.
The segmentation results are shown in Figure 5.

2.3. Keypoint Extraction and Pallet Localization

The ground where the forklift operates is flat, and pallets are placed flat on the flatbed truck. It is reasonable to assume that the pallet surface is parallel to the camera surface. In fact, the positioning of a pallet requires the determination of four degrees of freedom (DOFs), including 3D spatial coordinates and rotation angle. The pallets used in this paper have a common feature, that is, the front surface is planar and presents a regular rectangular shape, with four sides approximated as straight lines. The upper and lower edge contours of the pallet are longer and approximately horizontal, with an angle close to 0°, while the left and right edge contours are shorter and approximately vertical, with an angle close to 90°. We have designed a method to accurately extract keypoints on the front surface of the pallet from the 2D image that are used to calculate the pallet position and rotation angle. The geometric center of the pallet surface and the correction points of the boundary center points on both sides are used as keypoints, as shown in Figure 6.
We take a single pallet as an example to illustrate the keypoint extraction process. First, the contour detection method is performed to extract the pallet contour from the image segmentation result, followed by the extraction of the boundary straight lines using PPHT. The lines close to 0° and 90° are screened based on the horizontal angle threshold range T 1 ,   T 2 and the vertical angle threshold range T 3 ,   T 4 , respectively. The intersections of these two types of lines are then calculated and denoted as the set P . Furthermore, the intersection points in set P are classified into four categories using the K-means clustering algorithm, with the cluster centers representing the endpoints of the pallet boundaries. The intersection point of the diagonals serves as the keypoint for calculating the pallet position and is denoted as O . According to the four endpoints of the pallet, we calculate the pixel coordinates of the centers of the left and right boundaries, respectively. However, the coordinate data near the pallet boundary are not stable, and the two center points may not actually be located on the front surface of the pallet, which may result in an inability to accurately obtain the 3D coordinates of the pallet. Therefore, a straight line is computed through these two center points, labeled as L . Subsequently, the corrected coordinate points are obtained by moving the two center points along this straight line towards the inside of the pallet by a distance of d pixels. These two corrected points are used as keypoints for calculating the rotation angle of the pallet, labeled as E and F , respectively. The primary steps of keypoint extraction are shown in Figure 7 and Figure 8.
The alignment of the RGB image and the depth image is achieved through the camera’s intrinsic and extrinsic parameters, and the depth image is mapped to the point cloud data. This process is automatically completed inside the camera. The 3D coordinates of the keypoints under the camera coordinate system are determined by indexing the pixel coordinates in the point cloud data. After eliminating invalid values such as infinite values and null values, the average value of multiple valid measurements is taken, and the 3D coordinates of the keypoints are recorded as O ( X C , Y C , Z C ) , E ( X E , Y E , Z E ) , and F ( X F , Y F , Z F ) , respectively. A schematic diagram of rotation angle calculation is shown in Figure 9. The rotation angle of the pallet is calculated as shown in Formula (1).
α = a r c t a n Z F Z E X F X E / p i × 180 °
In Formula (1), α represents the rotation angle. Z F Z E represents the coordinate difference between E and F along the Z-axis direction. X F X E represents the coordinate difference between E and F along the X-axis direction.

2.4. Target Pallet Identification

In unloading scenarios, complex situations where pallets are stacked and closely arranged occur frequently. The unmanned forklift must be able to accurately search for the target pallet from multiple pallets presented in the camera’s field of view. To solve this problem, this paper proposes a method for searching for a target pallet. This method constrains the pixel coordinates and depth coordinates of the center points of all pallets and sets the priority of the pallets according to the unloading direction to locate the target pallet among many pallets.
Since goods are usually closely arranged, multiple pallets may be present in the camera’s field of view at the same time, as shown in Figure 10. The width pixels occupied by the complete pallet that appears in the camera’s field of view are counted and denoted as W P . Considering that the width pixel value occupied by the tilted pallet is often smaller than W P , we introduce the width pixel adjustment amount W T to extend the range of the width pixel coordinate constraint, denoted as W P / 2 W T ,   W W P / 2 + W T , as shown in Figure 11. Only pallets whose center point’s width pixel coordinates are within this range are considered as potential target pallets.
When there is a significant deviation between the unmanned forklift’s observation position and the pallet position, a greater number of complete pallets can be observed in the camera’s field of view, as shown in Figure 12. The center points of these pallets all meet the width pixel coordinate constraints. The priority of pallets is set according to the unloading direction to ensure that the forklift promptly handles the pallets at the highest priority position, as shown in Figure 13. When an unmanned forklift unloads goods from the rear of a flatbed truck, pallets on the left side of the camera’s field of view are given the highest priority. Conversely, when the unmanned forklift unloads goods from the front of the flatbed truck, those on the right side are prioritized. The opposite rule applies to pallets on the other side of the flatbed truck.
Unloading is carried out simultaneously by two unmanned forklifts from both sides of the flatbed truck, so the pallet on the opposite side may appear in the camera’s field of view, as shown in Figure 14. To minimize the impact of the rear pallet, with the forklift observation distance D as a reference, we introduce the depth distance adjustment amount D F and D B to expand the range of the depth distance coordinate constraint, denoted as D D F ,   D + D B , as shown in Figure 15. Only pallets whose center point’s depth coordinates are within this range are considered as potential target pallets.
For pallets stacked in multiple layers on the same side, the upper pallet has higher priority, and the lower pallet has lower priority, as shown in Figure 16. Therefore, the pallet on top is selected as the target pallet.
By setting constraints and priorities reasonably, unmanned forklifts can accurately find the target pallet in complex pallet arrangements. The positioning results of the target pallets are shown in Figure 17.

2.5. BDS Data Acquisition

A BDS reference station has been constructed in the business logistics park to provide high-precision positioning services for vehicle positioning devices. BDS-integrated inertial navigation outputs the position and yaw angle of the unmanned forklift at a high frequency. The dual-antenna BDS mobile receiver is installed on the roof along the longitudinal axis of the forklift as shown in Figure 18. The unmanned forklift relies on dual antennas to receive signals, so the forklift position coordinates actually refer to the longitude, latitude, and altitude coordinates of the main antenna in the China Geodetic Coordinate System 2000 (CGCS2000), denoted as ( B , L ,   H ) . The yaw angle is the angle between the longitudinal axis of the forklift and the true north direction, ranging from 0° to 360°, denoted as θ . In order to facilitate subsequent calculation and processing, we convert the yaw angle of the forklift to the angle between the longitudinal axis of the forklift and the true east direction and adjust the angle range between −180° and 180°.

2.6. Coordinate Transformation

To achieve the localization and navigation of the forklift, it is necessary to transform the camera coordinate system into the forklift navigation coordinate system. The camera coordinate system is defined in Figure 19. Since the forklift does not use height information for navigation, the height information is ignored, and the camera coordinate system is dimensionally reduced. Then, the coordinates of the target pallet in the camera plane coordinate system are simplified to ( X C , Z C ) .
The forklift navigation coordinate system is established as a plane rectangular coordinate system, with its Y-axis representing the true north direction and its X-axis representing the true east direction. The coordinates of the main antenna in the forklift navigation coordinate system ( X B , Y B ) can be obtained by transforming the coordinates of the main antenna in the CGCS2000 coordinate system  ( B , L ,   H ) . First, we transform the BeiDou positioning coordinates  ( B , L ,   H ) to the Earth-Centered, Earth-Fixed (ECEF) coordinates  ( X b , Y b , Z b ) . The transformation methods are shown in Formulas (2)–(4) [27,28].
X b = ( N + H ) c o s   B · c o s   L
Y b = ( N + H ) c o s   B · s i n   L
Z b = ( 1 e 2 ) N + H s i n   B
where N is the prime vertical radius of curvature and e is the first eccentricity of the Earth. The respective calculation methods are shown in Formulas (5) and (6).
N = a 1 e 2 s i n 2   B
e = 2 f f 2
Here, a is the Earth’s semi-major axis with a value of 6,378,137 m, and f is the Earth’s flattening with a value of 1/298.257222101.
Next, we transform the ECEF coordinates to the forklift navigation coordinate system. The origin of the forklift navigation coordinate system, which is the known reference point, has coordinates B O , L O , H O in the OGCS2000 coordinate system and ( X O , Y O , Z O ) in the ECEF coordinate system. The transformation methods are shown in Formulas (7) and (8) [28].
X B = s i n   L O X b X O + c o s   L O Y b Y O
Y B = s i n   B O · c o s   L O X b X O s i n   B O · s i n   L O Y b Y O + c o s   B O Z b Z O
For clarity, we denote the camera plane coordinate system as C C , the main antenna coordinate system as C G , and the forklift navigation coordinate system as C F . The spatial relationships among the three are illustrated in Figure 20.
Assuming that the origin of the camera plane coordinate system moves by X along the X-axis direction and Y along the Y-axis direction until it coincides with the antenna coordinate system, the coordinates of the target pallet in the antenna coordinate system are ( X C + X , Z C + Y ) . Furthermore, the forklift navigation coordinate system can be obtained from the antenna coordinate system through rotation and translation. The coordinates of the target pallet in the forklift navigation coordinate system can be calculated using the following formulas:
X = X B Z C + Y c o s   θ + ( X C + X ) s i n   θ
Y = Y B + Z C + Y s i n   θ + ( X C + X ) c o s   θ

3. Experiments

3.1. Forklift and Sensor Selection

The hardware used in this research included a forklift body, an RGB-D camera, and a dual-antenna BDS mobile receiver. The forklift body (model: TKA1535F905) was produced by Hefei Banyitong Science and Technology Developing Co., Ltd., located in Hefei, China, with a motion control precision of 3 cm. The RGB-D camera (model: TL460-S1-E1) was manufactured by Shanghai Tuyang Information Technology Co., Ltd., in Shanghai, China, with a range error of 9 mm at a distance of 2 m. The dual-antenna BDS mobile receiver (model: X1–5) was produced by Hunan Bynav Technology Co., Ltd., based in Changsha, China. This receiver can achieve a Real-Time Kinematic (RTK) positioning accuracy of 1.2 cm. Specifically, the system maintains its high precision when there are no obstacles within 15 degrees of the circumferential height angle above the receiver. During the loading process, ensuring that there are no obstructions above the forklift’s path is essential for the BDS receiver to reliably acquire satellite signals and achieve the stated positioning accuracy.

3.2. Experimental Site

The evaluation experiments of this method were conducted at Shunhe International Intelligent Logistics Park in Linyi City, Shandong Province, as shown in Figure 21. The flatbed truck was parked along the stop line and multiple pallets were placed on the flatbed truck. The unmanned forklift arrived at the observation position on the side of the flatbed truck to conduct pallet localization measurement and estimate rotation angle. The experiment completed the positioning measurement of pallets at three different positions. For each observation position, the position coordinates and rotation angle data of the pallet were recorded 20 times consecutively.

3.3. Pallet Image Segmentation Dataset

The dataset for pallet image segmentation consists of 1078 RGB images and 992 depth images. Eighty percent of the data was allocated to the training set, while the remaining twenty percent was used for validation, without establishing a separate test dataset. For RGB images with a resolution of 1184 × 720, the batch size was set to 32, while for depth images with a resolution of 640 × 480, the batch size was set to 16. The initial learning rate was set to 0.0003 and was dynamically adjusted using the cosine annealing function over 200 epochs. These parameter settings were chosen to balance training stability and convergence speed, ensuring effective segmentation performance for both RGB and depth images.

3.4. Parameter Setting

3.4.1. Preprocessing Parameter Selection

Considering the influence of factors such as the RGB-D camera’s field of view, measurement accuracy, and fork arm length, the vertical distance D between the observation position and the stop line was set to 2 m to ensure that the camera could acquire a complete image of the pallet. The depth threshold D T was set to 3 m to reduce interference of non-target-object pixels in the depth image by eliminating depth data greater than 3 m. A Gaussian filter was applied to smooth the image, with the window size set to 9 × 9.

3.4.2. Keypoint Extraction Parameter Settings

Taking into account that the pallet may not be strictly placed horizontally, the angle threshold range can be appropriately expanded to allow for a certain degree of tilt in the pallet. The horizontal angle threshold range [ T 1 , T 2 ] and vertical angle threshold range [ T 3 , T 4 ] were set to [−3°, 3°] and [85°, 95°], respectively. Horizontal line segments longer than 120 pixels and vertical line segments longer than 20 pixels were extracted from the RGB image. Horizontal line segments longer than 50 pixels and vertical line segments longer than 10 pixels were extracted from the depth image. According to the size of the pixels occupied by the pallet leg, the pixel distance moved by the midpoint of the short edge was appropriately adjusted to ensure that the corrected coordinate points were located on the pallet surface. In this experiment, d was set to 10 pixels.

3.4.3. Target Pallet Constraint Threshold Settings

The width pixels W P occupied by the pallet in the RGB image and depth image were approximately 460 pixels and 260 pixels, respectively. The width pixel adjustment amounts W T were set to 25 pixels and 20 pixels, so the width pixel coordinate constraints were [205, 979] and [115, 525], respectively. The depth distance adjustment amount D F and D B were set to 0.3 m and 0.5 m, respectively, so the depth distance coordinate constraint was [1.7, 2.5].

4. Evaluation

The absolute error ( A E ) is a commonly used error evaluation index and represents the absolute difference between the measured value and the reference value, as illustrated in Formula (11). A E i was calculated separately for multiple measurement data obtained at each observation position, and the maximum value A E m a x , the minimum value A E m i n , and the mean value M A E were further calculated, as shown in Formulas (12), (13), and (14), respectively.
A E i = r i R
A E m a x = m a x A E 1 , A E 2 , , A E n
A E m i n = m i n A E 1 , A E 2 , , A E n
M A E = 1 n i = 1 n ( A E 1 , A E 2 , , A E n )
For an assessment of the dispersion of the observed data, the standard deviation ( S T D ), presented in Formula (15), serves as an evaluation metric for quantifying the spread of measured values.
S T D = 1 n i = 1 n ( r i r ¯ ) 2
In the above formulas, R represents the reference value, r i represents the measurement value for the i-th time, n represents the number of measurements at the same observation position, and r ¯ represents the average value of the n measurements.
The results of target pallet segmentation and keypoint extraction based on RGB images and depth images are shown in Table 3 and Table 4, respectively.
The evaluation indices were calculated from the four degrees of freedom: X, Y, Z coordinates and rotation angle. The results are shown in Table 5 and Table 6.
Judging from the results, the A E of the localization results based on RGB images in the X, Y, and Z directions is less than 18 mm, with an S T D of less than 2.2 mm. The A E of the rotation angles is less than 0.34°, with an S T D of less than 0.20°, which indicates that the prediction results have a low degree of dispersion. In comparison, the localization results based on depth images show A E of less than 20 mm in the X, Y, and Z directions, with an S T D of less than 2.1 mm. The A E of the rotation angles is less than 0.37°, with an S T D of less than 0.22°, also demonstrating a low degree of dispersion. However, the results based on depth images exhibit a slight decrease in pallet localization accuracy. This difference can be attributed to the different resolutions of the two types of images, as well as variations in texture and color information. Specifically, the resolution of the depth images is 640 × 480, whereas the resolution of the RGB images is 1184 × 720. The lower resolution of the depth images results in less detailed information. Additionally, RGB images provide richer texture and color information, while depth images lack these detailed data, thereby affecting the accuracy of pallet detection and localization. Nevertheless, both methods provide accurate estimates for the rotation angles, which is related to their use of relative distances in calculations. The automatic unloading test conducted at the logistics site demonstrated that the accuracy meets the requirements.

5. Conclusions

This study aims to address the challenges of locating target pallets in outdoor unmanned forklift unloading scenarios, particularly due to uncontrollable lighting conditions and the close stacking of pallets. We propose a method for high-precision positioning and rotation angle estimation for a target pallet, utilizing the BDS and vision. This method has been successfully implemented in real unloading scenarios. The experimental results indicate that the pallet positioning error is less than 20 mm, and the rotation angle error is within 0.37°, thus meeting the accuracy requirements for automated forklift operations. Looking ahead, we plan to integrate depth and color channels to enhance pallet recognition. This approach aims to leverage the rich texture and color information available in RGB images and the strong resistance to lighting variations offered by depth images, ultimately improving both the accuracy and stability of pallet positioning. In particular, this integration seeks to address the specific challenge of recognizing pallets when two pallets are placed closely together with nearly aligned front faces. Additionally, some parameters related to keypoint extraction and target pallet constraints are still manually adjusted. Therefore, we intend to design and implement adaptive adjustment mechanisms for these parameters.

6. Patents

This paper has resulted in a patent with application number CN202310770170.7, titled “Method and Device for Segmentation and Positioning of Forklift Pallets, and Intelligent Forklift”.

Author Contributions

Conceptualization, D.M., Y.R. and X.Y. (Xinli Yu); methodology, D.M., X.Y., (Xinli Yu) and W.W.; software, D.M.; validation, W.W.; formal analysis, X.Y. (Xinli Yu) and W.W.; investigation, D.M.; resources, Y.R. and J.M.; data curation, D.M.; writing—original draft preparation, D.M.; writing—review and editing, D.M. and X.Y. (Xinli Yu); visualization, D.M.; supervision, X.Y. (Xiaoxv Yin); project administration, X.Y. (Xiaoxv Yin); funding acquisition, Y.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2021YFB1407003, and the Key Technology Research and Development Program of Shan-dong Province, grant number 2021SFGC0401.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Embargo on data due to commercial restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hanson, R.; Agrawal, T.; Browne, M.; Johansson, M.; Andersson, D.; Stefansson, G.; Katsela, K. Challenges and requirements in the introduction of automated loading and unloading. In Proceedings of the 11th Swedish Transportation Research Conference, Lund, Sweden, 18–19 October 2022. [Google Scholar]
  2. Carlan, V.; Ceulemans, D.; van Hassel, E.; Derammelaere, S.; Vanelslander, T. Automation in cargo loading/unloading processes: Do unmanned loading technologies bring benefits when both purchase and operational cost are considered? J. Shipp. Trade 2023, 8, 20. [Google Scholar] [CrossRef]
  3. Syu, J.L.; Li, H.T.; Chiang, J.S.; Hsia, C.H.; Wu, P.H.; Hsieh, C.F.; Li, S.A. A computer vision assisted system for autonomous forklift vehicles in real factory environment. Multimed. Tools Appl. 2017, 76, 18387–18407. [Google Scholar] [CrossRef]
  4. Wu, W.H.; Yang, M.; Wang, B.; Wang, C.X. Pallet detection based on contour matching for warehouse robots. J. Shanghai Jiaotong Univ. 2019, 53, 197–202. [Google Scholar] [CrossRef]
  5. Cui, G.Z.; Lu, L.S.; He, Z.D.; Yao, L.N.; Yang, C.X.; Huang, B.Y.; Hu, Z.H. A robust autonomous mobile forklift pallet recognition. In Proceedings of the 2010 2nd International Asia Conference on Informatics in Control, Automation and Robotics (CAR 2010), Wuhan, China, 6–7 March 2010. [Google Scholar] [CrossRef]
  6. Byun, S.; Kim, M. Real-time positioning and orienting of pallets based on monocular vision. In Proceedings of the 2008 20th IEEE International Conference on Tools with Artificial Intelligence, Dayton, OH, USA, 3–5 November 2008. [Google Scholar] [CrossRef]
  7. Chen, G.; Peng, R.; Wang, Z.; Zhao, W. Pallet recognition and localization method for vision guided forklift. In Proceedings of the 2012 8th International Conference on Wireless Communications, Networking and Mobile Computing, Shanghai, China, 21–23 September 2012. [Google Scholar] [CrossRef]
  8. Jia, F.; Tao, Z.; Wang, F. Wooden pallet image segmentation based on Otsu and marker watershed. In Proceedings of the 2021 International Conference on Computer, Communication, Control, Automation and Robotics (CCCAR 2021), Shanghai, China, 29–30 March 2021. [Google Scholar] [CrossRef]
  9. Seelinger, M.; Yoder, J.D. Automatic visual guidance of a forklift engaging a pallet. Robot. Auton. Syst. 2006, 54, 1026–1038. [Google Scholar] [CrossRef]
  10. Kesuma, E.S.; Rusmin, P.H.; Maharani, D.A. Pallet detection and distance estimation with YOLO and fiducial marker algorithm in industrial forklift robot. In Proceedings of the 2023 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Bali, Indonesia, 20–23 February 2023. [Google Scholar] [CrossRef]
  11. Li, T.; Huang, B.; Li, C.; Huang, M. Application of convolution neural network object detection algorithm in logistics warehouse. J. Eng. 2019, 2019, 9053–9058. [Google Scholar] [CrossRef]
  12. Zaccaria, M.; Monica, R.; Aleotti, J. A comparison of deep learning models for pallet detection in industrial warehouses. In Proceedings of the 2020 IEEE 16th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 3–5 September 2020. [Google Scholar] [CrossRef]
  13. Kirci, S.; Atali, G. EuroPallet Detection with RGB-D Camera Based on Deep Learning. Eur. J. Res. Dev. 2023, 3, 55–65. [Google Scholar] [CrossRef]
  14. Mok, C.; Baek, I.; Cho, Y.S.; Kim, Y.; Kim, S.B. Pallet Recognition with Multi-Task Learning for Automated Guided Vehicles. Appl. Sci. 2021, 11, 11808. [Google Scholar] [CrossRef]
  15. Molter, B.; Fottner, J. Real-time pallet localization with 3d camera technology for forklifts in logistic environments. In Proceedings of the 2018 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Singapore, 31 July–2 August 2018. [Google Scholar] [CrossRef]
  16. Xiao, J.; Lu, H.; Zhang, L.; Zhang, J. Pallet recognition and localization using an rgb-d camera. Int. J. Adv. Robot. Syst. 2017, 14, 1729881417737799. [Google Scholar] [CrossRef]
  17. Shao, Y.; Fan, Z.; Zhu, B.; Zhou, M.; Chen, Z.; Lu, J. A Novel Pallet Detection Method for Automated Guided Vehicles Based on Point Cloud Data. Sensors 2022, 22, 8019. [Google Scholar] [CrossRef] [PubMed]
  18. Shao, Y.; Fan, Z.; Zhu, B.; Lu, J.; Lang, Y. A Point Cloud Data-Driven Pallet Pose Estimation Method Using an Active Binocular Vision Sensor. Sensors 2023, 23, 1217. [Google Scholar] [CrossRef] [PubMed]
  19. Li, Y.Y.; Chen, X.H.; Ding, G.Y.; Wang, S.; Xu, W.C.; Sun, B.B.; Song, Q. Pallet detection and localization with RGB image and depth data using deep learning techniques. In Proceedings of the 2021 6th International Conference on Automation, Control and Robotics Engineering (CACRE), Dalian, China, 15–17 July 2021. [Google Scholar] [CrossRef]
  20. Iinuma, R.; Kojima, Y.; Onoyama, H.; Fukao, T.; Hattori, S.; Nonogaki, Y. Pallet handling system with an autonomous forklift for outdoor fields. J. Robot. Mechatron. 2020, 32, 1071–1079. [Google Scholar] [CrossRef]
  21. Iinuma, R.; Hori, Y.; Onoyama, H.; Fukao, T.; Kubo, Y. Pallet detection and estimation for fork insertion with rgb-d camera. In Proceedings of the 2021 IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 8–11 August 2021. [Google Scholar] [CrossRef]
  22. Rocha, E.D.S.; Chevtchenko, S.F.; Cambuim, L.F.; Macieira, R.M. Optimized Pallet Localization Using RGB-D Camera and Deep Learning Models. In Proceedings of the 2023 IEEE 19th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 26–28 October 2023. [Google Scholar] [CrossRef]
  23. Zhao, J.; Li, B.; Wei, X.; Lu, H.; Lü, E.; Zhou, X. Recognition and Location Algorithm for Pallets in Warehouses Using RGB-D Sensor. Appl. Sci. 2022, 12, 10331. [Google Scholar] [CrossRef]
  24. Pan, H.; Hong, Y.; Sun, W.; Jia, Y. Deep dual-resolution networks for real-time and accurate semantic segmentation of traffic scenes. IEEE Trans. Intell. Transp. Syst. 2022, 24, 3448–3460. [Google Scholar] [CrossRef]
  25. Galamhos, C.; Matas, J.; Kittler, J. Progressive probabilistic Hough transform for line detection. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), Fort Collins, CO, USA, 23–25 June 1999. [Google Scholar] [CrossRef]
  26. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 21 June–18 July 1965 and 27 December 1965–7 January 1966. [Google Scholar]
  27. Zhang, P.; Xu, C.; Hu, C.; Chen, Y. Coordinate transformations in satellite navigation systems. In Proceedings of the 2011 International Conference on Electronic Engineering, Communication and Management (EECM 2011), Beijing, China, 24–25 December 2011. [Google Scholar] [CrossRef]
  28. Kuna, D.; Santhosh, N.; Perumalla, N.K. Preliminary analysis of standalone Galileo and NavIC in the context of positioning performance for low latitude region. Procedia Comput. Sci. 2020, 171, 225–234. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the overall methodology.
Figure 1. Flowchart of the overall methodology.
Sensors 24 05330 g001
Figure 2. Schematic diagram of observation position.
Figure 2. Schematic diagram of observation position.
Sensors 24 05330 g002
Figure 3. RGB images and depth images of pallets under different lighting conditions. (a) Pallets under strong light condition; (b) pallets under no light condition; (c) pallets under good light condition.
Figure 3. RGB images and depth images of pallets under different lighting conditions. (a) Pallets under strong light condition; (b) pallets under no light condition; (c) pallets under good light condition.
Sensors 24 05330 g003aSensors 24 05330 g003b
Figure 4. RGB images and depth images of pallets under different conditions. (a) Pallets at different heights; (b) pallets partially covered by plastic film; (c) pallets at small tilt angles.
Figure 4. RGB images and depth images of pallets under different conditions. (a) Pallets at different heights; (b) pallets partially covered by plastic film; (c) pallets at small tilt angles.
Sensors 24 05330 g004
Figure 5. Segmentation results of pallets under different conditions based on RGB and depth images. (a) Pallets under strong light condition; (b) pallets under no light condition; (c) pallets under good light condition; (d) pallets at different heights; (e) pallets partially covered by plastic film; (f) pallets at small tilt angles.
Figure 5. Segmentation results of pallets under different conditions based on RGB and depth images. (a) Pallets under strong light condition; (b) pallets under no light condition; (c) pallets under good light condition; (d) pallets at different heights; (e) pallets partially covered by plastic film; (f) pallets at small tilt angles.
Sensors 24 05330 g005aSensors 24 05330 g005b
Figure 6. Schematic diagram of keypoint extraction.
Figure 6. Schematic diagram of keypoint extraction.
Sensors 24 05330 g006
Figure 7. Graphic presentation of primary steps of keypoint extraction based on RGB image. (a) Segmentation image; (b) contour detection; (c) line extraction; (d) intersection clustering; (e) keypoint extraction; (f) overlay image.
Figure 7. Graphic presentation of primary steps of keypoint extraction based on RGB image. (a) Segmentation image; (b) contour detection; (c) line extraction; (d) intersection clustering; (e) keypoint extraction; (f) overlay image.
Sensors 24 05330 g007
Figure 8. Graphic presentation of primary steps of keypoint extraction based on depth image. (a) Segmentation image; (b) contour detection; (c) line extraction; (d) intersection clustering; (e) keypoint extraction; (f) overlay image.
Figure 8. Graphic presentation of primary steps of keypoint extraction based on depth image. (a) Segmentation image; (b) contour detection; (c) line extraction; (d) intersection clustering; (e) keypoint extraction; (f) overlay image.
Sensors 24 05330 g008
Figure 9. Schematic diagram of pallet rotation angle calculation.
Figure 9. Schematic diagram of pallet rotation angle calculation.
Sensors 24 05330 g009
Figure 10. Multiple pallets appear in the camera’s field of view due to the close arrangement of goods. (a) Two columns of pallets are displayed, with only the column in the middle being complete. The column on the right side is incomplete and only partially visible, with no pallets on the left side. (b) Three columns of pallets are displayed, with only the column in the middle being complete. The columns on both the left and right sides are incomplete and only partially visible. (c) Two columns of pallets are displayed, with only the middle column being complete. The column on the right is incomplete and only partially visible, with no pallets on the right side.
Figure 10. Multiple pallets appear in the camera’s field of view due to the close arrangement of goods. (a) Two columns of pallets are displayed, with only the column in the middle being complete. The column on the right side is incomplete and only partially visible, with no pallets on the left side. (b) Three columns of pallets are displayed, with only the column in the middle being complete. The columns on both the left and right sides are incomplete and only partially visible. (c) Two columns of pallets are displayed, with only the middle column being complete. The column on the right is incomplete and only partially visible, with no pallets on the right side.
Sensors 24 05330 g010aSensors 24 05330 g010b
Figure 11. The width pixel coordinate constraint.
Figure 11. The width pixel coordinate constraint.
Sensors 24 05330 g011
Figure 12. More complete pallets appear in the camera’s field of view due to deviation between the unmanned forklift’s observation position and the pallet position. (a) When unloading from the front of the flatbed truck, two complete columns of pallets are displayed in the camera’s FOV; (b) when unloading from the rear of the flatbed truck, two complete columns of pallets are displayed in the camera’s FOV.
Figure 12. More complete pallets appear in the camera’s field of view due to deviation between the unmanned forklift’s observation position and the pallet position. (a) When unloading from the front of the flatbed truck, two complete columns of pallets are displayed in the camera’s FOV; (b) when unloading from the rear of the flatbed truck, two complete columns of pallets are displayed in the camera’s FOV.
Sensors 24 05330 g012aSensors 24 05330 g012b
Figure 13. Unloading direction constraint.
Figure 13. Unloading direction constraint.
Sensors 24 05330 g013
Figure 14. Pallet on the opposite side appears in the camera’s field of view.
Figure 14. Pallet on the opposite side appears in the camera’s field of view.
Sensors 24 05330 g014
Figure 15. The depth distance coordinate constraint.
Figure 15. The depth distance coordinate constraint.
Sensors 24 05330 g015
Figure 16. The height pixel coordinate constraint.
Figure 16. The height pixel coordinate constraint.
Sensors 24 05330 g016
Figure 17. Positioning results of the target pallet in complex pallet arrangements. (a) The image is derived from Figure 10a, with the target pallet being the topmost complete pallet in the middle column. (b) The image is derived from Figure 10b, with the target pallet being the topmost complete pallet in the middle column. (c) The image is derived from Figure 10c, with the target pallet being the topmost complete pallet in the middle column. (d) The image is derived from Figure 12a. When unloading from the front of the flatbed truck, the target pallet is the topmost complete pallet in the rightmost column. (e) The image is derived from Figure 12b. When unloading from the rear of the flatbed truck, the target pallet is the topmost complete pallet in the leftmost column. (f) The image is derived from Figure 14, with the target pallet being the one located on the current side.
Figure 17. Positioning results of the target pallet in complex pallet arrangements. (a) The image is derived from Figure 10a, with the target pallet being the topmost complete pallet in the middle column. (b) The image is derived from Figure 10b, with the target pallet being the topmost complete pallet in the middle column. (c) The image is derived from Figure 10c, with the target pallet being the topmost complete pallet in the middle column. (d) The image is derived from Figure 12a. When unloading from the front of the flatbed truck, the target pallet is the topmost complete pallet in the rightmost column. (e) The image is derived from Figure 12b. When unloading from the rear of the flatbed truck, the target pallet is the topmost complete pallet in the leftmost column. (f) The image is derived from Figure 14, with the target pallet being the one located on the current side.
Sensors 24 05330 g017aSensors 24 05330 g017b
Figure 18. Dual-antenna installation position diagram.
Figure 18. Dual-antenna installation position diagram.
Sensors 24 05330 g018
Figure 19. The camera coordinate system.
Figure 19. The camera coordinate system.
Sensors 24 05330 g019
Figure 20. The spatial relationship among the camera plane coordinate system, the antenna coordinate system, and the forklift navigation coordinate system.
Figure 20. The spatial relationship among the camera plane coordinate system, the antenna coordinate system, and the forklift navigation coordinate system.
Sensors 24 05330 g020
Figure 21. Map of the experimental field scene.
Figure 21. Map of the experimental field scene.
Sensors 24 05330 g021
Table 1. The architecture of DDRNet-23-slim.
Table 1. The architecture of DDRNet-23-slim.
StageOutputDDRNet-23-Slim
conv1112 × 1123 × 3, 32, stride 2
conv256 × 563 × 3, 32, stride 2
3 × 3 , 32 3 × 3 , 32 × 2
conv328 × 28 3 × 3 , 64 3 × 3 , 64 × 2
conv414 × 14, 28 × 28 3 × 3 , 128 3 × 3 , 128 × 2 3 × 3 , 64 3 × 3 , 64 × 2
Bilateral fusion
conv5_17 × 7, 28 × 28 3 × 3 , 256 3 × 3 , 256 × 2 3 × 3 , 64 3 × 3 , 64 × 2
Bilateral fusion
1 × 1 , 256 3 × 3 , 256 1 × 1 , 512 × 1 1 × 1 , 64 3 × 3 , 64 1 × 1 , 128 × 1
conv5_27 × 7High-to-low fusion
1 × 1, 1024
1 × 17 × 7 global average pool
1000-d fc, softmax
Table 2. Information about some pallet types in the dataset.
Table 2. Information about some pallet types in the dataset.
NumberMaterialColorDimensions (W × L × H) mmDiagram
1WoodWooden1000 × 1200 × 150Sensors 24 05330 i001
2PlasticBlue1000 × 1200 × 150Sensors 24 05330 i002
Table 3. Results of target pallet segmentation and keypoint extraction based on RGB images.
Table 3. Results of target pallet segmentation and keypoint extraction based on RGB images.
Serial NumberRGB ImagesImage Segmentation
Results
Keypoint 1 Recognition ResultsTarget Pallet Identification Results
1Sensors 24 05330 i003Sensors 24 05330 i004Sensors 24 05330 i005Sensors 24 05330 i006
2Sensors 24 05330 i007Sensors 24 05330 i008Sensors 24 05330 i009Sensors 24 05330 i010
3Sensors 24 05330 i011Sensors 24 05330 i012Sensors 24 05330 i013Sensors 24 05330 i014
1 Keypoints are depicted as red and green circles in the images.
Table 4. Results of target pallet segmentation and keypoint extraction based on depth images.
Table 4. Results of target pallet segmentation and keypoint extraction based on depth images.
Serial NumberDepth ImagesImage Segmentation
Results
Keypoint 1 Recognition ResultsTarget Pallet Identification Results
1Sensors 24 05330 i015Sensors 24 05330 i016Sensors 24 05330 i017Sensors 24 05330 i018
2Sensors 24 05330 i019Sensors 24 05330 i020Sensors 24 05330 i021Sensors 24 05330 i022
3Sensors 24 05330 i023Sensors 24 05330 i024Sensors 24 05330 i025Sensors 24 05330 i026
1 Keypoints are depicted as red and green circles in the images.
Table 5. Error between reference and predicted coordinates of target pallet from RGB images.
Table 5. Error between reference and predicted coordinates of target pallet from RGB images.
Serial
Number
Degrees of FreedomError (mm/°)
A E m i n A E m a x M A E S T D
1X9.68216.21812.4402.160
Y11.82616.99614.6691.695
Z10.40515.04113.0881.556
Angle0.0120.3180.1820.200
2X8.17214.18310.9642.004
Y10.53716.54313.1981.771
Z10.31315.67913.0611.750
Angle0.0150.3360.1720.190
3X11.15715.42613.0881.207
Y10.49615.37313.1561.678
Z11.50717.31914.2451.835
Angle0.0160.3130.1800.198
Table 6. Error between reference and predicted coordinates of target pallet from depth images.
Table 6. Error between reference and predicted coordinates of target pallet from depth images.
Serial
Number
Degrees of FreedomError (mm/°)
A E m i n A E m a x M A E S T D
1X13.12119.67915.8821.878
Y13.16519.57917.0931.784
Z11.15015.77713.7531.532
Angle0.0180.3490.1850.211
2X10.09115.83013.6651.671
Y12.54919.91416.6451.815
Z10.99316.96014.0852.033
Angle0.0190.3640.1690.181
3X12.97218.38815.6031.723
Y13.83019.76617.5791.991
Z12.38718.13515.8541.764
Angle0.0070.3580.1770.206
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Meng, D.; Ren, Y.; Yu, X.; Yin, X.; Wang, W.; Men, J. High-Precision Positioning and Rotation Angle Estimation for a Target Pallet Based on BeiDou Navigation Satellite System and Vision. Sensors 2024, 24, 5330. https://doi.org/10.3390/s24165330

AMA Style

Meng D, Ren Y, Yu X, Yin X, Wang W, Men J. High-Precision Positioning and Rotation Angle Estimation for a Target Pallet Based on BeiDou Navigation Satellite System and Vision. Sensors. 2024; 24(16):5330. https://doi.org/10.3390/s24165330

Chicago/Turabian Style

Meng, Deqiang, Yufei Ren, Xinli Yu, Xiaoxv Yin, Wenming Wang, and Junhui Men. 2024. "High-Precision Positioning and Rotation Angle Estimation for a Target Pallet Based on BeiDou Navigation Satellite System and Vision" Sensors 24, no. 16: 5330. https://doi.org/10.3390/s24165330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop