Next Article in Journal
Compound Fault Diagnosis of Wind Turbine Gearbox via Modified Signal Quality Coefficient and Versatile Residual Shrinkage Network
Previous Article in Journal
RFCS-YOLO: Target Detection Algorithm in Adverse Weather Conditions via Receptive Field Enhancement and Cross-Scale Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Grasp Pose Optimization for Robotic Arms Using Low-Cost Depth Sensors in Complex Environments

1
Faculty of Innovation Engineering, Macau University of Science and Technology, Taipa, Macau SAR 999078, China
2
The Institute of Systems Engineering, Macau University of Science and Technology, Macau SAR 999078, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 909; https://doi.org/10.3390/s25030909
Submission received: 14 November 2024 / Revised: 29 January 2025 / Accepted: 30 January 2025 / Published: 3 February 2025
(This article belongs to the Section Sensors and Robotics)

Abstract

:
This paper presents an efficient grasp pose estimation algorithm for robotic arm systems with a two-finger parallel gripper and a consumer-grade depth camera. Unlike traditional deep learning methods, which suffer from high data dependency and inefficiency with low-precision point clouds, the proposed approach uses ellipsoidal modeling to overcome these issues. The algorithm segments the target and then applies a three-stage optimization to refine the grasping path. Initial estimation fits an ellipsoid to determine principal axes, followed by nonlinear optimization for a six-degree-of-freedom grasp pose. Validation through simulations and experiments showed a target grasp success rate (TGSR) of over 83% under low noise, with only a 4.9% drop under high noise—representing a 68.0% and a 42.4% improvement over GPD and PointNetGPD, respectively. In real-world tests, success rates ranged from 95 to 100%, and the computational efficiency was improved by 56.3% compared to deep learning methods, proving its practicality for real-time applications. These results demonstrate stable and reliable grasping performance, even in noisy environments and with low-cost sensors.

1. Introduction

Dexterous manipulation of robots has broad applications in industries such as manufacturing, logistics, and healthcare. In industrial production, robots improve efficiency by performing tasks like sorting, packaging, and painting. In logistics systems, mobile robots and arms collaborate for intelligent de-stacking, picking, and sorting. In daily life, service robots assist with tasks such as serving tea or organizing, offering significant convenience.
Industrial robots based on artificial intelligence face significant challenges in efficiently grasping diverse objects [1,2,3], and while current grasping techniques are effective in specific scenarios, they often fail to meet the varied demands of industrial applications [4,5,6]. In both industrial and consumer-grade robot grasping tasks, the uncertainty in task execution can arise from several factors.
From a perception perspective, the grasping pose is derived from modeling and analyzing the scene data acquired by sensors. Whether using 2D or 3D data, strategies such as denoising, multi-frame accumulation, or improving sensor accuracy to reduce algorithmic strain may be employed to increase the success rate of task execution. However, these strategies often come with increased time, computational power, or equipment costs. To reduce these costs and make robotic arms more widely applicable, manufacturers have introduced miniaturized, low-cost 3D sensors, such as Intel’s RealSense and Microsoft’s Kinect. However, the perception accuracy of these devices is constrained by the cost, achieving only centimeter-level (5–10 cm) measurement precision. This has facilitated the widespread adoption of low-cost robots; however, the limited perception accuracy has resulted in a lack of robust algorithms for object grasping with these sensors. In fact, the trade-off between device cost, computational cost, and task execution robustness has been a central theme in the development of the robotics community. This paper focuses on the uncertainty in robotic arm grasping tasks caused by imprecise and insufficient perception, and aims to develop a cost-effective, efficient, and robust grasping algorithm, which is key to broader applications [7,8].
In terms of algorithmic principles, grasping solutions can be categorized into model-based and model-free approaches [9]. The former involves high maintenance costs [10,11], while the latter is more flexible but faces challenges in generalization [12,13]. Parallel computing has facilitated the application of deep learning in grasp detection [14], where the use of RGB-D sensors in conjunction with neural networks such as CNN and PointNet has enhanced the accuracy of grasp prediction [15,16,17,18]. However, methods like GPD and PointNet face high data dependency and computational demands, limiting their industrial applications [19,20,21,22,23]. Current methods achieve success rates of 75–95% under controlled conditions; however, this rate is insufficient for dealing with clutter and occlusion in real-world scenarios [7,19,24,25,26]. Improving the robustness and adaptability of algorithms remains a key challenge [27,28,29].
This paper proposes a robust, efficient grasp pose estimation algorithm for a robotic arm with a two-finger gripper and consumer-grade depth camera. Key contributions include the following:
  • A PCA-based point cloud processing method for diverse target types and orientations, offering superior generalization.
  • A grasp strategy considering both target pose and environment for successful, collision-free grasps in complex scenes.
  • Millisecond-level grasp estimation using low-cost depth sensors, ensuring deployability with minimal size, weight, and power (SwaP) requirements.

2. Methods

2.1. System Overview

The hardware and algorithm framework are shown in Figure 1. The system uses an eye-in-hand setup with a RealSense D455 (RealSense D455 camera, Intel Corporation, Santa Clara, CA, USA 2020.) camera on an Elfin5 (Elfin5, JAKA Robotics, Shenzhen, China, 2021.) robotic arm, effective within a 0.5–2 m range for small to medium grasping tasks.
Grasping begins with the accurate localization of the target using MobileSAM for pixel-wise segmentation, followed by point cloud alignment. The system then separates the environment and target point clouds. A multi-objective optimization approach estimates the optimal grasp pose by maximizing grasp success while avoiding collisions, as further explained in Section 2.2 and Section 2.3.

2.2. Data Acquisition and Preprocessing

In this study, we adopt MobileSAM [30], a lightweight image segmentation network optimized for mobile devices. MobileSAM uses a convolutional neural network to extract target regions from images and integrates information across different scales to accurately segment the targets. Compared to traditional image segmentation methods, this approach significantly reduces computational costs while maintaining accuracy, making it suitable for resource-constrained devices. The model processes normalized and resized images to produce segmentation results, followed by post-processing to generate the final mask. Test results are shown in Figure 2.
Pixel alignment between the RGB camera and the depth camera is a critical step in achieving multimodal data fusion. Camera calibration provides the intrinsic and extrinsic matrices for both the RGB and depth cameras. Specifically, a calibration board is used to calibrate the two cameras, obtaining the intrinsic matrices K RGB and K Depth for each camera, as well as the extrinsic matrices R and T between them. The pixel coordinates ( u d , ν d ) from the depth image are mapped to 3D space coordinates ( X ˙ , Y , Z ) , which are then transformed into the RGB camera coordinate system using the extrinsic matrices. Finally, these 3D coordinates are projected back into the RGB pixel coordinate system ( u r , ν r ) using the intrinsic matrix of the RGB camera. This mapping relationship can be expressed by the following equations:
u r ν r 1 = K R G B · R · X Y Z + T
Through the aforementioned process, the depth image can be reprojected into the coordinate system of the RGB image, achieving pixel-level alignment. Point cloud processing is a crucial step in analyzing and manipulating 3D point data obtained from depth images or 3D scanning devices. The first step involves generating a 3D point cloud using camera calibration parameters and the depth image. Specifically, the 3D coordinates ( X , Y , Z ) in the camera coordinate system are calculated for each pixel ( u , ν ) in the depth image along with its corresponding depth value d. This process can be achieved using the following equations:
X = ( u u 0 ) · d / f x
Y = ( v v 0 ) · d / f y
Z = d
where f x and f y are the focal lengths of the camera, and ( u 0 , ν 0 ) are the coordinates of the principal point.
After the above process, we can align the results from 2D image recognition with the 3D sensor data on a pixel-by-pixel basis to obtain the 3D information of the target object. Even in cases where the target overlaps with the background, the object can still be separated from the environment. However, in more complex scenarios, relying solely on instance segmentation from 2D images may not be robust enough. For example, when the object and background objects have similar colors or textures, and there is significant foreground-background overlap, the instance segmentation results on the image may suffer from boundary overflow, which could affect the extraction of the 3D information of the target. To address this, we employ a clustering strategy to exclude point clouds that do not belong to the target. The implementation principle is described below: We introduce the Euclidean clustering method, which effectively removes misidentified objects in the background under overlapping environments by considering the spatial distance information of objects in the point cloud. Even if instance segmentation confuses background objects with foreground objects, their positions in three-dimensional space typically differ significantly. Euclidean clustering calculates the Euclidean distance between points and groups closely located points into the same cluster, thereby effectively separating foreground and background objects. For each point, Euclidean clustering computes its neighborhood and assigns these points to the same cluster based on a preset distance threshold ϵ . If a background object is spatially far from a foreground object, even if it is mistakenly segmented as foreground in the 2D image, Euclidean clustering can exclude it from the foreground. After clustering, background objects, due to their greater distance from foreground objects, are typically assigned to a separate cluster or labeled as noise, thereby being discarded. Ultimately, only the point cloud close to the foreground is retained, ensuring the effective removal of background objects.
It is important to note that if the target we wish to extract is largely occluded by an unrelated object and becomes invisible, this will first pose a challenge to instance segmentation. Thus, this strategy is primarily designed for optimizing the segmentation of overlapping foreground objects. Regarding overlapping background objects, we may need to consider how to circumvent and grasp targets in a more reasonable posture, which will be discussed in later in Section 2.3.

2.3. Grasp Pose Estimation

This section outlines the method for estimating the grasp pose, as introduced in Section 2.2. The coordinate system of the robotic arm is illustrated in Figure 3. The prismatic axis defines the gripper’s movement, the tilt axis indicates the pitch angle, and the advance axis, derived from the cross product of the prismatic and tilt axes, defines the direction of gripper extension.

2.3.1. Ellipsoid Modeling

Grasp pose estimation aims to improve success by analyzing grasp factors. Objects are grasped by their thinnest part, modeled within an ellipsoid bounding box, as shown in Figure 4.
The ellipsoid model can be represented by the following general quadratic surface equation:
A x 2 + B y 2 + C z 2 + 2 D x y + 2 E x z + 2 F y z + 2 G x + 2 H y + 2 I z + J = 0
Let the point cloud data be denoted as P = { ( x i , y i , z i ) } i = 1 N . To apply the least squares method, a design matrix A and an observation vector b are constructed as follows:
A = x 1 2 y 1 2 z 1 2 2 x 1 y 1 2 x 1 z 1 2 y 1 z 1 2 x 1 2 y 1 2 z 1 x 2 2 y 2 2 z 2 2 2 x 2 y 2 2 x 2 z 2 2 y 2 z 2 2 x 2 2 y 2 2 z 2 x N 2 y N 2 z N 2 2 x N y N 2 x N z N 2 y N z N 2 x N 2 y N 2 z N , b = 1 1 1
Solving the linear system A X = b using the least squares method yields the parameter vector X = [ A , B , C , D , E , F , G , H , I , J ] T . These parameters define the general equation of the ellipsoid. The next step is to derive the ellipsoid’s geometric parameters, such as the center position and the lengths of its axes.
By completing the square for the ellipsoid equation, the coordinates of the center point Q t = ( x 0 , y 0 , z 0 ) can be calculated. Assuming the ellipsoid equation is in the standard form, as follows:
A ( x x 0 ) 2 + B ( y y 0 ) 2 + C ( z z 0 ) 2 + = 1
If the cross-product terms (i.e., the terms involving x y , x z , and y z , such as 2 D x y , 2 E x z , 2 F y z ) are non-zero, it indicates that the ellipsoid is rotated relative to the coordinate axes. Specifically, the non-zero cross-product terms imply that the principal axes of the ellipsoid are not aligned with the coordinate axes but are instead rotated. In this case, a rotation matrix can be derived to transform the ellipsoid into its canonical or standard form, where the axes are aligned with the coordinate system.
The center coordinates of the ellipsoid are then given by
x 0 = G A , y 0 = H B , z 0 = I C
Next, the quadratic form matrix Q is constructed as follows:
Q = A D E D B F E F C
The eigenvalue decomposition of Q is performed as follows:
Q = R Λ R T
where Λ is the diagonal matrix, with its diagonal elements being the eigenvalues λ 1 , λ 2 , λ 3 , and the square roots of the reciprocals of the eigenvalues provide the lengths of the semi-axes of the ellipsoid, denoted as a , b , c :
a = 1 λ 1 , b = 1 λ 2 , c = 1 λ 3
where λ 1 < λ 2 < λ 3 .
The transformation matrix from the ellipsoid coordinate system to the world coordinate system, T w t , includes both rotation and translation components:
T w t = R t 0 1
where R is the rotation matrix composed of the eigenvectors, and t is the translation vector, representing the center position of the ellipsoid ( x 0 , y 0 , z 0 ) .
( x 0 , y 0 , z 0 ) = G A , H B , I C
The lengths of the axes are
a = 1 λ 1 , b = 1 λ 2 , c = 1 λ 3
The complete transformation matrix is
T w t = R [ x 0 , y 0 , z 0 ] T 0 1
These geometric parameters, including the rotation matrix and axis lengths, are depicted in Figure 5.
Based on the previous analysis, we can define the three-axis coordinate system of the modeled ellipsoid, with the transformation matrix from the world coordinate system to this new system being denoted as T w t . The rotation matrix R w t = [ R x , R y , R z ] , where R x and R z correspond to the longest and shortest axes of the ellipsoid { C t } , respectively. The initial grasp pose is configured such that the marching axis aligns with the negative direction of the y-axis of { C t } , and the prismatic axis aligns with the z-axis. This configuration ensures that the gripper can effectively grasp larger objects within its limited finger opening and closing range. The tilt axis aligns with the x-axis, minimizing the risk of contact between the robotic arm and the target or surrounding objects. Consequently, the grasp pose transformation involves a ninety-degree rotation of the coordinate system around the x-axis, centered at the ellipsoid’s center, aligning the transformed y-axis with the original z-axis. The resulting coordinate vector is then multiplied by a rotation matrix representing this ninety-degree rotation about the x-axis. The grasp pose T g is subsequently defined as
T g 0 = R x ( π ) 0 0 1 · T w t , R x ( π ) = 1 0 0 0 0 1 0 1 0

2.3.2. Grasp Pose Optimization

The ellipsoid model and initial grasp pose were established, but a PCA-based model is insufficient for complex environments. Our algorithm adjusts the pitch angle based on the environment and target distance, expanding the workspace and enhancing robustness. Pitch and yaw angles are analyzed separately.
The pitch angle of the gripper is influenced by two factors: the target’s height relative to the robot. For a given grasping target, its height relative to the robot arm’s base can be represented as h, and the distance of the target from the robot arm’s base in the XOY plane can be expressed as d = x 2 + y 2 . The parameters α and β are set based on the operational range of the robot arm. The graphical representation of this function is shown in Figure 6.
In robotic grasping systems, the pitch angle of the grasping pose critically affects the success rate and stability. To optimize grasping performance, we propose a Gaussian-based pitch angle modeling method that accounts for variations in the horizontal distance and vertical height between the target and the robotic arm’s base.
First, as the horizontal distance d between the target and the robot arm’s base increases, the pitch angle θ should gradually decrease, allowing the robotic arm to extend further and cover a larger workspace. Similarly, as the vertical height h of the target increases, the pitch angle θ should also decrease appropriately to ensure stable grasping over a wider range of heights. Based on this design concept, we model the pitch angle θ as a two-dimensional Gaussian function of the horizontal distance d and vertical height h:
θ ( h , d ) = θ 0 · exp d d 0 2 2 σ d 2 h h 0 2 2 σ h 2
where θ 0 is the maximum value of the pitch angle, and d 0 and h 0 represent the horizontal distance and vertical height at which the pitch angle reaches its maximum value, respectively. The parameters σ d and σ h control the spread of the Gaussian distribution along the horizontal distance and vertical height, determining the rate at which the pitch angle decays as d and h change.
After obtaining the optimal pitch angle θ , the grasping pose is fine-tuned, and the transformation matrix is given by
R x ( θ ) = 1 0 0 0 cos θ sin θ 0 sin θ cos θ
With this modeling approach, the pitch angle reaches its maximum value when the target is near d = d 0 and h = h 0 , and it gradually decreases following a Gaussian distribution as the target deviates in horizontal distance or vertical height. This approach integrates target position with grasp pose, maximizing workspace and enhancing flexibility. Figure 6b illustrates grasp adjustments and optimizations for different heights and distances.

2.3.3. Obstacle Avoidance Design

This section builds upon the previous discussion, where we modeled the target and established the initial grasp pose, and optimized the pitch angle. Now, we focus on adjusting the yaw angle to improve obstacle avoidance in complex environments.
Figure 7 illustrates various grasping scenarios: (a) grasping a pen on a flat surface, (b) grasping a target on an inclined surface requiring yaw angle adjustment, and (c) grasping a target in a cluttered environment that demands flexible yaw optimization to avoid obstacles. This approach aims to enhance obstacle avoidance by adaptively adjusting the yaw angle based on the surrounding environment.
To optimize the grasp pose, we define the grasp pose as T g , where the approach direction aligns with the z-axis, and the gripper’s opening direction is along the x-axis. Given the point cloud of the target object P t and the neighborhood point cloud P r within a radius r around the object, we denote the yaw angle of the robotic arm as α .
To assess the potential for collisions between the robotic arm and obstacles during the grasping process, we introduce an optimization objective function based on the variance of the projection distribution of the neighborhood point cloud P r along the z-axis:
J ( α ) = Var { z i ( α ) | p i P r }
Here, p i = ( x i , y i , z i ) represents any point in the neighborhood point cloud P r , and z i ( α ) denotes the projection of point P i on the z-axis under the yaw angle α .
To make this variance more intuitive, the objective function is rewritten as
J ( α ) = 1 P r i = 1 | P r | z i ( α ) z ¯ ( α ) 2
where | P r | is the number of points in the neighborhood point cloud, and z ¯ ( α ) is the mean projection of the neighborhood point cloud on the z-axis under the yaw angle α , calculated as follows:
z ¯ ( α ) = 1 | P r | i = 1 | P r | z i ( α )
By minimizing the objective function J ( α ) , the projection of the neighborhood point cloud on the z-axis can be made more concentrated, thereby reducing the likelihood of collisions between the robotic arm and obstacles during the grasping process.
This optimization method not only enables the grasp pose to better adapt to the current environment but also significantly enhances the stability and success rate of the grasp in practical applications. The optimization process can be carried out using methods such as gradient descent, genetic algorithms, or simulated annealing, ultimately yielding the optimal yaw angle α to guide the actual grasping operation. In summary, the implementation details of our algorithm are shown in the pseudocode of Algorithm 1.
Algorithm 1: Grasp pose estimation and optimization.
Sensors 25 00909 i001

3. Experiment

The experiments included both simulation and real-world tests on a platform with an Intel Core i9-10700 CPU, NVIDIA GTX 2060Ti GPU, and Ubuntu 20.04.

3.1. Simulation Experiments

Grasp stability and success are influenced by point cloud noise and environmental complexity. Point cloud noise arises from sensor accuracy, environmental factors, and motion dynamics, which can significantly affect the performance of grasping algorithms. In real-world applications, sensor errors, variations in ambient lighting, surface reflectivity, and occlusions introduce various levels of noise, leading to inaccurate depth data or incomplete point clouds. This noise can reduce the accuracy of object recognition, pose estimation, and grasp planning, potentially resulting in grasp failures or collisions. Therefore, the robustness of the algorithm to noise is crucial for improving stability in practical scenarios.
To further assess the impact of point cloud noise on grasping performance, we conducted simulation experiments using Gazebo, a widely used platform for robotic simulation. Gazebo simulations were conducted with the robotic arm operating between 0.5 and 1.5 m, using an RGB-D camera. Gazebo, with ODE dynamics and ROS integration, provides realistic simulations and motion control.
Various common objects with different scales and shapes, including a pen, a banana, a mobile phone, an apple, a soda can, and a game controller, were selected as grasping targets. These objects were randomly placed on a table within the testing environment and tested at three different horizontal distances from the robotic arm’s origin, 0.6–0.8 m ( d 1 ), 0.8–1.0 m ( d 2 ), and 1.0–1.2 m ( d 3 ), which were used as the first control variable. The experimental setup is shown in Figure 8.
To evaluate the algorithm’s stability under different point cloud noise conditions, random noise following a normal distribution was added to the coordinates of each target point, defined as p i = p i + n i , where P i represents the ith point in the original point cloud, p i is the point with added noise, and n i N ( 0 , σ 2 ) denotes Gaussian noise with a mean of 0 and a variance of σ 2 . The noise levels were set at σ = 5 mm (low noise), σ = 10 mm (medium noise), and σ = 15 mm (high noise).
Each group was repeated 1000 times in the simulation. A successful grasp required lifting the target over 20 cm with 4 m/s² acceleration. TGSR was calculated and compared to learning-based methods GPD [19] and PointNetGPD [20]. The experimental results are shown in Figure 9, where (a), (b), and (c) represent the TGSR (%) results under point cloud noise levels of σ = 5 mm, σ = 10 mm, and σ = 15 mm, respectively. The results indicate that for medium-distance targets (0.8–1.0 m), PointNetGPD and the proposed method both achieved TGSRs above 70% in low noise, with the proposed method reaching over 83%. Under high noise, GPD and PointNetGPD rates dropped by 15.3% and 8.5%, respectively, due to reliance on point cloud stability. The proposed method showed only a 4.9% reduction, demonstrating greater robustness.
The averaged TGSR results under different noise levels are shown in Table 1. All methods performed best at medium distances, as short and long distances can cause kinematic singularities, leading to failures. The proposed method uses a nonlinear adjustment factor for adapting to different distances, achieving higher average TGSR across all scenarios.

3.2. Real-World Experiments

To validate the proposed method’s performance, two grasping scenarios were configured: static and dynamic. In the static scenario, an Elfin5 robotic arm with an AG_95 gripper (maximum opening width of approximately 12 cm) was used for tabletop grasping, as illustrated in Figure 10. Six objects—box, bottle, glasses case, scissors, pen, and tape—were selected as targets, representing common small to medium-sized desk items (volumes < 0.5 dm³).
The dynamic scenario, presenting greater challenges, employed a Robotiq 2F-85 gripper (maximum opening width 20 cm) mounted on a mobile platform for grasping from both the front and back of a long table. Target objects included a bottle, cup, game controller, mouse, and various fruits, representing medium-sized everyday items (volumes 0.2–1.5 dm³). Objects were densely arranged to increase the task complexity.
The Realsense-D455 was used for both experiments. The proposed method showed a 56.3% and 48.6% higher efficiency compared to GPD and PointNetGPD, relying on model optimization instead of high-performance GPUs, as shown in Figure 11.
In the static experiment, all objects were placed on a black mat, and the grasping performance of six objects was tested individually. Each target was repositioned 20 times for grasp attempts, and the success rates were evaluated. Figure 12 and Table 2 show the different grasp postures and test results for grasping a bottle, pen, and scissors.
The proposed method achieved a 100% grasp success rate for pen and glass case targets in all 20 attempts, demonstrating its stability and robustness. Other methods performed worse due to discrepancies between training and sensor-generated data, leading to incorrect grasps and poor repeatability, revealing a limitation of learning-based approaches.
In dynamic experiments, 20 targets were arranged as shown in Figure 13 and grasped sequentially, with three attempts per target. Experiment 1 grasped from the front, and Experiment 2 from the back. Success rates, calculated as the ratio of successful grasps to total attempts, are shown in Table 3.
The results presented in Table 3 indicate that the proposed method achieves a higher grasp success rate with fewer attempts, which is crucial for real-world robotic applications where precision is challenging.

4. Discussion

The grasp pose estimation algorithm proposed in this study represents a significant advancement in robotic arm manipulation, particularly for systems using consumer-grade RGB-D depth sensors. Extensive experiments conducted in both simulation environments and real-world scenarios demonstrate that this method significantly outperforms existing deep learning-based grasping algorithms, such as GPD and PointNetGPD, in terms of grasp success rate and computational efficiency. Specifically, under varying experimental conditions, the proposed algorithm exhibits stable performance in complex environments, while significantly enhancing computational efficiency and reducing reliance on high-performance computing resources, highlighting its broad applicability in both industrial and consumer robotics. In contrast to traditional deep learning methods, this approach does not rely on large annotated datasets or complex neural network training, resulting in lower computational costs and higher adaptability, enabling it to run on hardware platforms with limited resources.
The success of the algorithm can also be attributed to the application of nonlinear optimization techniques. By fitting ellipsoids to the point cloud, we are able to perform principal component analysis of the target’s geometric distribution, effectively describing the shape of the target object, particularly for objects with simple geometric structures. Additionally, this method incorporates the surrounding environment’s features into the nonlinear optimization for pose estimation, enabling the robotic arm to perform obstacle-avoiding grasps in the scene. However, despite its strong performance in most scenarios, the ellipsoidal model may not accurately capture the true shape of objects with complex topological structures or in dynamic environments, thereby affecting the grasp success rate. Specifically, when the surface geometry of the target object is irregular, the optimal grasp pose may not be derived solely from geometric distribution analysis, but often requires the integration of empirical constraints. For example, objects such as teapots or pots with handles may necessitate the introduction of neural network analysis at certain stages.
To enhance the stability of the algorithm in diverse environments, future research will focus on further optimizing the model to better address these challenges. Future work will concentrate on several key aspects: First, more precise geometric modeling methods, such as multimodal fusion, adaptive surface modeling, and point cloud completion, will be introduced to further improve the algorithm’s adaptability and accuracy. Second, hybrid models based on deep learning will be explored by combining traditional geometric modeling with deep learning approaches, thereby enhancing the algorithm’s generalization ability in complex scenarios.

5. Conclusions

In this study, the proposed grasp pose estimation algorithm demonstrated strong robustness and efficiency across multiple experiments. We simulated the impact of different sensor accuracies and environmental conditions on grasp performance by using varying levels of noise (low, medium, and high noise). Under low-noise conditions, the algorithm was able to estimate object poses with high accuracy, achieving a success rate close to 90%. As noise levels increased, the grasp success rates of other comparison methods dropped significantly, highlighting the non-negligible impact of noise on the algorithm’s performance. Under medium-noise conditions, the proposed algorithm achieved a grasp success rate (TGSR) exceeding 87%, outperforming GPD and PointNetGPD by 20% and 10%, respectively. This method effectively simplifies the modeling of object geometry, and the strategy of incorporating environmental constraints into pose optimization enables the robot to maintain high performance in cluttered environments. The algorithm also demonstrated good adaptability when handling objects of varying shapes and sizes, maintaining stable performance across different test objects.

Author Contributions

Conceptualization, A.C., X.L., K.C. and C.H.; methodology, A.C., X.L. and K.C.; software, A.C., X.L. and K.C.; validation, A.C. and C.H.; formal analysis, A.C., X.L. and K.C.; investigation, A.C., X.L. and K.C.; resources, A.C., X.L. and K.C.; data curation, A.C., X.L. and K.C.; writing—original draft preparation, A.C., X.L. and K.C.; writing—review and editing, A.C., X.L., K.C. and C.H.; visualization, A.C., X.L. and K.C.; funding acquisition, A.C. All authors have contributed to this work and have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Development Fund of Macau SAR, grant number FDCT 005/2022/ALC.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Belanche, D.; Casaló, L.V.; Flavián, C.; Schepers, J. Service robot implementation: A theoretical framework and research agenda. Serv. Ind. J. 2020, 40, 203–225. [Google Scholar] [CrossRef]
  2. Varley, J.; Weisz, J.; Weiss, J.; Allen, P. Generating multi-fingered robotic grasps via deep learning. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 4524–4530. [Google Scholar]
  3. Fang, H.-S.; Wang, C.; Fang, H.; Gou, M.; Liu, J.; Yan, H.; Liu, W.; Xie, Y.; Lu, C. Anygrasp: Robust and efficient grasp perception in spatial and temporal domains. IEEE Trans. Robot. 2023, 5, 3929–3945. [Google Scholar] [CrossRef]
  4. Mahler, J.; Pokorny, F.T.; Hou, B.; Roderick, M.; Laskey, M.; Aubry, M.; Kohlhoff, K.; Kröger, T.; Kuffner, J.; Goldberg, K. Dex-net 1.0: A cloud-based network of 3D objects for robust grasp planning using a multi-armed bandit model with correlated rewards. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1954–1961. [Google Scholar]
  5. Zhang, Y.; Adin, V.; Bader, S.; Oelmann, B. Leveraging Acoustic Emission and Machine Learning for Concrete Materials Damage Classification on Embedded Devices. IEEE Trans. Instrum. Meas. 2023, 72, 2525108. [Google Scholar] [CrossRef]
  6. Lenz, I.; Lee, H.; Saxena, A. Deep learning for detecting robotic grasps. Int. J. Robot. Res. 2015, 34, 705–724. [Google Scholar] [CrossRef]
  7. Xie, Z.; Liang, X.; Roberto, C.J. Learning-based robotic grasping: A review. Front. Robot. AI 2023, 10, 1038658. [Google Scholar] [CrossRef] [PubMed]
  8. Sundermeyer, M.; Mousavian, A.; Triebel, R.; Fox, D. Contact-graspnet: Efficient 6-DOF grasp generation in cluttered scenes. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 4244–4250. [Google Scholar]
  9. Bicchi, A.; Kumar, V. Robotic grasping and contact: A review. In Proceedings of the 2000 ICRA Millennium Conference IEEE International Conference on Robotics and Automation Symposia Proceedings, San Francisco, CA, USA, 24–28 April 2000; IEEE: Piscataway, NJ, USA, 2000; pp. 348–353. [Google Scholar]
  10. Boubekri, N.; Chakraborty, P.J. Robotic grasping: Gripper designs, control methods and grasp configurations—A review of research. Int. J. Mech. Sci. 2002, 13, 520–531. [Google Scholar] [CrossRef]
  11. Joshi, S.; Kumra, S.; Sahin, F. Robotic grasping using deep reinforcement learning. In Proceedings of the 2020 IEEE 16th International Conference on Automation Science and Engineering (CASE), Washington, DC, USA, 22–26 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 351–357. [Google Scholar]
  12. Guo, N.; Zhang, B.; Zhou, J.; Zhan, K.; Lai, S. Pose estimation and adaptable grasp configuration with point cloud registration and geometry understanding for fruit grasp planning. J. Comput. Agric. Eng. 2020, 179, 105818. [Google Scholar] [CrossRef]
  13. Ni, P.; Zhang, W.; Zhu, X.; Cao, Q. Pointnet++ grasping: Learning an end-to-end spatial grasp generation algorithm from sparse point clouds. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–4 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 4597–4603. [Google Scholar]
  14. Qian, K.; Jing, X.; Duan, Y.; Zhou, B.; Fang, F.; Xia, J.; Ma, X. Grasp pose detection with affordance-based task constraint learning in single-view point clouds. J. Robot. Autom. 2020, 100, 145–163. [Google Scholar] [CrossRef]
  15. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 77–85. [Google Scholar]
  16. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25 (NIPS 2012), Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
  17. Zhou, L.; Sun, G.; Li, Y.; Li, W.; Su, Z.J. Point cloud denoising review: From classical to deep learning-based approaches. J. Geom. Model. 2022, 121, 101140. [Google Scholar] [CrossRef]
  18. Nguyen, V.-D. Constructing force-closure grasp. J. Robot. Res. 1988, 7, 157–171. [Google Scholar] [CrossRef]
  19. Pas, A.; Gualtieri, M.; Saenko, K.; Platt, R. Grasp pose detection in point clouds. Int. J. Robot. Res. 2017, 36, 1455–1473. [Google Scholar]
  20. Liang, H.; Ma, X.; Li, S.; Görner, M.; Tang, S.; Fang, B.; Sun, F.; Zhang, J. Pointnetgpd: Detecting grasp configurations from point sets. In Proceedings of the 2019 IEEE International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1954–1961. [Google Scholar]
  21. Dharbaneshwer, S.; Subramanian, S.J.; Kohlhoff, K.J.M. Robotic grasp analysis using deformable solid mechanics. J. Mech. Eng. 2019, 54, 1767–1784. [Google Scholar] [CrossRef]
  22. Pierson, H.A.; Gashler, M.S. Deep learning in robotics: A review of recent research. J. Robot. Res. 2017, 31, 821–835. [Google Scholar] [CrossRef]
  23. Billard, A.; Kragic, D. Trends and challenges in robot manipulation. Science 2019, 364, eaat8414. [Google Scholar] [CrossRef] [PubMed]
  24. Caldera, S.; Rassau, A.; Chai, D. Review of deep learning methods in robotic grasp detection. J. Mach. Interact. 2018, 2, 57. [Google Scholar] [CrossRef]
  25. Mohammed, M.Q.; Chung, K.L.; Chyi, C.S. Review of deep reinforcement learning-based object grasping: Techniques, open challenges, and recommendations. J. Intell. Autom. 2020, 8, 178450–178481. [Google Scholar] [CrossRef]
  26. Hinterstoisser, S.; Holzer, S.; Cagniart, C.; Ilic, S.; Konolige, K.; Navab, N.; Lepetit, V. Multimodal templates for real-time detection of texture-less objects in heavily cluttered scenes. In Proceedings of the 2011 International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 858–865. [Google Scholar]
  27. Dong, M.; Zhang, J. A review of robotic grasp detection technology. J. Robot. Artif. Intell. 2023, 41, 3846–3885. [Google Scholar] [CrossRef]
  28. Hannah, R.; Aron, A.R. Towards real-world generalizability of a circuit for action-stopping. J. Neurosci. Res. 2021, 22, 538–552. [Google Scholar] [CrossRef] [PubMed]
  29. Zhang, C.; Han, D.; Qiao, Y.; Kim, J.U.; Bae, S.-H.; Lee, S.; Hong, C.S. Faster segment anything: Towards lightweight SAM for mobile applications. J. Comput. Vis. 2023, 1–12. [Google Scholar]
  30. Takaya, K.; Asai, T.; Kroumov, V.; Smarandache, F. Simulation environment for mobile robots testing using ROS and Gazebo. In Proceedings of the 2016 20th International Conference on System Theory, Control and Computing (ICSTCC), Sinaia, Romania, 19–21 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 157–162. [Google Scholar]
Figure 1. System algorithm flowchart.
Figure 1. System algorithm flowchart.
Sensors 25 00909 g001
Figure 2. Target localization and segmentation: (a) original image; (b) segmentation result.
Figure 2. Target localization and segmentation: (a) original image; (b) segmentation result.
Sensors 25 00909 g002
Figure 3. Two-finger parallel gripper coordinate system.
Figure 3. Two-finger parallel gripper coordinate system.
Sensors 25 00909 g003
Figure 4. Grasping scene modeling: (a) point cloud of the grasping scene with registered and fused color images; (b) ellipsoid modeling of the grasping scene.
Figure 4. Grasping scene modeling: (a) point cloud of the grasping scene with registered and fused color images; (b) ellipsoid modeling of the grasping scene.
Sensors 25 00909 g004
Figure 5. Ellipsoidal modeling of grasping target: (a) point cloud representation; (b) transformation matrix and axis lengths.
Figure 5. Ellipsoidal modeling of grasping target: (a) point cloud representation; (b) transformation matrix and axis lengths.
Sensors 25 00909 g005
Figure 6. Optimization of grasp pose: (a) influence of target distance and height on pitch angle; (b) grasp adjustments for varying heights and distances.
Figure 6. Optimization of grasp pose: (a) influence of target distance and height on pitch angle; (b) grasp adjustments for varying heights and distances.
Sensors 25 00909 g006
Figure 7. Grasp pose adjustments in complex environments: (a) grasping a pen on a flat surface; (b) grasping a target on an inclined surface; (c) grasping a target in a cluttered environment.
Figure 7. Grasp pose adjustments in complex environments: (a) grasping a pen on a flat surface; (b) grasping a target on an inclined surface; (c) grasping a target in a cluttered environment.
Sensors 25 00909 g007
Figure 8. Simulation experiment setup.
Figure 8. Simulation experiment setup.
Sensors 25 00909 g008
Figure 9. Grasp success rate (TGSR) comparison for different methods under varying noise levels: (a) TGSR at low noise; (b) TGSR at medium noise; (c) TGSR at high noise.
Figure 9. Grasp success rate (TGSR) comparison for different methods under varying noise levels: (a) TGSR at low noise; (b) TGSR at medium noise; (c) TGSR at high noise.
Sensors 25 00909 g009
Figure 10. Real-world grasping scenarios: (a) static tabletop grasping scenario; (b) dynamic grasping scenario with mobile platform.
Figure 10. Real-world grasping scenarios: (a) static tabletop grasping scenario; (b) dynamic grasping scenario with mobile platform.
Sensors 25 00909 g010
Figure 11. Box Plot comparing grasp pose estimation time for three methods. (The boxplot illustrates the distribution of the data, where the blue box represents the interquartile range (IQR) with the lower edge at the first quartile (Q1), the upper edge at the third quartile (Q3), and the middle line representing the median (Q2). The red line indicates the mean value of the data, while the red cross marks the outliers, which are data points significantly different from the majority of the data, typically located beyond 1.5 times the IQR from the quartiles.)
Figure 11. Box Plot comparing grasp pose estimation time for three methods. (The boxplot illustrates the distribution of the data, where the blue box represents the interquartile range (IQR) with the lower edge at the first quartile (Q1), the upper edge at the third quartile (Q3), and the middle line representing the median (Q2). The red line indicates the mean value of the data, while the red cross marks the outliers, which are data points significantly different from the majority of the data, typically located beyond 1.5 times the IQR from the quartiles.)
Sensors 25 00909 g011
Figure 12. Example of target grasp execution, with images captured by Realsense D455 Color Camera, with the gripper end at the bottom of the image. (Figure (a): Grasping the bottle on the left side of the robot; Figure (b): Grasping the pen; Figure (c): Grasping the scissors; Figure (d): Grasping the bottle on the right side of the robot; Figure (e): Grasping the pen placed directly next to the bottle; Figure (f): Grasping the scissors on the box.)
Figure 12. Example of target grasp execution, with images captured by Realsense D455 Color Camera, with the gripper end at the bottom of the image. (Figure (a): Grasping the bottle on the left side of the robot; Figure (b): Grasping the pen; Figure (c): Grasping the scissors; Figure (d): Grasping the bottle on the right side of the robot; Figure (e): Grasping the pen placed directly next to the bottle; Figure (f): Grasping the scissors on the box.)
Sensors 25 00909 g012
Figure 13. Target set in dynamic scenarios.
Figure 13. Target set in dynamic scenarios.
Sensors 25 00909 g013
Table 1. Mean TGSR (%) in three noise environments.
Table 1. Mean TGSR (%) in three noise environments.
Dist.Method\TargetPenBananaAppleBottleCanGamepadScissorsCaseAvg.
d1GPD46.254.772.173.356.952.863.850.846.8
PointNetGPD65.166.783.476.982.677.378.872.764.9
Our Method85.490.396.394.592.186.180.886.485.6
d2GPD52.156.171.270.960.463.473.568.350.8
PointNetGPD65.679.586.682.676.574.883.387.265.5
Our Method87.490.891.392.688.486.883.284.287.1
d3GPD40.237.336.550.347.231.134.940.841.1
PointNetGPD58.052.968.364.372.868.367.052.657.5
Our Method85.680.886.280.585.976.178.979.885.8
Table 2. Target grasp success rates (%) of three methods in real-world static experiments.
Table 2. Target grasp success rates (%) of three methods in real-world static experiments.
TargetBoxBottleGlass CaseScissorsPenGlue
GPD504545503055
PointNetGPD606575555075
Proposed Method95901009510095
Table 3. Attempts count and target grasp success rates (%) of three methods in real-world dynamic experiments.
Table 3. Attempts count and target grasp success rates (%) of three methods in real-world dynamic experiments.
GroupGPDPointNetGPDOur Method
Attempts CountSuccess Rate (%)Attempts CountSuccess Rate (%)Attempts CountSuccess Rate (%)
15150458027100
2476041753095
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, A.; Li, X.; Cen, K.; Hon, C. Adaptive Grasp Pose Optimization for Robotic Arms Using Low-Cost Depth Sensors in Complex Environments. Sensors 2025, 25, 909. https://doi.org/10.3390/s25030909

AMA Style

Chen A, Li X, Cen K, Hon C. Adaptive Grasp Pose Optimization for Robotic Arms Using Low-Cost Depth Sensors in Complex Environments. Sensors. 2025; 25(3):909. https://doi.org/10.3390/s25030909

Chicago/Turabian Style

Chen, Aiguo, Xuanfeng Li, Kerui Cen, and Chitin Hon. 2025. "Adaptive Grasp Pose Optimization for Robotic Arms Using Low-Cost Depth Sensors in Complex Environments" Sensors 25, no. 3: 909. https://doi.org/10.3390/s25030909

APA Style

Chen, A., Li, X., Cen, K., & Hon, C. (2025). Adaptive Grasp Pose Optimization for Robotic Arms Using Low-Cost Depth Sensors in Complex Environments. Sensors, 25(3), 909. https://doi.org/10.3390/s25030909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop