Next Article in Journal
Development of LPFG-Based Seawater Concentration Monitoring Sensors Packaged by BFRP
Previous Article in Journal
Design of an Angled Single-Excitation Elliptical Vibration System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-View Three-Dimensional Scanning Method for a Dual-Arm Hand–Eye System with Global Calibration of Coded Marker Points

1
School of Mechanical Engineering, Tiangong University, Tianjin 300387, China
2
College of Artificial Intelligence, Nankai University, Tianjin 300350, China
3
Ocean College, Zhejiang University, Zhoushan 316021, China
4
State Key Laboratory of Precision Measurement Technology and Instruments, Tianjin University, Tianjin 300072, China
5
College of Electronic Information and Optical Engineering, Nankai University, Tianjin 300350, China
*
Author to whom correspondence should be addressed.
Micromachines 2025, 16(7), 809; https://doi.org/10.3390/mi16070809 (registering DOI)
Submission received: 31 May 2025 / Revised: 1 July 2025 / Accepted: 7 July 2025 / Published: 13 July 2025

Abstract

To achieve robust and accurate collaborative 3D measurement under complex noise conditions, a global calibration method for dual-arm hand–eye systems and multi-view 3D imaging is proposed. A multi-view 3D scanning approach based on ICP (M3DHE-ICP) integrates a multi-frequency heterodyne coding phase solution with ICP optimization, effectively correcting stitching errors caused by robotic arm attitude drift. After correction, the average 3D imaging error is 0.082 mm, reduced by 0.330 mm. A global calibration method based on encoded marker points (GCM-DHE) is also introduced. By leveraging spatial geometry constraints and a dynamic tracking model of marker points, the transformation between multi-coordinate systems of the dual arms is robustly solved. This reduces the average imaging error to 0.100 mm, 0.456 mm lower than that of traditional circular calibration plate methods. In actual engineering measurements, the average error for scanning a vehicle’s front mudguard is 0.085 mm, with a standard deviation of 0.018 mm. These methods demonstrate significant value for intelligent manufacturing and multi-robot collaborative measurement.

1. Introduction

Driven by increasing market demand and the pressing need for technological innovation, the integration of robots with high-precision 3D vision has emerged as a crucial solution for flexible measurement challenges. In recent years, single-arm visual measurement systems have shown limitations, including low measurement speed, inefficiency, and an inability to perform full 3D inspection when target points fall outside the robotic workspace. To overcome these limitations, collaborative 3D measurement using dual robotic arms has gained increasing attention for its potential to achieve high-precision, efficient, and non-contact 3D inspection in complex measurement scenarios [1,2,3,4].
The widespread adoption of multi-camera systems and multi-robot measurement platforms in industrial manufacturing, healthcare, defense, and other sectors has attracted significant research attention to precise collaborative 3D measurement techniques. Hand–eye calibration is a critical prerequisite for collaborative 3D measurement in multi-robot systems. Its primary objective is to determine the precise geometric relationship between each robot end-effector and its corresponding vision sensor, thereby enabling coordinated measurement. Global calibration methods for multi-robot hand–eye systems have evolved into relatively mature frameworks encompassing theoretical modeling, visual inspection, data fusion, and optimization algorithms. However, they continue to face challenges in enhancing robustness, real-time performance, and accuracy under dynamic conditions [5,6,7]. To achieve global calibration in multi-robot vision systems, each robotic arm must acquire accurate and stable calibration data across diverse poses, followed by data processing and registration to establish coordinate transformations across multiple robot frames [8,9]. Wang et al. proposed an enhanced automatic TCP calibration method based on a laser displacement sensor and implemented on a cooperative robot with six degrees of freedom [10]. Daniele et al. proposed an attitude graph method to address the hand–eye calibration problem by leveraging the underlying graphic structure of the system, thus achieving global calibration of multiple robot hand–eye systems [11]. Wang proposed an algorithm based on bifurcation diagram analysis to investigate the system’s dynamic response under varying damping and stiffness parameters, effectively capturing stable symmetric orbits as well as periodic and chaotic motions [12]. Dominic et al. proposed a Kronecker–Product method to calibrate a dual-arm hand–eye system using a robot tracking tool by separating translation and rotation [13]. Erika et al. presented a multi-camera framework for collaborative robot vision servos, which enhanced the system’s robustness and efficiency [14]. Xu et al. developed a dual-robot calibration system using tooling components and established a mapping between the distance from reference points to the robot end-effectors and the base parameters, thereby accurately determining the relative pose between the two robots [15]. Davide et al. proposed a multi-camera hand–eye calibration method that optimizes the pose of each camera relative to the robot base and each other’s cameras [16]. Oliveira et al. proposed a novel calibration method for multi-sensor, multi-modal robot systems, utilizing the topological representation of coordinate systems to achieve global calibration across heterogeneous sensors [17]. Shahab et al. proposed a new control strategy for mobile manipulator systems (MMSs) that integrates image-based visual servoing (IBVS) to address operational limitations and safety constraints [18]. Zhou et al. proposed an optimization method based on the minimum spanning tree and graph for the calibration problem of multi-eye and multi-hand systems [19]. Wu et al. proposed a multi-sensor fusion calibration method, integrating hand–eye systems with IMUs for joint calibration of robot, camera, and IMU modules [20]. Wang et al. equated the multi-coordinate calibration problem of the dual-robot system to a matrix equation and employed a closed-loop method to obtain a rapid initial estimate for iterative refinement [21]. Wang et al. studied the visual tracking control problem in cooperative dual-robotic-arm systems and designed an adaptive algorithm to estimate the external parameters of the camera [22]. Duan et al. proposed a new novel hand–eye calibration method that combines spatial and out-of-plane constraints to iteratively optimize calibration parameters based on feature point relationships [23]. Yu et al. studied a hand–eye parameter estimation method based on 3D observation of a single marker, achieving global calibration for a large field-of-view 3D vision system [24]. Liu et al. proposed a vision-based synchronous calibration method that relies solely on marker position data, enhancing generalizability and reducing model dependency [25]. Tao et al. proposed a fast base frame calibration method based on the visual system, which can quickly and accurately calculate the relative pose between the hand–eye systems of a moving robotic arm [26].
At present, international researchers have conducted extensive studies on 3D measurement methods for robotic arm hand–eye systems, and several such visual systems have found applications in industrial metrology and intelligent manufacturing. However, under large field-of-view and high-noise conditions, collaborative 3D measurement using dual-arm hand–eye systems still encounters significant challenges in practical applications. Accurate motion pose estimation of the 3D vision system and proper normalization of the coordinate system are crucial for achieving high-precision collaborative 3D measurement with a dual-robotic-arm hand–eye system. The measurement field of view of the 3D vision sensor at the end of the robotic arm varies. In environments with background noise, field-of-view occlusions, and complex lighting, it is difficult to obtain real-time motion pose feedback from the 3D vision sensor. This limitation compromises the flexibility and accuracy required for global calibration, ultimately degrading the measurement performance of the dual-arm hand–eye system. Given the superior flexibility of encoded marker points, employing them for global calibration and multi-view 3D imaging of dual-arm hand–eye systems under non-overlapping fields of view and noisy conditions shows great potential. The main work and innovations of this paper are summarized as follows:
(1)
M3DHE-ICP is proposed to significantly improve the consistency of multi-view point cloud stitching by introducing a refined correction process based on the ICP (iterative closest point) algorithm, which compensates for residual errors caused by calibration and 3D sensor measurement inaccuracies. The average Euclidean distance errors in fitting the center of a standard sphere using the dual-arm hand–eye system in multi-view 3D imaging were 0.082 mm and 0.066 mm, respectively;
(2)
GCM-DHE is proposed to compute the global coordinates of encoded feature points through 3D reconstruction of a binocular motion structure. It further establishes the transformation relationships between the coordinate systems of two robotic arms by referencing encoded feature points and applying the Tsai–Lenz hand–eye calibration optimization algorithm. This method effectively addresses the challenges posed by non-overlapping fields of view and external noise. After calibration, the average error in multi-view 3D measurements using the dual-arm hand–eye system was 0.100 mm, demonstrating the effectiveness of the proposed global calibration approach;
(3)
We conducted a global calibration experiment of the dual-arm hand–eye system under complex noise conditions and compared the results with other calibration methods to verify the robustness of the proposed approach. Subsequently, we carried out a practical 3D measurement experiment on an automobile front fender component using a binocular structured light vision system combined with dual robotic arms. The results showed an average measurement error of 0.085 mm and a standard deviation of 0.018 mm, further demonstrating the accuracy of the proposed method in practical measurement applications.

2. M3DHE-ICP

The principle of M3DHE-ICP is shown in Figure 1. In the process of solving the hand–eye transformation matrix using the Tsai–Lenz algorithm, the relative poses between the robot coordinate system and the calibration plate coordinate system remain unchanged. Therefore, the transformation relationship between the robot coordinate system and the calibration plate coordinate system can be derived as shown in Equation (1).
H C a l i B a s e i = H T o o l i B a s e i · H C a m i T o o l i · H C a l i C a m i
where H C a l i B a s e i denotes the transformation matrix from the robot coordinate system of the i th robotic arm to the coordinate system of the calibration plate, H T o o l i B a s e i denotes the transformation matrix from the robot coordinate system of the i th robotic arm to the flange coordinate system thereof, H C a m i T o o l i denotes the transformation matrix from the flange coordinate system of the i th robotic arm to the camera coordinate system thereof, and H C a l i C a m i denotes the transformation matrix from the camera coordinate system of the i th robotic arm to the coordinate system of the calibration plate.
By controlling the robotic arm’s vision system to acquire images of the calibration plate from multiple poses, transformation matrices can be obtained. Since the matrix H C a m i T o o l i , representing the transformation from the flange coordinate system to the camera coordinate system, remains constant, Equation (2) can be derived from Equation (1).
( H T o o l i B a s e i j ) 1 · H T o o l i B a s e i k · H C a m i T o o l i = H C a m i T o o l i · H C a l i C a m i j · H C a l i C a m i k 1
where H C a m i T o o l i represents the transformation from the end-effector (flange) coordinate system to the camera coordinate system of the ith robotic arm. H T o o l i B a s e i j , H T o o l i B a s e i k , H C a l i C a m i j , and H C a l i C a m i k denote the transformation matrices from the robot coordinate system of the robotic arm to the flange coordinate system and from the camera coordinate system to the coordinate system of the calibrated plate in the j th and k th positional states, respectively.
By converting the matrix equation in Equation (2) into the standard form A X = X B , the hand–eye transformation matrix for each robotic arm’s vision system can be obtained independently. The 3D point clouds from each robotic arm’s hand–eye system can be fused under multi-view conditions using the coordinate transformation equation shown in Equation (3).
Ω B a s e j P i = H B a s e j T o o l i j R H T o o l j c a m j R Ω C a m j P i + H T o o l j c a m j T + H B a s e j T o o l i j T
where Ω B a s e j P i denotes the 3D point coordinates of the point cloud acquired by the j th robotic arm in the i th pose, expressed in its robot coordinate system. Ω C a m j P i denotes the 3D point coordinates of the point cloud acquired by the j th robotic arm in the i th pose, expressed in its camera coordinate system. A denotes the transformation matrix (including rotation and translation) from the robot flange coordinate system to the robot base coordinate system in the jth robotic arm’s hand–eye system at the j th pose. H B a s e j T o o l i j denotes the transformation matrix (including rotation and translation) from the robot flange coordinate system to the robot base coordinate system in the j th robotic arm hand–eye system at the i th pose. H T o o l j cam j R   and   H Tool j c a m j T denote the rotation and translation matrices, respectively, from the camera coordinate system to the robot flange coordinate system in the j th robotic arm’s hand–eye system.
The cumulative motion error of the robotic arm hand–eye system and the correction of the collected 3D point cloud are determined using the optimization function shown in Equation (4).
L o s s 3 d 3 d = a r g min θ , T i = 1 N w i · ( Ω B a s e j i R θ · Ω B a s e j 0 + T 2
where R θ and T denote the rotation and translation matrices obtained during the ICP iterative optimization process for aligning the 3D point cloud Ω B a s e j i (from the ith pose) to the reference point cloud Ω B a s e j 0 (from the 0th pose). w i denotes the weighting matrix that imposes a greater weight on the error in the Z-axis direction.
An iterative optimization model for multi-view 3D point cloud alignment in the robot coordinate system is constructed based on the Ceres solver, using the ICP correction loss function as shown in Equation (4). The hyperparameter settings of the iterative model are as follows: the error threshold in the Z-axis for transformed coordinates after stitching with the target point cloud is set to 0.1 to eliminate the influence of noise points; the weight coefficient w i is set to 20.0; the linear solver type in Ceres is configured as DENSE_QR; the output dimension of the residual is set to 3; and the maximum number of iterations is set to 200.

3. GCM-DHE

The principle of GCM-DHE is shown in Figure 2. The Tsai–Lenz algorithm is first applied to calibrate the hand–eye system of a single robotic arm. Then, the sub-pixel centers and 3D coordinates of the encoded feature points are localized using CSN-BSSCT [27], which generates a global 3D reference point cloud for the reconstruction of binocular motion via Structure from Motion by Binocular Stereo Vision (SFM-BSV). Next, SFM-BSV reconstructs the global 3D structure, synchronizes the calibration of external parameter matrices from multiple camera coordinate systems to the world coordinate system, and combines them with the camera-to-end-effector transformations obtained via the Tsai–Lenz algorithm to realize global calibration across. multiple robot coordinate systems.
The principle of SFM-BSV is illustrated in Figure 3. The left camera of the binocular vision system is selected as the reference coordinate system. The method constructs feature similarity relationships between multi-view encoded marker images in a graph structure and performs clustering of the encoded features. The feature similarity is quantified using a weight function, as defined in Equation (5), to characterize unordered image sets.
i = k = 1 m i l = 1 m j σ f i k , f j l σ f i k , f j l = 1 , i f f i k = f j l 0 , i f f i k f j l
where i denotes edge weight in the sparse graph structure. f i k and f j l denote the features of the i th and j th-coded marker point in the image captured in the k th and   l th pose state, and m i and m j denote the number of features of the coded marker point in the image captured in the i th and j th pose state, respectively.
Based on the feature similarity equation shown in Equation (5), an adjacency-based clustering algorithm (Equation (6)) is designed to determine the center image of each unordered image cluster under different motion poses.
ω i W , A d j ω i = ω i i j α c k C A d j c k = W , C = a r g min C C
where A d j c k denotes the set of encoded feature images in the k th similarity clustering cluster, W denotes the set of unordered images of encoded marker points, ω i denotes the feature image of the encoded marker points in the i th motion position state, i j denotes the feature similarity weight between the i th and j th images, C denotes the set of images of clustering centers capable of covering all nodes, and C denotes the number of clustering centers minimized.
By constructing matched pairs of encoded feature points, spatial constraint equations in the binocular vision system can be derived as shown in Equation (7).
u l m l 31 m l 11 x w + u l m l 32 m l 12 y w + u l m l 33 m l 13 z w = m l 14 u l m l 34 v l m l 31 m l 21 x w + v l m l 32 m l 22 y w + v l m l 33 m l 23 z w = m l 24 v l m l 34 u r m r 31 m r 11 x w + u r m r 32 m r 12 y w + u r m r 33 m r 13 z w = m r 14 u r m r 34 v r m r 31 m r 21 x w + v r m r 32 m r 22 y w + v r m r 33 m r 23 z w = m r 24 v r m r 34
where m l i j and m r i j denote the elements in the left and right camera projection matrices; ( u l , v l , 1 ) and u r , v r , 1 denote the chi-square coordinates of the encoded feature points in the left and right camera image coordinate system.
Based on the local partition clustering strategy defined in Equation (6), the center image of each local cluster of unordered images can be obtained. The coordinate system of the center images in multiple local clusters is used as the reference frame. The initial relative poses of other cluster nodes with respect to the cluster center are estimated using the PnP algorithm.
The feature similarity of each cluster is evaluated using the weight function from Equation (5). The two clusters with the highest similarity weights are selected as the initial matching module. An incremental global splicing model is then constructed using LM iterative optimization to reconstruct the 3D coordinates of encoded marker points in local clusters. The incremental splicing process transforms the 3D point clouds of all encoded marker clusters into the world coordinate system using Equation (8), thereby completing the global reconstruction.
P g l o b a l = R g l o b a l , i * P c e n t e r i + T g l o b a l , i * , P c e n t e r i ϵ P c e n t e r 1 , P c e n t e r 2 , , P c e n t e r n
where P g l o b a l denotes the global 3D point cloud of the encoded marker points in the large field of view; R g l o b a l , i * and T g l o b a l , i * denote the rotation and translation matrices of the i th cluster to the world coordinate system after iterative optimization by the LM algorithm, respectively; and P c e n t e r i denotes the 3D coordinates of the encoded marker points in the i th local cluster.
The spatial constraints among the image coordinate system, camera coordinate system, and world coordinate system in the dual-robotic-arm hand–eye system are established by encoding the position information of marker points in both the camera and image coordinate system, as shown in Equations (9) and (10).
r e s i d u a l u i = u i u i ^ f x r e s i d u a l v i = v i v i ^ f y E r e p r o j = i r e s i d u a l u i 2 + r e s i d u a l v i 2
where r e s i d u a l u i and r e s i d u a l v i denote the reprojection errors of the ith encoded marker point along the u and v directions in the image coordinate system. u i   and   v i are the observed subpixel coordinates, while u i ^   and   v i ^ are the reprojected subpixel coordinates obtained using the transformation matrix H C a m i w o r l d .
e r r o r e u c l i = X w i X c w i 2 + Y w i Y c w i 2 + Z w i Z c w i 2 E e u c l = i e r r o r e u c l i 2
where e r r o r e u c l i denotes the Euclidean distance error of the feature point, P w i ( X w i , Y w i , Z w i ) denotes the 3D coordinates of the encoded marker point in the world coordinate system, and P c w i ( X c w i , Y c w i , Z c w i ) denotes the 3D coordinates of the encoded marker point transformed from the camera coordinate system to the world coordinate system.
The 3D–3D Euclidean distance error is more strongly constrained to the absolute distance of the translation component, while the 3D–2D projection error is more sensitive to the direction of rotation and translation. Using the LM algorithm for iterative optimization of Equations (9) and (10), the position matrix between the corrected world coordinate system and the camera coordinate system can be solved for H C a m i w o r l d . Using the transformation relationships among the coordinate systems of the dual robotic arms (as illustrated in Figure 2), Equation (2) is employed to obtain the position transformation matrices for each hand–eye system, specifically H T o o l 1 c a m 1   and   H T o o l 2 c a m 2 . Subsequently, the transformation matrix between the robot coordinate systems of arms No. 1 and No. 2 is computed using Equation (11).
H B a s e 2 B a s e 1 = H T o o l 1 B a s e 1 · H c a m 1 T o o l 1 · H w o r l d c a m 1 · H c a m 2 w o r l d · H T o o l 2 c a m 2 · H B a s e 2 T o o l 2
where H B a s e 2 B a s e 1 denotes a transformation relationship between the robot coordinate system of the robotic arm No. 1 to the robotic arm No. 2, H C a m i T o o l i denotes a transformation matrix from the flange coordinate system of the i th robotic arm to its camera coordinate system, H B a s e i T o o l i denotes a transformation matrix from the flange coordinate system of the i th robotic arm to its robot coordinate system, and H C a m i w o r l d denotes a transformation matrix from the world coordinate system to the camera coordinate system of the i th robotic arm.

4. Experiments

4.1. Multi-View 3D Imaging of the Hand–Eye System with M3DHE-ICP

To evaluate the accuracy of the binocular structured light 3D vision system installed on the robotic arm, this study used a matte ceramic standard sphere (as shown in Figure 4a) as the measurement target. The binocular structured light vision system is shown in Figure 4b. As the perception module at the end of the robotic arm, the resolution of the camera is 1280 × 1024 pixels. The experiment adopted the improved three-wavelength six-step phase shift method to measure the three-dimensional shape of the standard sphere. The diameter of the matte ceramic standard ball A is 50.807 mm, and the diameter of the matte ceramic standard ball B is 19.998 mm.
The three-dimensional measurement results of a standard sphere using the binocular structured light vision system are shown in Figure A1. The No. 1 binocular structured light vision system achieved an average diameter fitting error of 0.022 mm for standard sphere A and 0.047 mm for standard sphere B. For the No. 2 system, the average diameter fitting errors were 0.018 mm for standard sphere A and 0.029 mm for standard sphere B. In summary, the average measurement error of the binocular structured light vision system in the dual-robotic-arm hand–eye system was 0.029 mm.
Two robotic arms equipped with binocular structured light vision systems were used separately for multi-view collaborative 3D measurements. To evaluate the multi-view 3D imaging accuracy of each hand–eye system, an error analysis experiment was conducted on the best-fit center coordinates of the standard sphere across multiple views, as shown in Figure A2. The fitted spherical center coordinates of the first group of 3D data were used as a reference. Each subsequent group of spherical center coordinates was corrected and aligned in the robot coordinate system. The corresponding center distance errors before and after correction are shown in Figure 5 and Figure 6. As shown in Figure 5, before correction, the No. 1 robotic arm’s hand–eye system achieved an average Euclidean distance error of 0.471 mm for target A and 0.352 mm for target B. After correction, the average Euclidean distance errors were reduced to 0.076 mm for target A and 0.088 mm for target B. Compared with the pre-correction values, these represent reductions of 0.395 mm and 0.264 mm, respectively. As shown in Figure 6, prior to correction, the No. 2 robotic arm’s hand–eye system produced an average Euclidean distance error of 0.553 mm for target A and 0.592 mm for target B. After correction, the average Euclidean distance errors were 0.071 mm for target A and 0.061 mm for target B, representing reductions of 0.482 mm and 0.531 mm, respectively, compared to the uncorrected values. In summary, the average Euclidean distance errors of the spherical center fitting for multi-view 3D imaging were 0.082 mm and 0.066 mm for the hand–eye systems of robotic arms No. 1 and No. 2, respectively.

4.2. Three-Dimensional Scanning Imaging Tests of Multi-View Fusion for GCM-DHE

The hardware system for the collaborative three-dimensional measurement of the dual-robotic-arm hand–eye system is shown in Figure A3. Firstly, the BSSCT feature points in the non-overlapping field of view were collected through the binocular vision scanner as shown in Figure A3a, and the BSSCT feature detection was carried out by the CSN-BSSCT dynamic tracking module in GCM-DHE. The global reference three-dimensional point cloud was constructed through SFM-BSV. Then, through the dual-robotic-arm hand–eye system as shown in Figure A3b, the BSSCT feature images were collected as in the scene shown in Figure A3c. Finally, the dual-arm hand–eye system was calibrated through the Tsai–Lenz algorithm to achieve multi-view three-dimensional measurement of the dual-arm hand–eye system.
GCM-DHE has the characteristics of a non-overlapping field of view and robustness. Supported by the NVIDIA GeForce RTX 3070 image processor, the BSSCT feature corner points extracted by GCM-DHE through CSN-BSSCT in complex scenes are shown in Figure A4, and the detection rate of each image was approximately 105.646 ms. The results of calibration of the dual-robotic-arm hand–eye system by GCM-DHE are shown in Figure A5. The calibration environment with multi-noise interference is shown in Figure A5a. Figure A5b displays the global reference 3D coordinate construction of coded marker points using SFM-BSV. The three-dimensional coordinates of the global point cloud of BSSCT constructed by the SFM-BSV module in GCM-DHE are shown in Table A1. Figure A5c presents the multi-view 3D measurement results of the standard sphere by the dual-robotic-arm hand–eye system after global calibration. Figure A5 shows that BSSCT can be detected rapidly under complex noise conditions through GCM-DHE, and the high-precision global calibration of the hand–eye system of the dual mechanical arms can be completed.
To verify the accuracy of the proposed method, feature images of the BSSCT were acquired using the binocular structured light vision system under three different motion poses of robotic arms No. 1 and No. 2, respectively, in order to compute the relative transformation between their coordinate systems. The 3D point cloud data acquired by the hand–eye system of robotic arm No. 1 from multiple viewpoints were transformed into the robot coordinate system of robotic arm No. 2. The average errors of the best-fit spherical center coordinates for these transformed point clouds are listed in Table 1. As shown in Table 1, the proposed global calibration algorithm enabled accurate multi-view 3D measurement of the standard sphere. The average fitting errors of the spherical center along the X-axis, Y-axis, and Z-axis were 0.044 mm, 0.028 mm, and 0.069 mm, respectively. Under three different motion pose groups, the Euclidean distance errors of the spherical centers from the multi-view 3D point clouds—transformed from the No. 1 to the No. 2 robot coordinate system—are given in Table 2. As shown in Table 2, the average Euclidean distance errors to the spherical center coordinates were 0.110 mm, 0.125 mm, and 0.065 mm for the first, second, and third motion pose states, respectively. The overall average error across all groups was approximately 0.100 mm.
To compare the measurement accuracy of multi-view 3D imaging under different global calibration methods, a high-precision rotary table and a binocular structured light vision system (Figure A6) were used to perform 3D measurements of standard sphere A. The corresponding results are shown in Figure A6. In the overlapping field of view of the dual-robotic-arm hand–eye system, the two robot coordinate systems were globally calibrated using a circular calibration plate and the RANSAC-PnP algorithm for multi-view 3D measurements of target A. Across all motion poses of the robotic arm hand–eye system, the spherical centers obtained from multi-view 3D imaging of standard sphere A were fitted using different global calibration methods.
The accuracy comparison of multi-viewpoint three-dimensional imaging of target A by different global calibration methods is shown in Table 3. As shown in Table 3, compared with the rotary table scanning, handheld 3D scanner, and circular target method, the average spherical center fitting errors along the X-axis, Y-axis, and Z-axis of the method proposed in this paper were reduced by 0.064 mm, 0.074 mm, 0.004 mm, 0.072 mm, 0.048 mm, and 0.048 mm, respectively. Compared with the turntable method, the average Euclidean distance error of this method was reduced by 0.088 mm. Compared with the handheld 3D scanner method, the average Euclidean distance error of this method was reduced by 0.112 mm. Compared with the circular target method, the average Euclidean distance error was reduced by 0.456 mm. To sum up, the multi-view imaging method of the dual-robotic-arm hand–eye system proposed in our method has good global calibration robustness and high multi-view three-dimensional measurement accuracy under complex noise interference conditions. Since the binocular structured light system is a fixed-focus vision system, although GCM-DHE has strong robustness, when the depth of field of the camera exceeds the range, or the camera is in an out-of-focus state, it will lead to the loss of global calibration accuracy or calibration failure.

4.3. Actual Measurement Experiments of the Hand–Eye System with Dual Mechanical Arms

In order to evaluate the practical performance of the multi-view three-dimensional imaging method using the dual-arm hand–eye system, we measured the three-dimensional surface topography of the automotive part shown in Figure 7 through the dual-arm hand–eye system measurement platform shown in Figure A3.
Figure 7a shows the front mudguard of the car to be tested, and Figure 7b shows the constructed dual-arm hand–eye system measurement platform. In the measurement experiment, the global calibration data of the hand–eye system of the dual mechanical arms are shown in Table A2. The three-dimensional point clouds collected from multiple perspectives are shown in Figure 8. Figure 8a shows the three-dimensional point clouds captured from multiple views and the global three-dimensional point clouds in each robot coordinate system, while Figure 8b presents the global three-dimensional point clouds in the world coordinate system. The error evaluation of the three-dimensional measurement results is shown in Figure 9. Figure 9a shows the global fusion point cloud optimized by ICP, and Figure 9b shows the average error distribution between the global three-dimensional point cloud and the standard model.
A coded marker-assisted global calibration method (GCM-DHE) for a dual-robotic-arm hand–eye system is proposed, combined with a multi-view 3D scanning method based on ICP correction (M3DHE-ICP). This approach effectively addresses the issues of inaccurate multi-coordinate system calibration and accumulated multi-view stitching errors under complex noise conditions. The GCM-DHE method establishes a robust global calibration relationship between the two robotic arms, and when combined with M3DHE-ICP, it significantly improves the accuracy of multi-view 3D measurements. In experiments, the spherical centroid fitting error of standard sphere imaging was controlled within 0.100 mm, and the average measurement error for the automobile front fender was 0.085 mm with a standard deviation of 0.018 mm, demonstrating the high accuracy and stability of the proposed method. The research results have important engineering application value and promotion prospects for multi-robot collaborative measurement and high-precision 3D reconstruction of complex workpieces in intelligent manufacturing environments.

5. Conclusions

A coded marker-assisted global calibration method (GCM-DHE) for a dual-robotic-arm hand–eye system is proposed, combined with a multi-view 3D scanning method based on ICP correction (M3DHE-ICP). This approach effectively addresses the problems of inaccurate multi-coordinate system calibration and accumulated multi-view stitching errors under complex noise conditions. The GCM-DHE method establishes a robust global calibration relationship between the two robotic arms. When combined with M3DHE-ICP, it significantly improves the accuracy of multi-view 3D measurements. In experiments, the spherical centroid fitting error for standard sphere imaging was controlled within 0.100 mm, verifying the high accuracy and stability of the proposed method. The research results have important engineering application value and promotion prospects for multi-robot collaborative measurement and high-precision 3D reconstruction of complex workpieces under an intelligent manufacturing environment. Based on this, in the subsequent research, the autonomous planning of the cooperative scanning path of the dual-robotic-arm hand–eye system will be the research goal.

Author Contributions

T.Z., validation and writing—original draft; X.F., investigation, writing—original draft, and writing—review and editing; S.W., writing—review and editing, conceptualization, and supervision; H.H., writing—review and editing; S.L., writing—original draft and writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 62303246).

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Appendix A

Figure A1. Three-dimensional measurement of a standard sphere using a binocular structured light vision system. (a) Deformation fringes of the standard sphere; (b) absolute phase map of the deformation fringes; (c) reconstructed 3D shape of the standard sphere; (d) spherical fitting error of the 3D measurement data.
Figure A1. Three-dimensional measurement of a standard sphere using a binocular structured light vision system. (a) Deformation fringes of the standard sphere; (b) absolute phase map of the deformation fringes; (c) reconstructed 3D shape of the standard sphere; (d) spherical fitting error of the 3D measurement data.
Micromachines 16 00809 g0a1
Figure A2. Optimal sphere coordinate fitting for multi-view 3D imaging of standard sphere by robotic-arm hand–eye system. (a) Calibration image of the hand–eye system; (b) 3D measurement results of the standard sphere using the binocular structured light vision system in multi-view; (c) multi-view 3D imaging of the standard sphere in the robotic coordinate system; (d) multi-view 3D point cloud data of the standard sphere after noise removal; (e) fitting results of the optimal sphere for the multi-view point cloud data.
Figure A2. Optimal sphere coordinate fitting for multi-view 3D imaging of standard sphere by robotic-arm hand–eye system. (a) Calibration image of the hand–eye system; (b) 3D measurement results of the standard sphere using the binocular structured light vision system in multi-view; (c) multi-view 3D imaging of the standard sphere in the robotic coordinate system; (d) multi-view 3D point cloud data of the standard sphere after noise removal; (e) fitting results of the optimal sphere for the multi-view point cloud data.
Micromachines 16 00809 g0a2
Figure A3. The hardware system for collaborative three-dimensional measurement of the hand–eye system of dual mechanical arms. (a) Dynamic three-dimensional tracking of BSSCT; (b) cooperative three-dimensional scanning of the hand–eye system of dual robotic arms; (c) global calibration of dual robotic arms for encoded marker points.
Figure A3. The hardware system for collaborative three-dimensional measurement of the hand–eye system of dual mechanical arms. (a) Dynamic three-dimensional tracking of BSSCT; (b) cooperative three-dimensional scanning of the hand–eye system of dual robotic arms; (c) global calibration of dual robotic arms for encoded marker points.
Micromachines 16 00809 g0a3
Figure A4. An example diagram of robustness detection of BSSCT feature points under complex noise conditions.
Figure A4. An example diagram of robustness detection of BSSCT feature points under complex noise conditions.
Micromachines 16 00809 g0a4
Figure A5. Global calibration results of GCM-DHE combined with Tsai–Lenz for a dual-robotic-arm hand–eye system. (a) Calibration environment under multi-noise interference; (b) construction of global reference 3D coordinates from BSSCT feature points using GCM-DHE; (c) multi-view 3D measurements of a standard sphere after global calibration.
Figure A5. Global calibration results of GCM-DHE combined with Tsai–Lenz for a dual-robotic-arm hand–eye system. (a) Calibration environment under multi-noise interference; (b) construction of global reference 3D coordinates from BSSCT feature points using GCM-DHE; (c) multi-view 3D measurements of a standard sphere after global calibration.
Micromachines 16 00809 g0a5
Figure A6. Results of multi-view 3D measurement of standard sphere by high-precision rotary table.
Figure A6. Results of multi-view 3D measurement of standard sphere by high-precision rotary table.
Micromachines 16 00809 g0a6
Table A1. BSSCT reference three-dimensional coordinates for global calibration of the hand–eye system of the dual robotic arms.
Table A1. BSSCT reference three-dimensional coordinates for global calibration of the hand–eye system of the dual robotic arms.
Encode the Serial Number of the Feature PointX (mm)Y (mm)Z (mm)
12684101218-11015-1174.287406.371121.330
12684101218-11015-10154.391401.486119.341
12684101218-11015-15199.009346.389106.283
12684101218-11015-A192.325402.493121.034
12684101218-11015-B206.563381.218115.801
12684101218-11015-C186.408383.694115.827
12684101218-11015-E142.76356.582107.078
12684101218-11015-O210.338398.670120.563
12684101218-5817-17169.153−30.8196.112
12684101218-5817-5222.681−7.00814.042
12684101218-5817-8195.773−0.52714.871
12684101218-5817-A216.184−32.9877.053
12684101218-5817-B193.861−45.9462.947
12684101218-5817-C196.881−26.5208.204
Table A2. Global calibration data of the hand–eye system of the dual robotic arms.
Table A2. Global calibration data of the hand–eye system of the dual robotic arms.
VariableThe Hand–Eye Matrix of Robotic Arm No. 1The Hand–Eye Matrix of Robotic Arm No. 2The Transformation Matrix of the Robot Coordinate System
Category
Rotation[0.133, 0.139, −1.568][−0.123, 0.231, 3.111][−7.582 × 10−3, −3.056 × 10−3, −0.746]
Translation[−96.046, 65.053, 90.697][103.499, 103.559, 72.669][807.320, −1339.438, 20.040]

References

  1. Ahn, S.J.; Rauh, W.; Kim, S.I. Circular coded target for automation of optical 3D-measurement and camera calibration. Int. J. Pattern Recognit. Artif. Intell. 2001, 15, 905–919. [Google Scholar] [CrossRef]
  2. Wang, L.; Zhang, L.; Yu, Z.; Fei, C.; Xiayan, S.; Dongrui, H. Precision circular target location in vision coordinate measurement system. In Proceedings of the Advanced Materials and Devices for Sensing and Imaging III, Beijing, China, 12–14 November 2007; SPIE: Bellingham, WA, USA; Volume 6829, pp. 90–97. [Google Scholar]
  3. Ahn, S.J.; Kotowski, R. Geometric image measurement errors of circular object targets. Opt. 3-D Meas. Tech. IV 1997, 463–471. Available online: https://www.researchgate.net/profile/Sung-Joon-Ahn/publication/272170844_GEOMETRIC_IMAGE_MEASUREMENT_ERRORS_OF_CIRCULAR_OBJECT_TARGETS/links/54dd66fd0cf28a3d93f912c4/GEOMETRIC-IMAGE-MEASUREMENT-ERRORS-OF-CIRCULAR-OBJECT-TARGETS.pdf (accessed on 6 July 2025).
  4. An, G.H.; Lee, S.; Seo, M.W.; Yun, K.; Cheong, W.S.; Kang, S.J. Charuco board-based omnidirectional camera calibration method. Electronics 2018, 7, 421. [Google Scholar] [CrossRef]
  5. Förstner, W.; Gülch, E. A fast operator for detection and precise location of distinct points, corners and centres of circular features. In Proceedings of the ISPRS Intercommission Conference on Fast Processing of Photogrammetric Data, Interlaken, Switzerland, 2–4 June 1987; Volume 6, pp. 281–305. [Google Scholar]
  6. Pan, X.; Ellis, T.J.; Clarke, T.A. Robust tracking of circular features. In Proceedings of the BMVC, Birmingham, UK, September 1995; Volume 95, pp. 553–562. [Google Scholar]
  7. Rocha, R.; Dias, J.; Carvalho, A. Cooperative multi-robot systems: A study of vision-based 3-D mapping using information theory. Robot. Auton. Syst. 2005, 53, 282–311. [Google Scholar] [CrossRef]
  8. Liu, Z.; Zhang, G.; Wei, Z.; Sun, J. A global calibration method for multiple vision sensors based on multiple targets. Meas. Sci. Technol. 2011, 22, 125102. [Google Scholar] [CrossRef]
  9. Wang, Y.; Li, Z.; Zhou, L.; Luo, J.; Han, X.; Wang, X.; Hu, L. Flexible technique for global calibration of multi-view stereo vision with a non-overlapping FOV using linear structured light projector. Opt. Express 2024, 32, 31405–31421. [Google Scholar] [CrossRef] [PubMed]
  10. Lin, C.J.; Wang, H.C.; Wang, C.C. Automatic calibration of tool center point for six degree of freedom robot. Actuators 2023, 12, 107. [Google Scholar] [CrossRef]
  11. Rozlivek, J.; Rustler, L.; Stepanova, K.; Hoffmann, M. Multisensorial robot calibration framework and toolbox. In Proceedings of the 2020 IEEE-RAS 20th International Conference on Humanoid Robots (Humanoids), Munich, Germany, 19–21 July 2021; IEEE: Piscataway, NJ, USA; pp. 459–466. [Google Scholar]
  12. Wang, C.C.; Zhu, Y.Q. Identification and Machine Learning Prediction of Nonlinear Behavior in a Robotic Arm System. Symmetry 2021, 13, 1445. [Google Scholar] [CrossRef]
  13. Evangelista, D.; Olivastri, E.; Allegro, D.; Menegatti, E.; Pretto, A. A graph-based optimization framework for hand-eye calibration for multi-camera setups. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; IEEE: Piscataway, NJ, USA; pp. 1474–11480. [Google Scholar]
  14. Heikkila, T.; Sallinen, M.; Matsushita, T.; Tomita, F. Flexible hand-eye calibration for multi-camera systems. In Proceedings of the 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000)(Cat. No. 00CH37113), Takamatsu, Japan, 31 October–5 November 2000; IEEE: Piscataway, NJ, USA; Volume 3, pp. 2292–2297. [Google Scholar]
  15. Hu, J.; Jones, D.; Valdastri, P. Coordinate calibration of a dual-arm robot system by visual tool tracking. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; IEEE: Piscataway, NJ, USA; pp. 11468–11473. [Google Scholar]
  16. Di Stefano, E.; Ruffaldi, E.; Avizzano, C.A. A multi-camera framework for visual servoing of a collaborative robot in industrial environments. In Proceedings of the 2017 22nd IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Limassol, Cyprus, 12–15 September 2017; IEEE: Piscataway, NJ, USA; pp. 1–8. [Google Scholar]
  17. Mao, J.; Xu, R.; Ma, X.; Hu, S.; Bao, X. Fast calibration method for base coordinates of the dual-robot based on three-point measurement calibration method. Appl. Sci. 2023, 13, 8799. [Google Scholar] [CrossRef]
  18. Heshmati-Alamdari, S.; Sharifi, M.; Karras, G.C.; Fourlas, G.K. Control barrier function based visual servoing for Mobile Manipulator Systems under functional limitations. Robot. Auton. Syst. 2024, 182, 104813. [Google Scholar] [CrossRef]
  19. Oliveira, M.; Pedrosa, E.; de Aguiar, A.P.; Rato, D.F.P.D.; dos Santos, F.N.; Dias, P.; Santos, V. ATOM: A general calibration framework for multi-modal, multi-sensor systems. Expert Syst. Appl. 2022, 207, 118000. [Google Scholar] [CrossRef]
  20. Schmidt, B.; Wang, L. Automatic work objects calibration via a global–local camera system. Robot. Comput.-Integr. Manuf. 2014, 30, 678–683. [Google Scholar] [CrossRef]
  21. Zhou, Z.; Ma, L.; Liu, X.; Cao, Z.; Yu, J. Simultaneously calibration of multi hand-eye robot system based on graph. IEEE Trans. Ind. Electron. 2023, 71, 5010–5020. [Google Scholar] [CrossRef]
  22. Wu, J.; Wang, M.; Jiang, Y.; Yi, B.; Fan, R.; Liu, M. Simultaneous hand–eye/robot–world/camera–IMU calibration. IEEE/ASME Trans. Mechatron. 2021, 27, 2278–2289. [Google Scholar] [CrossRef]
  23. Wang, G.; Li, W.; Jiang, C.; Zhu, D.H.; Xie, H.; Liu, X.J.; Ding, H. Simultaneous calibration of multicoordinates for a dual-robot system by solving the AXB= YCZ problem. IEEE Trans. Robot. 2021, 37, 1172–1185. [Google Scholar] [CrossRef]
  24. Wang, Y.; Liu, Z.; Zhang, Y. Visual servoing for cooperative dual-manipulator system. In Proceedings of the 2015 34th Chinese Control Conference (CCC), Hangzhou, China, 28–30 July 2015; IEEE: Piscataway, NJ, USA; pp. 3158–3163. [Google Scholar]
  25. Liu, Z.; Liu, X.; Duan, G.; Tan, J. Precise hand–eye calibration method based on spatial distance and epipolar constraints. Robot. Auton. Syst. 2021, 145, 103868. [Google Scholar] [CrossRef]
  26. Jin, G.; Yu, X.; Chen, Y.; Li, J. Hand-eye parameter estimation based on 3D observation of a single marker. IEEE Trans. Instrum. Meas. 2024, 18, 31405–31421. [Google Scholar]
  27. Song, L.; Zheng, T.; Li, Y.; Huang, H.; Yang, Y.; Zhu, X.; Zhang, Z. A novel dynamic tracking method for coded targets with complex background noise. Opt. Lasers Eng. 2025, 184, 108654. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of M3DHE-ICP.
Figure 1. Schematic diagram of M3DHE-ICP.
Micromachines 16 00809 g001
Figure 2. Multi-view 3D point cloud fusion method for dual-arm hand–eye system.
Figure 2. Multi-view 3D point cloud fusion method for dual-arm hand–eye system.
Micromachines 16 00809 g002
Figure 3. Schematic diagram of SFM-BSV.
Figure 3. Schematic diagram of SFM-BSV.
Micromachines 16 00809 g003
Figure 4. Equipment used in the 3D measurement experiment with the binocular structured light vision system: (a) standard sphere measurement target; (b) binocular structured light vision system and its components.
Figure 4. Equipment used in the 3D measurement experiment with the binocular structured light vision system: (a) standard sphere measurement target; (b) binocular structured light vision system and its components.
Micromachines 16 00809 g004
Figure 5. Euclidean distance errors of robotic arm hand–eye system No. 1 before and after correction for each set of spherical center coordinates in the multi-view fitting.
Figure 5. Euclidean distance errors of robotic arm hand–eye system No. 1 before and after correction for each set of spherical center coordinates in the multi-view fitting.
Micromachines 16 00809 g005
Figure 6. Euclidean distance errors of robotic arm hand–eye system No. 2 before and after correction for each set of spherical center coordinates in the multi-view fitting.
Figure 6. Euclidean distance errors of robotic arm hand–eye system No. 2 before and after correction for each set of spherical center coordinates in the multi-view fitting.
Micromachines 16 00809 g006
Figure 7. Three-dimensional measurement experiment of the car front fender using the dual-arm hand–eye system. (a) Car front fender; (b) dual-arm hand–eye system 3D measurement platform.
Figure 7. Three-dimensional measurement experiment of the car front fender using the dual-arm hand–eye system. (a) Car front fender; (b) dual-arm hand–eye system 3D measurement platform.
Micromachines 16 00809 g007
Figure 8. Multi-view 3D data acquisition and global point cloud. (a) Multi-view point cloud data collected by the dual-arm hand–eye system; (b) global 3D point cloud.
Figure 8. Multi-view 3D data acquisition and global point cloud. (a) Multi-view point cloud data collected by the dual-arm hand–eye system; (b) global 3D point cloud.
Micromachines 16 00809 g008
Figure 9. Accuracy analysis of 3D data acquisition for the car front fender by the dual-arm hand–eye system. (a) Global 3D point cloud corrected by ICP; (b) error mapping distribution map of the global 3D point cloud.
Figure 9. Accuracy analysis of 3D data acquisition for the car front fender by the dual-arm hand–eye system. (a) Global 3D point cloud corrected by ICP; (b) error mapping distribution map of the global 3D point cloud.
Micromachines 16 00809 g009
Table 1. Mean errors of best-fit spherical center coordinates for 3D measurement of the standard sphere using the dual-robotic-arm hand–eye system.
Table 1. Mean errors of best-fit spherical center coordinates for 3D measurement of the standard sphere using the dual-robotic-arm hand–eye system.
Mean Error of Fitted Spherocentric Coordinates (mm)/GroupΔXΔYΔZ
10.0220.0360.040
20.0730.0290.062
30.0360.0180.104
All groups0.0440.0280.069
Table 2. Euclidean distance fitting errors of spherical center coordinates for 3D measurements of the standard sphere using the dual-robotic-arm hand–eye system.
Table 2. Euclidean distance fitting errors of spherical center coordinates for 3D measurements of the standard sphere using the dual-robotic-arm hand–eye system.
Mean Error of Fitted Spherical Center Distance (mm)/Group3d Point Cloud of the 1st Motion Pose3d Point Cloud of the 2nd Motion Pose3d Point Cloud of the 3rd Motion Pose
10.0800.0660.044
20.0970.1470.092
30.1530.1630.059
All groups0.1100.1250.065
Table 3. Comparison of multi-view 3D imaging accuracy using different global calibration methods.
Table 3. Comparison of multi-view 3D imaging accuracy using different global calibration methods.
Method/Fitting Mean Error (mm)Rotary Table ScanningEva Handheld ScannerCircular Calibration TargetOurs
ΔX0.1080.1160.3120.044
ΔY0.1020.0760.2810.028
ΔZ0.0650.1170.2700.069
Euclidean distance0.1880.2120.5560.100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, T.; Feng, X.; Wang, S.; Huang, H.; Li, S. A Multi-View Three-Dimensional Scanning Method for a Dual-Arm Hand–Eye System with Global Calibration of Coded Marker Points. Micromachines 2025, 16, 809. https://doi.org/10.3390/mi16070809

AMA Style

Zheng T, Feng X, Wang S, Huang H, Li S. A Multi-View Three-Dimensional Scanning Method for a Dual-Arm Hand–Eye System with Global Calibration of Coded Marker Points. Micromachines. 2025; 16(7):809. https://doi.org/10.3390/mi16070809

Chicago/Turabian Style

Zheng, Tenglong, Xiaoying Feng, Siyuan Wang, Haozhen Huang, and Shoupeng Li. 2025. "A Multi-View Three-Dimensional Scanning Method for a Dual-Arm Hand–Eye System with Global Calibration of Coded Marker Points" Micromachines 16, no. 7: 809. https://doi.org/10.3390/mi16070809

APA Style

Zheng, T., Feng, X., Wang, S., Huang, H., & Li, S. (2025). A Multi-View Three-Dimensional Scanning Method for a Dual-Arm Hand–Eye System with Global Calibration of Coded Marker Points. Micromachines, 16(7), 809. https://doi.org/10.3390/mi16070809

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop