Next Article in Journal
Sub-Gigahertz Path Loss Measurement Campaign in Marine Environment: A Case Study
Previous Article in Journal
Adaptive Dataset Management Scheme for Lightweight Federated Learning in Mobile Edge Computing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Use of Phase-Angle Model for Full-Field 3D Reconstruction under Efficient Local Calibration

1
Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen 518055, China
2
Tsinghua-Berkeley Shenzhen Institute, Tsinghua University, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(8), 2581; https://doi.org/10.3390/s24082581
Submission received: 28 March 2024 / Revised: 15 April 2024 / Accepted: 15 April 2024 / Published: 18 April 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Currently, 3D reconstruction methods in structured light are generally implemented in a pre-calibrated area. To realize a full-field reconstruction, the calibration plate can be moved to multiple positions in a time-consuming manner, or the whole field can be calibrated with the help of a large calibration plate, which is more costly. In this paper, we address this problem by proposing a method for obtaining a global phase-angle model under a locally calibrated region, and based on this relationship, we investigate and analyze the reconstruction inside and outside of the calibrated zone. The results show that the method can reconstruct the object outside of the calibration zone completely, and can keep the planarity error around 0.1 mm and the sphericity error below 0.06 mm. The method only requires local calibration of the projected fringes at the two calibration positions to realize the 3D reconstruction of the full-field, which makes the method more advantageous.

1. Introduction

Fringe projection profilometry (FPP) has emerged as a powerful and versatile technique in the field of optical metrology and surface profilometry. Over the years, it has gained significant attention due to its high-accuracy [1,2], high-speed [3,4], and high-resolution [5,6] capabilities, enabling the accurate measurement of 3D shapes and surface topography. This technology has been widely applied in industrial measurement [7], cultural relic protection [8], surgical medicine [9], artificial intelligence [10], and other fields [11,12]. Its implementation process mainly consists of the following steps: the projector projects regular sinusoidal fringe patterns; then, the camera captures the modulated fringe patterns, and after phase calculation, 3D reconstruction is achieved with the help of system calibration parameters [13].
Since the calibration of a FPP system determines in which form the 3D reconstruction is carried out, it is closely related to the accuracy of the reconstruction and can affect the range of measurement to a certain extent. Different calibration strategies [14,15,16,17,18] have been widely and deeply investigated. Currently, there are mainly two classical calibration technology routes, including the stereo-vision-based calibration method and the phase-coordinate mapping model-based calibration method [16]. In the stereo-vision-based FPP system, the projector is regarded as an inverse camera [19,20], so it can search for matching points in a binocular framework. However, this type of calibration method is not applicable to the current emerging structured light projection devices, such as MEMS (micro-electro-mechanical system)-mirror-based projectors, that can only project unidirectional sinusoidal fringes because they need to determine the pixel positions of the markers of the calibration plate in the DMD image with the help of the orthogonal absolute phase. In addition, this type of method requires more computational resources for homologous point matching, which limits the extended use of this method. In the phase-coordinate mapping model [14,15,21,22], the coordinates X, Y, and Z can be directly obtained by bringing the absolute phase into a pre-calibrated function, which does not require orthogonal phases and thus is still applicable in FPP systems composed of MEMS-mirror-based projectors.
However, in the phase-coordinate mapping model phase, the direct independent variable generated by 3D coordinates needs to be pre-calibrated with points at known positions in space. Initially, the coordinates of these known points were represented in the world coordinate system, which meant that precise translation stages were needed to control the transformation of their coordinates, and these points were clearly always on the calibration board. Some existing methods [14,15] have relaxed restrictions and do not require precision translation stages or gage blocks for calibration. The coordinates of these points appear in the camera coordinate system, and they need to provide raw calibration data for the fitting of the phase-coordinate mapping model together with the phase of the corresponding pixels on the calibration board plane. After analysis, it can be found that both the phase and coordinates need to be located on the calibration board to participate in the calibration of the mapping model. This means that a reliable mapping model cannot be formed outside of the coverage range of the calibration board, which leads to the failure of 3D reconstruction.
Recently, researchers have proposed new characterization models to reveal the physical processes of unidirectional fringes projection, including plane models and ray models, etc. Yang et al. [17] have proposed a seven-parameter curved light surface model for unidirectional fringe projection to characterize non-ideal projections. However, this method requires the checkerboard to be large enough to allow all 1024 curved light surfaces to be projected onto it to obtain the 3D coordinates of enough spatial points, which means that a small calibration plate cannot efficiently create a global light surface model. Miao et al. [14] have proposed a new isophase plane model, which evolves into a reciprocal polynomial mapping model by merging calibration parameters. A look-up table (LUT) approach is used to achieve fast and highly accurate measurements; however, the coverage of the LUT is still limited to the calibration plate. Similar LUT mapping methods used in reference [13] have been tried, which implies that they have the same limitations. Wan et al. [18] have designed a large-scale calibration strategy based on the method of linear interpolation in the light plane. But, the spatial equation of the remaining light planes can only be calculated when it is between the calibrated light planes. To ensure the reliability of large-scale reconstruction, 11 sets of calibration data are needed to ensure full coverage of the projection range, which is inefficient. In addition to the plane model, the projection process can also be refined into a ray model. Yang et al. [16] have achieved 3D reconstruction using triangulation constructed from both the projected rays and the camera. However, each ray also needs to be constructed in the LUT, which means that the equation for the ray can only be pre-estimated within the local area covered by a calibration board. At present, the calibration strategies in unidirectional fringe projection profilometry mainly focus on the reconstruction effect within the calibration area. Inferring 3D reconstruction information beyond the range of the calibration plate in reconstructing full-field scenes has not yet received in-depth research and focused attention.
The phase-angle model [23] is a novel representation model applied in unidirectional fringe projection profilometry, and it can be used to describe the angle variation relationship between phase and isophase planes in space, providing a new approach for 3D reconstruction. The isophase plane represents a hypothetical plane formed by coordinates in projection space that are different in position but still have the same phase. In structured light systems based on MEMS-mirror-based projectors, the isophase plane can be understood as a plane formed by the reflection of laser vertical lines through the MEMS mirror. In the system composed of DLP projectors, the isophase plane is located in the direction reflected by a column of small micro mirrors in the array. After obtaining the phase-angle model, the position of the current phase plane can be located through phase positioning, and then the 3D coordinates can be obtained by combining the geometric constraints of the camera perspective instead of using phase as the direct independent variable to participate in coordinates generation. This approach avoids the problem of 3D coordinate-generation errors when the phase-coordinate mapping model is not fully fitted.
In this paper, a reconstruction method using a phase-angle model to infer 3D information outside of the calibration zone is proposed. It does not require the calibration plate to be large, but requires that the calibration plate positions of two projection fringes have overlapping areas to construct the relationship between the phase and the deflection angle of the isophase plane, which is then used to locate the isophase plane within the projection space rather than only within the calibration plate area. This endows the 3D reconstruction capability beyond the calibration zone. We have validated the feasibility of the method through experiments, comparing the reconstruction inside and outside of the calibration zone using similarity in reconstructed morphology and reconstruction accuracy. We have also compared the proposed method with existing phase-coordinate mapping model methods, demonstrating its advantages in reconstructing 3D information outside of the calibration zone. In addition, by reducing the size of the calibration area and comparing the reconstruction accuracy vertically, the results show that this method still has robust 3D reconstruction ability.

2. Principle

2.1. Camera Imaging Model

The camera’s imaging mechanism operates through perspective projection. During forward projection, the projection center casts points from the object onto the imaging plane, producing the camera’s image. In contrast, 3D reconstruction involves the inverse process of back-projecting image points to pinpoint their corresponding spatial coordinates. This process can be expressed mathematically as follows:
s c u v 1 = K c x c y c z c 1 = K c R T 0 1 x w y w z w 1 , K c = f u γ u 0 0 0 f v v 0 0 0 0 1 0 .
where the coordinates of point P in the world coordinate system (WCS) and camera coordinate system (CCS) are (xw, yw, zw) and (xc, yc, zc), respectively; rotation matrix R and a translation vector T represent the rigid body transformation between the WCS and the CCS; sc stands for the scaling factor; Kc are the camera’s internal parameters, where fu and fv, respectively, denote the equivalent focal lengths in the horizontal and vertical directions; γ denotes the skew factor between the u axis and v axis in the imaging plane; and (u0, v0) represents the principal point of the pixel plane. When we use (k1, k2, k3) to denote the radial aberration coefficients, (p1, p2) to denote the tangential aberration coefficients, ( u ˜ d , v ˜ d ) and ( u ˜ , v ˜ ) to denote the normalization terms for the actual imaging point ( u d , v d ) and the ideal imaging point (u, v), respectively, we can take the lens aberration into account with the following equations:
u ˜ d = u ˜ ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 ) + [ 2 p 1 u ˜ v ˜ + p 2 ( r 2 + 2 u ˜ 2 ) ] v ˜ d = v ˜ ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 ) + [ 2 p 2 u ˜ v ˜ + p 1 ( r 2 + 2 v ˜ 2 ) ] r 2 = u ˜ 2 + v ˜ 2 .

2.2. Basis of the Phase-Angle Model

In Figure 1, if the calibration board cannot fully cover the projection range in practical application scenarios for the reconstruction method of phase-coordinate mapping model, only the local calibration zone can participate in the reconstruction, which limits its reconstruction range ability for systems with large projection or field of view ranges. The phase-angle model mentioned below can cover all points in the projection field, which means that in the space outside of the calibration area, the phase-angle model can infer and assist in solving reconstruction problems cleverly.
It is worth noting that the definition of the deflection angle of the isophase plane requires a reference position, which can be taken arbitrarily but is fixed after the system calibration is completed. As shown in Figure 1, the green dashed line represents the position of the reference isophase plane, corresponding to phase φref, and its position can be arbitrarily taken within the projection range. Assuming the working distance of the projector is d, let Ox be the origin of the x axis. The reference phase φref corresponds to x1 on the x axis, and the angle between the reference isophase plane and the straight line AOx is θ1. The current phase φcur corresponds to xcur, and the angle between its isophase plane and the reference isophase plane is θ′. At the same time, its angle with the straight AOx is θ2.
For sinusoidal fringes projected on the orthogonal virtual plane, the linear relationship φ = kx + b is satisfied between the spatial coordinates x and the unwrapped phase φ because the spacing of the fringes is uniformly distributed. Then, it is clear that the reference phase φref and the spatial coordinates x1, as well as the current phase φcur and the corresponding spatial coordinates xcur, satisfy this linear relationship:
x 1 = φ r e f b k x c u r = φ c u r b k .
Based on all of the parameters available, the following relationship between angles and coordinates can be found:
tan θ = tan ( θ 1 + θ 2 ) = tan θ 1 + tan θ 2 1 tan θ 1 tan θ 2 tan θ 1 = x 1 d tan θ 2 = x c u r d .
Associating Equations (3) and (4), eliminating θ1, θ2, x1, and xcur, and combining the correlation coefficients yields the equation for the relationship between the deflection angle θ′ and the current phase φcur, which is the algebraic expression for the phase-angle model:
f φ = tan θ = φ c u r φ r e f k d + ( φ r e f b ) ( φ c u r b ) k d = φ c u r φ r e f a 1 φ c u r + a 2 .
where φref is a known quantity, a pre-determined value of a reference phase, and a1 and a2 are the model parameters to be calibrated.
In the process of proposing the phase-angle model, there must theoretically exist a plane perpendicular to the projection direction of the projector, on which the sinusoidal fringes are uniformly distributed. In practical application of the phase-angle model, the position of the calibration plane is not usually guaranteed to be orthogonal, especially when manually changing the position of the calibration plate. At this time, the projected fringes on the inclined calibration plane will appear as changes in the degree of sparsity. As shown in Figure 2a, with respect to the geometrical position of the inclined calibration plane and the orthogonal calibration plane in the system, the projected fringes on the side close to the projector are more tightly packed and have higher spatial frequency, while the fringes on the side far from the projector are sparser and the spatial frequency becomes lower. Then, as the change in spatial frequency occurs it results in the phase on the plane no longer showing a linear change trend. Therefore, it is necessary to prove that the corresponding phases of the same light plane projected on different positions of the inclined calibration plate are still equal, which guarantees the feasibility of the phase-angle model.
In the top view of the position of the calibration plane in Figure 2b, we can assume that there is an angular skew α s 1 between an inclined calibration plane and the orthogonal position, which intersect at a point O 1 , and that the distance from the point O 1 to the projector is L. Within the projection range, there is a projection plane with an angle of θ from the normal projection direction, which intersects with the orthogonal calibration plane at A. In addition, the angle between this projection plane and the orthogonal calibration plane is β , as shown in Figure 2b. The outgoing light plane intersects with the inclined calibration plane at point B, with an angle of O 1 B A = γ . When points A and B are defined on the right side of O 1 , it is specified that |O1B| = xskew and |O1A| = xp, so there exists the following geometric relationship:
β = 90     arctan ( x p L ) γ = 90     arctan ( x p L )     α s 1 x p sin γ = x s k e w sin ( 180 β ) .
By combining Equation (6), the equation relationship of xskew with respect to xp can be obtained:
x s k e w = x p cos α s 1 x p L sin α s 1 .
when A and B are on the left side of O 1 , it is specified that |O1B| = −xskew and |O1A| = −xp, and the same derivation method can be used to obtain the following:
x s k e w = x p cos α s 1 + x p L sin α s 1 ,
Equation (8) is expressed in the same form as Equation (7), so that the relationship between the transverse coordinate xp on the orthogonal calibration plane and the transverse coordinate xskew on the inclined calibration plane are related only to the angular skew α s 1 and the working distance L.
Assuming there is a (xp, y) point on the orthogonal calibration plane, the phase principal value can be obtained by solving the phase using the phase shift method at that point:
ϕ   =   arctan i = 1 N B cos ( 2 π f 0 x p + δ i ) sin [ 2 π ( i 1 ) / N ] i = 1 N B cos ( 2 π f 0 x p + δ i ) cos [ 2 π ( i 1 ) / N ] .
Among them, N represents the number of phase shift steps, B represents the modulation intensity, f0 is the spatial frequency of sinusoidal fringes, and δi is the phase shift increment of the i-th phase shift graph, which can be expressed as δi = 2π (i − 1)/N + φ0, where φ0 represents the initial phase that can be artificially specified.
On the inclined calibration plane, according to the mapping relationship of the coordinates, the fringes will be distorted and the spatial frequency will show corresponding changes: f = [cos(αs1) − (xp/L) sin(αs1)] f0. At this point, the variation in light intensity with xskew on the inclined calibration plane can be obtained:
I i ( x s k e w , y ) = c ( x p , y ) [ A + B cos ( 2 π f x s k e w + δ i ) ] ,   i = 1 , 2 , 3 , , N .
where c(xp, y) represents the attenuation factor of the grayscale value that occurs when the position of point (xp, y) corresponds to the inclined calibration plane. According to the calculation method using the phase principal values, the phase principal values on the inclined calibration plane can also be obtained, and f and xskew can be substituted to obtain Equation (11):
ϕ   =   arctan i = 1 N c ( x p , y )   [ A + B cos ( 2 π f x s k e w + δ i ) ] sin [ 2 π ( i 1 ) / N ] i = 1 N c ( x p , y )     [ A + B cos ( 2 π f x s k e w + δ i ) ] cos [ 2 π ( i 1 ) / N ] =   arctan i = 1 N   B cos ( 2 π f 0 x p + δ i ) sin [ 2 π ( i 1 ) / N ] i = 1 N   B cos ( 2 π f 0 x p + δ i ) cos [ 2 π ( i 1 ) / N ] .
It can be seen from above that the phase principal value ϕ′ on xp is equal to the corresponding phase principal value ϕ′ on xskew. The corresponding absolute phases φ and φ′ are also equal, and the two cause phase mismatch due to factors such as projection distance, tilt angle, and intensity attenuation factor. It can be concluded that when the fringe frequency changes due to the tilt of the calibration plane in the projection space, the corresponding phases of the same light plane projected at different positions remain the same, and the phase-angle model is still feasible.

3. System Calibration and 3D Reconstruction

3.1. Local Calibration

In order to obtain the phase-angle model relationship, the calibration process first needs to obtain the phase samples on the calibration plate. As shown in the system calibration diagram in Figure 3, the calibration plate may only occupy a small part of the projection range, and we need to sample enough absolute phase samples {φ0, φ1, φ2, …} in the limited range.
After that, to find the deflection angle corresponding to these absolute phase samples, it is necessary to locate their corresponding isophase planes. This process requires a total of three operations. Firstly, the pixels equal to the absolute phase samples are found on the two calibration plates, AP and BP, used for calibration. Since the calibration plate cannot completely occupy the camera image, it is necessary to set a mask to find the pixels with the same phase within the specified range. Secondly, according to the homography change relationship between the pixel plane and the calibration plate plane with HA and HB, and the rotation translation RT change relationship between the world coordinate system and the camera coordinate system, the coordinates of the pixel points on the plate are converted into 3D coordinates under the camera coordinate system. This operation can be expressed as the equation:
a p b p 1 = H u p v p 1 X p Y p Z p = R T a p b p 0 1 .
Among them, (up, vp) is the pixel retrieved from the phase map, the phase of which is equal to the given sample phase, and (Xp, Yp, Zp) is the 3D coordinate of the corresponding point on the calibration board. Finally, based on the 3D coordinates of the identified sample points, the solution of the equation for the sampled isophase planes can be completed. This can be achieved by solving a homogeneous linear system of Equation (13):
X i p Y i p Z i p 1 X M p Y M p Z M p 1 A F 0 B F 0 C F 0 D F 0 = 0 .
where A F 0 , B F 0 , C F 0 , and D F 0 represent the equation coefficients of a certain sample’s isophase plane, while the isophase plane space equations of all other absolute phase samples can be solved according to these three operations.
Up to this point, the samples {φ0, φ1, φ2, …} of the absolute phases and the plane equations {F0, F1, F2, …} of their corresponding isophase planes have been obtained and they are all Nφ in number. Here, we may choose the isophase plane F0 as the reference and calculate the Nφ − 1 of angles sample { θ φ 1 , θ φ 2 , …} based on the normal vector between the remaining isophase planes and the reference isophase plane F0. At this point, the phase sample {φ1, φ2, …} and deflection angle samples { θ φ 1 , θ φ 2 , …} are used to solve the model using the least square method, and the calibration of parameters a1 and a2 in the phase-angle model is completed.
In addition, we found that the constructed isophase planes are all deflected around a virtual rotation centerline L0, and that the extrapolation ability of the phase-angle model is reflected in the fact that the actual deflection angle can exceed the calibrated range when the isophase planes are based on this rotation centerline. Therefore, it is necessary to determine the position of the rotation centerline. Here, a vector d = (v1, v2, v3) and a point p = (p1, p2, p3) are used to uniquely determine the rotation centerline, where vector d represents the direction of the line and point p represents any point on the line. Since the rotation centerline is perpendicular to the normal vectors of the sampling isophase planes {F0, F1, F2, …} and is located on the isophase planes, the following relationship is satisfied:
A F i B F i C F i v 1 v 2 v 3 = [ 0 ] A F i p 1 + B F i p 2 + C F i p 3 + D F i = 0 .
Among them, A F i , B F i , C F i , and D F i represent the equation coefficients of any sampling isophase planes. By solving the equation, the spatial position of the rotation centerline L0 can be obtained.
In general, the calibration process can be divided into the following steps: step 1 is followed to obtain samples of absolute phase {φ0, φ1, φ2, …} and their corresponding isophase planes {F0, F1, F2, …}; step 2 is pe4rformed to complete the calibration of parameters a1 and a2 in the phase-angle model; step 3 is completed to obtain the spatial position of the rotation centerline L0.

3.2. Reconstruction for Global Scope

During the reconstruction process, the phase-angle model and rotation centerline determined during the calibration process are utilized, and these calibration results can be used for 3D reconstruction of all global pixels. The specific flowchart is shown in Figure 4. Firstly, extract the phase of each pixel on the reconstructed scene’s phase map one by one, and substitute it into the phase-angle model to obtain the current deflection angle θ′. Based on the rotation centerline, the position of the reference isophase plane and angle θ′, determine the spatial equation Fcur: A c u r X + B c u r Y + C c u r Z + D c u r = 0 for the current isophase plane. In this process, several conditions need to be realized: the rotation centerline is located on the current isophase plane, the angle between the current isophase plane and the reference isophase plane is θ′, and the normal vector modulus composed of the first three coefficients of Fcur can be artificially specified as 1. Specifically, the following equation system can be obtained:
A c u r v 1 + B c u r v 2 + C c u r v 3 = 0 A c u r p 1 + B c u r p 2 + C c u r p 3 + D c u r = 0 A c u r A + B c u r B + C c u r C = cos θ ( A c u r ) 2 + ( B c u r ) 2 + ( C c u r ) 2 = 1 .
Solve the equation system (15) to complete Part 1 of the reconstruction flowchart. In addition, the back-projection ray R corresponding to the current pixel will intersect with the normalization plane at a point x n = ( u ˜ , v ˜ , 1 ) T , and any point on R can be represented as follows:
X Y = Z u ˜ v ˜ .
At this point, Part 2 of the reconstruction has been completed. Finally, by combining the Fcur equation and the R ray equation, the intersection coordinates of the two can be solved, that is, the coordinates of the reconstructed points:
X = u ˜ D c u r u ˜ A c u r + v ˜ B c u r + C c u r Y = v ˜ D c u r u ˜ A c u r + v ˜ B c u r + C c u r Z = D c u r u ˜ A c u r + v ˜ B c u r + C c u r .

4. Experiment and Analysis

4.1. Calibration Results

To verify the feasibility of the proposed method for reconstruction in non-calibrated areas, we conducted experiments on an FPP system based on a DLP projector. The projector in the system was the Texas Instrument pro6500-rgb-235, with a resolution of 1920 × 1080 and a refresh rate of 100 fps @ 8 bit. And the camera used was the Balser aca 2040-120 um grayscale camera, with a resolution of 2064 × 1544 and a frame rate of 100 fps. The working distance of the system was approximately 450 mm, and the field of view of the camera was approximately 200 mm × 150 mm. In addition, we used a calibration board of the symmetrical circle grid type, which included eight rows and eleven columns of 12 mm spaced marker points.
As shown in Figure 5a,b, to free up more space outside of the calibration board for testing the 3D reconstruction effect, we placed the calibration board in the lower left corner area. By setting a mask, phase samples are selected within the specified rectangular area, as shown in the phase maps in Figure 5c,d. The same-colored circles represent matching sample points located on the same isophase plane, which are used for calculating the space equation of the isophase plane. There are a total of 11 colors corresponding to 11 samples of isophase planes, as shown in Figure 5e, and these are used as samples for fitting the phase-angle model to infer the localization of reconstruction points in areas outside of the coverage range of the calibration board.
Following the previous calibration process, the reference phase and the equation for the reference isophase plane were determined based on the selected sample points. Then, by solving the deflection angles corresponding to the isophase planes one by one, the optimization of the phase-angle model was completed, and the parameters a1 and a2 were obtained. Finally, the rotation centerline was obtained using the least squares method, and the relevant data are listed in Table 1.

4.2. Reconstruction Results

After completing the calibration using this method, the effectiveness of 3D reconstruction was tested. Firstly, the correctness of the method was verified based on the structured light system of the DLP projector. Secondly, the same calibration strategy used to test the applicability on emerging structured light projection devices, such as the FPP system of MEMS-mirror-based projectors, was used. As shown in Figure 6, (a) represents the fringe pattern of the object in the DLP-based FPP system, and (b) represents the 3D point cloud of the object reconstructed using this method; (c) represent the fringe pattern of the object in the MEMS-based FPP system, and (d) represents the corresponding 3D point cloud of the object. From the completeness of the results and the similarity of the shapes, it can be seen that this calibration method can achieve 3D reconstruction on different FPP systems.
By utilizing the scanning characteristics of the phase-angle model, as long as the projected fringes can cover the object, 3D reconstruction can be achieved efficiently and accurately, regardless of where the object is in the camera’s field of view. To verify the speculated reconstruction ability of this method, we compared the reconstruction results of the same object placed in the covered and uncovered areas of the calibration board and presented the feasibility of the proposed method in an intuitive form. As shown in Figure 7, (a) and (b) represent the positions of the two calibration boards, respectively, and (c) is the overlapping area of the two calibration boards, also known here as the calibration zone. The subgraphs (d), (h), (l), and (p) are the tested objects inside the calibration zone, and the same objects are also placed outside of the calibration zone, as shown in subgraphs (f), (j), (n), and (r). The subgraphs (e), (g), (i), (k), (m), (o), (q), and (s) show their reconstructed point cloud results, and, visually, there is no significant difference in the point cloud of the same object.
To highlight the speculated reconstruction ability of the proposed method outside of the calibration zone, the phase-coordinate mapping model was also compared with the proposed method. In this experiment, we placed the measured object at the junction of the calibration area and the non-calibration area, projected the fringes, calculated the phase, and used these two techniques to reconstruct the object separately. The specific results are shown in Figure 8. The subgraphs (a–d) represent the reconstruction results of the phase-coordinate mapping model shown by the polynomial model, while the subgraphs (e–h) represent the reconstruction results of the proposed method. It is obvious that the former shows the situation of reconstruction mapping failure outside of the calibration zone, and the proposed method avoids this problem.
In addition, the reconstruction accuracy in the calibration and non-calibration areas are quantitatively analyzed in the form of planarity error and sphericity error by reconstructing planes and spheres at different positions. Figure 9 shows the reconstructed planar and spherical point cloud results and presents their error distribution in the form of a 2D error distribution map. In Table 2, we reconstruct the plane and sphere at three positions inside and outside of the calibration zone and calculate the fitted planarity error and sphericity error. Although there is a certain increase in the reconstruction error outside of the calibration area, it can still be used as an ideal result in full-field reconstruction scenarios without being completely unable to implement 3D reconstruction like the phase-coordinate mapping model.
To further explore the reconstruction ability of the proposed method outside of the calibration zone, we set a mask to continuously reduce its coverage range, thereby reducing the actual calibration zone. As shown in Figure 10a, the mask was set on the calibration board, and the area decreased continuously from mask1 to mask2 and then to mask3. We sampled the phase within the respective set range and calibrated the phase-angle model curves, as shown in Figure 10b–d. It can be seen from the results that although the calibration area was reduced, the model curve could still be extended in the unsampled area and assisted in inferring the 3D reconstruction outside of the calibration zone.
A 2D error distribution map of the reconstructed standard object outside of the calibration zone is shown in Figure 11a–f, and the subgraphs (a–c) show the reconstructed sphericity error as the mask setting changes. The subgraphs (d–f) show the reconstructed planarity error as the mask setting changes. To avoid randomness, we placed the standard object at three different positions outside of the calibration zone and calculated the average planarity error and average sphericity error separately. The experiment was conducted under each different mask setting, as shown in Figure 11g,h, and the results show the changes in the average planarity error and average sphericity error under different mask settings. It can be seen that the reconstruction accuracy outside of the calibration area is high, and that this calibration method does not cause significant fluctuations in error due to the reduction in the calibration area. This also proves that it can still achieve robust 3D reconstruction ability outside of the calibration zone.

5. Conclusions

In this paper, we propose a method to infer the 3D information of objects outside of the calibration zone using a phase-angle model for the scenario of unidirectional fringe projection. By sampling the phase within the local calibration area and obtaining the sample isophase planes through coordinate transformation and least squares estimation, it can be used for calibrating the phase-angle model and the rotation centerline. This method can achieve full-field 3D reconstruction, and its innovative points can be summarized as follows: 1. the 3D information outside of the calibration zone can still be obtained; 2. this method is efficient and concise, requiring only the projection of fringes at two calibration positions; 3. reconstruction can still maintain high accuracy and robustness on a global scale. The feasibility and efficiency of the proposed method were experimentally verified.

Author Contributions

Conceptualization, F.L.; methodology, X.L. and F.L.; software, F.L. and R.M.; validation, R.M.; writing—original draft preparation, F.L. writing—review and editing, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Cross Funding of Tsinghua University Shenzhen International Graduate School, grant number JC2021003 and the Shenzhen Stable Supporting Program, grant number WDZC20200820200655001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lv, S.; Tang, D.; Zhang, X.; Yang, D.; Deng, W.; Kemao, Q. Fringe projection profilometry method with high efficiency, precision, and convenience: Theoretical analysis and development. Opt. Express 2022, 30, 33515–33537. [Google Scholar] [CrossRef]
  2. Xing, C.; Huang, J.; Wang, Z.; Duan, Q.; Li, Z.; Qi, M. A high-accuracy online calibration method for structured light 3D measurement. Measurement 2023, 210, 112488. [Google Scholar] [CrossRef]
  3. An, H.; Cao, Y.; Zhang, Y.; Li, H. Phase-shifting temporal phase unwrapping algorithm for high-speed fringe projection profilometry. IEEE Trans. Instrum. Meas. 2023, 72, 5009209. [Google Scholar] [CrossRef]
  4. Zhang, S. High-speed 3D shape measurement with structured light methods: A review. Opt. Lasers Eng. 2018, 106, 119–131. [Google Scholar] [CrossRef]
  5. Yang, G.; Wang, Y. High resolution laser fringe pattern projection based on MEMS micro-vibration mirror scanning for 3D measurement. Opt. Laser Technol. 2021, 142, 107189. [Google Scholar] [CrossRef]
  6. Wu, Y.; Wu, G.; Li, L.; Zhang, Y.; Luo, H.; Yang, S.; Yan, J.; Liu, F. Inner Shifting-Phase Method for High-Speed High-Resolution 3-D Measurement. IEEE Trans. Instrum. Meas. 2020, 69, 7233–7239. [Google Scholar] [CrossRef]
  7. Gong, Z.; Sun, J.; Zhang, G. Dynamic structured-light measurement for wheel diameter based on the cycloid constraint. Appl. Opt. 2016, 55, 198–207. [Google Scholar] [CrossRef] [PubMed]
  8. Sansoni, G.; Trebeschi, M.; Docchio, F. State-of-The-Art and Applications of 3D Imaging Sensors in Industry, Cultural Heritage, Medicine, and Criminal Investigation. Sensors 2009, 9, 568–601. [Google Scholar] [CrossRef] [PubMed]
  9. Haleem, A.; Javaid, M. 3D scanning applications in medical field: A literature-based review. Clin. Epidemiol. Glob. Health 2019, 7, 199–210. [Google Scholar] [CrossRef]
  10. Yu, H.; Zheng, D.; Fu, J.; Zhang, Y.; Zuo, C.; Han, J. Deep learning-based fringe modulation-enhancing method for accurate fringe projection profilometry. Opt. Express 2020, 28, 21692–21703. [Google Scholar] [CrossRef]
  11. Han, M.; Kan, J.; Yang, G.; Li, X. Robust Ellipsoid Fitting Using Combination of Axial and Sampson Distances. IEEE Trans. Instrum. Meas. 2023, 72, 2526714. [Google Scholar] [CrossRef]
  12. Gorthi, S.S.; Rastogi, P. Fringe projection techniques: Whither we are? Opt. Lasers Eng. 2010, 48, 133–140. [Google Scholar] [CrossRef]
  13. Feng, S.; Zuo, C.; Zhang, L.; Tao, T.; Hu, Y.; Yin, W.; Qian, J.; Chen, Q. Calibration of fringe projection profilometry: A comparative review. Opt. Lasers Eng. 2021, 143, 106622. [Google Scholar] [CrossRef]
  14. Miao, Y.; Yang, Y.; Hou, Q.; Wang, Z.; Liu, X.; Tang, Q.; Peng, X.; Gao, B.Z. High-efficiency 3D reconstruction with a uniaxial MEMS-based fringe projection profilometry. Opt. Express 2021, 29, 34243–34257. [Google Scholar] [CrossRef]
  15. Zhang, S. Flexible and high-accuracy method for uni-directional structured light system calibration. Opt. Lasers Eng. 2021, 143, 106637. [Google Scholar] [CrossRef]
  16. Yang, Y.; Miao, Y.; Cai, Z.; Gao, B.Z.; Liu, X.; Peng, X. A novel projector ray-model for 3D measurement in fringe projection profilometry. Opt. Lasers Eng. 2022, 149, 106818. [Google Scholar] [CrossRef]
  17. Yang, D.; Qiao, D.; Xia, C. Curved light surface model for calibration of a structured light 3D modeling system based on striped patterns. Opt. Express 2020, 28, 33240–33253. [Google Scholar] [CrossRef]
  18. Wan, T.; Liu, Y.; Zhou, Y.; Liu, X. Large-scale calibration method for MEMS-based projector 3D reconstruction. Opt. Express 2023, 31, 5893–5909. [Google Scholar] [CrossRef] [PubMed]
  19. Gao, W.; Wang, L.; Hu, Z. Flexible method for structured light system calibration. Opt. Eng. 2008, 47, 083602. [Google Scholar] [CrossRef]
  20. Zhu, H.; Xing, S.; Guo, H. Efficient depth recovering method free from projector errors by use of pixel cross-ratio invariance in fringe projection profilometry. Appl. Opt. 2020, 59, 1145–1155. [Google Scholar] [CrossRef]
  21. Zhang, S. High-Resolution, Real-Time 3-D Shape Measurement; State University of New York at Stony Brook: New York, NY, USA, 2005. [Google Scholar]
  22. Zhou, W.-S.; Su, X.-Y. A direct mapping algorithm for phase-measuring profilometry. J. Mod. Opt. 1994, 41, 89–94. [Google Scholar] [CrossRef]
  23. Lei, F.; Han, M.; Jiang, H.; Wang, X.; Li, X. A phase-angle inspired calibration strategy based on MEMS projector for 3D reconstruction with markedly reduced calibration images and parameters. Opt. Lasers Eng. 2024, 176, 108078. [Google Scholar] [CrossRef]
Figure 1. Schematic of the proposed global phase-angle relationship based on local calibration.
Figure 1. Schematic of the proposed global phase-angle relationship based on local calibration.
Sensors 24 02581 g001
Figure 2. (a) Schematic of the geometrical position of the inclined calibration plane and the orthogonal calibration plane in the system; (b) top view of the positions.
Figure 2. (a) Schematic of the geometrical position of the inclined calibration plane and the orthogonal calibration plane in the system; (b) top view of the positions.
Sensors 24 02581 g002
Figure 3. Schematic diagram of system calibration based on the phase-angle model.
Figure 3. Schematic diagram of system calibration based on the phase-angle model.
Sensors 24 02581 g003
Figure 4. Reconstruction process diagram.
Figure 4. Reconstruction process diagram.
Sensors 24 02581 g004
Figure 5. Calibration plate and phase sampling points in the phase maps. (a,b) are the distribution of sampling points on the original calibration images; (c,d) are the distribution of sampling points on the calibration plate phase maps; (e) is the distribution of calibration plane and isophase plane.
Figure 5. Calibration plate and phase sampling points in the phase maps. (a,b) are the distribution of sampling points on the original calibration images; (c,d) are the distribution of sampling points on the calibration plate phase maps; (e) is the distribution of calibration plane and isophase plane.
Sensors 24 02581 g005
Figure 6. Results of applying the proposed method to reconstruct objects in different FPP systems. (a) is the object projected under DLP based FPP system; (b) is the reconstruction result under DLP based FPP system; (c) is the object projected under MEMS based FPP system; (d) is the reconstruction result under MEMS based FPP system.
Figure 6. Results of applying the proposed method to reconstruct objects in different FPP systems. (a) is the object projected under DLP based FPP system; (b) is the reconstruction result under DLP based FPP system; (c) is the object projected under MEMS based FPP system; (d) is the reconstruction result under MEMS based FPP system.
Sensors 24 02581 g006
Figure 7. Reconstruction results of objects placed inside and outside of the calibration zone. (a,b) represent the positions of the two calibration boards; (c) is the overlapping area of the two calibration boards; (d,h,l,p) are the tested objects inside the calibration zone; (f,j,n,r) are the same objects outside the calibration zone; (e,g,i,k,m,o,q,s) are their corresponding reconstructed point cloud results.
Figure 7. Reconstruction results of objects placed inside and outside of the calibration zone. (a,b) represent the positions of the two calibration boards; (c) is the overlapping area of the two calibration boards; (d,h,l,p) are the tested objects inside the calibration zone; (f,j,n,r) are the same objects outside the calibration zone; (e,g,i,k,m,o,q,s) are their corresponding reconstructed point cloud results.
Sensors 24 02581 g007
Figure 8. Comparison of reconstruction results between the phase coordinate mapping model and the proposed method. (ad) represent the reconstruction results of the phase-coordinate mapping model shown by the polynomial model; (eh) represent the reconstruction results of the proposed method.
Figure 8. Comparison of reconstruction results between the phase coordinate mapping model and the proposed method. (ad) represent the reconstruction results of the phase-coordinate mapping model shown by the polynomial model; (eh) represent the reconstruction results of the proposed method.
Sensors 24 02581 g008
Figure 9. Point clouds and 2D error distribution maps for plane and sphere surfaces.
Figure 9. Point clouds and 2D error distribution maps for plane and sphere surfaces.
Sensors 24 02581 g009
Figure 10. (a) Schematic of the masks for phase sampling on the calibration plate; (bd) the phase-angle model curves under mask1, mask2, and mask3, respectively.
Figure 10. (a) Schematic of the masks for phase sampling on the calibration plate; (bd) the phase-angle model curves under mask1, mask2, and mask3, respectively.
Sensors 24 02581 g010
Figure 11. (ac) Sphericity errors 2D distribution map outside of the calibration zone at mask1, mask2, and mask3 settings; (df) planarity error 2D distribution map at mask1, mask2, and mask3 settings outside of the calibration zone; (g) the planarity mean error variation at different mask settings; (h) the sphericity mean error variation at different masks settings.
Figure 11. (ac) Sphericity errors 2D distribution map outside of the calibration zone at mask1, mask2, and mask3 settings; (df) planarity error 2D distribution map at mask1, mask2, and mask3 settings outside of the calibration zone; (g) the planarity mean error variation at different mask settings; (h) the sphericity mean error variation at different masks settings.
Sensors 24 02581 g011
Table 1. Phase-angle model-enabled calibration strategy parameters.
Table 1. Phase-angle model-enabled calibration strategy parameters.
System Calibration ParametersValue
Reference phase φ r e f 30.6732
Reference isophase plane A F 0 0.9093
B F 0 −0.0134
C F 0 0.4160
D F 0 −139.7620
Phase-angle modela10.2022
a2−382.4269
Rotational centerlinev1, v2, v3(0.0268, 0.9993, −0.0259)
p1, p2, p3(97.6335, −184.3396, 116.7032)
Table 2. Planarity or sphericity errors in different zones.
Table 2. Planarity or sphericity errors in different zones.
Planarity or Sphericity ErrorsPlane Inside
Calibration Zone
Plane Outside Calibration ZoneSphere, Inside Calibration ZoneSphere, Outside Calibration Zone
Position10.0484 mm0.1298 mm0.0432 mm0.0509 mm
Position20.0512 mm0.1291 mm0.0403 mm0.0535 mm
Position20.0411 mm0.1325 mm0.0388 mm0.0553 mm
Mean0.0469 mm0.1305 mm0.0408 mm0.0532 mm
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lei, F.; Ma, R.; Li, X. Use of Phase-Angle Model for Full-Field 3D Reconstruction under Efficient Local Calibration. Sensors 2024, 24, 2581. https://doi.org/10.3390/s24082581

AMA Style

Lei F, Ma R, Li X. Use of Phase-Angle Model for Full-Field 3D Reconstruction under Efficient Local Calibration. Sensors. 2024; 24(8):2581. https://doi.org/10.3390/s24082581

Chicago/Turabian Style

Lei, Fengxiao, Ruijie Ma, and Xinghui Li. 2024. "Use of Phase-Angle Model for Full-Field 3D Reconstruction under Efficient Local Calibration" Sensors 24, no. 8: 2581. https://doi.org/10.3390/s24082581

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop