Next Article in Journal
TDFusion: When Tensor Decomposition Meets Medical Image Fusion in the Nonsubsampled Shearlet Transform Domain
Next Article in Special Issue
A 256 × 256 LiDAR Imaging System Based on a 200 mW SPAD-Based SoC with Microlens Array and Lightweight RGB-Guided Depth Completion Neural Network
Previous Article in Journal
A Safety Warning Model Based on IAHA-SVM for Coal Mine Environment
Previous Article in Special Issue
A Fast and Reliable Solution to PnP, Using Polynomial Homogeneity and a Theorem of Hilbert
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Vision Measurement Technique with Large Field of View and High Resolution

1
Research Center of Advanced Microscopy and Instrumentation, Harbin Institute of Technology, Harbin 150001, China
2
Research Center of Basic Space Science, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6615; https://doi.org/10.3390/s23146615
Submission received: 19 June 2023 / Revised: 13 July 2023 / Accepted: 21 July 2023 / Published: 23 July 2023
(This article belongs to the Collection 3D Imaging and Sensing System)

Abstract

:
The three-dimensional (3D) displacement resolution of conventional visual measurement systems can only reach tens of microns in cases involving long measuring distances (2.5 m) and large fields of view (1.5 m × 1.5 m). Therefore, a stereo vision measurement technology based on confocal scanning is proposed herein. This technology combines macroscopic visual measurement technology with confocal microscopic measurement technology to achieve a long measuring distance, a large field of view, and micron-level measuring resolution. First, we analyzed the factors affecting the 3D resolution of the visual system and developed a 3D resolution model of the visual system. Subsequently, we fabricated a prototype based on the resolution model and the proposed stereo vision measurement technology. The 3D displacement resolution measurement results in the full field of view show that the displacement resolutions of the developed equipment in the x-, y-, and z-directions can reach 2.5, 2.5, and 6 μm, respectively.

1. Introduction

Stereoscopic vision technology [1,2,3,4,5,6] is a technology that mimics human eye imaging. It uses two or more cameras to image the same object from different orientations and then calculates the parallax through stereo matching. Finally, it calculates the three-dimensional (3D) geometric information of the object based on the calibrated internal and external parameters of the visual system. Visual technology is widely used and investigated owing to its characteristics of zero contact, high speed, and high measurement accuracy. In recent decades, visual technology has been increasingly used in many fields, such as robot vision, aerial surveying, medical imaging, and industrial testing.
However, most researchers of visual technology tend to focus on improving the measurement accuracy and success rate by investigating (i) different calibration algorithms [7,8,9,10,11] to improve the calibration accuracy, (ii) different matching algorithms to improve the matching accuracy [12,13,14,15], and (iii) three-dimensional reconstruction algorithms to improve the reconstruction accuracy [16,17,18]. In addition, researchers have focused on expanding the application scenarios [19,20]. Some researchers have studied the relationship between visual system structural parameters and measurement accuracy [21,22,23,24,25]. Resolution, a crucial indicator in the field of measurement, is typically disregarded in visual technology. The resolution of vision technology is defined as the minimum displacement of an object that can be recognized using vision technology. This is primarily because resolution is not an advantage of visual measurement technology, nor is it prioritized in most application fields of visual technology. Therefore, resolution is typically not considered in visual technology. However, in some application scenarios, not only the main indicators of visual technology—such as field of view (FOV) and accuracy—but also resolution must be considered. Gong et al. [26] and Li et al. [27] analyzed the effects of individual structural parameters of a visual system on its resolution; however, they did not disclose a method to overcome the inherent resolution limitations of the visual system. Under normal circumstances, when the test distance exceeds 2.5 m and the FOV exceeds 1.5 m × 1.5 m, the resolution of the conventional vision system is typically tens or hundreds of microns.
Confocal scanning imaging technology [28,29,30] is a two-dimensional (2D) optical measurement technology that can realize 3D measurements of objects by combining axial scanning with axial positioning technology. The FOV of confocal technology is typically in the order of millimeters or microns. Confocal technology uses the three-point conjugation principle (point illumination, point detection, and point object) to scan and image objects point by point. Unlike the case of conventional array detection technology, the sampling interval in confocal technology depends on the scanning and sampling frequencies. Based on the existing scanning and sampling equipment, the sampling interval can reach nanometers or sub-nanometers. Confocal technology is a micro measurement technology that offers FOV measurements not exceeding a few millimeters.
Herein, we propose a new vision measurement technology based on confocal scanning imaging that allows a large FOV and a high resolution to be achieved simultaneously in vision measurements. First, we develop a 3D resolution model of the visual system, through which the factors affecting the resolution of the visual system can be obtained. Subsequently, we combine confocal scanning technology with vision technology to propose a vision measurement technology based on confocal scanning imaging for the first time. This technology uses the point-scanning imaging characteristics of confocal technology to reduce the sampling interval and overcome the resolution limit of conventional vision technology.

2. Methods

2.1. Resolution Model of a Stereo Vision System

A binocular stereo vision system typically comprises two identical monocular vision systems, the principle of which is illustrated in Figure 1. The binocular system comprises two monocular systems, and each monocular vision system comprises an optical imaging system (c1 or c2) and an imaging detector (CCD1 or CCD2). The optical imaging system (c1 or c2) is not necessarily a camera and can be another complex optical imaging system equivalent to a thin lens without considering aberration. Any type of imaging detector can be used, provided that it can capture the image of the object. The two optical systems are assumed to be identical. The image distance is f; the center distance (baseline distance) of the two imaging planes is 2L; and o1 and o2 are the intersection points between the optical axes of the left and right monocular imaging systems and the corresponding imaging plane, respectively. The world coordinate system, O-xyz, was constructed by considering the midpoint of o1o2 as the origin and the extension line as the x-axis. The optical axis inclinations of left and right monocular imaging systems are α1 and α2, respectively, where α1 = α2 = α. The imaging coordinate systems of the left and right monocular imaging systems are o1x1y1z1 and o2x2y2z2, respectively, with o1 and o2 as the origins, respectively. The optical axes of the left and right monocular imaging systems are o1z1 and o2z2, respectively. Point P is the point to be measured and is located in the FOV of the binocular stereo vision system. The imaging points on the two imaging planes are points P1 and P2, whose coordinates are (x1, y1, 0) and (x2, y2, 0), respectively. Point p is the intersection of point P projected vertically onto the c1c2o1o2 plane; hence, point P has the same x- and z-coordinates as point p. Point p forms a straight line parallel to o1o2 on the c1c2o1o2 plane, which intersects o1z1 and o2z2 at points p1 and p2, respectively. A perpendicular line that passes point p intersects o1z1 at point b.
Before analyzing the factors affecting the resolution of the visual system, the definition of resolution should be presented. We define the resolution of a binocular stereo vision system as the minimum displacement of an object that can be effectively recognized by the visual system. Based on the characteristics of the visual system, resolution can be defined as the displacement of an object when a perceptible change occurs in the imaging plane of the left or right monocular system.
As shown in Figure 1, the coordinate of object point P (x, y, z) in the imaging coordinate system o1x1y1z1 of the left visual system is (s1 × cosα1, y, z × secα1 + s1 × sinα1), and the coordinate of object point P in the imaging coordinate system o2x2y2z2 of the right visual system is (s2 × cosα2, y, z × secα2s2 × sinα2), where:
s 1 = p 1 p = x + L z tan α 1 s 2 = p 2 p = x L + z tan α 2
Based on the similar triangle principle, we obtain:
x 1 s 1 cos α = f z sec α + s 1 sin α f = y 1 y x 2 s 2 cos α = f z sec α s 2 sin α f = y 2 y
Based on the definition of the resolution of the visual system, when point P propagates along any of the three directions (x, y, or z), the coordinates of the image points in the left and right cameras change accordingly. If the coordinate changes of one of the image points can be recognized, then it indicates that the vision system can distinguish changes in the position of point P, which is the resolution. The 3D resolutions of the visual system are independent of each other; therefore, we can calculate the resolution in three directions. The resolutions of the visual system in the x-, y-, and z-directions are denoted as ΔX, ΔY, and ΔZ, respectively.
Suppose that point P propagates only slightly along the x-direction. In this case, to solve the displacement of image points in the left and right visual systems, the coordinates of point P should first be represented by the left and right imaging coordinates. By substituting Equation (1) into Equation (2), we obtain:
x = ( L sin α + z cos α f ) x 1 L f cos α + z f sin α x 1 sin α + f cos α
x = y 1 f cos α z f y cos α y 1 sin α cos α L + z tan α
x = ( L sin α + z cos α f ) x 2 + L f cos α z f sin α x 2 sin α + f cos α
x = y 2 f cos α + z + f y cos α y 2 sin α cos α + L z tan α
When P propagates only along the x-direction and does not change along the y- and z-directions, the derivatives of Equations (3)–(6) with respect to x1, y1, x2, and y2 can be obtained as follows:
d x d x 1 = ( z f cos α ) f x 1 sin α + f cos α 2 = [ z cos α + ( x + L ) sin α f ] 2 ( z f cos α ) f
d x d x 2 = ( z f cos α ) f x 2 sin α + f cos α 2 = [ z cos α ( x L ) sin α f ] 2 ( z f cos α ) f
d x d y 1 = f y y 1 2 sin α = z cos α x + L sin α + f f y sin α 2
d x d y 2 = f y y 2 2 sin α = z cos α + x L sin α + f f y sin α 2
Assuming that the left and right visual systems are identical, the minimum perceptible change on the image plane of the left and right visual systems is Δw. Based on the definition of resolution, the resolution in the x-direction can be written as:
Δ X = min d x d x 1 · Δ w , d x d x 2 · Δ w , d x d y 1 · Δ w , d x d y 2 · Δ w
Here, Δw is a pixel without considering the subpixel algorithm. However, with the development of the subpixel algorithm, Δw can be one-tenth of a pixel or even smaller, where w denotes the sampling interval.
Similarly, for the resolutions in the y- and z-directions, the imaging coordinate system of the left and right visual systems is used to represent the y- and z-coordinates of point P, respectively, and the derivative can be obtained as follows:
d y d y 1 = ( x + L ) sin α z cos α + f f
d y d y 2 = ( x L ) sin α z cos α + f f
d z d x 1 = ( x + L f sin α ) f x 1 cos α + f sin α 2 = ( x + L ) sin α + z cos α f 2 ( x + L f sin α ) f
d z d x 2 = ( x L + f sin α ) f x 2 cos α + f sin α 2 = x sin α + L sin α + z cos α f 2 x + L f sin α f
d z d y 1 = f y y 1 2 cos α = ( x + L ) sin α z cos α + f 2 f y cos α
d z d y 2 = f y y 2 2 cos α = ( x L ) sin α z cos α + f 2 f y cos α
Subsequently, the resolutions in the y- and z-directions of the visual system are expressed as:
Δ Y = min d y d y 1 · Δ w , d y d y 2 · Δ w
Δ Z = min d z d x 1 · Δ w , d z d x 2 · Δ w , d z d y 1 · Δ w , d z d y 2 · Δ w
A 3D resolution model of the visual system was developed. The theoretical resolution of any visual system can be obtained by combining the requirements of the FOV. Based on the model, the resolution of the visual system is not only related to the internal parameters of the visual system but also directly related to the external parameters of the visual system. The left and right visual systems are typically placed in parallel in an application environment featuring a large FOV. Therefore, the resolutions of a visual system with parallel optical axes are expressed as follows:
Δ X = min z f f · Δ w , f z f · Δ w Δ Y = min z f f · Δ w , f z f · Δ w Δ Z = min z f 2 x + L f · Δ w , z f 2 L x f · Δ w , f z 2 f y · Δ w , f z 2 f y · Δ w
Based on Equation (20), one can intuitively form the following conclusions: 1. The resolution of the parallel optical axis vision system is proportional to the sampling interval. 2. The resolution of the parallel optical axis vision system in the x- and y-directions remains the same over the entire FOV. 3. The x- and y-resolutions of the parallel optical-axis vision system are proportional to the test distance, whereas the z-resolution is proportional to the square of the test distance. 4. The z-direction resolution of the parallel optical-axis vision system is inversely proportional to the baseline distance.
The relationship between the resolution and FOV of the visual system can be shown more intuitively. The focal length of the left and right lens was assumed to be 16 mm; the sampling interval was 0.2 μm, i.e., Δw = 0.2 μm; the FOV of the visual system was 1 m; the baseline distance, 2L, was 2 m; and the test distance of z was 2 m. Hence, the resolution of the monocular visual system in the x-direction was 24.8 μm. The resolution in the y-direction was also 24.8 μm, and the z-direction resolutions of the monocular and binocular vision systems are shown in Figure 2. As shown in Figure 2d, the binocular vision system with parallel optical axis indicated the worst resolution (49.2 μm) at the center of the FOV and the best resolution (32.8 μm) at the edge of the FOV. However, as shown in Figure 2c, the z-direction resolution of the monocular system was monotonic. Based on the left visual system as an example, the greater the distance from the optical axis of the left visual system, the higher the z-direction resolution. Figure 2a represents the change amount along the x1-direction on the image plane of the left visual system when point P changed along the z-direction, which is consistent with Figure 2c. Figure 2b shows the change in the value of point P along the y1-direction on the image plane of the left visual system when point P changed along the z-direction. When point P was in the center of the FOV, the displacement of point P in the z-direction did not reflect the change in the y1-direction. At the edge of the FOV, the minimum displacement of point P along the z-direction was 98.4 μm. In other words, when point P propagated along the z-direction, its image point was displaced along the x1- and y1-directions on the image plane, and the displacement along the x1-direction exceeded that along the y1-direction.
To analyze the influence of other parameters on the resolution, we first give the parameters of the visual system as follows: the measurement distance is 2.5 m, the FOV is 1.5 m, the focal length is 16 mm, the sampling interval is 0.5 μm, and the size of the detector is infinite. The variable parameters are the optical axis inclination and the baseline distance of the visual system. First, assume that the optical axis inclination is 0° and the baseline distance changes from 1.5 m to 2.5 m. At this time, according to the resolution model, the resolution in the x- and y-directions is 78.125 μm, and the resolution in the z-direction is shown in Figure 3. It can be seen that the larger the baseline distance, the higher the Z-direction resolution. The baseline distance increased from 1.5 m to 2.5 m and the z-direction resolution increased from 260.4 μm to 156.3 μm. In theory, the resolution can be improved by increasing the baseline distance, but in practical applications, the increase in the baseline distance will inevitably lead to reduction in the FOV, so the resolution cannot be greatly improved by this method. We assume that the baseline distance is 2 m, and the optical axis inclination increases from 0° to 10°. At this time, the resolution of the visual system in x-, y-, and z-directions can be calculated according to the resolution model, as shown in Figure 4. As can be seen from the figure, the resolution decreases with the increase in the optical axis inclination. With the optical axis inclination increased from 0° to 10°, the resolution in the x-direction decreased from 78.125 μm to 86.81 μm, the resolution in the y-direction decreased from 78.125 μm to 82.36 μm, and the resolution in the z-direction decreased from 195.3 μm to 217.7 μm. Therefore, from the above analysis, it can be seen that for a visual system, the simplest and most effective way to improve the resolution is to reduce the sampling interval. There are two methods to reduce the sampling interval: one is to reduce the pixel size of the detector, and the other is to develop a new subpixel algorithm to subdivide the pixel size. But, at present, both methods are difficult to further break through.
In order to demonstrate the predictive capabilities of the 3D resolution model of a vision system, we constructed a vision system. The focal length of the lens used in the vision system is 8 mm (product model: HN-0816-5M-C2/3X). The imaging area of the imaging detector is 2/3 inch and the pixel size is 3.45 μm (product model: MER2-503-36U3M). The parameters of the displacement platform are as follows: the maximum stroke is 200 μm and the resolution is 20 nm. We first calibrated the field of view of the visual system, as shown in Figure 5. We can see that the FOV in the x-direction is about 315 mm, and we assume that the FOV in the y-direction is also 315 mm. The measured distance can be calculated as 315 × 8/8.8 ≈ 286 mm. Due to the limitation of the stroke of the displacement platform, we placed the target at the edge of the FOV of the vision system, as shown in Figure 6. At this time, the distance of the object from the optical axis was 96 mm. Assume that the minimum displacement recognized by the camera is 0.5 pixels, that is, 1.725 μm. The theoretical resolution calculated by the resolution model is 61.7 μm, 61.7 μm, and 183.7 μm in the x-, y-, and z-directions, respectively. Due to the limitation of the stroke of the displacement platform, in this experiment, the movement of our displacement platform is carried out in a reciprocating way—that is, it first moves A μm along the x-direction, and then moves −a μm along the x-direction, repeated 5 times. The displacement platform moves in the same way in the y- and z-directions. In the actual test, the distance we shift in the x- and y-directions is 62 μm, i.e., a = 62 μm. The distance shifted in the z-direction is 184.1 μm, that is, a = 184.1 μm. The displacement results calculated by image-matching algorithm based on ZNCC are shown in Table 1. As can be seen from the table, these displacements can be effectively identified, that is, the three-dimensional displacement resolution is (62 μm, 62 μm, 184.1 μm). The actual measured resolution is very close to the theoretical value, and only slightly larger than the theoretical value. This is mainly due to parameter error, noise, and other factors. Therefore, our model can be effectively applied to the traditional vision system to predict its resolution.

2.2. Principle of Visual Measurement Technology Based on Confocal Scanning Imaging

Based on the analysis presented in the previous section, the resolution of the visual system depends on the sampling interval; that is, the smaller the sampling interval, the higher the resolution of the visual system. However, in conventional vision systems, fixed detectors such as CCDs are used as imaging detection devices, and the pixel size is the sampling interval. Owing to factors such as processing technology and detector material, further reducing the pixel size of the detectors is challenging; consequently, the resolution of the conventional vision system cannot be easily improved. As shown in the section above, under the meter-level FOV, the resolution of the binocular system typically exceeds 10 µm. However, the structure of the vision technology is simple, the 3D displacement of the object can be calculated only by image matching, and the measurement FOV is relatively large.
Confocal imaging systems are typically used for microscopic measurements. These systems scan objects point by point and then perform single-point detection imaging for each scanning point. Therefore, the sampling interval for this technology depends on the scanning interval. The scanning interval can be adjusted based on the scanning frequency of the galvanometer and the sampling frequency of the data acquisition card. For existing data acquisition cards, the sampling rate can typically reach tens of megahertz, and a sampling interval in the micron or submicron level can be achieved under the meter-level FOV by matching the scanning frequency. Furthermore, confocal technology can achieve nanometer or sub-nanometer resolutions; however, it has a maximum FOV on the order of millimeters. A more detailed, confocal technique is a point-by-point scanning imaging technique. It illuminates the object point through a point light source and then collects the reflected light of the object point to image the object point. The original confocal technique is stationary like a fixed sensor, while the object moves in the plane through an x/y displacement platform to achieve scanning of the object. At this time, the three-dimensional information of the object cannot be obtained, and the axial displacement of the object cannot be obtained. In order to obtain three-dimensional information about the object, it is necessary to use the axial displacement platform to drive the object along the axis direction, and then use the x/y displacement platform to move the object in the plane and image the object, then repeat the above steps. In general, the number of axial movements is dozens or even hundreds of times. And then the three-dimensional information of the object is calculated through a certain algorithm. Moreover, confocal technology uses a point light source to illuminate an object and only illuminates one object point at a time. As the object moves in the xy plane, the object point illuminated by the point light source also changes. The confocal system collects the reflected signals of these different object points through the data acquisition card, and the sampling rate of the data acquisition card is very high, resulting in a very small interval between the two object points. Therefore, when the object point moves in the xy plane, confocal technology can be easily identified. When the object has axial displacement, confocal technology needs to obtain the three-dimensional information of the object before displacement, and then obtain the three-dimensional information of the object after displacement and calculate the axial displacement of the object through the relative change in the three-dimensional information. Therefore, this method is very slow, and with the development of technology, it is now common to use galvanometers to achieve two-dimensional scanning, and the two scanning methods are exactly the same. But axial scanning still requires the use of a displacement platform.
Although visual and confocal measurement technologies are unrelated, we combined them to propose a technology based on confocal scanning. Our technology combines the advantages of the two technologies, using confocal technology to scan the object to improve the resolution, and then using vision technology to calculate the three-dimensional displacement of the object without axial scanning. A schematic illustration of the monocular system is shown in Figure 7. As shown in the schematic diagram, the technology is primarily composed of two components, i.e., a photographic lens, which serves to increase the measurement range of the system, and a confocal scanning imaging subsystem, which is composed of elements 3–12, as shown in the figure. A confocal scanning imaging module was used to replace the CCD imaging module of the conventional vision system, and a point-by-point scanning imaging method was used to realize the imaging measurement of the object. As shown in the schematic diagram, the imaging pixel was mapped to the object side, or the object was mapped to the image side. When the object point shifted slightly, the amount of movement was less than the corresponding size of the CCD pixel. Therefore, in the conventional visual system, the object point is imaged in the same pixel before and after displacement. Thus, even if the object point is displaced, the conventional visual system cannot recognize it. For vision technology based on confocal scanning imaging, the sampling interval can be reduced based on the sampling rate and scanning frequency, and submicron or smaller sampling intervals can be realized under a meter-level FOV. Therefore, for the confocal vision system, because the sampling interval was reduced, the object point was imaged at the 7th sampling interval (the first black pixel from left to right) before it shifted, and at the 10th sampling interval after it shifted. This implies that a slight displacement can be recognized.
We can treat the entire confocal module as an imaging detector whose role is to scan the image plane of the photographic lens. This is feasible because the diffraction effect is not taken into account when the resolution model is established, that is, the aperture of the photographic lens is treated as infinite. Therefore, imaging of the object by the photographic lens does not lose any information about the object, but only changes the size of the object. The scanning of the confocal module on the image plane is equivalent to a single-pixel detector scanning on the image plane. It is assumed that the focal length of the objective lens is f2, the tube lens is f3, the scanning lens is f4, the scanning angle range of the galvanometer is θ, the scanning frequency of the galvanometer is m Hz, and the sampling frequency of the data acquisition card is n Hz. According to the confocal scanning principle, the sampling interval is m/n × θ. Then, the minimum sampling interval on the front focal plane of the scanning lens is f4 × m/n × θ, and the sampling interval on the front image plane of the objective lens is:
Δ c = f 4 f 2 f 3 m n θ = f 4 k m n θ
where k represents the magnification of the objective lens and the tube lens, which is called the first-order magnification. When k = 1, the objective and tube lenses can be removed. Therefore, the 3D resolution model of vision measurement technology based on confocal scanning is expressed as follows:
Δ X = min d f · f 4 k m n θ , d f · f 4 k m n θ Δ Y = min d f · f 4 k m n θ , d f · f 4 k m n θ Δ Z = min d 2 x + L f · f 4 k m n θ , d 2 L x f · f 4 k m n θ , d 2 f y · f 4 k m n θ , d 2 f y · f 4 k m n θ
where d represents the test distance. The following can be seen from the formula: 1. The resolution of the vision measurement technology based on confocal scanning is proportional to the scanning frequency; that is, the smaller the scanning frequency, the higher the resolution of the system; 2. The resolution is inversely proportional to the sampling frequency; that is, the higher the sampling frequency, the higher the resolution of the system; 3. The resolution is proportional to the focal length of the scanning lens; that is, the smaller the focal length, the higher the resolution; 4. The resolution is inversely proportional to the first-order magnification rate; that is, the greater the first-order magnification rate, the higher the resolution.
Assuming that the focal length of the telephoto lens is 16 mm, the resolution of the matching algorithm has a sampling interval of 0.1, the field of view of the visual system is 1.5 m, the baseline distance is 2 m, the test distance d is 2 m, the scanning angle of the galvanometer is 25°, the first-order magnification is 5, the focal length of the scanning lens is 367 mm, the sampling frequency of the data acquisition card is 20 MHz, and the scanning frequency is 20 Hz. The ratio between the sampling frequency of the data acquisition card and the scanning galvanometer is called the scanning sampling ratio, that is, r = n/m. Through calculation, the resolution of the system in the x-direction is 0.4 μm, and the resolution in the y-direction is also 0.4 μm. The z-direction resolution of the confocal binocular vision system is shown in Figure 8. It can be seen that, compared with the traditional vision system, the resolution of the vision measurement system based on confocal scanning is increased by more than 50 times under the same telephoto lens. When the optical system remains unchanged, the resolution of the system increases with the increase in the scan sampling ratio. Since the x- and y-direction resolutions of the system are consistent in the full field of view, and the z-direction resolution is consistent at the same x position, the relationship between the system resolution and the scan sampling ratio is shown at y = 0, as shown in Figure 9. Among them, cs0.2 represents the resolution of the traditional vision system with a sampling interval of 0.2 μm. As can be seen from the figure, as long as the scanning sampling ratio is greater than 1.6 × 104, the resolution of the vision measurement system based on confocal scanning is theoretically superior to that of the traditional vision system. When the scanning sampling ratio is greater than 105, the resolution of the system can be broken down to less than 10 μm, and with the increase in the scanning sampling ratio, the theoretical resolution of the system will increase proportionally.
However, as can be seen from the schematic diagram, the function of the photographic lens is to ensure the FOV and measurement distance. In order to ensure that the public FOV is not less than 1.5 m × 1.5 m, the diameter of the FOV of the photographic lens is designed to be not less than 3.5 m, which also leads to the relatively large diameter of the image plane of the photographic lens. Therefore, if we use traditional optical design methods to design the entire optical part to achieve the fusion of vision technology and confocal technology, the aperture of some optical lenses in the instrument will be too large. To solve this problem, we designed the lens in the instrument as a telecentric lens, as shown in Figure 10. This can effectively reduce the aperture of the lens, and the aperture of the lens can be limited to less than 180 mm by a certain design method. In the design, it is necessary to ensure that the FOV and pupil match between the lenses so as to make full use of the performance of each lens.

3. Test Methods and Experiments

3.1. Experimental Equipment

Figure 11 shows the binocular and monocular vision equipment used for confocal scanning imaging. The two monocular vision systems were placed in parallel. The baseline distance was 2 m, the test distance was 2.5 m, and the public FOV exceeded 1.5 m × 1.5 m. The optical lenses used in the equipment were designed independently. The F-number of the photographic lens was 2, and the focal length was 16 mm. According to the principles of pupil matching and FOV matching, the numerical aperture of the objective lens was set to 0.25, and the focal length was designed to be 50 mm. To increase the freedom of the optical path placement, the focal length of the tube lens was designed to be 375 mm, and the diameter of the incident pupil was designed to be 26 mm to match the pupil. To match the FOV and reduce design difficulty, the focal length of the scanning lens was designed to be 225 mm, and the pupil diameter was 15.6 mm. Owing to the 2D galvanometer, the convergent lens was a converging beam on the axis without an off-axis incident beam; therefore, its diameter was designed to be 20 mm.

3.2. Resolution Measuring Equipment and Methods

To verify the resolution of the proposed method in an FOV of no less than 1.5 m × 1.5 m, the size of the object to be measured must be at least 1.5 m × 1.5 m, and a micron-scale 3D displacement must be able to be generated. Controlling such small displacements precisely on a large object is challenging. Hence, a high-precision large-stroke 2D guide rail frame was used to replace large objects, and a high-precision nano-displacement platform was used to achieve high-precision and controllable 3D displacement, as shown in Figure 12. The large 2D guide rail frame was composed of a guide rail and an aluminum alloy frame. The length of each guide rail was 2 m, and a small platform that could shift freely was placed on the guide rail; a magnetic scale was added to the small platform and guide rail to read the position of the small platform. The high-precision micro-displacement platform was a nanometer-scale 3D displacement platform that could exhibit 3D motions in intervals of 50 nm and propagate up to 100 µm.
The experiment was carried out in the laboratory environment and the main measurement conditions were as follows: 1. During the experiment, it is necessary to ensure that the vibration is very small, so a certain vibration isolation condition is required, and the measurement error caused by the fluctuation of the instrument was not more than 1 μm. 2. The test object must have high reflectivity to ensure that the image of the tested object has a high signal-to-noise ratio, so the surface of the object was sprayed with micro-bead reflective film.
After the developed instrument was fixed, the 2D guide rail frame was placed 2.5 m away from the instrument; the center of the frame was located at the center of the instrument’s FOV, and the FOV and resolution were subsequently tested. The test methods and steps were as follows:
  • The target with the characteristic structure was placed on the displacement platform, and the displacement platform was placed in the center of the 2D guide rail frame, which is the center of the FOV of the developed instrument; the reading of the grating ruler was recorded at this time.
  • The position of the small target was adjusted such that the target could be clearly imaged by the developed instrument, and the image of the target was recorded at this time (A).
  • A high-precision micro-displacement platform was used to drive the small target to generate a certain amount of micro-displacement (a μm) along the x-direction, and an image (B) of the target was recorded at this time.
  • The pixel movement of images B and A was calculated using a Zero-normalized cross-correlation (ZNCC)-based image-matching algorithm.
  • Steps 3 and 4 were repeated. If the signs of pixel displacement measured repeatedly are the same, then this implies that the resolution of the measuring instrument in the x-direction is no less than a μm.
  • The y- and z-direction resolutions of the instrument at the center of the FOV were tested using the same method.
  • The position of the target was changed, and the distance between the target and the center of the FOV was ensured to exceed 0.75 m. Steps 2 to 6 were repeated.
By performing the steps above, the resolution of the system could be determined within a 1.5 m × 1.5 m FOV.

3.3. Experiments

3.3.1. Resolution and Field of View Test Experiment

First, the resolution at the center of the FOV was tested, and the target was placed at the center of the FOV. At this time, the magnetic scale read (953.2 mm, 1072.1 mm). The parameters of the left and right visual systems at the center of the FOV were set as follows: the scanning range was approximately 5 mm × 5 mm, and the number of sampling points was 500 × 500. The target was shifted to the right edge of the FOV, and the reading of the magnetic scale was recorded at this time. The distance from the center of the FOV can be calculated as 1707.5 − 953.2 = 754.3 mm. The scanning range was approximately 5 mm × 5 mm, and the number of sampling points was 500 × 500. Similarly, the target was shifted to the upper edge of the FOV, the reading of the magnetic scale was recorded, and the moving distance can be calculated as 1832.5 − 1072.1 = 760.4 mm. The scanning range was approximately 5 mm × 5 mm, and the number of sampling points was 500 × 500. The resolution was tested at only three locations, owing to the circular symmetry of the monocular vision system. Therefore, the actual FOV was 1.521 m × 1.509 m. Finally, the target shifted to the bottom left corner of the FOV, and the distance from the center of the FOV was exactly 760.4 mm and 754.3 mm.
At the center and edge of the FOV, a high-precision displacement platform was used to realize displacements at intervals of 2.5 μm in the x- and y-directions and at intervals of 6 μm in the z-direction. The movement was repeated six times for each position and direction, and the calculation results are listed in Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7. Figure 13 shows an image of the target obtained using the left visual system. Figure 13a,b show the images before and after the object shifted, respectively. Almost no visual change was indicated from the figures; therefore, the imaging results are not shown in subsequent tests. Table 2, Table 3, Table 4 and Table 5 show that for the left and right visual systems, when the target shifted monotonically along the x- and y-directions at intervals of 2.5 μm, the symbols of pixel movement amount obtained by the matching algorithm based on ZNCC were the same, both in the center and at the edge of the FOV. That is, when the target propagated 2.5 μm along the x- or y-direction, the vision system was able to recognize this displacement. This indicates that the resolution of the developed system can reach 2.5 μm in the x- and y-directions. Although the resolution was measured at only three locations, one can conclude that the resolution of the system in the x- or y-direction is 2.5 μm in the full FOV because the system is circularly symmetric. As shown in Table 6 and Table 7, the resolution of the left visual system in the three FOVs in the z-direction can reach 6 μm, but the right visual system cannot distinguish this displacement at the right edge of the FOV. This is because the target is near the optical axis of the right visual system at this time; therefore, it cannot be recognized, which is consistent with the simulation results presented in Section 2.1. Although the right visual system cannot recognize this displacement, the left visual system can. Therefore, based on the definition of the resolution of the visual system, the z-direction resolution of the visual system can reach 6 μm. Furthermore, because the system is symmetrical, the z-resolution of the system in the full FOV can reach 6 μm. Therefore, the resolution of the visual system in the full FOV (1.521 m × 1.509 m) can reach 2.5, 2.5, and 6 μm in the x-, y-, and z-directions, respectively. When the measurement field of view is 1.52 m × 1.51 m, the measurement time is 43.7 s.
According to Table 2 and Table 4, the average pixel displacement of the left system in the x- and y-directions can be calculated, and the results are 0.1841 and 0.1667, respectively. The actual displacement corresponding to them is 2.50 μm. Through simple calculation, the actual displacement per pixel can be obtained as 2.5/0.1841 ≈ 13.5796 μm/pixel and 2.5/0.1667 ≈ 14.9970 μm/pixel. The measurement results of resolution can be calculated, as shown in Table 2 and Table 4. Similarly, the average pixel displacement of the right system in the x- and y-directions is 0.1829 and 0.1867, so the corresponding displacement per pixel is 13.6687 μm/ pixel and 13.3905 μm/ pixel. The measurement results of resolution can also be calculated, as shown in Table 3 and Table 5. The average pixel displacement of the left system and the right system in the z-direction is 0.1468 and 0.1720, and the corresponding displacement is 6 μm, so the corresponding displacement per pixel is 40.8719 μm/pixel and 34.8837 μm/pixel. The measurement results of resolution can also be calculated, as shown in Table 6 and Table 7. We used the mean of the measurement results of the left and right systems as the accurate measurement results, as shown in Table 8, Table 9 and Table 10. These three tables represent precise measurements of displacement in the x-, y-, and z-directions, respectively. Through calculation, the maximum measurement errors of the proposed algorithm in the x-, y-, and z-directions are 1.3946 μm, 1.3210 μm and 3.9541 μm, respectively, and the standard deviations are 0.7259 μm, 0.6677 μm, and 2.2936 μm.

3.3.2. Comparative Experiments

To further illustrate the performance of our technique, we perform comparative experiments in this section. The type of vision-measuring instrument owned by our laboratory is FARO cobalt. The measuring field of view of the instrument is 260 mm × 200 mm, and the testing distance is 505 mm, which cannot reach the corresponding measuring field of view and distance of our method. In fact, the resolution is directly related to the test distance and field of view; the closer the test distance and the smaller the field of view, the higher the resolution of the visual system. Therefore, we do not directly use resolution for comparison but use relative resolution (resolution/field of view), just like the resolution indicators given by some commercial vision instruments.
Due to the different illumination methods, we could not use the previous target as the test object for the FARO instrument. So, in this experiment, we used the standard round ball as the test object, and the resolution test procedure was the same as that in Section 3.2. In order to increase the reliability of the experiment, the software of the FARO instrument was used to calculate the spherical center, and the resolution of the system was judged by the changing law of the position of the spherical center. In this experiment, we only tested the resolution at the center and edge of the field of view. In addition, according to the previous analysis, the resolution of the visual system in the x-direction and y-direction is the same or similar, so only the resolution of the system in the x- and z-direction is tested in this experiment. The round ball captured by the system is shown in Figure 14. The FARO instrument takes about 13 s to complete a measurement. A high-precision displacement platform was used to realize displacements at intervals of 2.5 μm in the x-direction and at intervals of 5 μm in the z-direction. The movement was repeated six times for each position and direction, and the calculation results are listed in Table 11 and Table 12. Table 11 shows that for the FARO visual system, when the target shifted monotonically along the x-direction at intervals of 2.5 μm, the vision system was able to recognize this displacement. That is, its relative resolution is 2.5/260,000 ≈ 0.0000096 × FOV. Its relative resolution in the y-direction is also approximately equal to this value. This indicates that the relative resolution of the FARO system can reach 0.0000096 × FOV in the x- and y-directions. Similarly, it can be seen from Table 12 that for the FARO vision system, its relative resolution in the z-direction is 5/260,000 ≈ 0.000019 × FOV. For our system, the relative resolution in the x- and y-directions is 2.5/1,500,000 ≈ 0.0000017 × FOV and the relative resolution in the z-direction is 6/1,500,000 = 0.000004 × FOV. In other words, our system has 5.8 times better resolution in the x- and y-directions and 4.75 times better resolution in the z-direction compared to the FARO system.

3.3.3. Deformation Test Experiment

To further demonstrate the practicability of the method studied in this paper, UAV was taken as the measurement object. One end of the wing of the drone was fixed, and the other end was fixed on the displacement platform. When the displacement table moved axially, the wing appeared to have continuous deformation due to the elasticity of the wing. The UAV was placed in the center of the field of view, and the confocal vision system was used to image the UAV. The results are shown in Figure 15b,c. Then, using the semi-global stereo matching algorithm, the parallax map of the UAV can be calculated, as shown in Figure 15d. It can be seen from the parallax map that there are some mismatching points, which are mainly caused by occlusion, but this is not the content of this paper.
When the displacement table moved 6 μm axially, one end of the wing moved 6 μm. Since the other end of the wing was fixed, the wing underwent continuous deformation. When measuring continuous deformation, the deformation can be measured point by point, or it can be obtained by measuring several points and fitting. Here, in order to save time, we chose two positions for calculation, which are defined as positions 1 and 2, as shown in Figure 15d. The size of each position is 100 px × 100 px. Since position 1 is far away from the displacement table and position 2 is on the displacement table, the deformation at position 2 should be greater than that at position 1. It should be noted that measuring only once may not be accurate enough due to the effects of vibration and noise, so we measured five times after deformation. We took out part of the images corresponding to position 1 and position 2 and then used the image-matching algorithm to match them with the measured images before deformation, respectively, and calculated the pixel displacement, as shown in Table 13. As can be seen from Table 13, the five calculation results at position 1 are different signs, indicating that the deformation at this position cannot be distinguished, so the deformation there is set to 0. The calculation results at position 2 are of the same sign, so it is believed that the deformation there can be distinguished. As can be seen from Table 13, the average pixel movement is 0.278. Considering that the wing deformation is continuous, it is assumed that the deformation conforms to a quadric surface, so the wing deformation diagram can be obtained through fitting, as shown in Figure 15e. To be more intuitive, only one line from position 1 to position 2 is shown. Therefore, the developed system can effectively identify the small continuous deformation in the axial direction of 6 μm, which is consistent with the resolution experimental results.

4. Conclusions

In this paper, we propose a new large-field and high-resolution measurement technology—a vision measurement technology based on confocal scanning imaging. Firstly, we built a 3D resolution model of the visual system, through which we analyzed the factors that affect the resolution of the visual system, especially the 3D resolution of the parallel optical axis binocular vision system. Then, the resolution model was used to find a way to improve the resolution of the visual system. Combining confocal scanning imaging technology with vision technology, confocal scanning imaging can effectively compress the sampling interval and achieve a breakthrough in resolution. Finally, the corresponding optical system was designed and the field of view and resolution were measured. The test results show that when the test distance is 2.5 m, the field of view of the developed system can reach 1.521 m × 1.509 m, and the three-dimensional resolutions are 2.5 μm, 2.5 μm, and 6 μm, respectively.
In theory, the smaller the sampling interval, the higher the resolution; however, due to noise, vibration, and other reasons, the resolution cannot be infinitely improved. Therefore, it is necessary to analyze the effects of noise and vibration in order to further improve the resolution. And our technology requires scanning of objects, so the measurement efficiency is lower than traditional vision technology. Therefore, it is necessary to research how to improve the measurement efficiency.

Author Contributions

Y.L. and J.L. proposed the confocal vision technology. X.Y. and C.L. conducted the experiments and wrote the manuscript under the supervision of Y.L. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China under Grant No. 51975159, the National Key Research and Development Program of China under Grant No.2021YFF0700400, the Foundation Strengthening Program—Key Basic Research Projects under Grant No. 2019JCJQZD38500, and the National Defense Basic Research Program under Grant No. JCKY2020208B021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, upon reasonable request.

Acknowledgments

We would like to express our thanks to anonymous reviewers for their valuable and insightful suggestions, which helped us to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Deng, H.; Wang, J.; Zhang, J.; Liang, C.-J.; Ma, M.-C.; Zhong, X.; Yu, L.-D. A stereovision measurement for large deformation of light structures. Measurement 2019, 136, 387–394. [Google Scholar] [CrossRef]
  2. Cai, L.; He, L.; Xu, Y.; Zhao, Y.; Yang, X. Multi-object detection and tracking by stereo vision. Pattern Recognit. 2010, 43, 4028–4041. [Google Scholar] [CrossRef]
  3. Ren, J.; Jian, Z.; Wang, X.; Mingjun, R.; Zhu, L.; Jiang, X. Complex surface reconstruction based on fusion of surface normals and sparse depth measuremen. IEEE Trans. Instrum. Meas. 2021, 70, 2506413. [Google Scholar] [CrossRef]
  4. Ju, Y.; Jian, M.; Guo, S.; Wang, Y.; Zhou, H.; Dong, J. Incorporating lambertian priors into surface normals measurement. IEEE Trans. Instrum. Meas. 2021, 70, 5012913. [Google Scholar] [CrossRef]
  5. Hu, Y.; Rao, W.; Qi, L.; Dong, J.; Cai, J.; Fan, H. A Refractive Stereo Structured-Light 3-D Measurement System for Immersed Object. IEEE Trans. Instrum. Meas. 2022, 72, 5003613. [Google Scholar] [CrossRef]
  6. Yang, D.S.; Gao, T.H.; Lu, F. Optical three-dimensional shape measurement based on structured light and a binocular vision system. JOSA A 2022, 39, 2009–2015. [Google Scholar] [CrossRef]
  7. Shan, B.; Yuan, W.; Xue, Z. A calibration method for stereovision system based on solid circle target. Measurement 2019, 132, 213–223. [Google Scholar] [CrossRef]
  8. Zhang, C.; Zhang, X.; Tu, D.; Jin, P. On-site calibration of underwater stereo vision based on light field. Opt. Lasers Eng. 2019, 121, 252–260. [Google Scholar] [CrossRef]
  9. Chen, Z.; Wang, R.; Ji, W.; Zong, M.; Fan, T.; Wang, H. A novel monocular calibration method for underwater vision measurement. Multimed. Tools Appl. 2019, 78, 19437–19455. [Google Scholar] [CrossRef]
  10. Liu, X.; Liu, Z.; Duan, G.; Cheng, J.; Jiang, X.; Tan, J. Precise and robust binocular camera calibration based on multiple constraints. Appl. Appl. Opt. 2018, 57, 5130–5140. [Google Scholar] [CrossRef]
  11. Meng, Z.; Zhang, H.; Guo, D.; Chen, S.; Huo, J. Defocused calibration for large field-of-view binocular cameras. Autom. Constr. 2023, 147, 104737. [Google Scholar] [CrossRef]
  12. Yin, Z.; Xiong, J. Stereovision measurement of layer geometry in wire and arc additive manufacturing with various stereo matching algorithms. J. Manuf. Process. 2020, 56, 428–438. [Google Scholar] [CrossRef]
  13. Zhang, J.; Zhang, Y.; Wang, C.; Yu, H.; Qin, C. Binocular stereo matching algorithm based on MST cost aggregation. Math. Biosci. Eng. 2021, 18, 3215–3226. [Google Scholar] [CrossRef] [PubMed]
  14. Rao, Y.; Ju, Y.; Wang, S.; Gao, F.; Fan, H.; Dong, J. Learning Enriched Feature Descriptor for Image Matching and Visual Measurement. IEEE Trans. Instrum. Meas. 2023, 72, 5008512. [Google Scholar] [CrossRef]
  15. Yang, G.; Liao, Y. An improved binocular stereo matching algorithm based on AANet. Multimed. Tools Appl. 2023, 82, 1–17. [Google Scholar] [CrossRef]
  16. Zhang, J.; Zhang, P.; Deng, H.; Wang, J. High-accuracy three-dimensional reconstruction of vibration based on stereo vision. Opt. Eng. 2016, 55, 091410. [Google Scholar] [CrossRef]
  17. Xiong, J.; Zhong, S.; Liu, Y.; Tu, L.-F. Automatic three-dimensional reconstruction based on four-view stereo vision using checkerboard pattern. J. Cent. South Univ. 2017, 24, 1063–1072. [Google Scholar] [CrossRef]
  18. Li, J.; Liu, T.; Wang, X. Advanced pavement distress recognition and 3D reconstruction by using GA-DenseNet and binocular stereo vision. Measurement 2022, 201, 111760. [Google Scholar] [CrossRef]
  19. Hu, Y.; Chen, Q.; Feng, S.; Tao, T.; Asundi, A.; Zuo, C. A new microscopic telecentric stereo vision system—Calibration, rectification, and three-dimensional reconstruction. Opt. Lasers Eng. 2019, 113, 14–22. [Google Scholar] [CrossRef]
  20. Brunken, H.; Gühmann, C. Road Surface Reconstruction by Stereo Vision. PFG—J. Photogramm. Remote Sens. Geoinf. Sci. 2020, 88, 433–448. [Google Scholar] [CrossRef]
  21. Hu, Q.; Feng, Z.; He, L.; Shou, Z.; Zeng, J.; Tan, J.; Bai, Y.; Cai, Q.; Gu, Y. Accuracy improvement of binocular vision measurement system for slope deformation monitoring. Sensors 2020, 20, 1994. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Yang, L.; Wang, B.; Zhang, R.; Zhou, H.; Wang, R. Analysis on location accuracy for the binocular stereo vision system. IEEE Photonics J. 2017, 10, 7800316. [Google Scholar] [CrossRef]
  23. Jin, D.; Yang, Y. Sensitivity analysis of the error factors in the binocular vision measurement system. Opt. Eng. 2018, 57, 104109. [Google Scholar] [CrossRef]
  24. Zhang, B.; Zhu, D. Improved Camera Calibration Method and Accuracy Analysis for Binocular Vision. Int. J. Pattern Recognit. Artif. Intell. 2021, 35, 2155010. [Google Scholar] [CrossRef]
  25. Ning, S.; Zhu, Y.; Lv, X.; Song, H.; Zhang, R.; Zhang, G.; Zhang, L. Analysis and optimization of the performance parameters of non cooperative target location detection system. Optik 2021, 227, 166100. [Google Scholar] [CrossRef]
  26. Gong, H.; Zhang, F.; Ji, Q. Analysis of the effective field of view and resolution on the binocular measuring system of contact wire’s parameters. Mechnical 2012, 39, 55–59. [Google Scholar] [CrossRef]
  27. Li, Z.; Li, H.; Zhang, Z.; Liu, X. Space Resolution and Structural Parameters of Stereo Vision System. Opto-Electron. Eng. 2012, 39, 48–53. [Google Scholar] [CrossRef]
  28. Abbasi, M. Polymer blends analyzed with confocal laser scanning microscopy. Polym. Bull. 2023, 80, 5929–5964. [Google Scholar] [CrossRef]
  29. Chang, S.; Hong, Y. Three-dimensional confocal reflectance microscopy for surface metrology. Meas. Sci. Technol. 2021, 32, 102002. [Google Scholar] [CrossRef]
  30. Gao, P.; Ulrich, N. Confocal laser scanning microscopy with spatiotemporal structured illumination. Opt. Lett. 2016, 41, 1193–1196. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of binocular vision.
Figure 1. Schematic diagram of binocular vision.
Sensors 23 06615 g001
Figure 2. Z-resolution of parallel optical axis vision system. (a) The x-direction component of the z-direction displacement, (b) the y-direction component of the z-direction displacement, (c) z-direction resolution of the left system, and (d) z-resolution of parallel optical axis vision system.
Figure 2. Z-resolution of parallel optical axis vision system. (a) The x-direction component of the z-direction displacement, (b) the y-direction component of the z-direction displacement, (c) z-direction resolution of the left system, and (d) z-resolution of parallel optical axis vision system.
Sensors 23 06615 g002
Figure 3. Effect of baseline distance on Z-direction resolution.
Figure 3. Effect of baseline distance on Z-direction resolution.
Sensors 23 06615 g003
Figure 4. Effect of optical axis inclinations on resolution. (a) The effect of optical axis inclinations on x-direction resolution, (b) the effect of optical axis inclinations on y-direction resolution, and (c) the effect of optical axis inclinations on z-direction resolution.
Figure 4. Effect of optical axis inclinations on resolution. (a) The effect of optical axis inclinations on x-direction resolution, (b) the effect of optical axis inclinations on y-direction resolution, and (c) the effect of optical axis inclinations on z-direction resolution.
Sensors 23 06615 g004
Figure 5. Calibration figure of the field of view.
Figure 5. Calibration figure of the field of view.
Sensors 23 06615 g005
Figure 6. Location figure of the target.
Figure 6. Location figure of the target.
Sensors 23 06615 g006
Figure 7. Schematic diagram of monocular vision system based on confocal scanning imaging.
Figure 7. Schematic diagram of monocular vision system based on confocal scanning imaging.
Sensors 23 06615 g007
Figure 8. Z-resolution of confocal binocular vision system.
Figure 8. Z-resolution of confocal binocular vision system.
Sensors 23 06615 g008
Figure 9. 3D resolution of the system with different scan sampling ratios. (a) x-/y-resolution of the system with different scan sampling ratios, (b) z-resolution of the system with different scan sampling ratios.
Figure 9. 3D resolution of the system with different scan sampling ratios. (a) x-/y-resolution of the system with different scan sampling ratios, (b) z-resolution of the system with different scan sampling ratios.
Sensors 23 06615 g009
Figure 10. Confocal scanning imaging vision system based on telecentric connection.
Figure 10. Confocal scanning imaging vision system based on telecentric connection.
Sensors 23 06615 g010
Figure 11. Equipment diagram of a vision system based on confocal scanning imaging.
Figure 11. Equipment diagram of a vision system based on confocal scanning imaging.
Sensors 23 06615 g011
Figure 12. Test system for field of view and resolution. (a) Aluminum frame with high-precision and large-stroke 2-D guide rail, and (b) 3D high-precision nano-displacement platform.
Figure 12. Test system for field of view and resolution. (a) Aluminum frame with high-precision and large-stroke 2-D guide rail, and (b) 3D high-precision nano-displacement platform.
Sensors 23 06615 g012
Figure 13. Image of the target. (a) Image before moving; (b) image after moving 2.5 μm.
Figure 13. Image of the target. (a) Image before moving; (b) image after moving 2.5 μm.
Sensors 23 06615 g013
Figure 14. Image of the round ball.
Figure 14. Image of the round ball.
Sensors 23 06615 g014
Figure 15. UAV test results. (a) The UAV, (b) imaging results of the left system, (c) imaging results of the right system, (d) disparity map, and (e) curve of wing deformation.
Figure 15. UAV test results. (a) The UAV, (b) imaging results of the left system, (c) imaging results of the right system, (d) disparity map, and (e) curve of wing deformation.
Sensors 23 06615 g015
Table 1. Resolution of the vision system.
Table 1. Resolution of the vision system.
Number of Moves12345
x-resolutionDisplacement (pixel)0.51720.50440.50260.50450.5038
y-resolutionDisplacement (pixel)0.54240.53690.53860.53700.5450
z-resolutionDisplacement (pixel)0.54990.56820.56310.56590.5641
Table 2. X-direction resolution of the left system at the center and edge of the FOV.
Table 2. X-direction resolution of the left system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)0.20380.08550.17710.25110.1901
Displacement (μm)2.76751.16112.40493.40982.5815
Upper edge of FOVDisplacement (pixel)0.13500.10320.27450.15390.1965
Displacement (μm)1.83321.40143.72762.08992.6684
Right edge of FOVDisplacement (pixel)0.09690.28640.18350.13800.2269
Displacement (μm)1.31593.88922.49191.87403.0812
Lower left of FOVDisplacement (pixel)0.15210.21130.16630.26320.1857
Displacement (μm)2.06552.86942.25833.57422.5217
Table 3. X-direction resolution of the right system at the center and edge of the FOV.
Table 3. X-direction resolution of the right system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)−0.1856−0.0768−0.2059−0.2356−0.2727
Displacement (μm)2.53691.04982.81443.22043.7275
Upper edge of FOVDisplacement (pixel)−0.2745−0.1353−0.2126−0.0850−0.1773
Displacement (μm)3.75211.84942.90601.16182.4235
Right edge of FOVDisplacement (pixel)−0.1369−0.2539−0.1566−0.1234−0.1482
Displacement (μm)1.87133.47052.14051.68672.0257
Lower left of FOVDisplacement (pixel)−0.1022−0.1689−0.2698−0.2553−0.1812
Displacement (μm)1.39692.30863.68783.48962.4768
Table 4. Y-direction resolution of the left system at the center and edge of the FOV.
Table 4. Y-direction resolution of the left system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)−0.1473−0.0567−0.2336−0.1577−0.1250
Displacement (μm)2.20910.85033.50332.36501.8746
Upper edge of FOVDisplacement (pixel)−0.2036−0.1583−0.0921−0.2123−0.1752
Displacement (μm)3.05342.37401.38123.18392.6275
Right edge of FOVDisplacement (pixel)−0.1033−0.1851−0.1425−0.2036−0.2567
Displacement (μm)1.54922.77592.13713.05343.8497
Lower left of FOVDisplacement (pixel)−0.2231−0.1187−0.1361−0.2482−0.1555
Displacement (μm)3.34581.78012.04113.72232.3320
Table 5. Y-direction resolution of the right system at the center and edge of the FOV.
Table 5. Y-direction resolution of the right system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)−0.2020−0.1126−0.2859−0.1743−0.2216
Displacement (μm)2.70451.50783.82832.33402.9673
Upper edge of FOVDisplacement (pixel)−0.3025−0.2718−0.1478−0.2356−0.0636
Displacement (μm)4.05063.63951.97913.15480.8516
Right edge of FOVDisplacement (pixel)−0.2985−0.1356−0.1026−0.1461−0.1510
Displacement (μm)3.99711.81581.37391.95642.0220
Lower left of FOVDisplacement (pixel)−0.1993−0.2512−0.1386−0.1915−0.1011
Displacement (μm)2.66873.36371.85592.56431.3538
Table 6. Z-direction resolution of the left system at the center and edge of the FOV.
Table 6. Z-direction resolution of the left system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)−0.0431−0.1157−0.2651−0.0968−0.1711
Displacement (μm)1.76164.728910.83513.95646.9932
Upper edge of FOVDisplacement (pixel)−0.1164−0.1562−0.2126−0.0215−0.2243
Displacement (μm)4.75756.38428.68940.87879.1676
Right edge of FOVDisplacement (pixel)−0.1816−0.1531−0.1082−0.2347−0.1013
Displacement (μm)7.42236.25754.42239.59264.1403
Lower left of FOVDisplacement (pixel)−0.1210−0.08520.13320.0878−0.1026
Displacement (μm)\\\\\
Table 7. Z-direction resolution of the right system at the center and edge of the FOV.
Table 7. Z-direction resolution of the right system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (pixel)0.06680.14250.20410.08110.2988
Displacement (μm)2.33024.97097.11982.829110.4233
Upper edge of FOVDisplacement (pixel)0.17780.21570.26910.11510.1373
Displacement (μm)6.20237.52449.38724.01514.7895
Right edge of FOVDisplacement (pixel)0.1132−0.1054−0.05890.15330.0637
Displacement (μm)\\\\\
Lower left of FOVDisplacement (pixel)0.18820.21250.10590.12820.2371
Displacement (μm)6.56517.41283.69424.47218.2709
Table 8. Accuracy measurements of x-direction resolution at the center and edge of the FOV.
Table 8. Accuracy measurements of x-direction resolution at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (μm)2.65221.10542.60973.31513.1545
Upper edge of FOVDisplacement (μm)2.79271.62543.31681.62592.5459
Right edge of FOVDisplacement (μm)1.59363.67982.31621.78042.5535
Lower left of FOVDisplacement (μm)1.73122.58902.97313.53192.4993
Table 9. Accuracy measurements of y-direction resolution at the center and edge of the FOV.
Table 9. Accuracy measurements of y-direction resolution at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (μm)2.45701.17913.66582.34952.4210
Upper edge of FOVDisplacement (μm)3.55203.00681.68013.16931.7396
Right edge of FOVDisplacement (μm)2.77312.29591.75552.50492.9359
Lower left of FOVDisplacement (μm)3.00732.57191.94853.14331.8429
Table 10. Accuracy measurements of z-direction resolution at the center and edge of the FOV.
Table 10. Accuracy measurements of z-direction resolution at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (μm)2.04594.84998.97753.39278.7082
Upper edge of FOVDisplacement (μm)5.47996.95439.03832.44696.9786
Right edge of FOVDisplacement (μm)7.42236.25754.42239.59264.1403
Lower left of FOVDisplacement (μm)6.56517.41283.69424.47218.2709
Table 11. X-direction resolution of the FARO system at the center and edge of the FOV.
Table 11. X-direction resolution of the FARO system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (μm)2.004.001.004.003.00
Edge of FOVDisplacement (μm)4.004.001.005.002.00
Table 12. Z-direction resolution of the FARO system at the center and edge of the FOV.
Table 12. Z-direction resolution of the FARO system at the center and edge of the FOV.
Number of Moves12345
Center of FOVDisplacement (μm)7.004.006.008.005.00
Edge of FOVDisplacement (μm)3.009.003.002.004.00
Table 13. The deformation test of UAV wing.
Table 13. The deformation test of UAV wing.
Number of Measurements 12345
Position 1Deformation (pixel)left system−0.11330.15160.1308−0.11760.0532
right system0.18540.0887−0.1023−0.20010.1682
Position 2Deformation (pixel)left system−0.1512−0.0885−0.1362−0.2431−0.1147
right system0.09890.10350.08760.15530.2113
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Liu, C.; You, X.; Liu, J. A New Vision Measurement Technique with Large Field of View and High Resolution. Sensors 2023, 23, 6615. https://doi.org/10.3390/s23146615

AMA Style

Li Y, Liu C, You X, Liu J. A New Vision Measurement Technique with Large Field of View and High Resolution. Sensors. 2023; 23(14):6615. https://doi.org/10.3390/s23146615

Chicago/Turabian Style

Li, Yong, Chenguang Liu, Xiaoyu You, and Jian Liu. 2023. "A New Vision Measurement Technique with Large Field of View and High Resolution" Sensors 23, no. 14: 6615. https://doi.org/10.3390/s23146615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop