Next Article in Journal
LiDAR Echo Gaussian Decomposition Algorithm for FPGA Implementation
Previous Article in Journal
HARNU-Net: Hierarchical Attention Residual Nested U-Net for Change Detection in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Global Fundamental Matrix Estimation Method of Planar Motion Based on Inlier Updating

School of Electrical Engineering and Automation, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4624; https://doi.org/10.3390/s22124624
Submission received: 8 April 2022 / Revised: 28 May 2022 / Accepted: 15 June 2022 / Published: 19 June 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
A fundamental matrix estimation based on matching points is a critical problem in epipolar geometry. In this paper, a global fundamental matrix estimation method based on inlier updating is proposed. Firstly, the coplanar constraint was incorporated into the solution of the fundamental matrix to reduce the number of parameters to be solved. Subsequently, an inlier updating matrix was introduced according to the threshold of the epipolar geometry distance to eliminate the potential outliers and obtain a reliable initial value of the fundamental matrix. On this basis, we employed a four-point iterative method to estimate the fundamental matrix and make it satisfy the rank constraint at the same time. Finally, the epipolar geometry in binocular vision was extended to triple-view, and the fundamental matrix obtained in the previous step was globally optimized by minimizing the coordinate deviation between the intersection point and feature point in each group of images. The experiments show that the proposed fundamental matrix estimation method is robust to noise and outliers. In the attitude measurement, the maximum static error was 0.104° and dynamic measurement error was superior to 0.273°, which improved the reconstruction accuracy of feature points. Indoor images were further used to test the method, and the mean rotation angle error was 0.362°. The results demonstrate that the estimation method proposed in this paper has a good practical application prospect in multi-view 3D reconstruction and visual localization.

1. Introduction

The estimation of a fundamental matrix is a basic step for computer vision applications and is widely used in visual localization, camera calibration [1,2], object recognition, motion analysis [3,4], 3D reconstruction [5,6], etc. Specifically, with the rapid development of 3D reconstruction technology, the demand and applications of 3D models in virtual reality, digital twin, the metaverse and other fields are rapidly growing [7,8]. The 3D reconstruction method using the images has been widely researched. In multi-view 3D reconstructions based on monocular vision, local invariance is used to detect image feature information and match feature points, which can be used to recover camera position, direction and scene structure. The matching of feature points and the relative pose between cameras directly affect the accuracy of reconstruction [9,10,11]. In robotics, the estimation of the fundamental matrix is the premise and foundation of visual localization. This accurately estimated fundamental matrix can be used to effectively construct the matching between feature points, and the relative pose between cameras is included in the fundamental matrix. Therefore, it is of great significance to develop a high precision and robust estimation method of the fundamental matrix [12,13].
The estimation accuracy of the fundamental matrix is mainly related to the extraction and matching accuracy of the feature points [14,15,16]. Among them, the extraction error is usually caused by the noise. When the feature points are not accurately matching, a small number of outliers will seriously affect the estimation accuracy of the fundamental matrix. To solve the above problems, many fundamental matrix estimation methods have emerged in recent years, including the linear method, iterative method, and robust method.
The linear method mainly includes a seven-point method, eight-point method, improved eight-point method, etc. [17,18]. These methods use least square and singular value decomposition to estimate the fundamental matrix by solving a set of linear equations. Under the conditions that feature point extraction and matching are accurate, linear methods are usually more efficient. However, the accuracy of linear methods is seriously affected when correspondences are abnormal. The iterative method can be divided into two categories: one is based on minimizing the epipolar geometric distance and the other is based on gradient [19,20]. Compared with the linear method, the iterative method improves the estimation accuracy and effectively reduces the influence of noise, although this method has a high computational complexity and is not suitable for cases with many outliers. Comparatively, the robust method has been widely used and studied for its advantages of eliminating outliers and strong anti-noise ability. Typically, M-estimator [21,22], LMedS [23,24] and RANSAC [25,26] are considered to be the most effective robust methods. To obtain reliable results, these methods usually screen matching points before calculating the fundamental matrix and take out the matching points with small geometric errors as inliers. Then, these inliers are used to estimate the fundamental matrix. Compared with these three methods, the M-estimator method reduces the influence of outliers by assigning different weights to each point, but this method has higher requirements for the initial values. The LMedS method uses the median distance from the point to the corresponding epipolar line to optimize the fundamental matrix, but this kind of method is quite time-consuming. The RANSAC method is used to estimate the fundamental matrix by iteratively selecting inliers. However, as the proportion of outliers increases, the computational efficiency significantly decreases.
To further improve the robustness and computational efficiency of the fundamental matrix estimation, many scholars made some improvements to the methods mentioned above. For example, Chum [27] proposed the randomized RANSAC method, which effectively reduces the amount of computation. The authors of [28,29] used different geometric constraints to iteratively solve the fundamental matrix. Moreover, Xiao et al. [30] proposed a fundamental matrix estimation method based on inlier set sample optimization. By adopting guided sampling and local optimization, this method can effectively deal with outliers. The above methods make up for the deficiency of the traditional robust method to some extent, but they also have a fatal problem, that is, the accuracy of these methods deteriorates sharply with the increase in outlier ratio. Later, Yan et al. [31] proposed a robust fundamental matrix estimation method based on epipolar geometric error criterion, which eliminated the outliers in the calculation process of the fundamental matrix and improved the calculation efficiency; however, the calculation accuracy of such methods needs to be further improved.

2. Related Work in Robotic Applications

Visual localization and multi-view 3D reconstruction of planar motion robots are important to indoor service robotic applications. Visual localization utilizes feature correspondences in the environment to recover the camera pose, and the relative pose between cameras can be used in 3D reconstruction. Among them, the fundamental matrix obtained by point correspondence can provide more constraints for camera pose estimation [32,33]. He et al. [34] systematically introduced the latest achievements of robot navigation in the literature and studied the simplified fundamental matrix of planar motion. Jiao et al. [35] introduced the fundamental matrix into 3D–2D correspondences to calculate the camera pose, which incorporated a planar motion constraint and enhanced the robustness of pose estimation. Choi et al. [36] proposed a 2-point non-iterative method based on epipolar geometry under a planar motion to estimate the camera pose. Choi et al. [37] proposed a minimal solver for pose estimation by establishing the fundamental matrix constraint and perspective projection function under a planar motion. Wang et al. [38] presented the importance of the fundamental matrix in robot active vision technology, and Dong et al. [39] devised a collaborative dense scene reconstruction method for multi-planar motion robot. These cases reduce the complexity of the algorithm and improve visual localization accuracy and 3D reconstruction accuracy by exploiting the fundamental matrix constraint under planar motion, but they do not conduct a thorough study on the solution of the fundamental matrix. The fundamental matrix is estimated by the traditional method.
Motivated by the above analysis, this paper proposes a minimal solution for estimating the fundamental matrix from point correspondence by taking the planar motion constraint into consideration. Among the existing methods, the linear method and iterative method are usually poor in robustness and not suitable for cases with many outliers. Although the robust method can eliminate outliers, it has the problem of unstable calculation when the outlier proportion is high. In addition, the number of iterations exponentially increases with the number of inliers required. When the camera moves on the plane, the accuracy of the traditional fundamental matrix estimation method decreases, and the required number of inliers is relatively large. The main contribution of this study is that it addresses the problem of a fundamental matrix estimation from multi-view images under planar motion, a robust null space estimation method is proposed. On this basis, the four-point iterative method is derived. Thus, the problem of rank constraint and local optimum caused by inlier random selection and an artificial threshold in the process of fundamental matrix estimation is successfully solved. Furthermore, the global optimization function of the fundamental matrix further improves the estimation accuracy.
The remainder of this paper is organized as follows. Section 3 introduces the epi-polar geometry in binocular vision and the traditional robust fundamental matrix estimation method. In Section 4, the global fundamental matrix estimation method based on the inlier updating of planar motion is proposed. Simulations and practical experiments are provided to verify the robustness and effectiveness of the proposed method in Section 5, and finally, the paper is concluded in Section 6.

3. Fundamental Matrix Estimation Method Based on RANSAC Method

As shown in Figure 1, the projection points of a 3D point P w i on image planes are m i and m i , respectively. O 1 , O 2 indicate optical centers of the cameras, and the line between them is the baseline. The baseline intersects with image planes at two points e 1 and e 2 , which are called epipoles. The plane formed by P w i , O 1 and O 2 is defined as the epipolar plane Π , which intersects with image planes at lines l and l . l is the epipolar line corresponding to m i on the image plane I , and m i is on the line l . Similarly, l is the epipolar line corresponding to m i on the image plane I . We define the constraint as epipolar geometry in binocular vision.
Suppose that the homogenous coordinates of the projection points can be expressed as:
m i = [ u i v i 1 ] T m i = [ u i v i 1 ] T
According to the epipolar geometry constraint:
m i T F m i = 0 .
The fundamental matrix F is defined as:
F = [ F 11 F 12 F 13 F 21 F 22 F 23 F 31 F 32 1 ] .
By combining Equations (1)–(3), we develop the following equation:
U f = 0
where:
U = [ u 1 u 1 u 1 v 1 u 1 v 1 u 1 v 1 v 1 v 1 u 1 v 1 1 u n u n u n v n u n v n u n v n v n v n u n v n 1 ] .
f = [ F 11 F 12 F 13 F 21 F 22 F 23 F 31 F 32 1 ] T .
The solution of the fundamental matrix is transformed into the process of solving the least squares of min U f with f = 1 . Due to the influence of noise and mismatching, the fundamental matrix cannot be directly obtained by solving Equation (4). Therefore, in this case, we often use the RANSAC method to estimate the fundamental matrix.
The specific calculation process is as follows: Firstly, eight pairs of matching points are randomly selected from correspondences of feature point pairs, and Equation (4) is utilized to solve the fundamental matrix. Then, the points with geometric distances of less than the designed threshold value are judged as the inliers. By repeating the first two steps, the model with the largest number of inliers is selected as the fundamental matrix. The accuracy of the fundamental matrix estimation method based on RANSAC method depends on the proportion of inliers. When the ratio of inliers is low, it is difficult to find enough inlier sets to estimate the fundamental matrix, and the uneven distribution of matching points selected by random sampling will also affect the accuracy and stability of the estimation.

4. The Improved Fundamental Matrix Estimation Method of Planar Motion

4.1. Robust Null Space Estimation Method Based on Inlier Updating

The fundamental matrix of planar motion is analyzed in this section and the simplified fundamental matrix contains the constraint of coplanarity.
As shown in Figure 2, when the camera moves on the horizontal plane, the relative rotation and translation relationship between camera coordinate systems at location 1 and location 2 can be expressed as:
R = [ cos φ 0 sin φ 0 1 0 sin φ 0 cos φ ] ,   t = [ sin θ 0 cos θ ] .
where φ is the rotation angle and θ is the direction of translation.
The essential matrix E is defined as:
E = t × R .
By substituting Equation (7) into Equation (8), the following equation can be obtained:
E = [ 0 cos φ 0 cos ( φ θ ) 0 sin ( φ θ ) 0 sin θ 0 ] .
Assuming that the intrinsic matrix of the camera is K, and the intrinsic parameters remain constant during camera movement, then K can be expressed as follows:
K = [ f x s x 0 0 f y y 0 0 0 1 ] .
Here, f x and f y are the normalized focal length on the X axis and Y axis; ( x 0 , y 0 ) is the principal point coordinate; s is the non-perpendicular factor between the X axis and Y axis.
According to the relationship between fundamental matrix and essential matrix, we obtain:
F = K T E K 1 = [ 0 f 1 f 2 f 3 0 f 4 f 5 f 6 f 7 ] .
For Equation (11):
d = 1 / f x f y f 1 = d cos θ f 2 = d y 0 cos θ f 3 = d cos ( φ θ ) f 4 = d f x sin ( φ θ ) d x 0 cos ( φ θ ) f 5 = d y 0 cos ( φ θ ) f 6 = d f x sin θ + d x 0 cos θ f 7 = d x 0 y 0 ( cos ( φ θ ) cos θ ) d f x y 0 ( sin ( φ θ ) + sin θ )
By substituting Equation (11) into Equation (2), we obtain:
M f = 0 .
where:
M = [ u 1 v 1 u 1 v 1 u 1 v 1 u 1 v 1 1 u n v n u n v n u n v n u n v n 1 ] ,
f = [ f 1 , f 2 , f 3 , f 4 , f 5 , f 6 , f 7 ] T .
When there is no noise in the point correspondence, the rank of the null-space of M is 1. However, affected by noise and outliers, the rank of null-space does not equal one and M has no zero singular values. Motivated by the paper [40], to eliminate the outliers, let the matrix L represent the matrix M without noise, and the process of solving L can be converted into the following minimization:
arg min W , L W ( M L ) 2 rank ( L ) = rank ( M ) 1
In Equation (16), W = d i a g ( w 1 , w 2 , , w n ) is an inlier updating matrix. If the i-th pair of matching points are inliers, set w i = 1 , otherwise set w i = 0 . Since L satisfies L f = 0 , Equation (16) can be transformed into a minimization problem of Equation (17):
arg min W , f W M f 2 .
Initially, let us assume that all the corresponding points are inliers, and W is a n × n identity matrix, ξ = Inf . Before iteration, W is updated according to the position similarity of feature points in the images. Subsequently, singular value decomposition is performed on M T W M , and f is the eigenvector associated with the minimum singular value. If the condition ε max > ξ is not met, W and ξ are updated according to Equation (18). Ultimately, W and f are substituted into the four-point iterative method as initial values:
w i = { 1 ε i τ 0 otherwise ,   ξ = ε max ,
ε i = ( 1 ( F m i ) 1 2 + ( F m i ) 2 2 + 1 ( F T m i ) 1 2 + ( F T m i ) 2 2 ) m i F m i .
where ε i denotes the epipolar geometry distance; ε max = Q 25 % ( ε 1 , ε n ) is the lowest quartile of the epipolar geometry distance; and τ = max ( ε max , δ max ) ; δ max is the maximal geometric error.

4.2. Four-Point Iterative Method for Global Fundamental Matrix Estimation

In the estimation method based on traditional RANSAC method, the artificial threshold is usually utilized to iteratively update the model, and the fundamental matrix does not satisfy the rank constraint. Therefore, we used a singular value decomposition correction to solve the rank constraint problem.
In the four-point iterative method, primarily, the average value of the epipolar geometry distance calculated by the initial value f is taken as the threshold. Furthermore, four pairs of matching points are randomly selected from the inliers obtained by the initial value W and substituted into Equation (13):
A f = 0 .
where:
A = [ u 1 v 1 u 1 v 1 u 1 v 1 u 1 v 1 1 u 4 v 4 u 4 v 4 u 4 v 4 u 4 v 4 1 ] .
A is a 4 × 7 matrix, f 1 , f 2 , f 3 are vectors that span the right null space of A. Therefore, f can be expressed as:
f = a f 1 + b f 2 + c f 3 .
These vectors are transformed into the fundamental matrixes and assume to be c = 1 :
F = a F 1 + b F 2 + F 3 .
When the camera moves on the plane, the intrinsic matrix of the camera K remains unchanged, and the rank of the fundamental matrix is 2. According to [26], the fundamental matrix satisfies:
det ( F ) = 0 det ( F + F T ) = 0
By substituting Equation (23) into Equation (24), we obtain:
C [ a 3 , a 2 b , a b 2 , b 3 , a 2 , a b , b 2 , a , b , 1 ] T = 0 .
where C is a parameter matrix constituted by the elements of f 1 , f 2 , f 3 .
The parameters a and b can be solved by the minimum automatic generator method [41], and the fundamental matrix is estimated. Meanwhile, the number of matching points that satisfy the threshold is calculated and then these points are defined as a new inlier set. Eventually, through repeated use of the four-point iterative method, the model with the largest number of inliers is selected as the estimated fundamental matrix, which is stable and satisfies the rank constraint.
The epipolar geometry in triple-view is shown in Figure 3, where the camera captures the object from different angles. The projection points of a 3D feature point P w i on image planes are m i , m i and m i , respectively; O 1 , O 2 , O 3 are the optical centers of the cameras. Three images taken with public view as a group and F 12 , F 32 are the fundamental matrices estimated by the four-point iterative method between the two images, respectively. Then, m i , m i in the first and third images are mapped to the epipolar lines l 1 i = F 12 m i , l 3 i = F 32 m i in the intermediate image, and the intersection point of epipolar lines is set as p i . Due to the influence of noise, there is a distance error between p i and m i . Therefore, F 12 and F 32 are simultaneously optimized by bundle adjustment:
arg min F 13 , F 23 p i m i 2 .
Figure 4 shows the flow chart of the fundamental matrix estimation method in this paper. Firstly, the fundamental matrix is simplified by analyzing the motion characteristics of the camera, and the parameters to be solved are determined. Furthermore, the robust null space estimation method is applied to calculate the initial model of the fundamental matrix, and on this basis, a four-point automatic generator is derived. Ultimately, according to the epipolar geometry in triple-view, global optimization is carried out.

5. Experimental Results and Analysis

Two datasets were used in the experiment: one is the real dataset, and the other is a simulated dataset containing Gaussian noise and outliers. The proposed method is compared with ISSO [30], EGEC [31] and RANSAC [26] methods. In the subsequent experiments, the mean epipolar geometry distance and the mean distance between feature points and the intersection points of epipolar lines are used as the evaluation criteria to evaluate the accuracy of various methods.

5.1. Experiments on the Simulated Dataset

In the simulation test, we generated 300 pairs of corresponding points, which were uniformly distributed in the synthetic images. Gaussian noise with a mean value of 0, the standard deviation of σ and outliers with different proportions were added to simulate real-world conditions. We conducted 100 independent tests, taking the average value as the final result.
In Figure 5a, the standard deviation of Gaussian noise increases from 0 to 2 without adding any outliers to the simulated data. The results show the epipolar geometric distance obtained by various fundamental matrix estimation methods under Gaussian noise. With the increase in noise standard deviation, the accuracy of the fundamental matrix estimated by all methods linearly decreases. Among them, the performance of the RANSAC method and EGEC method seems to be similar, and both of them sharply decline with the increase in noise. In contrast, the method proposed in this paper deteriorates slowly with the increase in noise. The experimental results in Figure 5b show that the other three methods have a stronger robustness to outliers than the RANSAC method when the corresponding points are noise-free. Moreover, it is clear that the epipolar geometry distance of the proposed method is essentially independent from the proportion of outliers.
In Figure 3, p i should coincide with m i . Gaussian noise and outliers are added to the simulation points in Figure 6a,b as Figure 5, respectively. As can be seen from Figure 6, under the same experimental conditions, the distance between p i and m i obtained by the proposed method is the smallest, which indicates that this method has a good anti-interference performance and can better adapt to the uncertainty of noise and outliers.
In the second simulation experiment, the proportion of outliers is set at 10%, while Gaussian noise with mean value of 0 and standard deviation from 0 to 2 is added to the synthetic data. The experimental results of four methods for synthetic data are summarized in Table 1 and Table 2. When the ratio of outliers is fixed, the errors of the four estimation methods increase with the noise intensity. However, in comparison, the proposed method has the best performance when there are outliers and noises in the dataset.

5.2. Experiments on Real Dataset

Next, in order to verify the effectiveness of the proposed method under planar motion scenes, we randomly selected three images in the Middlebury dataset [42]. The images are taken from equally spaced viewpoints along the x-axis of camera coordinate system from left to right. Meanwhile, to further evaluate the performance of different fundamental matrix estimation methods in computer vision fields, we added a set of satellite images, as shown in Figure 7d, and a sequence of indoor images in the real experimental environment, as shown in Figure 7e,f.
The satellite was placed on the turntable and images were taken from nine perspectives from equiangular viewpoints. In the process, the camera was horizontal. The sequence of indoor images was collected by a camera mounted on a mobile robot horizontally, and the pose of the camera was provided by the OptiTrack system. Feature points were obtained by applying a Harris corner detector, and the correspondence of feature points was obtained by the optical flow method. The fundamental matrix was estimated by the proposed method, ISSO and EGEC methods.
Figure 7 shows the inlier feature points in the first and the intermediate images of the proposed method, which are marked as red ‘*‘. Additionally, the epipolar lines are recovered using the fundamental matrix estimated by the proposed method. As we can see, feature points fall exactly on the corresponding epipolar lines, indicating that the proposed method is relatively accurate for the estimation of the fundamental matrix. The mean epipolar geometry distance is shown in Table 3.
Figure 8 shows the inlier feature points in the intermediate images, which are labeled by ‘o’, and intersection points are labeled by ‘+’. It can be clearly seen that the intersection points and feature points calculated by the proposed method correctly coincide. Table 4 lists the mean distance between the intersection points and feature points calculated by each method. Compared with the other two methods, the proposed method is more robust.
From the above experiments, it can be included that the proposed fundamental matrix estimation method has a higher accuracy under different planar motion scenes. In order to display the application of the fundamental matrix estimation method in the field of 3D reconstruction and visual localization more intuitively, we designed the satellite reconstruction and indoor localization experiments.
In 3D reconstruction, the static measurement and dynamic measurement are devised to verify the reconstruction accuracy of feature points by calculating the rotation angle of the satellite. Firstly, we extract the feature points in the satellite dataset images and obtain the correspondences of feature points. The camera coordinate system in the first image is defined as the world coordinate system by default. The fundamental matrix estimation method proposed in this paper and the ISSO method, respectively, are used to reconstruct the feature points on the satellite and establish the 3D feature point models of the satellite. Figure 9 shows the satellite 3D feature point model obtained by our method and the relative position relationship between camera coordinate systems.
Afterwards, the turntable is rotated every 2° for a total of 10°, and the control accuracy is 0.010°. The satellite 3D feature point models obtained by the proposed method and ISSO method are used to measure the static rotation angle by applying the perspective-n-point algorithm. The measured static errors are shown in Table 5. It can be seen that the error gradually increases with the increase in the rotation angle, and the estimation method proposed in this paper can control the maximum error of the rotation angle within 0.104°, which is better than the result of the ISSO method (0.138°).
The motion trajectory of the turntable is a cosine curve with amplitudes rising from 0° to 10°, and the average speed of the turntable is 1°/s. Figure 10 shows the dynamic errors of the rotation angle in one cycle of motion. We can see that the proposed method in this paper can control the dynamic error of satellite up to 0.273°, but the dynamic error of ISSO method is relatively high.
In visual localization, after we obtain the point correspondence, the proposed method and ISSO method are used to estimate the fundamental matrix between the indoor images. The rotation angle of the camera calculated by the fundamental matrix at different moments is used to verify the accuracy of visual localization. We calculate the camera rotation angle in 30 s with the angle amplitudes rising from 0° to 43°. Figure 11a,b show the ground-true value and error of the rotation angle, respectively. In Figure 11b, the average rotation error obtained by the ISSO method is 0.517°; however, the error obtained by the proposed method is 0.362°. It can be seen that the error distribution of the proposed method is more concentrated than that of the ISSO method.

6. Conclusions

In this paper, we propose a robust method for estimating the fundamental matrix from point correspondence in multi-view images. Firstly, stable and reliable initial values are obtained by combining the outlier elimination with the fundamental matrix estimation. Even if there is a large number of outliers, the calculated values will soon become stable. Then, the rank constraint is introduced, and we solve the problem of reducing the accuracy of the fundamental matrix in a traditional singular value decomposition correction. The global optimization function under triple-view is constructed, which further improves the estimation accuracy of the fundamental matrix. The experimental results show that the proposed method is more accurate and robust than traditional methods. It can solve the fundamental matrix estimation problem under planar motion scenes. This kind of method is beneficial for the research of high precision non-cooperative target pose measurement and has important significance for the planning research of mobile robots.

Author Contributions

Writing—original draft, L.W.; Writing—review & editing, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61473100.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khattar, F.; Franck, L.; Larroque, B.; Dornaika, F. Visual localization and servoing for drone use in indoor remote laboratory environment. Mach. Vis. Appl. 2021, 32, 32. [Google Scholar] [CrossRef]
  2. Yu, H.Y.; Zhu, H.J.; Huang, F.R. Visual simultaneous localization and mapping (SLAM) based on blurred image detection. J. Intell. Robot. Syst. 2021, 103, 12. [Google Scholar] [CrossRef]
  3. Zia, M.Z.; Stark, M.; Schiele, B.; Schindler, K. Detailed 3d representations for object recognition and modelling. IEEE. Trans. Pattern Anal. Mach. Intell. 2013, 35, 2608–2623. [Google Scholar] [CrossRef] [PubMed]
  4. Pizarro, D.; Bartoli, A. Feature-based deformable surface detection with self-occlusion reasoning. Int. J. Comput. Vis. 2012, 97, 54–70. [Google Scholar] [CrossRef]
  5. Wang, J.H.; Zhou, Z.H. The 3D reconstruction method of a line-structured light vision sensor based on composite depth images. Meas. Sci. Technol. 2021, 32, 075101. [Google Scholar] [CrossRef]
  6. Ingale, K.; Udayan, D.J. Real-time 3D reconstruction techniques applied in dynamic scenes: A systematic literature review. Comput. Sci. Rev. 2021, 39, 100338. [Google Scholar] [CrossRef]
  7. Park, S.M.; Kim, Y.G. A metaverse: Taxonomy, components, applications, and open challenges. IEEE Access 2022, 10, 4209–4251. [Google Scholar] [CrossRef]
  8. Li, W.; Zhu, D.L.; Wang, Q. A single view leaf reconstruction method based on the fusion of ResNet and differentiable render in plant growth digital twin system. Comput. Electron. Agric. 2022, 193, 106712. [Google Scholar] [CrossRef]
  9. Matuzevicius, D.; Serackis, A. Three-dimensional human head reconstruction using smartphone-based close-range video photogrammetry. Appl. Sci. 2022, 12, 229. [Google Scholar] [CrossRef]
  10. Wang, F.Y.; Ma, X.D.; Liu, M.; Wei, B.X. Three-dimensional reconstruction of soybean canopy based on multivision technology for calculation of phenotypic traits. Agronomy 2022, 12, 692. [Google Scholar] [CrossRef]
  11. Ortín, D.; Montiel, J.M.M. Indoor robot motion based on monocular images. Robotica 2001, 19, 331–342. [Google Scholar] [CrossRef]
  12. Li, X.S.; Yuan, X.D. Fundamental matrix computing based on 3D metrical distance. Algorithms 2021, 14, 89. [Google Scholar] [CrossRef]
  13. Schonberger, J.L.; Frahm, J.M. Structure-from-motion revisited. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  14. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  15. He, Z.X.; Shen, C.T.; Wang, Q.Y.; Zhao, X.Y.; Jiang, H.L. Mismatching removal for feature-point matching based on triangular topology probability sampling consensus. Remote Sens. 2022, 14, 706. [Google Scholar] [CrossRef]
  16. Chen, H.Y.; Lin, Y.Y.; Chen, B.Y. Robust feature matching with alternate hough and inverted hough transforms. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  17. Valgaerts, L.; Bruhn, A.; Mainberger, M.; Weickert, J. Dense versus sparse approaches for estimating the fundamental matrix. Int. J. Comput. Vis. 2012, 96, 212–234. [Google Scholar] [CrossRef] [Green Version]
  18. Nistér, D. An efficient solution to the five-point relative pose problem. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 756–770. [Google Scholar] [CrossRef] [PubMed]
  19. Torr, P.H.S.; Murray, D.W. The development and comparison of robust methods for estimating the fundamental matrix. Int. J. Comput. Vis. 1997, 24, 271–300. [Google Scholar] [CrossRef]
  20. Armangué, X.; Salvi, J. Overall view regarding fundamental matrix estimation. Image Vis. Comput. 2003, 21, 205–220. [Google Scholar] [CrossRef]
  21. Zou, Y.X.; Chan, S.C.; Ng, T.S. Least mean M-estimate algorithms for robust adaptive filtering in impulse noise. IEEE Trans. Circuits Syst. II-Analog Digit. Signal Process. 2002, 47, 1564–1569. [Google Scholar]
  22. Zhen, Y.; Liu, X.J.; Wang, M.Z. Precise fundamental matrix estimation based on inlier distribution constraint. In Proceedings of the 2012 International Conference on Information Technology and Software Engineering, Beijing, China, 8–10 December 2012. [Google Scholar]
  23. Jian, B.; Vemuri, B.C. Robust point set registration using gaussian mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1633–1645. [Google Scholar] [CrossRef]
  24. Zhang, Z.Y. Determining the epipolar geometry and its uncertainty: A review. Int. J. Comput. Vis. 1998, 27, 161–195. [Google Scholar] [CrossRef]
  25. Xu, M.; Lu, J. Distributed RANSAC for the robust estimation of three-dimensional reconstruction. IET Comput. Vis. 2012, 6, 324–333. [Google Scholar] [CrossRef]
  26. Andrew, A.M. Multiple view geometry in computer vision. Kybernetes 2001, 30, 1333–1341. [Google Scholar] [CrossRef]
  27. Chum, O.; Matas, J. Optimal randomized RANSAC. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1472–1482. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Ghergherehchi, M.; Kim, S.Y.; Afarideh, H.; Kim, Y.S. Random sample consensus algorithm for enhancing overlapped etched track counting. IET Image Process. 2015, 9, 97–106. [Google Scholar] [CrossRef]
  29. Fu, Q.; Mu, X.X.; Wang, Y. Minimal solution for estimating fundamental matrix under planar motion. Sci. China-Inf. Sci. 2021, 64, 209203. [Google Scholar] [CrossRef]
  30. Xiao, C.B.; Feng, D.Z.; Yuan, M.D. An efficient fundamental matrix estimation method for wide baseline images. Pattern Anal. Appl. 2018, 21, 35–44. [Google Scholar] [CrossRef]
  31. Yan, K.; Zhao, R.J.; Liu, E.H.; Ma, Y.B. A robust fundamental matrix estimation method based on epipolar geometric error criterion. IEEE Access 2019, 7, 147523–147533. [Google Scholar] [CrossRef]
  32. Fraundorfer, F.; Tanskanen, P.; Pollefeys, M. A minimal case solution to the calibrated relative pose problem for the case of two known orientation angles. In Proceedings of the 11th European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010. [Google Scholar]
  33. Chou, C.C.; Wang, C.C. 2-point RANSAC for scene image matching under large viewpoint changes. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
  34. He, W.; Li, Z.J.; Chen, C.L.P. A survey of human-centered intelligent robots: Issues and challenges. IEEE-CAA J. Autom. Sin. 2017, 4, 602–609. [Google Scholar] [CrossRef]
  35. Jiao, Y.M.; Liu, L.L.; Fu, B.; Ding, X.Q.; Wang, M.H.; Wang, Y.; Xiong, R. Robust localization for planar moving robot in changing environment: A perspective on density of correspondence and depth. In Proceedings of the IEEE International Conference on Robotics and Automation, Xi’an, China, 30 May–5 June 2021. [Google Scholar]
  36. Choi, S.; Kim, J.H. Fast and reliable minimal relative pose estimation under planar motion. Image Vis. Comput. 2018, 69, 103–112. [Google Scholar] [CrossRef]
  37. Choi, S.I.; Park, S.Y. A new 2-point absolute pose estimation algorithm under planar motion. Adv. Robot. 2015, 29, 1005–1013. [Google Scholar] [CrossRef]
  38. Wang, C.G.; Li, G.; Wang, X.; Zhang, Y.Y. Research on visual odometer of wheeled robot with motion constraints. J. Phys. Conf. Ser. 2022, 2171, 012064. [Google Scholar]
  39. Dong, S.Y.; Xu, K.; Zhou, Q.; Tagliasacchi, A.; Xin, S.Q.; Niessner, M.; Chen, B.Q. Multi-robot collaborative dense scene reconstruction. ACM Trans. Graph. 2019, 38, 1–16. [Google Scholar] [CrossRef]
  40. Ferraz, L.; Binefa, X.; Moreno-Noguer, F. Very fast solution to the pnp problem with algebraic outlier rejection. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  41. Kukelova, Z.; Bujnak, M.; Pajdla, T. Automatic generator of minimal problem solvers. In Proceedings of the 10th European Conference on Computer Vision, Marseille, France, 12–18 October 2008. [Google Scholar]
  42. Scharstein, D.; Pal, C. Learning conditional random fields for stereo. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar]
Figure 1. Epipolar geometry in binocular vision.
Figure 1. Epipolar geometry in binocular vision.
Sensors 22 04624 g001
Figure 2. Camera motion scheme on the plane.
Figure 2. Camera motion scheme on the plane.
Sensors 22 04624 g002
Figure 3. Epipolar geometry in triple-view.
Figure 3. Epipolar geometry in triple-view.
Sensors 22 04624 g003
Figure 4. Flow chart of fundamental matrix estimation method.
Figure 4. Flow chart of fundamental matrix estimation method.
Sensors 22 04624 g004
Figure 5. Comparisons of the epipolar geometric distances. (a) The mean geometric distance with varying Gaussian noise; (b) The mean geometric distance with varying outliers proportion.
Figure 5. Comparisons of the epipolar geometric distances. (a) The mean geometric distance with varying Gaussian noise; (b) The mean geometric distance with varying outliers proportion.
Sensors 22 04624 g005
Figure 6. Comparisons of the distance between intersection points and feature points. (a) The mean distance with varying Gaussian noise; (b) The mean distance with varying outliers proportion.
Figure 6. Comparisons of the distance between intersection points and feature points. (a) The mean distance with varying Gaussian noise; (b) The mean distance with varying outliers proportion.
Sensors 22 04624 g006
Figure 7. The epipolar geometry in different images. (a) Feature points and epipolar lines in Image1; (b) Feature points and epipolar lines in Image2; (c) Feature points and epipolar lines in Image3; (d) Feature points and epipolar lines in Image4; (e) Feature points and epipolar lines in Image5; (f) Feature points and epipolar lines in Image6.
Figure 7. The epipolar geometry in different images. (a) Feature points and epipolar lines in Image1; (b) Feature points and epipolar lines in Image2; (c) Feature points and epipolar lines in Image3; (d) Feature points and epipolar lines in Image4; (e) Feature points and epipolar lines in Image5; (f) Feature points and epipolar lines in Image6.
Sensors 22 04624 g007
Figure 8. The feature points and intersection points in the intermediate images. (a) Feature points and intersection points in Image1; (b) Feature points and intersection points in Image2; (c) Feature points and intersection points in Image3; (d) Feature points and intersection points in Image4; (e) Feature points and intersection points in Image5; (f) Feature points and intersection points in Image6.
Figure 8. The feature points and intersection points in the intermediate images. (a) Feature points and intersection points in Image1; (b) Feature points and intersection points in Image2; (c) Feature points and intersection points in Image3; (d) Feature points and intersection points in Image4; (e) Feature points and intersection points in Image5; (f) Feature points and intersection points in Image6.
Sensors 22 04624 g008
Figure 9. Satellite 3D feature point model.
Figure 9. Satellite 3D feature point model.
Sensors 22 04624 g009
Figure 10. Dynamic rotation angle error.
Figure 10. Dynamic rotation angle error.
Sensors 22 04624 g010
Figure 11. Ground-true value and error of rotation angle. (a) The ground-true value of the rotation angle; (b) Error of rotation angle.
Figure 11. Ground-true value and error of rotation angle. (a) The ground-true value of the rotation angle; (b) Error of rotation angle.
Sensors 22 04624 g011
Table 1. The mean epipolar geometric distance of synthetic data (pixel).
Table 1. The mean epipolar geometric distance of synthetic data (pixel).
Noise LevelProposed MethodISSO MethodEGEC MethodRANSAC Method
00.0310.1910.0861.645
0.20.0560.3080.3031.896
0.40.0940.4120.5112.105
0.60.1470.5030.7312.333
0.80.2770.6070.9552.556
10.4060.7241.1642.75
1.20.510.8931.3963.001
1.40.6161.0361.6123.206
1.60.7971.2131.8173.411
1.80.9711.4112.0453.65
21.1321.7412.2543.851
Table 2. The mean distance between feature point and intersection point of synthetic data (pixel).
Table 2. The mean distance between feature point and intersection point of synthetic data (pixel).
Noise LevelProposed MethodISSO MethodEGEC MethodRANSAC Method
00.1910.5340.3542.812
0.20.3450.9830.9343.524
0.40.4981.1351.5024.013
0.60.6411.4321.7524.292
0.80.8751.6522.1414.733
11.191.8362.3524.95
1.21.432.0472.5535.163
1.41.5742.2072.8515.514
1.61.7232.4913.2075.87
1.82.132.7133.5246.183
22.2313.1353.8576.488
Table 3. The mean epipolar geometric distance of real images (pixel).
Table 3. The mean epipolar geometric distance of real images (pixel).
DataProposed MethodISSO MethodEGEC Method
Image10.6180.9921.615
Image20.8151.2411.826
Image31.0381.412.346
Image40.5090.9761.587
Image51.1611.8172.455
Image61.2912.0192.532
Table 4. The mean distance between feature point and intersection point of real images (pixel).
Table 4. The mean distance between feature point and intersection point of real images (pixel).
DataProposed MethodISSO MethodEGEC Method
Image11.5132.1622.893
Image21.7752.6853.381
Image32.3092.7873.502
Image41.4022.0542.586
Image52.2713.0173.883
Image62.4063.1374.038
Table 5. Static rotation angle error (°).
Table 5. Static rotation angle error (°).
DataMeasured AngleStandard AngelAbsolute Error
Proposed MethodISSO MethodProposed MethodISSO Method
10.0130.02300.0130.023
22.0341.94220.0340.058
33.9493.92340.0510.077
46.0756.08660.0750.086
58.0917.88780.0910.113
69.89610.138100.1040.138
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wei, L.; Huo, J. A Global Fundamental Matrix Estimation Method of Planar Motion Based on Inlier Updating. Sensors 2022, 22, 4624. https://doi.org/10.3390/s22124624

AMA Style

Wei L, Huo J. A Global Fundamental Matrix Estimation Method of Planar Motion Based on Inlier Updating. Sensors. 2022; 22(12):4624. https://doi.org/10.3390/s22124624

Chicago/Turabian Style

Wei, Liang, and Ju Huo. 2022. "A Global Fundamental Matrix Estimation Method of Planar Motion Based on Inlier Updating" Sensors 22, no. 12: 4624. https://doi.org/10.3390/s22124624

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop