Next Article in Journal
Assessment of Drivers’ Mental Workload by Multimodal Measures during Auditory-Based Dual-Task Driving Scenarios
Previous Article in Journal
Multi-Stage Learning Framework Using Convolutional Neural Network and Decision Tree-Based Classification for Detection of DDoS Pandemic Attacks in SDN-Based SCADA Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multilayer Perceptron-Based Error Compensation for Automatic On-the-Fly Camera Orientation Estimation Using a Single Vanishing Point from Road Lane

1
Electrical and Computer Engineering, Inha University, 100 Inha-ro, Michuhol-gu, Incheon 22212, Republic of Korea
2
Future Vehicle Engineering, Inha University, 100 Inha-ro, Michuhol-gu, Incheon 22212, Republic of Korea
3
Integrated System Engineering, Inha University, 100 Inha-ro, Nam-gu, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(3), 1039; https://doi.org/10.3390/s24031039
Submission received: 5 December 2023 / Revised: 22 January 2024 / Accepted: 2 February 2024 / Published: 5 February 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
This study introduces a multilayer perceptron (MLP) error compensation method for real-time camera orientation estimation, leveraging a single vanishing point and road lane lines within a steady-state framework. The research emphasizes cameras with a roll angle of 0°, predominant in autonomous vehicle contexts. The methodology estimates pitch and yaw angles using a single image and integrates two Kalman filter models with inputs from image points (u, v) and derived angles (pitch, yaw). Performance metrics, including avgE, minE, maxE, ssE, and Stdev, were utilized, testing the system in both simulator and real-vehicle environments. The outcomes indicate that our method notably enhances the accuracy of camera orientation estimations, consistently outpacing competing techniques across varied scenarios. This potency of the method is evident in its adaptability and precision, holding promise for advanced vehicle systems and real-world applications.

1. Introduction

1.1. Background

With the advancements in vision sensor technology, the utilization of vision sensors in advanced driving-assistant systems (ADAS) has become prevalent. The organic incorporation of vision-based sensing technologies such as lane-keeping systems (LKS) and forward collision warnings (FCW) into self-driving scenarios has begun to improve. Most functions are based on vision systems, such as cameras, to obtain the visual context of human perception. However, this pipeline approach of streamlining vision-based inputs to obtain plausible outcomes is common in the ADAS-based industry for accomplishing multiple tasks, such as object detection and segmentation. This is continuously widening the sensing capabilities of self-driving systems [1]. Lane detection is a basic task of an ADAS to identify lane boundaries and extend these to attain LKS, lane departure warning (LDW), and vanishing point detection. Commercial products such as open pilots can be considered successful cases of combining object detection and semantic segmentation into a product with vertical coverage from sensing to control. The simulator also contributes significantly to the regulation and simulation of control systems [2].
Camera calibration is also a sub-problem in computer vision. It mainly analyzes known scenes from a sensor from various perspectives and uses these to calibrate the sensor. For example, in the FCW, a practical calculation of the distance between the vehicle and the object ensures a safe distance for the vehicle. This can be affected by the terrain and weather conditions. Single-camera distance estimation assumes that the object and vehicle are on the same plane and projects a 3D object onto a 2D image using calibrated intrinsic and extrinsic parameters of the camera [3]. Because the intrinsic parameters would have been established at the time of production, these do not vary under normal conditions. Meanwhile, the extrinsic parameters represent the relationship between the target coordinate system and the camera coordinate system. The extrinsic parameters are generally updated according to the motion of the vehicle.
The accurate camera model and the relationship between it and the world are helpful for sensing 3D worlds such as object-based distance estimation [4]. To describe the camera model as intrinsic parameters including the focal length and principal point, and to describe the relationship between the camera model and the world, always use rigid body transformation, also known as extrinsic parameters. In a self-driving vehicle environment, the camera intrinsic parameters are always stable with temperature changes and vehicle shaking, and the camera extrinsic parameters always change when the camera changes. However, conventional camera calibration methods require a calibration target (typically a 2D plane of known size and shape, such as a checkerboard) to determine the camera parameters, thus constituting off-line calibration. As for online calibration, it has a long processing time or needs more images as a reference for better performance [5,6]. Therefore, accurate on-the-fly camera calibration is needed on the vehicle for several applications.

1.2. Purpose of Study

The proposed research was conducted in relevance to the error compensation method for on-the-fly camera orientation estimation such as multilayer perceptron (MLP). The overall study analysis is depicted in Figure 1. Often, traditional online camera orientation methods accumulate errors in the camera orientation estimation and do not consider error compensation factors, as shown in Figure 2 and Figure 3, which was rectified in the proposed method. Accordingly, the contributions of this study are as follows:
  • A working pipeline for the MLP-based adaptive error correction automatic on-the-fly camera orientation estimation algorithm with a single VP from a road lane is proposed, along with relevant quantitative analysis.
  • A stable camera on-the-fly orientation estimation variant is proposed. It uses a Kalman filter that can estimate the angular pitch and yaw concerning the road lane.
  • The residual error of using VP to estimate the camera orientation is compensated for, and several related compensation modules are compared.
The remainder of the paper is organized as follows. The related work is presented in Section 2. This is followed by the proposed method, including the rotation estimation, system of estimation, and evaluation metrics in Section 3. The experimental results are presented in Section 4. The performance of the proposed system and ablation study are discussed in Section 5. Section 6 discusses the limitations of the present study and future work. Section 7 concludes this study.

2. Related Work

Fundamental camera calibration can be categorized into two types: intrinsic (focal length, camera center, and distortions) [7] and extrinsic, which involves mapping rotations and translations from the camera to the world coordinates [8,9] or other sensor coordinates [10,11]. There are two methods involved in on-the-fly the camera orientation estimation for camera extrinsic calibration using target-less methods. One is to determine the area of the object to estimate the yaw and pitch in [12]. The other is to identify the vanishing point (VP) to estimate the orientations. Hold et al. used periodic dashed lanes to estimate the initial extrinsic parameters [13]. Paula et al. developed a model to estimate using the VP [14]. Lee et al. estimated the pitch and yaw using a new method in conjunction with VPs [15]. Jang et al. estimated the three angles using three VPs [16]. Guo et al. also developed indoor applications using VPs [17]. This study focuses on the orientation between the camera and world coordinates. To obtain rotations and translations for online camera calibration, targets such as registered objects [18,19], road marker lanes [20], and objects with apparent appearances [21] are commonly used. However, these methods cannot cover the case regarding varying extrinsic parameters. The authors of [14] proposed another model to estimate the pitch, yaw, and height for the target-less calibration methods. The generated case was used for extrinsic parameter estimation. The authors of [15] proposed an online extrinsic camera calibration method that estimates pitch, yaw, roll angles, and camera height from road surface observations by using a combination of VP estimation and lane width prior. The authors of [17] used a more detailed version to estimate the absolute orientation angle. It included the pitch and yaw. They used a method similar to that of the authors of [15] and analyzed the failure cases. Although they utilized specific applications considering auto-calibration, their approach had significant errors. These methods cannot prevent projection errors even with an appropriate VP. The authors of [16] proposed an online extrinsic calibration approach for estimating camera orientation using motion vectors and line structures in an urban driving environment from three VPs using 3-line RANSAC [22] on a Gaussian sphere. However, the method required more reference lines in the image, and the algorithm was easy to use to obtain failed results. To solve this problem, the concept of the reprojection root-mean-squared error [23] for pixels was used. It is a metric used to express the calibration error. This method is a camera-setup-independent error metric used to measure the performance of the calibration algorithm (our model) while omitting extrinsic influences.
For the calibration system, the authors of [15,16] proposed stabilization systems using an extended Kalman filter (EKF) to solve the nonlinear system. However, Ref. [16] needs more reference lines to obtain several sets of candidate VPs from 3-line RANSAC on a Gaussian sphere. Therefore, a considerably long time is required to complete the process. The VPs also fail because the selected lines intersect at infinity. Study [15] has a residual error based on the distance from VP and the center point(CP) of the image.
This study proposes an MLP-based error compensation method for automatic on-the-fly camera orientation estimation with a single VP from the road lane and steady-state systems. The method considers the commonly available scenes around a self-driving vehicle. It estimates the pitch and yaw with the correction part using an image.

3. Proposed Method

3.1. Camera Orientation Estimation

It is established that parallel lines do not intersect in the world coordinate system. Thus, we can conclude that the VPs are located at an infinite distance from each other. If we consider the forward direction of a vehicle parallel to the lane lines, we can determine that the lane lines intersect at z = . The road reference frame is denoted by the point (X, Y, Z), where Z corresponds to the forward axis of the vehicle. It should be noted that the road reference frame is fixed to the vehicle, as shown in Figure 4. Assumptions regarding the alignment of the vehicle with the lane and lane straightness are crucial. This is because the VP, the intersection point of the lane lines in the image, can provide information regarding camera mounting. Specifically, it reveals the orientation of the camera relative to a vehicle. However, if this assumption is not satisfied, the VP would only provide information regarding the orientation of the vehicle to the lane lines and not that regarding the camera orientation.
The camera projection equation with P i as a point in the image coordinates and P r as a point in the real-world coordinates is as follows:
s P i = K [ R | T ] P r
P i = ( u , v , 1 ) T
P r = ( X , Y , Z , 1 ) T
where K refers to the intrinsic parameter, R to the rotation matrix, and T to the translation matrix. Because this is an equation in homogeneous coordinates, we can multiply both sides by a scalar value s and absorb this value into Z on the left-hand side of the equation:
s u v 1 = K R | T X / Z Y / Z 1 1 / Z
The z-axis value of the VP in the world coordinates is infinite. Thus, (4) becomes
s u v 1 = K R | T 0 0 1 0
The rotation matrix is composed of pitch, yaw, and roll angles. During the setup, the camera is mounted on the windshield, which is parallel to the road trajectory, and under such circumstances, the roll angle defaults to zero. Thereby, the rotation matrix is as follows:
R = r 11 r 21 r 31 r 12 r 22 r 32 r 13 r 23 r 33
As previously assumed for the roll degree, the rotation matrix represents the yaw and pitch as follows:
R = c o s ( α ) s i n ( α ) s i n ( β ) s i n ( α ) c o s ( β ) 0 c o s ( β ) s i n ( β ) s i n ( α ) c o s ( α ) s i n ( β ) c o s ( α ) c o s ( β )
As shown in (5), when (6) is substituted into (5), only the third column of information remains. Therefore, the following formula can be used for the two angles:
α c β c = a t a n 2 ( r 31 , r 33 ) a r c s i n ( r 32 )
where α c is yaw and β c is pitch radian angle.
Accordingly, the reprojection errors can be estimated using the calculated VPs, pitch, and yaw information. Furthermore, the reprojection error ( e p , e y ) distribution of angles with respect to various camera FOVs can be obtained (see Figure 2 and Figure 3). The function that obtains the reprojection error is shown as:
e p e y = i = 1 n ( f ( u , v ) ( α c , β c ) )
where the function f ( u , v ) is defined as follows:
( α ^ , β ^ ) T = f ( u , v ) T
where α ^ , β ^ represent the estimated yaw and pitch angle.

3.2. Error Compensation of Orientation Estimation

As shown in Equation (9), the error residuals are determined by the camera pose angles as depicted in Equation (8), referring to the position of the VP. The pitch angle β c and yaw angle α c are mathematically represented by the sin() and tan() elements as shown in Equation (8); accordingly, the residuals used in the process of estimating the reprojection error as shown in Equation (8) use this pitch and yaw information. During this, the residual pitch and yaw angles are estimated using the distance from the center of the picture to the edge, the residual pitch is approximated by a quadratic function as shown in Figure 2, and the residual yaw is approximated as an even function at the zero point as shown in Figure 3.
MLP [24] is a conventional method for determining nonlinear coefficients. Considering that the complexity of the nonlinear relationship is moderate and that the correlation is quasi-linear, this study used a shallow network with a rectified linear unit (ReLU) activation function (see Figure 5). The MLP model is for the VP to estimate the pitch and yaw errors. The two inputs are the distances from the VP to the CP. These are described as D u and D v in the image coordinates. The two outputs of the model are the pitch e p and yaw e y errors for error compensation and there by Equation (9) can be extended to Equation (13) using image coordinates of the distances from the VP to the CP whuch are D u and D v .
After the MLP model, the error compensation part would function as follows:
H ( D u , D v ) = ( α ^ , β ^ ) T M L P ( D u , D v ) T
( D u , D v ) = v p c p
e p e y = M L P ( D u , D v ) T

3.3. Multilayer Perceptron

In this section, we utilize an MLP model as the core component for the error compensation of the angle estimation from the VP. The MLP architecture comprises an input layer, at least one hidden layer, and an output layer. This neural network design aims to generate a specific flow of information between the layers to learn complex representations of the input data.
Figure 6 and Figure 7 show the performance of each network with different FOV data. The layers of all networks are progressive to the power of 2. For example, the first layer is 2 5 , and the second layer is 2 6 . Considering the loss and inferencing time, the network with 5 layers are best one to build the on-the-fly system. The first hidden layer receives two input features and maps these to a 64-dimensional feature space. The subsequent layers increase the feature space to 128 dimensions, gradually decrease it to 64 and 32 dimensions, and produce a 2-dimensional output. The layers are connected through linear transformations (weights and biases) and activation functions such as ReLU. The number of neurons in each layer enables the network to learn higher-level features and abstractions to fit the error distribution. And the loss almost always converges at the 50th epoch.

4. Experimental Result

4.1. Experimental Environment: Simulation and Real

This section presents the settings and scenarios used in this study. First, to ensure the authenticity of the algorithm verification environment while considering the accuracy and operability of the ground truth (GT) data, the current city simulator was selected (see Table 1). The table outlines the main features and functions of certain highly effective emulators, such as the image resolution, FOV, camera model, camera pose, map generation engine, and minimum hardware requirements. Among these, MORAI has relatively high accuracy. This map of South Korea was selected as the experimental simulator. An experiment was conducted on IONIQ5, which is used for self-driving research.
Table 1 shows the differences. Here, the platforms offer a range of image resolutions. Most of these support a maximum of 1920 × 1920 pixels. The FOV for these platforms also varies. A few provide a more comprehensive range of up to 180 and 200 . The camera models utilized by these platforms can be ideal or physical. Specific platforms support both types. The rotation indicates the camera poses R and translation T values. These demonstrate the precision in the simulation environment.
From the survey of the simulators, the setting of the MORAI simulator was more suitable for this study. Moreover, it supports the city map of South Korea. The experimental setup of the emulator used in this study is shown in Figure 8. The map was that of Sangam-dong in Seoul, Republic of Korea. The vehicle model was the Kia Niro(EV) 2020 version. The camera model was the default one that supports undistorted image information with adjustable FOV and image resolution. The sensor position and orientation were based on the car coordinates.
To evaluate the stability of the pitch and yaw angle estimation and verify the performance of angle prediction, one angle was fixed, and the other was varied. The scenarios of the experiments are designed as follows in Figure 9:
  • Scenario 1: For the algorithm to follow the pitch motion with a 0.1° variation, it modifies the pitch angle continuously while the yaw angle is fixed to 0 .
  • Scenario 2: For the algorithm to follow the yaw motion with a 0.1° variation, the yaw angle is modified continuously while the pitch angle is fixed to 0 .
  • Scenario 3: For the system to converge in certain frames to correct the pitch angle with a 10 variation, the pitch angle is modified by 10 in each frame while the yaw angle is fixed to 0 .
  • Scenario 4: For the system to converge in certain frames to correct the yaw angle with a 10 variation, the yaw angle is modified by 10 in each frame while the pitch angle is fixed to 0 .
  • Scenario 5: For the system to converge in certain frames to correct the angle with a 10 variation, the pitch and yaw angle is modified by 1 in each frame.
To better verify the algorithm in the real world, experiments related to real data were conducted on IONIQ5 vehicles equipped with a GMSL camera (manufactured by SEKONIX, Dongducheon-si, Gyeonggi-do, South Korea). The camera was located above the windshield. There were two cameras with FOVs of 60 and 120 . In this study, the intrinsic parameters and image undistortion of the camera with 60 were preconfigured fully. Therefore, a camera with a FOV of 60 was used to verify this experiment (see Figure 10). This study used linear lane lines that exist at a short distance from the camera (within about 20 m), such as cornerstones or obstacles along the outer edge of the road.

4.2. Results

For each scenario, the table provides the average error (avgE), minimum error (minE), maximum error (maxE), and standard deviation (Stdev) for all the control algorithms. This enabled the assessment of their performance in terms of accuracy and stability.
To ensure the validity and efficiency of the proposed method, in the system design, a KF is added at the end of the VP and angle estimation process to filter out false VP detection or miscalculations. The performance of [14] consistently exhibits larger errors than the method proposed by [15]; therefore, its results are often not visible in the graphs.
Table 2 presents a quantitative analysis of the pitch angle estimation with a fixed yaw angle. As shown in the table, the proposed method outperformed the methods of [14,15] regarding the avgE, minE, maxE, and Stdev. This indicated that the proposed method is more accurate and stable for pitch angle estimation. This is crucial for robotics and computer vision applications. The angle estimation and error curve results are shown in Figure 11.
Table 3 presents a quantitative analysis of the yaw angle estimation with a fixed pitch angle. Compared with the method of Paula et al., the proposed method demonstrated superior performance in the minE, maxE, and Stdev. Although the method of [15] performed marginally better in terms of the avgE, the proposed method maintained a balanced performance across all the metrics. This demonstrated its effectiveness for yaw angle estimation. The results of the angle estimation and error curves are shown in Figure 12.
Table 4 and Table 5 focus on the step scenarios with pitch angle estimation under a fixed yaw angle and with yaw angle estimation under a fixed pitch angle. The proposed method exhibited advantages regarding specific error metrics such as the minE and maxE in both cases. Although the proposed method did not outperform [15] in all the categories, its overall performance remained competitive and stable. This further emphasized the robustness and reliability of the proposed method in various scenarios. The system convergence interval was approximately 30 frames. The angles estimation and error curve result are shown in Figure 13 and Figure 14.
It can be observed from Figure 15 that the proposed method is very close to the lines of the GT and, therefore, has a high accuracy in estimating the pitch angle. The methods in [14,15] seem to have larger errors, especially in some frames, where the errors are pretty significant. In Figure 16, the proposed method again shows characteristics that are very close to the GT, especially in the error plot; the error of the method is relatively small. However, the yaw angle error of the method of [15] is significant in most frames, and although the method of [14] seems to be close to GT in some frames, the error is also large in other frames.
Table 6 presents a quantitative analysis of the orientation using three estimation methods to pitch and yaw for real vehicle datasets. The performances of the methods of [14,15] and that of the proposed method were compared in terms of the avgE, steady-state error (ssE), and Stdev for both pitch and yaw angles.
As observed from the table, the proposed method demonstrates superior performance in terms of the avgE and Stdev for both pitch and yaw angles compared with the other two methods. The proposed method achieved the lowest avgE for pitch and yaw estimation. This indicated its higher accuracy in orientation estimation for real vehicle datasets. Furthermore, the proposed method displayed the lowest Stdev for yaw estimation and a competitive Stdev for pitch estimation. This illustrated the stability and reliability of the method.

5. Ablation Study

This study introduces a compensation module based on the foundational mathematical model proposed by [15]. Furthermore, an ablation study was conducted to evaluate the efficacy of various compensation modules. The following methods were analyzed: the original model proposed by [15] (designated as “No process”), Linear Regression (designated as “Linear”), Plane Function Compensation (designated as “Function”), and MLP.
Figure 17, Figure 18, Figure 19 and Figure 20 elucidate the estimation performance of pitch and yaw angles across cameras with different fields-of-view (FOVs), specifically, 60°, 90°, 120°, and 150°. This exploration covers multiple pivotal aspects:
  • Pitch Analysis: The MLP method consistently outperformed other techniques in estimating pitch angles across all FOV ranges. Relative to GT, the estimations from MLP consistently exhibited remarkable accuracy. Conversely, the “No process” method presented substantial deviations from true values.
  • Yaw Analysis: Several methods achieved commendable precision for yaw estimations across most FOVs. Nonetheless, the “Linear” encountered marginal error increments at specific angles, while the accuracy of the MLP remained relatively invariant.
  • FOV Assessment: Across varied FOVs, the MLP method stood out for its pitch and yaw angle estimations accuracy. While the linear and other methods demonstrated efficacy within certain angular ranges, they exhibited noticeable deviations under particular conditions.
Table 7 conveys a comprehensive quantitative evaluation of the ablation studies across different FOVs. This table contrasts the performance metrics of the four methodologies on a simulated dataset. Key insights from this analysis include:
  • Estimation Error for Pitch: Across all FOVs, MLP consistently registered the most minor error. Notably, at a 150° FOV, its performance superiority was markedly evident. The error of the Linear Regression trajectory steeply ascended with FOV increments. Notably, the error magnitude for the “Function” surged most significantly with FOV enlargement.
  • Estimation Error for Yaw: Astonishingly, both Linear Regression and Function Compensation methods yielded zero error across all FOVs, epitomizing impeccable estimations. Similarly, the performance of MLP mirrored this perfection. In comparison, the “No process” method, while competent, exhibited a marginal error increase as the FOV expanded.
  • Processing Time Assessment: The “Function” consistently achieved the swiftest processing times across all FOVs, signifying optimal resource efficiency. In contrast, the “Linear” generally demanded more prolonged processing intervals. MLP and the “No process” methods displayed commendable consistency in processing durations across all FOVs.
While the MLP methodology showcased exemplary precision in pitch and yaw estimations, its consistency in processing durations substantiated its robustness. “Function”, while unparalleled in processing speed, witnessed occasional challenges in estimation accuracy, especially under expanded FOVs. This analysis underscores the intricate balance between speed, resource allocation, and estimation accuracy, suggesting the potential superiority of MLP in wide-ranging practical applications.

6. Discussions

6.1. Practical Application and Limitations

The practical application of this study is monocular camera object-based distance estimation. The sample image is shown in Figure 21. There are four scenarios for distance estimation: go through the bump, back through the bump, turn left, and turn right. Here, just discuss the scenario result of going through the bump and turning left as an example as shown in Figure 22. Figure 22 shows two cases of distance estimation with the proposed method. When the vehicle goes through the bump, the compensated one works more than the uncompensated one and the errors of the compensated one are less than 5 % . And when the vehicle turns left, the compensated one and uncompensated one are the same: their errors are less than 5 % . The total sample videos, MLP weights, and figures of application are shared in the GitHub link https://github.com/leolixingyou/on-the-fly_camera_calibration_MLP_compensation (accessed on 5 December 2023).
Although the compensation part effectually increases the orientation estimation performance, the additional errors happened from different parts. The findings, as represented in Figure 11, show that when the pitch is 10 there is a spike. The reason is it has false VP detection is because of lane detection. A similar case is shown in Figure 23. When the camera enters or detects a curve, the above-mentioned error phenomenon will often occur, or its predicted value will fluctuate within a considerable range, so its performance has yet to be determined; when the image is without the lane or any landmark then the outline of the road can be detected so that estimation can work normally. The main idea of this study is to obtain one vanishing point from the target plane with the camera heading direction; this is shown in Figure 21.

6.2. Future Work

Considering the accuracy of the proposed method and the content mentioned from Section 6.1, future work will be conducted from the following four aspects:
  • Accurate Lane Detection and Its Process: For more accurate VP prediction, accurate lane instance segmentation or similar instance classification functions are needed. Building an end-to-end deep neural network from VP detection to orientation estimation may be a good research direction.
  • More Accurate Angle Estimation: The performance of the compensated angles will show relative advantages compared to other methods. However, as far as the method itself is concerned, there is still room for improvement, such as the prediction performance of pitch in the real environment, the generalization of the compensation part to the camera FOV ability, etc.
  • FOV Assessment: Across varied FOVs, the MLP method stood out for its pitch and yaw angle estimations accuracy. While the linear and other methods demonstrated efficacy within certain angular ranges, they exhibited noticeable deviations under particular conditions.
  • Curve Case Process: The proposed method fails when cornering, which is a practical shortcoming. Therefore, defining the necessity and application of orientation estimation during curves is also necessary.

7. Conclusions

This study introduces an MLP-based error compensation method for camera orientation estimation using lane lines and a vanishing point. The aim is to identify an algorithm for Automatic On-the-Fly Camera Orientation Estimation and assess its accuracy. An ablation study compared various compensation modules, including Not Processed, Linear Regression, Function, and MLP. The results showed that the MLP compensation improved the original accuracy of the algorithm, especially in estimating pitch and yaw angles. Future research could expand the calibration approach to include the full rotation matrix and camera elevation. With its potential in autonomous vehicles and driver-assistance systems, the method offers promise for real-time camera orientation tasks.

Author Contributions

Conceptualization, X.L., V.K. and H.K. (Hakil Kim); methodology, X.L., V.K. and H.K. (Hyoungrae Kim); validation, X.L., V.K. and H.K. (Hakil Kim); formal analysis, X.L., V.K. and H.K. (Hakil Kim) and H.K. (Hyoungrae Kim); writing—original draft preparation, X.L.; writing—review and editing, V.K., X.L. and H.K. (Hakil Kim); visualization, X.L., H.K. (Hyoungrae Kim) and V.K.; supervision, H.K. (Hakil Kim); and project administration, H.K. (Hakil Kim). All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the BK21 Four Program funded by the Ministry of Education (MOE, Korea) and National Research Foundation of Korea (NRF).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to ongoing validations and continuous improvements.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gupta, A.; Choudhary, A. A Framework for Camera-Based Real-Time Lane and Road Surface Marking Detection and Recognition. IEEE Trans. Intell. Veh. 2018, 3, 476–485. [Google Scholar] [CrossRef]
  2. Chen, L.; Tang, T.; Cai, Z.; Li, Y.; Wu, P.; Li, H.; Qiao, Y. Level 2 autonomous driving on a single device: Diving into the devils of openpilot. arXiv 2022, arXiv:2206.08176. [Google Scholar]
  3. Kim, J.K.; Kim, H.R.; Oh, J.S.; Li, X.Y.; Jang, K.J.; Kim, H. Distance Measurement of Tunnel Facilities for Monocular Camera-based Localization. J. Inst. Control Robot. Syst. 2023, 29, 7–14. [Google Scholar] [CrossRef]
  4. Vajgl, M.; Hurtik, P.; Nejezchleba, T. Dist-YOLO: Fast object detection with distance estimation. Appl. Sci. 2022, 12, 1354. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Zhou, L.; Liu, H.; Shang, Y. A flexible online camera calibration using line segments. J. Sens. 2016, 2016, 2802343. [Google Scholar] [CrossRef]
  6. Hold-Geoffroy, Y.; Sunkavalli, K.; Eisenmann, J.; Fisher, M.; Gambaretto, E.; Hadap, S.; Lalonde, J.F. A perceptual measure for deep single image camera calibration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2354–2363. [Google Scholar]
  7. Ge, D.-Y.; Yao, X.-F.; Xiang, W.-J.; Liu, E.-C.; Wu, Y.-Z. Calibration on Camera’s Intrinsic Parameters Based on Orthogonal Learning Neural Network and Vanishing Points. IEEE Sens. J. 2020, 20, 11856–11863. [Google Scholar] [CrossRef]
  8. Lébraly, P.; Deymier, C.; Ait-Aider, O.; Royer, E.; Dhome, M. Flexible extrinsic calibration of non-overlapping cameras using a planar mirror: Application to vision-based robotics. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 5640–5647. [Google Scholar]
  9. Ly, D.S.; Demonceaux, C.; Vasseur, P.; Pégard, C. Extrinsic calibration of heterogeneous cameras by line images. Mach. Vis. Appl. 2014, 25, 1601–1614. [Google Scholar] [CrossRef]
  10. Jansen, W.; Laurijssen, D.; Daems, W.; Steckel, J. Automatic calibration of a six-degrees-of-freedom pose estimation system. IEEE Sens. J. 2019, 19, 8824–8831. [Google Scholar] [CrossRef]
  11. Domhof, J.; Kooij, J.F.P.; Gavrila, D.M. A Joint Extrinsic Calibration Tool for Radar, Camera and Lidar. IEEE Trans. Intell. Veh. 2021, 6, 571–582. [Google Scholar] [CrossRef]
  12. Lee, J.H.; Lee, D.-W. A Hough-Space-Based Automatic Online Calibration Method for a Side-Rear-View Monitoring System. Sensors 2020, 20, 3407. [Google Scholar] [CrossRef] [PubMed]
  13. Wu, Z.; Fu, W.; Xue, R.; Wang, W. A Novel Line Space Voting Method for Vanishing-Point Detection of General Road Images. Sensors 2016, 16, 948. [Google Scholar] [CrossRef] [PubMed]
  14. de Paula, M.B.; Jung, C.R.; da Silveira, L.G. Automatic on-the-fly extrinsic camera calibration of onboard vehicular cameras. Expert Syst. Appl. 2014, 41, 1997–2007. [Google Scholar] [CrossRef]
  15. Lee, J.K.; Baik, Y.K.; Cho, H.; Yoo, S. Online Extrinsic Camera Calibration for Temporally Consistent IPM Using Lane Boundary Observations with a Lane Width Prior. arXiv 2020, arXiv:2008.03722. [Google Scholar]
  16. Jang, J.; Jo, Y.; Shin, M.; Paik, J. Camera Orientation Estimation Using Motion-Based Vanishing Point Detection for Advanced Driver-Assistance Systems. IEEE Trans. Intell. Transp. Syst. 2021, 22, 6286–6296. [Google Scholar] [CrossRef]
  17. Guo, K.; Ye, H.; Gu, J.; Tian, Y. A Fast and Simple Method for Absolute Orientation Estimation Using a Single Vanishing Point. Appl. Sci. 2022, 12, 8295. [Google Scholar] [CrossRef]
  18. Wang, L.L.; Tsai, W.H. Camera calibration by vanishing lines for 3-D computer vision. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 370–376. [Google Scholar] [CrossRef]
  19. Meng, L.; Li, Y.; Wang, Q.L. A calibration method for mobile omnidirectional vision based on structured light. IEEE Sens. J. 2021, 21, 11451–11460. [Google Scholar] [CrossRef]
  20. Bellino, M.; de Meneses, Y.L.; Kolski, S.; Jacot, J. Calibration of an embedded camera for driver-assistant systems. In Proceedings of the IEEE International Conference on Industrial Informatics, Perth, Australia, 10–12 August 2005; pp. 354–359. [Google Scholar]
  21. Zhuang, S.; Zhao, Z.; Cao, L.; Wang, D.; Fu, C.; Du, K. A Robust and Fast Method to the Perspective-n-Point Problem for Camera Pose Estimation. IEEE Sens. J. 2023, 23, 11892–11906. [Google Scholar] [CrossRef]
  22. Bazin, J.C.; Pollefeys, M. 3-line RANSAC for orthogonal vanishing point detection. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 4282–4287. [Google Scholar]
  23. Tabb, A.; Yousef, K.M.A. Parameterizations for reducing camera reprojection error for robot-world hand-eye calibration. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; pp. 3030–3037. [Google Scholar]
  24. Taud, H.; Mas, J.F. Multilayer perceptron (MLP). In Geomatic Approaches for Modeling Land Change Scenarios; Springer: Cham, Switzerland, 2018; pp. 451–455. [Google Scholar]
  25. Rong, G.; Shin, B.H.; Tabatabaee, H.; Lu, Q.; Lemke, S.; Možeiko, M.; Kim, S. Lgsvl simulator: A high fidelity simulator for autonomous driving. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece, 20–23 September 2020; pp. 1–6. [Google Scholar]
  26. Hong, C.J.; Aparow, V.R. System configuration of Human-in-the-loop Simulation for Level 3 Autonomous Vehicle using IPG CarMaker. In Proceedings of the 2021 IEEE International Conference on Internet of Things and Intelligence Systems (IoTaIS), Bandung, Indonesia, 23–24 November 2021; pp. 215–221. [Google Scholar]
  27. Dosovitskiy, A.; Ros, G.; Codevilla, F.; Lopez, A.; Koltun, V. CARLA: An open urban driving simulator. In Proceedings of the Conference on Robot Learning (CoRL), Mountain View, CA, USA, 13–15 November 2017; pp. 1–16. [Google Scholar]
  28. Samak, T.V.; Samak, C.V.; Xie, M. Autodrive simulator: A simulator for scaled autonomous vehicle research and education. In Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System, Qingdao, China, 20–22 August 2021; pp. 1–5. [Google Scholar]
  29. Rojas, M.; Hermosilla, G.; Yunge, D.; Farias, G. An Easy to Use Deep Reinforcement Learning Library for AI Mobile Robots in Isaac Sim. Appl. Sci. 2022, 24, 8429. [Google Scholar] [CrossRef]
  30. Shah, S.; Dey, D.; Lovett, C.; Kapoor, A. AirSim: High-Fidelity Visual and Physical Simulation for Autonomous Vehicles. Field and Service Robotics: Results of the 11th International Conference; Springer International Publishing: Cham, Switzerland, 2017; pp. 621–635. [Google Scholar]
  31. Park, C.; Kee, S.C. Online local path planning on the campus environment for autonomous driving considering road constraints and multiple obstacles. Appl. Sci. 2021, 11, 3909. [Google Scholar] [CrossRef]
Figure 1. The framework for automatic on-the-fly camera orientation estimation system.
Figure 1. The framework for automatic on-the-fly camera orientation estimation system.
Sensors 24 01039 g001
Figure 2. f ( u , v ) , e p for pitch performance with different camera FOVs.
Figure 2. f ( u , v ) , e p for pitch performance with different camera FOVs.
Sensors 24 01039 g002
Figure 3. f ( u , v ) , e p for yaw performance with different camera FOVs.
Figure 3. f ( u , v ) , e p for yaw performance with different camera FOVs.
Sensors 24 01039 g003
Figure 4. Camera orientation to road plane and coordinates definition for each coordinate system.
Figure 4. Camera orientation to road plane and coordinates definition for each coordinate system.
Sensors 24 01039 g004
Figure 5. The MLP model to estimate the error with the VP: The two inputs are the distance from the CP of the image to the VP, whose description as D u and D v on the image coordinates. The two outputs are the pitch and yaw error for error compensation, whose description is e p and e y on the image coordinates.
Figure 5. The MLP model to estimate the error with the VP: The two inputs are the distance from the CP of the image to the VP, whose description as D u and D v on the image coordinates. The two outputs are the pitch and yaw error for error compensation, whose description is e p and e y on the image coordinates.
Sensors 24 01039 g005
Figure 6. Loss curves: The training data comprise the distance from VP to CP and the related error. The validation data comprise different FOV data. Networks are built by different layers.
Figure 6. Loss curves: The training data comprise the distance from VP to CP and the related error. The validation data comprise different FOV data. Networks are built by different layers.
Sensors 24 01039 g006
Figure 7. Average inferencing time for each network.
Figure 7. Average inferencing time for each network.
Sensors 24 01039 g007
Figure 8. The settings of MORAI environment and camera.
Figure 8. The settings of MORAI environment and camera.
Sensors 24 01039 g008
Figure 9. Scene settings based on transformed angles and changing angles.
Figure 9. Scene settings based on transformed angles and changing angles.
Sensors 24 01039 g009
Figure 10. The settings of real vehicle and camera.
Figure 10. The settings of real vehicle and camera.
Sensors 24 01039 g010
Figure 11. Performance comparison of three algorithms with Scenario 1.
Figure 11. Performance comparison of three algorithms with Scenario 1.
Sensors 24 01039 g011
Figure 12. Performance comparison of three algorithms with Scenario 2.
Figure 12. Performance comparison of three algorithms with Scenario 2.
Sensors 24 01039 g012
Figure 13. Performance comparison of three algorithms with Scenario 3.
Figure 13. Performance comparison of three algorithms with Scenario 3.
Sensors 24 01039 g013
Figure 14. Performance comparison of three algorithms with Scenario 4.
Figure 14. Performance comparison of three algorithms with Scenario 4.
Sensors 24 01039 g014
Figure 15. Pitch performance comparison of three algorithms with Scenario 5.
Figure 15. Pitch performance comparison of three algorithms with Scenario 5.
Sensors 24 01039 g015
Figure 16. Yaw performance comparison of three algorithms with Scenario 5.
Figure 16. Yaw performance comparison of three algorithms with Scenario 5.
Sensors 24 01039 g016
Figure 17. Compensate pitch performance comparison with different camera FOVs.
Figure 17. Compensate pitch performance comparison with different camera FOVs.
Sensors 24 01039 g017
Figure 18. Compensate yaw performance comparison with different camera FOVs.
Figure 18. Compensate yaw performance comparison with different camera FOVs.
Sensors 24 01039 g018
Figure 19. Compensate pitch error comparison with different camera FOVs.
Figure 19. Compensate pitch error comparison with different camera FOVs.
Sensors 24 01039 g019
Figure 20. Compensate yaw error comparison with different camera FOVs.
Figure 20. Compensate yaw error comparison with different camera FOVs.
Sensors 24 01039 g020
Figure 21. Application sample image of the proposed method.
Figure 21. Application sample image of the proposed method.
Sensors 24 01039 g021
Figure 22. Distance estimation performance comparison between off-line calibration and the proposed method.
Figure 22. Distance estimation performance comparison between off-line calibration and the proposed method.
Sensors 24 01039 g022
Figure 23. Limitations in terms of VP detection.
Figure 23. Limitations in terms of VP detection.
Sensors 24 01039 g023
Table 1. Comparative analysis of seven simulators.
Table 1. Comparative analysis of seven simulators.
SimulatorImage ResolutionFOVCamera ModelCamera PoseMap GeneratingCPU/GPU Minimum Requirements
LGSVL [25]0∼1920 × 0∼108030∼90IdeaR (0.1)/T (0.1)-4 GHz Quad core CPU/GTX 1080 8 GB
Carmaker [26]0∼1920 × 0∼192030∼180PhysicalR (0.1)/T (0.1)UnityRAM 4 GB 1 GHz CPU/-
CARLA [27]0∼1920 × 0∼192030∼180PhysicalR (0.1)/T (0.1)Unity/UnrealRAM 8 GB Inter i5/GTX 970
NVIDIA DRIVE Sim [28]0∼1920 × 0∼192030∼200Physical-Unity/UnrealRAM 64 GB/RTX 3090
Isaac Sim [29]0∼1920 × 0∼19200∼90PhysicalR (0.001)/
T (10 × 10−5)
Unity/UnrealRAM 32 GB Inter i7 7th/RTX 2070 8 GB
Air Sim [30]0∼1920 × 0∼192030∼180PhysicalR (0.1)/T (0.1)Unity/UnrealRAM 8 GB Inter i5/GTX 970
MORAI [31]0∼1920 × 0∼192030∼179Idea/PhysicalR (10 × 10−6)/
T (10 × 10−6)
UnityRAM 16 GB I5 9th/RTX 2060 Super
Table 2. Quantitative analysis of pitch estimation with Scenario 1.
Table 2. Quantitative analysis of pitch estimation with Scenario 1.
Paula et al. [14]Lee et al. [15]Proposed Method
avgE2.5510.0200.015
minE3.8650.1300.120
maxE4.9580.2360.220
Stdev2.5251.8171.352
Table 3. Quantitative analysis of yaw estimation with Scenario 2.
Table 3. Quantitative analysis of yaw estimation with Scenario 2.
Paula et al. [14]Lee et al. [15]Proposed Method
avgE−29.109−0.105−0.566
minE0.3070.0170.002
maxE34.22111.8106.211
Stdev5.5763.2851.726
Table 4. Quantitative analysis of pitch estimation with Scenario 3.
Table 4. Quantitative analysis of pitch estimation with Scenario 3.
Paula et al. [14]Lee et al. [15]Proposed Method
avgE−10.3921.4651.731
minE−13.210−0.088−0.057
maxE−3.7663.0734.317
Stdev5.0135.0515.009
Table 5. Quantitative analysis of yaw estimation with Scenario 4.
Table 5. Quantitative analysis of yaw estimation with Scenario 4.
Paula et al. [14]Lee et al. [15]Proposed Method
avgE−20.3344.5234.804
minE0.3070.6730.112
maxE38.27020.76314.877
Stdev13.3409.0496.347
Table 6. Quantitative analysis of orientation using three different pitch and yaw estimations for real vehicle datasets.
Table 6. Quantitative analysis of orientation using three different pitch and yaw estimations for real vehicle datasets.
MetricPaula et al. [14]Lee et al. [15]Proposed Method
PitchYawPitchYawPitchYaw
avgE9.241−22.823−2.573−0.297−2.567−0.256
ssE−14.26624.854−0.836−1.400−0.847−1.395
Stdev2.6802.6300.1650.1990.1720.156
Table 7. Computational cost analysis of each algorithm.
Table 7. Computational cost analysis of each algorithm.
FOVAccuracyMethod
No ProcessLinearFunctionMLP
60Pitch/degree1.5721.9447.6060.881
yaw/degree0.111000
Processing time/ms0.0250.0980.0030.063
FLOPs708012037,318
90Pitch/degree1.5723.34413.1741.155
yaw/degree0.142000
Processing time/ms0.0250.0970.0030.063
FLOPs708012037,318
120Pitch/degree1.5725.79322.8171.336
yaw/degree0.164000
Processing time/ms0.0240.0940.0030.061
FLOPs708012037,318
150Pitch/degree1.57212.48249.1641.473
yaw/degree0.181000
Processing time/ms0.0240.0960.0030.062
FLOPs708012037,318
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Kim, H.; Kakani, V.; Kim, H. Multilayer Perceptron-Based Error Compensation for Automatic On-the-Fly Camera Orientation Estimation Using a Single Vanishing Point from Road Lane. Sensors 2024, 24, 1039. https://doi.org/10.3390/s24031039

AMA Style

Li X, Kim H, Kakani V, Kim H. Multilayer Perceptron-Based Error Compensation for Automatic On-the-Fly Camera Orientation Estimation Using a Single Vanishing Point from Road Lane. Sensors. 2024; 24(3):1039. https://doi.org/10.3390/s24031039

Chicago/Turabian Style

Li, Xingyou, Hyoungrae Kim, Vijay Kakani, and Hakil Kim. 2024. "Multilayer Perceptron-Based Error Compensation for Automatic On-the-Fly Camera Orientation Estimation Using a Single Vanishing Point from Road Lane" Sensors 24, no. 3: 1039. https://doi.org/10.3390/s24031039

APA Style

Li, X., Kim, H., Kakani, V., & Kim, H. (2024). Multilayer Perceptron-Based Error Compensation for Automatic On-the-Fly Camera Orientation Estimation Using a Single Vanishing Point from Road Lane. Sensors, 24(3), 1039. https://doi.org/10.3390/s24031039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop