Next Article in Journal
High-Precision Measurement of Sea Surface Temperature with Integrated Infrared Thermometer
Previous Article in Journal
The Way to Modern Shutter Speed Measurement Methods: A Historical Overview
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Landing System Development Based on Inverse Homography Range Camera Fusion (IHRCF)

by
Mohammad Sefidgar
* and
Rene Landry, Jr.
LASSENA Laboratory, École de Technologies Supérieure (ÉTS), Montreal, QC H3C 1K3, Canada
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 1870; https://doi.org/10.3390/s22051870
Submission received: 10 January 2022 / Revised: 11 February 2022 / Accepted: 25 February 2022 / Published: 27 February 2022
(This article belongs to the Section Navigation and Positioning)

Abstract

:
The Unmanned Aerial Vehicle (UAV) is one of the most remarkable inventions of the last 100 years. Much research has been invested in the development of this flying robot. The landing system is one of the more challenging aspects of this system’s development. Artificial Intelligence (AI) has become the preferred technique for landing system development, including reinforcement learning. However, current research is more focused is on system development based on image processing and advanced geometry. A novel calibration based on our previous research had been used to ameliorate the accuracy of the AprilTag pose estimation. With the help of advanced geometry from camera and range sensor data, a process known as Inverse Homography Range Camera Fusion (IHRCF), a pose estimation that outperforms our previous work, is now possible. The range sensor used here is a Time of Flight (ToF) sensor, but the algorithm can be used with any range sensor. First, images are captured by the image acquisition device, a monocular camera. Next, the corners of the landing landmark are detected through AprilTag detection algorithms (ATDA). The pixel correspondence between the image and the range sensor is then calculated via the calibration data. In the succeeding phase, the planar homography between the real-world locations of sensor data and their obtained pixel coordinates is calculated. In the next phase, the pixel coordinates of the AprilTag-detected four corners are transformed by inverse planar homography from pixel coordinates to world coordinates in the camera frame. Finally, knowing the world frame corner points of the AprilTag, rigid body transformation can be used to create the pose data. A CoppeliaSim simulation environment was used to evaluate the IHRCF algorithm, and the test was implemented in real-time Software-in-the-Loop (SIL). The IHRCF algorithm outperformed the AprilTag-only detection approach significantly in both translational and rotational terms. To conclude, the conventional landmark detection algorithm can be ameliorated by incorporating sensor fusion for cameras with lower radial distortion.

1. Introduction

Landing system design and UAV descending landmark detection have been the focus of many studies. The prime aim of these efforts is to develop highly accurate and computationally lightweight algorithms to meet the needs of businesses and emergency UAVs that serve in marine environments.

1.1. Problem Statement

The main drawback of landing system development research thus far is that it has been centered around less complex landing platforms such as static or minimally fluctuating ship decks. Hence, there is a need to investigate landing surfaces with higher levels of motion complexity, such as a Stewart table, for pose estimation. The work presented here is a new range and camera sensor fusion technique applicable to complex landing tasks that addresses the pose estimation problem of a descending surface by utilizing Inverse Homography Range Camera Fusion (IHRCF). IHRCF operates by first calculating the homography between pixels and world coordinates. Then, by inverse homography, Apriltag pixels are transformed to world coordinates. Next, knowing the real-world coordinates of the AprilTag corners, the rigid body transformation between the camera frame and real-world landmark coordinates is calculated using range sensor data. Finally, depending on the sensor installation and the type of acquisition, which is here dependent on the CoppeliaSim convention of image data acquisition, a transformation step is required. This proposed IHRCF technique was evaluated, and its performance indicates it is a viable algorithm for UAV pose estimation.

1.2. Literature Review

The methods found in the literature can be categorized into two main sections: conventional and modern approaches. While modern methods use deep learning and machine learning for positioning, conventional methods use algorithms such as Kalman Filter (KF), Extended Kalman Filter (EKF), Unscented Kalman Filter, and Particle Filter (PF). There are some drawbacks to address the issue of pose estimation using these methods. Regarding modern methods, they utilize black box for modeling the complex system. This significantly increases the computational expenses of the designed system, leading to including a Graphical Processing Units (GPUs) which can handle heavy calculations [1]. In addition, training a deep model requires the availability of large datasets for training which is sometimes impractical or imprecise [2]. Power and memory demands are escalating when deep learning model becomes deeper, resulting in a weighty system design and lowering flight time, in the case of aerial autonomous design [1]. The training process also for the Ultra Deep Neural Network (UDNN) is resource intensive and time consuming [3]. Furthermore, concerning conventional methods, KF is very well known to handle the linear system with high accuracy. However, it cannot be used for the nonlinear system. Hence EKF is used for systems with moderate nonlinearity [4]. EKF can, to some extent, handle system nonlinearity up to the point that the noise statistic such as covariance matrixes of the process noise is accurate, otherwise the estimation can either be inaccurate or diverge [5]. Therefore, it can be inferred that EKF error estimates tend to undervalue state uncertainties [6]. UKF, to a great extent, compensates the deficiencies of the EKF in system nonlinearity, but it still suffers from the fact that rounding error and approximation error become larger and larger when nonlinearity of the system increases [7]. PF was developed to handle even more complex system with very high nonlinearity. The accuracy of PF largely depends on the number of particles, this can boost the computations, leading to the expensive and heavy system design [8]. We developed a sensor fusion technique that can improve AprilTag detection algorithms (ATDA) pose estimation in terms of rotational and translational estimation. The designed system is very lightweight in compared to heavy LIDAR system and generates only four single line range data, which significantly lower the computations.
Sensors used for pose estimation and location include monocular and stereo cameras, Inertial Measurement Units (IMU), Global Positioning System (GPS), and Impulse Radio Ultra-Wideband (IR-UWB). A simple landing system consists of a rope and a tether for landing platform pose estimation. Most conventional methods apply an estimation theory to compensate for signal noise. Two works focused on the vision and inertial measurements-based algorithm almost at the raw-sensor level to conduct EKF sensor fusion leading to a robust and rapid pose estimation [9,10]. A subsequent study employed a camera and an IMU sensor fusion algorithm using EKFs that made use of the more robust features from the camera data for object detection: the speeded-up robust feature (SURF) and random sample consensus (RANSAC). The algorithm proved to be computationally rapid, robust and reliable, and thus can be considered for practical applications [11]. A more nonlinear estimation would be to use an UKF that can better handle the nonlinearity of the system models. Researchers used a variation of the UKF called the Square-root Unscented Kalman Filter (SR_UKF) combined with the concept of homography between the geoinformation of the landing platform and data from vision and IMU fusion sensors. They found that their proposed method could be used for the accurate landing of commercial aircraft in GPS-denied environments and in low visibility [12]. Another work used a small-sized drone to conduct pose estimation, with the camera installed on the ship deck to reduce the size of the drone for test purposes. An even more nonlinear filter, cited as a particle filter, was used to extract the bounding box feature from each frame of the camera. These features are important for sampling, weighting and resampling. A UKF and an Unscented Bingham Filter (UBiF) were used for the translational motion filtering and the rotational motion filtering, respectively. This study showed that the developed system performance satisfied the automatic landing system requirements [13]. What is more, the stereo camera found its place in drone-based localization and navigation of UAVs, as it offers 3D data. Kim and Lee (2003) proposed a system based on the stereo camera, using a Kalman filter to reduce drift from IMU data and making use of Differential GPS (DGPS). They concluded that the technique could provide accurate attitude and position data for GPS-denied environments [14]. Another study focused on stereo vision and inertial navigation data fusion using basic Kalman filtering and featured matching between stereo frames for the aims of takeoff, hovering and landing. The main advantage of this work was that it provided an algorithm independent of the GPS or landmark coordinates for positioning [15]. The fisheye camera is another camera variant, one that furnishes a wider angle of view compared to monocular cameras. The fusion of the features data from the camera and stereo vision provides a system with sufficient accuracy and robustness to successfully predict landings on a moving Unmanned Ground Vehicle (UGV) [16]. Besides, a work by Wang and She (2017) proposed a multi-landmark method for moving rover landing research by utilizing a gimbaled camera instead of a fisheye camera. This approach could prompt the camera’s Field of View (FOV) [17]. Finally, a sensor fusion technique based on image and IMU data has met the challenging task of guiding beyond visual line of sight (BVLOS) by First Person View (FPV) automatic navigational equipment. Estimated poses were amended by adding a term cited as a correction function calculation to the estimated pose [18].
While the above-mentioned conventional techniques applied filters such as the EKF and the UKF, modern research approaches utilize Artificial Intelligence (AI) [19,20,21,22,23]. Powerful Graphical Processing Units now allow for rapid calculations [24,25,26]. Most basic intelligent landing systems use neural networks. At least one uses a neural network trained by fuzzy logic for descending landmark detection [27]. Another study focuses on developing a safe landing detection algorithm from 3D lidar point clouds, aiming to outperform the conventional heuristic methods [28]. Reinforcement learning is a technique that employs a base agent which interacts with the environment through actions. The agent is awarded or penalized for every action it takes. Therefore, the agent is trained until it finds a way to maximize its reward while minimizing its penalty through its experience obtained through its actions [29,30,31]. Another approach investigates deep reinforcement learning based on Deep Q-Networks (DQNs) as a control method to navigate towards the landing landmark [32]. The following sub-sections discuss the settings, installations, and methodologies utilized in this work.

2. Proposing Inverse Homography Range Camera Fusion (IHRCF) Methodology

Figure 1 presents an overview of the developed IHRCF system, including its evaluation process. The landing position of the drone is vertical, and the drone processes data from AprilTag via the proposed algorithm. The IHRCF only requires four-point pixel coordinates and their distances in the real world. Therefore, the algorithm does not require a camera calibration phase. The overview of the technique requires A ToF and Camera calibration technique in advance of the IHRCF to relate distances from ranges sensors to the pixel in the image. The technique starts by collecting images from the camera and detection of four corners of the AprilTag. Knowing world dimensions of the sensors and the touch points in the image, the planar homography is calculated. Since the homography and pixel coordinate of the AprilTag corners are known, X and Y directions of the landmark can be found using inverse homography in the next step. Then these directions data are used to find altitude of the corner points by the planar surface equation found by 4 ranges sensors. Subsequently, rigid body transformation between found world coordinate of the corners and known dimension of the AprilTag is calculated. In the final stage, the simulator coordinate system is converted to the Cartesian coordinate translations and Euler angles.
Figure 2 represents the pixel coordinate of the camera, while Figure 3 and Figure 4 illustrate the Cartesian coordinates of the drone and landing surfaces.

2.1. Camera Calibration

Camera calibration is used here in the AprilTag detection algorithm (ATDA) to evaluate the performances of the proposed IHRCF approach. The camera calibration process is the attempt to find the camera’s mathematical model. Generally, a camera has the following relations:
s [ u v 1 ] = [ f x 0 c x 0 f y c y 0 0 1 ] [ r 11 r 12 r 13 t x r 21 r 22 r 23 t y r 31 r 32 r 33 t z ] [ X w Y w Z w 1 ]
where fx, fy, cx, and cy are the camera focal lengths and camera principal points, respectively, in the x and y directions. The matrix that contains the focal lengths and principal points is called the intrinsic parameter of the camera. Moreover, rij and tx, ty, and tz are the camera rotations and translations. The matrix that includes the rotation elements and the translation is called the extrinsic parameter of the camera. Equation (1) can be used to calculate the pixel coordinates of any point in the world if the X, Y, and Z coordinates of that point in the world are known.
The standard camera calibration method follows the steps listed below:
  • Obtain the chessboard images with different rotations and translations in the camera frame.
  • Calculate the grayscale image from the acquired chessboard images.
  • Apply the corner detection algorithm; and
  • Use the real-world camera position to calculate the homography, using the chessboard’s size, which has squares of 2.5 mm and 7 × 9 corners [33,34,35,36].

2.2. Camera Range Sensor Calibration

In the simulation, the drone is commanded to increase the altitude, stop and take a picture in 120 stages every 3 s, allowing it to stabilize after each altitude increase, up to the maximum sensor range, refer to Figure 4.
The motivation for camera range sensor calibration is to find the coordinates of the pixels at different altitudes. Therefore, four circular calibration pads, each of a different color: blue, green, yellow, and red, are chosen and centralized with each sensor. In this attempt the blue, green, yellow, and red pads are centralized with sensors ps1, ps2, ps3, and ps4 touch points, respectively. Figure 5 illustrates the relative location of range sensors and their allocated color pads only for one frame.
In the next phase, images are collected via the downward camera and each color pad is segmented by the conversion of images from RGB to YCBCR color format using Equation (2) [37]:
[ 601 Y 219 C B C R ] = [ 16 128 128 ] + [ 65.481 128.553 24.966 37.797 74.203 112 112 93.786 18.214 ] [ R G B ] T
The next step is to filter each Cy, CB, and CR channel. Table 1 shows these color thresholding data; Please refer to Figure 6a,b for the result of color based segmentation.
Each image convex hull is then extracted from each frame, and the best convex hull is cited as the smallest convex set that contains all the region’s pixels in the Euclidean plane [38,39,40].
X k i = ( X k 1 B i ) A i = 1 , 2 , 3 , 4   and   k = 1 , 2 , 3 ,
X 0 i = A   and   D l = X con   i
X k i = X k 1 i
C ( A ) = i = 1 4 D i
where Bi and Xk are the structuring elements of the binary image and the set of convex sets, respectively. In Equation (4), the subscript “con” denotes the convergence that satisfies Equation (5). Finally, considering Equation (6), the union of all the subsets of the convex set is the binary convex of all the image blobs.
Next, the centroid of each color pad is extracted by the following relationship:
{ ϑ x = 1 N i N x i ϑ y = 1 N i N y i
where N is the number of the binary image points [41].
Finally, all the collected centroid points and range data, containing close to 120 points, are stored, and the lookup table is made from this data. To project the data that are not included in the table, data is interpolated through the Piecewise Cubic Hermite Interpolating Polynomial (PCHIP). PCHIP is a very rapid and efficient algorithm and takes less than 0.03 s to calculate data interpolation.
To obtain f k at the x k point via the PCHIP algorithm:
Let us consider h k = x k + 1 x k ,   and   d k = ( y k + 1 y k ) / h k as the slope between two interpolating points x k + 1 , x k . The slope of the point between them has the Equation (9) as indicated.
f k = { ( w 1 + w 2 ) ( w 1 d k 1 + w 2 d k ) 0 d k , d k 1 0   &   S i g n ( d k ) S i g n ( d k 1 ) d k , d k 1 = 0   &   S i g n ( d k ) = S i g n ( d k 1 )
s i g n ( x ) = { 1 0 1 x > 0 x = 0 x < 0
h k = x k + 1 x k ,
d k = ( y k + 1 y k ) / h k
where Sign is the signal function [42]. In Equation (11), d k is the slope between two points, x k + 1 , and x k . Finally, a one-sided scheme is used to calculate the end slope of the interpolating shape [43].

2.3. Image Range Acquisition

In this experiment, the image acquisition device is a monocular camera attached to the drone and which has the same coordinate frame as the drone, but in the downward direction. The resolution of the camera is 512 × 512 pixels. In addition, as depicted in Figure 7, there are four range sensors located in the same plane as the camera. There is a translational shift for these four range sensors. The coordinate system for the sensors is:
Ps 1 = [        0             0             0 ] ; Ps 2 = [ a 1 × cos ( θ 1 ) + a 2 × cos ( θ 2 )          0             0 ] ; Ps 3 = [ a 2 × cos ( θ 2 ) + a 3 × cos ( θ 3 )      a 2 × sin ( θ 2 ) + a 3 × sin ( θ 3 )         0 ] ; Ps 4 = [        0            a 1 × cos ( θ 1 ) + a 4 × cos ( θ 4 )        0 ] ;
where Ps1, Ps2, Ps3, and Ps4, are the sensors’ locations in the camera frame, with the origin of the first sensor. Also, ai are the distances of the sensors from the camera and θ i are assumed to be nearly 45°.
Figure 7 illustrates these angles and distances for the designed in Equation (12). As can be seen, each sensor is located at the same distance, presented in dark blue line with the magnitude of 0.1 m, from the camera and at the same angle magnitude, depicted in white curve with 45°, with respect to the camera’s axes.

2.4. Calculate the Homography between the Pixels and the World Coordinates

This step calculates the relation between an image pixel and the corresponding world coordinate points. The projective transformation has eight degrees of freedom [44]. Figure 8 shows the range sensor touch points, which are calculated through camera range sensor calibration (refer to Section 2.2), in dark blue. These are transformed by projective transformation, as illustrated in Figure 9 in image plane.
Figure 10 demonstrates the relationship of a 2D image plane, presented in red and in left, and its projected 2D world plane, depicted in light green and in right.
The homogeneous coordinate system is used for this mapping with the following relation [45,46,47,48]:
u = H u
[ x 1 y 1 1 ] = H [ x 2 y 2 1 ] = [ h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 ] [ x 2 y 2 1 ]
in which it is assumed that h 22 is taken as one, hence there are 8 variables, the 8 Degrees of Freedom (DoF), in the homography matrix, denoted by H in Equation (14) [49].
x 1 = h 11 x 2 + h 12 y 2 + h 13 h 31 x 2 + h 32 y 2 + h 33
y 1 = h 21 x 2 + h 22 y 2 + h 23 h 31 x 2 + h 32 y 2 + h 33
In Equations (15) and (16), x 1 , y 1 , x 2 , and y 2 are corresponding points that undergo projective transformation. To solve Equations (15) and (16) at least four points are required. The order of the points is vitally important to achieve a meticulous transformation. The world coordinates of the sensor are as stated in Equation (12) with the order as given below in Equation (17):
PSWorld xy = [ ps 1 xy ps 2 xy ps 4 xy ps 3 xy ]
ProjectedPoints v u = [ P x l ( B L U E ) v u P x l ( GREEN ) v u P x l ( RED ) v u P x l ( YELLOW ) v u ]
where PSWorldxy are the 2D world points with X-Y coordinates only, and ProjectedPoints v u are the projected points in the pixel plane that is flipped in the u and v directions.

2.5. Mapping from Apriltag Pixels to World Coordinates by Inverse Homography

Now that we have the relation between the world plane and its corresponding image, by using inverse planar homography, all points on the 2D image plane can be mapped to 2D world points, presented in worlds’ X and Y coordinates. This can be conducted easily through inverse homography of the projective transform found in the previous subsection and by the following Equation (19) [50,51,52]
u l = H 1 u
where u l represents the 2D coordinates of the landmark, here AprilTag, with four known points.

2.6. Calculation of Points’ Altitude in the Camera Frame

Since there are four range sensors mounted on the quadrotor’s arm, it is easy to find the landing platform planar Equation (20) [53,54,55]:
n ^ = ( p s t 1 p s t 2 ) × ( p s t 2 p s t 3 )
p s t = [ p s x y d ]
where p s t represents the sensors’ touch coordinates in the camera frame. In addition, n ^ and p s x y denote the planar normal vector of the landing surface and the sensor coordinates in the X and Y directions, respectively.
Having the planar Equation (20) and the landmark’s 2D coordinates, u l in the Equation (19), it is convenient to find the Z coordinates of the points on the landing surface by means of Equation (22).
u l Z = ( d n ^ 1 u l X n ^ 2 u l Y ) n ^ 3

2.7. Estimate Rigid Body Transformation

At this stage, all the point coordinates of the landmark are known. Hence, the landmark is considered as a rigid body in the real world, and the transformation between the real-world coordinates and the calculated points is in the frame of the camera. The following Equations (23) and (24) are used to find the coordinates of the transformed points [56].
x = R x + t = R X R Y R Z x + t
t = [ t X t Y t Z ]
In Equation (23), R is the rotation matrix and is expressed in Euler-angle order (pitch, roll and yaw) in “ZYX” order, with the angles expressed as in Equation (25). Moreover, variable t is the translational transformation in the X, Y, and Z directions, as stated in Equation (24). Euler angle vector, which are ϕ (roll), θ (pitch), and ψ (yaw), of the landing surface is indicated in Equation 25.
A n g l e ( R ) = [ ϕ θ ψ ]

2.8. Transformation of the Coordinates

Considering the pixel coordinate flip in the homography calculation due to the order convention used in the CoppeliaSim simulator, a transformation would be indispensable to calculate the final corrected transformation.
R c o r = R x ( ϕ ) · R y ( ψ ) · R z ( θ )
Using Equation (26), the corrected rotational components are calculated in Tait–Bryan angles with the order of ‘xyz’, with the result expressed in Euler-angle order (ZYX). Additionally, the translation is negated in the X-direction.
t c o r = [ t X t Y t Z ]
The results of this stage are the 3D coordinates of the landmark in the real world. Figure 11 depicts the correction of the frame along with its calculated points, and calculated touch points through calibration process in the image coordinate. The cyan and red dots are the landmark corners and sensors’ touch points, respectively, on the surface, which are in a lower altitude than that of the sensor-camera platform. The drone-mounted sensors are depicted at zero altitude. Finally, yellow dots indicate the stable landmark coordinates in the camera frame.

3. Experimental Design

The settings of the experiment and software implementations are elaborated in the following sections.

3.1. Test Platform

Figure 12 illustrates the overview of the test bench. The platform has dynamic translation and rotation. In addition, the AprilTag frame and the landing platform frame are assumed to be the same attitude. It is assumed that there is a motionless flying quadrotor and variable attitude landing platform.

3.2. Software Implementation

The simulator used in this experiment is called CoppeliaSim which has the potential to be controlled through the remote Application Programming Interface (API) using CoppeliaSim API framework and to be sent data in real time to a variety of software including MATLAB, Python, etc. From Figure 13, MATLAB software has the control of the drone dynamics, while sensors and camera data are sent back to MATLAB for the calculations. Ground truth data, including Euler angles and translational movements data of the landing platform are sent back to MATLAB for the evaluation of the proposing technique and ATDA for a comparative evaluation. The simulation engine employed were Bullet physic engine version 2.78 with the simulation time step of 50 ms and balanced speed. d1, d2, d3, and d4 are ranges measured through four range sensors. ϕ, θ, ψ, tX, tY, and tZ are roll, pitch, yaw, translation in X, translation in Y, and translation in Z, for landing platform, respectively; The subscript IHRCF, ATDA, and GT indicate the type of data in Figure 13.

4. Results and Discussion

The IHRCF pose estimation conducted in this research is examined in its translational and rotational components. Figure 14 shows the angular comparison of the IHRCF, Ground Truth, and ATDA approaches in the simulation environment, presented in blue, green, and red, respectively. The results confirm a robust tracking of the ground truth by the IHCRF, and by the ATDA for roll and pitch. From Figure 15, IHRCF proved a significant improvement in the altitude estimation, while it slightly ameliorated the estimation in X and Y directions. ATDA shows major inaccuracy for Z direction distance calculation. Figure 16 illustrates the translational results of the IHRCF (in blue) compared to those of the ATDA, in red, and the Ground truth (GT), in green. The IHRCF offers a better tracking of the GT than the AprilTag algorithm. At 40 and 120 s ATDA technique proved slight inaccuracy, while IHRCF rectified the error using range sensor data.
Figure 16 depicts the IHRCF trajectory plots (in blue) of the descending XZ, YZ, XY, and XYZ surfaces in (a–d), respectively, compared to those of the AprilTag/camera pose detection technique (in red) and ground truth (in green). The pose estimation errors are portrayed next.
Figure 17 shows the IHRCF transitional errors, revealing that they do not exceed 0.04 m. Disregarding initial errors, the proposing technique pose estimation error stays below 0.01 m, with the maximum error in the Z direction at around 110 s.
Figure 17 shows the IHRCF Euler angle estimations errors along the X, Y and Z axes, while Figure 18 demonstrates the IHRCF translational errors. Regarding Figure 17, excluding initial inaccuracy of the IHRCF, maximum error belongs to pitch calculation at 120 s and in the remaining period the errors are beneath 2 degrees. Considering Figure 18, the maximum error is, for Z direction, in 110 s with 0.009 m and in the rest of the periods the errors stayed less than 0.0085 in all axes.
As it can be inferred from Figure 19 the AprilTag detection algorithm has a significant inaccuracy in the Z-direction calculation, while the rest of the directions’ error is below 0.02 m. Figure 20 illustrates the striking angular estimation error at 120 s with 5 degrees and at 4 s with approximately 4.3 degrees in pitch. In addition, roll inaccuracy happens at 83 s with 3 degrees. However, in the rest of the simulation error remained less than 3 degrees.
As shown in Figure 19 and Figure 20, the ATDA demonstrates high angular and translational estimation errors. Equation (28) shows Mean Absolute Error (MAE) calculation over the whole simulation process.
MAE = t = t 1 t n | y t y ˜ t | n = t = t 1 t n | e t | n
yt, y ˜ t , and et are ground truth, estimated pose, and error at time t respectively. The MAEs of the proposed IHRCF system are summarized and compared to those of the ATDA in Table 2.
Table 2 shows the improved error values in both angular terms and translational estimations realized by the IHRCF in relation to those of the ATDA. The estimation errors for the translational results’ peak are at close to 0.073 m and 0.0085 m for the ATDA, in the Z direction, and for the IHRCF, in the X direction, respectively. In addition, the ATDA had maximum angular errors of 5 degrees for the pitch angle, while the angular error peak for IHRCF was 2 degrees for the yaw angle.

5. Conclusions and Future Work

The prime aim of the current research was to develop a process that can support emergency landing systems in highly complex contexts, such as drone marine deck landing. The proposed IHCFR algorithm addresses the problem of pose estimation. This research showed that single camera-based pose estimation is not adequate to calculate the attitude of a descending platform, as the camera model is not obtained correctly via the camera calibration process. A precise intrinsic camera parameter is vitally important to find the real-world coordinates of the landmark. The IHCFR algorithm performed pose estimation more accurately than the ATDA, but it requires four accurate range sensors. The algorithm employs range sensors and proposing camera-ToF calibration data to calculate planar homography. Then the inverse of homography, normal vector found by ranges data, and pixel coordinate of the corners are used to calculate the world coordinate of the corners for the AprilTag. A rigid body estimation method is used to find rotational and translational behavior of the landing platform. Finally, a transformation correction of the coordinate from simulator allocated coordinates is calculated. Overall, the IHRCF algorithm outperformed the ATDA in these experiments in rotational and translational pose estimation. An interesting aspect of this proposed algorithm is that it is independent of the camera calibration process and thus does not have the deficiencies of inaccurate camera calibration. As a future work, it is recommended that Inertial Measurement Units (IMUs) are included in the landing system design to address its the weaknesses, such as sensor interaction noise among the ToF range sensors.

Author Contributions

Conceptualization, M.S.; methodology, M.S.; writing—original draft preparation, M.S.; writing—review and editing, M.S.; supervision, R.L.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to acknowledge the Lassena lab and the École de Technologie Supérieure for making resources available for conducting this research. In addition, I would like to acknowledge the contribution of Farnoush Falahatraftar for his first revision of this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aspri, M.; Tsagkatakis, G.; Tsakalides, P. Distributed Training and Inference of Deep Learning Models for Multi-Modal Land Cover Classification. Remote Sens. 2020, 12, 2670. [Google Scholar] [CrossRef]
  2. Sehgal, A.; Kehtarnavaz, N. Guidelines and Benchmarks for Deployment of Deep Learning Models on Smartphones as Real-Time Apps. Mach. Learn. Knowl. Extr. 2019, 1, 27. [Google Scholar] [CrossRef] [Green Version]
  3. Ang, L.; Seng, K. GPU-Based Embedded Intelligence Architectures and Applications. Electronics 2021, 10, 952. [Google Scholar] [CrossRef]
  4. Ristic, B.; Arulampalam, S.; Gordon, N. Beyond the Kalman Filter: Particle Filters for Tracking Applications; Artech House: London, UK, 2003. [Google Scholar]
  5. Dong, L.; Xu, H.; Feng, X.; Han, X.; Yu, C. An Adaptive Target Tracking Algorithm Based on EKF for AUV with Unknown Non-Gaussian Process Noise. Appl. Sci. 2020, 10, 3413. [Google Scholar] [CrossRef]
  6. Jeon, J.; Hwang, Y.; Jeong, Y.; Park, S.; Kweon, I.S.; Choi, S.B. Lane Detection Aided Online Dead Reckoning for GNSS Denied Environments. Sensors 2021, 21, 6805. [Google Scholar] [CrossRef]
  7. Liu, X.; Wen, C.; Sun, X. Design Method of High-Order Kalman Filter for Strong Nonlinear System Based on Kronecker Product Transform. Sensors 2022, 22, 653. [Google Scholar] [CrossRef]
  8. Wang, D.; Zhang, H.; Ge, B. Adaptive Unscented Kalman Filter for Target Tacking with Time-Varying Noise Covariance Based on Multi-Sensor Information Fusion. Sensors 2021, 21, 5808. [Google Scholar] [CrossRef]
  9. Tanskanen, P.; Naegeli, T.; Pollefeys, M.; Hilliges, O. Semi-Direct EKF-Based Monocular Visual-Inertial Odometry. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015. [Google Scholar] [CrossRef]
  10. Liang, Q.; Liu, M. A Tightly Coupled VLC-Inertial Localization System by EKF. IEEE Robot. Autom. Lett. 2020, 5, 3129–3136. [Google Scholar] [CrossRef]
  11. Alatise, M.B.; Hancke, G.P. Pose Estimation of a Mobile Robot Based on Fusion of IMU Data and Vision Data Using an Extended Kalman Filter. Sensors 2017, 17, 2164. [Google Scholar] [CrossRef] [Green Version]
  12. Zhang, L.; Zhai, Z.; He, L.; Wen, P.; Niu, W. Infrared-Inertial Navigation for Commercial Aircraft Precision Landing in Low Visibility and GPS-Denied Environments. Sensors 2019, 19, 408. [Google Scholar] [CrossRef] [Green Version]
  13. Santos, N.P.; Lobo, V.; Bernardino, A. Unmanned Aerial Vehicle Tracking Using a Particle Filter Based Approach. In Proceedings of the 2019 IEEE Underwater Technology (UT), Kaohsiung, Taiwan, 16–19 April 2019. [Google Scholar] [CrossRef]
  14. Kim, S.-B.; Lee, S.-Y.; Choi, J.-H.; Choi, K.-H.; Jang, B.-T. A Bimodal Approach for GPS and IMU Integration for Land Vehicle Applications. In Proceedings of the 2003 IEEE 58th Vehicular Technology Conference, VTC 2003-Fall, (IEEE Cat. No.03CH37484), Orlando, FL, USA, 6–9 October 2003. [Google Scholar] [CrossRef]
  15. Carrillo, L.R.G.; López, A.E.D.; Lozano, R.; Pégard, C. Combining Stereo Vision and Inertial Navigation System for a Quad-Rotor UAV. J. Intell. Robot. Syst. 2011, 65, 373–387. [Google Scholar] [CrossRef]
  16. Yang, T.; Ren, Q.; Zhang, F.; Xie, B.; Ren, H.; Li, J.; Zhang, Y. Hybrid Camera Array-Based UAV Auto-Landing on Moving UGV in GPS-Denied Environment. Remote Sens. 2018, 10, 1829. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, Z.; She, H.; Si, W. Autonomous Landing of Multi-Rotors UAV with Monocular Gimbaled Camera on Moving Vehicle. In Proceedings of the 2017 13th IEEE International Conference on Control & Automation (ICCA), Orhid, Macedonia, 3–6 July 2017. [Google Scholar] [CrossRef]
  18. Marut, A.; Wojtowicz, K.; Falkowski, K. ArUco Markers Pose Estimation in UAV Landing Aid System. In Proceedings of the 2019 IEEE 5th International Workshop on Metrology for AeroSpace (MetroAeroSpace), Torino, Italy, 19–21 June 2019; pp. 261–266. [Google Scholar]
  19. Lee, S.; Shim, T.; Kim, S.; Park, J.; Hong, K.; Bang, H. Vision-Based Autonomous Landing of a Multi-Copter Unmanned Aerial Vehicle Using Reinforcement Learning. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Dallas, TX, USA, 12–15 June 2018; pp. 108–114. [Google Scholar]
  20. Dinh, T.H.; Hong, H.L.T. Detection and Localization of Helipad in Autonomous UAV Landing: A Coupled Visual-Inertial Approach with Artificial Intelligence. Tạp Chí Khoa Học Giao Thông Vận Tải 2020, 71, 828–839. [Google Scholar]
  21. Arantes, J.D.S.; Arantes, M.D.S.; Missaglia, A.B.; Simoes, E.D.V.; Toledo, C.F.M. Evaluating Hardware Platforms and Path Re-planning Strategies for the UAV Emergency Landing Problem. In Proceedings of the 2017 IEEE 29th International Conference on Tools with Artificial Intelligence (ICTAI), Boston, MA, USA, 6–8 November 2017; pp. 937–944. [Google Scholar]
  22. Kuchár, D.; Schreiber, P. Comparison of UAV Landing Site Classifications with Deep Neural Networks. In Computer Science On-line Conference; Springer: Cham, Switzerland, 2021; pp. 55–63. [Google Scholar]
  23. Sayfeddine, D. Control of Fixed-Wing UAV at Levelling Phase Using Artificial Intelligence. IOP Conf. Ser. Mater. Sci. Eng. 2018, 327, 022092. [Google Scholar] [CrossRef]
  24. Ayoub, N.; Schneider-Kamp, P. Real-Time On-Board Deep Learning Fault Detection for Autonomous UAV Inspections. Electronics 2021, 10, 1091. [Google Scholar] [CrossRef]
  25. Ayoub, N.; Schneider-Kamp, P. Real-time On-Board Detection of Components and Faults in an Autonomous UAV System for Power Line Inspection. In Proceedings of the 1st International Conference on Deep Learning Theory and Applications, Paris, France, 8–10 July 2020; pp. 68–75. [Google Scholar]
  26. Fu, X.; Zhu, F.; Wu, Q.; Sun, Y.; Lu, R.; Yang, R. Real-Time Large-Scale Dense Mapping with Surfels. Sensors 2018, 18, 1493. [Google Scholar] [CrossRef] [Green Version]
  27. De Souza, J.P.C.; Marcato, A.L.M.; de Aguiar, E.P.; Jucá, M.A.; Teixeira, A.M. Autonomous Landing of UAV Based on Artificial Neural Network Supervised by Fuzzy Logic. J. Control. Autom. Electr. Syst. 2019, 30, 522–531. [Google Scholar] [CrossRef]
  28. Maturana, D.; Scherer, S. 3D Convolutional Neural Networks for Landing Zone Detection from LiDAR. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015. [Google Scholar] [CrossRef]
  29. Sadhu, A.K.; Konar, A. Multi-Agent Coordination: A Reinforcement Learning Approach; John and Wiley and Sons: Hoboken, NJ, USA, 2020. [Google Scholar]
  30. Wang, L.; Wang, K.; Pan, C.; Xu, W.; Aslam, N.; Hanzo, L. Multi-Agent Deep Reinforcement Learning-Based Trajectory Planning for Multi-UAV Assisted Mobile Edge Computing. IEEE Trans. Cogn. Commun. Netw. 2020, 7, 73–84. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Mou, Z.; Gao, F.; Jiang, J.; Ding, R.; Han, Z. UAV-Enabled Secure Communications by Multi-Agent Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2020, 69, 11599–11611. [Google Scholar] [CrossRef]
  32. Polvara, R.; Patacchiola, M.; Sharma, S.; Wan, J.; Manning, A.; Sutton, R.; Cangelosi, A. Toward End-to-End Control for UAV Autonomous Landing via Deep Reinforcement Learning. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Dallas, TX, USA, 12–15 June 2018. [Google Scholar] [CrossRef]
  33. Kraus, K.; Harley, I.A.; Kyle, S. Photogrammetry: Geometry from Images and Laser Scans; Walter De Gruyter: Berlin, Germany, 2007. [Google Scholar]
  34. Gruen, A.; Huang, S.T.; Gruen, A. Calibration and Orientation of Cameras in Computer Vision; Gruen, A., Huang, S.T., Eds.; Springer: Berlin/Heidelberg, Germany, 2021; Available online: https://www.springer.com/gp/book/9783540652830 (accessed on 29 September 2021).
  35. Luhmann, T.; Robson, S.; Kyle, S.; Boehm, J. Close-Range Photogrammetry and 3D Imaging; Walter De Gruyter: Berlin, Germany, 2019. [Google Scholar]
  36. El-Ashmawy, K. Using Direct Linear Transformation (DLT) Method for Aerial Photogrammetry Applications. ResearchGate, 16 October 2018. Available online: https://www.researchgate.net/publication/328351618_Using_direct_linear_transformation_DLT_method_for_aerial_photogrammetry_applications (accessed on 29 September 2021).
  37. Poynton, C.A. Digital Video and HD: Algorithms and Interfaces; Morgan Kaufmann: Waltham, MA, USA, 2012. [Google Scholar]
  38. Liang, Y. Salient Object Detection with Convex Hull Overlap. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 4605–4612. Available online: https://arxiv.org/abs/1612.03284 (accessed on 4 October 2021).
  39. Lin, S.; Garratt, M.A.; Lambert, A.J. Monocular Vision-Based Real-Time Target Recognition and Tracking for Autonomously Landing an UAV in a Cluttered Shipboard Environment. Auton. Robot. 2016, 41, 881–901. [Google Scholar] [CrossRef]
  40. Yadav, A.; Yadav, P. Digital Image Processing; University Science Press: New Delhi, India, 2021. [Google Scholar]
  41. Arthur, D.; Vassilvitskii, S. K-means++: The Advantages of Careful Seeding. Available online: http://ilpubs.stanford.edu:8090/778/1/2006-13.pdf (accessed on 10 December 2021).
  42. Fritsch, F.N.; Butland, J. A Method for Constructing Local Monotone Piecewise Cubic Interpolants. SIAM J. Sci. Stat. Comput. 1984, 5, 300–304. [Google Scholar] [CrossRef] [Green Version]
  43. Cleve, B.M. Numerical Computing with MATLAB; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2004. [Google Scholar]
  44. Bunke, H.; Bruckstein, A.; Dori, D. Shape, Structure and Pattern Recognition; Bunke, H., Bruckstein, A., Dori, D., Eds.; World Scientific: Singapore, 1995. [Google Scholar]
  45. Valavanis, K.P.; Oh, P.; Piegl, L.A. Unmanned Aircraft Systems: International Symposium on Unmanned Aerial Vehicles, UAV’08; Valavanis, K.P., Oh, P., Piegl, L.A., Eds.; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  46. Ma, Y.; Soatto, S.; Košecká, J.; Sastry, S. An Invitation to 3-D Vision: From Images to Geometric Models; Springer: New York, NY, USA, 2004; Volume 26. [Google Scholar]
  47. Aghajan, H.; Cavallaro, A. Multi-Camera Networks: Principles and Applications; Aghajan, H., Cavallaro, A., Eds.; Academic Press: Cambridge, MA, USA, 2009. [Google Scholar]
  48. Yang, D. Informatics in Control, Automation and Robotics: Volume 2; Yang, D., Ed.; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 33. [Google Scholar]
  49. Hernandez-Matas, C.; Argyros, A.A.; Zabulis, X. Chapter 4: Retinal Image Preprocessing, Enhancement, and Registration. In Computational. Retinal Image Analysis; Elsevier: Amsterdam, The Netherlands, 2019; pp. 59–77. [Google Scholar] [CrossRef]
  50. Faugeras, O.; Luong, Q.T. The Geometry of Multiple Images: The Laws that Govern the Formation of Multiple Images of a Scene and Some of their Applications; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  51. Nath, V.; Levinson, S.E. Autonomous Robotics and Deep Learning; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  52. Triggs, B.; Zisserman, A.; Szeliski, R. Vision Algorithms: Theory and Practice: In Proceedings of the International Workshop on Vision Algorithms, Corfu, Greece, 21–22 September 1999; Triggs, B., Zisserman, A., Szeliski, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  53. Strang, G.; Borre, K. Linear Algebra, Geodesy, and GPS; Wellesley-Cambridge Press: Wellesley, MA, USA, 1997. [Google Scholar]
  54. Agoston, M.K.; Agoston, M.K. Computer Graphics and Geometric Modeling; Springer: Berlin/Heidelberg, Germany, 2005; Volume 1, pp. 301–304. [Google Scholar]
  55. Carter, N. Introduction to the Mathematics of Computer Graphics; MAA Press: New Denver, BC, Canada, 2016; Volume 51. [Google Scholar]
  56. Bronstein, A.M.; Bronstein, M.M.; Kimmel, R. Numerical Geometry of Non-Rigid Shapes; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
Figure 1. Overview of the IHRCF algorithm and evaluation process.
Figure 1. Overview of the IHRCF algorithm and evaluation process.
Sensors 22 01870 g001
Figure 2. Drone downward camera view of the landing platform-u and v Pixel coordinate.
Figure 2. Drone downward camera view of the landing platform-u and v Pixel coordinate.
Sensors 22 01870 g002
Figure 3. Cartesian coordinates of the landing platform and the designated drone.
Figure 3. Cartesian coordinates of the landing platform and the designated drone.
Sensors 22 01870 g003
Figure 4. Camera range calibration procedures in each stage.
Figure 4. Camera range calibration procedures in each stage.
Sensors 22 01870 g004
Figure 5. Range sensors’ touch points centralization with their corresponding color pads.
Figure 5. Range sensors’ touch points centralization with their corresponding color pads.
Sensors 22 01870 g005
Figure 6. Color segmentation result. (a) Color image-(b) Masked color pads’ image.
Figure 6. Color segmentation result. (a) Color image-(b) Masked color pads’ image.
Sensors 22 01870 g006
Figure 7. Range sensor positions and angles w.r.t the camera.
Figure 7. Range sensor positions and angles w.r.t the camera.
Sensors 22 01870 g007
Figure 8. Touch points of the range sensor (blue), AprilTag center (cyan), AprilTag corners (cyan), and of the sensor unit principle point (yellow).
Figure 8. Touch points of the range sensor (blue), AprilTag center (cyan), AprilTag corners (cyan), and of the sensor unit principle point (yellow).
Sensors 22 01870 g008
Figure 9. Transformation from world-corresponding points to sensor touch points in the image plane.
Figure 9. Transformation from world-corresponding points to sensor touch points in the image plane.
Sensors 22 01870 g009
Figure 10. Image plane, red plane, and world plane, in light green, relations by homography.
Figure 10. Image plane, red plane, and world plane, in light green, relations by homography.
Sensors 22 01870 g010
Figure 11. Estimation of the real-world 3D points in the camera frame.
Figure 11. Estimation of the real-world 3D points in the camera frame.
Sensors 22 01870 g011
Figure 12. Experiment bench for the assessment of the technique along with Coordinate system used.
Figure 12. Experiment bench for the assessment of the technique along with Coordinate system used.
Sensors 22 01870 g012
Figure 13. Software-in-the-Loop (SIL) data transmission and the connection protocol.
Figure 13. Software-in-the-Loop (SIL) data transmission and the connection protocol.
Sensors 22 01870 g013
Figure 14. Euler angles, namely roll-pitch-yaw for the Stewart platform using IHRCF, GT, and ATDA.
Figure 14. Euler angles, namely roll-pitch-yaw for the Stewart platform using IHRCF, GT, and ATDA.
Sensors 22 01870 g014
Figure 15. Translational results in X, Y, and Z directions for the Stewart platform using IHRCF, GT, and ATDA.
Figure 15. Translational results in X, Y, and Z directions for the Stewart platform using IHRCF, GT, and ATDA.
Sensors 22 01870 g015
Figure 16. Plots of 3D trajectories of the Stewart platform using the IHRCF, ATDA base pose estimation and the GT, with the X-Z view, Y-Z view, X-Y view, and 3D trajectory displayed in (ad), respectively.
Figure 16. Plots of 3D trajectories of the Stewart platform using the IHRCF, ATDA base pose estimation and the GT, with the X-Z view, Y-Z view, X-Y view, and 3D trajectory displayed in (ad), respectively.
Sensors 22 01870 g016aSensors 22 01870 g016b
Figure 17. IHRCF-based Stewart platform Euler angles estimation error along the X (blue), Y (red), and Z (yellow) axes.
Figure 17. IHRCF-based Stewart platform Euler angles estimation error along the X (blue), Y (red), and Z (yellow) axes.
Sensors 22 01870 g017
Figure 18. IHRCF translational pose estimation errors for the Stewart platform along the X (blue), Y (red), and Z (yellow) axes.
Figure 18. IHRCF translational pose estimation errors for the Stewart platform along the X (blue), Y (red), and Z (yellow) axes.
Sensors 22 01870 g018
Figure 19. ATDA base Stewart platform translational estimation error along the X (blue), Y (red), and Z (yellow) axes.
Figure 19. ATDA base Stewart platform translational estimation error along the X (blue), Y (red), and Z (yellow) axes.
Sensors 22 01870 g019
Figure 20. ATDA base Stewart platform angular estimation error along the X (blue), Y (red), and Z (yellow) axes.
Figure 20. ATDA base Stewart platform angular estimation error along the X (blue), Y (red), and Z (yellow) axes.
Sensors 22 01870 g020
Table 1. YCBCR base color thresholding for blue, green, yellow, and red.
Table 1. YCBCR base color thresholding for blue, green, yellow, and red.
ColorChannel YChannel CBChannel CR
Blue0 ≤ Y ≤ 165139 ≤ CB ≤ 2550 ≤ CR ≤ 255
Green0 ≤ Y ≤ 2550 ≤ CB ≤ 1550 ≤ CR ≤ 90
Yellow103 ≤ Y≤ 2550 ≤ CB ≤ 950 ≤ CR ≤ 255
Red0 ≤ Y ≤ 1600 ≤ CB ≤ 255167 ≤ CR ≤ 255
Table 2. Mean absolute error comparison of AprilTag and IHRCEFcalibration algorithms.
Table 2. Mean absolute error comparison of AprilTag and IHRCEFcalibration algorithms.
Absolute Error. ATDAIHRCF
Translations
(m)
0.01620.01340.06970.00350.00390.0041
Angles
(degree)
2.98431.6571.77430.981.37311.180
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sefidgar, M.; Landry, R., Jr. Landing System Development Based on Inverse Homography Range Camera Fusion (IHRCF). Sensors 2022, 22, 1870. https://doi.org/10.3390/s22051870

AMA Style

Sefidgar M, Landry R Jr. Landing System Development Based on Inverse Homography Range Camera Fusion (IHRCF). Sensors. 2022; 22(5):1870. https://doi.org/10.3390/s22051870

Chicago/Turabian Style

Sefidgar, Mohammad, and Rene Landry, Jr. 2022. "Landing System Development Based on Inverse Homography Range Camera Fusion (IHRCF)" Sensors 22, no. 5: 1870. https://doi.org/10.3390/s22051870

APA Style

Sefidgar, M., & Landry, R., Jr. (2022). Landing System Development Based on Inverse Homography Range Camera Fusion (IHRCF). Sensors, 22(5), 1870. https://doi.org/10.3390/s22051870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop