Next Article in Journal
A Lightweight Neural Network for the Real-Time Dehazing of Tidal Flat UAV Images Using a Contrastive Learning Strategy
Previous Article in Journal
Fed4UL: A Cloud–Edge–End Collaborative Federated Learning Framework for Addressing the Non-IID Data Issue in UAV Logistics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Visual Navigation Algorithm for UAV Based on Visual-Geography Optimization

Xi’an Research Institute of Hi-Tech, Xi’an 710025, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(7), 313; https://doi.org/10.3390/drones8070313
Submission received: 8 April 2024 / Revised: 4 July 2024 / Accepted: 8 July 2024 / Published: 10 July 2024

Abstract

:
The estimation of Unmanned Aerial Vehicle (UAV) poses using visual information is essential in Global Navigation Satellite System (GNSS)-denied environments. In this paper, we propose a UAV visual navigation algorithm based on visual-geography Bundle Adjustment (BA) to address the challenge of missing geolocation information in monocular visual navigation. This algorithm presents an effective approach to UAV navigation and positioning. Initially, Visual Odometry (VO) was employed for tracking the UAV’s motion and extracting keyframes. Subsequently, a geolocation method based on heterogeneous image matching was utilized to calculate the geographic pose of the UAV. Additionally, we introduce a tightly coupled information fusion method based on visual-geography optimization, which provides a geographic initializer and enables real-time estimation of the UAV’s geographical pose. Finally, the algorithm dynamically adjusts the weight of geographic information to improve optimization accuracy. The proposed method is extensively evaluated in both simulated and real-world environments, and the results demonstrate that our proposed approach can accurately and in real-time estimate the geographic pose of the UAV in a GNSS-denied environment. Specifically, our proposed approach achieves a root-mean-square error (RMSE) and mean positioning accuracy of less than 13 m.

1. Introduction

In recent years, UAVs have been extensively utilized in various domains, including military operations, search and rescue missions, and terrain exploration [1,2,3,4,5]. While GNSS technology can provide precise geolocation information for UAVs, it is susceptible to signal blockage and interference from adversarial sources [6]. Therefore, it is imperative to develop methodologies for UAV localization in GNSS-denied environments [7]. In such challenging scenarios, UAVs must rely solely on their onboard camera to infer their pose without any external infrastructure. Hence, developing robust techniques for accurate and independent UAV positioning becomes paramount.
In GNSS-denied environments, UAVs can rely on visual sensors for precise pose estimation [8], enabling autonomous navigation and positioning. Visual sensors play a crucial role in environmental perception by capturing image information that provides UAVs with abundant contextual data. The technologies for UAV visual navigation and positioning can be primarily categorized into two approaches: VO, which operates without a map, and geolocation based on image matching, which utilizes an existing map.
The UAV can estimate its relative pose using VO, which solely relies on a cost-effective and lightweight vision sensor to accurately determine the local pose of the UAV in real-time. Depending on different approaches for associating features between frames, it can be categorized into feature-based, direct, and hybrid methods. The feature-based method is employed to extract features from the image for matching, while camera pose estimation is achieved by minimizing the reprojection error. PTAM [9] and ORB-SLAM [10] are both classical SLAM systems that utilize the feature-based approach. Among them, ORB-SLAM leverages ORB [11] features in matching, tracking, relocation, and loop closing processes, thereby enhancing information interaction efficiency across different threads. Subsequently, ORB-SLAM2 [12] and ORB-SLAM3 [13] also employ ORB features. On the other hand, direct methods directly use the pixel intensities in the images and estimate motion by minimizing a photometric error. DSO [14] and VINS-Mono [15] are advanced SLAM systems that employ the direct method. SVO [16] is a hybrid VO approach that initially extracts FAST features but uses a direct method to track features and any pixel with a nonzero intensity gradient from frame to frame.
Currently, state-of-the-art VO algorithms employ keyframe-based techniques [17], which provide higher accuracy compared to filtering methods at the same computational cost [18]. Keyframe-based techniques utilize a multi-threaded approach for tracking the robot’s motion, comprising a tracking thread responsible for the UAV’s motion tracking and keyframe selection and a mapping thread that performs computationally expensive yet more accurate BA optimization at the keyframe rate. This enables the VO system to achieve real-time performance while enhancing its localization accuracy. Although VO has demonstrated reliable precision in short-term operations, it tends to accumulate drift over long-term operations. Moreover, VO fails to provide a more comprehensive geographic pose, which limits its applicability in aerial scenarios. Therefore, it is crucial to incorporate supplementary techniques to address these challenges.
With the continuous advancement of image-matching technology, sophisticated image-matching networks can effectively address the challenges posed by significant variations in seasons, lighting conditions, and other factors between satellite images and aerial images [19,20]. Consequently, these networks facilitate establishing correspondences between aerial and satellite images to accurately determine the geographic pose of the UAV. Various methods [21] have been developed for image matching and localization in GNSS-denied environments. Goforth et al. [22] employed a deep convolutional neural network (CNN) with an iterative closest landmark keypoints (ICLK) layer to align aerial images and satellite images. Chen et al. [23] proposed an image-based geolocation method that is adaptable to a downward-tilted camera, which necessitates prior offline dataset preparation.
However, a drawback of geolocation methods based on image matching is their significant computational cost, leading to substantial delays and rendering them unsuitable for meeting the real-time positioning requirements of UAVs [24]. To enhance the positioning performance of visual navigation algorithms, certain studies have focused on integrating VO with image-matching methods. Kinnari et al. [25] orthorectified UAV images based on VIO and planarity assumptions, which is also applicable to a downward-tilted camera configuration. The ortho-projected image is utilized for matching with the satellite images, and the geolocation results are combined with the pose estimation from VIO in a particle filter framework. Hao et al. [26] proposed a geolocation method that utilizes global pose graph optimization to integrate RVIO measurements and image registration. Zhang et al. [27] proposed an iterative trajectory fusion pipeline that integrates vSLAM measurements and image matching through solving a scaled pose graph problem.
The aforementioned studies employ a trajectory fusion method to integrate the measurements from VO and image matching, treating them as two relatively independent modules. While this design simplifies the system architecture, it also results in a disconnection between VO and image matching, impeding the immediate feedback of image-matching results in the VO estimation process. This limitation may potentially impact the overall accuracy and robustness of the system.
To fulfill the aforementioned requirements and address existing concerns, we have developed a tightly coupled visual navigation algorithm for UAVs based on visual-geography optimization. As depicted in Figure 1, the algorithm is structured into three concurrent threads inspired by the SLAM multi-threading concept: tracking thread, mapping thread, and image-matching thread. The tracking thread enhances real-time performance by monitoring the UAV’s movement, while the mapping thread optimizes the VO map through the fusion of visual information from VO and geographic positioning data obtained via image matching. Additionally, the image-matching thread provides globally unbiased geographic positioning information. The image-matching thread incorporates the existing coarse-to-fine image-matching method [23], while also introducing a prior-based approach. The prior-based image-matching method not only reduces computational costs but also enhances positioning accuracy, which holds significant implications for the fusion algorithm with VO. In the mapping thread, we have developed a robust geographic initializer and map fusion algorithm. The geographic initializer accurately estimates initialization parameters to align the VO world coordinate system with the geographic coordinate system. The map fusion algorithm utilizes geographical positioning information obtained from image matching for real-time optimization of the VO map. In summary, our algorithm enables real-time and precise estimation of a UAV’s geographic pose through visual-geography optimization.
This work makes the following contributions:
  • We propose a UAV visual navigation algorithm that combines the merits of VO and image matching.
  • We introduce a geolocation method based on heterogeneous image matching, which employs the coarse-to-fine and prior-based image-matching methods to enhance the accuracy of geolocation. This method effectively leverages the operational characteristics of VO to provide precise geolocation information.
  • We present a tightly coupled information fusion method based on visual-geography optimization that jointly optimizes the visual and geolocation information of keyframes to facilitate tightly coupled geolocation in algorithms. Compared with the existing trajectory fusion method, our method achieves higher positioning accuracy.
The remaining sections of this paper are organized as follows. Section 2 presents a detailed explanation of the visual navigation algorithm for the UAVs proposed in this study. In Section 3, we discuss our real-world experiments conducted to validate our approach. Concluding remarks are provided in Section 4. Finally, an extensive discussion is presented in Section 5.

2. Method

The proposed visual navigation algorithm for the UAV is detailed in this section. This paper presents a visual-geography optimization method for integrating visual information (VO map consists of the UAV’s poses and map points) and geographic information (the geographical poses of the UAV are determined through image matching), enabling geographic initializer and real-time pose estimation. More details are shown in Algorithm 1. The following section introduces the fusion of visual and geographic information with tight coupling as well as geolocation based on heterogeneous image matching.
Algorithm 1 UAV visual navigator algorithm
Input: UAV images I U and satellite map tiles.
Output: Estimated UAV trajectory.
  1:  for all UAV images do
  2:    Extract ORB features.
  3:    VO initialization.
  4:    Track UAV movement.
  5:    if keyframe then
  6:      if  ! S g w then
  7:        Visual BA optimization.
  8:         T g i = CtoF _ Image _ Matching T U i   Sec .   2.2 . 1  
  9:        Calculate the initial value of S g w   by   Sec .   2.1 . 2 ( 2 )  
10:        Reliability check of S g w
11:      else if !GeoInitialized then
12:        Visual BA optimization.
13:         T g i =   Prior _ Image _ Matching T U i , T w i   Sec .   2.2 . 2
14:        Geographic initialization.   Sec .   2.1 . 2 ( 3 )  
15:      else
16:         T g i = Prior _ Image _ Matching T U i , T w i   Sec .   2.1 . 1
17:        Visual-geography BA optimization.
18:      end if
19:    end if
20:  end for

2.1. Tightly Coupled Visual and Geographic Information Fusion

This paper proposes a tightly coupled visual-geographic fusion method based on visual-geography optimization, aiming to enhance the algorithm’s navigation and positioning performance by effectively utilizing both visual and geolocation information from keyframes. The section of the code is available at https://github.com/XuWeibo-code/UAV_Visual_Navigation.git (accessed on 4 June 2024).

2.1.1. Visual-Geography Optimization

The visual-geography optimization proposed in this paper is essentially a nonlinear optimization, enabling the joint optimization of visual information and geographic information.
The 3D transformation involved in the UAV’s motion primarily consists of the pose transformation matrix T and the similarity transformation matrix S , where T SE 3 and its tangent vector δ s e 3 . Similarly, S Sim 3 and its tangent vector ω s i m 3 . The specific form of this transformation is as follows:
T = R t 0 1 SE 3 ; δ = ω υ T s e 3 S = s R t 0 1 Sim 3 ; ε = ω υ λ T s i m 3 ,
where R is the 3 × 3 rotation matrix, t is the 3 × 1 translation vector, and s represents the scale. The components in the tangent space are denoted as follows: ω is a 3 × 1 vector, v is a 3 × 1 vector, and λ is a scalar representing the rotation component, translation component, and scale component, respectively.
The optimization method employs four coordinate systems, namely the SLAM world coordinate system, geographic coordinate system, local camera coordinate system, and global camera coordinate system. In this paper, S g w S i m 3 is utilized to denote the similarity transformation matrix from the SLAM world coordinate system to the geographic coordinate system. S c S i m 3 represents the similarity transformation matrix from the local camera coordinate system to the global camera coordinate system. T g i S E 3 signifies the pose transformation matrix from the global camera coordinate system to the geographic coordinate system in frame i. T w i S E 3 denotes the pose transformation matrix from the local camera coordinate system to the SLAM world coordinates in frame i. The variables S g w and S c share a common scale parameter, which can be expressed as follows:
S g w = s g w R g w t g w 0 1 S c = s g w I 3 × 3 0 0 1 ,
The conventional VO estimates the pose of a UAV, denoted as T w i , by optimizing the visual information through visual BA (Figure 2a). In contrast, the proposed visual-geography BA (Figure 2b) jointly optimizes the parameters T w i and S g w using both visual and geolocation information from keyframes.
First, the visual residual r i j v is defined as the reprojection error of the map point j in frame i.
r i j v = x ^ j π T w i X j ,
where x ^ j is the observed value of the map points, X j represents the 3D location of the map point j, and π · corresponds to the camera projection function.
Then, the geographic residual r i g of frame i is defined as follows:
r i g = Log T ^ g i 1 S g w T w i S c 1 ,
where Log : SE 3 6 is a mapping from the Lie group Sim 3 to the vector space s i m 3 , T ^ g i is an observability measure, and S g w and T w i are the optimization variables.
Given k keyframes and their states T ¯ k T 0 T k - 1 , l 3D map points and their states χ l X 0 X k - 1 , the cost function of this optimization is as follows:
T i , X j , S g w | T i T ¯ k , X j χ l = arg min T i , X j i = 0 k 1 j = 0 l 1 ρ r k j c c 2 + ρ r k g g 2 ,
where C and g represent the weight factors of visual residual and geographical residual, respectively. The value of C is obtained from [13], while the derivation of g is elaborated in detail in Section 2.1.3. For visual and geographic residuals, we use a robust Huber kernel function ρ · to reduce the influence of spurious matchings.
To enhance the optimization efficiency of visual-geography BA, the algorithm must optimize the Jacobian matrix of the residual for the variable. Among them, the Jacobian matrix of the visual residual concerning the optimization variables T w i and X j can be obtained from [13]. Therefore, this paper only derives the Jacobian matrix of the geographic residual to the optimization variables S g w and T w i .
In Equation (4), the geographic residual r i g is a 7-dimensional vector, the tangent vector ε g w of the S g w is also a 7-dimensional vector, and the tangent vector δ w i of T w i is a 6-dimensional vector. The Jacobian matrices of the geographic residual concerning ε g w and δ w i are as follows:
r i g δ g w = A d j T ^ g i 1 A d j r i g 0 6 × 6 0 0 1 r i g δ w i = A d j T ^ g i 1 S g w I 6 × 6 0 ,
where A d j S is the adjoint matrix of S , which takes the following form:
A d j S = R 0 0 t × R s R t 0 0 1 ,
The optimization process of this BA requires the assignment of initial values. The initial values for the variables T w i and X j are provided by VO, while the initial value for the variable S g w is calculated during the geographic initializer.

2.1.2. Geographic Initializer

The algorithm utilizes a [13]-based VO technique to track the motion of the UAV in real-time and perform a visual initializer. To enhance the real-time performance and accuracy of the SLAM, our algorithm employs a subset of representative frames as keyframes for BA optimization and image matching. During the selection process, temporal intervals, visual content disparities, camera movements, and co-visibility relationships between the current frame and other keyframes are all taken into consideration.
The algorithm then employs a rapid and robust geographic initializer, which aims to accurately estimate the initialization parameter S g w that aligns the SLAM world coordinate system with the geographic coordinate system. The transformation relationship between the local pose in the SLAM world coordinate system and the geographical pose is described as follows:
T g i = S g w T w i S c 1 ,
We state the geographic initializer as a map estimation problem, which can be divided into three sequential steps.
(1)
Visual BA optimization
The algorithm utilizes a vision-only BA (Figure 2a) to optimize the UAV’s pose and map points during the initial phase of its operation, when sufficient geolocation information is not yet available.
(2)
Calculate the initial value of S g w
The algorithm calculates the initial value of S g w after receiving reliable geolocation information from the two keyframes. At this stage, the local poses of the two keyframes are denoted as T w 1 and T w 2 , while their geographical poses are represented by T g 1 and T g 2 . The scale component s g w of the similarity transformation matrix S g w is as follows:
s g w = t g 2 t g 1 t w 2 t w 1 ,
The rotation component R g w of the similarity transformation matrix S g w is as follows:
R g w = R g 1 R w 1 T R g w = R g 2 R w 2 T ,
The translation component t g w of the similarity transformation matrix S g w is as follows:
t g w = R g 1 R w 1 T t w 1 + s g w 1 t g 1 t g w = R g 2 R w 2 T t w 2 + s g w 1 t g 2 ,
The initializer calculates the initial value of S g w at this stage. It is important to note that the geolocation information obtained during this process relies on a coarse-to-fine image-matching method, which may result in mismatching. Therefore, the algorithm utilizes an error coefficient to check the reliability of S g w . We construct S g w and S g w using R g w , t g w , s g w and R g w , t g w , s g w , respectively, and the error coefficient is calculated as follows:
r S = Log S g w S g w 1 ,
The reliability of S g w is considered when r S < t h ; otherwise, the algorithm will wait for the keyframe to continue receiving geolocation information of the UAV and repeat the aforementioned steps.
(3)
Visual-geography BA
When the algorithm obtains a reliable initial value of S g w , the prior geographic pose of the UAV can be calculated based on Equation (11). At this stage, the geolocation method based on heterogeneous image matching can directly extract the satellite image according to the prior geographic pose and perform more precise fine matching. Once geolocation information has been received for N keyframes, the algorithm optimizes the UAV pose, map points, and initialization parameter S using visual-geography BA (Figure 2a) of Equation (5).

2.1.3. Map Fusion Method and Geographic Weights Update

After the completion of the geographical initialization, it remains necessary for the system to output real-time and accurate geographical pose of the UAV. To achieve this, we employ visual-geography BA to optimize both UAV poses and map points. Unlike visual-geography BA with a geographical initializer, at this stage, our algorithm does not optimize the initialization parameter.
The geolocation information obtained by image matching is subject to errors, necessitating the assignment of weights to each geographic edge in the visual-geography BA. Therefore, we propose a calculation method for quantifying the magnitude of geographic residuals.
Refine the weight of the geographic edge i in BA according to the following criteria:
q = β n φ n + β p Log T g c 1 T g c 2 + β e e i g 2 ,
where the balance factors β n , β p , and β e represent the weights assigned to different components; the variable φ n represents the number of remaining inner points after removing outer points from the matching points obtained by the RANSAC algorithm [28]; T g c and T g c denote the prior geolocation information input in the image matching and the geolocation information output by image matching, respectively; and e i g represents the geographic residual of geographic edge i in BA.
The first two terms in Equation (13) are generated during the image matching, while the last term is dynamically updated through BA. In Equation (5), q I 7 × 7 is introduced into g to fine-tune the weighting of geographic residuals in the BA. By adjusting the weight of geographic edges in BA, this algorithm enables a more accurate estimation of the UAV’s geographic pose.

2.2. Geolocation Based on Heterogeneous Image Matching

The geolocation method based on heterogeneous image matching is utilized to provide precise geolocation information for keyframes. Considering the characteristics of VO, appropriate adjustments are made to the geolocation method. In the initial stage of the algorithm, a coarse-to-fine image-matching method is employed for UAV geolocation. After the algorithm calculates the S g w , the satellite image is intercepted using the prior pose of the UAV and then directly processed by the fine matching.

2.2.1. Coarse-to-Fine Image-Matching Method

The geolocation of the UAV is achieved in this paper by implementing the coarse-to-fine image-matching method utilized in [23], even in the absence of prior geolocation information. The pipeline of the coarse-to-fine image-matching method is shown in Figure 3.
During the coarse matching stage, satellite images are preprocessed offline. These images, which have varying resolutions, are clipped into smaller map tiles and then mapped into 4096-dimensional global descriptors using NetVLAD [29]. During UAV execution, the extracted global descriptors from aerial images are also obtained using NetVLAD. The l2-norm of the differences between the aerial image’s global descriptor and every single global descriptor in the prepared array is computed and sorted in ascending order. The top 3 candidates in the sorted list are selected for subsequent fine matching.
After the coarse matching stage, the algorithm proceeds to conduct fine matching between the aerial image and the satellite image. This process involves utilizing SuperPoint [30] for extracting local descriptors from both images, generating 2D matching points between the aerial image and the candidate matching images through SuperGlue [31]. The matching result of SuperPoint + SuperGlue is shown in Figure 4.
Assuming a planar ground with zero height, these 2D-2D correspondences are transformed into a camera projection relation from the 3D coordinates in the geographic coordinate system to the 2D-pixel coordinates in the aerial image. The EPnP+RANSAC algorithm [28] is then utilized to accurately determine the geographic pose of the UAV.

2.2.2. Prior-Based Image-Matching Method

The coarse-to-fine image-matching method exhibits drawbacks, such as significant latency and a high mismatching rate. In this study, leveraging the operational characteristics of VO, we propose a prior-based image-matching approach for precise UAV geolocation.
After obtaining the geographic initialization parameter S g w , the algorithm calculates the prior geographic pose T i g T g i 1 of the keyframe i based on Equation (8). Assuming a planar ground with zero height, the transformation relationship between the geographic plane coordinates of the satellite image and pixel coordinates of the aerial image is described using a homography matrix H i g .
H i g = 1 λ K r i g 1 , r i g 2 , t i g ,
where R i g and t i g are the rotational and translational components of T i g , r i g 1 and r i g 2 are the first and second columns of R i g , respectively, K represents the camera internal reference, and λ denotes the depth of the feature point, which is equivalent to the absolute value of the z coordinate of R i g 1 t i g .
The aerial image can be transformed into a geographic plane coordinate by applying the homography matrix H i g , extracting a rectangular region in the geographic plane coordinate system for intercepting the satellite image. Subsequently, utilizing the homography matrix, the captured satellite image is projected onto the pixel coordinate of the aerial image. Figure 5 illustrates the flowchart depicting satellite image extraction.
After extracting the satellite image, the algorithm will perform fine matching, as described in Section 2.2.1, to accurately determine the geographic pose of the UAV.

3. Experimental Setups and Results

This section outlines the experimental setup and presents the test results. To simulate real-flight conditions, we have configured the Ubuntu environment on NVIDIA Jetson Xavier NX and conducted experimental validation.

3.1. Simulation Dataset

3.1.1. Setups

To evaluate the efficacy of the algorithm proposed in this paper, we generated a simulation dataset using a custom simulation program. This program employs the camera projection principle to transform satellite images captured in visible light mode into image sequences by configuring camera parameters and executing trajectories.
The trajectory of the UAV in the simulation dataset is depicted in Figure 6, with a satellite image serving as the background. Specific characteristics of the dataset are elaborated upon in Table 1.

3.1.2. Geolocation Performance

To evaluate the performance of our proposed algorithm, we compare it with various algorithms in the simulation dataset, including the VO algorithm (the monocular vision mode of ORB-SLAM3) [13], the VO algorithm after geographic initialization, the geolocation algorithm based on coarse-to-fine image matching [23], and our proposed algorithm. The 2D errors between the different algorithms and the ground-truth trajectory are depicted in Figure 7. The trajectory obtained from the VO algorithm is converted into the geographic coordinate through a similarity transformation matrix, which is derived using multiple ground truth poses by solving a least-squares problem [32].

3.2. Real-World Dataset

3.2.1. Setups

To evaluate the performance of the proposed algorithm in a real environment, we employed the DJI M300 RTK UAV equipped with an H20 pan-tilt camera to generate two distinct datasets for visual navigation and positioning of the UAV.
The trajectory of the UAV in real-world datasets is depicted in Figure 8, with the accurate trajectory obtained by RTK represented by the red curve. Furthermore, two videos were captured from a downward-facing perspective by the UAV, and their specific characteristics are detailed in Table 2.

3.2.2. Geolocation Performance

To evaluate the performance of our proposed algorithm, we conducted a comparative analysis with various state-of-the-art algorithms using real-world datasets. These include the VO algorithm, the VO algorithm after geographic initialization, the geolocation algorithm based on coarse-to-fine image matching, and our proposed algorithm. The 2D errors between the different algorithms and the ground-truth trajectory are depicted in Figure 9.

3.3. Ablation Study

We conduct an ablation study to evaluate the contribution of each module in the proposed algorithm. The translation errors by removing different configurations are present in Table 3. We can observe that the result using the full proposed framework achieves the best performance.

3.4. Analysis of Experimental Findings

We conducted geolocation experiments in a simulated dataset and two segments of real-world datasets. Throughout the experiment, we evaluated the proposed geographic initialization method, which successfully accomplished geographic initialization within 5 s. Comparative analysis with the VO algorithm revealed that performing geographic initialization significantly enhances stability. The mean and RMSE of the different algorithms in the geolocation experiment are compared in Table 4 and Table 5. It is evident that the algorithm proposed in this paper exhibits superior accuracy, enabling precise estimation of the geographic pose of the UAV.
The geolocation method based on coarse-to-fine image matching in the experiment exhibits high accuracy. However, the output positioning data exhibit a delay of 3 s and a mismatching probability not lower than 25%, making it unsuitable for real-time UAV positioning. The proposed algorithm achieves a data update frequency of 20 Hz/s, while the visual navigation algorithm presented in this paper enhances performance. Specifically, our proposed visual navigation algorithm enables accurate and real-time estimation of the geographic pose of the UAV in GNSS-denied environments.

4. Conclusions

This paper presents a visual navigation algorithm for UAVs based on visual-geography optimization, which effectively integrates visual and geolocation information from keyframes to achieve tightly coupled geolocation. Moreover, the algorithm incorporates a heterogeneous image-matching approach for geolocation, which combines coarse-to-fine image matching and prior-based methods. Based on the experimental results, the following conclusions can be drawn:
  • In the geolocation method based on heterogeneous image matching, our proposed prior-based image-matching method utilizes the prior information to enhance the accuracy and efficiency of geolocation for the algorithm.
  • The fusion method based on visual-geography optimization achieves stable and reliable estimation of geographic initialization parameters within 5 s, enabling real-time estimation of the UAV’s geographic pose.
  • We propose a tightly integrated fusion method that effectively combines the visual information from VO with the geolocation information obtained through the image-matching method. Experimental results demonstrate that our proposed algorithm accurately and in real-time estimates the UAV’s geolocation information solely relying on the vision sensor, even in GNSS-denied environments.

5. Discussion

Although promising results have been achieved, there are several limitations. Firstly, the proposed algorithm is more suitable for scenarios within the range of tens to hundreds of meters. For higher or lower scenarios, significant errors may arise due to the image-matching method employed in the algorithm. Secondly, to calculate the geographic pose of the UAV, we assume a planar ground with zero elevation. However, this assumption introduces a significant discrepancy in estimating the altitude of the UAV. Lastly, when confronted with low image texture or substantial variations in lighting conditions, inevitable disruptions occur in the visual navigation and positioning algorithm utilized by the UAV.
To enhance the positioning performance of UAVs in complex scenarios, our future research will focus on developing visual-inertial navigation algorithms for UAV positioning and incorporating barometers or laser rangefinders to accurately measure UAV altitude, thereby providing precise and reliable positioning information.

Author Contributions

Conceptualization, W.X. and Y.L.; methodology, W.X., D.Y. and Y.L.; software, W.X. and Y.L.; validation, Y.X, D.Y. and J.L.; formal analysis, W.X.; investigation, W.X. and D.Y.; resources, D.Y. and Y.L.; data curation, W.X., Y.L. and M.Z.; writing, original draft preparation, W.X.; writing, review and editing, W.X. and M.Z.; visualization, W.X.; supervision, D.Y. and J.L.; project administration, D.Y. and J.L.; funding acquisition, D.Y and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Foundation of China (Grant Nos. 42301535, 61673017, 61403398) and the Natural Science Foundation of Shaanxi Province (Grant Nos. 2017JM6077).

Data Availability Statement

The data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Scherer, J.; Yahyanejad, S.; Hayat, S.; Yanmaz, E.; Andre, T.; Khan, A.; Vukadinovic, V.; Bettstetter, C.; Hellwagner, H.; Rinner, B. An Autonomous Multi-UAV System for Search and Rescue. In Proceedings of the MobiSys’15: The 13th Annual International Conference on Mobile Systems, Applications, and Services, Florence, Italy, 18 May 2015. [Google Scholar]
  2. Messinger, M.; Silman, M. Unmanned aerial vehicles for the assessment and monitoring of environmental contamination: An example from coal ash spills. Environ. Pollut. 2016, 218, 889–894. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, Y.; Meng, Z.; Zou, Y.; Cao, M. Visual Object Tracking and Servoing Control of a Nano-Scale Quadrotor: System, Algorithms, and Experiments. IEEE/CAA J. Autom. Sin. 2021, 8, 344–360. [Google Scholar] [CrossRef]
  4. Ganesan, R.; Raajini, X.M.; Nayyar, A.; Sanjeevikumar, P.; Hossain, E.; Ertas, A.H. BOLD: Bio-Inspired Optimized Leader Election for Multiple Drones. Sensors 2020, 20, 3134. [Google Scholar] [CrossRef]
  5. Yayli, U.; Kimet, C.; Duru, A.; Cetir, O.; Torun, U.; Aydogan, A.; Padmanaban, S.; Ertas, A. Design optimization of a fixed wing aircraft. Int. J. Adv. Aircr. Spacecr. Sci. 2017, 4, 65–80. [Google Scholar] [CrossRef]
  6. Huang, K.W.; Wang, H.M. Combating the Control Signal Spoofing Attack in UAV Systems. IEEE Trans. Veh. Technol. 2018, 67, 7769–7773. [Google Scholar] [CrossRef]
  7. Li, Y.; Yang, D.; Wang, S.; He, H.; Hu, J.; Liu, H. Road-Network-Based Fast Geolocalization. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6065–6076. [Google Scholar] [CrossRef]
  8. Fragoso, A.T.; Lee, C.T.; McCoy, A.S.; Chung, S.-J. A seasonally invariant deep transform for visual terrain-relative navigation. Sci. Robot. 2021, 6, eabf3320. [Google Scholar] [CrossRef] [PubMed]
  9. Klein, G.; Murray, D. Parallel Tracking and Mapping for Small AR Workspaces. In Proceedings of the 2007 6th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Nara, Japan, 13–16 November 2007. [Google Scholar]
  10. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  11. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  12. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  13. Campos, C.; Elvira, R.; Rodriguez, J.J.G.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial and Multi-Map SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  14. Engel, J.; Koltun, V.; Cremers, D. Direct Sparse Odometry. IEEE Trans. Pattern Anal. 2018, 40, 611–625. [Google Scholar] [CrossRef] [PubMed]
  15. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  16. Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  17. Leutenegger, S.; Lynen, S.; Bosse, M.; Siegwart, R.; Furgale, P. Keyframe-based visual–inertial odometry using nonlinear optimization. Int. J. Robot. Res. 2015, 34, 314–334. [Google Scholar] [CrossRef]
  18. Strasdat, H.; Montiel, J.M.; Davison, A.J. Visual SLAM: Why filter? Image Vis. Comput. 2012, 30, 65–77. [Google Scholar] [CrossRef]
  19. Sarlin, P.-E.; Unagar, A.; Larsson, M.; Germain, H.; Toft, C.; Larsson, V.; Pollefeys, M.; Lepetit, V.; Hammarstrand, L.; Kahl, F.; et al. Back to the Feature: Learning Robust Camera Localization from Pixels to Pose. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021. [Google Scholar]
  20. Shetty, A.; Gao, G.X. UAV Pose Estimation using Cross-view Geolocalization with Satellite Imagery. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–22 May 2019. [Google Scholar]
  21. Shi, Y.; Li, H. Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization Using Satellite Image. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  22. Goforth, H.; Lucey, S. GPS-Denied UAV Localization using Pre-existing Satellite Imagery. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
  23. Chen, S.; Wu, X.; Mueller, M.W.; Sreenath, K. Real-time Geo-localization Using Satellite Imagery and Topography for Unmanned Aerial Vehicles. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021. [Google Scholar]
  24. Hu, S.; Lee, G.H. Image-Based Geo-Localization Using Satellite Imagery. Int. J. Comput. Vis. 2020, 128, 1205–1219. [Google Scholar]
  25. Kinnari, J.; Verdoja, F.; Kyrki, V. GNSS-denied geolocalization of UAVs by visual matching of onboard camera images with orthophotos. In Proceedings of the 2021 20th International Conference on Advanced Robotics (ICAR), Ljubljana, Slovenia, 6–10 December 2021. [Google Scholar]
  26. Hao, Y.; He, M.; Liu, Y.; Liu, J.; Meng, Z. Range–Visual–Inertial Odometry with Coarse-to-Fine Image Registration Fusion for UAV Localization. Drones 2023, 7, 540. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Shi, Y.; Wang, S.; Vora, A.; Perincherry, A.; Chen, Y.; Li, H. Increasing SLAM Pose Accuracy by Ground-to-Satellite Image Registration. In Proceedings of the 2024 International Conference on Robotics and Automation (ICRA), Tokyo, Japan, 13–17 May 2024. [Google Scholar]
  28. Lepetit, V.; Moreno-Noguer, F.; Fua, P. EPnP: An Accurate O(n) Solution to the PnP Problem. Int. J. Comput. Vis. 2009, 81, 155–166. [Google Scholar] [CrossRef]
  29. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN Architecture for Weakly Supervised Place Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1437–1451. [Google Scholar] [PubMed]
  30. DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperPoint: Self-Supervised Interest Point Detection and Description. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  31. Sarlin, P.-E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperGlue: Learning Feature Matching with Graph Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  32. Horn, B.K.P. Closed-form solution of absolute orientation using unit quaternions. J. Opt. Soc. Am. 1987, 4, 629. [Google Scholar] [CrossRef]
Figure 1. Pipeline of the proposed localization method. Aerial image refers to real-time images captured by a UAV, while satellite images need to be pre-stored in the UAV’s onboard computer before the mission. The algorithm operates in real-time for tracking the movement of the UAV using aerial images and selects a subset of aerial images for matching with satellite images, facilitating geographic initialization and visual-geography BA.
Figure 1. Pipeline of the proposed localization method. Aerial image refers to real-time images captured by a UAV, while satellite images need to be pre-stored in the UAV’s onboard computer before the mission. The algorithm operates in real-time for tracking the movement of the UAV using aerial images and selects a subset of aerial images for matching with satellite images, facilitating geographic initialization and visual-geography BA.
Drones 08 00313 g001
Figure 2. Factor graph representation for different optimizations. The nodes in the factor graph correspond to optimization variables, including the UAV’s relative pose in the SLAM world coordinate system, its geographic pose, map points, and initialization parameters. Meanwhile, the edges represent constraints through reprojection residuals and geographic residuals. By optimizing these nodes using BA, we aim to minimize these residuals.
Figure 2. Factor graph representation for different optimizations. The nodes in the factor graph correspond to optimization variables, including the UAV’s relative pose in the SLAM world coordinate system, its geographic pose, map points, and initialization parameters. Meanwhile, the edges represent constraints through reprojection residuals and geographic residuals. By optimizing these nodes using BA, we aim to minimize these residuals.
Drones 08 00313 g002
Figure 3. Pipeline of the coarse-to-fine image-matching method.
Figure 3. Pipeline of the coarse-to-fine image-matching method.
Drones 08 00313 g003
Figure 4. The matching result of SuperPoint + SuperGlue. The green lines depict the extracted feature correspondences between the aerial and satellite images using SuperPoint + SuperGlue.
Figure 4. The matching result of SuperPoint + SuperGlue. The green lines depict the extracted feature correspondences between the aerial and satellite images using SuperPoint + SuperGlue.
Drones 08 00313 g004
Figure 5. Flowchart of satellite image extraction.
Figure 5. Flowchart of satellite image extraction.
Drones 08 00313 g005
Figure 6. Trajectories of the UAV in the simulation dataset.
Figure 6. Trajectories of the UAV in the simulation dataset.
Drones 08 00313 g006
Figure 7. Localization errors of various algorithms in the simulated dataset.
Figure 7. Localization errors of various algorithms in the simulated dataset.
Drones 08 00313 g007
Figure 8. (a) The trajectories of UAVs in the real-world dataset 1; (b) The trajectories of UAVs in the real-world dataset 2. Red lines represent the accurate trajectory obtained by RTK.
Figure 8. (a) The trajectories of UAVs in the real-world dataset 1; (b) The trajectories of UAVs in the real-world dataset 2. Red lines represent the accurate trajectory obtained by RTK.
Drones 08 00313 g008
Figure 9. (a) Localization errors of various algorithms in the real-world dataset 1; (b) Localization errors of various algorithms in the real-world dataset.
Figure 9. (a) Localization errors of various algorithms in the real-world dataset 1; (b) Localization errors of various algorithms in the real-world dataset.
Drones 08 00313 g009aDrones 08 00313 g009b
Table 1. Characteristics of the simulation datasets.
Table 1. Characteristics of the simulation datasets.
DatasetLength (km) Altitude (m)Speed (m/s)Duration (s)ResolutionFrame Rate (fps)
simulation29.150030960960 × 54020
Table 2. Characteristics of the real-world datasets.
Table 2. Characteristics of the real-world datasets.
DatasetLength (km) Altitude (m)Speed (m/s)Duration (s)ResolutionFrame Rate (fps)
17.75009.5808960 × 54020
27.850012.2640960 × 54020
Table 3. Ablation study on different configurations.
Table 3. Ablation study on different configurations.
DatasetMetricGeo-InitImage MatchingTrajectory FusionFull
SimulationMean (m)161.1214.3615.3610.32
RMSE (m)181.0414.9417.2111.11
Real-World 1Mean (m)241.1617.3619.3612.47
RMSE (m)275.7919.2121.3312.72
Real-World 2Mean (m)150.3716.3314.137.53
RMSE (m)161.9918.9616.528.31
Geo-init: VO with geographic initialization; Image matching: pose estimation of UAV via SuperPoint + SuperGlue; Trajectory fusion: optimizing only trajectory in map fusion; Full: results using all proposed modules.
Table 4. The mean and RMSE of various algorithms in the simulated dataset.
Table 4. The mean and RMSE of various algorithms in the simulated dataset.
MetricVOGeo-Initialized VOCoarse-to-Fine MethodProposed
Algorithm
Mean (m)254.58161.1220.5910.32
RMSE (m)282.35181.0425.4011.11
Table 5. The mean and RMSE of various algorithms in the real-world dataset.
Table 5. The mean and RMSE of various algorithms in the real-world dataset.
DatasetMetricVOGeo-Initialized VOCoarse-to-Fine MethodProposed Algorithm
1Mean (m)330.63241.1625.4512.47
RMSE (m)376.23275.7937.1112.72
2Mean (m)255.54150.3719.087.53
RMSE (m)293.14161.9926.478.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, W.; Yang, D.; Liu, J.; Li, Y.; Zhou, M. A Visual Navigation Algorithm for UAV Based on Visual-Geography Optimization. Drones 2024, 8, 313. https://doi.org/10.3390/drones8070313

AMA Style

Xu W, Yang D, Liu J, Li Y, Zhou M. A Visual Navigation Algorithm for UAV Based on Visual-Geography Optimization. Drones. 2024; 8(7):313. https://doi.org/10.3390/drones8070313

Chicago/Turabian Style

Xu, Weibo, Dongfang Yang, Jieyu Liu, Yongfei Li, and Maoan Zhou. 2024. "A Visual Navigation Algorithm for UAV Based on Visual-Geography Optimization" Drones 8, no. 7: 313. https://doi.org/10.3390/drones8070313

APA Style

Xu, W., Yang, D., Liu, J., Li, Y., & Zhou, M. (2024). A Visual Navigation Algorithm for UAV Based on Visual-Geography Optimization. Drones, 8(7), 313. https://doi.org/10.3390/drones8070313

Article Metrics

Back to TopTop