Next Article in Journal
Influence of Trabecular Bone Presence on Osseodensification Instrumentation: An In Vivo Study in Sheep
Next Article in Special Issue
A Pneumatic Soft Exoskeleton System Based on Segmented Composite Proprioceptive Bending Actuators for Hand Rehabilitation
Previous Article in Journal
Human–Robot Intimacy: Acceptance of Robots as Intimate Companions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Target Tracking: A Novel Grid-Based Beetle Antennae Search Algorithm and Confusion-Aware Detection

by
Yixuan Lu
,
Chencong Ma
and
Dechao Chen
*
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(9), 567; https://doi.org/10.3390/biomimetics9090567
Submission received: 13 August 2024 / Revised: 16 September 2024 / Accepted: 17 September 2024 / Published: 19 September 2024
(This article belongs to the Special Issue Optimal Design Approaches of Bioinspired Robots)

Abstract

:
Unmanned aerial vehicle target tracking is a complex task that encounters challenges in scenarios involving limited computing resources, real-time requirements, and target confusion. This research builds on previous work and addresses challenges by proposing a grid-based beetle antennae search algorithm and designing a lightweight multi-target detection and positioning method, which integrates interference-sensing mechanisms and depth information. First, the grid-based beetle antennae search algorithm’s rapid convergence advantage is combined with a secondary search and rollback mechanism, enhancing its search efficiency and ability to escape local extreme areas. Then, the You Only Look Once (version 8) model is employed for target detection, while corner detection, feature point extraction, and dictionary matching introduce a confusion-aware mechanism. This mechanism effectively distinguishes potentially confusing targets within the field of view, enhancing the system’s robustness. Finally, the depth-based localization of the target is performed. To verify the performance of the proposed approach, a series of experiments were conducted on the grid-based beetle antennae search algorithm. Comparisons with four mainstream intelligent search algorithms are provided, with the results showing that the grid-based beetle antennae search algorithm excels in the number of iterations to convergence, path length, and convergence speed. When the algorithm faces non-local extreme-value-area environments, the speed is increased by more than 89%. In comparison with previous work, the algorithm speed is increased by more than 233%. Performance tests on the confusion-aware mechanism by using a self-made interference dataset demonstrate the model’s high discriminative ability. The results also indicate that the model meets the real-time requirements.

1. Introduction

Unmanned aerial vehicle (UAV) autonomous tracking technology [1,2] is an innovative technology based on advanced sensors, intelligent algorithms, and autonomous decision-making capabilities. Its primary objective is to enable UAVs to independently identify, track, and pursue targets without human intervention, achieving efficient and accurate target tracking. Autonomous aerial tracking faces challenging tasks in fields such as search and rescue, pursuit, photography, and surveillance. Given the urgent need for UAV autonomous tracking solutions, the real-time tracking system design is considered a fundamental and crucial issue in UAV research. To achieve autonomous tracking, the UAV real-time tracking system relies on advanced sensors and high-performance computing platforms, utilizing deep learning and artificial intelligence algorithms to rapidly identify, locate, and predict the dynamic behavior of targets. It then plans a safe and reliable path within a short time frame. Moreover, to handle unpredictable situations, the system requires highly responsive, accurate, and intelligent path planning algorithms [3,4,5].
Some state-of-the-art aerial tracking planners [6,7,8] addressing the above issues have shown significant robustness and impressive agility. However, several limitations remain with these approaches. Firstly, the system’s high hardware demands may constrain its deployment on UAV platforms with limited computational capabilities. Secondly, the path search algorithm is relatively outdated and lacks efficiency compared to more advanced contemporary methods. Additionally, while the system offers a lightweight design, its target detection capabilities are limited to a narrow range, reducing its adaptability across broader application scenarios. Finally, the system demonstrates shortcomings in handling complex environments; for instance, the presence of visually similar objects within the field of view can result in tracking failures, ultimately diminishing the overall system performance.
In this study, two significant technical innovations are proposed: a confusion-aware multi-class object detection mechanism and a grid-based path planning algorithm utilizing the beetle antennae search (BAS) approach. For object detection, a lightweight framework based on machine learning and deep learning techniques is introduced. This framework effectively tracks multiple object categories in real time and incorporates a confusion-aware mechanism, enabling it to differentiate between similar-category targets by employing feature point matching and target position estimation. This significantly enhances robustness and real-time performance, particularly in scenarios with multiple similar objects within the field of view. For path planning, a grid-based algorithm is developed that leverages the rapid convergence properties of bionic algorithms. A distance-feedback-based step-size update strategy is employed to improve the efficiency of navigation. Additionally, a dynamically constructed grid index enables secondary searches, effectively reducing the overall path length. A fallback mechanism is also implemented to assist the system in escaping local minima, ensuring adaptive and efficient tracking in complex environments.
These methods are integrated into a comprehensive UAV tracking system developed in previous work. As illustrated in Figure 1, the system consists of the following key components: (1) multi-class object detection with a confusion-aware mechanism, (2) future motion prediction for targets without explicit motion intent, (3) a grid-based BAS algorithm for the tracking path search, (4) the generation of safe and dynamically feasible tracking trajectories, and (5) an onboard UAV system equipped with omnidirectional vision capabilities. The primary focus of this work is on improving components (1) and (3), as they directly influence the system’s tracking accuracy and path planning efficiency. The contributions of this study can be summarized as follows:
  • A grid-based path planning algorithm inspired by the BAS algorithm.
  • A secondary search mechanism based on a dynamic grid index and a distance-based step update strategy to improve the path quality and search speed.
  • A grid fallback mechanism to escape local extreme regions.
  • A confusion-aware multi-class object detection and localization framework.
The remainder of this paper is organized as follows. Section 2 provides a comprehensive literature review, summarizing relevant works on UAV autonomous tracking, path planning algorithms, and object detection methods, particularly focusing on those using confusion-aware techniques and bionic algorithms. Section 3 presents the improved grid-based beetle antennae search (GBAS) algorithm and its implementation. This section also discusses the design and integration of the confusion-aware mechanism and depth information for multi-class object tracking. Section 4 details the experimental setup and evaluation results, providing insights into the performance improvements achieved by our proposed methods. Finally, Section 5 concludes the paper and suggests potential directions for future research.

2. Related Work

2.1. Neural Networks in UAV Target Tracking

In recent years, the development of neural networks has had a profound impact on UAV autonomous tracking systems [9]. The rise of deep learning technology has led to significant advancements in image recognition [10,11] and target tracking [12,13] for UAVs. Through deep learning models like convolution neural networks (CNNs) [14,15], UAVs can efficiently and accurately identify targets such as humans, vehicles, and animals from the image data captured by sensors, laying a solid foundation for subsequent tracking tasks. Additionally, deep learning methods allow UAVs to learn target motion patterns and features. By employing algorithms based on long short-term memory networks (LSTMs) [16,17] or CNNs, the tracking robustness and reliability have been improved, effectively addressing issues like rapid target changes, occlusions, and complex backgrounds. Optimized network structures and algorithms have significantly increased the speed of target recognition and tracking, meeting the real-time decision-making requirements for UAVs in dynamic environments. Through end-to-end learning, neural networks enable UAVs to directly learn from perceptual data and make autonomous decisions, reducing reliance on human intervention and achieving a higher degree of autonomy. However, the application of neural networks also brings challenges, such as increased model complexity and computational requirements. Efficient computation in embedded systems is necessary to meet the real-time demands of UAV autonomous tracking systems. Moreover, acquiring and annotating a large amount of real-world data, especially in complex and changing environments, remains a challenging aspect of training neural network models for this application.
You Only Look Once (YOLO) [18] has established itself as one of the leading algorithms for real-time object detection due to its ability to balance speed and accuracy. Since its introduction, YOLO has been extensively studied and applied in various fields, including surveillance, autonomous driving, and robotics, where real-time performance is critical. Its capability to detect multiple objects in a single frame and its efficient computation make it a preferred choice for tasks requiring high-speed detection and classification. With the growing interest in technology cross-fusion, researchers have integrated YOLO with UAV systems, forming what is known as YOLO-based UAV technology (YBUT) [19]. YBUT leverages the fast object detection capabilities of YOLO and applies them to the aerial perspectives provided by UAVs. This combination has enabled more advanced UAV applications, such as autonomous navigation, surveillance, and search-and-rescue operations, where real-time decision-making based on object detection is crucial. YBUT has demonstrated excellent performance in dynamic target detection and classification, making it particularly suitable for UAV-based moving target tracking. Many studies have shown that YBUT is capable of identifying and tracking multiple moving targets in real time, even in complex environments where targets may be occluded or where multiple objects of similar appearance are present [20,21,22]. This adaptability and real-time processing capability make YBUT a popular choice for UAV applications, particularly in dynamic and challenging environments.

2.2. Path Planning in UAV Target Tracking

In the field of path planning, many studies have shown that efficiency and robustness are equally important in complex environments [23,24,25]. Various bio-inspired methods [26,27] have been developed to enhance planning performance, such as the artificial bee colony (ABC) algorithm based on bee foraging behavior [28,29]; the ant colony (ACO) algorithm, which utilizes a cooperative swarm search [30,31,32]; the genetic algorithm (GA) [33,34]; and the particle swarm optimization (PSO) algorithm, suitable for large-scale problems [35,36]. These algorithms have demonstrated remarkable robustness and impressive efficiency in different scenarios. However, they still suffer from certain inefficiencies. In [37], a spherical vector-based PSO (SPSO) algorithm was proposed, which used a PSO method based on ball support to transform the path planning problem into an optimization problem. Experimental results proved the excellent performance of SPSO. The IACO-IABC was proposed in [38], which combined the ACO and ABC algorithms and had impressive stability. The goal-distance-based rapidly exploring random tree* (GDRRT*) [39] improved the traditional RRT* algorithm by introducing an intelligent sampling mechanism for the target distance and had a faster convergence speed than RRT*. The above studies are novel attempts at path planning problems using heuristic algorithms.
Additionally, in [40], Qing Wu et al. proposed a path planning algorithm based on the beetle antennae search (BAS), considering requirements such as the path length, maximum turning angle, and obstacle avoidance. By ignoring optimization along the Z-axis and integrating the minimum threat surface (MTS), they applied the proposed path planning algorithm to UAV path planning. The results show that it significantly shortened the path length and maintained a high success rate in generating excellent paths, even in non-convex scenarios. In [41], Wang et al. combined swarm intelligence algorithms with a feedback-based step-size update strategy in the beetle swarm antenna search (BSAS) algorithm. They search in k directions at each iteration to find positions with smaller objective function values while evaluating the possibility of the beetle missing better positions of the objective function. By comparing with random numbers, they decide whether to update the step size. The results demonstrated that BSAS outperforms BAS, showing better performance. In [42], Khan et al. introduced a nature-inspired robust meta-heuristic algorithm. Their algorithm utilizes a novel beetle structure and optimization mechanism, assuming that the beetle has multiple antennae distributed with the same angular spread. Gradient estimation is represented as the difference between the average values of the objective function for two sets of non-overlapping antennae. They employ the adaptive moment estimation (ADAM) update rule to adaptively adjust the step size at each iteration. The results showed a significant improvement in convergence speed compared to the original BAS, with a more accurate gradient estimation model. All of the mentioned algorithms challenge the inherent limitations of the BAS algorithm, and each of them has valuable insights that can be learned and applied in further research.

2.3. UAV Target Tracking System

Unmanned aerial vehicles (UAVs) have seen significant advancements in autonomous tracking technology, thanks to the development of sophisticated path planning algorithms and advanced object detection mechanisms. The main challenge of UAV-based target tracking lies in the ability to efficiently plan paths while maintaining real-time performance, especially when dealing with complex environmental constraints and dynamic targets. Over the years, researchers have proposed various methods to optimize these tracking systems. This review highlights key developments and methodologies in UAV target tracking, focusing on path planning, object detection, and tracking algorithms.
One of the primary challenges in UAV target tracking is optimizing path planning algorithms to strike a balance between path length, computational complexity, and convergence speed. Efficient path planning must ensure that the UAV follows short paths while keeping computational overhead low. This is crucial for ensuring real-time responsiveness, especially when UAVs need to process large-scale environmental data. To address these challenges, Han [43] proposed an aerial tracking framework that uses polynomial regression to predict target motion based on past observations. The trajectories generated from this method are extrapolated to predict future movements of the target, allowing the UAV to quickly adjust its position. However, the reliance on AprilTag for target detection restricts its applicability, and the system’s inability to perform simultaneous target and environmental perception effectively with limited onboard vision highlights a key limitation. The challenge of maintaining real-time performance while addressing complex environmental factors remains a critical concern for UAV-based tracking systems.
With the rise of deep learning technologies, more advanced techniques for object detection and tracking have been introduced into UAV systems. Pan [7] developed a lightweight human-body detection method using deep learning and nonlinear regression to enhance the UAV’s ability to detect and localize targets. They further improved the UAV’s tracking module by incorporating an occlusion-aware mechanism, allowing the UAV to anticipate potential obstructions and plan accordingly. These improvements enabled a more robust and flexible tracking system. Despite these advances, the system faced challenges when attempting to balance the multiple constraints needed for effective path planning. Ensuring that trajectories meet criteria like safety, visibility, smoothness, robustness, and efficiency in real-time target tracking can be difficult, especially when these properties conflict with each other in certain scenarios. The inconsistency among these constraints necessitates adaptive capabilities within the constraint-solving algorithm to maintain high performance.
To address these multi-faceted challenges, ji [8] introduced the elastic tracker, which presents a more adaptive solution for UAV target tracking. The elastic tracker utilizes an occlusion-aware path search mechanism and introduces the concept of the intelligent safe flight corridor (SFC). By evaluating occlusion costs and developing a specialized mechanism to avoid obstructions, the system improves tracking accuracy, maintaining an appropriate observation distance between the UAV and its target. However, one of the key drawbacks of this system is its dependence on the A* algorithm [44,45] for the hierarchical multi-target path search. While the A* algorithm is reliable, it is relatively outdated compared to more modern algorithms, which could limit the system’s performance in dynamic and complex environments. Additionally, while the use of OpenPose [46] offers lightweight human body detection, its applicability to other entity categories remains limited, reducing the system’s flexibility.
While existing UAV tracking systems have made significant strides in addressing challenges like path planning and object detection, there is still room for further improvements. For instance, real-time UAV tracking systems must continue to optimize computational efficiency, especially in embedded systems with limited processing power. Balancing the demand for fast, real-time decision-making with the need for robust and accurate tracking of multiple dynamic targets under varying environmental conditions remains a challenge. Moreover, the integration of more sophisticated detection methods, such as multi-modal data fusion, could further enhance the adaptability and accuracy of UAV systems in tracking various types of targets. As UAV technology continues to evolve, future research should focus on improving algorithmic adaptability to diverse scenarios and enhancing the robustness of UAV systems in more extreme environmental conditions.
In summary, UAV target tracking systems have evolved significantly with advances in path planning and deep learning-based object detection. However, despite numerous proposed frameworks and methodologies, balancing real-time computational demands with robustness and adaptability in dynamic environments remains a challenge. While solutions such as polynomial regression-based tracking, occlusion-aware mechanisms, and elastic trackers offer promising results, further work is required to refine these systems for broader applications and to address their current limitations.

3. Methodology

The overall mission of this plan is to help the drone find a feasible path to reach the target and complete the tracking task. For the tracking scenarios, we make the following reasonable assumptions during the simulation phase:
(1)
It is assumed that the displacement of the target in the Z-axis direction is negligible during its movement.
(2)
The initial positions and velocities of the UAV and the targets are both assumed to be known and within a certain range.
In this study, it is assumed that the displacement of the target in the Z-axis direction is negligible during its movement. This assumption is grounded in typical UAV target tracking scenarios where targets, such as vehicles or pedestrians, predominantly move within a two-dimensional plane, making vertical movement minimal. Therefore, focusing on the X- and Y-axes simplifies the tracking process without significantly impacting the overall performance. However, for scenarios involving substantial Z-axis displacement, such as aerial or climbing targets, this assumption would need to be reconsidered. In future work, incorporating depth sensors or 3D tracking algorithms could address this limitation and improve tracking performance in such contexts.

3.1. Front-End Optimization

3.1.1. Preliminaries of Bionic Algorithm

In the BAS algorithm, the foraging behavior of beetles is modeled, and before each movement, a random unit vector b R n is used to represent the direction. Beetles rely on antennae on both sides of their bodies to sense the odor of food and choose the direction with the stronger food odor to move. They continue searching for food using this mechanism until they find it. In the BAS algorithm, the antennae are fixed on both sides of the beetle, and the coordinates of the left and right antennae of the beetle are represented as
x l = x t d t b , x r = x t + d t b ,
where the position of x l R n represents the location of the left antenna of the beetle, while x r R n represents the location of the right antenna of the beetle. In addition, d represents the distance from the antennae to the centroid, and  b R n is used to represent the direction. Specifically, d t R 1 represents the search range of the beetle, which decreases over time, as described in Equation (3). The value of d 0 ensures that d t does not decrease to zero. It is worth mentioning that the magnitude of d 0 depends on the scale of the specific problem. Once x l and x r are obtained, the algorithm uses a heuristic function f ( x ) to evaluate the quality of the two points and selects the smaller one as the candidate point. The foraging behavior of the beetle is described by Equation (2).
x t = x t 1 δ t b sign ( f ( x l ) f ( x r ) ) ,
where the symbol δ t represents the step size, which decreases over time, as indicated in Equation (4). The update rules for d and δ are described in Equations (3) and (4) as follows:
d t = η d d t 1 + d 0 ,
δ t = η δ δ t 1 + δ 0 ,
where the symbols η d and η δ represent the decay coefficients for the antennae length d and the step size δ , respectively. Jiang and Li [47] pointed out that the predefined values of parameters (e.g., d and δ ) have a significant impact on the performance of the BAS. A schematic diagram of the BAS algorithm is shown in Figure 2. After each movement, the beetle randomly changes its orientation, and its position is updated according to Equation (2). The red dashed lines with arrows indicate the direction of movement. It is evident that the length of the red dashed line representing the step size gradually decreases over time, following Equation (4). It is important to note that the process depicted in Figure 2 does not include any additional constraints.

3.1.2. Grid-Based Beetle Antennae Search

The BAS algorithm faces two inherent problems in its application to path planning. Firstly, due to the random orientation of the beetles, it cannot guarantee that the beetle’s movement will lead to a decrease in the objective function. This can result in generated paths that may not meet our expectations or even fail to converge. Therefore, the algorithm uses a quadratic search mechanism to enable the beetle to search for grids with lower costs more efficiently. The search step length S t can be defined as follows:
S t = δ t .
Before each beetle searches for the next position, we define a square with the current position of the beetle as the centroid and 2 S t as the side length. The grids on this square form the search area, which we represent using a set of mappings M ( · ) here. As shown in Equation (5), the search step length S t is determined by taking the ceiling of the step size δ t , representing the beetle’s search step length. The side length of a square is defined as twice the search step length 2 S t , with the beetle’s current position at the center of this square. This square is then overlaid on a grid map, ensuring that the center of the square coincides with the beetle’s current position. Consequently, the number of grid cells touched by this square is exactly 8 S t , which represents the search range for the beetle in the current iteration. The range of [ 1 , 8 S t ] ensures that the beetle searches within a well-defined area, facilitating effective exploration without excessive overlap and thus maintaining computational efficiency. We randomly select an integer ϵ from the range [ 1 , 8 S t ] and use ϵ as a random index instead of a random direction as in the BAS. The left and right antennae of the beetle are now represented as
n l t = M ( ϵ ) , n r t = M ( mod ( ϵ + 4 S t , 8 S t ) ) .
After the beetle relies on its left and right antennae to find a grid with a smaller objective function, it does not move immediately. Instead, it adjusts its orientation to face that grid. The value of ϵ is updated to the index of the grid toward which the beetle is oriented. The beetle then turns its left and right antennae forward, and the grid where the antennae are located is represented as
c l t = M ( mod ( ϵ S t , 8 S t ) ) , c r t = M ( mod ( ϵ + S t , 8 S t ) ) .
At this time, the beetle has three candidates: M ( ϵ ) , c l t , and  c r t . The beetle then removes unavailable nodes from the candidates. This includes Discarded nodes and Occluded nodes during algorithm iterations. Discarded nodes refer to the set of nodes where the beetle was before the rollback mechanism started. Occluded nodes refer to nodes that the detected beetle cannot reach in a straight line from its current location. The beetle uses this to escape the local extreme areas and ensure the safety of the search. Then, the node with the smallest cost function value is selected among the remaining candidate nodes as the destination. Therefore, the secondary search mechanism of the GBAS is shown in Figure 3. Under the conditions of search step size s t = 1 and no unavailable nodes, Figure 3a shows the search range of the beetle. Figure 3b shows the beetle’s first search. Figure 3c shows the beetle’s second search. Figure 3d shows the beetle moving to its destination.
Secondly, in the BAS, the beetle’s step length decreases over time to ensure the effective convergence of the algorithm. However, when the objective function shows good improvement, it is unnecessary to reduce the step length. Instead, we only need to reduce the step length when the beetle is about to reach its destination to ensure the convergence of the algorithm. Therefore, we have designed a step-length update strategy based on feedback distance, as shown below:
δ t ( D t ) = 1 1 + e k 1 ( r D t ) δ 0 , if D t > r , k 2 ( D t ) 3 ( r ) 3 + 1 2 δ 0 , if 0 < D t < r .
The variable D t represents the distance between the current position of the beetle and the target. The variable r represents the desired convergence distance. Both k 1 and k 2 are small constants. When the distance is greater than the desired distance, the step size decreases slowly. When the distance is less than the desired distance, the step size decreases rapidly, as shown in Figure 4. Based on the characteristics of the GBAS, the cost function is designed as follows:
f t ( n ) = ( d x y t ( n ) d d ) 2 + d z t ( n ) 2 + α b t O t .
In the given expression, d x y t represents the distance between node n and the target in the X-plane and Y-plane. In addition, d d is a constant representing the desired tracking distance, d z t represents the distance between node n and the target along the Z-axis, α is a constant coefficient, b t represents the vector pointing from the current position of the beetle to the target, and  O t represents the vector pointing from the current position of the beetle to node n. Therefore, we have designed a GBAS enhancement algorithm, as shown in Algorithm 1. In addition, the GBAS flowchart is shown in Figure 5; the parameter list of the GBAS is shown in Table 1.
Algorithm 1 GBAS algorithm for path planning.
Input: Given the target position z 0 and the current position of the UAV x 0 . Establish the cost function f t ( n ) . The variables n, δ 0 , δ t , S t , ϵ , M, D t , r, and  d d are initialized.
Output: The list of waypoints X = [ x 0 , , x n ] T .
  1: while  t < T max and D t > d d  do
  2:    Calculate δ t and S t according to Equations (8) and (5).
  3:    Update the mapping M.
  4:     ϵ = random ( 1 , 8 S t ) .
  5:    if  f t ( M ( ϵ ) ) > f t ( M ( mod ( ϵ + 4 S t , 8 S t ) ) )  then
  6:        ϵ = mod ( ϵ + 4 S t , 8 S t ) .
  7:    end if
  8:    Calculate c l t and c r t according to Equation (7).
  9:    Detect and remove Occluded or Discarded nodes.
10:    if No candidate nodes available then
11:      Update D t and x t to the previous state.
12:      Continue.
13:    end if
14:     x t = argmin { f t ( c l t ) , f t ( c r t ) , f t ( M ( ϵ ) ) } .
15:     X = [ X , x t ] .
16: end while
17: return X
The authors would like to clarify that the grid-based beetle antennae search (GBAS) algorithm, as a path search algorithm, is designed to generate safe and efficient paths without directly considering the aircraft’s dynamics. The role of the GBAS is to ensure that an optimal and feasible trajectory is planned.
However, the aircraft’s dynamic model, including the turn radius and other motion constraints, is addressed in a subsequent path optimization step. For path optimization, we followed previous work [8]. This step ensures that the generated path can be followed by the UAV in accordance with its physical limitations.

3.1.3. Time Complexity Analysis

Time complexity is a method of measuring the merits and complexity of the gradient algorithm and is also a tool to reflect the performance of the algorithm [48,49]. The main determinants of time complexity are the spatial dimensions of the model and the objective function.
The GBAS algorithm consists of a main loop that iterates until a maximum number of iterations T max is reached or the distance to the goal is less than a desired distance d d . The main loop of the GBAS algorithm mainly includes the step-size update, the generation of the search range M r , the calculation cost, and node elimination. The step-size update is completed in constant time, so its time complexity is O ( 1 ) . Generating the search range M r means generating 8 S t nodes around the current position, so its time complexity is O ( S t ) . The calculation cost is also completed in constant time, so its time complexity is also O ( 1 ) . Node elimination checks whether a node is out of bounds or occluded and involves traversing a set of nodes n occ and performing checks. The time complexity is O ( n occ ) , with n occ being a constant.
Therefore, assuming that the number of iterations used by the GBAS is t, the time complexity of the GBAS is depicted as follows:
T = O ( t ( 1 + S t + 1 + n occ ) ) = O ( t S t ) .
In general, the time complexity of the GBAS algorithm is mainly consumed in the search range generation.

3.2. Confusion-Aware Object Localization

3.2.1. Multi-Class Object Detection

Considering real-time performance and accuracy, we applied YOLOv8 [50]. The You Only Look Once version 8 (YOLOv8) is a significant technological advancement in the field of object detection. It integrates the strengths of the YOLO series and introduces further innovations. Its distinguishing feature lies in combining efficient real-time detection with accurate localization capabilities, achieving object detection in a single forward pass. This not only results in high speed but also maintains a high level of precision.
YOLOv8 achieves improved object classification and position prediction by incorporating a more powerful neural network architecture and enhanced training strategies. This enables it to better handle challenges such as detecting small-sized objects and densely arranged objects in complex scenes. Furthermore, YOLOv8 has been optimized in areas like data augmentation and model fusion, enhancing the model’s generalization capabilities.
For multi-class object detection, we tested various object detection networks, and the results are presented in Section 4.4.

3.2.2. Confusion-Aware Mechanism

In a real-world scenario, relying solely on target detection to obtain the position information for a UAV tracking system presents issues with insufficient robustness. In particular, when the UAV’s field of view includes similar target types, target detection may generate multiple target bounding boxes, posing challenges of perceptual confusion in the UAV tracking process. To overcome this issue, further measures can be taken to enhance the accuracy and reliability of tracking.
One strategy is to introduce target feature recognition. Assuming stable environmental lighting during UAV tracking, given that targets can change in scale and that real-time performance demands are high, we choose to implement target recognition using oriented FAST and rotated BRIEF (ORB) [51] feature points. After target detection, we extract ORB feature points from the targets and establish a feature point dictionary. Through feature point matching, we can identify the targets. This approach distinguishes targets in the field of view with similar targets by utilizing the uniqueness and stability of feature points, thus enhancing tracking accuracy.
Another strategy is to combine target tracking algorithms with the initial position information provided by target detection. During tracking, we can use the initial position information provided by the target detection algorithm as the starting point for tracking and then employ motion estimation techniques to continuously track the targets in subsequent frames. This strategy yields continuous target trajectories, overcoming potential issues from multiple bounding boxes introduced by target detection and enhancing the robustness and accuracy of tracking.
To address the challenge of perceptual confusion due to the presence of similar targets within the UAV’s field of view, we adopt the approach of ORB feature point extraction and dictionary matching to correctly identify and track the targets. The ORB algorithm possesses characteristics such as rotational invariance, scale invariance, and high computational speed, making it well suited for real-time visual systems.
In this mechanism, we first employ the target detection algorithm to detect the objects within the input image, obtaining their positions and bounding box information. Then, for each bounding box region, we extract the ORB feature point set KP k and generate a descriptor set D k :
KP k = { kp 1 k , kp 2 k , , kp n k } ,
D k = { des 1 k , des 2 k , , des n k } .
In the above expression, kp n k represents the pixel coordinate of the n-th feature point in the k-th bounding box. des n k represents the descriptor of the n-th feature point in the k-th bounding box. ORB uses the BRIEF algorithm to calculate the descriptor of a feature point. The core idea of the BRIEF algorithm is to select multiple point pairs in a certain pattern around key points and combine the comparison results of these multiple point pairs as descriptors. We gradually construct these trusted descriptors of the tracking target bounding box into a descriptor dictionary Dic as a reference for the target features:
Dic = Dic { des i r e f } , if des i r e f Dic ,
where des i r e f represents the descriptor of the i-th feature point in the trusted bounding box.
In terms of finding trusted bounding boxes, in subsequent image frames, we also perform target detection and bounding box extraction. Then, we extract the ORB feature points and descriptors within the bounding box regions. We match the descriptors corresponding to the feature points within the bounding boxes of the subsequent image frames with the previously constructed descriptor dictionary Dic. Specifically, for the descriptor des i k corresponding to the i-th feature point in the k-th bounding box b k t in the t-th frame image, if there is a descriptor des j Dic in the dictionary Dic, and this des i k is calculated through the similarity matching function and the result is greater than the confidence CS, the match is successful. The descriptor des i k is considered trusted. This increases the feature confidence N k t of the current bounding box:
N k t = i = 1 m 1 , if j : sim ( des i k , des j Dic ) > C S , 0 , otherwise .
where m is the total number of detected bounding boxes in the image. The confusion-aware mechanism uses this method to calculate the score of each bounding box. All calculated bounding box scores are combined into a set N G t :
NG t = { N 1 t , N 2 t , , N k t } ,
In this case, the confusion-aware module selects the highest-scoring bounding box as the final output bounding box B t of the t-th frame image.
B t = arg max k N G t .
This feature matching algorithm accomplishes target identification and tracking. Poorly matched bounding boxes are considered invalid. To enhance robustness, in situations where multiple targets exist in the field of view, target positioning information is compared to predicted positions, and targets with large positional discrepancies are marked as invalid. This perceptual confusion mechanism offers more stable and accurate target tracking capabilities, particularly suitable for scenarios with substantial variations in target scale and appearance.
By integrating ORB feature point extraction, dictionary construction, matching, and position information comparison, we are able to accurately identify and track targets, mitigating confusion scenarios. This mechanism provides a reliable target tracking solution for real-time UAV visual systems. The flowchart of the confusion-aware mechanism for distinguishing tracked targets is shown in Figure 6.

3.2.3. Depth-Based Object Localization

Regarding object localization, we employ a common depth-based localization method [52]. Firstly, considering the influence of noise and outliers on depth estimation, we use the average depth value S t of multiple feature points within the bounding box obtained from object detection to represent the depth of the target. The depth information is obtained from a depth camera. Secondly, we convert the target pixel coordinates ( u t , v t ) T to normalized coordinates ( x t , y t , 1 ) T , as depicted below:
x t = u t c x f x , y t = v t c y f y .
The coordinates ( c x , c y ) T represent the origin coordinates, which correspond to the position of the camera’s optical center on the image plane. The values ( f x , f y ) T represent the focal lengths. To obtain the coordinates ( x c t , y c t , z c t ) T of the target in the camera coordinate system, it can multiply the normalized coordinates by the depth value S t , as shown in the following equation:
x c t = x t S t , y c t = y t S t , z c t = S t .
From this, the UAV system obtains the coordinates in the coordinate system of the target camera. Therefore, the workflow of confusion-aware object localization is as shown in Figure 7. Figure 7a shows the output bounding box of object detection. In addition, Figure 7b shows the ORB feature points extracted in the confusion-aware mechanism. Figure 7c shows the calculation of the average depth. Figure 7d shows the target location information obtained by the drone.

4. Results and Discussion

4.1. Parameter-Sensitive Testing

This section presents a performance test of the GBAS algorithm and the chaos perception module. In addition, it also compares the performance of some current mainstream target detectors to optimize the target detector for the tracking system. All experiments were conducted on a computer configured with an Intel i5-12400 CPU, an NVIDIA RTX 3060 12 GB GPU, and 64 GB RAM.
In this section, we validate the performance and convergence of the GBAS in both the Matlab and ROS environments to verify the algorithm’s effectiveness and its application in elastic trackers. In Matlab, we conducted extensive tests on cost using different parameter settings, such as δ 0 , r, k 1 , and k 2 , and starting from various positions. Generally, on a 50 × 50 grid map, the maximum iteration count for the GBAS algorithm was set as T max = 100 , and the desired distance as d d = 1 . The predefined parameters for the feedback-based step-size update strategy were set as δ 0 = 3 , r = 2 , k 1 = 3 , and k 2 = 0.15 . For clarity, in subsequent discussions, we refer to this set of data as the control group.
To investigate the influence of predefined parameters on the quality of paths and convergence, we set up comparison experiments with multiple groups of different parameters. We modified the k 2 parameter for the comparison group, ensuring that the size of the root does not exceed the convergence radius for efficient path convergence. The predefined experimental parameters for the comparison group were δ 0 = 3 , r = 2 , k 1 = 3 , and k 2 = 0.005 , referred to as the experimental group. The comparison experiments are shown in Figure 8. In Figure 8a, the path of the control group is displayed in an open environment. In Figure 8b, the path of the experimental group in an open environment is shown. Figure 8c presents the variation in cost functions for the experimental and control groups with iterations, where blue represents the control group, and red represents the experimental group. Figure 8d illustrates the relationship between the step size δ t and the distance d t to the target for both the experimental and control groups, with blue indicating the control group and red indicating the experimental group. From the experiments, we found that k 2 primarily affects the size of the root in the step-size update strategy, with smaller k 2 leading to larger roots. In practical applications, k 2 should not be too small to ensure quick path convergence. However, setting it too large should be avoided to prevent the agent from losing its ability to act right after entering the convergence area r. We also tested parameters like k 1 , r, and δ 0 , and the results indicated that their values all require empirical guidance to align with practical applications.

4.2. Number Function Evaluation

The number function evaluation for the optimization algorithm was carried out using the CEC test functions, which incorporate experiments conducted across dimensions of 10, 30, 50, and 100. The CEC test suite consists of 30 truly challenging benchmark problems, including 1–3 single-modal, 4–10 multi-modal, 11–20 hybrid, and 21–30 composite features. Note that the results of F2 are usually abandoned in the CEC test. This dataset is a recent and highly complex one and includes all major types of optimization problems. A general discussion on the definition of these benchmark problems can be found in [53].
In order to adapt the proposed GBAS algorithm for CEC testing, several mechanisms are optimized within the algorithm environment. This includes the integration of a quadratic search and step-size update mechanism. Specifically, the proposed GBAS algorithm performs a quadratic search in each dimension after initial random searches, focusing subsequent efforts on regions with lower fitness values. The step-size function is designed to decrease with experimental error reduction, as defined below:
δ t = δ i δ i C 1 1 + exp k ( e c m ) ,
where δ t denotes the step size of the algorithm, δ i represents the initial step size, C is a constant adjusting the lower limit of the step-size function, k reflects the gradient when the step size decreases, m adjusts the influence of error, and e c represents the current position’s error in the algorithm.
The parameter settings are based on the requirements of the CEC and the characteristics of the GBAS optimization algorithm. The parameter settings for the algorithm are shown in Table 2. Table 3 presents the experimental results of the proposed GBAS algorithm on the CEC test functions of varying dimensions.
The above refinements enhance the adaptability and performance of the GBAS across the diverse settings of the CEC test functions, as demonstrated in the experimental outcomes. Judging from the experimental results, the GBAS algorithm performs better when dealing with low-dimensional problems, but as the dimension increases, both the error and standard deviation may increase accordingly, especially on specific functions.

4.3. Performance Comparisons and Statistical Tests

In order to test the performance of the proposed GBAS algorithm, different environmental models were designed in the experiment, as shown in Table 4. On the map, black represents obstacles. The experimental platform is MATLAB. The GBAS algorithm is also compared with other intelligent path planning algorithms. The parameter settings of the GBAS algorithm are as follows: δ 0 = 3 , k 1 = 3 , k 2 = 0.15 , and r = 5 . In order to ensure fairness, the general parameters of the GBAS are the same as those of the other compared algorithms; the maximum number of iterations is set to 100. The population number of the population-based methods is consistent, including PSO and IACO-IABC, and their population number is set to S N = 20 . Methods with range search mechanisms include the GBAS and GDRRT*. Their initial step sizes are the same. The initial step size is set to δ 0 = 3 . The special parameters of ACO in the IACO-IABC algorithm are the same as those of the traditional ACO, and the parameters of the ABC algorithm are the same as those of the traditional ABC algorithm. For example, the parameters of ACO include number of ants M = 10 , pheromone importance α = 1 , heuristic importance β = 7 , pheromone evaporation rate ρ = 0.3 , and pheromone intensification constant Q = 1 . The parameters of ABC include number of bees N = 10 , and the maximum number of trials is set to 10. The special parameters of GDRRT* are neighborhood radius r n = 5 , search spherical area diameter d = 5 , random minimum size d min = 2 , and minimum obstacle distance O d = 1 . The special parameters of SPSO are inertia weight w = 1 , cognitive coefficient η 1 = 1.5 , social coefficient η 2 = 1.5 , and number of intermediate points n i = 50 . Then, each algorithm was run 30 times on each map for path planning. The optimal paths obtained by each compared algorithm are shown in Figure 9, Figure 10 and Figure 11, and the experimental results are shown in Table 5. In addition, the speedup depicted in T imp represents the average running time difference compared to other algorithms, expressed as a percentage, which is positive if there is an improvement:
T imp = ( T com T pro ) / T pro · 100 % ,
where T com represents the running time of the other algorithms, and T pro represents the running time of the proposed GBAS algorithm.
As shown in Figure 9, Figure 10 and Figure 11 and Table 5, the path length of the GBAS algorithm and the path lengths of the other four algorithms were tested through the Wilcoxon rank-sum test. It is shown that, except for SPSO, the path length of the proposed GBAS algorithm and other comparative algorithms are better in different environments. All differences are significant. In addition, the average path length of the GBAS algorithm is shorter than those of the other algorithms except for A*, and the standard deviation is also within an acceptable range. In addition, the running time of the GBAS is the shortest among all environments, and the standard deviation is also better, which shows the good real-time performance of the GBAS. The Wilcoxon rank-sum test shows that there is a significant difference in running time between the GBAS algorithm and the other algorithms in different environments. In Map 1, the running time of the algorithm is shortened by more than 100%, reaching a maximum of 495%. In Map 2’s environment with a large number of discrete obstacles and Map 3 with concave obstacles, the performance is not as good as in Map 1, but what is interesting is that the GBAS algorithm performs well in both Map 2 and Map 3. This is due to the quadratic search mechanism and the distance-based step-size adjustment strategy, which allow the proposed GBAS algorithm to converge quickly.
In order to verify the effectiveness of the improvement, on the ROS platform, we integrated the GBAS algorithm into the elastic tracker and used the benchmark method in Ji et al. [8] to conduct comparative testing in a simulated environment. The experimental map contained a discrete obstacle, the target’s speed was set to 1 m/s and 2 m/s, and the same series of target points were set to drive the tracked target to move. There were 40 target points in total, distributed in different locations on the map. The new system was tested against the elastic tracker as the benchmark. The comparative experimental results are shown in Table 6.
Under the condition of target v max = 1 m/s, the average search time of the GBAS in the path search phase of the two tracking systems is 0.18 ms, which is 233% faster than that of A*, which is 0.96 ms. Under the condition of target v max = 2 m/s, the average search time of the GBAS in the path search phase of the two tracking systems is 0.22 ms, which is 555% faster than that of A*, which is 1.41 ms.
The initial positions and velocities of the UAV and targets are crucial for trajectory planning, as they determine the starting conditions for tracking. Defining appropriate ranges for these parameters ensures that the UAV can effectively initiate and maintain tracking. In this study, these ranges were used to simulate various scenarios, influencing the UAV’s response time and the complexity of path planning.
Higher velocities require faster decision-making from the UAV to maintain smooth tracking and avoid overshooting. At high speeds, even minor errors can significantly impact the tracking accuracy. Therefore, velocity was carefully evaluated in our experiments.
In the experiments, the target’s maximum speeds were set to 1 m/s and 2 m/s to test the UAV’s ability to track targets at different speeds, providing a benchmark for its performance in varied scenarios.

4.4. Object Detection Tests

Regarding the object detection dataset, we chose the Det-Fly dataset [54] as the foundation for our training. This dataset is renowned for its systematic and comprehensive nature, encompassing various factors, such as diverse background scenarios, different shooting perspectives, variations in object distances, flight altitudes, and lighting conditions. This makes it an ideal dataset, particularly suitable for training object detection models focused on identifying drones. We selected multiple object detection networks from the Det-Fly dataset for training, including some popular real-time object detectors and end-to-end object detection models. To assess the performance of these models, we constructed a test dataset, as shown in Figure 12, comprising 3194 images, all of which were captured in real-world drone tracking scenarios.
The test results on real-time performance and detection accuracy are presented in Table 7 and Table 8. Table 7 provides insight into real-time performance, showcasing YOLOv8’s ability to balance high FPS and a manageable parameter size. It highlights that the proposed model achieves higher FPS (62 FPS) while keeping parameters (43 M) relatively low compared to similar models, demonstrating efficiency in computation. Table 8 further compares the detection performance and computational cost of different models, where the FLOPs (floating point operations per second) serve as an indicator of computational complexity. Although YOLOv8 (ours) has slightly higher GFLOPs (164 GFLOPs) than other YOLO versions, it compensates with better precision, recall, and F1-score values (85.2%, 92.7%, and 0.89, respectively), indicating that the additional computational complexity leads to superior detection accuracy. Therefore, while YOLOv8’s computational complexity is slightly higher, the trade-off in terms of improved real-time performance and enhanced detection capabilities justifies its inclusion. Among various tested networks, YOLOv8 developed in our model exhibited exceptional performance, making it the chosen object detector for the UAV system.
In addition to object detection, we conducted a series of tests on the confusion-aware module. To maintain real-time performance, we limited the number of feature points extracted per image to a maximum of 50 and set the confidence threshold for descriptors to 50. For benchmark testing, we selected drone images with different shapes and colors as experimental subjects, ensuring clear distinctions between drones. Furthermore, we selected drone images with the same shape but different colors, as well as drone images with different shapes but similar colors for experimentation. Ultimately, we also tested drones of the same model, with all test results outlined in Figure 13.
The experimental results indicated that the confusion-aware module achieved a high success rate when dealing with significant differences in color or shape features. Even in cases where the differences in color and shape features were less pronounced, the module maintained a reasonable success rate. Apart from success rate testing, we also evaluated the performance of the confusion-awareness. It is important to note that when the object detector failed to detect additional objects, the confusion-aware module remained in a dormant state. During continuous operation, the average processing time of the confusion-aware module was 40 milliseconds. The average computation time for subsequent target localization remained below 6 milliseconds, which is sufficient to meet real-time requirements.

4.5. Managerial Implications and Applications

The UAV target tracking system developed in this work holds significant managerial implications, particularly in sectors where real-time monitoring, decision-making, and automation are crucial. For industries like logistics, surveillance, and agriculture, efficient target tracking can reduce human intervention, improve operational safety, and enhance productivity. For example, in logistics, UAVs equipped with advanced tracking systems can autonomously follow delivery routes, detect obstacles, and ensure the timely delivery of packages. Similarly, in surveillance, the system enables UAVs to continuously monitor large areas with minimal human oversight, providing real-time information critical for security operations. These improvements in automation can lead to reduced labor costs and faster response times, thereby increasing overall operational efficiency.
Furthermore, this system’s adaptability to different environmental conditions and its robustness in tracking multiple targets simultaneously make it ideal for dynamic environments where conditions can change rapidly. This capability is particularly beneficial for industries like search-and-rescue operations or agricultural monitoring, where UAVs need to swiftly identify and track moving objects (e.g., humans in distress, wildlife, or livestock) and provide critical data to operators in real time. The integration of a confusion-aware mechanism in the tracking system ensures that UAVs can distinguish between similar targets, thus avoiding errors that could lead to operational inefficiencies or missed objectives.
Beyond UAV tracking, the technologies and algorithms proposed in this work are applicable to other domains where real-time object detection and decision-making are essential. For instance, in autonomous driving, vehicles equipped with similar tracking algorithms could improve their ability to detect and respond to pedestrians or other vehicles in complex urban environments. Similarly, in smart surveillance systems, such algorithms could enhance real-time threat detection, providing quicker alerts and responses in situations requiring immediate attention. Additionally, the framework’s adaptability to handle dynamic, multi-target tracking scenarios positions it as a valuable tool for industrial automation, where it could be applied to track machinery or goods on factory floors, improving workflow efficiency and safety.

5. Conclusions

In this paper, the elastic tracker has been enhanced by integrating a novel GBAS algorithm, which improves path efficiency and convergence speed. The addition of a backtracking mechanism further allows the algorithm to avoid local minima, leading to more optimal path planning. The proposed lightweight multi-object detection and localization method, which incorporates a confusion-aware mechanism and depth information, can efficiently handle multiple targets with minimal cost. In particular, the proposed GBAS algorithm has successfully enhanced the elastic tracker by improving path efficiency and convergence speed, as demonstrated through extensive experimental results and statistical tests. The inclusion of the backtracking mechanism also allows the algorithm to avoid local minima, fulfilling the objective of optimizing path planning in complex environments. Furthermore, the confusion-aware multi-object detection and localization method meets the objective of efficiently handling multiple targets with minimal computational cost while still maintaining real-time performance. These contributions, validated through comparative experiments, address both theoretical and practical aspects of UAV target tracking. These advancements contribute significantly to the theory and practice of UAV target tracking, enhancing both performance and reliability. Future work could focus on the compression approach of the algorithm and model and the realization of a lightweight embedded UAV tracking and detection system.

Author Contributions

Y.L.: Methodology Design, Experiment, Writing—original draft, Review and editing. C.M.: Investigation, Data collection, Review and editing. D.C.: Analysis, Writing—review and editing, Funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article. The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China under Grant 62276085 and Grant 61906054 and in part by the Zhejiang Provincial Natural Science Foundation of China under Grant LR24F030002 and Grant LY21F030006.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, P.; Zhou, Y. The Review of Target Tracking for UAV. In Proceedings of the 2019 14th IEEE Conference on Industrial Electronics and Applications (ICIEA), Xi’an, China, 19–21 June 2019; pp. 1800–1805. [Google Scholar] [CrossRef]
  2. Han, Y.; Liu, H.; Wang, Y.; Liu, C. A Comprehensive Review for Typical Applications Based Upon Unmanned Aerial Vehicle Platform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9654–9666. [Google Scholar] [CrossRef]
  3. Jurn, Y.N.; Mahmood, S.A.; Aldhaibani, J.A. Anti-Drone System Based Different Technologies: Architecture, Threats and Challenges. In Proceedings of the 2021 11th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 27–28 August 2021; pp. 114–119. [Google Scholar] [CrossRef]
  4. Patle, B.; Babu L, G.; Pandey, A.; Parhi, D.; Jagadeesh, A. A review: On path planning strategies for navigation of mobile robot. Def. Technol. 2019, 15, 582–606. [Google Scholar] [CrossRef]
  5. Yang, L.; Qi, J.; Song, D.; Xiao, J.; Han, J.; Xia, Y. Survey of Robot 3D Path Planning Algorithms. J. Control Sci. Eng. 2016, 2016, 7426913. [Google Scholar] [CrossRef]
  6. Jeon, B.; Lee, Y.; Kim, H.J. Integrated Motion Planner for Real-time Aerial Videography with a Drone in a Dense Environment. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 1243–1249. [Google Scholar] [CrossRef]
  7. Pan, N.; Zhang, R.; Yang, T.; Cui, C.; Xu, C.; Gao, F. Fast-Tracker 2.0: Improving autonomy of aerial tracking with active vision and human location regression. IET Cyber-Syst. Robot. 2021, 3, 292–301. [Google Scholar] [CrossRef]
  8. Ji, J.; Pan, N.; Xu, C.; Gao, F. Elastic Tracker: A Spatio-temporal Trajectory Planner for Flexible Aerial Tracking. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 47–53. [Google Scholar] [CrossRef]
  9. Chen, D.; Li, S.; Liao, L. A recurrent neural network applied to optimal motion control of mobile robots with physical constraints. Appl. Soft Comput. 2019, 85, 105880. [Google Scholar] [CrossRef]
  10. Li, C.; Li, H.; Gao, G.; Liu, Z.; Liu, P. An accelerating convolutional neural networks via a 2D entropy based-adaptive filter search method for image recognition. Appl. Soft Comput. 2023, 142, 110326. [Google Scholar] [CrossRef]
  11. Wu, Z.Z.; Wan, S.H.; Wang, X.F.; Tan, M.; Zou, L.; Li, X.L.; Chen, Y. A benchmark data set for aircraft type recognition from remote sensing images. Appl. Soft Comput. 2020, 89, 106132. [Google Scholar] [CrossRef]
  12. Wang, J.; Zhang, P.; Wang, Y. Autonomous target tracking of multi-UAV: A two-stage deep reinforcement learning approach with expert experience. Appl. Soft Comput. 2023, 145, 110604. [Google Scholar] [CrossRef]
  13. Hu, C.; Qu, G.; Zhang, Y. Pigeon-inspired fuzzy multi-objective task allocation of unmanned aerial vehicles for multi-target tracking. Appl. Soft Comput. 2022, 126, 109310. [Google Scholar] [CrossRef]
  14. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  15. Krichen, M. Convolutional Neural Networks: A Survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  16. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  17. Smagulova, K.; James, A.P. A survey on LSTM memristive neural network architectures and applications. Eur. Phys. J. Spec. Top. 2019, 228, 2313–2324. [Google Scholar] [CrossRef]
  18. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo Algorithm Developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  19. Chen, C.; Zheng, Z.; Xu, T.; Guo, S.; Feng, S.; Yao, W.; Lan, Y. YOLO-Based UAV Technology: A Review of the Research and Its Applications. Drones 2023, 7, 190. [Google Scholar] [CrossRef]
  20. Çintaş, E.; Özyer, B.; Şimşek, E. Vision-Based Moving UAV Tracking by Another UAV on Low-Cost Hardware and a New Ground Control Station. IEEE Access 2020, 8, 194601–194611. [Google Scholar] [CrossRef]
  21. Phung, K.P.; Lu, T.H.; Nguyen, T.T.; Le, N.L.; Nguyen, H.H.; Hoang, V.P. Multi-model Deep Learning Drone Detection and Tracking in Complex Background Conditions. In Proceedings of the 2021 International Conference on Advanced Technologies for Communications (ATC), Ho Chi Minh City, Vietnam, 14–16 October 2021; pp. 189–194. [Google Scholar] [CrossRef]
  22. Hong, T.; Liang, H.; Yang, Q.; Fang, L.; Kadoch, M.; Cheriet, M. A Real-Time Tracking Algorithm for Multi-Target UAV Based on Deep Learning. Remote Sens. 2023, 15, 2. [Google Scholar] [CrossRef]
  23. Meng, L.; Qing, S.; Qinjun, Z.; Yongliang, Z. Route planning for unmanned aerial vehicle based on rolling RRT in unknown environment. In Proceedings of the 2016 IEEE International Conference on Computational Intelligence and Computing Research (ICCIC), Chennai, India, 15–17 December 2016; pp. 1–4. [Google Scholar] [CrossRef]
  24. Lei, L.; Kun, Z.; Dewei, W.; Kun, H.; Hailin, L.; Qiurong, Z. A Method of Hybrid Intelligence for UAV Route Planning Based on Membrane System. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 1317–1320. [Google Scholar] [CrossRef]
  25. Huang, Y.; Wang, Y.; Zatarain, O. Dynamic Path Optimization for Robot Route PlanningDynamic Path Optimization for Robot Route Planning. In Proceedings of the 2019 IEEE 18th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC), Milan, Italy, 23–25 July 2019; pp. 47–53. [Google Scholar] [CrossRef]
  26. Ma, L.; Zhu, Y.; Liu, Y.; Tian, L.; Chen, H. A novel bionic algorithm inspired by plant root foraging behaviors. Appl. Soft Comput. 2015, 37, 95–113. [Google Scholar] [CrossRef]
  27. Liu, J.; Chen, Y.; Liu, X.; Zuo, F.; Zhou, H. An efficient manta ray foraging optimization algorithm with individual information interaction and fractional derivative mutation for solving complex function extremum and engineering design problems. Appl. Soft Comput. 2024, 150, 111042. [Google Scholar] [CrossRef]
  28. Karaboga, D.; Basturk, B. On the performance of artificial bee colony (ABC) algorithm. Appl. Soft Comput. 2008, 8, 687–697. [Google Scholar] [CrossRef]
  29. Sharma, A.; Sharma, A.; Choudhary, S.; Pachauri, R.; Shrivastava, A.; Kumar, D. A review on artificial bee colony and it’s engineering applications. J. Crit. Rev. 2020, 7, 2020. [Google Scholar]
  30. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  31. Dorigo, M.; Stützle, T. Ant Colony Optimization: Overview and Recent Advances; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  32. Wu, H.; Gao, Y. An ant colony optimization based on local search for the vehicle routing problem with simultaneous pickup–delivery and time window. Appl. Soft Comput. 2023, 139, 110203. [Google Scholar] [CrossRef]
  33. Mirjalili, S. Evolutionary algorithms and neural networks. In Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2019; Volume 780, pp. 43–53. [Google Scholar] [CrossRef]
  34. Lambora, A.; Gupta, K.; Chopra, K. Genetic Algorithm- A Literature Review. In Proceedings of the 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon), Faridabad, India, 14–16 February 2019; pp. 380–384. [Google Scholar] [CrossRef]
  35. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  36. Gad, A.G. Particle swarm optimization algorithm and its applications: A systematic review. Arch. Comput. Methods Eng. 2022, 29, 2531–2561. [Google Scholar] [CrossRef]
  37. Phung, M.D.; Ha, Q.P. Safety-enhanced UAV path planning with spherical vector-based particle swarm optimization. Appl. Soft Comput. 2021, 107, 107376. [Google Scholar] [CrossRef]
  38. Li, G.; Liu, C.; Wu, L.; Xiao, W. A mixing algorithm of ACO and ABC for solving path planning of mobile robot. Appl. Soft Comput. 2023, 148, 110868. [Google Scholar] [CrossRef]
  39. Aslan, M.F.; Durdu, A.; Sabanci, K. Goal distance-based UAV path planning approach, path optimization and learning-based path estimation: GDRRT*, PSO-GDRRT* and BiLSTM-PSO-GDRRT*. Appl. Soft Comput. 2023, 137, 110156. [Google Scholar] [CrossRef]
  40. Wu, Q.; Shen, X.; Jin, Y.; Chen, Z.; Li, S.; Khan, A.H.; Chen, D. Intelligent Beetle Antennae Search for UAV Sensing and Avoidance of Obstacles. Sensors 2019, 19, 1758. [Google Scholar] [CrossRef]
  41. Wang, J.; Chen, H. BSAS: Beetle Swarm Antennae Search Algorithm for Optimization Problems. arXiv 2018, arXiv:1807.10470. [Google Scholar] [CrossRef]
  42. Khan, A.H.; Cao, X.; Li, S.; Katsikis, V.N.; Liao, L. BAS-ADAM: An ADAM based approach to improve the performance of beetle antennae search optimizer. IEEE/CAA J. Autom. Sin. 2020, 7, 461–471. [Google Scholar] [CrossRef]
  43. Han, Z.; Zhang, R.; Pan, N.; Xu, C.; Gao, F. Fast-Tracker: A Robust Aerial System for Tracking Agile Target in Cluttered Environments. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 328–334. [Google Scholar] [CrossRef]
  44. Hart, P.E.; Nilsson, N.J.; Raphael, B. A Formal Basis for the Heuristic Determination of Minimum Cost Paths. IEEE Trans. Syst. Sci. Cybern. 1968, 4, 100–107. [Google Scholar] [CrossRef]
  45. Foead, D.; Ghifari, A.; Kusuma, M.B.; Hanafiah, N.; Gunawan, E. A Systematic Literature Review of A* Pathfinding. Procedia Comput. Sci. 2021, 179, 507–514. [Google Scholar] [CrossRef]
  46. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar] [CrossRef]
  47. Jiang, X.; Li, S. BAS: Beetle Antennae Search Algorithm for Optimization Problems. arXiv 2017, arXiv:1710.10724. [Google Scholar] [CrossRef]
  48. Liu, C.; Wu, L.; Xiao, W.; Li, G.; Xu, D.; Guo, J.; Li, W. An improved heuristic mechanism ant colony optimization algorithm for solving path planning. Knowl.-Based Syst. 2023, 271, 110540. [Google Scholar] [CrossRef]
  49. Liu, C.; Wu, L.; Huang, X.; Xiao, W. Improved dynamic adaptive ant colony optimization algorithm to solve pipe routing design. Knowl.-Based Syst. 2022, 237, 107846. [Google Scholar] [CrossRef]
  50. Jocher, G. YOLOv8. 2023. Available online: https://github.com/ultralytics/ultralytics/tree/main (accessed on 16 September 2024).
  51. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar] [CrossRef]
  52. Huang, A.S.; Bachrach, A.; Henry, P.; Krainin, M.; Maturana, D.; Fox, D.; Roy, N. Visual Odometry and Mapping for Autonomous Flight Using an RGB-D Camera. In Robotics Research: The 15th International Symposium ISRR; Springer International Publishing: Cham, Switzerland, 2017; pp. 235–252. [Google Scholar] [CrossRef]
  53. Wu, G.; Mallipeddi, R.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real-Parameter Optimization; Technical Report; National University of Defense Technology: Changsha, China; Kyungpook National University: Daegu, Republic of Korea; Nanyang Technological University: Singapore, 2017. [Google Scholar]
  54. Zheng, Y.; Chen, Z.; Lv, D.; Li, Z.; Lan, Z.; Zhao, S. Air-to-Air Visual Detection of Micro-UAVs: An Experimental Evaluation of Deep Learning. IEEE Robot. Autom. Lett. 2021, 6, 1020–1027. [Google Scholar] [CrossRef]
  55. Jocher, G. YOLOv5 Release v7.0. 2022. Available online: https://github.com/ultralytics/yolov5/tree/v7.0 (accessed on 16 September 2024).
  56. Li, C.; Li, L.; Geng, Y.; Jiang, H.; Cheng, M.; Zhang, B.; Ke, Z.; Xu, X.; Chu, X. YOLOv6 v3.0: A full-scale reloading. arXiv 2023, arXiv:2301.05586. [Google Scholar] [CrossRef]
  57. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar] [CrossRef]
  58. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection. arXiv 2022, arXiv:2203.03605. [Google Scholar] [CrossRef]
  59. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. arXiv 2024, arXiv:2304.08069. [Google Scholar] [CrossRef]
Figure 1. The architecture of the quadrotor system with target detection and a tracking planner.
Figure 1. The architecture of the quadrotor system with target detection and a tracking planner.
Biomimetics 09 00567 g001
Figure 2. The bionic working principle, with the intensity of color being the odor distribution.
Figure 2. The bionic working principle, with the intensity of color being the odor distribution.
Biomimetics 09 00567 g002
Figure 3. Illustrates one iteration of the GBAS algorithm under the assumption of a step size of 1. (a) The red region represents the search space of the beetle. (b) The beetle randomly orients itself toward a grid, and the green region represents the grid where the beetle’s antennae are located. (c) The beetle reorients itself toward a grid with a lower objective function value and adds it to the candidate list. Then, the beetle tilts its antennae forward and adds the grid where the antennae are located to the candidate list as well. (d) The beetle moves to the grid with the minimum objective function value.
Figure 3. Illustrates one iteration of the GBAS algorithm under the assumption of a step size of 1. (a) The red region represents the search space of the beetle. (b) The beetle randomly orients itself toward a grid, and the green region represents the grid where the beetle’s antennae are located. (c) The beetle reorients itself toward a grid with a lower objective function value and adds it to the candidate list. Then, the beetle tilts its antennae forward and adds the grid where the antennae are located to the candidate list as well. (d) The beetle moves to the grid with the minimum objective function value.
Biomimetics 09 00567 g003
Figure 4. In order to achieve convergence within the desired distance, we have designed a step-size update based on distance feedback between the drone and the target. When the distance is greater than the desired distance, the step size decreases slowly. When the distance is less than the desired distance, the step size decreases rapidly.
Figure 4. In order to achieve convergence within the desired distance, we have designed a step-size update based on distance feedback between the drone and the target. When the distance is greater than the desired distance, the step size decreases slowly. When the distance is less than the desired distance, the step size decreases rapidly.
Biomimetics 09 00567 g004
Figure 5. Flowchart of proposed GBAS algorithm.
Figure 5. Flowchart of proposed GBAS algorithm.
Biomimetics 09 00567 g005
Figure 6. Flowchart of confusion-aware mechanism for distinguishing tracked targets.
Figure 6. Flowchart of confusion-aware mechanism for distinguishing tracked targets.
Biomimetics 09 00567 g006
Figure 7. The confusion-aware target localization method workflow. (a) Object detection output bounding boxes. (b) Feature point extraction and matching. (c) Calculating the average depth using a depth map. (d) Locating based on the average depth.
Figure 7. The confusion-aware target localization method workflow. (a) Object detection output bounding boxes. (b) Feature point extraction and matching. (c) Calculating the average depth using a depth map. (d) Locating based on the average depth.
Biomimetics 09 00567 g007
Figure 8. The influence of different parameters on algorithm convergence on an empty map. (a) The path of the GBAS when k 2 = 0.15 . (b) The path of the GBAS when k 2 = 0.005 . (c) In the case of different parameters, the decline in the objective function. (d) The change curve of the step size with different parameters.
Figure 8. The influence of different parameters on algorithm convergence on an empty map. (a) The path of the GBAS when k 2 = 0.15 . (b) The path of the GBAS when k 2 = 0.005 . (c) In the case of different parameters, the decline in the objective function. (d) The change curve of the step size with different parameters.
Biomimetics 09 00567 g008
Figure 9. The optimal paths of the tested algorithms in Map 1.
Figure 9. The optimal paths of the tested algorithms in Map 1.
Biomimetics 09 00567 g009
Figure 10. The optimal paths of the tested algorithms in Map 2.
Figure 10. The optimal paths of the tested algorithms in Map 2.
Biomimetics 09 00567 g010
Figure 11. The optimal paths of the tested algorithms in Map 3.
Figure 11. The optimal paths of the tested algorithms in Map 3.
Biomimetics 09 00567 g011
Figure 12. Object detection test set example.
Figure 12. Object detection test set example.
Biomimetics 09 00567 g012
Figure 13. Confusion-aware failure rates in different scenarios: the first group represents experimental sets with varying colors and shapes, the second group consists of experimental sets with different colors but the same shape, the third group includes experimental sets with the same color but varying shapes, and the fourth group comprises experimental sets with both the same color and shape.
Figure 13. Confusion-aware failure rates in different scenarios: the first group represents experimental sets with varying colors and shapes, the second group consists of experimental sets with different colors but the same shape, the third group includes experimental sets with the same color but varying shapes, and the fourth group comprises experimental sets with both the same color and shape.
Biomimetics 09 00567 g013
Table 1. GBAS algorithm parameters and descriptions.
Table 1. GBAS algorithm parameters and descriptions.
ParameterDescription
δ 0 Initial movement distance.
k 1 Decay rate as beetle nears target.
k 2 Adjusts step size based on distance.
rArea within which search stops.
Table 2. Parameter settings for the CEC tests.
Table 2. Parameter settings for the CEC tests.
ParametersValueDescription
i Max 2000*dimThe maximum number of iterations of the algorithm
ε 1 × 10 8 The convergence error of the algorithm
δ i 10The initial step size
C−6A constant used to adjust the lower limit of the step size
k0.0001Reflects the gradient when the step size decreases
m5000The role of adjustment error effects
L5The maximum number of times without significant improvement
α 1.5The step-size growth rate
P t 1The total group number
Table 3. Number function evaluation results on varying dimensions via proposed GBAS algorithm.
Table 3. Number function evaluation results on varying dimensions via proposed GBAS algorithm.
Dimension103050100
FunctionMeanStdMeanStdMeanStdMeanStd
F14.07e+033.46e+035.27e+035.64e+039.36e+039.98e+031.61e+041.88e+04
F30.00e+000.00e+001.19e+053.78e+042.32e+055.25e+046.19e+051.12e+05
F49.49e+001.57e+014.25e+013.00e+015.90e+013.80e+011.07e+024.87e+01
F52.47e+021.20e+027.56e+021.57e+021.08e+031.87e+022.29e+032.89e+02
F69.36e+012.06e+019.81e+011.30e+019.80e+011.09e+019.70e+017.53e+00
F79.37e+022.79e+023.85e+038.03e+027.09e+039.33e+021.42e+041.10e+03
F81.61e+026.82e+016.15e+021.65e+021.09e+031.75e+022.48e+033.23e+02
F93.61e+031.45e+031.36e+042.90e+033.00e+046.01e+035.90e+047.06e+03
F101.99e+033.92e+025.63e+036.36e+028.96e+039.54e+021.74e+041.33e+03
F119.45e+015.19e+011.95e+026.31e+012.89e+026.64e+013.30e+041.71e+04
F124.47e+063.66e+063.12e+061.95e+069.59e+064.22e+062.88e+071.44e+07
F131.63e+041.16e+041.66e+042.05e+041.21e+041.07e+041.43e+047.52e+03
F141.06e+049.63e+036.84e+056.14e+058.71e+054.48e+051.24e+065.82e+05
F156.78e+038.01e+037.25e+031.00e+041.00e+047.28e+033.76e+033.67e+03
F166.86e+022.49e+021.67e+034.03e+022.74e+034.97e+025.06e+037.41e+02
F175.17e+021.84e+021.17e+033.78e+022.50e+034.70e+024.65e+035.81e+02
F181.50e+041.18e+042.02e+061.81e+061.93e+061.34e+061.45e+066.26e+05
F196.53e+038.53e+031.14e+041.29e+042.57e+041.46e+044.85e+035.07e+03
F205.83e+021.78e+021.50e+033.73e+021.98e+033.90e+024.52e+036.70e+02
F212.66e+025.29e+014.51e+026.43e+017.09e+027.65e+011.48e+031.60e+02
F221.20e+036.26e+025.11e+039.15e+028.18e+031.06e+031.79e+041.41e+03
F234.06e+021.75e+026.16e+021.37e+029.84e+028.16e+011.38e+031.00e+02
F243.33e+021.29e+029.94e+022.06e+022.01e+032.00e+021.98e+031.32e+02
F254.34e+027.90e+014.12e+022.77e+015.53e+024.55e+017.87e+026.42e+01
F261.49e+036.03e+023.75e+031.47e+036.70e+032.09e+031.68e+042.59e+03
F274.63e+024.09e+015.96e+022.96e+011.21e+032.38e+021.37e+031.77e+02
F285.21e+021.44e+024.44e+023.66e+015.07e+023.05e+015.84e+023.05e+01
F294.33e+021.03e+021.38e+033.28e+022.32e+033.64e+025.52e+036.42e+02
F307.32e+057.94e+051.93e+052.46e+057.85e+064.34e+078.49e+056.07e+05
Table 4. Feature descriptions on different maps for path planning.
Table 4. Feature descriptions on different maps for path planning.
EnvironmentFeaturesDesign Goals
Map 1Obstacles with large shapesTo evaluate the ability to handle specific passing areas
Map 2Obstacles with discrete distributionsTo evaluate the efficiency of the algorithm
Map 3Obstacles with concave shapesTo evaluate the ability to handle concave obstacles
Table 5. Comparisons with different existing algorithms and statistical tests.
Table 5. Comparisons with different existing algorithms and statistical tests.
MapAlgorithmPath LengthRun Times (ms)Speedup
Mean ± Std ↓p-Value ↓Mean ± Std ↓p-Value ↓Mean ↑
Map 1A * [45]49.18 ± 0.003.02e-111.81 ± 1.301.21e-12321%
SPSO [37]57.20 ± 10.178.91e-022.56 ± 2.343.01e-11495%
IACO-IABC [38]58.18 ± 0.003.35e-111.68 ± 1.413.01e-11291%
GDRRT * [39]58.48 ± 5.254.98e-042.20 ± 1.124.07e-11412%
GBAS (Ours)55.67 ± 2.88-0.43 ± 0.17--
Map 2A * [45]54.60 ± 0.001.33e-081.29 ± 0.294.20e-10135%
SPSO [37]60.41 ± 10.391.44e-021.60 ± 0.783.02e-11191%
IACO-IABC [38]61.60 ± 0.002.84e-081.98 ± 1.213.02e-11260%
GDRRT * [39]61.43 ± 2.955.32e-031.04 ± 0.461.19e-0689%
GBAS (Ours)59.10 ± 3.53-0.55 ± 0.16--
Map 3A * [45]57.43 ± 0.007.47e-101.20 ± 0.115.09e-0646%
SPSO [37]62.89 ± 8.652.17e-012.61 ± 1.513.02e-11218%
IACO-IABC [38]63.67 ± 0.007.87e-111.84 ± 1.313.02e-11124%
GDRRT * [39]62.49 ± 3.825.08e-030.99 ± 0.322.92e-0221%
GBAS (Ours)61.59 ± 6.41-0.82 ± 0.40--
Table 6. Comparison of computational time with the benchmark.
Table 6. Comparison of computational time with the benchmark.
Target v max MethodMean Times ↓ (ms)Speedup ↑
1 m/sJi et al. [8]0.96233%
Ours0.18-
2 m/sJi et al. [8]1.41555%
Ours0.22-
Table 7. Comparison of real-time performance of different models for object detectors.
Table 7. Comparison of real-time performance of different models for object detectors.
ModelBackboneEpochsFPS ↑Params ↓ (M)
YOLOv5-L [55]-1004246
YOLOv6-L [56]-1005035
YOLOv7-L [57]-1003836
Ours-100 62 43
DINO-Deformable-DETR [58]R50365046
RT-DETR-R50 [59]R5072642
RT-DETR-L [59]HGNetv27214 32
Table 8. Comparison of key detection performance of different models for object detectors.
Table 8. Comparison of key detection performance of different models for object detectors.
ModelGFLOPs ↑Precision ↑Recall ↑F1-Score ↑
YOLOv5-L [55]10968.687.40.77
YOLOv6-L [56]15164.588.20.75
YOLOv7-L [57]10470.790.40.79
Ours16485.2 92 . 7 0.89
DINO-Deformable-DETR [58] 276 69.288.60.78
RT-DETR-R50 [59]135 88 . 2 92.1 0 . 90
RT-DETR-L [59]10887.592.40.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, Y.; Ma, C.; Chen, D. Enhancing Target Tracking: A Novel Grid-Based Beetle Antennae Search Algorithm and Confusion-Aware Detection. Biomimetics 2024, 9, 567. https://doi.org/10.3390/biomimetics9090567

AMA Style

Lu Y, Ma C, Chen D. Enhancing Target Tracking: A Novel Grid-Based Beetle Antennae Search Algorithm and Confusion-Aware Detection. Biomimetics. 2024; 9(9):567. https://doi.org/10.3390/biomimetics9090567

Chicago/Turabian Style

Lu, Yixuan, Chencong Ma, and Dechao Chen. 2024. "Enhancing Target Tracking: A Novel Grid-Based Beetle Antennae Search Algorithm and Confusion-Aware Detection" Biomimetics 9, no. 9: 567. https://doi.org/10.3390/biomimetics9090567

APA Style

Lu, Y., Ma, C., & Chen, D. (2024). Enhancing Target Tracking: A Novel Grid-Based Beetle Antennae Search Algorithm and Confusion-Aware Detection. Biomimetics, 9(9), 567. https://doi.org/10.3390/biomimetics9090567

Article Metrics

Back to TopTop