Next Article in Journal
Effects of Flight and Smoothing Parameters on the Detection of Taxus and Olive Trees with UAV-Borne Imagery
Next Article in Special Issue
Bearing-Based Distributed Formation Control of Unmanned Aerial Vehicle Swarm by Quaternion-Based Attitude Synchronization in Three-Dimensional Space
Previous Article in Journal
Preliminary Clinical Validation of a Drone-Based Delivery System in Urban Scenarios Using a Smart Capsule for Blood
Previous Article in Special Issue
A Four-Dimensional Space-Time Automatic Obstacle Avoidance Trajectory Planning Method for Multi-UAV Cooperative Formation Flight
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-UAV Collaboration to Survey Tibetan Antelopes in Hoh Xil

School of Energy and Power Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
*
Author to whom correspondence should be addressed.
Drones 2022, 6(8), 196; https://doi.org/10.3390/drones6080196
Submission received: 13 July 2022 / Revised: 31 July 2022 / Accepted: 4 August 2022 / Published: 6 August 2022
(This article belongs to the Special Issue Intelligent Coordination of UAV Swarm Systems)

Abstract

:
Reducing the total mission time is essential in wildlife surveys owing to the dynamic movement of animals throughout their migrating environment and potentially extreme changes in weather. This paper proposed a multi-UAV path planning method for counting various flora and fauna populations, which can fully use the UAVs’ limited flight time to cover large areas. Unlike the current complete coverage path planning methods, based on sweep and polygon, our work encoded the path planning problem as the satisfiability modulo theory using a one-hot encoding scheme. Each instance generated a set of feasible paths at each iteration and recovered the set of shortest paths after sufficient time. We also flexibly optimized the paths based on the number of UAVs, endurance and camera parameters. We implemented the planning algorithm with four UAVs to conduct multiple photographic aerial wildlife surveys in areas around Zonag Lake, the birthplace of Tibetan antelope. Over 6 square kilometers was surveyed in about 2 h. In contrast, previous human-piloted single-drone surveys of the same area required over 4 days to complete. A generic few-shot detector that can perform effective counting without training on the target object is utilized in this paper, which can achieve an accuracy of over 97%.

1. Introduction

Located between the Tanggula Mountains and the Kunlun Mountains, Hoh Xil is one of the main water sources of the Yangtze River and the Yellow River [1]. It is a significant habitat for wildlife, such as Tibetan antelopes, black-necked cranes, lynxes, and wild yaks. Hoh Xil National Nature Reserve (HXNNR) was established in 1995. It was majorly divided into four national nature reserves, the Qiangtang, the Arjinshan, the Sanjiangyuan, and the Hoh Xil, located in Qinghai province. Furthermore, HXNNR joined the World Heritage List in 2017 for its unique biodiversity and environmental conditions [2], of which Tibetan antelopes constitute a highly representative population.
However, wild Tibetan antelopes have been listed as endangered by the International Union for Conservation of Nature (IUCN) [3]. Facing this heated issue, they are protected by the Wildlife Protection Law of China, and the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES). Over the past century, hunting and grazing have been the main threats to the survival rate of the Tibetan antelope population. Worse, Tibetan antelopes’ poachers killed large numbers of antelopes for their hides and skins due to the enormous financial profits. From 1900 to 1998, the Tibetan antelope population declined from a million to less than 70,000, with a mortality rate of nearly 85%. In recent two decades, the Chinese government and wildlife conservation organizations have acted to constrain poaching, which has gained progress. Still, the deterioration of the antelopes’ natural habitat and environmental degradation continue to threaten their survival [4]. Owing to the shortage of funds and labor, reserve staff can only monitor limited areas but not the entire reserve. However, artificial ground investigation caused potential damage to the protection of fragile regions [5].
Traditional survey methods mainly include field surveys by vehicle or on foot. Nevertheless, Tibetan antelopes, especially females in pregnancy, are sensitive to the appearance of humans and vehicles, so surveys are usually conducted from a distance. This method can be appropriate during the non-migratory season when small groups are evenly distributed across the plateau. Still, it may lead to biased data where terrain features block sight lines, or animals are herded [6]. Moreover, some population surveys were conducted by land-based rovers remotely piloted within survey sites [7], but this method is not suitable for surveys in larger areas or uneven, rocky, and steeply sloping terrain. It is arduous to determine the population size with ground survey methods, as animal movements can cause violent fluctuations in density estimates. In addition, severe weather conditions in unpopulated areas often create much uncertainty at high altitudes, making it difficult for observers to access investigation areas [8]. Moreover, some zoologists tracked individuals through collars with global positioning system devices to identify calving grounds, migration corridors, and suitable habitats [1,9]. However, they cannot accurately and quickly estimate the number of animals or their distribution range, so new observation methods are needed.
In recent years, aerial surveys by UAVs in unpopulated areas, poor-developed transportation, and a fragile environment have become promising options [5,6]. Compared to full-size helicopters, small UAVs produce less noise flying at altitudes that can effectively obtain ground information and significantly reduce the disturbance to wildlife [10]. Fixed-wing UAVs have also been applied to wildlife surveys [11]. However, they are not suitable for the large-area survey missions in this survey, considering the take-off and landing environments and portability. Moreover, fixed-wing UAVs may also disturb wildlife as they may be perceived as avian predators [12]. In contrast, rotary-wing UAVs are more controllable and can obtain higher-resolution images, as UAVs can fly at lower speeds and altitudes without disturbing Tibetan antelopes [6]. Besides, using commercial rotary-wing UAVs can reduce the difficulty of flight operations for scientific investigations and allow researchers to focus on data analysis [13]. Moreover, UAV airports have been successfully applied in multiple fields recently [14], making regular unmanned aerial surveys possible. Utilizing multiple UAVs in an autonomous system can reduce manual requirements and conduct surveys with higher frequency and efficiency with the help of the short window period of weather, environments, and animal migration.
The efficiency of UAVs in conducting aerial surveys depends largely on the strategies the area coverage paths are planned. Generally, the complete coverage path-planning (CCPP) problem of a single UAV can be solved by sweep-style patterns, which guide the UAV back and forth over a rectangular space [15]. Other similar spatial sweep-style patterns include spiral patterns [16] and Hilbert curves [17]. Robotics designed to solve cyclic paths may encounter many tracebacks in sweep-style patterns [18]. Meanwhile, some tracebacks also occur in graphic search methods, such as wavefront methods [19,20]. Approaches that focus on minimizing the number of turns are often improved based on the minimum energy method [21,22], which is more suitable for fixed-wing or fast-moving aircraft. In contrast, we pay more attention to rotary-wing UAVs, consuming less energy during low-speed turns. In these path planning methods, the distance between the starting and ending points of the trajectories is often long. At the same time, unnecessary tracebacks may exist if the take-off and landing positions are far from the survey area. Moreover, it is time-consuming and inefficient for multiple sorties by a single UAV to perform coverage survey tasks in large areas. Presently, it is becoming a trend for multiple UAV formations to accomplish tasks jointly [23].
Most multi-UAV CCPP algorithms are extended based on the single-UAV area coverage research [24]. The common multi-UAV CCPP can be divided into two ways according to whether the task assignment is performed or not. In the multi-UAV CCPP without task assignment, each UAV may avoid collision with others by communicating with or regarding them as moving obstacles. After that, they complete area coverage tasks wholly and jointly. For instance, the path planning of each UAV is randomly generated by random coverage methods [25] and with a high traceback rate. There is also an improvement in the spanning tree coverage algorithm that employs multiple UAVs for complete coverage of a task area where there is acquired environmental information [26]. However, this method has poor robustness, and a planned trajectory’s quality depends on the robot’s initial position. In contrast, the research idea of the task assignment-based multi-UAV CCPP method is to divide a task area according to a certain way and assign the multiply divided sub-regions to each UAV. Then, each UAV plans its sub-region coverage trajectory within the region by a single UAV CCPP algorithm. For instance, task assignments of the sweep-style pattern [18] and polygonal area coverage [27] have been widely studied with high backtracking rates. Moreover, many of these methods are often used to find UAVs that satisfy appointed duration requirements. Moreover, many of these methods are often used to find UAVs that meet some endurance requirements. However, this paper mainly emphasizes generating optimal paths that satisfy a maximum path length constraint decided by UAV’s endurance conditions.
Therefore, we formulate the multi-robot CCPP problem according to the satisfiability modulo theory (SMT), which is a generalization of Boolean satisfiability because it allows us to encode all the specific survey-related constraints. Compared to the traditional traveling salesman problem (TSP), our method focuses more on finding feasible paths that satisfy a maximum path length constraint, allowing for vertex revisitation. Because the UAV’s battery constraints dictate the maximum path length, we can find solutions tailored to the limitations of the chosen UAV rather than finding a vehicle that would need to meet some endurance requirement. Although, our planning method is inherently a non-deterministic polynomial complete problem. The coverage problem can also be encoded as a MILP [28,29], it takes 4 to 6 days to solve problems for the same area, even using the most advanced solver on a powerful workstation. In contrast, the SMT method can generate available solutions in a few hours on a laptop [30].
The raw image shot by UAVs were assembled to form a large patchwork map. Then, this larger image is adopted for analysis. For example, ecologists leverage it to count flora and fauna for population analysis [31] or wildfire risk assessment in forests and grasslands [32]. However, most current counting methods can only be applied to specific types of objects, such as people [33], plants [34], and animals [31]. Meanwhile, they usually require images with tens of thousands or even millions of annotated object instances for training. Some works tackle the issue of expensive annotation cost by adopting a counting network trained on a source domain to any target domain using labels for only a few informative samples from the target domain [35]. However, even these approaches require many labeled data in the source domain. Moreover, obtaining this type of annotation is costly and arduous by annotating millions of objects on thousands of training images, especially when the photos of endangered animals are scarcer. The few-shot image classification task only requires learning about known or similar categories to classify images without training in a test, which is a promising solution [36,37]. Most existing works for few-shot object counting involve the dot annotation map with a Gaussian window of a fixed size, typically 15 × 15, generating a smoothed target density map for training the density estimation network. However, there are huge variations in the size of different target objects, and using a fixed-size Gaussian window will lead to significant errors in the density map. Therefore, to solve this problem, this paper uses Gaussian smoothing with adaptive window size to generate target density maps suitable for objects of different sizes. Once UAV aerial survey data have been collected and images have been analyzed to acquire accurate survey data, which can provide information on species distribution and population size, it also enables a deeper understanding of animals’ seasonal migrations and habitat changes. That is critical for preparing proper wildlife conservation policies and provides affluent data for ecologists. The focus of this paper is on the development of a multi-UAV system for conducting aerial population counting.
In this paper, we considered a path planning method with a set of aerial robots for population counting, which plans the trajectories of these robots for aerial photogrammetry. UAVs acquire a set of images covering the entire area with overhead cameras. It requires the images to have higher resolution and ensure enough overlap to stitch the images further together. Besides, aerial surveys should be conducted as soon as possible, owing to changing weather, light, and dynamic ground conditions. So it is necessary to require all UAVs to perform their tasks simultaneously. The collected UAV data and analyzed images will promote the sustainable development of population ecology. This paper focuses on developing a multi-UAV system for conducting aerial surveys to provide an autonomous multi-robot solution for UAV surveys and planning suitably for most survey areas.
This research is aimed at: (1) developing and testing a multi-UAVs CCPP algorithm for fast and repeatable unmanned aerial surveys on a large area that optimizes mission time and distance. Meanwhile, it compensates for the short battery life of small UAVs by minimizing tracebacks when investigating large areas. (2) Developing and testing a population counting method applicable to various species that does not require pre-training and large data sets.
The method in this paper is demonstrated in an extensive field survey of Tibetan antelope populations at the HXNNR, China. In this task, the celerity and repeatability of aerial surveys were conducted under extreme conditions on a large area. So, it can demonstrate the effectiveness of our system in facilitating the development of population ecology. In this paper, an investigating system of four UAVs observed a herd containing approximately 20,000 Tibetan antelopes distributed over an area of more than 6 km2, which reduced the survey time from 4 to 6 days to about 2 h. The method put forward in this paper can complete aerial surveys in shorter distance and less mission time. For instance, this method reduced the total distance and mission time by 2.5% and 26.3%, respectively. Furthermore, the effectiveness increased by 5.3% compared to the sweep-style pattern over an area of the same size for the Tibetan antelope herd survey described in this paper.

2. Materials and Methods

2.1. Survey Area

Situated in southwest Qinghai Province, China (top right of Figure 1), the unpopulated areas of Hoh Xil are 83,000 square kilometers with an average altitude of 4800 m. The Zonag Lake and Sun Lake in northwest HXNNR (lower left of Figure 1) have been identified as the current principal calving grounds for Tibetan antelopes. The birth season reaches its peak from mid-June to July every year, when tens of thousands of female Tibetan antelopes from the TRHR (Three-Rivers Headwater Region), the Qiangtang Plateau of Tibet, and Hoh Xil migrate about 300 km southwest to the Zonag Lake and the Sun Lake for parturition [4]. Therefore, the Zonag Lake is also known as the “great delivery room of antelopes”. After calving, the antelope population will return to the northeast and arrive at the major overwintering pastures in mid-August.
Zonag Lake most frequently sees poachers. It is difficult to hunt antelopes because they usually are strong and vigorous. Most poachers decide to besiege Tibetan antelopes during the calving period around the Lake and kill them, which seriously affects the ecological balance of Tibetan antelope reproduction and leads to a sharp decline in population quantity. Tibetan antelopes are almost extinct.
One of the main calving grounds is presented in Figure 2. The water area of the Zonag Lake shrank drastically by 39% after the flood occurred in 2011. The western, southern, and eastern shorelines of the lake significantly altered, and the calving grounds of Tibetan antelopes on the southwestern shore also changed [2]. Recently, the government and organizations have gradually paid attention to conserving Tibetan antelopes. However, specific conservation efforts are often accompanied by the danger of entry into unpopulated areas, which requires high human costs and may cause unavoidable human interventions on the antelopes. In order to reduce costs in antelope population conservation and human interventions, we arranged a team of four rotary-wing UAVs to conduct aerial surveys and population counting in one of the main calving grounds of Tibetan antelopes located on the southwestern shore of the Zonag Lake. Meanwhile, we can learn more about the Tibetan antelope population’s reproduction and changes in calving grounds.

2.2. CCPP Methods

2.2.1. Path-Planning Problem Encoding

The multi-UAV CCPP problem was described as SMT in this paper, which is a popularization of SAT, as it allows us to encode all the specific survey-related constraints. SMT allows predicate logic in addition to the propositional logic used in the SAT solver.
B i , a , t { 0 , 1 } was set to be a Boolean variable where i { 1 , , N r o b o t s } , a A = { 1 , , N v e r t e x } , a n d   t { 0 , , T max } . The UAV i is at the vertex a A when the step is t if B i , a , t = 1 . The parameter T max is the maximum trajectory allowed by any UAV. For instance, a UAV will pass through a maximum of 20 vertices if T max = 20 . For simplicity, it was assumed in this paper that all UAVs share the same T max and there were enough UAVs to cover the entire map. The Figure G ( A , E ) was constructed based on the survey areas, where there are various types of logical statements to describe the expected behaviors allowed by UAVs. The solver attempted to assign a value of 0 (False) or 1 (True) to each Boolean variable. By doing so, the values of these statements formed in the predicate logic are true. These logical statements can be considered constraints, like expression approaches to the optimization problem.
To strengthen the physical constraints of coverage, it is required to define that
( i , t )   a   B i , a , t = 1
works only for an a . This expression confines the UAV i to the point a at any given step t . Then, the UAV is only allowed to move to the vertex y at the step t + 1 , given that it is in the vertex a at the step t . Therefore, there is a line ( a ~ y ) between these two vertices.
¬ B i , a , t A a ~ y B i , a , t + 1 = 1
Only the starting/ending vertex is connected to itself, and all other vertices have no self-loops. Next, it is defined that:
a   ( i , t ) B i , a , t = 1
It enforces constraint requirements of the coverage task. For each vertex a in the figure, there is at least a step t where there is at least a UAV i occupying the space. If there is only a pair of ( i , t ) in a spot where the constrains must be satisfied, it will be equivalent to traveling salesman problem. The anti-collision constraint between UAVs may be achieved by, at most, one i as required.
a , t   B i , a , t = 1
At the end, the starting and the ending points were coded in this paper by setting a i 0 as the starting and ending vertices of the UAV i :
B i , a i 0 , 0 = B i , a i 0 , T max = 1
It is expected that there is a closed-loop trajectory to ensure that UAVs do not have to return to the take-off location with repeated paths through the area. A separate boundary condition can be set if a closed-loop trajectory is not required, for the maximum trajectory length is determined by a UAV’s battery use. To find a feasible trajectory means that the UAV has enough power to arrive at the end of the trajectory. In an emergency recall, the UAV will have enough power to return, since the straight-line distance between any point of the trajectory and its ending point is at most the length of its remaining part. This feature allows the UAV to safely return to the starting point of the survey path when the vehicle cannot safely complete the survey (e.g., adverse weather conditions or poor visibility). It also encourages the UAV to end the survey at a spot near the starting point of the trajectory.

2.2.2. Model Optimization

  • Variables reduction
The number of Boolean variables and the computing time can be reduced by leveraging the problem structure although these four constraints have fully encoded the multi-UAV coverage problem on SMT. It is assumed that each UAV can only move to an adjacent vertex at each time step, then the binary variables representing unreachable states can be removed from the problem to reduce the problem-solving time. Algorithm 1 describes the method.
2.
Sequential SMT
An SMT instance does not directly optimize the objective function but only finds a feasible solution at any time. The standard descending (or ascending) method was adopted in this paper to solve the objective function until the problem is infeasible to transform the feasibility framework into an optimization framework. Algorithm 1 performs this iterative process by repeatedly solving an SMT instance to identify the shortest trajectory. Each step is sequentially decremented by T max according to a certain search dispatch until the user terminates the process or the problem is infeasible. Regarding problem structure, each subsequent iteration guarantees that the generated trajectory is no longer than that in the previous iteration if a solution is found. For instance, linear search dispatch for T max is provided to Algorithm 1. However, in effect, any reasonable search dispatch is applicable.
3.
Parameter optimization
A parameter optimization algorithm (POA) [38] was adopted for the optimal trajectory. In this paper, the variational design of variables was performed by adding a normally distributed random value to each design variable, based on the evolutionary strategy of Schwefel’s [39] work. The goal of the POA is to minimize f ( x ) . We use x i to denote the candidate solution at the i-th generation. For each generation, we randomly mutate x i to obtain x i . If f ( x i ) < f ( x i ) , then we set x i + 1 = x i .If f ( x i ) > f ( x i ) , then we use the following logic to determine x i + 1 . If we had set x k + 1 = x k at the previous generation k at which f ( x k ) > f ( x k ) , then we set x i + 1 = x i with a 10% probability, and x i + 1 = x i with a 90% probability. However, if we had set x k + 1 = x k at the previous generation k at which f ( x k ) > f ( x k ) , then we set x i + 1 = x i with a 50% probability, and we set x i + 1 = x i with a 50% probability. This POA is greedy as it always accepts a beneficial mutation. However, it also includes some exploration in that it sometimes accepts a detrimental mutation. The probability of accepting a detrimental mutation varies, depending on whether or not the previous detrimental mutation was accepted. The algorithm for this POA is shown in Algorithm 1.
Algorithm 1 Model optimization algorithm
Drones 06 00196 i001
AcceptFlag indicates if the previous detrimental mutation replaced the candidate solution. The function of Distance() refers to time steps required to return to two vertices. In a linear grid, the function of Distance() refers to Manhattan distance required to return to two points.

2.2.3. Evaluation Indexes

The focus of this study is to reduce the total survey mission time and the total course flown by UAVs in order to accomplish a survey quickly while saving energy consumption. Therefore, the total course length was regarded as a constraint, rather than the result of a path planner’s calculation. The planning aims to satisfy the trajectory coverage requirement while minimizing the distance of tracebacks. The efficiency of a set of planned trajectories R is defined as:
η path = i R S i i R L i i R S i
S i is the total course length flown by the UAV i , L i is the sum of the UAV’s i repeated flight paths and the distance from the take-off point to the survey area (also called transit distance). When the efficiency is 1, it means that the UAV took off within the survey area and had no repeated flight path.

2.3. Counting Methods

2.3.1. Adaptation and Matching Counting Network

In Figure 3, the network architecture of the animal counting algorithm adopted in this paper is depicted. Input to the network is the image X H × W × 3 and several examples describing bounding boxes of the objects to be counted from the same image. Output from the network is the predicted density map Z H × W , and the counts of the interesting objects are obtained by summing up all density values. The network consists of two crucial modules: (1) a multi-scale feature extraction module; (2) a density prediction module. These two modules were designed in this paper for them to handle new categories in a test. ImageNet was adopted to extract features for such a network that can handle a wide range of visual categories. The feature extractor consists of a multi-layer convolution operation that maps the input into d-channel features. For the image X , it outputs a downsampled feature map F ( X ) d × h × w . For an exemplar box a , the output feature map is further processed with global average pooling to form a feature vector F ( a ) d .The density prediction module was designed to be incognizable to visual categories. The multi-scale feature extraction module consists of the first four blocks from a pre-trained ResNet-50 backbone [40]. Images were represented in this paper by convolutional feature mapping in the third and fourth blocks. In addition, ROI pooling was performed in this paper for the convolutional feature mapping from the third and fourth Resnet-50 blocks to obtain the examples’ multi-scale features.
The features obtained from the feature extraction module were not directly adopted in density prediction, in order to make the density prediction module incognizable to visual categories. Instead, the correlation mapping between sample features and image features was input to the density prediction module. The sample features were resized to different scales given the target objects at different scales should be considered. The resized sample features were associated with the image features to obtain multiple correlation maps, each of which corresponds to one scale. Scales of 0.9 and 1.1, as well as the original scales, were adopted in all experiments. The correlation maps were connected in series’ and fed to the density prediction module. The density prediction module consists of five convolutional blocks and three upsampling layers placed after the first, second, and third convolutional layers, respectively. The last layer is a 1 × 1 convolutional layer that predicts the two-dimensional density map. The size of the predicted density map is the same as the size of the input image.

2.3.2. Adaptative Window

The training images of a dataset were leveraged to train the network. Each training image contains multiple target objects. Moreover, most objects were only annotated with dots, while only the example objects were annotated with bounding boxes. However, it is difficult to train the density prediction network based on the training losses defined by dotted annotations. Most existing visual counting efforts, especially crowd counting [41], use a fixed size Gaussian window (usually 15 × 15) to convolve dotted annotated maps, thus generating smooth target density maps for training the density prediction network. The selected training dataset consists of 147 different categories in which the object size varies considerably [42]. Therefore, Gaussian smoothing with a self-adaptive window size was employed to generate the target density map. Firstly, the object size was estimated with dotted annotations. In a given dotted annotated map, each dot is nearly located at the center of the object. The horizontal distance x i and vertical distance y i between each dot i and its nearest neighbor were calculated. With the modes x , y ^ of X and X , Y ( x i X , y i Y , i = 1 , 2 , 3 n ) , the distance d = ( x ) 2 + ( y ^ ) 2 can be calculated as the size of the Gaussian window to generate the target density map. The Gaussian standard deviation was set to be one quarter of the window size. The mean square error between the density prediction map and the ground truth density map was minimized to train the network. To match it, we adopted Adam optimizer with a learning rate of 10 5 and a batch size of 1. The height of each image was fixed at 384, and the width was accordingly adjusted to maintain the aspect ratio of the original image.

2.3.3. Loss Function

The above network structure only uses the samples’ example bounding boxes to extract their feature information. It does not fully use the position information provided by the exemplar boxes, while annotations by human eyes are crucial in this process. So we define the combination of two losses here to correct the model’s errors during the test.
  • Count Loss
For each example bounding box, a, the sum of the density values within the box should be at least 1 because the exemplar box may also contain parts of other objects or overlap other objects with the same type to some extent. The following count loss was defined to quantify the number of constraint violations.
c o u n t = a A max ( 0 , 1 Z a 1 )
In this equation, A is the set of bounding boxes, and Z a 1 the sum of all values in Z a .
  • Perturbation Loss
The density prediction map Z in essence is a correlation map between a sample and an image. Therefore, the density values near the sample should be similar to Gaussian distribution in an ideal circumstance. G h × w was set to be a two-dimensional Gaussian window with the size of h × w . The perturbation loss was defined as follows:
p = a A Z a G h × w 2 2
  • The Combined Adaptation Loss
The self-adaptation loss for testing is a weighted combination of the minimum count loss and the perturbation loss.
A d a p t = λ 1 c o u n t + λ 2 p
In the equation, λ 1 and λ 2 are coefficients of the count loss and the perturbation loss respectively. In testing, we referred to parameters setting [42], so as to perform 100 gradient descent steps for each test image and optimize the weighted loss of Equation (9). The learning rate adopted in this paper is 10 7 . The values of λ 1 and λ 2 are, respectively, 10 9 and 10 4 . The self-adaptation loss was only employed in testing. The loss function during training is the root mean square error between the density prediction map and the actual density map.

2.3.4. Evaluation indexes

The mean absolute error (MAE) and the root mean square error (RMSE) were adopted in this paper to measure the accuracy of counting methods, which are defined as follows.
MAE = 1 n i = 1 n | c i c ^ i |
RMAE = 1 n i = 1 n ( c i c ^ i ) 2
In these equations, n refers to the quantity of test images, c i and c ^ i are the ground truth and the predicted counts, respectively.

3. Results

3.1. Multi-UAVs CCPP

3.1.1. Summary of the Surveys

Considering the altitude of the animal survey area is around 5000 m, we chose a high-altitude low noise propeller. In this way, the UAVs can better adapt to the flight conditions at high altitudes and meet the noise limits of animal surveys. During the execution of the survey, the path spacing and flight altitude of the UAV are set according to the camera parameters and image stitching requirements. During non-mission execution (transit distance transit distance to and from the takeoff point and mission start point), the flight speed of the UAV is 10 m/s, which is determined by the environmental conditions and the specifications of the UAV. We want the speed to be as fast as possible under the premise of ensured safety and low noise. In addition, considering the safety withdrawal requirements and the actual hovering test results in abnormal weather, we set the safe flight time of the UAV at 28 min. Its detailed parameters are shown in Table 1.
During each survey, more than 6000 images were taken by a group of UAVs and stitched together to form a large, stitched image covering the entire survey area, as presented in Figure 4d. In this paper, these stitched images were employed to identify the number of Tibetan antelopes for estimating the population density of these groups. In our work, CCPP methods and population counting were emphasized rather than ecology.
Figure 4 shows the workflow of our system in the survey area (Figure 2). Firstly, we specified the interested area and divided it into grids with the spacing of 100 m between vertices, as shown in Figure 4a. Secondly, the CCPP algorithm is used to generate offline trajectories and optimize the starting and ending positions of the UAVs, as shown in Figure 4b. The four starting points are distributed in the vicinity of the survey area to ensure the optimal total flight distance and mission time for a group of UAVs. Thirdly, UAVs that execute these routes were activated to collect survey images at regular intervals. Figure 4c shows the image coverage area of each UAV. The images are stitched according to the UAVs’ time intervals and flight paths. The four different color blocks represent the captured areas by each of the four UAVs. Finally, these images were stitched together to generate a mosaic image, as shown in Figure 4d. The mosaic images will be used for animal counts to estimate population size, as described in Section 3.2.
12~13 flights (the flight number depends on the UAV position) were required to survey the area with four deployed UAVs in advance. It cost 2 days for a single manually piloted UAV to survey. However, in this paper, the survey can be completed in about 2 h in some regions of Zonag Lake by the proposed multi-UAV collaborative survey method. Using the autonomous system reduces the burden on manual requirements and realizes faster surveys with higher frequencies, thus enabling the exploitation of short windows in favor of the survey. That makes it more likely to make arrangements in advance and perform unmanned observation and data collection with the knowledge of animal migration routes.
The paths in Figure 4b were input to universal ground control software (UGCS) [43] for arranging the interface between a route planner and the UAV on-board flight computer. Although a trajectory is generated offline, its self-adaption allows a user to place the starting node at any position on it without recalculating the entire path. The entire course of each path is displayed in Figure 5. The UGCS will send a UAV command after it takes off so that the drone can execute the route and automatically land back at the starting point for recharging, thus preparing for the next flight.

3.1.2. Comparisons Ignoring Transit Distance

In this paper, we employed the SMT algorithm employed with the Z3 solver [44], the experimental calculations, which are performed under CPU:12th Gen Intel(R) Core(TM) i9-12900KF 3.20 GHz and the operating system of Windows11. Each computation takes about 15 s, and optimized calculation can get the best result in less than 100 epochs. We developed the algorithms on python and the commercial software ISIGHT, and we eventually ported them to python for ease of deployment and fast computation.
The proposed method in this paper was compared with both the common sweep-style method in the literature [18] and recently cell decomposition polygon coverage path planning method [27]. The sweep-style pattern was planned by using the area coverage function already realized in the UGCS [43]. The polygon coverage method divides the required region into polygons first, then plans coverage routes by using the sweep-style method, and last stitches these paths together with the TSP solver at the end. Moreover, Polygon division allows the solver to locally optimize the sweep direction of each polygon cell, which leads to shorter trajectories in most cases.
The efficiency of the polygon coverage method is higher than that of the simple sweep-style method, but there are still considerable tracebacks in both two methods because the core planning strategy of two methods lies are the sweep-style pattern. We compared routes of our trajectory planning algorithm, the standard sweep-style method, and the polygon division method in Figure 6. Although the sweep-style method has covered enough areas, it is influenced greatly by tracebacks, as presented in Figure 6b.
The path efficiency of our method, the sweep-style method, and the polygon method are demonstrated in Table 2. The most effective set of paths is presented in bold. Algorithms in this paper realize the best efficiency in all but one case, with an average efficiency improvement of 8.2% compared to the sweep planner and 8.1% compared to the polygon coverage method. Without considering the transit distance of the planned trajectory, the sweep-style method results in a total course that is about 700 m longer than the trajectory generated by our method. The route generated by the polygon method is 400 m longer than that planned by our method. In all cases, the method proposed in this paper can reduce 366 m and 1.5 min compared to the polygon method, which will be 666 m and 2.2 min, compared to the sweep-style method.

3.1.3. Comparisons Considering Transit Distance

In an actual flight survey, there is a long distance between the UAV take-off point and the survey area. In the event of regular multi-UAV surveys, the take-off point may be flexibly arranged in advance before the survey window, which can reduce the impact of tracebacks and the consumption of UAV power to a certain extent. The total UAV flight distance was optimized in this paper within a given take-off area to make full use of the advantages of multiple UAVs.
We demonstrated the actual survey results of our method, the sweep-style method, and the polygon method in parts of the Zonag Lake (Figure 2) in Table 3, where the effective set of trajectories are presented in bold. In Table 3, the take-off point is designated at the geometric center of the take-off area, based on empirical judgment. It is evident that the trajectory efficiency and the mission time are improved for different methods by flexibly configuring take-off points (number of starting points, restrictions on the starting area). Differences among the methods are few and finding the optimal set of paths is easy when there are two different take-off points. As the differences among algorithms grow when plans become more flexible, there are more starting points available for choosing. Moreover even the take-off area is no longer restricted to be outside the survey area (which is undesirable in animal surveys because the noise from the UAV may be perceived as a threat). The method proposed in this paper demonstrates the best path’s efficiency and the shortest time consumption under all three different plans, with a 22.72% increase in trajectory efficiency and a 34.62% reduction in mission time with unrestricted starting area. The method may reduce at most 54 min of the mission time, when compared to the manual setting method (designate starting points according to experience), which is equivalent to the time about twice as long as the maximum UAV endurance time.
During short-term and quick surveys, the flight time of UAVs is a factor to be fully considered to make full use of the valuable survey window. Shorter aerial survey time can help accomplish more surveys and capture more data throughout the breeding season. Based on the above methods, the total UAV flight distance and total mission time were optimized in this paper within a given planned take-off area.
Table 4 shows the actual test results of our method and other methods in some parts of Zonag Lake. If the configuration of starting points is more flexible (number of starting points, restrictions on the starting area), the metrics of the different methods will improve accordingly. This is similar to the results of the method mentioned above, that optimizes only the total flight distance of a group of aircrafts. The variation between algorithms is increasing as plans become more flexible, starting points become more selective, and even take-off areas are no longer restricted to survey areas. Our method shows the best trajectory efficiency and the shortest task time among all three different schedules, with a 23.63% improvement in path efficiency and a 47.83% reduction in task time as the starting area is unrestricted. Our method’s flight distance is unchanged, compared to methods only optimizing the total flight distance while the total mission time decreases from 2.6 h to 1.2 h at most, with an improvement of 19.23%. The method proposed in this paper witnesses an average improvement of 12.82% with three different take-off plans. Meanwhile, it can reduce mission times by up to 66 min, compared to manual setup methods, which are about twice the safe endurance time.

3.2. Population Counting

3.2.1. Comparison with Few-Shot Approaches

The performance of counting effects of our method was compared with that of other methods. There are two common baselines, namely (1) always outputting the average object quantity of the training images; (2) always outputting the mid-value counts of the training images. The two popular algorithms are: feature reweighting (FR) few-shot detectors [36] and few-shot object detection (FSOD) [37]. The method adopted in this paper is more advance than others, as presented in Table 5.

3.2.2. Comparison with the Object Detector

One of the other counting methods is to use detectors to count objects after they have been detected, which is only applicable to certain categories of objects with pre-trained detectors. In general, training object detectors requires thousands of examples, so it is not a practical approach for general visual counting. The performance of the network on a validation set with pre-trained object detectors and on categorized subsets of a test set was evaluated in this paper. To be specific, we compared our method with Faster R-CNN [45] and Mask R-CNN [46]. These two pre-trained detectors can be found in the Detectron2 library [47]. The comparison results are shown in Table 6. It can be seen that our method outperforms target detectors that have been pre-trained with thousands of annotated examples.

3.2.3. Qualitative Analysis

Figure 7, Figure 8, Figure 9 and Figure 10 shows the visual prediction results of our algorithm. The input was the image to be tested and several bounding boxes of counting objects. In order to analyze the correctness of the density map expediently, we showed the input image and the density map superimposed. Figure 7 and Figure 8 show examples of Tibetan antelope images with different densities from the aerial view of the UAVs, which has proved the algorithm works well in the case of a flat background from an overhead view with more than 97% correct counting rate. Furthermore, the model still has an accuracy rate of more than 97% in the case of complex backgrounds (Figure 9), where both water and land are present in the overhead view and the color of the land is close to that of Tibetan antelope. Figure 10 shows an example of a Tibetan antelope herd image taken from the side view of a UAV flight, which has a correct counting rate of over 96%. The overhead and side view results demonstrate the robustness of the method utilized in this paper for counting Tibetan antelopes photographed from different viewpoints.

4. Discussion

The study demonstrates that daily monitoring in protected areas is greatly complemented by surveys of wildlife, which are conducted by small, low-cost, multiple rotary-wing UAVs in large areas. However, the influence of UAVs’ noise and appearance on wildlife should be considered during surveys. As we only surveyed and analyzed Tibetan antelopes, prior detailed testing on the UAV’s influence of noise and movements should be conducted when this method is applied to other animals. Moreover, our method extends the field of UAV application and provides an autonomous multi-robot solution for UAV path planning, which is suitable for any size of the survey area. It can also select the appropriate path according to the existing UAV parameters and allows them to optimize the take-off point of each flight.
The flexible deployment of multiple UAVs can significantly improve the path efficiency for completing area coverage missions and reducing manual survey time and manpower consumption, especially in remote and unpopulated areas. The results of multi-UAV trajectory planning (Table 2 and Table 3) have been visualized in Figure 11, Figure 12 and Figure 13. Regarding the total distance and the mission time, our method has more edges over the current multi-UAV CCPP methods. In one aspect, the shorter total distance means less energy consumption, which indicates a group of recharged UAVs can complete more flights when deployed at a UAV airport. In another aspect, a shorter mission time means that a quicker survey, allowing more surveys to be done within a window period to reduce da-ta errors and learn about timing-related issues. It can provide ecologists with more options for their survey methods. Compared to the method where only the total distance is done, both the total distance and mission time decline when these two indicators are optimized simultaneously. This is because the planner divides and distributes the entire survey route task to each with different take-off points evenly while acquiring the optimal solution for the area to survey on its own.
The few-shot object counting method can be used without pre-training the model via images of the target population. This improves the universality of the counting method to a certain extent, and it can be applied to many different fields. When using this method to test the photos of Tibetan antelope taken by UAVs, it has good applicability to images of different densities and perspectives. It can meet the requirements of population counting. A particular animal can be repeatedly photographed when images are stitched together because of the possible animal movements during a multi-UAV survey. Ecologists focusing on data research need to correct the survey results to ensure the accuracy of data results, which has become a necessary research subject [48]. Therefore, it is essential to reduce the mission time of such investigations. The information obtained directly from traditional ground surveys is often one-dimensional. The distribution of all animals may be visually linear (Figure 14) if interpreted from an image perspective when surface observation is conducted on the ground [49]. These factors often lead to significant errors in the results, which is not conducive to animal counting. When compared with traditional field survey methods by vehicle or on foot, UAV surveys have a better field of view (Figure 7, Figure 8 and Figure 9) and have reduced the impact of animal movements on the data to a large extent, which can improve the survey accuracy.
With the current development of commercial multi-rotary-wing UAVs and UAV airports, an untrained layman may acquire the skills to control a rotary-wing UAV quickly. The aerial animal counting system with multi-UAVs, developed by us, aims to enable scientists to focus on experimental design and data analysis rather than flight control skills. The survey deployment can be made before the survey window through a simple setup that determines the surveyed areas, the take-off areas, and the UAVs’ information. The current artificial intelligence methods can quickly assist humans in completing many of tedious counting tasks. The approach proposed in this paper can replace humans in performing uninterrupted unmanned count surveys, and we only need to ensure that UAVs will not influence the ecology of the survey areas.
However, more sorties of UAVs are often required for conducting simultaneous surveys of large areas, due to UAV endurance limitations. Besides, excessively high-flight altitudes, which means reducing the impact of UAV noise on animals, may not be conducive to animal counting in feature identification. It does harm to the applicability of the multi-UAV survey method because of camera resolution. UAV aerial surveys will also integrate more features by combining visible light and multispectral and thermal infrared cameras to improve information accuracy further.

5. Conclusions

A general and flexible CCPP method was provided in this paper for a group of UAVs to count animals. This planning algorithm enables scientists to employ multiple drones to conduct regular unmanned surveys in the uninhabited region in fragile environments with poor access. Moreover, multiple UAVs allow faster surveys with higher frequencies, lower costs, and greater resolution than helicopters and better leverage for rare observation opportunities by reducing wildlife disturbance from human activities. We have implemented our algorithms on a set of UAVs and conducted field experiments. The rapid and repeatable aerial surveys of large-scale environments demonstrated the utility of our system in facilitating population ecology. The results unveil that, in this paper, a herd containing approximately 20,000 Tibetan antelopes distributed over an area of more than 6 km2 was observed by a measuring system of four UAVs, which reduced the survey time from 4 to 6 days to about 2 h. Our method can complete aerial surveys in shorter trajectories and mission time. With this method, the trajectories and time were reduced by 2.5% and 26.3%, respectively, and the effectiveness increased by 5.3% when compared to sweep-style paths over an area of the same size.
Different from the target detector methods commonly used in current counting methods, the current work mainly focuses on one specific category at a time, such as people, cars, and plants. In this paper, we use a general few-shot adaptive matching counting method, which can count effectively without pre-training through the objects to be measured. We tested the photos of Tibetan antelope taken from the UAV, and the accuracy can reach more than 97%. The survey data will facilitate the development of animal protection policies. The results of the later survey can contribute to the development of animal protection policy, which is conducive to protecting and promoting ecological sustainability. However, animals may move in the process of multi-drone surveys, which makes photos of them possible to be duplicated when stitched together. In the future, reasonable corrections based on the survey results can further improve the accuracy of the survey. Other investigations that require fast aerial surveys under unpopulated circumstances can also benefit from our algorithms, such as surveys of other wildlife populations, forests, and shrublands.

Author Contributions

Methodology, H.S., R.H., H.Z. and T.L.; software R.H. and T.L.; Investigation, H.S. and H.Z.; data curation, H.Z. and T.L.; writing—original draft preparation visualization, H.S., R.H., H.Z. and T.L.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Funding of National Key Laboratory of Rotorcraft Aeromechanics (No. 61422202108), National Natural Science Foundation of China (No.51906103, No.52176009).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Manayeva, K.; Hoshino, B.; Igota, H.; Nakazawa, T.; Sumiya, G. Seasonal migration and home ranges of Tibetan antelopes (Pantholops hodgsonii) based on satellite tracking. Int. J. Zool. Res. 2016, 13, 26–37. [Google Scholar] [CrossRef] [Green Version]
  2. Lu, S.; Chen, F.; Zhou, J.; Hughes, A.C.; Ma, X.; Gao, W. Cascading implications of a single climate change event for fragile ecosystems on the Qinghai-Tibetan Plateau. Ecosphere 2020, 11, e03243. [Google Scholar] [CrossRef]
  3. IUCN Website. Available online: www.redlist.org (accessed on 13 July 2022).
  4. Lin, X. Conservation and Monitoring of Tibetan Antelopes in Hoh-Xil Nature Reserve; BP Conservation Programme BPCP: Beijing, China, 2014. [Google Scholar]
  5. Bollard, B.; Doshi, A.; Gilbert, N.; Poirot, C.; Gillman, L. Drone technology for monitoring protected areas in remote and fragile environments. Drones 2022, 6, 42. [Google Scholar] [CrossRef]
  6. Hu, J.; Wu, X.; Dai, M. Estimating the population size of migrating Tibetan antelopes Pantholops hodgsonii with unmanned aerial vehicles. Oryx 2020, 54, 101–109. [Google Scholar] [CrossRef] [Green Version]
  7. Le Maho, Y.; Whittington, J.D.; Hanuise, N.; Pereira, L.; Boureau, M.; Brucker, M.; Chatelain, N.; Courtecuisse, J.; Crenner, F.; Friess, B. Rovers minimize human disturbance in research on wild animals. Nat. Methods 2014, 11, 1242–1244. [Google Scholar] [CrossRef]
  8. Schaller, G.B.; Aili, K.; Xinbin, C.; Yanlin, L. Migratory and calving behavior of Tibetan antelope population. Acta Theriol. Sin. 2006, 26, 105. [Google Scholar]
  9. Hongping, Z.; Wei, H.; Dong, W.; Jie, J. A wildlife monitoring system based on Tianditu and Beidou: In case of the Tibetan antelope. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 259–262. [Google Scholar] [CrossRef] [Green Version]
  10. Christie, K.S.; Gilbert, S.L.; Brown, C.L.; Hatfield, M.; Hanson, L. Unmanned aircraft systems in wildlife research: Current and future applications of a transformative technology. Front. Ecol. Environ. 2016, 14, 241–251. [Google Scholar] [CrossRef]
  11. Linchant, J.; Lisein, J.; Semeki, J.; Lejeune, P.; Vermeulen, C. Are unmanned aircraft systems (UAS s) the future of wildlife monitoring? A review of accomplishments and challenges. Mammal Rev. 2015, 45, 239–252. [Google Scholar] [CrossRef]
  12. McEvoy, J.F.; Hall, G.P.; McDonald, P.G. Evaluation of unmanned aerial vehicle shape, flight path and camera type for waterfowl surveys: Disturbance effects and species recognition. PeerJ 2016, 4, e1831. [Google Scholar] [CrossRef] [Green Version]
  13. Pirotta, V.; Hocking, D.P.; Iggleden, J.; Harcourt, R. Drone observations of marine life and human–wildlife interactions off Sydney, Australia. Drones 2022, 6, 75. [Google Scholar] [CrossRef]
  14. Dajiang Airport. Available online: https://www.dji.com/cn/dock (accessed on 13 July 2022).
  15. Maza, I.; Ollero, A. Multiple UAV cooperative searching operation using polygon area decomposition and efficient coverage algorithms. In Distributed Autonomous Robotic Systems 6; Springer: Tokyo, Japan, 2007; pp. 221–230. [Google Scholar] [CrossRef]
  16. Cabreira, T.M.; Di Franco, C.; Ferreira, P.R.; Buttazzo, G.C. Energy-aware spiral coverage path planning for UAV photogrammetric applications. IEEE Robot. Autom. Lett. 2018, 3, 3662–3668. [Google Scholar] [CrossRef]
  17. Artemenko, O.; Dominic, O.J.; Andryeyev, O.; Mitschele-Thiel, A. Energy-aware trajectory planning for the localization of mobile devices using an unmanned aerial vehicle. In Proceedings of the 2016 25th international conference on computer communication and networks (ICCCN), Waikoloa, HI, USA, 1–4 August 2016; pp. 1–9. [Google Scholar] [CrossRef]
  18. Avellar, G.S.; Pereira, G.A.; Pimenta, L.C.; Iscold, P. Multi-UAV routing for area coverage and remote sensing with minimum time. Sensors 2015, 15, 27783–27803. [Google Scholar] [CrossRef] [Green Version]
  19. Nam, L.; Huang, L.; Li, X.J.; Xu, J. An approach for coverage path planning for UAVs. In Proceedings of the 2016 IEEE 14th international workshop on advanced motion control (AMC), Auckland, New Zealand, 22–24 April 2016; pp. 411–416. [Google Scholar] [CrossRef] [Green Version]
  20. Pérez-González, A.; Benítez-Montoya, N.; Jaramillo-Duque, Á.; Cano-Quintero, J.B. Coverage path planning with semantic segmentation for UAV in PV plants. Appl. Sci. 2021, 11, 12093. [Google Scholar] [CrossRef]
  21. Li, Y.; Chen, H.; Er, M.J.; Wang, X. Coverage path planning for UAVs based on enhanced exact cellular decomposition method. Mechatronics 2011, 21, 876–885. [Google Scholar] [CrossRef]
  22. Valente, J.; Del Cerro, J.; Barrientos, A.; Sanz, D. Aerial coverage optimization in precision agriculture management: A musical harmony inspired approach. Comput. Electron. Agric. 2013, 99, 153–159. [Google Scholar] [CrossRef] [Green Version]
  23. Yang, J.; Thomas, A.G.; Singh, S.; Baldi, S.; Wang, X. A semi-physical platform for guidance and formations of fixed-wing unmanned aerial vehicles. Sensors 2020, 20, 1136. [Google Scholar] [CrossRef] [Green Version]
  24. Shafiq, M.; Ali, Z.A.; Israr, A.; Alkhammash, E.H.; Hadjouni, M.; Jussila, J.J. Convergence analysis of path planning of multi-UAVs using max-min ant colony optimization approach. Sensors 2022, 22, 5395. [Google Scholar] [CrossRef] [PubMed]
  25. Batalin, M.A.; Sukhatme, G.S. Spreading out: A local approach to multi-robot coverage. In Distributed Autonomous Robotic Systems 5; Springer: Tokyo, Japan, 2002; pp. 373–382. [Google Scholar] [CrossRef] [Green Version]
  26. Hazon, N.; Kaminka, G.A. On redundancy, efficiency, and robustness in coverage for multiple robots. Robot. Auton. Syst. 2008, 56, 1102–1114. [Google Scholar] [CrossRef]
  27. Bähnemann, R.; Lawrance, N.; Chung, J.J.; Pantic, M.; Siegwart, R.; Nieto, J. Revisiting boustrophedon coverage path planning as a generalized traveling salesman problem. In Field and Service Robotics; Springer: Singapore, 2021; pp. 277–290. [Google Scholar] [CrossRef]
  28. Yu, J.; LaValle, S.M. Planning optimal paths for multiple robots on graphs. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 3612–3617. [Google Scholar] [CrossRef] [Green Version]
  29. Nedjati, A.; Izbirak, G.; Vizvari, B.; Arkat, J. Complete coverage path planning for a multi-UAV response system in post-earthquake assessment. Robotics 2016, 5, 26. [Google Scholar] [CrossRef] [Green Version]
  30. Shah, K.; Ballard, G.; Schmidt, A.; Schwager, M. Multidrone aerial surveys of penguin colonies in Antarctica. Sci. Robot. 2020, 5, eabc3000. [Google Scholar] [CrossRef]
  31. Fettermann, T.; Fiori, L.; Gillman, L.; Stockin, K.A.; Bollard, B. Drone surveys are more accurate than boat-based surveys of bottlenose dolphins (Tursiops truncatus). Drones 2022, 6, 82. [Google Scholar] [CrossRef]
  32. Ouattara, T.A.; Sokeng, V.-C.J.; Zo-Bi, I.C.; Kouamé, K.F.; Grinand, C.; Vaudry, R. Detection of forest tree losses in Côte d’Ivoire using drone aerial images. Drones 2022, 6, 83. [Google Scholar] [CrossRef]
  33. Fan, J.; Yang, X.; Lu, R.; Xie, X.; Li, W. Design and implementation of intelligent inspection and alarm flight system for epidemic prevention. Drones 2021, 5, 68. [Google Scholar] [CrossRef]
  34. Rominger, K.R.; Meyer, S.E. Drones, deep learning, and endangered plants: A method for population-level census using image analysis. Drones 2021, 5, 126. [Google Scholar] [CrossRef]
  35. Ranjan, V.; Wang, B.; Shah, M.; Hoai, M. Uncertainty estimation and sample selection for crowd counting. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020. [Google Scholar] [CrossRef]
  36. Kang, B.; Liu, Z.; Wang, X.; Yu, F.; Feng, J.; Darrell, T. Few-shot object detection via feature reweighting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8420–8429. [Google Scholar] [CrossRef] [Green Version]
  37. Fan, Q.; Zhuo, W.; Tang, C.-K.; Tai, Y.-W. Few-shot object detection with attention-RPN and multi-relation detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4013–4022. [Google Scholar] [CrossRef]
  38. Dan, S. Evolutionary Optimization Algorithms; Genetic Programming; John Wiley &Sons, Inc.: Hoboken, NJ, USA, 2013. [Google Scholar]
  39. Strubel, D. Coverage Path Planning Based on Waypoint Optimization, with Evolutionary Algorithms. Ph.D. Thesis, Université Bourgogne Franche-Comté, Besançon, France, 2019. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  41. Zhang, Y.; Zhou, D.; Chen, S.; Gao, S.; Ma, Y. Single-image crowd counting via multi-column convolutional neural network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 589–597. [Google Scholar] [CrossRef]
  42. Ranjan, V.; Sharma, U.; Nguyen, T.; Hoai, M. Learning to count everything. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3394–3403. [Google Scholar] [CrossRef]
  43. SPH Engineering, Universal Ground Control Station (UGCS). Available online: www.ugcs.com (accessed on 13 July 2022).
  44. Moura, L.D.; Bjørner, N. Z3: An efficient SMT solver. In Proceedings of the International conference on Tools and Algorithms for the Construction and Analysis of Systems, Budapest, Hungary, 29 March–6 April 2008; pp. 337–340. [Google Scholar] [CrossRef] [Green Version]
  45. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar] [CrossRef]
  47. Wu, Y.; Kirillov, A.; Massa, F.; Lo, W.; Girshick, R. Detectron2. 2019. Available online: https://github.com/facebookresearch/detectron2/ (accessed on 13 July 2022).
  48. Dunstan, A.; Robertson, K.; Fitzpatrick, R.; Pickford, J.; Meager, J. Use of unmanned aerial vehicles (UAVs) for mark-resight nesting population estimation of adult female green sea turtles at Raine Island. PLoS ONE 2020, 15, e0228524. [Google Scholar] [CrossRef] [PubMed]
  49. Wen, D.; Su, L.; Hu, Y.; Xiong, Z.; Liu, M.; Long, Y. Surveys of large waterfowl and their habitats using an unmanned aerial vehicle: A case study on the Siberian crane. Drones 2021, 5, 102. [Google Scholar] [CrossRef]
Figure 1. Survey area. Location of Zonag Lake and the migration route of Tibetan Antelopes.
Figure 1. Survey area. Location of Zonag Lake and the migration route of Tibetan Antelopes.
Drones 06 00196 g001
Figure 2. One of the calving grounds of Tibetan Antelopes.
Figure 2. One of the calving grounds of Tibetan Antelopes.
Drones 06 00196 g002
Figure 3. Few-shot adaptation and matching network.
Figure 3. Few-shot adaptation and matching network.
Drones 06 00196 g003
Figure 4. Planning workflow. (a) The coverage grides of survey area; (b) Planned flight path of UAV. The planned UAV path is a solid line, and the transit route is a dotted line. The drone logo represents the take-off and landing location; (c) The resulting photo coverage area allocated to the four UAVs with different starting points in (b); (d) The last stitched image.
Figure 4. Planning workflow. (a) The coverage grides of survey area; (b) Planned flight path of UAV. The planned UAV path is a solid line, and the transit route is a dotted line. The drone logo represents the take-off and landing location; (c) The resulting photo coverage area allocated to the four UAVs with different starting points in (b); (d) The last stitched image.
Drones 06 00196 g004
Figure 5. Survey routes. All routes to the survey area are shown here. The blue-green lines are routes, which are planned by the method in this paper, and are executed by four UAVs at different starting points. The blue area is the optional takeoff and landing area. The red no fly zone marks the mountainous area or flight interference area.
Figure 5. Survey routes. All routes to the survey area are shown here. The blue-green lines are routes, which are planned by the method in this paper, and are executed by four UAVs at different starting points. The blue area is the optional takeoff and landing area. The red no fly zone marks the mountainous area or flight interference area.
Drones 06 00196 g005
Figure 6. Comparison of route planning methods. (a) Ours; (b) Sweep-style path; (c) Polygon coverage. The red arrow is the flight path of backtracking.
Figure 6. Comparison of route planning methods. (a) Ours; (b) Sweep-style path; (c) Polygon coverage. The red arrow is the flight path of backtracking.
Drones 06 00196 g006
Figure 7. Low-density photos of Tibetan antelope taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 34, the predicted count is 35, and the accuracy is 97.1%.
Figure 7. Low-density photos of Tibetan antelope taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 34, the predicted count is 35, and the accuracy is 97.1%.
Drones 06 00196 g007
Figure 8. Medium-density photos of Tibetan antelope taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 53, the predicted count is 52, and the accuracy is 98.1%.
Figure 8. Medium-density photos of Tibetan antelope taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 53, the predicted count is 52, and the accuracy is 98.1%.
Drones 06 00196 g008
Figure 9. Medium-density photos of Tibetan antelope with complex background taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 45, the predicted count is 46, and the accuracy is 97.8%.
Figure 9. Medium-density photos of Tibetan antelope with complex background taken by UAVs above. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 45, the predicted count is 46, and the accuracy is 97.8%.
Drones 06 00196 g009
Figure 10. Network prediction results of medium density photos of Tibetan antelope taken by UAVs from the side. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 55, the predicted count is 53, and the accuracy is 96.4%.
Figure 10. Network prediction results of medium density photos of Tibetan antelope taken by UAVs from the side. (a) Input image and a few exemplar bounding boxes depicting the object to be counted; (b) Overlapping of the input image and the predicted density map; (c) Predicted density map. The value in the yellow box represents the predicted density value. The ground truth count is 55, the predicted count is 53, and the accuracy is 96.4%.
Drones 06 00196 g010
Figure 11. The total flight distance of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Figure 11. The total flight distance of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Drones 06 00196 g011
Figure 12. The total mission time of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Figure 12. The total mission time of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Drones 06 00196 g012
Figure 13. Trajectory efficiency of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Figure 13. Trajectory efficiency of a group of UAVs. (a) Set the total flight distance as the optimization parameter; (b) Set the weighted sum of total flight path and total mission time as optimization parameters.
Drones 06 00196 g013
Figure 14. Tibetan antelope was photographed on the ground. In the sun, four Tibetan antelopes overlap, so there seem to be only three. Therefore, artificial ground measurement is not conducive to accurate counting.
Figure 14. Tibetan antelope was photographed on the ground. In the sun, four Tibetan antelopes overlap, so there seem to be only three. Therefore, artificial ground measurement is not conducive to accurate counting.
Drones 06 00196 g014
Table 1. Technical parameters of the aircraft.
Table 1. Technical parameters of the aircraft.
Technical ParametersDescriptionTechnical ParametersDescription
Safe endurance time28 minSurvey area flight speed5 m/s
Maximum flight time55 minTransit distance flight speed10 m/s
Maximum takeoff weight9 kgMaximum flight altitude7000 m
Body weight with battery6.3 kgFlight height from ground200 m
AutopilotDJI flight control unitSize (folded, including paddles)430 × 420 × 430 mm (L × W × H)
Table 2. Comparison of route efficiency. Bold entries indicate the most efficient set of routes.
Table 2. Comparison of route efficiency. Bold entries indicate the most efficient set of routes.
MethodParametersZone 1Zone 2Zone 3Zone 4Zone 5Zone 6Average
OursLength5.9 km4.7 km4.4 km3.6 km2.5 km1.5 km-
N vertex 132109102826256-
N path 136112104856561-
η path 0.9710.9730.9810.9650.9540.9180.960
SweepLength6.5 km5.5 km4.9 km4.5 km3.4 km1.8 km-
Backtrack0.57 km0.98 km0.37 km0.63 km0.52 km0.08 km-
η path 0.9120.8220.9240.8600.8470.9560.887
PolygonLength6.8 km4.7 km4.5 km4.3 km2.9 km1.6 km-
Backtrack0.80 km0.51 km0.38 km0.61 km0.33 km0.17 km-
η path 0.8820.8910.9160.8580.8860.8940.888
Table 3. Comparison of route parameters (distance, efficiency, and mission time). Set the total flight distance as the optimization parameter. Bold entries indicate the most efficient set of routes. The data are the average of 50 experiments.
Table 3. Comparison of route parameters (distance, efficiency, and mission time). Set the total flight distance as the optimization parameter. Bold entries indicate the most efficient set of routes. The data are the average of 50 experiments.
Take-Off PlanMethodDistanceBacktrack η path Efficiency ImprovementMission TimeTime Reduction
Two different take-off points within the area restrictionsManual setting94.2 km20.8 km0.779Control group2.6 hControl group
Sweep94.1 km20.8 km0.7790.00%2.6 h0%
Polygon94.0 km20.7 km0.7800.13%2.4 h7.69%
Ours94.0 km20.7 km0.7800.13%2.4 h7.69%
Four different take-off points within the area restrictionsManual setting94.2 km20.8 km0.779Control group2.6 hControl group
Sweep91.8 km18.4 km0.7990.00%1.9 h26.92%
Polygon91.6 km18.3 km0.8000.13%1.9 h26.92%
Ours88.7 km15.5 km0.8255.91%1.7 h34.62%
Four different take-off points without the area restrictionsManual setting95.8 km22.4 km0.766Control group2.6 hControl group
Sweep92.4 km16.2 km0.7953.79%1.8 h30.77%
Polygon89.0 km15.6 km0.8257.70%1.8 h30.77%
Ours78.0 km4.7 km0.94022.72%1.7 h34.62%
Table 4. Comparison of route parameters (distance, efficiency, and mission time). Set the weighted sum of total flight path and total mission time as optimization parameters. Bold entries indicate the most efficient set of routes. The data are the average of 50 experiments.
Table 4. Comparison of route parameters (distance, efficiency, and mission time). Set the weighted sum of total flight path and total mission time as optimization parameters. Bold entries indicate the most efficient set of routes. The data are the average of 50 experiments.
Take-Off PlanMethodDistanceBacktrack η path Efficiency ImprovementMission TimeTime Reduction
Two different take-off points within the area restrictionsManual setting94.2 km21.6 km0.771Control group2.6 hControl group
Sweep95.0 km21.7 km0.7720.13%2.5 h3.85%
Polygon95.0 km21.3 km0.7760.65%2.3 h11.54%
Ours94.4 km20.8 km0.7801.17%2.2 h15.38%
Four different take-off points within the area restrictionsManual setting94.2 km20.8 km0.779Control group2.6 hControl group
Sweep91.8 km18.5 km0.7990.00%1.9 h26.92%
Polygon92.7 km19.0 km0.7952.05%1.7 h34.62%
Ours89.5 km16.1 km0.8205.26%1.4 h46.15%
Four different take-off points without the area restrictionsManual setting95.8 km22.4 km0.766Control group2.5 hControl group
Sweep90.8 km21.2 km0.86611.55%1.5 h42.31%
Polygon80.3 km7.0 km0.91319.19%1.4 h46.15%
Ours77.4 km4.1 km0.94723.63%1.2 h53.85%
Table 5. Comparing our method to two simple baselines (mean, median) and two popular algorithms (FR few-shot detector, FSOD few-shot detector). These are few-shot methods that have been adapted and trained for population counting. Our method has the lowest MAE and RMSE on Tibetan antelope dataset. Bold entries indicate the most efficient method.
Table 5. Comparing our method to two simple baselines (mean, median) and two popular algorithms (FR few-shot detector, FSOD few-shot detector). These are few-shot methods that have been adapted and trained for population counting. Our method has the lowest MAE and RMSE on Tibetan antelope dataset. Bold entries indicate the most efficient method.
MethodMAERMSE
Mean48.7146.7
Median47.7152.4
FR few-shot detector41.7141.0
FSOD few-shot detector35.5140.7
Ours23.298.4
Table 6. Comparing our method with pre-trained object detectors. Bold entries indicate the most efficient method.
Table 6. Comparing our method with pre-trained object detectors. Bold entries indicate the most efficient method.
MethodMAERMSE
Faster R-CNN36.579.2
Mask R-CNN35.279.8
Ours23.945.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, R.; Zhou, H.; Liu, T.; Sheng, H. Multi-UAV Collaboration to Survey Tibetan Antelopes in Hoh Xil. Drones 2022, 6, 196. https://doi.org/10.3390/drones6080196

AMA Style

Huang R, Zhou H, Liu T, Sheng H. Multi-UAV Collaboration to Survey Tibetan Antelopes in Hoh Xil. Drones. 2022; 6(8):196. https://doi.org/10.3390/drones6080196

Chicago/Turabian Style

Huang, Rui, Han Zhou, Tong Liu, and Hanlin Sheng. 2022. "Multi-UAV Collaboration to Survey Tibetan Antelopes in Hoh Xil" Drones 6, no. 8: 196. https://doi.org/10.3390/drones6080196

APA Style

Huang, R., Zhou, H., Liu, T., & Sheng, H. (2022). Multi-UAV Collaboration to Survey Tibetan Antelopes in Hoh Xil. Drones, 6(8), 196. https://doi.org/10.3390/drones6080196

Article Metrics

Back to TopTop