Next Article in Journal
Effects of Injection Rate Shape on Performance and Emissions of a Diesel Engine Fuelled by Diesel and Biodiesel B20
Previous Article in Journal
Implementing a GIS-Based Digital Atlas of Agricultural Plastics to Reduce Their Environmental Footprint; Part I: A Deductive Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Task Allocation Strategy of the UAV Swarm Based on Multi-Discrete Wolf Pack Algorithm

1
School of Computer and Information, Hohai University, Nanjing 211100, China
2
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, Hohai University, Nanjing 211100, China
3
Suma Technology Co., Ltd., Kunshan, Suzhou 215300, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(3), 1331; https://doi.org/10.3390/app12031331
Submission received: 21 December 2021 / Revised: 22 January 2022 / Accepted: 24 January 2022 / Published: 26 January 2022

Abstract

:
With the continuous development of artificial intelligence, swarm control and other technologies, the application of Unmanned Aerial Vehicles (UAVs) in the battlefield is more and more extensive, and the UAV swarm is increasingly playing a prominent role in the future of warfare. How tasks are assigned in the dynamic and complex battlefield environment is very important. This paper proposes a task assignment model and its objective function based on dynamic information convergence. In order to resolve this multidimensional function, the Wolf Pack Algorithm (WPA) is selected as the alternative optimization algorithm. This is because its functional optimization of high-dimensional complex problems is better than other intelligent algorithms, and the fact that it is more suitable for UAV swarm task allocation scenarios. Based on the traditional WPA algorithm, this paper proposes a Multi-discrete Wolf Pack Algorithm (MDWPA) to solve the UAV task assignment problem in a complex environment through the discretization of wandering, calling, sieging behavior, and new individual supplement. Through Orthogonal Experiment Design (OED) and analysis of variance, the results show that MDWPA performs with better accuracy, robustness, and convergence rate and can effectively solve the task assignment problem of UAVs in a complex dynamic environment.

1. Introduction

Due to its high flexibility and wide adaptability, the UAV swarm has more and more extensive application potential and has received great attention at home and abroad [1]. With the continuous development of technologies such as artificial intelligence, swarm control, and collaborative interaction, the application of Unmanned Aerial Vehicles (UAVs) in the battlefield is more and more extensive, and the UAV swarm is increasingly playing a prominent role in the future of warfare [2]. Among them, the purpose of UAV task assignment is to find a distribution plan that optimizes the overall target effect, which is of great significance in the application scenario of UAV swarm combat. In general, it is a sophisticated combinatorial optimization problem with complex constraints in terms of task precedence and coordination, timing constraints, and flyable trajectories. In the real environment, it is difficult to get the optimal solution to this problem, especially for the uncertain complex dynamic environment [3]. In essence, the task allocation problem of the UAV swarm is also a large-scale, high-dimensional, and complex function optimization problem. So it is important to work out how to optimize this complex combinatorial optimization problem. At present, relevant scholars have done a lot of research on the task allocation problem of the UAV swarm.
UAV task assignment methods are mainly divided into centralized task assignment and distributed task assignment [1]. Among them, the distributed task assignment algorithm has the characteristics of good fault tolerance and strong scalability. However, due to the limited grasp of the overall information of the battlefield by each decision maker, it is often difficult to guarantee the quality of the solution. Given that, in the actual combat environment, more precise assignment plans are often required to achieve the best results from battlefield operations. Therefore, the distributed task allocation plan is not very applicable in the process of optimizing the best combat plan. Corresponding to the distributed task assignment method, is the centralized task assignment method. The control center has the global information and makes the optimal allocation plan on behalf of the overall interests and then notifies each UAV. The centralized task assignment method has the potential to generate the global optimal solution [4], which mainly includes optimization methods [5,6,7,8] and heuristic methods. The optimization algorithms can flexibly adjust the constraint conditions to solve the actual problem and have the theoretical optimal solution; however, the scale should not be too large, and it is generally used for the discrete task allocation of the UAV swarm.
The heuristic methods are independent of the mathematical performance of the problem, they do not have strict requirements on the initial value and can effectively deal with high-dimensional complex optimization problems, including dynamic lists [9], clustering algorithms [10], and smart algorithms. Among them, the characteristics of biological groups in decentralization, neighboring individual information interactions, and overall self-organization are consistent with the pursuit of the UAV swarm. Therefore, the application of intelligent algorithms in UAV swarm task assignment is relatively common. It mainly includes intelligent algorithms such as the particle swarm algorithm (PSO) [11,12], genetic algorithm (GA) [13,14], wolf pack algorithm (WPA) [15], and so on. PSO is easy to implement and has fewer control parameters. However, its problems are low search efficiency and fast premature birth [16], and as the dimensionality of the problem increases, the convergence speed of the algorithm will inevitably be affected. In recent years, there has been a lot of research on evolutionary algorithms to improve the search ability of the entire search stage. Although it has a good search performance, it cannot avoid sticking into local optima, thereby obtaining the inferior solutions [14]. Most of these algorithms have global optimization capabilities, which are very helpful for the small-scale optimization of complex functions. However, in the UAV swarm combat scenario with high real-time dynamics and complexity, the above algorithms have certain shortcomings for the optimization of high-dimensional complex functions.
Among them, the wolf pack algorithm is a new bionic intelligent algorithm proposed by Wu et al. [17] in 2013. With clever division of labor, wolves can complete a series of group activities such as complex cooperative hunting, feeding cubs, and territorial maintenance [18], showing high efficiency, flexibility, and dynamic adaptation. Compared with other algorithms, the wolf pack algorithm is a random probability search algorithm, which can quickly find the optimal solution with a high probability. WPA also has parallelism, which can search from multiple points at the same time without affecting each other, thus improving the efficiency of the algorithm. WPA has better global convergence and computational robustness and is especially suitable for solving complex functions with high dimensions and multiple peaks [19]. Therefore, the processing and adaptability of WPA has greater advantages in handling the task allocation of the UAV swarm in a dynamic and complex battlefield environment. However, the wolf pack algorithm has disadvantages such as slow convergence speed and the fact that it can easy to fall into local optimum. Many scholars have done a lot of research on how to optimize the wolf pack algorithm. The optimization of the wolf pack algorithm can be divided into three categories.
(1)
Enhance the global search ability;
Literature [20] aims to improve the optimization quality, stability and convergence speed of the wolf pack algorithm. All wolves in the pack are allowed to compete as the leading wolf to improve the probability of finding the global optimal solution, thus enhancing the search ability; In order to promote information exchange between artificial wolves, improve wolves’ grasp of global information and enhance their exploration ability, reference [21] introduces drift operators and wave operators into wolves’ reconnaissance behavior and summation behavior. An adaptive dynamic adjustment factor strategy is proposed for besieging behavior. Literature [22] discusses an optimization strategy of WPA based on opposition learning, which not only enhances the local search ability of WPA under normal conditions, but also increases the population diversity under catastrophic conditions. In order to study the parameter optimization problem of complex nonlinear functions, an adaptive variable step size chaotic wolf optimization algorithm (CWOA) is proposed in literature [23]. Combining the adaptive variable step search strategy with the chaos optimization strategy, gives the algorithm high robustness and global search ability. In order to accelerate the convergence of WPA algorithm and enhance its search ability, the differential evolution (DE) elite-set strategy is adopted in literature [24]. In literature [25], on the premise of ensuring the solution accuracy, optimization time is reduced. A jump raid strategy is proposed, which directly airlifts half of the wolves to the lead wolf, which improves the performance of the raid process and speeds up the convergence speed of the algorithm.
(2)
The balance between local search ability and global search ability;
In order to solve the disadvantages of WPA, the strategy of opposition learning is adopted to initialize the population in literature [26], so as to enhance the population diversity in the process of global search. At the same time, genetic algorithm is used to select the head wolf to avoid local optimization. This combined optimization strategy harmonizes the global exploration capability and local development capability well. Literature [27] proposes a quantum-excited wolf pack algorithm based on quantum coding: quantum rotation and quantum collapse. The first step makes the population move to the global optimal value, the second step helps to avoid individuals falling into the local optimal value. The global search capability and local search capability can be balanced.
(3)
Optimization of high-dimensional complex problems;
In order to study the parameter optimization problem of complex nonlinear functions, an adaptive variable step size Chaotic Wolf Optimization Algorithm (CWOA) is proposed in literature [23]. Combining the adaptive variable step search strategy with chaos optimization strategy, the algorithm has high robustness and global search ability. For the optimization of high-dimensional functions, a new wolf pack algorithm is proposed in literature [28], which can accurately calculate the optimal value of high-dimensional functions. Firstly, the chaotic inverse initialization is designed to improve the quality of the initial solution. Secondly, interference factors are added in the search process to enhance the search ability of wolves, and the adaptive step size is designed to enhance the global search ability of wolves, which effectively prevents wolves from falling into local optimal. Literature [29] proposes a Multi-population Parallel Wolf Algorithm (MPPWPA) in order to solve the problem of too long convergence time of the wolf algorithm in the optimization of high-dimensional complex problems. By dividing the wolf population into multiple sub-populations and independently optimizing the sub-populations at the same time, the size of the wolf population is reduced. Literature [30] proposes a combination of the principle of particle swarm optimization and genetic algorithm to optimize the wolf pack algorithm and solve the task allocation problem of the UAV swarm with a faster convergence speed. The improved algorithm shows good search quality in high-dimensional space. Literature [31] applies the improved Wolf Pack Search (WPS) algorithm to calculate the quasi-optimal trajectory of rotorcraft UAV in a complex three-dimensional space (including real and false three-dimensional space) and to adopt a multi-objective cost function. In the process of path planning, some concepts of the genetic algorithm are applied to implement the WPS algorithm.
The optimization strategies can effectively improve the performance of WPA, but there are few literature studies on the wolf pack algorithm dealing with complex discrete problems, such as whether it could be better applied to discrete issues such as UAV task allocation.
In order to enable the wolf pack algorithm to better solve the discrete problem, this paper proposes a multiple discretized wolf pack algorithm (MDWPA). The contributions of this paper can be summarized as follows.
(1)
In order to solve the shortcomings of traditional WPA in solving discrete problems, a multiple discrete optimization strategy is proposed to improve the traditional wolf pack algorithm. The improved WPA is compared with other intelligent algorithms in the simulation experiment, and its performance is far superior to other algorithms.
(2)
Looking into the multi-UAV intelligence collection scenario in a dynamic battlefield environment, a corresponding multi-UAV task allocation model is established and, at the same time, resource consumption and mission revenue are considered. The objective function is constructed to solve the complex dynamic task assignment problem.

2. Task Assignment Model

In combat scenarios, in order to better attack enemy targets, UAVs are often required to collect enemy information and other intelligence in the enemy area, and after mastering comprehensive intelligence, they are required to perform tasks such as precision strikes on the enemy area. At the initial moment, the UAV departs from the base, is required to successfully detect the enemy target and obtain useful information, and execute the task sequence in turn before returning to the base. All UAVs are not initially dispatched, some UAVs will be kept at the base for emergencies.

2.1. UAV and Task Modeling

This article models the above actual combat scenarios, which can be expressed as a four-tuple U , T , M , C , where U is the set of UAVs, T is the set of targets in the combat area, and M is the set of tasks on the target. C is the set of constraints in the problem model. Based on the literature [32], this paper further considers the problem of the cooperative task allocation of UAVs in actual combat area scenarios, further considers multiple constraints and optimization objectives, and constructs a cooperative multi-task assignment model for UAVs in combat scenarios. Table 1 shows the attributes of three objects in the model.
The definition of the UAV system has N u reconnaissance UAVs. In this paper, for any UAV u i i   1 , N u , given its constant flight speed v i , maximum navigation distance d i m a x and maximum flight time τ i   m a x .
Due to the dynamic update of information in the battlefield environment, the intelligence information of each target point may be updated after reconnaissance. Therefore, each target point T j   j   1 , N t involved in this article contains at least one reconnaissance mission. Among them, define N t as the number of targets. I j is the reconnaissance time required to complete the intelligence collection work at the target point T j .
On the premise of ensuring the validity of the model, in order to simplify the model and reduce the difficulty of solving the model, this article puts forward the following reasonable assumptions about the model:
(1)
The number and approximate location coordinates of all targets are known to the command center of UAV base;
(2)
The UAV flies at a constant speed, and the flying heights of the UAVs are not consistent in order to prevent collisions between the UAVs;
(3)
The same task can be completed by multiple UAVs, and it is assumed that the UAVs working in coordination reach the target point at the same time, and there is no waiting time for each other.
Based on the above symbolic interpretation and model assumptions, the scheme of cooperative multi-task allocation for UAVs can be expressed as a vector, as shown in Table 2. The vector has three rows, the first row represents the target sequence, the second row represents the task sequence on the target (mainly multiple intelligence searches, taking two searches as an example), and the third row represents the result of task assignment.
In Table 2, M K represents the K th mission of target T j and it is assigned to u i , so the third row X i j   i   1 , N u is 1. For other i i i ,   i   1 , N u ,   X i j is 0, which means that T j is not assigned to u i .

2.2. Objective Function

In order to make the plan of task allocation more scientific and reasonable, this paper constructs a task allocation model under multiple constraints such as limited resources. The objective function of the constructed model mainly considers the minimization of resource consumption and the optimization of the task completion effect. Therefore, we construct a cost function and an evaluation function of task completion effect to quantify resource consumption and task completion effects respectively.

2.2.1. Cost Function

Most studies only consider the traveled distance or time in the evaluation of resource consumption, which makes the obtained task assignment result one-sided. To address this defect, we jointly consider the impact of distance and time on resource consumption. Following this idea, the distance cost and time cost of the UAV system is constructed as follows.
a. Distance Cost
The distance cost of the UAV system is denoted by d t o t a l , which can be achieved by
d t o t a l = i = 1 N u j = 1 N t d j k j ^ i x i j k
where d j k j ^ i represents the flight distance of the UAV u i from the target point t j to the target point t j ^ ; where x i j k is a binary variable that represents the k execution of target t j by UAV u i , and the specific expression is shown as follows.
x i j k = 1 ,   i f   M k   o n   t j   i s   a s s i g n e d   t o   u i   0 ,   o t h e r s
If there are no obstacles in the flight path from the task point t j to the next task point t j ^ , the distance d j k j ^ can be easily obtained according to the calculation Formula (3).
d j k j ^ i = x t j k u i x t j ^ k u i 2 + y t j k u i y t j ^ k u i 2
where [ x t j k u i ,   y t j k u i ] and [ x t j ^ k u i , y t j ^ k u i ] represent the position coordinates of the UAV u i in executing the task M k of the target t j and the task M k of the target t j ^ , respectively.
b. Time Cost
In order to ensure that the tasks are completed as quickly as possible, the model also considers the completion time of all tasks as another optimization index, where the shortest completion time of all tasks can ensure that the entire task can be completed quickly. The time cost of the UAV system to perform the task is represented by τ t o t a l .
τ t o t a l = m a x τ i j k e + τ i j k f
where τ i j k e and τ i j k f represents the time that u i execute M k on t j and the flight time of the UAV u i from t j to the next target point t j ^ , respectively. The τ t o t a l   is the maximum sum of τ i j k e and τ i j k f . τ i j k e is a predetermined parameter, and the calculation formula of τ i j k f is shown in Formula (5).
τ i j k f = d j k , j ^ i v i
This paper defines a cost function to quantify the resource consumption of the UAV system. Distance and time are the main indicators to measure resource consumption. Therefore, the distance cost and time cost of the UAV system should be the main components of the cost function. However, the evaluation indicators of distance and time are in different numerical dimensions, the direct addition of these two indicators affects the fairness of the cost function. Therefore, these two indicators should be standardized to eliminate dimensional effects. Based on the above analysis, the cost function of the UAV system can be obtained as followed.
f c o s t = d t o t a l i = 1 N u d i m a x + τ t o t a l max i 1 , N u τ i m a x
The first and second terms of Formula (6) measure resource consumption in terms of distance and time, respectively. In order to eliminate the dimensional influence between distance and time indicators, divide the different indicators by their respective upper bounds (the maximum flight distance of the UAV system i = 1 N u d i m a x and the endurance time of the UAV system max i 1 , N u τ i m a x ) for normalization.

2.2.2. Task Revenue Function

In a battle scenario, mastering the enemy’s important intelligence is the key to victory. Assuming that, at the initial moment, the UAV only knows the approximate location of the target. When the UAV detects the target, the accurate location information of the target P j = x j , y j , j 1 , N t and the initial value of the target v a l u e t j can be obtained. This article uses the initial value of the target to express the importance of intelligence.
In actual scenarios, the initial value of the task tends to decrease as the waiting time increases, so this paper considers that the effect of task execution will decrease with the increase of time. That is to say, the function value is inversely proportional to the time. To eliminate the dimensional influence between costs and benefits, we also normalize the benefits by dividing the actual revenue by the sum of the initial values of all the tasks. The specific operation is shown in Formula (7).
f r e w = i = 1 N u j = 1 N t x i j q · v a l u e t j · φ t j = 1 N t v a l u e t j
where x i j q is a binary value, indicating whether UAV u i is assigned to a task t j . Only when μ q is an attack task, the value of x i j q is 1; v a l u e t j is the importance of the target at the initial moment, and φ t   0 , 1 decreases with time, indicating that the value obtained by destroying the target varies with time, and the calculation formula of   φ t is as follows.
φ t = e ρ t
where   ρ represents the rate of decline in revenue, the larger the ρ , the faster the revenue from destroying the target decreases over time, and vice versa.
However, in a real scenario, enemy information is dynamic and constantly updated. UAVs may be required to perform repeated or multiple intelligence searches at the same target point to ensure effectiveness and accuracy. But in a dynamic mission scenario, the optimal solution is to cover as many target points as possible in a short time. Therefore, UAV should try to select unsearched target points, rather than repeatedly search the same mission point. Therefore, we adjust the function of task revenue according to the actual situation. When UAVs conduct repeated searches of the same target, the information they find will largely overlap, so the value of the mission should be halved.
The change of the function of task value is shown in Formula (9).
v a l u e t j = ( 1 2 ) k 1 v a l u e t j
where k is the number of UAV searches for intelligence on task point t j .

2.3. Task Allocation Model

Comprehensive consideration of the UAV body constraints, task requirements, cost consumption, effects of task completion and other factors, a task allocation model under multiple constraints is established.
max f = α f r e w f c o s t
s . t .   i = 1 N u j = 1 N t k = 1 3 x i j k = 3
i   1 , N u , d i d i m a x , τ i τ i m a x
i = 1 N u τ i j k e X i j I j  
where α is the binary flag, which can be expressed as follows.
α = 1 ,   i f 9 ~ 14 a r e   a l l   s a t i s f i e d 0 , o t h e r s
Formula (11) is used to ensure that each task must be completed; in Formula (12), d i and τ i , respectively, represent the total flight distance and total time consumed by the UAV u i to complete all assigned tasks, neither of which can exceed the prescribed upper limit.
Formula (13) is to ensure that each task is successfully completed, and the resources consumed by the UAV performing the task need to meet the conditions required for the completion of the task.
In the formulated task allocation model, the cost function and the function of task benefit evaluation are used as the numerator and denominator of the objective function, respectively. By maximizing the objective function proposed, the contradictory variables of task revenue and resource consumption are optimized at the same time. Meanwhile, constraints such as limited resources and task priority are introduced to improve the applicability of the model. The introduction of multiple constraints makes the established model more accurate and further improves the mission execution efficiency of the UAV system.

3. The Wolf Pack Algorithm

3.1. Traditional Wolf Pack Algorithm

The wolf pack algorithm is a relatively new swarm intelligence algorithm based on the wolf pack behavior mechanism [17]. The entire wolf pack algorithm mainly contains two mechanisms, namely, the head wolf generation mechanism, and the wolf pack update mechanism, and three behaviors, namely, wandering behaviors, summoning behaviors, and siege behaviors. The following gives a brief description of the two mechanisms and three behaviors in the wolf pack algorithm to prepare for the specific analysis of the improved wolf pack algorithm.

3.1.1. The Head Wolf Generation Mechanism

The head wolf generation mechanism means that, according to the “victor is king” rule, when a more powerful leader appears in the wolves to prey on prey, a more capable wolf will become a new head wolf. The prey odor concentration perceived by the artificial wolf is Y = f X , where Y is the value of the objective function. That is, when the fitness of the i t h wolf in the pack Y i (prey concentration at its location) is greater than the fitness Y l e a d of the head wolf, this wolf becomes the new head wolf.

3.1.2. Wolf Pack Update Mechanism

The wolf pack update mechanism refers to removing the R   wolf pack individuals with the worst value of objective function, and then adding R wolf pack individuals at the same time.
R = r a n d i ( N 2 β , N β )
among them, β   is the population update scale factor; N is the number of wolves.

3.1.3. Wandering Behavior

Except for the head wolf, the most elite S n u m wolves are called the detection wolf. If α is the scale factor of the detection wolf, the number of the detection wolf S n u m is a random number between N 1 + α , N α . The detective wolf walks one step in the surrounding h directions with its walking step s t e p a , and respectively senses the odor concentration of the location. Every time it travels in one direction, the odor concentration is sensed and recorded, and then it returns to the original coordinates. If there is a direction where the odor concentration is greater than the original coordinate, then it moves to the direction with the largest odor concentration. After detective wolf i advances in the p p = 1 , 2 , , h direction, its position in the d t h dimension is shown as follows.
Z i d p = Z i d + λ s t e p a d
among them, Z i d is the coordinates of the i t h wolf in the d-dimensional space;   λ is a random number in the range 1 , 1 . Compare the updated fitness of the wolf detection with the fitness of the head wolf. If Y i > Y l e a d , then update Y l e a d = Y i , and detective wolf i will replace the head wolf and re-initiate the call; if Y i < Y l e a d , continue the wandering behavior.

3.1.4. Summoning Behavior

As the wolves enter the summoning phase, the fierce wolves summoned by the head wolf rush towards the coordinates of the head wolf with a larger step length s t e p b , the position change of the fierce wolf i in the d-dimensional space in the k + 1 iteration is shown in Formula (17).
Z i d k + 1 = Z i d k + s t e p b d Z l e a d d k Z i d k / Z l e a d d k Z i d k
among them, Z i d k is the current position of the wolf; Z l e a d d k is the coordinate of the k t h generation wolf in the d-dimensional space. In the process of running, if the adaptability of the wolf   i is greater than that of the head wolf, the wolf i becomes a new generation of the head wolf, the position of the head wolf is updated, and the calling behavior is initiated; if the adaptability of the wolf i is less than the wolf, and the distance d i s between wolf i and the head wolf is less than d n e a r , they switch to siege behavior; supposing the value range of the d t h variable to be optimized is m i n d , m a x d , then the judgment distance d n e a r can be determined by the Formula (18).
d n e a r = 1 / D ω d = 1 D m a x d m i n d
among them, ω is the distance determination factor, and its different values will affect the convergence speed of the algorithm. m a x d and m i n d are the upper and lower bounds, respectively.

3.1.5. Sieging Behavior

When the wolves are close enough to the prey, the fierce wolves initiate a siege on the prey. At this time, the updated position Z i d k + 1 of the wolf can be expressed by Formula (19).
Z i d k + 1 = Z i d k + λ s t e p c d Z l e a d d k Z i d k
among them, λ is a random number in the range 1 , 1 ; s t e p c d is the siege step length. If the fitness of the wolf i after the siege behavior is greater than the fitness of the original position, its position will be updated; otherwise, the wolf will keep the original position unchanged.
Assuming that the value range of the d t h variable to be optimized is m i n d , m a x d , then in the wolf pack algorithm process, three steps are involved: walking step size s t e p a , rush step size s t e p b , siege step size s t e p c , the step size in the d-dimensional space has the following relationship.
s t e p a d = s t e p b d 2 = 2 s t e p c d = m a x d m i n d / S
where S is the step size factor and represents the fineness of the artificial wolf searching for the optimal solution in the solution space.
The optimization process of the WPA algorithm is divided into wandering, calling, and sieging. In the WPA, the exploring wolves select H directions and explore them. There is no communication between them and so there will be repetition in exploring space, so there is a lack of necessary information communication between artificial wolves, and the algorithm is not global enough. At present, it is difficult to find the global optimal solution for the wolf pack algorithm. Therefore, the continuous exploration of swarm intelligence optimization algorithm has become a hot topic for many scholars.

3.2. Proposed Algorithm

The wolf pack algorithm is mainly composed of three behavior mechanisms: wandering, summoning and sieging. Through these three behaviors, the position of each wolf is gradually optimized and the global optimum is finally found. The traditional wolf pack algorithm has strong convergence and global search ability and is not easy to fall into invalid search. WPA usually has good performance for continuous problems, but is not good at discrete problem optimization. The task assignment problem of UAV is a discrete problem, in which the variables of each dimension are integers belonging to a set. Therefore, based on the WPA rules, it is necessary to improve the WPA to match the integer discrete characteristics of task assignment. In order to improve solution accuracy and efficiency of the algorithm, this paper uses the multi-step discrete optimization strategy to improve the WPA, which mainly includes individual coding, wandering strategy, calling strategy, siege strategy, and recruitment of new individuals. The details of the MDWPA are as follows and the algorithm is shown in Algorithm 1.

3.2.1. Individual Coding and Initialization

In the improved algorithm, a hybrid coding method based on task assignment and path planning is adopted for mixed variables. The position vectors of individuals contain discrete variables representing the results of task assignment. In order to generate a feasible task allocation plan more effectively, according to the characteristics of UAV task allocation problem, the algorithm adopts a mixed variable coding method based on task allocation and path planning and designs an appropriate constraint processing method to initialize the position vector of wolf individuals according to the coding method. At the same time, new individuals are generated by a recombination method based on structure learning to further improve the quality of the population. In addition, in the process of population renewal, in order to obtain more excellent individuals, the improved algorithm adopts the mechanism of coevolution, that is, a coevolutionary population and the original population compete to generate new individuals.
At present, the existing coding methods can only describe the final assignment scheme but cannot represent the executing sequence of tasks, that is, the path planning result of the task performed by UAV. Therefore, in order to represent both the result of task assignment and the result of UAV’s path planning during task execution, a hybrid variable coding method based on task assignment and path planning is proposed in this paper. This method codes according to the specific situation of the UAV in the execution of the task. The code of each individual contains two parts, that is, P =   M N ,   U N , where the first part M N is a discrete variable, recording the task number, and the second part U N is a discrete variable, which records the number of UAVs performing corresponding tasks.
Figure 1 shows the code of an individual corresponding to the model of Section 2. It contains 6 targets, there are 3 UAVs, numbered 1–3. The two lines of the position vector of the individual wolf are used to represent the task assignment result. For example, the targets T1, T4 and T6 are assigned to the UAV U 1 for execution. At this time, this is the corresponding task assignment plan vector: X 1 k = 1     k       1   6     .
In addition, in order to show the results of the UAV’s path planning, the order in which the task numbers appear in the position vector represents the order in which the UAV performs tasks. The blue curve in Figure 1 indicates that UAV   U 1 performs tasks in sequence T 1 , T 6 . The orange curve indicates that the execution tasks of the UAV   U 2 are T 2 ,   T 3 . The green curve indicates that the execution tasks of the UAV U 3 are T 4   and T 5 , in turn. Because the coding method uses a hybrid variable coding method based on task allocation and path planning, this variable can not only efficiently convert the position vector of the individual wolf pack into a task allocation plan of the model, but can also fully consider the path uncertainty during the execution of the UAV task and use the order of tasks in the position vector to explicitly show the order in which UAVs perform tasks. This coding method can generate a definite task allocation plan and the path planning results of the UAV when the individual is initialized.
In the actual initialization operation in the UAV task allocation, we first initialize the task number in the first row. In order to provide a global orientation to the algorithm, randomly scramble 0 , 1 , , W and assign it to the first row (W is the last mission number), and then the UAV number in the second line is initialized by random assignment.

3.2.2. Improvement on Walking Behavior

The walking behavior of wolves is essentially an active exploration of the unknown environment, which determines the global search capability of the algorithm. In the WPA, each exploring wolf scouts the prey from h directions divided equally by 360 degrees, which is represented by Formula (16), and adjusts the coverage of the scout by changing the size of h. The bigger h is, the larger the coverage of the scout will be, but the speed of the scout will be relatively slower.
In order to better solve the convergence speed of the optimal solution in discrete problems, this paper proposes that the updating method of Formula (16) is no longer used when each exploring wolf finds the optimal forward direction. Instead, it states that all exploring wolves, in turn, select the position variables in each column, read the number of the UAV performing that task, find all the task numbers assigned to that UAV, and then randomly rank those task numbers and compare them to the original individual. If the change is better, keep the updated plan, otherwise leave it unchanged. The specific operation steps are shown in Figure 2.

3.2.3. Improvement on Calling Behavior

The calling behavior is an essential process for fierce wolves to converge to the current optimal solution which is the position of the leader wolf. This behavior determines the convergence rate of the algorithm. The traditional wolf pack algorithm’s calling behavior, when the fierce wolves approach the leader wolf in step length, will lead to the convergence speed of the algorithm being too slow. Inspired by replication of gene fragment in genetic algorithm, each of the fierce wolf copy part of the leader’s position to replace its own position, so as to achieve fast proximity to the leader wolf and achieve fast convergence of the algorithm.
The position of the fierce wolf should be updated by learning the position of the leader wolf. The specific operations are as follows: Firstly, keep the order of the tasks in the first line of the position variable of the fierce wolf unchanged; finally, update the UAV task assignment scheme in the second line by learning the UAV task assignment scheme of the leader wolf. The specific steps are shown in Figure 3.
Given that the calling step of the improved strategy is a direct learning of the head wolf, the distance judgment operation of the traditional wolf pack algorithm is not used here, and the elite strategy is used instead, that is, if the fierce wolf is worse than the original, it will not be updated.

3.2.4. Improvement on Sieging Behavior

The main purpose of the improved siege strategy is to enhance the algorithm’s global search capabilities. The operations are as follows.
In order to make the algorithm have better convergence effect and optimization ability in the later stage of the iteration, this article will adopt two strategies to improve the siege behavior. First, traverse all individuals with a probability of 0.5, randomly select a column of random integers between n / 2 × ϑ ,   n / ϑ for each individual, ϑ is a scale factor, and re-assign the UAVs number under these columns randomly. Or perform a mutation operation with a probability of 0.5. That is, an individual with a higher fitness value in the population is randomly selected, a few columns are randomly selected from the position information of the individual variable, and the two elements of the column are randomly allocated to obtain a new individual. After improvement, if the effect is better, update it, otherwise it will remain unchanged. The specific steps are shown in the Figure 4 and Figure 5, respectively.

3.2.5. Replenishment of New Individual

The diversity of the wolf population algorithm in the later stage of the population is maintained by the “survival of the strong” mechanism, which is not effective for complex problems. The reason is that it only relies on the initialization mechanism to supplement new individuals, and it is not competitive with other individuals that have been screened later. However, the traditional optimization strategy only uses the position information of the copy of the wolf, and only a slight exchange strategy. In the later stage of the algorithm iteration, the new individuals generated by this method do not have enough diversity, which can easily cause the algorithm to enter local convergence. Therefore, this article proposes a new individual generation strategy.
Among them, individual wolves perform exchange operations with a probability of 0.5, that is, they randomly find two tasks performed by different UAVs in the position variables of the head wolves, and then exchange the numbers of the UAV that perform these two tasks. Or they perform mutation operations with a probability of 0.5. That is, an individual with a higher fitness value in the population is randomly selected, a few columns are randomly selected from the position information of the individual variable, and the two elements of the column are randomly allocated to obtain a new individual. The specific steps are shown in Figure 6 and Figure 7, respectively. If the mutant individuals of the head wolf are used for individual supplementation, in the later stage of the iteration, the similarity between each individual and the head wolf becomes high, which makes it easy for the population to fall into the local optimum. Therefore, this paper adopts the probability of 0.5 to use mutant individuals with higher fitness values to supplement new individuals. This method maintains the diversity of the population and prevents the algorithm from falling into a local optimum.
Algorithm 1: MDWPA
Input: initial position of UAVs: posV, position of targets: posT, iterations, population size: N, Tmax
Output: solution xbest
Generate initial solution X;
for iterations do
   Select the leader wolf Lx;
   /* walking behavior (See Section 3.2.2)   */
Select S exploring wolves;
walking times = 0
   while walking times < Tmax do:
     for exploring wolves Xew do
        [
          individual variation according to Section 3.2.2 and Figure 2;
]
     if F(Xew) >F(Xew) then
        update position: Xew  Xew
     end
     if F(Xew) >F(Xleader) then
        update position: Xleader  Xew
        break;
     end
   end
/* calling behavior (See Section 3.2.3)   */
Select M fierce wolves;
   for fierce wolves Xfw do
     [
     get close to the leader wolf according to Section 3.2.3;
     ]
     if F(Xfw) >F(Xfw) then
        update position: Xfw  Xfw
     end
     if F(Xfw) >F(Xleader) then
        replace leader wolf: Xleader  Xfw
        break;
     end
end
/* sieging behavior (See Section 3.2.4)    */
   for all wolves except the leader wolf Xw do
     [
     r = rand (0,1)
     if r < 0.5 then
        individual variation according to Section 3.2.4 and Figure 4;
   else
        get close to the wolf with better fitness according to Section 3.2.4 and Figure 5;
     end
]
     if F(Xw) >F(Xw) then
        update position: Xw  Xw
     end
end
/* generating new individuals (See Section 3.2.5)    */
Select R weakest wolves XN
   for XN do
     [
     r = rand (0,1)
     if r < 0.5 then
get close to the leader wolf according to Section 3.2.5 and Figure 6;
     else
        get close to the wolf with better fitness according to Section 3.2.5 and Figure 7;
     end
]
     if F(XN) >F(XN) then
        update position: XN  XN
     end
end
Select optimal individual xbest in the population.
end

4. Simulation

4.1. Simulation Setup

The simulation is performed on a 2.30 GHz Intel i5 processor and Python 3.8.0 on a computer with 4 GB memory.
Experimental setting parameters are shown in Table 3.
The setting of related parameters of the wolf pack algorithm has been analyzed in the literature [17], and the parameter settings in Table 3 can optimize the performance of the algorithm.

4.2. Parameter Analysis of MDWPA

In order to obtain the best performance of MDWPA, the optimization strategy of each stage is worth further study. There are four optimization steps in MDWPA, namely, discrete optimization of wandering behavior, discrete optimization of calling behavior, discrete optimization of besieging behavior, and discrete optimization strategy of new individual supplement. In order to verify the effect of the above four optimization steps and optimal performance, this paper uses the orthogonal experimental design (OED) method [33], with the objective function value and the average distances as the evaluation index, 16 kinds of algorithm have been designed, which run independently under the UAV task assignment model 200 times in order to verify which combination of discrete optimization operations would achieve the best optimization performance of MDWPA. In order to avoid contingency, we conduct 30 independent repeated experiments for each scheme, and record the mean, max, and variance of the objective function optimization. The settings of each parameter in the algorithm are shown in Table 3, and the relevant parameter settings of the algorithm are set according to literature [17], in which a large number of experiments prove that the data in Table 3 is the optimal parameter combination.
In Table 4, P1 represents the discrete optimization strategy of wandering behavior. P2 represents the discrete optimization strategy of calling behavior. P3 represents the discrete optimization strategy of siege behavior. P4 represents the new individual recruitment strategy. Through the above 16 different algorithm combinations, simulation experiments are conducted to determine which phase of the optimization strategy has the greatest impact on the algorithm performance. As can be seen from the results in Table 4, the combined optimization strategy in the first and fourth stages has the greatest impact on algorithm performance. It can be seen from the results that MDWPA has certain advantages over the other 15 algorithms in terms of optimal value and average value of objective function.

4.3. Performance of MDWPA

In order to fully analyze the performance and advantages of MDWPA, this paper takes the objective function in Section 2 as the optimization object. By comparing and analyzing the experimental data, the performance of MDWPA algorithm is evaluated from efficiency and stability.

4.3.1. Efficiency

In the UAV task allocation model described in Section 2, for each target’s military mission, the UAV needs to start from the base and rush to the coordinates of the target in turn, the coordinate’s unit is kilometers. The UAV needs to spend a certain period of time at the target to complete the intelligence search task (the unit of time is seconds). Each instance contains five UAVs and 25 military mission targets, which are randomly set in a fixed position. The attributes of each UAV include flight speed and initial coordinates. The attributes of each target are the starting coordinates, initial value, and the execution time. Table 5 is the attributes of targets in Example 1; Table 6 is the attributes of UAVs in Example 1.
In order to prove the efficiency of the MDWPA algorithm in solving the UAV task allocation model, a variety of optimization algorithms are selected to compare with the MDWPA. Some of the algorithms involved are for continuous problems and others are for discrete problems. Therefore, it is difficult to use continuous benchmark functions for unified performance evaluations. In this paper, we select the most widely used intelligent algorithms, PSO [34], GA [35], WPA, PSO-GA-WPA [31] (For brevity, PSO-GA-WPA is abbreviated as GPWPA), CJADE [36] and EBOCMAR [37].
For the fairness of the experiment, the population size is set to 100. In addition, the learning factors c1 and c2 in the particle swarm algorithm are set to 1; the crossover probability in the genetic algorithm is set to 0.8, and the mutation probability is set to 0.2. In order to ensure the accuracy of the results of the simulation experiment, each algorithm will be run 200 times on each instance individually, and the iterative effect analysis curve of each algorithm will be obtained. Figure 6 is the comparison of iterative curves of algorithms on test cases 1.
In Figure 8, the blue, red, orange, purple, green, pink, and brown curves represent the iterative curves of MDWPA, GA, GPWPA, PSO, WPA, CJADE and EBOCMAR, respectively. The purpose of the comparison with the traditional WPA and PSO is to illustrate our efforts on solving discrete problems with traditional biological optimization algorithms. But unfortunately, direct reference to these continuous algorithms is not ideal for simulation results. Comparing MDWPA with the three algorithms for discrete problems, GA, CJADE and EBOCMAR, our proposed MDWPA has advantages on convergence speed and accuracy. GA is a traditional optimization algorithm for discrete problems, and the improved optimization algorithm CJADE and EBOCMAR are newer and can solve discrete problems very well. It is clear from the figure that the GPWPA algorithm has a fast convergence rate in the initial stage, but with the increase of the number of iterations, the algorithm falls into a local convergence state. Although the genetic algorithm has good optimization ability, the convergence speed of genetic algorithm becomes slow after 75 iterations, resulting in local convergence problem. Among them, CJADE can only solve discrete problems, into both continuous and discrete processing. However, the results are not as good as the improved wolf pack algorithm in dealing with the high-dimensional complex task assignment of UAV. As other algorithms deal with continuous problems, they are not effective in dealing with the discrete problem of task allocation for UAVs. Obviously, the MDWPA algorithm has a better convergence speed and assignment effect than other algorithms when dealing with the UAV task assignment model. This is mainly related to the new individual recruitment strategy. With a probability of 0.5, the mutation of the head wolf individual is used for the recruitment of new individuals, and the mutation of the individual with a higher fitness value is used for the recruitment of new individuals. In this way, the diversity of population is kept and the algorithm is prevented from falling into local optimum.
This paper uses variance analysis and crunches other data to compare and analyze the performance of MDWPA and six other algorithms. We carried out the experiments in the first scenario of the model, and the parameters of each algorithm were set as they were in Table 3. A total of 30 independent repeated experimental samples were taken for each algorithm. The algorithm performance was analyzed by function mean, function standard deviation, function maximum, and mean time of each algorithm. The experimental results are shown in Table 7.
By analyzing the data in Table 7, we can see that CJADE, EBOCMAR, PSO and WPA do not have good effects when processing the model in this paper. The function values generally converge at a low level and fall into the local optimal solution. GPWPA improves the function search accuracy in a huge cost of running time, but it is still not ideal. GA has good search accuracy and short running time, but the search accuracy is not good compared with MDWPA proposed in this paper, and it is still locally optimal. In addition, although the MDWPA in this paper has a longer running time, it is also within a reasonable range, and the time complexity does not increase compared with the original WPA, especially for the task assignment scenarios with low real-time requirements.
In summary, the MDWPA has better results in dealing with the model of UAV task assignment, whether it is compared with other intelligent algorithms or compared with the existing improved wolf pack algorithm.

4.3.2. Stability

In order to further verify the stability of the proposed algorithm in complex scenarios. We redesigned two examples to increase the number of UAVs and military targets, respectively. And further compare and analyze with other algorithms.
Example 2 includes 10 UAVs and 50 military targets; each time it runs, the maximum number of iterations is set to 400; the search space dimension is set to 100; Example 3 includes 15 UAVs and 75 military targets; the maximum number of iterations is set to 600; the search space dimension is set to 150; Figure 9a,b show the experimental results of Example 2 and Example 3, respectively.
In the Figure 9, the blue, red, orange, purple, green, pink, and brown curves represent the iterative curves of MDWPA, GA, GPWPA, PSO, WPA, CJADE, and EBOCMAR, respectively. It can be clearly seen from the Figure 9a,b that, with the increase of iterations and task complexity, GPWPA performs better than the genetic algorithm compared with Example 1, which also proves the advantage of the wolf pack algorithm in dealing with complex problems. At the end of iteration, GPWPA still fell into local convergence, but it is obvious from the figure that MDWPA not only has good optimization effect, but also still has high search ability at the end of iteration.
It is worth noting that as the number of UAVs and targets increases, the complexity and dimension of the problem also increase. Therefore, GA is inferior to the original WPA algorithm in Figure 9, which also confirms that the wolf pack algorithm has more advantages when it comes to solving high-dimensional complex problems than other optimization algorithms.

4.3.3. Comprehensive Analysis of MDWPA Performance

In the previous two subsections, the efficiency and stability of the proposed algorithm are compared, and it can be seen that MDWPA has obvious performance advantages in the UAV task assignment model. However, in order to more fully prove the practicability of the algorithm, the average flight distance is further used as an evaluation index, and the MDWPA algorithm is compared and analyzed with other intelligent algorithms. Figure 10 shows the average flight distance of the UAV of each algorithm in the three example scenarios.
In Figure 10, the abscissa is the five intelligent algorithms, and the ordinate is the average flight distance of the UAV. The red (set1), blue (set2), and green rectangles (set3) represent the aforementioned Example 1, Example 2, and Example 3, respectively. Each intelligent algorithm has three bar graphs to represent the average flight distance of the UAV under the three mission scenarios. It can be clearly seen from Figure 10 that the MDWPA has obvious advantages compared to the other four intelligent algorithms. Regardless of the mission scenario, its average flying distance is the shortest, which proves that in the actual scenario, it can always find the best and shortest flight path for the UAV.
The above experimental results show that the efficiency and stability of the MDWPA in solving the UAV task allocation problem has considerable advantages compared with other algorithms. And it solves the shortcomings of the traditional wolf pack algorithm that can easily fall into the local optimum. Therefore, the improved wolf pack algorithm is suitable for large-scale and complex UAV task assignment and has relatively large performance advantages.

5. Conclusions

This paper mainly studies the swarm intelligence algorithm based on the behavior mechanism of wolves and its application in the task assignment problem of UAV swarms. Firstly, a UAV task allocation model under the actual battlefield environment is established, and two evaluation indicators of resource consumption and task execution effect will be considered at the same time. The paper then summarizes and analyzes the principles and steps of the original WPA. After gaining a full understanding of the theoretical basis of the WPA, by analyzing the shortcomings of the wolf pack algorithm, in order to better solve the discrete problem of task allocation, a better performance multi-discrete wolf pack algorithm is proposed: MDWPA.
Through the simulation experiments, it was found that compared with WPA, PSO, GA GPWPA, and other algorithms, MDWPA has a greater performance advantage in terms of solving the problem of complex UAV task allocation.
In the future, more practical situations will be considered to further improve the existing model, such as building a model for a situation where too many tasks cause UAVs to complete all tasks due to limited resources. The next step of the research work will combine these new technical theoretical methods to improve existing algorithms to further improve the solution performance of the UAV cooperative multi-task assignment.

Author Contributions

Conceptualization, S.X., L.L., J.H. and Y.M.; methodology, S.X., L.L. and Z.Z.; validation, L.L. and Z.Z.; formal analysis, L.L., Z.Z. and S.X.; investigation, S.X. and L.L.; writing—original draft preparation, L.L., Z.Z. and S.X.; writing—review and editing, L.L.; supervision, S.X.; project administration, S.X. and Y.M.; funding acquisition, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the Gusu Innovation Leading Talents Special Project (the 2020 Kunshan Zu Chong Key Tack Project), grant number ZXL2020210.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jia, G.; Wang, J. A Review on Mission Planning Methods of UAV Cluster. Syst. Eng. Electron. 2021, 43, 99–111. [Google Scholar]
  2. Zhong, W.; Li, X.; Chang, H.; Liang, F. Anti-UAV cluster air defense deployment model based on nested PSO algorithm. Electro-Opt. Control. 2021, 28, 1–7. [Google Scholar]
  3. Chen, Y.; Yang, D.; Yu, J. Multi-UAV Task Assignment With Parameter and Time-Sensitive Uncertainties Using Modified Two-Part Wolf Pack Search Algorithm. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 2853–2872. [Google Scholar] [CrossRef]
  4. Wei, R.; Wu, Z. Research on Real-time Task Assignment method of UAV Cluster. J. Syst. Simul. 2021, 33, 1574–1581. [Google Scholar]
  5. Alidaee, B.; Wang, H.; Landram, F. A Note on Integer Programming Formulations of the Real-Time Optimal Scheduling and Flight Path Selection of UAVs. IEEE Trans. Control. Syst. Technol. 2009, 17, 839–843. [Google Scholar] [CrossRef]
  6. Lim, G.J.; Kim, S.; Cho, J.; Gong, Y.; Khodaei, A. Multi-UAV Pre-Positioning and Routing for Power Network Damage Assessment. IEEE Trans. Smart Grid 2018, 9, 3643–3651. [Google Scholar] [CrossRef]
  7. Bisis, R.S.; Pal, A.; Werho, T.; Vittal, V. A Graph Theoretic Approach to Power System Vulnerability Identification. IEEE Trans. Power Syst. 2021, 36, 923–935. [Google Scholar] [CrossRef]
  8. Tootooni, M.S.; Rao, P.K.; Chou, C.; Kong, Z.J. A Spectral Graph Theoretic Approach for Monitoring Multivariate Time Series Data From Complex Dynamical Processes. IEEE Trans. Autom. Sci. Eng. 2018, 15, 127–144. [Google Scholar] [CrossRef]
  9. Wang, H.; Zhang, L.; Shi, C.; Che, F.; Zhang, P. Modeling of Equipment Support Task Assignment and Solution of DLS-BCIWBA Algorithm. Syst. Eng. Electron. 2018, 40, 1979–1985. [Google Scholar]
  10. Almannaa, M.H.; Elhenawy, M.; Rakha, H.A. A Novel Supervised Clustering Algorithm for Transportation System Applications. IEEE Trans. Intell. Transp. Syst. 2020, 21, 222–232. [Google Scholar] [CrossRef]
  11. Wang, Z.J.; Zhan, Z.H.; Kwong, S.; Jin, H.; Zhang, J. Adaptive Granularity Learning Distributed Particle Swarm Optimization for Large-Scale Optimization. IEEE Trans. Cybern. 2021, 51, 1175–1188. [Google Scholar] [CrossRef] [PubMed]
  12. Li, L.; Chang, L.; Gu, T.; Sheng, W.; Wang, W. On the Norm of Dominant Difference for Many-Objective Particle Swarm Optimization. IEEE Trans. Cybern. 2021, 51, 2055–2067. [Google Scholar] [CrossRef] [PubMed]
  13. Huanca, D.H.; Pareja, L.A.G. Chu and Beasley Genetic Algorithm to Solve the Transmission Network Expansion Planning Problem Considering Active Power Losses. IEEE Lat. Am. Trans. 2021, 19, 1967–1975. [Google Scholar] [CrossRef]
  14. Pradhan, D.; Wang, S.; Ali, S.; Yue, T.; Liaaen, M. CBGA-ES+: A Cluster-Based Genetic Algorithm with Non-Dominated Elitist Selection for Supporting Multi-Objective Test Optimization. IEEE Trans. Softw. Eng. 2021, 47, 86–107. [Google Scholar] [CrossRef]
  15. Davari, S.A.; Nekoukar, V.; Garcia, C.; Rodriguez, J. Online Weighting Factor Optimization by Simplified Simulated Annealing for Finite Set Predictive Control. IEEE Trans. Ind. Inform. 2021, 17, 31–40. [Google Scholar] [CrossRef]
  16. Duan, H.B.; Qiao, P.X. Pigeon-inspired optimization: A new swarm intelligence optimizer for air robot path planning Int. J. Intell. Comput. Cybern 2014, 07, 24–37. [Google Scholar] [CrossRef]
  17. Wu, H.; Zhang, F.; Wu, L. A new swarm intelligence algorithm-Wolf pack algorithm. Syst. Eng. Electron. 2013, 35, 2430–2438. [Google Scholar]
  18. Wu, H.; Xiao, R. New research on swarm intelligence: Role-matching Wolf Division of Labor. J. Intell. Syst. 2021, 16, 125–133. [Google Scholar]
  19. Xu, J.; Kang, X.; Fan, Z.; Zhang, Z.; Li, Y.; Dong, X.; Gao, X. Design of Highly Uniform Magnetic Field Coils With Wolf Pack Algorithm. IEEE Sens. J. 2021, 21, 4412–4424. [Google Scholar] [CrossRef]
  20. Wang, D.; Ban, X.; Ji, L.; Guan, X.; Liu, K.; Qian, X. An Adaptive Shrinking Grid Search Chaotic Wolf Optimization Algorithm Using Standard Deviation Updating Amount. Comput. Intell. Neurosci. 2020, 2020, 1–15. [Google Scholar] [CrossRef]
  21. Cao, Q.K.; Yang, K.W.; Ren, X.Y. Vehicle routing optimization with multiple fuzzy time windows based on improved wolf pack algorithm. Adv. Prod. Eng. Manag. 2017, 12, 401–411. [Google Scholar] [CrossRef] [Green Version]
  22. Li, H.; Wu, H. An oppositional wolf pack algorithm for parameter identification of the chaotic systems. Optik 2016, 127, 9853–9864. [Google Scholar] [CrossRef]
  23. Zhu, Y.; Jiang, W.; Kong, X.; Quan, L.; Zhang, Y. A chaos wolf optimization algorithm with self-adaptive variable step-size. Aip Adv. 2017, 7, 105024. [Google Scholar] [CrossRef]
  24. Chen, X.; Tang, C.; Wang, J.; Zhang, L.; Meng, Q. Improved Wolf Pack Algorithm Based on Differential Evolution Elite Set. IEICE Trans. Inf. Syst. 2018, 101, 1946–1949. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, D.; Qian, X.; Liu, K.; Ban, X.; Guan, X. An Adaptive Distributed Size Wolf Pack Optimization Algorithm Using Strategy of Jumping for Raid(September 2018). IEEE Access 2018, 6, 65260–65274. [Google Scholar] [CrossRef]
  26. Chen, X.; Cheng, F.; Liu, C.; Cheng, L.; Mao, Y. An improved Wolf pack algorithm for optimization problems: Design and evaluation. PLoS ONE 2021, 16, e0254239. [Google Scholar] [CrossRef]
  27. Gao, Y.; Zhang, F.; Zhao, Y.; Li, C. Quantum-inspired wolf pack algorithm to solve the 0-1 knapsack problem. Math. Probl. Eng. 2018, 2018, 1–10. [Google Scholar] [CrossRef]
  28. Zhu, Q.; Wu, H.; Li, N.; Hu, J. A Chaotic Disturbance Wolf Pack Algorithm for Solving Ultrahigh-Dimensional Complex Functions. Complexity 2021, 2021, 1–15. [Google Scholar] [CrossRef]
  29. Lu, Y.; Ma, Y.; Wang, J. Multi-Population Parallel Wolf Pack Algorithm for Task Assignment of UAV Swarm. Appl. Sci. 2021, 11, 11996. [Google Scholar] [CrossRef]
  30. Lu, Y.; Ma, Y.; Wang, J.; Han, L. Task assignment of UAV swarm based on Wolf Pack algorithm. Appl. Sci. 2020, 10, 8335. [Google Scholar] [CrossRef]
  31. YongBo, C.; YueSong, M.; JianQiao, Y.; XiaoLong, S.; Nuo, X. Three-dimensional unmanned aerial vehicle path planning using modified wolf pack search algorithm. Neurocomputing 2017, 266, 445–457. [Google Scholar] [CrossRef]
  32. Wang, F.; Zhang, H.; Han, M.; Xing, L. Co-evolution Based Mixed-variable Multi-objective Particle Swarm Optimization for UAV Cooperative Multi-task Allocation Problem. Chin. J. Comput. 2021, 44, 1967–1983. [Google Scholar]
  33. Gao, S.; Zhou, M.; Wang, Y.; Cheng, J.; Yachi, H.; Wang, J. Dendritic Neuron Model With Effective Learning Algorithms for Classification, Approximation, and Prediction. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 601–614. [Google Scholar] [CrossRef]
  34. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. MHS’95. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  35. Grefenstette, J.J. Optimization of Control Parameters for Genetic Algorithms. IEEE Trans. Syst. Man Cybern. 1986, 16, 122–128. [Google Scholar] [CrossRef]
  36. Gao, S.; Yu, Y.; Wang, Y.; Wang, J.; Cheng, J.; Zhou, M. Chaotic Local Search-Based Differential Evolution Algorithms for Optimization. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 3954–3967. [Google Scholar] [CrossRef]
  37. Kumar, A.; Misra, R.K.; Singh, D. Improving the local search capability of Effective Butterfly Optimizer using Covariance Matrix Adapted Retreat Phase. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), Donostia, Spain, 5–8 June 2017; pp. 1835–1842. [Google Scholar]
Figure 1. The hybrid coding method.
Figure 1. The hybrid coding method.
Applsci 12 01331 g001
Figure 2. The walking strategy.
Figure 2. The walking strategy.
Applsci 12 01331 g002
Figure 3. The calling strategy.
Figure 3. The calling strategy.
Applsci 12 01331 g003
Figure 4. The siege strategy using random assignment.
Figure 4. The siege strategy using random assignment.
Applsci 12 01331 g004
Figure 5. The siege strategy using mutation operations.
Figure 5. The siege strategy using mutation operations.
Applsci 12 01331 g005
Figure 6. Create new individuals by replicating the head wolf.
Figure 6. Create new individuals by replicating the head wolf.
Applsci 12 01331 g006
Figure 7. Generate new individuals in a mutated way.
Figure 7. Generate new individuals in a mutated way.
Applsci 12 01331 g007
Figure 8. Comparison of iterative curves of algorithms on test cases 1.
Figure 8. Comparison of iterative curves of algorithms on test cases 1.
Applsci 12 01331 g008
Figure 9. Comparison of iterative curves of algorithms on test cases. (a) is the comparison of iterative curves of algorithms on Example 2; (b) is the comparison of iterative curves of algorithms on Example 3.
Figure 9. Comparison of iterative curves of algorithms on test cases. (a) is the comparison of iterative curves of algorithms on Example 2; (b) is the comparison of iterative curves of algorithms on Example 3.
Applsci 12 01331 g009
Figure 10. The average flight distance of the UAV in the three example scenarios.
Figure 10. The average flight distance of the UAV in the three example scenarios.
Applsci 12 01331 g010
Table 1. Object attributes and symbol interpretation.
Table 1. Object attributes and symbol interpretation.
Attributes
UAV U
(For any UAV u i )
The total number of UAVs: N u
Initial position (the base) P i = x i , y i i = 0
Flying speed: v i
The maximum travel distance of UAV: d i m a x
The maximum travel time of UAV: τ i   m a x
Target T
(For any target T j )
The number of targets: N t
Position coordinates: P j = x j , y j
Reconnaissance time required for completion of missions: I j
Table 2. The vector of cooperative task allocation for UAVs.
Table 2. The vector of cooperative task allocation for UAVs.
1 T j
2 M K
3 X i j
Table 3. Experimental parameter settings.
Table 3. Experimental parameter settings.
ParameterValue
Number of wolves: N
Dimension of the search space:
100
50/100/150
The maximum number of iterations: k m a x 200/400/600
Distance judgement factor: ω 800
Step factor: S 1000
Detective Wolf Ratio Factor: α 4
Update scale factor: β 6
Maximum number of walks: T m a x 10
Table 4. The design scheme of orthogonal experiment design.
Table 4. The design scheme of orthogonal experiment design.
No.AlgorithmAverageMaxStd
1WPA0.0943118280.0985155620.003123479
2WPA + P10.0944068830.0997378270.003602062
3WPA + P20.1000994290.1049960640.003406232
4WPA + P30.2558776330.2773448510.012403991
5WPA + P40.0899912020.0998914480.003966736
6WPA + P1 + P20.1013468730.1044144140.002610191
7WPA + P1 + P30.1285439160.1331542610.002759905
8WPA + P1 + P40.2667372650.2922528010.017772005
9WPA + P2 + P30.2596152610.2918029990.012754875
10WPA + P2 + P40.0915928710.0996380710.003605891
11WPA + P3 + P40.2521971810.2703498450.011637504
12WPA + P1 + P2 + P30.2633252870.2952717520.013132305
13WPA + P1 + P2 + P40.1214764440.1253327040.003436636
14WPA + P1 + P3 + P40.2633256110.2946005590.018923418
15WPA + P2 + P3 + P40.2551409280.2767469830.016519906
16WPA + P1 + P2 + P3 + P40.2757759870.3154694380.011828138
Table 5. The attributes of targets in Example 1.
Table 5. The attributes of targets in Example 1.
No.CoordinateInitial ValueExecution Time (s)No.CoordinatesInitial ValueExecution Time (s)
0(13,41)80813(90,22)9215
1(61,83)901114(95,20)9512
2(79,12)40315(81,86)9312
3(41,98)901416(16,65)687
4(23,65)952217(85,51)829
5(12,91)791218(23,72)6216
6(68,54)731719(70,83)6018
7(82,50)841820(23,30)904
8(20,50)81321(65,30)6016
9(92,48)691822(87,77)9411
10(69,59)711023(93,18)8312
11(59,10)701524(19,19)908
12(12,51)6220
Table 6. The attributes of UAVs in Example 1.
Table 6. The attributes of UAVs in Example 1.
No.Initial CoordinateVelocity (km/s)
0(0,0)0.1
1(0,0)0.12
2(0,0)0.16
3(0,0)0.11
4(0,0)0.1
Table 7. The experimental results of variance analysis.
Table 7. The experimental results of variance analysis.
AlgorithmAverageStdMaxTime (Avg)
GPWPA0.1628477720.0073167360.1784012715900.154593
CJADE0.1019511090.0069017080.115765696649.6378847
EBOCMAR0.117672230.0063585090.138806691161.0406503
GA0.2183057640.0246355960.28258295561.89460519
MDWPA0.2635485690.015905240.2908385722827.711819
PSO0.1228226770.0073410320.14646144572.66523946
WPA0.0937250060.0031002260.101597478397.2063249
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, S.; Li, L.; Zhou, Z.; Mao, Y.; Huang, J. A Task Allocation Strategy of the UAV Swarm Based on Multi-Discrete Wolf Pack Algorithm. Appl. Sci. 2022, 12, 1331. https://doi.org/10.3390/app12031331

AMA Style

Xu S, Li L, Zhou Z, Mao Y, Huang J. A Task Allocation Strategy of the UAV Swarm Based on Multi-Discrete Wolf Pack Algorithm. Applied Sciences. 2022; 12(3):1331. https://doi.org/10.3390/app12031331

Chicago/Turabian Style

Xu, Shufang, Linlin Li, Ziyun Zhou, Yingchi Mao, and Jianxin Huang. 2022. "A Task Allocation Strategy of the UAV Swarm Based on Multi-Discrete Wolf Pack Algorithm" Applied Sciences 12, no. 3: 1331. https://doi.org/10.3390/app12031331

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop