Next Article in Journal
GTrXL-SAC-Based Path Planning and Obstacle-Aware Control Decision-Making for UAV Autonomous Control
Previous Article in Journal
Safe and Optimal Motion Planning for Autonomous Underwater Vehicles: A Robust Model Predictive Control Framework Integrating Fast Marching Time Objectives and Adaptive Control Barrier Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Optimization of Task Completion Time and Energy Consumption in UAV-Enabled Mobile Edge Computing

National Key Laboratory of Information Systems Engineering, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(4), 274; https://doi.org/10.3390/drones9040274
Submission received: 24 February 2025 / Revised: 17 March 2025 / Accepted: 1 April 2025 / Published: 3 April 2025

Abstract

:
Unmanned Aerial Vehicles (UAVs) hold great promise for Mobile Edge Computing (MEC) owing to their flexible mobility, rapid deployment, and low-cost characteristics. However, UAV-enabled MEC still faces challenges in terms of the real-time arrival of computational tasks, energy reservation, and the actual response efficiency of the system. In this study, we focus on a UAV-enabled MEC scenario, where multiple UAVs function as airborne edge servers, offering computation services to multiple ground-based user devices (UDs). We aim to minimize the cost of the MEC system by optimizing the computation offloading policy. Specifically, we take task latency into account to ensure the timeliness of real-time tasks. The Lyapunov optimization method is employed to maintain a uniform and stable queue for energy consumption. Additionally, we draw on the concept of maximum completion time in shop-floor scheduling to optimize the actual response latency. To this end, we propose a joint optimization algorithm. First, the joint optimization problem is transformed into a per-time-slot real-time optimization problem (PROP) using the Lyapunov optimization framework. Then, a reinforcement learning method, LyraRD, is proposed to solve the PROP. Experimental results verify that the proposed approach outperforms the benchmarks in terms of system performance.

1. Introduction

With the widespread adoption of the Internet of Things (IoT) and wireless communication technologies, numerous intelligent applications with strict computational latency requirements have emerged, such as real-time video, autonomous driving, and online gaming [1]. Nevertheless, the terminal devices for these applications are limited in computing power and battery capacity, making it arduous to meet the low-latency demands. Mobile Edge Computing (MEC) is regarded as a promising solution due to its real-time computing and energy-saving capabilities [2,3,4]. Based on MEC, user devices can delegate computationally intensive tasks to edge servers instead of relying on remote data centers. This not only compensates for the computing power shortage of user devices but also reduces the response time.
Initially, most MEC servers were deployed on the ground [5,6,7]. In fact, these ground-based MEC servers have provided convenient services to users. However, they typically cover only a portion of the user population [2,8,9]. Fortunately, the rapid development of Unmanned Aerial Vehicles (UAVs) offers new opportunities to address the above-mentioned issues. UAVs can carry computing devices and hover in the air, enabling MEC servers to be deployed flexibly to meet user needs. As depicted in Figure 1, two UAVs hover in air, with different levels of battery capacity, computing resources, and bandwidth resources. We use a three-tuple to represent the different normalized resource values of the UAVs, i.e., U A V 1 : ( 20 ,   2 ,   80 % ) and U A V 2 : ( 20 ,   1.5 ,   60 % ) . Two user devices, U D 1 and U D 2 , each with unit computing resources, generate computation tasks at 5-second intervals in different time slots. U D 1 generates tasks { a 1 = 5 ,   a 2 = 3 } , { a 3 = 6 ,   a 4 = 4 } , { a 5 = 2 ,   a 6 = 3 } and U D 2 generates tasks { b 1 = 3 ,   b 2 = 1 } , { b 3 = 5 ,   b 4 = 4 } , and { b 5 = 6 ,   b 6 = 3 } , where the values represent the units required of computing resources in three slots, respectively. Without loss of generality, we assume that UAVs consume one unit of energy for executing one unit of a computation task.
Figure 2 presents two solutions to offload the computing tasks. In Figure 2a, U D 1 offloads tasks { a 1 ,   a 3 } to U A V 1 , and { a 4 ,   a 6 } to U A V 2 , while U D 2 offloads tasks { b 4 ,   b 5 } to U A V 1 , and { b 1 } to U A V 2 . In the first slot (i.e., T = 1 ), task b 1 is executed in U A V 2 , taking up 3 / 1.5 = 2 seconds and 3 / 1 = 3 units of energy; task a 1 is executed in U A V 1 , taking up 5 / 2 = 2.5 seconds and 5 / 1 = 5 units of energy; task a 2 is executed in U D 1 , taking up 3 / 1 = 3 seconds; and task b 2 is executed in U D 2 , taking up 1 / 1 = 1 seconds. Then, the makespan of all tasks of U D 1 is m a x ( 2.5 ,   3 ) = 3 seconds, while the makespan of all tasks of U D 2 is m a x ( 2 ,   1 ) = 2 seconds. Similarly, we can easily obtain the energy cost and makespan in the next two slots (see the detailed values in Figure 2a). Figure 2b presents the other solution. We can easily find that the total percentages of energy use of U A V 1 and U A V 2 in Figure 2a are 5 + 10 + 6 20 = 105 % and 3 + 4 + 3 20 = 50 % , while in Figure 2b, they are 5 + 8 + 6 20 = 95 % and 3 + 6 + 3 20 = 60 % . The average makespan in Figure 2a is 3 + 2 + 3 + 5 + 2 + 3 6 = 3 , and in Figure 2b, it is 3 + 4 + 2 + 2 + 5 + 3 6 = 3.16 . Considering that the scenario in Figure 1 is online, meaning that users will constantly generate tasks, to ensure that users have a good experience in offloading tasks to the UAVs, the system should have rapidity to finish tasks in each slot (i.e., minimize the average makespan) as soon as possible, and durability to serve (i.e., minimize the maximum energy cost) as long as possible.
In this paper, we are committed to exploring the optimization of both efficiency and durability to identify an optimal task offloading solution in UAV-enabled MEC. However, solving this problem poses three major challenges:
(i) The existing research [10,11,12] predominantly relies on offline task scheduling or prior knowledge of tasks. In contrast, practical applications such as autonomous driving and real-time video analytics typically involve the arrival of stochastic real-time task data. This implies that the current decision making must consider the potential influence of future tasks.
(ii) Most research [4,7,13] efforts concentrate on keeping energy consumption below a certain limit or minimizing the overall energy consumption throughout the process. However, effective energy management should not only aim at minimization but also strive for balance. This is crucial to ensure that UAVs equipped with high computational resources can operate for extended durations.
(iii) It is commonly assumed that task offloading occurs at the beginning of each time slot [3,5,6]. Given that transmitting the task to the UAV will inevitably consume the available time within the time slot, careful bandwidth allocation is essential.
To address these challenges, we model the joint optimization problem of task scheduling in an online fashion instead of an offline one. This approach optimizes the average values across all time slots. Next, to ensure the stable operation of the MEC system with randomly arriving tasks, we utilize Lyapunov optimization theory to introduce a virtual queue that can evenly distribute energy consumption. To solve the online model more accurately, we further decompose the multi-stage stochastic problem into a series of deterministic problems, which can be effectively solved by reinforcement learning. The main contributions of this study are as follows:
  • We present a novel online optimization framework to help users complete tasks quickly and provide them with more durable services. The framework integrates Lyapunov optimization and Mixed Integer Nonlinear Programming (MINLP), which dynamically adapts to real-time task arrivals while jointly optimizing task response time (by minimizing makespan) and energy consumption. Through Lyapunov optimization, we ensure a uniform energy distribution across the UAV, thereby effectively mitigating energy consumption fluctuations caused by bursty tasks while responding in a timely manner in the absence of a priori knowledge of the data.
  • We introduce Lyapunov virtual energy queues, decompose the problem into a real-time optimization problem (PROP) for each time slot via Lyapunov drift, and design the reinforcement learning-based algorithm LyraRD. LyraRD uniquely combines Lyapunov-induced stability constraints with adaptive decision making to reduce the computational complexity to millisecond response times.
  • We investigate the performance of our method under different parameter settings. With extensive experiments, we demonstrate that our method performs better than the benchmarks.
The rest of the paper is organized as follows. Section 2 summarizes the related work. Section 3 introduces the related system model and problem formulation in detail. Then, we describe the Lyapunov-based problem transformation in detail in Section 4. In Section 5, we introduce the two-stage optimization algorithm and theoretical analysis. Our evaluation method and experimental results are shown and analyzed in Section 6. Finally, this paper is concluded in Section 7.

2. Related Work

Due to the high mobility and flexibility, UAV-enabled MEC systems have attracted widespread attention in recent years.

2.1. Real-Time Task Arrival

Most studies have based their strategies on user task requirements given as a priori knowledge [10,11]. However, in most MEC applications, such as autonomous driving and online games, computational tasks arrive randomly in real time [12], and there is a lack of access to the a priori knowledge of the tasks. Therefore, considering the random arrival of task data, Wan et al. [14] developed a path planning algorithm based on deep reinforcement learning by discretizing the action space into a finite set, including eight steering directions and hovering modes. Yang et al. [13] considered the stochastic nature of user tasks and developed a Lyapunov optimization based on perturbation-based algorithm based on an online algorithm to jointly optimize the energy consumption and task processing rate of UAVs while maintaining the stability of long-term data queues. When considering user mobility and random task arrivals, Yang et al. proposed an online control algorithm for UAV trajectory and resource scheduling to minimize a certain budgeted energy consumption of the UAV in a certain budgeted energy consumption by converting the long-term optimization problem into an optimization problem for the real-time control of a real-time control system in [15]. Yan et al. [16] proposed an artificial noise-based UAV-assisted non-orthogonal multiple access (NOMA) secure communication scheme, whose channel dynamics analysis and resource allocation mechanism provide a theoretical basis for future extension of security enhancement strategies. Ali et al. [17] proposed a novel multi-objective adaptive learning framework called MOALF-UAV-MEC, tailored for dynamic IoT environments, which provides a scalable and adaptive solution for deploying computational resources in infrastructure-limited regions, ad hoc events or emergencies. However, they simply add the time delays linearly as the completion delay of the task, which is biased compared to the actual response delay of the task. For this reason, in this paper, we introduce the maximum completion delay (i.e., makespan) in the scheduling problem to represent the actual response delay of the task and improve the response efficiency of MEC by reducing it.
Meanwhile, most of the existing studies consider that the UAV computes the offloaded tasks at the initial moment of each time slot. Zeng et al. [11] optimized resource allocation and UAV trajectory to reduce the waiting latency of the task queue. Sheng et al. [18] considered the simultaneous occurrence of the time-varying/random ground channel and the air-to-ground line-of-sight channel, and optimized the UAV trajectory and the task offload strategy to ensure the robustness of task processing and, at the same time, minimize energy consumption.
All of the above studies theoretically assume that the computational delay of the tasks offloaded to the UAV should be less than or equal to the time slot length. However, when the computational delay of the offloaded tasks happens to be equal to the time slot length, due to the effect of the task transmission delay, it will result in part of the tasks not being able to become computed in all time slots (e.g., task b 2 on the U A V 1 at T = 2 in Figure 2). If this remaining task is ignored, it will lead to part of the task losing packets, so this paper considers deferring the remaining task to the next time slot, which creates a period of deferred delay, thus causing the UAV pair to start calculating the task not at the initial moment of the time slot. The above effects should be taken into account when calculating the task response delay.

2.2. UAV Energy Management

Traditional UAV energy control mainly considers the available energy consumption of a given UAV and ensures that the energy consumption of the entire process is less than this [11,19,20], and this approach of energy control is mostly applicable to task scheduling problems with given a priori knowledge, where the allocation of the energy consumption of the whole MEC process is adjusted with the number of tasks unloaded in the time slot. New studies on UAV energy management and control are also underway. Michailidis et al. [21] maximize the minimum security computational efficiency (SCE) in a UAV-RIS-assisted MEC-IoT network in the presence of airborne and ground eavesdroppers by jointly optimizing the transmission power allocation, time-slot scheduling, task assignment, and phase-shift of the RIS to ensure that the system’s energy consumption and security performance are optimized. Lakew et al. [22] proposed a MEC model assisted in LEO satellites, in which UAVs can acquire a series of energy packets at the beginning of each time slot by the energy harvesting (EH) technique, which in turn efficiently manages the energy harvesting and consumption of the EH-UAVs to optimize the performance of the whole system and the mission accomplishment rate. However, when there is a lack of a priori knowledge about the tasks, i.e., the tasks arrive randomly, this energy management approach will lead to the uneven distribution of energy consumption and thus insufficient reservation, which will reduce the computational efficiency in the face of unexpected large-scale tasks and thus increase the response latency of the whole MEC system. For this reason, Lyapunov optimization methods are commonly used to stabilize the energy use of each time slot.
Wang et al. [23] considered the joint optimization problem of UAV trajectories and offloading decisions and used Lyapunov optimization methods to satisfy the long-time energy constraints of UAVs while making the energy allocation uniform to minimize the sum of the service latency linearity of each time slot. Qin et al. [24] considered task cache policies, UAV trajectories, and resource allocation policies to reduce the sum of service delay linearity.
In recent years, reinforcement learning has shown significant potential in UAV energy optimization. For example, Li et al. [25] balanced the multidimensional constraints of ground–air cooperative communication using multi-objective deep reinforcement learning, and proposed a ground–air cooperative communication framework based on Evolutionary Multi-objective Deep Reinforcement Learning (EMODRL), which balanced the multidimensional objectives of latency, energy consumption, and coverage through Pareto frontier search. Prakhar et al. [26] presented the SFRL method for the purpose of distributed ML model learning in wireless UAV-assisted MEC AI model networks. Compared to existing distributed learning algorithms, the proposed method yields comparable high test accuracy performance while consuming less energy. And Wu et al. [27] designed an evolutionary reinforcement learning framework for incomplete information scenarios to improve the strategy robustness through action sequence search, and its evolutionary exploration mechanism can provide a reference for the improvement of strategy generalizability under stochastic task arrival.

2.3. Differences of Our Work

Sequential Task Execution with Dynamic Deferral. Unlike existing approaches that either ignore deferred tasks or assume instantaneous computation starts at slot boundaries, we explicitly model the inter-slot dependency of task execution. By deferring unfinished tasks to subsequent slots and accounting for their cascading delays, our framework minimizes the actual response latency (via makespan optimization) while maintaining queue stability. This contrasts with methods like  [13,15], which linearly aggregate delays and fail to capture the nonlinear impact of deferred computations.
Non-Uniform Energy Balancing via Lyapunov-Guided RL. While previous studies can address problems such as UAV task offloading based on Lyapunov optimization [23,24] stabilize energy queues, they do not integrate reinforcement learning (RL) to adaptively balance energy consumption across heterogeneous UAVs. Our LyraRD algorithm uniquely combines Lyapunov optimization with RL to dynamically adjust offloading decisions, ensuring that UAVs with higher resource capacities sustain longer operational durations. This prevents energy “hotspots” and improves the resilience of the system, addressing a critical gap in works such as  [17,25], which treat energy allocation statically.
Inspired by these gaps, we consider a UAV-enabled MEC system with randomly arriving tasks. Task timeliness is ensured by deferring unprocessed tasks from the previous moment to the next moment, while UAV energy usage is controlled throughout the process to minimize the UAV energy consumption and actual response time of the system.

3. System Model and Problem Formulation

In this section, we first introduce a UAV-enabled MEC framework that is designed to offload computational tasks from UDs. Next, we model the latency and energy consumption of task offloading, transmission, and computation. We then present a joint optimization problem with the goal of minimizing the weighted sum of the actual task response latency and energy usage of UAVs across the MEC system.

3.1. System Overview

As illustrated in Figure 1, a MEC system enabled by UAVs consists of I UDs and J rotation UAVs, where the set I = { 1 , 2 , , I } , and J = { 1 , 2 , , J } . The system timeline is discretized into T equal time slots [28], i.e., t T = { 1 , 2 , , T } , where the duration of each time slot is denoted as T s . The UAV acts as an aerial edge server of MEC, providing computation offloading services for UDs in each time slot.
UD Model. We assume that there are several computation tasks that arrive at each UD in each time slot [29]. For U D i ( i I ) , the properties of this UD in time slot t can be expressed as M i ( t ) = ( f i , Φ i ( t ) , P i ( t ) ) , where f i represents the local computation capability of U D i , and P i ( t ) represents the coordinates of U D i . The task that reaches U D i at time t is Φ i ( t ) , which consists of k tasks: Φ i ( t ) = ( α i 1 ( t ) , α i 2 ( t ) , α i 3 ( t ) , , α i k ( t ) ) . These tasks are expressed in the form α i k ( t ) = { D i k ( t ) , η i k ( t ) } , where D i k ( t ) and η i k ( t ) represent the amount of data and computing intensity of task α i k ( t ) generated at time t.
UAV Model. The attributes of U A V j ( j J ) are represented as U j ( t ) = ( r j , ω j , P j ( t ) , H ) , where r j and ω j represent the proportion of computing resources and bandwidth, respectively. P j ( t ) represents the coordinates of the U A V j at time t, and H represents the altitude [30].
Decision Variables. Decision variables include the task offloading strategy and the UAV resource allocation strategy. (i) Task Offloading Decision. For task Φ i ( t ) , define a binary variable a i , j k ( t ) to represent the decision to load U D i in the time period t, where a i , j k ( t ) = 0 means that the k t h task of U D i in the time period t is processed locally, and a i , j k ( t ) = 1 means that the k t h task of U D i in time period t is loaded into U A V j for processing. (ii) Resource Allocation Decision. For UAVs, the communication resources and bandwidth resources allocated to the task k unloaded at time t are r j i , k ( t ) and ω j i , k ( t ) .
These factors affect each other, so joint decision making is required.

3.2. Communication Model

The probabilistic line-of-sight (LoS) channel model is used to simulate the communication between U D i and U A V j  [19]. First, the probability of LoS and NLoS communication between U D i and U A V j in time period t is shown in (1) and (2) [31]:
P i , j los ( t ) = 1 1 + ξ 1 exp ξ 2 θ i , j ( t ) ξ 1
and
P ¯ i , j l o s ( t ) = P i , j los ( t ) + 1 P i , j los ( t ) κ
where ξ 1 and ξ 2 are constants depending on the propagation environment, θ i , j ( t ) = 180 π arcsin H d i , j ( t ) represents the elevation angle, d i , j ( t ) represents the straight-line distance between U D i and U A V j , and κ represents the additional attenuation factor.
According to  [32], the communication rate of U D i in time slot t can be expressed as Equation (3):
R i , j k ( t ) = ω j i , k ( t ) φ i , j ( t )
where φ i , j ( t ) is the spectrum efficiency of U D i , which can be calculated as (4)–(5):
φ i , j ( t ) = log 2 1 + θ i ( t ) ( | | P j ( t ) P i ( t ) | | 2 + H 2 ) μ
θ i ( t ) = P i β 0 P ¯ i , j l o s ( t ) N 0 , μ = μ ¯ 2
where β 0 represents the channel gain at the reference distance of 1 m , N 0 represents the noise power, P i is the transmission power of U D i , and μ ¯ represents the path loss exponent.

3.3. Computation Model

Local Computation. The local completion delay of the task α i k ( t ) arriving at U D i at time slot t can be calculated as Equation  (6):
T i k ( t ) l o c = η i k ( t ) D i k ( t ) f i
Correspondingly, its energy consumption can be calculated as (7) [33]
E i k ( t ) l o c = k 0 ( f i ) 3 T i k ( t ) l o c
where k 0 represents the hardware architecture parameter.
Edge Computation. The transmission delay of the task α i k ( t ) arriving at U D i at time slot t to be offloaded to U A V j can be calculated as Equation  (8):
T i , j k ( t ) P = D i k ( t ) R i , j k ( t )
And the edge computing delay of this process can be calculated as Equation  (9):
T i , j k ( t ) C = η i k ( t ) D i k ( t ) r j i , k ( t )
Accordingly, the energy consumption of transmission can be expressed as (10):
E i , j k ( t ) P = P i D i k ( t ) R i , j k ( t )
where P i represents the transmission power.
The energy consumption of edge computing can be expressed as Equation  (11):
E i , j k ( t ) C = ϖ η i k ( t ) D i k ( t )
where ϖ represents the energy consumption of the UAV per unit CPU cycle.

3.4. Cost Model

UD Cost. Similar to [34,35], from the user’s perspective, in each time period, the task completion delay T i ( t ) and UD energy consumption E i ( t ) are considered the cost of UD, which reflects the quality of experience of the UD.
In fact, the task calculations between each UD and UAV are parallel within a certain period of time, so simply optimizing the linear sum of all delays is meaningless to reduce the delay of the entire MEC process. Therefore, this paper borrows the concept of the minimum maximum completion time in the flexible workshop scheduling problem.
T i ( t ) can be formulated as (12)
T i ( t ) = Min Max T i l o c ( t ) , T i m e c ( t ) , T s
where
T i l o c ( t ) = k = 1 K ( 1 a i , j k ( t ) ) T i k ( t ) l o c
and
T i m e c ( t ) = Max { j J | Δ t j + i = 1 I k = 1 K a i , j k ( t ) T i , j k ( t ) C ) }
T i l o c ( t ) represents the sum of the latency of all local computing tasks of U D i . T i m e c ( t ) represents the edge task completion delay of U D i , that is, the time from the start of the time slot to the completion of the last task of U D i calculated by each UAV (makespan). k = 1 K ( 1 a i , j k ( t ) ) T i k ( t ) l o c is the total local processing time of U D i . The maximum value of T i m e c ( t ) and k = 1 K ( 1 a i , j k ( t ) ) T i k ( t ) l o c is the task response delay of time slot t. Since the task response delay cannot exceed the time slot length, the smaller one of this value and the time slot length is taken.
E i ( t ) can be presented as (15)
E i ( t ) = k = 1 K ( 1 a i , j k ( t ) ) E i k ( t ) l o c + k = 1 K a i , j k ( t ) E i , j k ( t ) P
Then, the cost of U D i at time slot t can be given as (16) [30,35]
C i ( t ) = γ i T i ( t ) + ( 1 γ i ) E i ( t )
The weighted parameters γ i of T i ( t ) and ( 1 γ i ) of E i ( t ) can be flexibly set according to the preference of the UD for latency and energy consumption.
UAV Energy Cost. The cost E j ( t ) of the U A V j in time slot t includes the computing energy consumption and movement energy consumption (17):
E j ( t ) = E j C ( t ) + E j P ( t )
where the computing energy consumption can be represented as (18)
E j C ( t ) = i = 1 I k = 1 K a i , j k ( t ) E i , j k ( t ) C
and the movement energy consumption can be represented as (19) [36]
E j P ( t ) = W j 2 ( v j ( t ) ) 2 T P
where W j is the weight of the U A V j , v j ( t ) is the flight speed of the U A V j , and T P is the travel time. Due to the randomness and uncertainty of the arrival of UD tasks, we consider that each UAV travels along a certain regular route, so this part of the energy consumption is fixed in [21,22].

3.5. Data Queue Model

A i ( t ) represents the task data that arrive at U D i at time t. In the actual process, they are random. Let Q i ( t ) denote the queue length of the i t h UD at the beginning of the time frame. Then, the queue dynamics can be modeled as
Q i ( t + 1 ) = max Q i ( t ) D ¯ i ( t ) + A i ( t ) , 0 , i I
where D ¯ i = m i n ( Q i ( t ) , D i ( t ) ) , D i ( t ) represents the amount of task data processed at time t, and Q i ( 1 ) = 0 . In this paper, we consider the infinite queuing capacity for analytical tractability. We force Q i ( t ) 0 to always hold. Then, the queue dynamics are simplified to
Q i ( t + 1 ) = Q i ( t ) D i ( t ) + A i ( t ) , i I
Definition 1.
A discrete time queue Q i ( t ) is strongly stable if the time average queue length lim K 1 T t = 1 T E [ Q i ( t ) ] , where the expectation is taken with respect to the system random events [30], i.e., task data arrivals in this paper.
By the Little’s law, the average delay is proportional to the average queue length. Therefore, a strongly stable data queue translates into a bounded processing delay for each task data bit.

3.6. Problem Formulation

Based on the above model, we minimize the average cost of the entire MEC system over time by jointly optimizing unloading decisions and resource allocation. The joint optimization problem can be expressed as (22) [30]:
P : min A , F , W 1 T t = 1 T i = 1 I C i ( t )
s . t . lim T + 1 T t = 1 T E { E j ( t ) } E j ¯
a i , j k ( t ) { 0 , 1 } , i I , k K , j J , t T
0 r j i , k ( t ) r j , j J , t T
i = 1 I k = 1 K a i , j k ( t ) r j i , k ( t ) r j , i I , k K , j J , t T
lim T 1 T · t = 1 T E [ Q i ( t ) ] , t T
where E j ¯ is the energy upper limit of a single time slot, and F max is the total amount of computing resources. Constraint (23) is to ensure that the energy consumption of each UAV in each time slot does not exceed the upper limit. The goal of (24) is to ensure that all computing tasks arriving at the UD can be processed. (25) and (26) are communication resource constraints. (27) is the data queue stability constraint.
The solution to problem P depends on long-term information, such as the arrival of tasks at different times. More seriously, problem P contains two decision variables, task offloading and resource allocation, and is a Mixed Integer Nonlinear Programming problem (MINLP). Even if the future information is known, it is challenging to solve the problem directly.

4. Lyapunov-Based Decoupling of the Multi-Slot MINLP

In this section, we apply Lyapunov optimization to decouple the problem P into a per-frame deterministic problem. Since the problem P is related to future information, we use Lyapunov optimization to decompose the conditions with long-term constraints (i.e., the long-term energy constraints of the UAV (23)) into each time slot.
First, in order to satisfy constraint (23), a virtual energy queue is introduced: UAV computing energy queue Y j c ( t ) . Assume that the queue is set to zero at the initial time slot, Y j c ( t ) = 0 . And it is updated as (28)
Y j c ( t + 1 ) = max Y j c ( t ) + E j c ( t ) E ¯ j c , 0 , t T
where E ¯ j c represents the computational energy budget of U A V j in each time slot. The backlog state of the energy queue is represented by the vector Θ j ( t ) = { Y j c ( t ) } , and the scalar measure of the backlog of the energy queue can be represented by the function L ( Θ j ( t ) ) (29):
L ( Θ j ( t ) ) = ( Y j c ( t ) ) 2
Secondly, we define the Lyapunov drift, which is used to calculate the increment δ Θ j ( t ) of the energy queue backlog from time slot t to time slot t + 1 (30):
δ θ j ( t ) = L ( Θ j ( t + 1 ) ) L ( Θ j ( t ) )
The drift penalty expression for problem P can be given as (31)
P 1 : min A , F , W δ θ j ( t ) + V i = 1 I C i ( t ) s . t . ( 24 ) ,   ( 25 ) ,   ( 26 ) ,   ( 27 )
where V is the parameter that weighs the total cost against the stability of the queue. Then, the scaling transformation of problem P is as follows:
δ θ j ( t ) + V P ( t ) = 1 2 j = 1 J ( ( Y j ( t + 1 ) ) 2 ( Y j ( t ) ) 2 ) + V P ( t )
1 2 j = 1 J ( y j ( t ) ) 2 + j = 1 J Y j ( t ) y j ( t ) + V P ( t )
B m a x + j = 1 J Y j ( t ) y j ( t ) + V P ( t )
where Y j ( t ) represents the energy queue, y j ( t ) represents the optimization goal, P ( t ) represents the constraint condition, and B m a x represents the positive constant of the constraint condition.
Finally, the drift penalty expression for problem P after scaling is shown as (35)
P 2 : min A t , F t , W t j = 1 J Y j c ( t ) E j c ( t ) + V i = 1 I C i ( t )
s . t . a i , j k ( t ) { 0 , 1 } , i I , k K , j J , t T
0 r j i , k ( t ) r j , j J , t T
i = 1 I k = 1 K a i , j k ( t ) r j i , k ( t ) r j , i I , k K , j J , t T
and we remove the irrelevant constant term from P 2 and obtain the problem P 2 (39):
P 2 : min A t , F t , W t j = 1 J Y j c ( t ) E j c ( t ) V + i = 1 I C i ( t ) s . t . ( 36 ) ,   ( 37 ) ,   ( 38 )
Since the Lyapunov drift relies on information in the t + 1 time slot, the problem P is scaled to become a real-time optimization problem that only needs information in a certain time slot. Due to the limited space, please refer to this article [37] for the related reasoning and proof.
Nevertheless, the new problem P 2 is still an MINLP problem. The two decision variables, task offloading strategy and resource allocation strategy, still affect each other. In order to solve the optimal solution of the P 2 problem, we transform the complex P 2 problem into two sub-problems: P 3 and P 4 . Then, we design a joint optimization method to solve it.

5. Joint Optimization Algorithm

5.1. Optimal Resource Allocation Algorithm

Although the problem P 2 is still a multi-slot MINLP optimization problem, when considering a certain slot, we are surprised to find a subtle relationship between the two decision variables. That is, given the slot task offloading strategy, there is and will be a unique set of resource allocation strategies corresponding to it so that the objective function value of the slot reaches the optimal value. The derivation of this subtle relationship is as follows:
When the task offload strategy A t ϕ is given, the UAV resource allocation strategy is adjusted to optimize the objective function P 3 (40):
P 3 : min F t ϕ , W t ϕ i = 1 I γ i T i ( t ) + i = 1 I k = 1 K ( 1 γ i ) a i , j k ( t ) E i , j k ( t ) P s . t . ( 36 ) ,   ( 37 ) ,   ( 38 )
Substituting Formulas (8)–(10) into problem P 3 , we can obtain problem P 3 after removing the fixed constant term and simplifying the coefficients (41)
P 3 : min k 1 n 1 + k 2 n 2 + k 3 n 3 + + k ζ n ζ +
s . t . n 1 ,   n 2 ,   n 3 , ,   n ζ , ,   n l a s t 0
n 1 + n 2 + n 3 + + n ζ + + n l a s t N ,
where n ζ represents the computing resources allocated to the task numbered ζ of all tasks unloaded to U A V j , k ζ (known quantity) is the simplified coefficient in the objective function, and N represents the total computing resources of U A V j .
Taking computing resources as an example, problem P 3 becomes problem P 3 , which can be processed using the Lagrange multiplier method. It is proved that problem P 3 has only one solution (Appendix A) [38]. Based on this, we firstly design Algorithm 1 to solve the optimal resource and bandwidth allocation strategy when a given task offloading strategy is given.   
Algorithm 1: Algorithm for optimal resource allocation of ( P 3 ).
Drones 09 00274 i001
Now that we have the optimal resource and bandwidth allocation strategies F t and W t , solving the problem P 2 is equivalent to finding the optimal task offloading strategy A t . To do this, we introduce a new subproblem P 4 of P 2 (44):
P 4 : ( A t ) = arg min F t , W t j = 1 J Y j c ( t ) E j c ( t ) V + i = 1 I C i ( t ) s . t . ( 36 )
Considering the different ways of offloading tasks of U D i , we can introduce the optimization cost function O i ( A t ) (45) of the U D i task offloading strategy to simplify the problem P 4 into P 4 :
O i ( A t ) = j = 1 J Y j c ( t ) E i , j k ( t ) C V + γ i T i m e c ( t ) + ( 1 γ i ) E i m e c ( t ) a i , j k ( t ) = 1 γ i T i l o c ( t ) + ( 1 γ i ) E i l o c ( t ) a i , j k ( t ) = 0
Then, we can obtain P 4  (46):
P 4 : ( A t ) = arg min F t , W t O i ( A t ) s . t . ( 36 )
To tackle problem P 4 , we propose a Lyapunov-guided reinforcement learning-based rapid and durable optimization algorithm (LyraRD) framework. Traditional branch and bound and block coordinate descent methods require enumerating 2 N unloading strategies, which brings great complexity even when N is a moderate value (for example, N = 10). Using RL techniques, we propose a LyraRD algorithm to construct a policy π that maps from input I ( t ) to optimal actions x ( t ) with very low complexity, for example, requiring tens of milliseconds of computation time (that is, the duration from observing a state to producing a control action x ( t ) ) when N = 10.

5.2. LyraRD Algorithm Description

As shown in Figure 3, LyraRD consists of four main modules: the actor module accepts input I ( t ) and produces a set of candidate offloading actions x ( t ) , the critic module evaluates x ( t ) and selects the best offloading action x ( t ) , the policy update module improves the actor module’s policy over time, and the queue module updates the state of the system queue Y ( t ) after performing an offloading action. Through repeated interactions with the stochastic environment A ( t ) , the four modules operate sequentially and iteratively as described below.
(1) Actor module: The actor module consists of a DNN and an action quantizer. The observation value I ( t ) is taken as input, where it includes the amount of task data A ( t ) generated by each UD at the current time slot and the virtual energy queue value Y ( t ) for each UAV that performs long-term energy constraints. The network architecture of DNN consists of an input layer, a convolutional layer, and a fully connected layer. The features are extracted by convolving the input I ( t ) with three layers through the ReLU activation function. The output of the convolutional layer is flattened for input to the fully connected layer. The fully connected layer, containing 64 neurons with the ReLU activation function, is used to integrate the global features. The DNN outputs a relaxed unloading decision x ( t ) [ 0 , 1 ] N . Then, a sigmoid activation function is used in the output layer to quantize the continuous x ( t ) into feasible candidate binary offloading actions M t , where M t is a time-dependent design parameter. The quantization function is expressed as
x ( t ) g = x i ( t ) | x i ( t ) { 0 , 1 } N , i = 1 , , M t
where g denotes the set of candidate offloading actions in the t t h time frames.
(2) Critic module: Followed by the actor module, the critic module evaluates x i ( t ) and selects the best unloading action x ( t ) according to formula (46):
L S ( θ t ) = 1 | S t | T S t x T ( t ) log Π θ t ( I ( t ) ) + ( 1 x ( t ) ) T log ( 1 Π θ t ( I ( t ) ) )
(3) Policy Update Module: LyraRD uses ( I ( t ) , x ( t ) ) as the labeled input–output samples to update the DNN policy. In particular, we maintain a replay memory that only stores the most recent data samples q. In practice, with the memory initially empty, we start training the DNN after collecting more than q / 2 data samples. The DNN then is periodically trained every δ t time interval to avoid overfitting the model. We then use the Adam algorithm [39] to minimize the average cross entropy loss function L S ( θ t ) (48) on the data samples to update the parameters of the DNN.
(4) Queuing module: As a by-product of the critic module, we obtain the optimal allocation of resources associated with x ( t ) . Then, we update the energy queue based on them and the newly arrived task data and iterate again from the actor module as the next round of system input.
The entire process of LyraRD algorithm is shown in Algorithm 2. The algorithm has been shown to have lower complexity and better convergence (LyraRD algorithm has been proven to have extremely low complexity and excellent convergence. For details, refer to [40]).
Algorithm 2: The online LyraRD algorithm for solving (P4’).
Drones 09 00274 i002

6. Simulation Results

In this section, we perform simulations to verify the effectiveness of the proposed method.

6.1. Platforms and Tasks Data

We conduct simulation experiments on a computer with the following specifications: CPU: Intel i7-13650HX 3.2 GHz; GPU: NVIDIA RTX 4060; RAM: 64 GB; OS: 64-bit Ubuntu 16.04. All algorithms are developed using Python 3.7.
We consider a MEC system enabled by UAVs consisting of 4 UAVs and 20 UDs, where the initial horizontal positions of the UAVs are set to P1(0) = [10, 10], P2(0) = [10, 80], P3(0) = [90, 20], P4(0) = [75, 85], respectively, with a fixed height of H = 100 m. The UDs appear randomly in an area of 100 × 100 m 2 , and UAVs fly along fixed routes. The system timeline is discretized into 8000 time slots, each of which is 1 s long [41]. The remaining parameter settings are shown in the table in Appendix B.
In order to simulate the random arrival of data, we assume that each user reaches no more than three tasks at the same time, and the arrival forms of these tasks obey exponential, Poisson, and uniform distributions, respectively. That is, the task queue Q 1 E ( λ ) , Q 2 π ( λ ) , Q 3 U ( a , b ) . To compare the impact of different types of data arrival on the stability of the data queue, we make the expectation the same for all three distributions, i.e., λ = λ = ( a + b ) / 2 , where λ = λ = 3 , a = 0 , b = 6 . Next, we obtain the simulation results shown in Figure 4, Figure 5 and Figure 6.

6.2. Overall Performance Analysis

Considering the randomness of data arrival, we fully conduct several experiments and select the four most representative experimental results to verify the performance of the proposed method. Figure 4 shows the training loss of the algorithm, which fully illustrates the convergence of the LyraRD algorithm.
Further, we observe the use of energy as shown in Figure 5, where the energy cohort fluctuates over the initial time period and gradually stabilizes. In Figure 5b–d, the energy queue still has a second peak after the initial fluctuation, but it gradually plateaus after that.
Based on the fluctuations in the results, we further investigate the arrival and processing of the data. First, observe the data queue, as shown in Figure 6; the data queue basically reaches a stable state after training. However, the data cohorts in Figure 6b–d all exhibit secondary fluctuations to varying degrees. In response to these fluctuations, we recover the specific values of the data in the queue during that period and analyze them.
Single UD data accumulation. In Figure 6b, the energy queue of the second peak is given by Figure 7. We find that the data queue of the 20th user has an abnormal accumulation that is significantly higher than that of other users. This abnormal accumulation may be due to the excessive amount of random data arriving from this user in the previous moment, resulting in untimely data calculation. In the short period of time after that, our algorithm processes these backlogged data, thus ensuring that the data queue returns to stability in the subsequent moments. This also confirms the energy consumption in Figure 5b.
Task-oriented offloading. The second peak value in Figure 6c is lower, so we speculate that there is no backlog in the data queue, but it is caused by more data being offloaded to the UAV for calculation. For this reason, we also observe the user data queue at this time. We can see in Figure 8 that the amount of data that need to be processed in the user’s data queue is very small, indicating that most of the data are offloaded to the UAV for calculation. The peak of the test energy consumption comes from the large amount of calculation of the UAV.
Large-scale distributed tasks. Then, we retrieve the abnormal data queue value in Figure 6d, as shown in Figure 9. It can be found that the pending value of each user’s subtask queue data is between 20 and 45. However, the sum of the pending data of Q1, Q2, and Q3 is 560.52, 556, and 576.38 respectively, which exceeds the data processing limit of the user and UAV in a period of time, thus causing data accumulation. As the task is continuously unloaded, the data queue returns to stability, and the energy consumption also tends to be stable.
The above results fully demonstrate that our method maintains stable data processing capabilities when dealing with randomly arriving data, while ensuring uniform and stable energy usage throughout the entire process, and can effectively cope with the sudden arrival of large-scale data.

6.3. Baseline Methods

Moreover, we compare LyraRD with the following benchmark schemes:
  • Average latency of tasks ( A L T ): The optimization goal is the linear sum of the average task completion delay.
  • Equal resource allocation ( E R A ): The UAV allocates computing and communication resources equally.

6.4. Evaluation Result

Figure 10 shows the case of cost in the three scenarios. The scheme proposed in this paper is O B J . First, when the value of V is not large (not more than 40), the cost case of O B J is not significantly lower than the other two comparative cases as expected, which is caused by the value of V being small and the cost being too focused on the stability of the energy-consuming queue. When the value of V is greater than or equal to 50, it can be seen that the cost of O B J is significantly lower than the cost case of A L T , and O B J shows good performance in combining the queue stability and the time overhead of the MEC system with the energy overhead of the UAV. Counterintuitively, E R A has a lower cost than O B J and A L T in some cases compared to O B J and A L T . This is due to the fact that computational task arrivals are randomized, and when generating offloading actions based on these task data, E R A outperforms O B J and A L T in terms of UAV resource allocation in the following scenario, i.e., a chaotic scenario where very few tasks are offloaded to the UAV. In spite of this, the E R A scheme does not demonstrate a clear superiority at cost, especially in resource-constrained scenarios with large-scale task arrivals. Finally, it can be seen that the proposed O B J scheme is able to adapt to different task data sizes and has relatively superior performance in terms of cost for MEC systems.

7. Conclusions

In this paper, we study a UAV-enabled MEC system that satisfies long-term energy consumption constraints with high response efficiency. The joint optimization of the offloading decision and resource allocation strategy is investigated to minimize the system cost. To solve this problem, we propose a joint optimization algorithm. Specifically, the multi-stage MINLP problem is transformed into a MINLP problem within a single time period using Lyapunov optimization, which is later solved using a reinforcement learning approach. Simulation results show that the method significantly reduces the system cost compared to the benchmark solution. However, our work has some limitations that need to be addressed in future work. First, the current model does not explicitly consider the motion patterns of the models, which may limit the applicability of the results in highly dynamic scenarios. Second, adaptive collaboration mechanisms for multi-UAV systems have been understudied. In the future, we will integrate realistic UD motion models and explore the interactions between dynamic UD behaviors and UAV trajectory optimization, as well as further investigating the adaptive collaboration of UAVs to enhance the utility of the framework. These extensions aim to provide more comprehensive solutions for complex UAV-based MEC networks in the real world.

Author Contributions

Conceptualization, H.Z. and T.C.; methodology, H.Z., B.R. and T.C.; software, H.Z.; validation, H.Z. and T.C.; formal analysis, H.Z. and B.R.; investigation, H.Z., T.C. and B.R.; resources, H.Z., T.C. and B.R.; writing—original draft preparation, H.Z., T.C., B.R. and R.L.; writing—review and editing, H.Z., T.C., B.R. and H.Y.; supervision, B.R. and T.C.; funding acquisition, B.R. and T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partically supported by the National Natural Science Foundation of China (No. 62302510). The authors would like to thank the support by the COSTA: complex system optimization team of the College of System Engineering at NUDT.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of the First Zonklar Equation

f ( n 1 ,   n 2 ,   n 3 , ,   n i ) = k 1 n 1 + k 2 n 2 + k 3 n 3 + + k i n i
subject to the constraints that n 1 ,   n 2 ,   n 3 , ,   n i 0 and n 1 + n 2 + n 3 + + n i N .
To solve this problem using the method of Lagrange multipliers, we introduce a Lagrange multiplier λ and construct the Lagrangian function:
L ( n 1 ,   n 2 , ,   n i ,   λ ) = j = 1 i k j n j + λ N j = 1 i n j
Taking the partial derivatives of L with respect to each n j and λ , and setting them to zero, we obtain:
L n j = k j n j 2 λ = 0 for j = 1 , 2 , , i
L λ = N n 1 n 2 n 3 n i = 0
From the first set of equations, we can solve for λ :
λ = k j n j 2
Since λ is the same for all j, we can set the expressions for λ equal to each other:
k 1 n 1 2 = k 2 n 2 2 = k 3 n 3 2 = = k i n i 2
This implies
k 1 n 1 2 = k 2 n 2 2 = k 3 n 3 2 = = k i n i 2
Let this common value be C, then
n j 2 = k j C
n j = k j C
Substituting n j into the constraint n 1 + n 2 + n 3 + + n i = N ,
k 1 C + k 2 C + k 3 C + + k i C = N
j = 1 i k j = N C
C = j = 1 i k j N 2
Finally, substituting C back into the expression for n j ,
n j = k j C = k j j = 1 i k j N 2 = N k j j = 1 i k j
Therefore, the values of n j are
n j = N k j j = 1 i k j

Appendix B. Parameter Settings

Drones 09 00274 i003

References

  1. Dash, S.; Ahmad, M.; Iqbal, T. Mobile cloud computing: A green perspective. In Intelligent Systems: Proceedings of ICMIB 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 523–533. [Google Scholar]
  2. Xu, Y.; Zhang, T.; Liu, Y.; Yang, D.; Xiao, L.; Tao, M. UAV-assisted MEC networks with aerial and ground cooperation. IEEE Trans. Wirel. Commun. 2021, 20, 7712–7727. [Google Scholar] [CrossRef]
  3. Ranaweera, P.; Jurcut, A.; Liyanage, M. MEC-enabled 5G use cases: A survey on security vulnerabilities and countermeasures. ACM Comput. Surv. (CSUR) 2021, 54, 186. [Google Scholar]
  4. Zhao, R.; Fan, C.; Ou, J.; Fan, D.; Ou, J.; Tang, M. Impact of direct links on intelligent reflect surface-aided MEC networks. Phys. Commun. 2022, 55, 101905. [Google Scholar]
  5. Dai, P.; Song, F.; Liu, K.; Dai, Y.; Zhou, P.; Guo, S. Edge intelligence for adaptive multimedia streaming in heterogeneous internet of vehicles. IEEE Trans. Mob. Comput. 2021, 22, 1464–1478. [Google Scholar]
  6. Khan, M.A.; Baccour, E.; Chkirbene, Z.; Erbad, A.; Hamila, R.; Hamdi, M.; Gabbouj, M. A survey on mobile edge computing for video streaming: Opportunities and challenges. IEEE Access 2022, 10, 120514–120550. [Google Scholar]
  7. Yang, C.; Xu, X.; Zhou, X.; Qi, L. Deep Q network–driven task offloading for efficient multimedia data analysis in edge computing–assisted IoV. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2022, 18, 124. [Google Scholar] [CrossRef]
  8. Pervez, F.; Sultana, A.; Yang, C.; Zhao, L. Energy and latency efficient joint communication and computation optimization in a multi-UAV assisted MEC network. IEEE Trans. Wirel. Commun. 2023, 23, 1728–1741. [Google Scholar]
  9. Du, J.; Wang, J.; Sun, A.; Qu, J.; Zhang, J.; Wu, C.; Niyato, D. Joint optimization in blockchain and mec enabled space-air-ground integrated networks. IEEE Internet Things J. 2024, 11, 31862–31877. [Google Scholar]
  10. Zhan, C.; Hu, H.; Sui, X.; Liu, Z.; Niyato, D. Completion time and energy optimization in the UAV-enabled mobile-edge computing system. IEEE Internet Things J. 2020, 7, 7808–7822. [Google Scholar] [CrossRef]
  11. Zeng, Y.; Chen, S.; Cui, Y.; Yang, J.; Fu, Y. Joint resource allocation and trajectory optimization in UAV-enabled wirelessly powered MEC for large area. IEEE Internet Things J. 2023, 10, 15705–15722. [Google Scholar] [CrossRef]
  12. Jiang, F.; Wang, K.; Dong, L.; Pan, C.; Xu, W.; Yang, K. AI driven heterogeneous MEC system with UAV assistance for dynamic environment: Challenges and solutions. IEEE Netw. 2020, 35, 400–408. [Google Scholar] [CrossRef]
  13. Yang, Z.; Bi, S.; Zhang, Y.J.A. Dynamic offloading and trajectory control for UAV-enabled mobile edge computing system with energy harvesting devices. IEEE Trans. Wirel. Commun. 2022, 21, 10515–10528. [Google Scholar] [CrossRef]
  14. Wan, S.; Lu, J.; Fan, P.; Letaief, K.B. Toward big data processing in IoT: Path planning and resource management of UAV base stations in mobile-edge computing system. IEEE Internet Things J. 2019, 7, 5995–6009. [Google Scholar] [CrossRef]
  15. Yang, Z.; Bi, S.; Zhang, Y.J.A. Online trajectory and resource optimization for stochastic UAV-enabled MEC systems. IEEE Trans. Wirel. Commun. 2022, 21, 5629–5643. [Google Scholar] [CrossRef]
  16. Yan, P.; Cao, Z.; Duan, W.; Li, B.; Zou, Y.; Li, C.; Wang, J. Securing UAV-Aided NOMA Wireless Powered Communications via Artificial Noise. IEEE Trans. Wirel. Commun. 2025. [Google Scholar] [CrossRef]
  17. AL-Bakhrani, A.A.; Li, M.; Obaidat, M.S.; Amran, G.A. MOALF-UAV-MEC: Adaptive Multi-Objective Optimization for UAV-Assisted Mobile Edge Computing in Dynamic IoT Environments. IEEE Internet Things J. 2025, 1. [Google Scholar] [CrossRef]
  18. Sheng, Z.; Hu, H.; Nasir, A.A.; Fang, Y.; da Costa, D.B. Online Trajectory Planning and Resource Allocation of UAV-Enabled MEC Networks Empowered by RIS. IEEE Trans. Green Commun. Netw. 2024, 1. [Google Scholar] [CrossRef]
  19. Zhang, L.; Zhang, Z.Y.; Min, L.; Tang, C.; Zhang, H.Y.; Wang, Y.H.; Cai, P. Task Offloading and Trajectory Control for UAV-Assisted Mobile Edge Computing Using Deep Reinforcement Learning. IEEE Access 2021, 9, 53708–53719. [Google Scholar] [CrossRef]
  20. Yuan, H.; Wang, M.; Bi, J.; Shi, S.; Yang, J.; Zhang, J.; Zhou, M.; Buyya, R. Cost-Efficient Task Offloading in Mobile Edge Computing With Layered Unmanned Aerial Vehicles. IEEE Internet Things J. 2024, 11, 30496–30509. [Google Scholar] [CrossRef]
  21. Michailidis, E.T.; Volakaki, M.G.; Miridakis, N.I.; Vouyioukas, D. Optimization of Secure Computation Efficiency in UAV-Enabled RIS-Assisted MEC-IoT Networks With Aerial and Ground Eavesdroppers. IEEE Trans. Commun. 2024, 72, 3994–4009. [Google Scholar] [CrossRef]
  22. Lakew, D.S.; Tran, A.T.; Dao, N.N.; Cho, S. Intelligent Self-Optimization for Task Offloading in LEO-MEC-Assisted Energy-Harvesting-UAV Systems. IEEE Trans. Netw. Sci. Eng. 2024, 11, 5135–5148. [Google Scholar] [CrossRef]
  23. Wang, J.; Wang, L.; Zhu, K.; Dai, P. Lyapunov-Based Joint Flight Trajectory and Computation Offloading Optimization for UAV-Assisted Vehicular Networks. IEEE Internet Things J. 2024, 11, 22243–22256. [Google Scholar] [CrossRef]
  24. Qin, P.; Wu, X.; Fu, M.; Ding, R.; Fu, Y. Latency Minimization Resource Allocation and Trajectory Optimization for UAV-Assisted Cache-Computing Network with Energy Recharging. IEEE Trans. Commun. 2025, 1. [Google Scholar] [CrossRef]
  25. Li, J.; Sun, G.; Wu, Q.; Niyato, D.; Kang, J.; Jamalipour, A.; Leung, V.C. Collaborative ground-space communications via evolutionary multi-objective deep reinforcement learning. IEEE J. Sel. Areas Commun. 2024, 42, 3395–3411. [Google Scholar]
  26. Consul, P.; Budhiraja, I.; Garg, D.; Garg, S.; Kaddoum, G.; Hassan, M.M. SFL-TUM: Energy efficient SFRL method for large scale AI model’s task offloading in UAV-assisted MEC networks. Veh. Commun. 2024, 48, 100790. [Google Scholar] [CrossRef]
  27. Wu, X.; Zhu, Q.; Chen, W.N.; Lin, Q.; Li, J.; Coello, C.A.C. Evolutionary reinforcement learning with action sequence search for imperfect information games. Inf. Sci. 2024, 676, 120804. [Google Scholar]
  28. Qu, Y.; Dai, H.; Wang, H.; Dong, C.; Wu, F.; Guo, S.; Wu, Q. Service provisioning for UAV-enabled mobile edge computing. IEEE J. Sel. Areas Commun. 2021, 39, 3287–3305. [Google Scholar]
  29. Jiang, H.; Dai, X.; Xiao, Z.; Iyengar, A. Joint task offloading and resource allocation for energy-constrained mobile edge computing. IEEE Trans. Mob. Comput. 2022, 22, 4000–4015. [Google Scholar] [CrossRef]
  30. He, L.; Sun, G.; Sun, Z.; Wang, P.; Li, J.; Liang, S.; Niyato, D. An Online Joint Optimization Approach for QoE Maximization in UAV-Enabled Mobile Edge Computing. arXiv 2024, arXiv:2404.02166. [Google Scholar]
  31. Sun, G.; Zheng, X.; Sun, Z.; Wu, Q.; Li, J.; Liu, Y.; Leung, V.C. UAV-enabled secure communications via collaborative beamforming with imperfect eavesdropper information. IEEE Trans. Mob. Comput. 2023, 23, 3291–3308. [Google Scholar] [CrossRef]
  32. Ndikumana, A.; Tran, N.; Ho, T.; Han, Z.; Saad, W.; Niyato, D.; Hong, C. Joint Communication, Computation, Caching, and Control in Big Data Multi-access Edge Computing. IEEE Trans. Mob. Comput. 2018, 19, 1359–1374. [Google Scholar] [CrossRef]
  33. Zhang, X.; Zhang, J.; Xiong, J.; Zhou, L.; Wei, J. Energy-Efficient Multi-UAV-Enabled Multiaccess Edge Computing Incorporating NOMA. IEEE Internet Things J. 2020, 7, 5613–5627. [Google Scholar] [CrossRef]
  34. Chen, Y.; Zhao, J.; Wu, Y.; Huang, J.; Shen, X. QoE-Aware Decentralized Task Offloading and Resource Allocation for End-Edge-Cloud Systems: A Game-Theoretical Approach. IEEE Trans. Mob. Comput. 2024, 23, 769–784. [Google Scholar] [CrossRef]
  35. Ding, Y.; Li, K.; Liu, C.; Li, K. A Potential Game Theoretic Approach to Computation Offloading Strategy Optimization in End-Edge-Cloud Computing. IEEE Trans. Parallel Distrib. Syst. 2022, 33, 1503–1519. [Google Scholar] [CrossRef]
  36. Xu, B.; Kuang, Z.; Gao, J.; Zhao, L.; Wu, C. Joint offloading decision and trajectory design for UAV-enabled edge computing with task dependency. IEEE Trans. Wirel. Commun. 2022, 22, 5043–5055. [Google Scholar] [CrossRef]
  37. Neely, M. Stochastic Network Optimization with Application to Communication and Queueing Systems; Springer Nature: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  38. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press eBooks; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  39. Marsland, S. Machine Learning: An Algorithmic Perspective; Chapman and Hall/CRC: Boca Raton, FL, USA, 2009. [Google Scholar]
  40. Bi, S.; Huang, L.; Wang, H.; Zhang, Y.J.A. Lyapunov-guided deep reinforcement learning for stable online computation offloading in mobile-edge computing networks. IEEE Trans. Wirel. Commun. 2021, 20, 7519–7537. [Google Scholar] [CrossRef]
  41. Wang, L.; Wang, K.; Pan, C.; Xu, W.; Aslam, N.; Nallanathan, A. Deep Reinforcement Learning Based Dynamic Trajectory Control for UAV-assisted Mobile Edge Computing. IEEE Trans. Mob. Comput. 2022, 21, 3536–3550. [Google Scholar] [CrossRef]
Figure 1. The UAV-enabled MEC system consists of several UAVs and multiple ground-based user devices (UDs). UAVs provide computing services to UDs through allocating communication and computing resources. Each UD randomly arrives at several tasks at the beginning of a time slot and independently decides to compute tasks locally or offload tasks to UAVs.
Figure 1. The UAV-enabled MEC system consists of several UAVs and multiple ground-based user devices (UDs). UAVs provide computing services to UDs through allocating communication and computing resources. Each UD randomly arrives at several tasks at the beginning of a time slot and independently decides to compute tasks locally or offload tasks to UAVs.
Drones 09 00274 g001
Figure 2. (a) The task offloading method with the smallest makespan. (b) A reasonable task offloading method with energy consumption balance and makespan.
Figure 2. (a) The task offloading method with the smallest makespan. (b) A reasonable task offloading method with energy consumption balance and makespan.
Drones 09 00274 g002
Figure 3. LyraRD algorithm flow.
Figure 3. LyraRD algorithm flow.
Drones 09 00274 g003
Figure 4. Four cases of Training loss.
Figure 4. Four cases of Training loss.
Drones 09 00274 g004
Figure 5. Four cases of loss energy queue.
Figure 5. Four cases of loss energy queue.
Drones 09 00274 g005
Figure 6. Four cases of loss data queue.
Figure 6. Four cases of loss data queue.
Drones 09 00274 g006
Figure 7. Exception energy queue of Figure 6b.
Figure 7. Exception energy queue of Figure 6b.
Drones 09 00274 g007
Figure 8. Exception energy queue of Figure 6c.
Figure 8. Exception energy queue of Figure 6c.
Drones 09 00274 g008
Figure 9. Exception energy queue of Figure 6d.
Figure 9. Exception energy queue of Figure 6d.
Drones 09 00274 g009
Figure 10. Cost values of the three scenarios for different number of UDs with different values of v.
Figure 10. Cost values of the three scenarios for different number of UDs with different values of v.
Drones 09 00274 g010
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Chen, T.; Ren, B.; Li, R.; Yuan, H. Joint Optimization of Task Completion Time and Energy Consumption in UAV-Enabled Mobile Edge Computing. Drones 2025, 9, 274. https://doi.org/10.3390/drones9040274

AMA Style

Zhang H, Chen T, Ren B, Li R, Yuan H. Joint Optimization of Task Completion Time and Energy Consumption in UAV-Enabled Mobile Edge Computing. Drones. 2025; 9(4):274. https://doi.org/10.3390/drones9040274

Chicago/Turabian Style

Zhang, Hanwen, Tao Chen, Bangbang Ren, Ruozhe Li, and Hao Yuan. 2025. "Joint Optimization of Task Completion Time and Energy Consumption in UAV-Enabled Mobile Edge Computing" Drones 9, no. 4: 274. https://doi.org/10.3390/drones9040274

APA Style

Zhang, H., Chen, T., Ren, B., Li, R., & Yuan, H. (2025). Joint Optimization of Task Completion Time and Energy Consumption in UAV-Enabled Mobile Edge Computing. Drones, 9(4), 274. https://doi.org/10.3390/drones9040274

Article Metrics

Back to TopTop