Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (306)

Search Parameters:
Keywords = offloading decision

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 1501 KB  
Article
MA-JTATO: Multi-Agent Joint Task Association and Trajectory Optimization in UAV-Assisted Edge Computing System
by Yunxi Zhang and Zhigang Wen
Drones 2026, 10(4), 267; https://doi.org/10.3390/drones10040267 - 7 Apr 2026
Abstract
With the rapid development of applications such as smart cities and the industrial internet, the computation-intensive tasks generated by massive sensing devices pose significant challenges to traditional cloud computing paradigms. Unmanned aerial vehicle (UAV)-assisted edge computing systems, leveraging their high mobility and wide-area [...] Read more.
With the rapid development of applications such as smart cities and the industrial internet, the computation-intensive tasks generated by massive sensing devices pose significant challenges to traditional cloud computing paradigms. Unmanned aerial vehicle (UAV)-assisted edge computing systems, leveraging their high mobility and wide-area coverage capabilities, offer an innovative architecture for low-latency and highly reliable edge services. However, the practical deployment of such systems faces a highly complex multi-objective optimization problem featured by the tight coupling of task offloading decisions, UAV trajectory planning, and edge server resource allocation. Conventional optimization methods are difficult to adapt to the dynamic and high-dimensional characteristics of this problem, leading to suboptimal system performance. To address this critical challenge, this paper constructs an intelligent collaborative optimization framework for UAV-assisted edge computing systems and formulates the system quality of service (QoS) optimization problem as a mixed-integer non-convex programming problem with the dual objectives of minimizing task processing latency and reducing overall system energy consumption. A multi-agent joint task association and trajectory optimization (MA-JTATO) algorithm based on hybrid reinforcement learning is proposed to solve this intractable problem, which innovatively decouples the original coupled optimization problem into three interrelated subproblems and realizes their collaborative and efficient solution. Specifically, the Advantage Actor-Critic (A2C) algorithm is adopted to realize dynamic and optimal task association between UAVs and edge servers for discrete decision-making requirements; the multi-agent deep deterministic policy gradient (MADDPG) method is employed to achieve cooperative and energy-efficient trajectory planning for multiple UAVs to meet the needs of continuous control in dynamic environments; and convex optimization theory is applied to obtain a closed-form optimal solution for the efficient allocation of computational resources on edge servers. Simulation results demonstrate that the proposed MA-JTATO algorithm significantly outperforms traditional baseline algorithms in enhancing overall QoS, effectively validating the framework’s superior performance and robustness in dynamic and complex scenarios. Full article
(This article belongs to the Section Drone Communications)
Show Figures

Figure 1

28 pages, 2619 KB  
Article
A Dynamic Clustering Framework for Intelligent Task Orchestration in Mobile Edge Computing
by Mona Alghamdi, Atm S. Alam and Asma Cherif
Computers 2026, 15(4), 214; https://doi.org/10.3390/computers15040214 - 1 Apr 2026
Viewed by 282
Abstract
Mobile edge computing (MEC) enables resource-constrained mobile devices to execute delay-sensitive and compute-intensive applications by offloading tasks to nearby edge servers. However, task orchestration in MEC is challenged by the highly dynamic system conditions, unreliable networks, and distributed edge environments. Moreover, as the [...] Read more.
Mobile edge computing (MEC) enables resource-constrained mobile devices to execute delay-sensitive and compute-intensive applications by offloading tasks to nearby edge servers. However, task orchestration in MEC is challenged by the highly dynamic system conditions, unreliable networks, and distributed edge environments. Moreover, as the number of mobile users, tasks, and distributed computing resources (edge/cloud servers) increases, the task orchestration process becomes more complex due to the expanded decision space and the need to efficiently allocate heterogeneous resources under latency and capacity constraints. As the decision space grows, exhaustive-search-based orchestration becomes computationally infeasible. Clustering approaches often rely on proximity-only grouping, while learning-based solutions require extensive training and parameter tuning. To address these challenges, this paper proposes a Multi-Criteria Hierarchical Clustering-based Task Orchestrator (MCHC-TO), a novel framework that integrates multi-criteria decision making with divisive hierarchical clustering for preference-aware and adaptive workload orchestration. Edge servers are first evaluated using multiple decision criteria, and the resulting preference rankings are exploited to form hierarchical preference-based clusters. Incoming tasks are then assigned to the most suitable cluster based on task requirements, enabling efficient resource utilization and dynamic decision-making. Extensive simulations conducted using an edge computing simulator demonstrate that the proposed MCHC-TO framework consistently outperforms benchmark approaches, achieving reductions in average service delay and task failure rate of up to 48% and 92%, respectively. These results highlight the effectiveness of combining multi-criteria evaluation with hierarchical clustering for robust and dynamic task orchestration in MEC environments. Full article
(This article belongs to the Special Issue Mobile Fog and Edge Computing)
Show Figures

Graphical abstract

32 pages, 2463 KB  
Review
Artificial Intelligence and Youth: Cognitive, Educational, and Behavioral Impacts
by Daniele Giansanti and Claudia Cosenza
AI 2026, 7(4), 121; https://doi.org/10.3390/ai7040121 - 1 Apr 2026
Viewed by 694
Abstract
Background: Artificial Intelligence (AI) and Generative AI (GenAI) are increasingly integrated into educational and professional settings, offering personalized learning, productivity gains, and enhanced engagement. However, excessive reliance may compromise critical thinking, autonomous problem-solving, and emotional regulation among youth (i.e., adolescents and young adults) [...] Read more.
Background: Artificial Intelligence (AI) and Generative AI (GenAI) are increasingly integrated into educational and professional settings, offering personalized learning, productivity gains, and enhanced engagement. However, excessive reliance may compromise critical thinking, autonomous problem-solving, and emotional regulation among youth (i.e., adolescents and young adults) and early-career professionals. Aim: This review examines the cognitive, educational, and behavioral impacts of AI and GenAI use in youth, highlighting implications for their responsible integration in learning and professional development. Methods: A narrative review was conducted, synthesizing empirical studies, psychometric instruments, and international policy frameworks addressing AI engagement. Emphasis was placed on cognitive, behavioral, educational, and ethical dimensions across youth and early-career professionals. Results: AI enhances learning efficiency, creativity, and professional decision-making but may also foster cognitive offloading, dependency, and addiction-like behaviors. Instruments such as the Conversational AI Dependence Scale (CAIDS) and the Problematic ChatGPT Use Scale (PCGUS) help identify maladaptive patterns. Effective strategies include structured pedagogy, human oversight, reflective practice, AI literacy, and ethical guidance. Paradoxically, higher AI competence and trust may increase reliance, underscoring the need for guided and balanced engagement. Conclusions: Responsible AI integration requires multidimensional approaches combining instructional scaffolding, metacognitive strategies, supervision, and governance to preserve autonomy, professional judgment, and cognitive development in youth. Full article
(This article belongs to the Special Issue How Is AI Transforming Education?)
Show Figures

Figure 1

28 pages, 4715 KB  
Article
Techno-Economic and SLA-Aware Control of 5G Cloud-RAN via Multi-Objective and Penalty-Constrained Reinforcement Learning
by Sherif M. Aboul, Hala M. Abd El Kader, Esraa M. Eid and Shimaa S. Ali
Network 2026, 6(2), 20; https://doi.org/10.3390/network6020020 - 31 Mar 2026
Viewed by 220
Abstract
Fifth-generation (5G) mobile networks must simultaneously satisfy stringent latency targets, high user density, and energy-aware operation across heterogeneous services. Cloud Radio Access Networks (C-RAN) provide architectural flexibility through centralized baseband processing, but they also introduce new control challenges related to fronthaul constraints, dynamic [...] Read more.
Fifth-generation (5G) mobile networks must simultaneously satisfy stringent latency targets, high user density, and energy-aware operation across heterogeneous services. Cloud Radio Access Networks (C-RAN) provide architectural flexibility through centralized baseband processing, but they also introduce new control challenges related to fronthaul constraints, dynamic traffic variations, and joint radio–compute coordination with Mobile Edge Computing (MEC). This paper proposes a unified AI-driven optimization framework for adaptive 5G C-RAN management, where the controller dynamically tunes key system decisions—including functional split selection, TDD downlink ratio, user–RU association, fronthaul load management, and MEC offloading proportion. To enable fair benchmarking under identical simulation settings, a static baseline policy is compared against five adaptive control strategies: Deep Q-Network (DQN), Proximal Policy Optimization (PPO), Deep Deterministic Policy Gradient (DDPG), Multi-Objective Reinforcement Learning (MORL), and a Deterministic Service-Level Agreement (SLA)-aware controller Penalty-Constrained Hierarchical Action Controller (PCHAC). Performance evaluation across techno-economic and service KPIs shows that intelligent control significantly improves operational profit, tail-latency behavior, and energy efficiency while enhancing SLA compliance compared with non-adaptive operation. The results highlight the practicality of multi-objective and constraint-aware learning for next-generation C-RAN orchestration under scaling traffic demand. Full article
Show Figures

Figure 1

31 pages, 1333 KB  
Article
Optimal Security Task Offloading in Cognitive IoT Networks: Provably Optimal Threshold Policies and Model-Free Learning
by Ning Wang and Yali Ren
IoT 2026, 7(2), 30; https://doi.org/10.3390/iot7020030 - 26 Mar 2026
Viewed by 332
Abstract
The proliferation of Internet of Things (IoT) devices has introduced significant security challenges. Resource-constrained devices face sophisticated threats but lack the computational capacity for advanced security analysis. This study investigates optimal security task allocation in Cognitive IoT (CIoT) networks. It specifically examines when [...] Read more.
The proliferation of Internet of Things (IoT) devices has introduced significant security challenges. Resource-constrained devices face sophisticated threats but lack the computational capacity for advanced security analysis. This study investigates optimal security task allocation in Cognitive IoT (CIoT) networks. It specifically examines when IoT devices should process security tasks locally or offload them to Mobile Edge Computing (MEC) servers. The problem is formulated as a Continuous-Time Markov Decision Process (CTMDP). The study demonstrates that the optimal offloading policy has a threshold structure. Security tasks are offloaded to MEC servers when the offloading queue length is below a critical threshold, k. Otherwise, tasks are processed locally. This structural property is robust to changes in MEC server configurations and threat arrival patterns. It ensures an optimal and easily implementable security policy under the exponential model. Theoretical analysis establishes upper bounds on the performance of AI-based security controllers using the same models. The results also show that standard model-free Q-learning algorithms can recover optimal thresholds without any prior knowledge of the system parameters. Simulations across multiple reinforcement learning architectures, including Q-learning, State–Action–Reward–State–Action (SARSA), and Deep Q-networks (DQN), confirm that all methods converge to the predicted threshold. This empirically validates the analytical findings. The threshold structure remains effective under practical imperfections such as imperfect sensing and parameter estimation errors. Systems maintain 85% to 93% of their optimal performance. This work extends threshold Markov Decision Process (MDP) analysis from classical queuing theory to the context of CIoT security offloading. It provides optimal and practical policies and model-free algorithms for use by resource-constrained devices. Full article
Show Figures

Figure 1

29 pages, 5613 KB  
Article
Sustainability Performance of FPSO Recycling
by Júlia Fernandes Sant’ Ana, Lino Guimarães Marujo and Carlos Eduardo Durange de Carvalho Infante
Sustainability 2026, 18(7), 3204; https://doi.org/10.3390/su18073204 - 25 Mar 2026
Viewed by 269
Abstract
The recycling of Floating Production Storage and Offloading (FPSO) units has become an important economic and environmental challenge as a growing number of offshore assets reach end-of-life. This study evaluates the comparative economic, environmental, and social performance of alternative FPSO recycling scenarios evaluated [...] Read more.
The recycling of Floating Production Storage and Offloading (FPSO) units has become an important economic and environmental challenge as a growing number of offshore assets reach end-of-life. This study evaluates the comparative economic, environmental, and social performance of alternative FPSO recycling scenarios evaluated using a stochastic Monte Carlo simulation, focusing on five FPSOs that operated in Brazil and were scheduled for recycling either domestically or in Denmark. Twelve performance indicators were aggregated into sustainability indices using a Monte Carlo simulation with 100,000 iterations, enabling analysis of robustness and variability across ten recycling scenarios. The results indicate that Brazilian recycling scenarios (P-32 and P-33) outperform the Danish scenarios in terms of global performance, with Global Sustainability Index values predominantly ranging from 0.59 to 0.75, compared to 0.37 to 0.61 for the Danish cases. Differences in performance are mainly associated with towing distance, cost structure, and emissions. Social indicators show limited variability and act as a stabilizing component across scenarios. Plasma cutting presents slightly better environmental and economic results than LPG cutting, although it does not alter the overall ranking of scenarios. These findings support decision-making on FPSO recycling scenarios by highlighting the role of uncertainty and contextual factors, particularly in emerging recycling markets. Full article
Show Figures

Figure 1

20 pages, 543 KB  
Article
EdgeGuard-AI: Zero-Trust and Load-Aware Federated Scheduling for Secure and Low-Latency IoT Edge Networks
by Abdulaziz G. Alanazi and Haifa A. Alanazi
Sensors 2026, 26(6), 1989; https://doi.org/10.3390/s26061989 - 23 Mar 2026
Viewed by 269
Abstract
Edge computing is now widely used to support real-time and safety-critical IoT services. However, current edge schedulers usually optimize only performance, while security verification and trust assessment are handled as separate modules. This separation creates a practical risk: tasks may be assigned to [...] Read more.
Edge computing is now widely used to support real-time and safety-critical IoT services. However, current edge schedulers usually optimize only performance, while security verification and trust assessment are handled as separate modules. This separation creates a practical risk: tasks may be assigned to lightly loaded but compromised edge nodes, or secure nodes may become overloaded, violating latency requirements. We propose EdgeGuard-AI, a unified trust-driven and load-aware scheduling framework inspired by zero-trust security principles for next-generation IoT edge networks. The framework jointly learns dynamic node trust and short-term workload patterns from distributed edge data and integrates both signals into scheduling decisions. Experimental results on a realistic IoT edge security dataset show a task success rate of 97.3 percent, average scheduling latency of 58.1 ms during stress periods, unsafe offloading below 2 percent, and trust discrimination AUC of 0.971. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

20 pages, 1680 KB  
Article
Efficient Inference of Neural Networks with Cooperative Integer-Only Arithmetic on a SoC FPGA for Onboard LEO Satellite Network Routing
by Bogeun Jo, Heoncheol Lee, Bongsoo Roh and Myonghun Han
Aerospace 2026, 13(3), 277; https://doi.org/10.3390/aerospace13030277 - 16 Mar 2026
Viewed by 236
Abstract
Low Earth orbit (LEO) satellite networks require real-time routing to cope with dynamic topology variations caused by continuous orbital motion. As an alternative to conventional routing approaches, deep reinforcement learning (DRL) has recently gained attention as an effective means for optimizing routing paths. [...] Read more.
Low Earth orbit (LEO) satellite networks require real-time routing to cope with dynamic topology variations caused by continuous orbital motion. As an alternative to conventional routing approaches, deep reinforcement learning (DRL) has recently gained attention as an effective means for optimizing routing paths. To solve routing problems modeled as a grid-based Markov decision process (grid-based MDP), DRL methods such as CNN-based Dueling DQN have been proposed. However, these approaches are difficult to implement in practice. In particular, the substantial floating-point computation and memory traffic of CNN inference make real-time onboard inference challenging under the stringent power and resource constraints of satellite platforms. To address these constraints, this paper proposes an INT8 quantization and hardware–software co-design framework using heterogeneous SoC FPGA acceleration. We offload compute-intensive CNN inference to the programmable logic (PL), while the processing system (PS) orchestrates overall control and data movement, forming a collaborative PS–PL architecture. Furthermore, we integrate the NITI-style two-pass scaling with PS–PL exponent propagation to preserve end-to-end integer consistency without floating-point conversion. To demonstrate its practical onboard feasibility, we employ standard accelerator implementation choices—such as output-stationary scheduling and on-chip prefetching—and conduct an ablation study over independently tunable axes (PE array size and PS-side buffer reuse) to quantify their incremental contributions. Experimental results show that the proposed PS–PL cooperative scheme dramatically reduces computation time compared to a PS-only reference implementation on the same platform. Full article
(This article belongs to the Section Astronautics & Space Science)
Show Figures

Figure 1

20 pages, 315 KB  
Systematic Review
Green Scheduling and Task Offloading in Edge Computing: A Systematic Review
by Adriana Rangel Ribeiro, Ana Clara Santos Andrade, Gabriel Leal dos Santos, Guilherme Dinarte Marcondes Lopes, Edvard Martins de Oliveira, Adler Diniz de Souza and Jeremias Barbosa Machado
Network 2026, 6(1), 17; https://doi.org/10.3390/network6010017 - 16 Mar 2026
Viewed by 278
Abstract
This paper presents a Systematic Literature Review (SLR) on green scheduling and task offloading strategies for energy optimization in edge computing environments. The evolution of low-latency, high-performance applications has driven the widespread adoption of distributed computing paradigms such as Edge Computing, Fog-Cloud architectures, [...] Read more.
This paper presents a Systematic Literature Review (SLR) on green scheduling and task offloading strategies for energy optimization in edge computing environments. The evolution of low-latency, high-performance applications has driven the widespread adoption of distributed computing paradigms such as Edge Computing, Fog-Cloud architectures, and the Internet of Things (IoT). In this context, Mobile Edge Computing (MEC) is often combined with Unmanned Aerial Vehicles (UAVs) to extend computational capabilities to areas with limited infrastructure, bringing processing closer to the data source to reduce latency and improve scalability. Nevertheless, these systems encounter substantial energy-related challenges, particularly in battery-powered or resource-constrained environments. To address these concerns, green computing strategies—especially energy-efficient scheduling and task offloading—have emerged as promising approaches to optimize energy usage in edge environments. Green scheduling optimizes task allocation to minimize energy consumption, whereas offloading redistributes workloads from resource-constrained devices to edge or cloud servers. Increasingly, these techniques are enhanced through artificial intelligence (AI) and machine learning (ML), enabling adaptive and context-aware decision-making in dynamic environments. This paper conducts a systematic literature review (SLR) to synthesize the most widely adopted strategies for energy-efficient scheduling and task offloading in edge computing, highlighting their impact on sustainability and performance. The analysis provides a comprehensive view of the state of the art, examines how architectural contexts influence energy-aware decisions, and highlights the role of AI/ML in enabling intelligent and sustainable edge systems. The findings reveal current research gaps and outline future directions to advance the development of robust, scalable, and environmentally responsible computing infrastructures. Full article
Show Figures

Figure 1

43 pages, 6922 KB  
Article
Multi-Flow Hybrid Task Offloading Scheme for Multimodal High-Load V2I Services
by Weiqi Luo, Yaqi Hu, Maoqiang Wu, Yijie Zhou, Rong Yu and Junbin Qin
Electronics 2026, 15(6), 1229; https://doi.org/10.3390/electronics15061229 - 16 Mar 2026
Viewed by 380
Abstract
In the Internet of Vehicles (IoV), connected vehicles generate high-load perception tasks with large-scale and multimodal sensitive data, imposing strict requirements on latency, computing, and privacy. Existing solutions still suffer from high task service latency and privacy risks. To address these issues, this [...] Read more.
In the Internet of Vehicles (IoV), connected vehicles generate high-load perception tasks with large-scale and multimodal sensitive data, imposing strict requirements on latency, computing, and privacy. Existing solutions still suffer from high task service latency and privacy risks. To address these issues, this paper proposes an integrated framework that jointly considers multi-flow task offloading, adaptive privacy preservation, and latency-aware resource incentive mechanism. Specifically, we propose a Location-Aware and Trust-based (LA-Trust) dual-node task offloading algorithm based on deep reinforcement learning (DRL), which treats pre-partitioned subtasks as multiple parallel flows and enables flow-level collaborative offloading optimization across neighboring nodes, allows subtask data uploading and processing to proceed concurrently, and incorporates node security into decision making. To further enhance privacy protection, a Distribution-Aware Local Differential Privacy (DA-LDP) algorithm is designed to adaptively inject artificial noise according to data heterogeneity, balancing privacy protection and task execution accuracy. In addition, a Delay-Cost Reverse Auction (DC-RA) algorithm is proposed to further reduce latency by introducing wireless channel modeling between idle vehicles and edge nodes into the incentive mechanism. Experimental results show that the proposed framework improves task execution accuracy by 38% and reduces offloading cost, delay, incentive cost, and auction communication latency by 64.41%, 64.64%, 19%, and 44%, respectively, while more than 60% of tasks are offloaded to high-trust nodes. Full article
Show Figures

Figure 1

21 pages, 1611 KB  
Article
Mobility-Aware Cooperative Optimization for Task Offloading and Resource Allocation in Multi-Edge Computing
by Dong Chen, Ximing Zhang, Kequan Lin, Chunhua Mei and Ru Huo
Algorithms 2026, 19(3), 221; https://doi.org/10.3390/a19030221 - 16 Mar 2026
Viewed by 252
Abstract
The rapid proliferation of mobile Internet of Things (IoT) devices has introduced significant resource scheduling challenges in multi-edge computing networks, where device mobility leads to dynamic network connectivity and load imbalance, complicating task offloading and resource management. To address these issues, this paper [...] Read more.
The rapid proliferation of mobile Internet of Things (IoT) devices has introduced significant resource scheduling challenges in multi-edge computing networks, where device mobility leads to dynamic network connectivity and load imbalance, complicating task offloading and resource management. To address these issues, this paper presents a mobility-driven hierarchical optimization framework for task offloading and computation resource allocation in multi-region edge computing environments, a functionally coupled hierarchical framework that integrates mobility-aware heuristic offloading with multi-agent deep deterministic policy gradient (MADDPG)-based resource allocation. Devices are first clustered according to their mobility patterns, and offloading decisions are dynamically made based on trajectory and dwell-time characteristics. Each edge server is modeled as an autonomous agent, and an MADDPG framework is adopted to collaboratively optimize resource allocation, with the joint objective of minimizing task processing delay and system energy consumption. Experimental evaluations under diverse mobility and workload conditions show that the proposed approach achieves a 19.0% reduction in task delay compared to the Multi-Objective Gray Wolf Optimization (MOGWO) method at the largest device scale (60 devices) and maintains comparable energy efficiency. Furthermore, it exhibits stronger adaptability and scheduling performance across varying mobility group distributions. These results confirm the effectiveness of the proposed method in enhancing system performance within dynamic mobile edge computing scenarios. Full article
Show Figures

Figure 1

30 pages, 1414 KB  
Article
Graph-Attention Constrained DRL for Joint Task Offloading and Resource Allocation in UAV-Assisted Internet of Vehicles
by Peiying Zhang, Xiangguo Zheng, Konstantin Igorevich Kostromitin, Wei Zhang, Huiling Shi and Lizhuang Tan
Drones 2026, 10(3), 201; https://doi.org/10.3390/drones10030201 - 13 Mar 2026
Viewed by 359
Abstract
Unmanned aerial vehicles (UAVs) acting as mobile aerial edge platforms can deliver on-demand communication and computing for the Internet of Vehicles (IoV) via flexible deployment and line-of-sight (LoS) links, improving reliability and reducing latency. However, high vehicle mobility, time-varying channels, and limited onboard [...] Read more.
Unmanned aerial vehicles (UAVs) acting as mobile aerial edge platforms can deliver on-demand communication and computing for the Internet of Vehicles (IoV) via flexible deployment and line-of-sight (LoS) links, improving reliability and reducing latency. However, high vehicle mobility, time-varying channels, and limited onboard energy make task offloading and resource coordination challenging. This paper studies joint task offloading and resource allocation in a UAV-assisted IoV system, where the UAV selects its hovering position from discrete candidate sites each time slot and splits vehicular tasks between the UAV and a roadside unit (RSU) to relieve backhaul congestion and enhance edge resource utilization. Considering vehicle mobility, multi-stage queue dynamics, and UAV energy consumption for communication, computation, and movement, the online optimization of position selection, task splitting, and bandwidth allocation is formulated as a constrained Markov decision process (CMDP). The goal is to maximize the number of tasks completed within the latency deadlines while satisfying the UAV energy budget. To solve this CMDP, we propose a graph-attention-based constrained twin delayed deep deterministic policy gradient (GAT-CTD3) algorithm. A graph attention network captures spatial correlations and resource competition among active vehicles, while a Lagrangian TD3 framework enforces long-term energy constraints and improves learning stability via twin critics, delayed policy updates, and target smoothing. The simulation results demonstrate that it outperforms the comparative scheme in terms of task completion rate, delay, and energy consumption per completed task, and exhibits strong robustness in situations with dense traffic. Full article
Show Figures

Figure 1

22 pages, 1506 KB  
Article
Task Offloading Based on Virtual Network Embedding in Software-Defined Edge Networks: A Deep Reinforcement Learning Approach
by Lixin Ma, Peiying Zhang and Ning Chen
Information 2026, 17(3), 278; https://doi.org/10.3390/info17030278 - 10 Mar 2026
Viewed by 310
Abstract
The advent of 5G/6G technologies and the pervasive deployment of IoT devices are driving the emergence of demanding applications that necessitate ultra-low latency, high bandwidth, and significant computational power. Traditional cloud computing models fall short in meeting these stringent requirements. To address this, [...] Read more.
The advent of 5G/6G technologies and the pervasive deployment of IoT devices are driving the emergence of demanding applications that necessitate ultra-low latency, high bandwidth, and significant computational power. Traditional cloud computing models fall short in meeting these stringent requirements. To address this, Software-Defined Edge Networks (SDENs) have emerged as a promising architecture, yet efficiently managing their heterogeneous and geographically distributed resources poses substantial challenges for optimal application provisioning. In response, this paper proposes a novel framework for intelligent task offloading, which reframes the intricate multi-component application task offloading problem as a Virtual Network Embedding (VNE) challenge within a SDEN environment. We introduce a comprehensive model where complex applications are represented as Virtual Network Requests (VNRs). In this model, each VNR consists of virtual nodes that demand specific computing and storage resources, as well as virtual links that demand specific bandwidth and must adhere to maximum tolerable delay constraints. To dynamically solve this NP-hard VNE problem in the face of stochastic VNR arrivals and dynamic network conditions, we leverage Deep Reinforcement Learning (DRL). Specifically, a Soft Actor-Critic (SAC) agent is employed at the SDN controller. This agent learns a sequential decision-making policy for mapping virtual nodes to physical edge servers and virtual links to network paths. To guide the agent towards efficient resource utilization, we define the reward for each successful embedding as the long-term revenue-to-cost ratio. By learning to maximize this reward, the agent is naturally driven to find economically viable allocation strategies. Comprehensive simulation experiments demonstrate that our SAC-based VNE approach significantly outperforms other baselines across key metrics, affirming its efficacy in dynamic SDEN environments. Full article
(This article belongs to the Section Information and Communications Technology)
Show Figures

Figure 1

26 pages, 1169 KB  
Article
HyAR-PPO: Hybrid Action Representation Learning for Incentive-Driven Task Offloading in Vehicular Edge Computing
by Wentao Wang, Mingmeng Li and Honghai Wu
Sensors 2026, 26(6), 1743; https://doi.org/10.3390/s26061743 - 10 Mar 2026
Viewed by 319
Abstract
Vehicular Edge Computing (VEC) can effectively guarantee the service experience of user vehicles, but resource-limited Roadside Units (RSUs) may face insufficient computing capacity during task peak periods. Utilizing Assisting Vehicles (AVs) with idle resources to share computing power can alleviate the pressure on [...] Read more.
Vehicular Edge Computing (VEC) can effectively guarantee the service experience of user vehicles, but resource-limited Roadside Units (RSUs) may face insufficient computing capacity during task peak periods. Utilizing Assisting Vehicles (AVs) with idle resources to share computing power can alleviate the pressure on RSUs. However, existing studies often fail to adequately incentivize selfish assisting vehicles to contribute resources and frequently lack a global optimization perspective from the overall system welfare. To address these challenges, this paper proposes an incentive-driven utility-balanced task offloading framework that aims to maximize social welfare while jointly optimizing resource allocation and profit pricing. Specifically, we first formulate the resource allocation as a Mixed-Integer Nonlinear Programming (MINLP) problem. To solve this problem, we introduce hybrid action representation learning to VEC for the first time and propose the HyAR-PPO algorithm to jointly optimize discrete offloading decisions and continuous resource allocation. This algorithm maps heterogeneous hybrid actions to a unified latent representation space through a Variational Autoencoder for the solution. Subsequently, equilibrium prices among user vehicles, Computation Service Providers (CSPs), and assisting vehicles are determined through Nash bargaining games, satisfying individual rationality constraints and achieving Pareto-optimal fair profit distribution. Experimental results demonstrate that the proposed framework can effectively coordinate multi-party interests. Compared with mainstream methods, the approach based on hybrid action representation learning achieves a significant improvement in social welfare, with its advantages being more pronounced in medium-to-large-scale scenarios. Full article
(This article belongs to the Special Issue Edge Computing for Resource Sharing and Sensing in IoT Systems)
Show Figures

Figure 1

27 pages, 2344 KB  
Article
Cloud-Edge Resource Scheduling and Offloading Optimization Based on Deep Reinforcement Learning
by Lili Yin, Yunze Xie, Ze Zhao and Jie Gao
Sensors 2026, 26(5), 1704; https://doi.org/10.3390/s26051704 - 8 Mar 2026
Viewed by 338
Abstract
In the context of smart manufacturing, with the widespread deployment of Industrial Internet of Things (IoT) devices, a large number of computation tasks that are highly sensitive to latency and have strict deadlines have emerged, requiring real-time processing. Effectively offloading tasks to address [...] Read more.
In the context of smart manufacturing, with the widespread deployment of Industrial Internet of Things (IoT) devices, a large number of computation tasks that are highly sensitive to latency and have strict deadlines have emerged, requiring real-time processing. Effectively offloading tasks to address the issues of increased latency and task dropouts caused by dynamic changes in edge node load has become a key challenge in the cloud–edge–end collaborative environment of smart manufacturing. To tackle the complex issues of unknown edge node loads and dynamic system state changes, this paper proposes a distributed algorithm based on deep reinforcement learning, utilizing convolutional neural networks (CNN) and the Informer architecture. The proposed algorithm leverages CNN to extract local features of edge node loads while utilizing Informer’s self-attention mechanism to capture long-term load variation trends, thereby effectively handling the uncertainty and dynamics inherent in node loads. Furthermore, by integrating the Dueling Deep Q-Network (DQN) and Double DQN techniques, the algorithm achieves a precise approximation of the state–action value function, further enhancing its capability to perceive system temporal characteristics and adapt to heterogeneous tasks. Each mobile device can independently make task offloading decisions and scheduling strategies based on its observations, enabling dynamic task allocation and optimization of execution order. Simulation results show that, compared to various existing algorithms, the proposed method reduces task dropout rates by 82.3–94% and average latency by 28–39.2%. Experimental results validate the significant advantages of this method in intelligent manufacturing scenarios with high load and latency-sensitive tasks. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

Back to TopTop