Next Article in Journal
A Novel Group Decision-Making Method with Adjustment Willingness in a Distributed Hesitant Fuzzy Linguistic Environment
Previous Article in Journal
Weak Solutions to Leray–Lions-Type Degenerate Quasilinear Elliptic Equations with Nonlocal Effects, Double Hardy Terms, and Variable Exponents
Previous Article in Special Issue
Adaptive Real-Time Transmission in Large-Scale Satellite Networks Through Software-Defined-Networking-Based Domain Clustering and Random Linear Network Coding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Satellite Task Parallelism via Priority-Aware Decomposition and Dynamic Resource Mapping

1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
Big Data Institute, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(7), 1183; https://doi.org/10.3390/math13071183
Submission received: 7 March 2025 / Revised: 1 April 2025 / Accepted: 2 April 2025 / Published: 3 April 2025
(This article belongs to the Special Issue New Advances in Network and Edge Computing)

Abstract

:
Multi-satellite collaborative computing has achieved task decomposition and collaborative execution through inter-satellite links (ISLs), which has significantly improved the efficiency of task execution and system responsiveness. However, existing methods focus on single-task execution and lack multi-task parallel processing capability. Most methods ignore task priorities and dependencies, leading to excessive waiting times and poor scheduling results. To address these problems, this paper proposes a task decomposition and resource mapping method based on task priorities and resource constraints. First, we introduce a graph theoretic model to represent the task dependency and priority relationships explicitly, combined with a novel algorithm for task decomposition. Meanwhile, we construct a resource allocation model based on game theory and combine it with deep reinforcement learning to achieve resource mapping in a dynamic environment. Finally, we adopt the theory of temporal logic to formalize the execution order and time constraints of tasks and solve the dynamic scheduling problem through mixed-integer nonlinear programming to ensure the optimality and real-time updating of the scheduling scheme. The experimental results demonstrate that the proposed method improves resource utilization by up to about 24% and reduces overall execution time by up to about 42.6% in large-scale scenarios.

1. Introduction

With the rapid development of satellite communication and networking technologies, satellite networks have become a crucial part of the global information infrastructure, supporting applications in communications, navigation, and remote sensing [1,2]. In particular, the deployment of large Low Earth Orbit (LEO) satellite constellations in recent years has further underscored the vital role of satellite networks in worldwide connectivity and Earth observation [3]. Traditional satellite task processing methods rely on ground stations for centralized mission planning and single-satellite execution [4]. This paradigm requires frequent satellite–ground communications and has limited real-time responsiveness [5]. Recent advances in integrated space–ground network architectures enable resource sharing and cooperative processing directly between satellites [6,7]. Consequently, multi-satellite cooperative computing has emerged as a new research focus, wherein multiple satellites interact via inter-satellite links to distribute and execute computational tasks collaboratively, essentially forming a “space-based cloud” computing environment [8,9,10]. By decomposing complex missions and distributing subtasks across satellites for parallel execution, multi-satellite cooperation markedly reduces the burden on any single satellite and diminishes reliance on ground control, thereby significantly improving real-time responsiveness [11]. Therefore, incorporating multi-satellite cooperative computing for task decomposition and resource allocation has become a key avenue to meet the demands of future large-scale and complex space missions [12].
With the rapid development of satellite communication and networking technologies, satellite networks have become a crucial part of the global information infrastructure, supporting applications in communications, navigation, and remote sensing [1]. However, efficient multi-satellite cooperative computing is challenging regarding task decomposition, resource allocation, and real-time scheduling [13,14]. First, determining how to partition a complex mission into subtasks and assign them to different satellites is a complicated decision problem that must account for heterogeneous satellite resources and inter-task dependencies [15]. Second, the multi-satellite network environment is highly dynamic: satellite positions and link conditions change over time, and task requests may arrive unpredictably. This dynamism demands that scheduling strategies adapt rapidly [16]; otherwise, a precomputed plan can quickly become suboptimal or invalid. In addition, the joint optimization problem of multi-satellite collaboration is typically non-convex, involving discrete task assignment and continuous resource allocation variables, which makes it exceptionally computationally complex. Indeed, the complexity of typical multi-satellite scheduling grows exponentially with the number of satellites and tasks, rendering it an NP-hard problem. This complexity hampers the ability to obtain exact optimal solutions within real-time constraints.
In response to the above challenges, researchers have proposed various heuristic and meta-heuristic algorithms (e.g., genetic algorithms [17], tabu search [18], multi-objective evolutionary algorithms [19,20]) to tackle non-convex task allocation challenges. These algorithms have achieved specific progress in computational efficiency. However, most such studies rely on simplified assumptions and static scenarios without adequately accounting for real-time changes in satellite states and task demands. As a result, their scheduling strategies may perform suboptimally in actual dynamic environments and often focus on optimizing only a single performance metric (e.g., minimizing task completion time) without considering other objectives such as energy consumption. Overall, current methods still exhibit shortcomings in aspects like real-time responsiveness, solving complex non-convex optimizations, and cross-system collaboration. On one hand, many algorithms have high computational complexity and struggle to meet immediate response requirements. On the other hand, most research has been confined to cooperation within satellite networks alone, lacking consideration of integrated satellite–ground computing, which limits their applicability to broader collaborative scenarios.
In response to the abovementioned challenges, we introduce a comprehensive approach that integrates three main layers to achieve efficient task decomposition and allocation. The task decomposition layer utilizes a graph-theoretical model to represent complex task dependency and priority relationships. Then, it is addressed by mathematical tools such as Integer Linear Programming to derive optimal or near-optimal task decomposition solutions. The resource mapping and allocation layer models the heterogeneous resource competition within the multi-satellite system as a game-theoretic process by combining game theory with Deep Reinforcement Learning (DRL), gradually converging toward Nash equilibrium. Finally, the dynamic scheduling and temporal constraints layer formalizes task execution order and time constraints using Temporal Logic theory. It solves the dynamic scheduling problem to ensure feasibility and optimality during execution. The proposed method effectively overcomes the challenges of real-time dynamic environments and complex non-convex optimizations.
The main contributions of this paper are as follows:
1.
The task dependencies and priorities are explicitly represented through a directed acyclic graph (DAG) model based on graph theory, and the NP-hard decomposition problem caused by task dependencies and resource constraints is solved by integer linear programming. This method can reasonably allocate subtasks among multiple heterogeneous satellites, effectively reduce the overall execution time, and ensure the optimization of task decomposition.
2.
We propose a model based on game theory to construct a resource competition game between satellites. We combine this with a Deep Q-Learning Network (DQN), and the Nash equilibrium of the game is gradually approximated through the dynamic learning of intelligence, thus achieving the efficient adaptive allocation of resources. This method can adapt to changes in system resources in real time and optimize resource utilization efficiency.
3.
We introduce an LTL-based model to formally describe the dependencies and timeframes of tasks, and the dynamic scheduling problem can be solved through a mixed integer nonlinear programming (MINLP) model, which incorporates real-time feedback for online updating of the scheduling scheme, which adapts to environmental changes and resource fluctuations in the execution process, ensuring the optimality of task scheduling.

2. Related Work

Driven by the demand for multi-satellite collaboration, satellite networks have become the core platform for achieving broad area observation and efficient computing. In recent years, research on complex task decomposition, allocation, and resource mapping has mainly focused on static task scheduling and single-satellite resource optimization. However, existing methods generally lack dynamic multi-tasking parallel processing capabilities, neglect the joint optimization of priority and dependency relationships, and have rigid resource mapping mechanisms, leading to prominent issues of system response delay and resource fragmentation, making it difficult to adapt to high real-time and heterogeneous satellite collaboration scenarios.

2.1. Dynamic Decomposition and Hierarchical Optimization

To address the challenges of high-dimensional task conflicts and dynamic scheduling in the complex task decomposition of satellite networks, researchers have proposed a dynamic decomposition and hierarchical optimization strategy that reduces the complexity of the problem through decoupling in the spatiotemporal dimension [21,22,23,24,25,26].
For instance, Wang et al. [21] dynamically decomposes the target area into candidate observation bands and optimizes the combination using the adaptive genetic algorithm. Moreover, Yao et al. [24] adopts a hierarchical structure of task allocation merging and utilizes dynamic programming to reduce task conflicts. Liu et al. [25] proposed a dynamic task decomposition and resource allocation optimization method based on the MILP algorithm, which improves the efficiency and real-time performance of complex task scheduling and resource mapping in satellite networks through spatiotemporal decoupling and multi-level optimization strategies. Also, Bittencourt et al. [26] utilized a lookahead-based variant of HEFT (HEFT-LA) to dynamically adjust task scheduling in satellite networks, improving task execution efficiency by anticipating future resource demands. However, such methods still face the problems of insufficient task conflict handling and scheduling flexibility, and require exploration of more precise resource allocation and scheduling strategies.

2.2. Intelligent Optimization and Collaborative Scheduling

In addition to traditional optimization methods, researchers have combined intelligent computational methods to solve the problem of balancing resource competition in multi-objective optimization in task decomposition [23,27,28,29,30,31,32]. For example, Chai et al. [23] improved resource utilization by balancing power and rate through a two-stage optimization framework. Similarly, Ikoma et al. [29] proposed a resource allocation method based on ACO, which improves task execution efficiency and real-time performance by optimizing computation and network resource allocation. Additionally, Chen et al. [32] introduced the PA-LBIMM algorithm, which integrates user-priority-based Min-Min scheduling to enhance load balancing in cloud computing environments, offering better performance in terms of task completion time under multi-objective optimization. However, non-convex optimization solutions lack stability, so further research is needed on weight-adaptive allocation methods for high-dimensional objective spaces.
For the coupling problem of task relevance in convex optimization, some researchers have proposed task aggregation and graph network methods based on the concept of intelligent computing [33,34,35,36]. Typically, Fan et al. [33] reduced task conflicts through task network graph aggregation. And Cai et al. [36] proposed a task decomposition and resource scheduling method based on the network flow algorithm, which optimizes task offloading and resource allocation in cloud–edge collaborative computing through LSTM prediction and hierarchical scheduling. Although these methods effectively optimize task scheduling, they still face problems such as poor real-time performance, high computational complexity, and insufficient flexibility in resource scheduling.
In the decentralized scheduling scenario, researchers adopt multi-agent collaboration frameworks [37,38]. For example, Zilberstein et al. [37] reduces task latency through deep reinforcement learning. However, the communication overhead and policy consistency between intelligent agents are still bottlenecks, so designing lightweight collaboration protocols and heterogeneous verification mechanisms is necessary. Faced with the challenge of limited resources, Xiaolu et al. designed a cross-layer resource allocation model [28,39]. However, existing methods have not been able to solve the problem of the dynamic matching of satellite ground resources, so combining them with digital twin technology is necessary to achieve virtual real linkage optimization. Although these methods have made some progress in task decomposition, scheduling, and resource optimization, they still face challenges in real-time performance, stability, and intelligence.

2.3. Dynamic Scheduling and Resource Allocation Optimization Methods

Dynamic scheduling mainly focuses on solving problems encountered in complex task decomposition in satellite networks, such as task correlation coupling, and researchers have proposed methods based on dynamic scheduling and resource allocation optimization to address the dynamic resource coupling and real-time challenges faced by complex task allocation in satellite networks. Researchers have adopted DRL and hierarchical multi-time-scale methods to solve the delay and resource allocation efficiency problems in dynamic task scheduling through constrained Markov decision process (CMDP) modeling and mixed near-end policy optimization [40,41,42,43]. Based on multi-agent collaboration and genetic programming, they have constructed autonomous satellite collaboration models and centralized distributed architectures, which have improved the task planning capability of multi-satellite systems [24,44,45,46].
In addition to innovative research work from an algorithmic perspective, some researchers have focused on network architecture innovation and optimized load balancing and economic benefits by designing multi-layer edge computing frameworks and online scheduling mechanisms [47,48,49,50]. For example, Li et al. [42] proposed a dual decoder strategy network based on the attention mechanism, which utilizes the virtual task adjustment mechanism to achieve flexible scheduling sequence generation; Lyu et al. [47] proposed a hybrid cloud satellite multi-layer architecture that dynamically coordinates node selection and computing resource allocation. However, there is an inherent contradiction between real-time assurance and multi-objective trade-offs in existing methods for cross-system collaborative optimization, and further exploration of low-complexity joint optimization theory is needed to balance the global efficiency and local resource constraints of heterogeneous satellite networks.

2.4. Intelligent Beam Matching and Cross-Layer Resource Optimization

While focusing on the decomposition and allocation of complex tasks in satellite networks, resource mapping has also become a core challenge in achieving efficient service provision. In order to address the challenges of non-uniform traffic demand and dynamic topology adaptation in satellite networks, researchers have solved the problems of improving spectral efficiency and dynamic resource allocation through flexible user beam mapping, multi-objective resource optimization, and cross-layer modeling, improving the responsiveness of satellite systems. Ramírez et al. [51] proposed an intelligent beam user matching scheme that breaks through traditional cell boundaries and combines dynamic beam coverage with adjustable bandwidth allocation, achieving efficient resource adaptation in non-uniform traffic scenarios; Du et al. [52] introduces an iterative optimization framework based on TDM, which combines time score and power allocation to reduce resource fragmentation while meeting differentiated requirements. In addition, Chen et al. [27] proposed a multi-objective modeling method that combines region decomposition and resource allocation, using the NSGA-II algorithm to balance coverage and resource consumption, providing a cross-domain reference for resource mapping. However, existing research lacks theoretical support for scenarios such as real-time reconstruction and multi-track layer resource collaboration in dynamic environments, and breakthroughs need to be sought in stochastic optimization models and distributed dynamic decision-making mechanisms.
In conclusion, existing methods have significant bottlenecks in real-time performance assurance, non-convex optimization solution efficiency, cross-domain collaboration capability (multi-orbital layer resource linking), and lack systematic solutions to policy consistency and resource fragmentation issues in heterogeneous network environments. In the future, breakthroughs are needed in low-complexity joint optimization theory, lightweight distributed decision-making mechanisms, and virtual accurate linkage digital twin modeling technology to address the challenges of high dynamics and multi-constraint satellite network collaboration. To address these issues, this paper proposes hierarchical adaptive decomposition and reinforcement learning-based resource mapping (HADRM), which models the complex dependencies between tasks and employs deep reinforcement learning-based resource mapping to cope with real-time changes in task dependencies and on-board resources. Ultimately, timing scheduling ensures the optimization of the execution order of the final tasks without affecting global optimality.

3. Problem Description and Formulation

3.1. System Model

The multi-satellite cooperative computing system consists of multiple heterogeneous satellites, each tasked with various remote sensing operations. The primary objective is to achieve efficient task execution through task decomposition and resource mapping techniques. Each satellite S = { s 1 , s 2 , , s n } has unique computational, storage, and communication capabilities, enabling resource sharing to enhance overall system performance. The system architecture comprises satellite nodes, communication links, and a task scheduling center, where components interact via inter-satellite communication for data transmission and processing.
The core components of the system include satellite nodes, the task scheduling center, and the communication network. Each satellite node s i possesses resources, characterized as computational resources r b ( s i ) , communication resources c ( s i ) , and payload resources r p ( s i ) . A task m i is decomposed into k subtasks { T i 1 , T i 2 , , T i k } , with each subtask T i j requiring specific resources, expressed as
R ( T i j ) = r p ( T i j ) , r b ( T i j ) , c ( T i j ) ,
The task allocation matrix X represents the assignment of subtasks to satellites:
X i j = 1 , if subtask T i j is assigned to satellite s i , 0 , otherwise
Data flows begin with the observation phase, generating observation data D o , which are transmitted via communication links to other nodes or the task scheduling center, followed by compression and processing phases. The amount of observation data is represented as
D o = ω s p · r s p ( t ) .
The compressed data can be represented as
D c = η · D o ,
where η is the compression efficiency. The data flow model can be represented through task dependency matrices Z and time constraints.

3.2. Problem Definition

The goal of task decomposition is to find an optimal decomposition strategy that minimizes the total execution time of the tasks while satisfying the task dependencies and time window constraints. The total execution time of a task m i , denoted as φ m ( t ) , is the sum of the completion times of all its subtasks. For each subtask T i j , its completion time φ i j ( t ) is given by the following formula:
φ i j ( t ) = max ( φ i k ( t ) + Δ t i k , φ i j sub ) ,
where φ i k ( t ) represents the completion time of a dependent subtask T i k , and  Δ t i k is the time delay between the completion of subtask T i k and the start of subtask T i j due to their dependency. The term φ i j sub represents the subtask completion time, which is determined by the resources required for the subtask. Specifically, it is the time needed to execute subtask T i j independently, considering the computational, communication, and payload resources required. This can be formally expressed as
φ m ( t ) = j = 1 k φ i j ( t ) ,
where φ i j ( t ) represents the completion time of subtask T i j . The completion time of each subtask depends on Remark 1.
Remark 1
(Task Dependencies and completion time). When a subtask T i j depends on the result of another subtask T i k , its completion time must be after the completion of the dependent task. The completion time of a dependent subtask is determined by both the completion time of the preceding subtask and the delay introduced by the dependency. Specifically, subtask T i j can only begin after T i k has completed, with an additional delay Δ t i k that reflects the time required for data transfer, communication, or synchronization between the tasks. This relationship can be mathematically expressed as
t i j t i k + φ i k ( t ) + Δ t i k , if subtask T i j depends on T i k ,
And the completion time of subtask T i j must be after the completion time of subtask T i k , where t i j is the start time of subtask T i j , t i k is the completion time of subtask T i k , and  φ i k ( t ) is the completion time of subtask T i k . The additional delay Δ t i k accounts for the necessary synchronization and data handling before T i j can begin.
In scenarios where there is no dependency, the completion time of subtask T i j is independent of other tasks. The completion time φ i j sub of a subtask T i j is determined solely by the resources required for its execution, including computational, communication, and payload resources. This is expressed as
φ i j sub = r p ( T i j ) R s p + r b ( T i j ) R s b + c ( T i j ) C s ,
where r p ( T i j ) , r b ( T i j ) , and  c ( T i j ) are the resource requirements for payload, computation, and communication, respectively, and  R s p , R s b , and  C s are the available capacities of the satellite.
In task scheduling, we take the maximum of the two values—the earliest possible completion time determined by task dependencies and the independent completion time of the subtask itself. Therefore, the completion time of subtask T i j is given by Equation (5).
This ensures that the task execution respects both the dependencies between tasks and the intrinsic resource requirements of each subtask, optimizing the overall scheduling of tasks.
The objective is to find the optimal task decomposition strategy that minimizes the total task execution time. The total cost function can be expressed as the sum of the execution times of the observation and data processing tasks:
μ m ( t ) = φ m ( t ) .
The total execution time is further decomposed into observation and compression tasks in Equations (10)–(12). Then, the observation task execution time is calculated as follows:
φ m p ( t ) = s S ω s p · r s p ( t ) ,
where ω s p is the priority of the observation task, and  r s p ( t ) is the execution time for observation tasks.
The compression task execution time is calculated as follows:
φ m b ( t ) = s S r s b ( t ) ν ,
where r s b ( t ) is the execution time for compression tasks, and  ν is the efficiency parameter for computational resources. Our goal is to minimize the total execution time of tasks in the multi-satellite cooperative computing system. Inputs include resource characteristics R ( T i j ) and task requirements M , with the optimal task allocation strategy X being the output. The objective function is
min m = 1 M φ m p ( t ) + φ m b ( t ) ,
where φ m p ( t ) denotes the observation time, and  φ m b ( t ) denotes the time for data processing and compression. This combinatorial optimization problem involves complex task decomposition and resource mapping.

3.3. Constraints

Based on the proposed problem scenario, we consider following constraints:

3.3.1. Task Dependency Constraint

If task T i j depends on task T i k , then
t i j t i k , if T i j depends on T i k .

3.3.2. Priority Constraint

Higher-priority tasks must be executed first:
ω i j ω i k , if ω i j > ω i k .

3.3.3. Time Window Constraint

Each subtask must be executed within its predefined time window:
t i j S t a r t t i j t i j E n d .

3.3.4. Resource Constraints

Resources allocated to each subtask must not exceed the total resources of the satellite:
r p ( T i j ) R s p , r b ( T i j ) R s b , c ( T i j ) C s .

3.3.5. Load Balancing Constraint

Resource utilization across all satellites must be balanced:
r b ( s i ) R s b r b ( s j ) R s b , s i , s j S .

3.3.6. Fault Tolerance Constraint

Considering that individual satellites or their resources (e.g., computation, storage, or payload) may fail in a multi-satellite system, once a failure or partial malfunction is detected in a satellite’s resources, the system will trigger a dynamic task reallocation process. Based on the existing resource constraints (16) and load balancing constraints (17), the task allocation matrix X is adjusted to promptly reassign the affected tasks, thereby ensuring the continuity and stability of the overall task scheduling.

3.3.7. Dependency and Execution Time Constraints

In scenarios where subtasks have dependencies, the completion time of a subtask T i j is determined not only by its own resource consumption, but also by the completion of preceding tasks. This constraint ensures that the completion time of T i j is bounded by the completion of task T i k and the delay Δ t i k that accounts for synchronization and data transfer between tasks. The following relation must hold for Equation (7):
Remark 2
(Fault Tolerance). The above fault tolerance mechanism is predicated on the system’s capability for real-time fault detection and response, ensuring that when a single satellite or some of its resources fail, the remaining satellites can utilize their available resources to complete the tasks, thereby maintaining the overall system performance without significant degradation.

3.4. Objective Function

The objective is to minimize the total execution time of tasks. The objective function is
min m = 1 M t = 1 T φ m p ( t ) + φ m b ( t ) .
For multi-objective optimization, a weighted method is used and we obtain the final objective function as Equation (19):
min F = α m = 1 M φ m p ( t ) + β m = 1 M φ m b ( t ) , s . t . C 1 : t i j t i k , if T i j depends on T i k , C 2 : ω i j ω i k , if ω i j > ω i k , C 3 : t i j S t a r t t i j t i j E n d , C 4 : r p ( T i j ) R s p , r b ( T i j ) R s b , c ( T i j ) C s , C 5 : r b ( s i ) R s b r b ( s j ) R s b , s i , s j S .
The objective function aims to minimize the total execution time of observation and compression tasks while ensuring efficient resource utilization and system stability. The function considers both the observation processing time φ i j p and compression processing time φ i j b , with weighting factors α and β to balance their contributions, where α and β are weight coefficients for different task phases. The execution time of each task is constrained by resource availability, time windows, and task dependencies. Let t i j be the execution time of an observation task, with feasible execution times defined in the set T i j . The deviation from the target execution time is denoted as d i , tar ( t i j ) , while resource allocations for compression tasks are represented by r p ( T i j ) and are constrained by the maximum allowed resource capacity R s p . Power consumption during execution is modeled by r b ( T i j ) and c ( T i j ) . Additional constraints include distance or deviation limits t i j S t a r t t i j t i j E n d , as well as capacity constraints for computation and transmission, given by r p ( T i j ) R s p . Task scheduling follows dependency constraints, ensuring that a compression task t i j does not start before the observation task t i k , with dependencies represented by ω i j ω i k . Furthermore, load balancing is maintained by ensuring that resource utilization across all satellites satisfies r b ( s i ) R s b r b ( s j ) R s b , s i , s j S .
The analysis shows that this problem is NP-hard. As the number of tasks M and satellite nodes N increases, the combinatorial complexity of task allocation and resource mapping grows exponentially. Polynomial-time algorithms cannot effectively solve this problem.
Through reduction from the classical task scheduling problem, the complexity of the problem is proven, classifying it as NP-hard. The task dependencies, resource allocation, and time constraints make the problem highly challenging, requiring heuristic or approximation algorithms to solve it.

4. Methodology

Figure 1 illustrates the overall framework of the HADRM method, which is composed of three main modules working iteratively to enhance system efficiency. (i) Task decomposition considers task dependencies and priorities to ensure the logical breakdown of complex tasks. (ii) Resource mapping and allocation integrates both static and adaptive strategies to optimize resource utilization across the system. (iii) Dynamic scheduling continuously adjusts task execution based on real-time system state updates, ensuring adaptability and efficiency. Together, these modules form a cohesive approach to optimizing task execution and resource management.

4.1. Task Decomposition Strategy Based on Graph Theory and Lagrangian Relaxation

To address the challenges of efficient task decomposition considering complex task dependencies, we model the remote sensing tasks as a DAG and apply integer linear programming (ILP) combined with Lagrangian relaxation to minimize the total execution time while satisfying task dependencies and resource constraints.
Task dependencies play a critical role in determining the execution order and time of tasks in a multi-satellite system. In our approach, tasks are modeled as a DAG, where each node represents a subtask T i j and directed edges represent the dependency relationships between subtasks. A key aspect of this model is the handling of dependencies that may impose constraints on the execution order of subtasks.
For instance, if subtask T i j depends on the results of subtask T i k , then T i j cannot start until T i k is completed, along with an additional delay Δ t i k to account for data transfer or synchronization. This relationship can be formally expressed as Equation (7).
However, in cases where there is no strict dependency, subtasks can be executed concurrently, which can reduce the overall execution time. Therefore, adjusting the execution sequence of tasks based on their dependencies is crucial for optimizing the total execution time. By carefully selecting which task should be executed first and considering the delays introduced by dependencies, the overall system efficiency can be maximized.
For example, if task T i j is dependent on T i k , but there is no significant data transfer between the two, the system might allow T i j to start earlier, assuming other resources are available. This adjustment is particularly important in a distributed multi-satellite environment, where communication and processing delays can be significant.
This approach ensures that the task execution respects both the dependency relationships and the intrinsic resource requirements of each subtask, optimizing the overall scheduling of tasks.
The task decomposition problem in multi-satellite cooperative computing is modeled as a DAG, where each node represents a subtask T i j and edges represent dependency relationships between subtasks. Each subtask T i j has an completion time t i j and resource requirements R ( T i j ) . The task dependencies and priorities can be expressed as
If T i j depends on T i k , t i j t i k + δ i k ,
where δ i k is the minimum time interval between tasks.
Based on the DAG model, an ILP model is established to minimize the total execution time. The decision variables are defined as follows:
x i j s is a binary variable, where x i j s = 1 if subtask T i j is assigned to satellite s, and 0 otherwise. t i j represents the start time of subtask T i j .
We obtain the following objective function:
min max i , j t i j + d i j ,
where d i j is the duration of subtask T i j .
The optimization problem includes the following constraints (22)–(26):

4.1.1. Task Allocation Constraint

Each subtask must be assigned to exactly one satellite:
s S x i j s = 1 , T i j

4.1.2. Resource Constraint

At any given time, the total resource demand of subtasks assigned to satellite s must not exceed its available resources:
T i j M R ( T i j ) x i j s R s , s S

4.1.3. Task Dependency Constraint

If subtask T i j depends on T i k ,
t i j t i k + d i k , if T i j depends on T i k

4.1.4. Time Window Constraint

Each subtask must be executed within its predefined time window:
t i j start t i j t i j end

4.1.5. Task Priority Constraint

Higher-priority tasks should be executed first:
If ω i j > ω i k , t i j t i k
Due to the NP-hard nature of the ILP problem, a Lagrangian relaxation approach is employed to improve the solution efficiency. The resource and task dependency constraints are relaxed into the objective function using Lagrange multipliers to construct the Lagrangian function:
L ( x , t , λ , μ ) = max i , j t i j + d i j + s S λ s T i j M R ( T i j ) x i j s R s + ( i j , i k ) E μ i j , i k t i j t i k d i k
where λ s represents the Lagrange multipliers for the resource constraints, μ i j , i k represents the Lagrange multipliers for the task dependency constraints, and E represents the set of task dependency relationships.
This Algorithm 1 decomposes the complex integer linear programming problem into smaller subproblems and employs Lagrange multipliers to coordinate global constraints with local optimization, ensuring scientific validity. It begins by initializing the iteration counter and Lagrange multipliers (Line 1). During each iteration, the algorithm solves a relaxed subproblem for each subtask T i j M (Line 3), minimizing the objective function of task start time, delay penalties, and resource constraint violations weighted by Lagrange multipliers. The solution must satisfy binary constraints, time windows, and other feasibility conditions. By accounting for task dependencies, resource constraints, time windows, and priorities, this approach guarantees feasible and near-optimal task allocation. Next, the Lagrange multipliers are updated to enforce resource and temporal constraints (Line 4). The resource-based multiplier λ s is adjusted based on the difference between the total allocated and available resources for each satellite, while the temporal constraint multiplier μ i j , i k is updated based on task precedence relationships, ensuring the overall allocation remains balanced and efficient. A stopping criterion is then evaluated (Line 5): if met, the algorithm terminates; otherwise, it increments the iteration counter and continues. Once convergence is achieved, a feasible solution is constructed (Line 6). The final task assignments and start times (or their averaged values) form the basis of the initial solution, which a repair heuristic (e.g., simulated annealing) refines by mitigating constraint violations and improving efficiency. Finally, the algorithm returns the optimized task assignments and start times x i j s , t i j (Line 7).
Algorithm 1: Task Decomposition Using Lagrangian Relaxation (LRCTD)
Mathematics 13 01183 i001

4.2. Resource Mapping and Allocation Strategy Based on Game Theory and Reinforcement Learning

To tackle the heterogeneity and dynamic nature of satellite resources in multi-satellite cooperative computing, a resource mapping and allocation strategy based on game theory and reinforcement learning is proposed. We model the resource allocation problem as a resource competition game among satellite nodes. The optimal allocation is then derived by solving the Nash equilibrium of the game. Furthermore, a deep reinforcement learning algorithm is integrated to dynamically adjust resource allocation in response to the changing environment, ensuring adaptive and efficient resource mapping.
In a multi-satellite cooperative system, each satellite node competes for limited computational, communication, and storage resources to maximize its utility. This competition can be naturally modeled as a non-cooperative game where each satellite s i selects a resource allocation strategy r i = [ r i b , r i c , r i p ] . The utility function for each satellite is defined as
U i ( r i , r i ) = f i ( r i ) g ( r i , r i ) ,
where f i ( r i ) represents the benefit derived from the allocated resources, and  g ( r i , r i ) represents the cost influenced by other satellites’ resource strategies r i .
By modeling the allocation as a game, the Nash equilibrium condition ensures that no satellite can improve its utility by unilaterally changing its resource strategy, providing a theoretically sound framework for optimal resource distribution.
To find the optimal resource allocation, we define the Nash equilibrium of this game. A strategy profile ( r 1 , r 2 , , r n ) is a Nash equilibrium if
U i ( r i , r i ) U i ( r i , r i ) , i , r i
We use the best response dynamics to iteratively adjust each satellite’s resource strategy toward equilibrium. However, due to the dynamic nature of the environment, solving for Nash equilibrium in real time is computationally expensive.
Considering the dynamic and complex nature of resource competition, we introduce DRL to adaptively adjust resource allocation. Each satellite node is modeled as an agent in a Markov decision process (MDP) where the state s t captures the current resource usage and task demand, the action a t represents the selected resource allocation strategy, and the reward r t is derived from the utility function U i . The objective of the agent is to learn an optimal policy π ( s t ) that maximizes its expected cumulative reward.
The learning process employs DQN to approximate the optimal action-value function:
Q ( s t , a t ) = max π E k = 0 γ k r t + k s t , a t .
The proposed resource allocation algorithm combines game theory and DRL to achieve an optimal and adaptive resource distribution. Algorithm 2 outlines the iterative learning and allocation process.
This algorithm begins by initializing the Q-network and the replay buffer (Lines 2–3). During each iteration, every satellite first observes its current state (Line 6) and selects a resource allocation action according to the ϵ -greedy policy derived from the Q-network (Line 7). The chosen action is then executed, yielding a reward and a subsequent state (Line 8), and the transition tuple is stored in the replay buffer (Line 9). A minibatch is subsequently sampled from the replay buffer (Line 11), and for each transition in the minibatch, a target Q-value is computed using the Bellman update (Line 13). The Q-network parameters are then optimized by minimizing the loss function (Line 14). Finally, for each satellite, the optimal resource allocation is obtained by selecting the action that maximizes the Q-network’s output (Line 16). This iterative process enables the resource allocation strategy to continuously improve, converging towards a near-optimal solution in a dynamic environment.
Algorithm 2: Resource Allocation Using Game Theory and DRL (GDRL)
Mathematics 13 01183 i002

4.3. Dynamic Scheduling and Task Dependency Constraint Handling Based on Temporal Logic

In the multi-satellite cooperative computing system, the dynamic nature of task execution and the complexity of task dependencies necessitate a robust scheduling mechanism. In the dynamic scheduling process, it is essential to consider not only the task dependencies but also how these dependencies evolve in real time as the system state changes. By using Linear Temporal Logic (LTL), we can express the timing and sequence constraints that govern the relationships between tasks. This formal framework ensures that the execution order of tasks adheres to the precedence relationships dictated by their dependencies. Furthermore, the dynamic nature of task execution requires continuous adjustments based on the current system state and task progress. As the system evolves, the scheduling process needs to re-evaluate task dependencies and resource availability to maintain optimal execution. This is achieved by using MINLP, which provides a flexible and adaptive framework for real-time scheduling. This section proposes a dynamic scheduling strategy using temporal logic and MINLP to handle the constraints imposed by task dependencies. Temporal logic provides a formal framework to model the sequential relationships between tasks.
To accurately represent the task dependencies in a multi-satellite environment, we use LTL to express the sequence and timing constraints among tasks. Each task T i j is associated with an LTL formula that defines its start and end conditions based on other dependent tasks:
G ( F T i k F T i j ) ,
where G denotes the “globally” operator, ensuring that the task relationship holds at all times, and F represents the “eventually” operator, indicating that T i j can only start after the completion of T i k . This formalism allows the encoding of complex task dependencies and timing constraints into the scheduling model.
We define the decision variables x i j ( t ) as binary variables indicating whether task T i j is executed at time t:
x i j ( t ) = 1 , if T i j is executed at time t , 0 , otherwise
The objective function aims to minimize the total execution time while satisfying the LTL constraints. The formulation includes constraints for task dependencies, resource limitations, and time windows. For example, the dependency constraint between tasks T i j and T i k can be formulated as
t = 1 T t · x i j ( t ) t = 1 T t · x i k ( t ) + τ i k ,
where τ i k represents the minimum time delay required between the completion of task T i k and the start of task T i j .
Theorem 1.
Given a set of task dependencies modeled by LTL formulas and the available resources of the multi-satellite system, the dynamic scheduling problem has a feasible solution if the MINLP formulation satisfies all temporal logic constraints and resource limitations.
Proof. 
We first encode each LTL dependency of the form
φ i k i j = G ( F T i k F T i j ) .
By discretizing time into slots t = 1 , , T . We define the binary decision variables as follows:
x i j ( t ) { 0 , 1 } , T i j , t = 1 , , T ,
Let the start time of task T i j be given by
t i j = t = 1 T t x i j ( t ) .
For every dependency pair ( T i k , T i j ) with a required minimum delay τ i k , introduce an auxiliary binary variable z i j k { 0 , 1 } and enforce the constraints
t i j t i k + τ i k M ( 1 z i j k ) ,
t = 1 T x i k ( t ) T z i j k , t = 1 T x i k ( t ) z i j k ,
where M is a sufficiently large constant ensuring that when z i j k = 0 , the constraint (35) is inactive.
Resource limitations for each satellite s are imposed by the linear constraint
T i j M R ( T i j ) x i j s R s , s ,
where R ( T i j ) denotes the resource requirement of task T i j , and R s is the resource capacity of satellite s.
Thus, the overall MINLP formulation is
min x i j ( t ) , t i j max T i j M t i j + d i j , s . t . t i j = t = 1 T t x i j ( t ) , T i j M , t = 1 T x i j ( t ) = 1 , x i j ( t ) { 0 , 1 } , T i j M , t i j t i k + τ i k M ( 1 z i j k ) , ( T i k , T i j ) , t = 1 T x i k ( t ) T z i j k , t = 1 T x i k ( t ) z i j k , ( T i k , T i j ) , T i j M R ( T i j ) x i j s R s , s ,
Since all LTL formulas are systematically transformed into the linear inequalities Equations (35) and (36) via auxiliary variables and big-M constraints, and the resource constraints (37) are inherently linear, the MINLP in Equation (38) comprises a finite set of linear inequalities. Therefore, if there exists an assignment of the binary variables { x i j ( t ) , z i j k } satisfying Equation (38), a feasible schedule exists. Consequently, under the condition that all temporal logic and resource constraints are met, the dynamic scheduling problem admits a feasible solution.    □
Based on the concept above, we proposed a scheduling algorithm that employs MINLP to dynamically adjust task scheduling in real time based on the system state and task dependencies. The main idea of the proposed TL-MIS is shown in Algorithm 3.
Algorithm 3: Dynamic Scheduling Using Temporal Logic and MINLP (TL-MIS)
Mathematics 13 01183 i003
This algorithm starts by initializing the task states and available resources (Lines 2–3). In each iteration, it constructs a scheduling problem based on the current set of tasks and constraints (Line 5). Then, for every task T i in the set M, if the task dependencies ϕ i j (as specified by temporal logic) are satisfied, the algorithm assigns T i to a satellite from S, determines its start time t i and duration Δ t i (Lines 6–9), and updates the available resources by subtracting the resources required for T i (Line 10). After processing all tasks in the current iteration, the task states are updated (Line 11), and completed tasks are removed from M (Line 12). Finally, the algorithm returns the optimal task scheduling x i j ( t ) , which specifies the execution schedule that meets all the temporal and resource constraints.
The proposed dynamic scheduling approach leverages temporal logic to encode task dependencies directly into the scheduling model, allowing for a flexible and robust adaptation to the dynamic nature of multi-satellite cooperative computing. The use of MINLP ensures that the solution satisfies all the constraints, providing a feasible and efficient scheduling strategy.

4.4. Overall Algorithm and Complexity Analysis

In Section 4, we propose a priority-aware decomposition and dynamic resource mapping approach for optimizing multi-satellite cooperative computing. The method integrates three primary components: task decomposition based on graph theory and ILP, resource allocation modeled as a game-theoretic problem solved by DRL, and dynamic scheduling formulated with LTL and MINLP.
The task decomposition module utilizes a DAG representation. It transforms the problem into an ILP formulation, where a Lagrangian relaxation-based heuristic efficiently explores feasible task allocation solutions while satisfying dependencies and resource constraints. The resource mapping and allocation module employs a game-theoretic model, where satellites compete for resources, and a Nash equilibrium-based strategy is determined through a DRL approach. This enables adaptive real-time optimization in response to system dynamics. The dynamic scheduling module leverages LTL to model task dependencies explicitly and incorporates an MINLP solver to iteratively adjust task execution in real time, ensuring the temporal constraints are met.
At each scheduling step, the proposed method executes three sub-algorithms sequentially. The complexity analysis shows that the dominant computational cost arises from the ILP-based task decomposition, which has a complexity of O ( N 3 ) , given that the graph-based task dependency representation typically requires cubic complexity for solving the associated ILP problems. Meanwhile, the reinforcement learning-based resource mapping algorithm contributes an additional complexity of O ( K · N 2 ) per iteration, where K denotes the number of training iterations, and N is the number of satellites. The dynamic scheduling algorithm based on temporal logic and MINLP further contributes a complexity of approximately O ( N 2 log N ) due to the nonlinear constraints and temporal logic encoding.
Therefore, combining these analyses, the overall time complexity per iteration of the proposed method is O ( N 3 ) , dominated by the ILP component. Thus, the overall computational complexity of the algorithm across all iterations is O ( T max · N 3 ) , where T m a x denotes the maximum number of iterations until convergence.

5. Experiment Results

In order to evaluate the performance of the proposed method, we construct an experimental scenario oriented to the decomposition of remote sensing tasks based on the self-developed Common Satellite Toolkit (CSTK) simulation platform. The CSTK platform calculates the real-time positions and orbits of the satellites within a 3D scenario for the satellite’s TLE data and assigns a unique computation, storage, and observation payload resource to each satellite for mission execution.

5.1. Simulation Scenario and Dataset Details

This scenario simulates the task of remote sensing data decomposition and allocation in a dynamic satellite network, with a duration from 04:00:00.00 UTC on 3 September 2024 to 04:00:00.00 UTC on 5 September 2024.
When constructing remote sensing tasks, we thoroughly considered the satellites’ operational status and the target area’s geographical characteristics. Each remote sensing task is represented as a structured data unit designed to facilitate efficient scheduling and resource allocation in a dynamic satellite network. The dataset captures both the satellites’ operational status and the target areas’ geographical characteristics, enabling accurate imaging and timely task execution. The main fields include the task ID, satellite ID assigned to the task, task time window, target area, expected data volume, task priority, and task dependencies. Regarding the target area, we describe it in the form of latitude and longitude ranges, clearly specifying the minimum and maximum latitude and longitude of the imaging area to ensure accurate coverage of the predetermined area during task execution. The “task dependency relationship” field clearly describes the sequential dependencies between the current task and other tasks.
Specifically, in each task, the observable angle and altitude requirements between the satellite and the target area are determined through precise calculations based on the satellite’s orbit information and ground station position. We use a formula to calculate the observation angle that meets the minimum elevation angle requirement, in order to derive the most suitable imaging time window. Subsequently, we set different priorities for each task based on its importance and resource consumption. The resource requirements of the task are mainly reflected in the working status of the imaging payload and the data transmission requirements. During the task execution process, the satellite needs to complete multi-stage operations such as data acquisition, compression, and transmission in real time, and each stage has strict time and resource limitations.
To further illustrate the structure of the task dataset used in this study, Table 1 provides a more detailed breakdown of actual task instances, including task IDs, target observation regions, discrete time ranges, and resource requirements. The interface for adding remote sensing tasks is shown in Figure 2.
We combined precise satellite data and using the CSTK simulation platform to construct a real multi-satellite network environment. The actual dataset was constructed using precise satellite data from multiple authoritative platforms, including Celestrak (CELESTRAK same as below. satellite database: https://celestrak.org/) and SatNOGS (SatNOGS open-source ground station network: https://satnogs.org/). Celestrak provides precise orbit parameters (comprising the semi-major axis, eccentricity, inclination, ascending node longitude, and perigee angle) that meet the standards of the North American Aerospace Defense Command, ensuring reliable orbit information. SatNOGS provides detailed ground station observation data, such as coordinates and communication link status.
This dataset covers various key parameters in satellite networks, including the computing, storage, and payload resources of each satellite, as well as task identification, subtask details, task time windows, task priorities, and resource requirements in task data. Specifically, the satellite data section provides a detailed record of the resource distribution of each satellite. The task data section decomposes each remote sensing task into several subtasks, which not only indicate the type (such as observation or compression), priority, and required resources, but also clearly define the start and end time windows of the task and the dependencies between tasks. As shown in Table 2, we can intuitively understand how the dataset comprehensively reflects the satellite orbit characteristics, resource distribution, and key parameters in task scheduling.
In order to evaluate the performance of this paper under each evaluation metric, we conducted simulation experiments on 14 datasets with different numbers of satellites and numbers of missions. Table 3 shows the details of the datasets used for the experiments.

5.2. Comparative Algorithm and Performance Metrics

To address the problem’s characteristics of temporal variation, complex resource allocation, block structure preservation, network delay, and hierarchical scheduling, we select five comparative algorithms: Optimizing Scheduled Virtual Machine Requests Placement in Cloud Environments: A Tabu Search Approach (TS-SVMPP) [31] with time complexity O ( n 2 ) ; Cooperative Mapping Task Assignment of Heterogeneous Multi-UAV Using an Improved Genetic Algorithm (IDCE-GA) [30] with time complexity O ( P G ) (in the genetic algorithm, P represents the population size in each generation, and G represents the number of generations); MILP-StuDio: MILP Instance Generation via Block Structure Decomposition (MILP-STUDIO) [25] with time complexity O ( n 3 ) ; Resource Allocation Considering Impact of Network on Performance in a Disaggregated Data Center (RA-CNP) [29] with time complexity O ( n log n ) ; and Task Decomposition and Hierarchical Scheduling for Collaborative Cloud–Edge–End Computing (VNE-AC) [36] with time complexity O ( n ) . In addition, the HEFT-LA algorithm [26] is included in the scheduling category with time complexity O ( n 2 ) , which improves task execution efficiency through a lookahead variant, and the PA-LBIMM algorithm [32] with time complexity O ( P G ) , which integrates user-priority-based Min-Min scheduling to enhance load balancing. These algorithms are commonly used to solve resource allocation and scheduling problems and have efficient optimization effects. TS-SVMPP and IDCE-GA are often used to optimize virtual machine placement and UAV task allocation, demonstrating good adaptability, while MILP-STUDIO, RA-CNP, and VNE-AC exhibit strong optimization capabilities in large-scale and dynamic environments.
For performance evaluation, we employ the following metrics. Task execution rate T C R ( t ) is calculated using the equation
T C R ( t ) = N completed ( t ) N total ( t )
which evaluates the proportion of successfully executed tasks. The objective function value defined in Equation (19) is used to quantify the overall optimization effect of the system. Additionally, resource utilization U r ( t ) serves as a metric to evaluate the efficiency of resource allocation, defined as
U r ( t ) = s S R s used s S R s total
To measure communication costs, we introduce the following metrics:
C C = s S i , j M s c ( T i j ) s S C s
where c ( T i j ) represents the communication cost for task T i j in the communication resource request, and C s represents the total communication capacity of the resource. This metric reflects the communication cost incurred during the execution of tasks, with lower values indicating better communication efficiency. The CC helps evaluate the optimization of communication resources and their impact on the overall system.
In addition, our evaluation criteria also include EQ (task execution quality, where higher values indicate better completion), UQ (execution uncertainty, where lower values indicate smaller errors or delays), and SR (success rate, measured as the ratio of executed tasks to total tasks, expressed as a percentage).

5.3. Numerical Analysis

For each dataset, we performed multiple executions to ensure the robustness and stability of the algorithm, facilitating subsequent comparison and analysis. Our data is based on simulated and synthetic data, reflecting remote sensing satellite tasks in the real world, providing insights into the performance of algorithms in real-world environments.
Figure 3 provides the objective function values and optimization trends for 14 datasets, offering a detailed view of the performance of each algorithm at different task scales. From Table 4, it can be seen that HADRM consistently produces lower objective function values in larger-scale scenarios. For example, in dataset D12, the value of HADRM is 1,523,658, while the value of IDCE-GA is 2,254,874, a decrease of approximately 32.4%. In dataset D14, the value of HADRM2038836 increases by nearly 44.7% compared to the 3,685,412 reported by IDCE-GA. These quantitative differences highlight the enhanced ability of HADRM in managing large-scale complex task environments.
The outstanding performance of HADRM can be attributed to its hierarchical adaptive decomposition and reinforcement learning-based resource mapping mechanism. Specifically, HADRM dynamically decomposes complex task sets into smaller subproblems, simplifying scheduling and reducing computational overhead. Its reinforcement learning component continuously adjusts resource mapping decisions in real time based on constantly changing network conditions and task dependencies. This integrated, data-driven approach enables HADRM to quickly adapt and maintain efficient resource allocation, even in the face of a surge in workload.
In contrast, IDCE-GA and TS-SVMPP, although they perform reasonably in medium-sized scenarios, lack this dynamic adjustment capability. This limitation makes it difficult for them to maintain stable optimization when the task load increases. Similarly, MILP-STUDIO and RA-CNP rely on complex integer programming formulas, which can lead to a rapid increase in computational overhead in large-scale environments. Although VNE-AC sometimes produces competitive results in smaller networks, its static scheduling mechanism limits its adaptability to rapidly changing conditions in a wider and more complex environment. Though HEFT-LA and PA-LBIMM perform reasonably well, they still do not achieve the same level of optimization as HADRM. HEFT-LA lacks the dynamic resource mapping of HADRM, making it less effective in large-scale, dynamic environments. PA-LBIMM, while incorporating user-priority awareness, relies on static scheduling, limiting its adaptability in complex, fluctuating conditions.
Table 5 shows the resource utilization percentages of each algorithm at different task scales, providing us with a detailed quantitative evaluation. For example, in dataset D14, the resource utilization rate of HADRM reached 77.5143%, higher than IDCE-GA’s 53.2478% and TS-SVMPP’s 62.8745%; in dataset D10, the resource utilization rate of HADRM reached 92.2541%, which is also better than the other comparative algorithms. These data fully demonstrate the ability of HADRM to more efficiently utilize available resources in large-scale complex task environments.
As shown in Figure 4a, we first evaluate each algorithm’s resource utilization in a static scenario by gradually increasing the task size. The results indicate that HADRM consistently maintains a high utilization rate, reflecting its advantage in real-time resource allocation as the task size grows.
Next, to investigate how the iteration count affects performance, we conduct a parameter-tuning experiment. Specifically, we adjust the number of iterations and observe the algorithms’ dynamic changes in resource usage. As illustrated in Figure 4b, HADRM continues to achieve the highest utilization among all methods over successive iterations, demonstrating robust and adaptive resource allocation under changing workload conditions.
Its excellent performance is mainly attributed to its hierarchical adaptive decomposition and reinforcement learning-based resource mapping mechanism. Specifically, HADRM simplifies the scheduling process and reduces computational overhead by dynamically decomposing complex task sets into more manageable subproblems; meanwhile, its reinforcement learning module can adjust resource allocation decisions in real time based on constantly changing network conditions and task dependencies.
However, although IDCE-GA and TS-SVMPP perform well in medium-scale scenarios, they lack the necessary dynamic adjustment capabilities, making it difficult for them to maintain stable optimization effects when task loads increase; however, MILP-STUDIO and RA-CNP rely on complex integer programming methods, leading to a rapid increase in computational overhead in large-scale environments. In addition, the static scheduling mechanism of VNE-AC also limits its adaptability in rapidly changing and complex environments. Although HEFT-LA and PA-LBIMM performed well, especially the PA-LBIMM algorithm, which showed competitive results in multiple datasets, HADRM maintained more balanced and efficient resource utilization overall.
As shown in Table 6 and Figure 5a, we measure the task execution time of each algorithm at different task scales, providing a comprehensive quantitative comparison. From these results, it is evident that algorithms such as MILP-STUDIO and TS-SVMPP perform well in small-scale datasets with shorter execution times; however, as the task scale increases, HADRM demonstrates notable advantages. For example, in dataset D14, the execution time of HADRM is 29,652.15, which is substantially lower than MILP-STUDIO’s 37,254.12 and the times of IDCE-GA and TS-SVMPP, both around 39,874.12.
Next, to investigate how the iteration count affects task execution time, we conducted a parameter-tuning experiment. As shown in Figure 4b, HADRM not only achieves lower execution times, but also converges more rapidly over successive iterations, indicating that in large-scale, complex task scenarios, HADRM can more efficiently handle task scheduling and resource allocation.
The advantage of HADRM lies in its adoption of a scheduling strategy that integrates hierarchical adaptive decomposition and reinforcement learning. Specifically, HADRM reduces the dimensionality of problems by decomposing large-scale complex tasks into multiple easily manageable subtasks, and utilizes reinforcement learning modules to monitor network status and task dependency changes in real time, dynamically adjusting scheduling strategies and effectively shortening overall execution time. In contrast, IDCE-GA is based on genetic algorithms, and its evolutionary process converges slowly and is prone to becoming stuck in local optima, making it difficult to search for global optimal solutions in a timely manner when facing large-scale tasks. Although TS-SVMPP uses tabu search to improve local optimization, the search space rapidly expands when the number of tasks increases, resulting in a significant increase in computational burden. MILP-STUDIO relies on integer programming for solving, and its computational complexity increases exponentially, leading to a sharp increase in computation time in large-scale scenarios. RA-CNP adopts a static network performance model, and the scheduling strategy lacks responses to real-time changes in task load and network status. The fixed static scheduling mechanism of VNE-AC cannot effectively cope with the dynamic changes in task dependencies in complex environments, and its overall scheduling efficiency is greatly limited.
For clarity of presentation, we conducted experiments on all 14 datasets but only show representative results for four of them in this table. The Table 7 provides a detailed display of two key metrics for task execution for each algorithm in four datasets ( D 1 D 4 ). From the table, it can be seen that HADRM achieved the best performance in all datasets. For example, in dataset D 1 , HADRM’s EQ reached 78 and its UQ was only 2, significantly better than those of IDCE-GA (EQ 75, UQ 5), TS-SVMPP (EQ 77, UQ 3), MILP-STUDIO (EQ 76, UQ 4), and VNE-AC (EQ 73, UQ 7). In dataset D 2 , HADRM led again with EQ 146 and UQ 14. In datasets D 3 and D 4 , HADRM achieved EQs of 215 and 264, respectively, while its UQs were only 25 and 36. These data fully demonstrate the advantages of HADRM in terms of task execution quality and stability.
The trend of task execution rate in Figure 6 further illustrates the dynamic changes in the performance of various algorithms at different task scales. From the graph, it can be seen that as the task size increases, and HADRM consistently maintains high execution quality and low uncertainty, demonstrating its ability to effectively schedule tasks and reduce errors in high-load environments. This outstanding performance is mainly attributed to the advanced task decomposition and real-time scheduling mechanism adopted by HADRM, which can dynamically split complex tasks into manageable subtasks and flexibly adjust scheduling strategies based on real-time feedback, thereby ensuring efficient and stable execution in large-scale task environments.
In contrast, IDCE-GA relies on traditional genetic algorithms, while TS-SVMPP introduces tabu search to improve local optimization, but the search space rapidly expands when the task size increases. The integer programming method used by MILP-STUDIO has high computational complexity, while RA-CNP is limited by static network models, and the fixed scheduling mechanisms of VNE-AC cannot easily balance efficiency and low uncertainty in large-scale complex environments. HEFT-LA and PA-LBIMM demonstrate competitive performance in terms of makespan and resource utilization but still fall behind HADRM. While they offer improvements in load balancing and user-priority handling, their task execution time and resource utilization are less optimized compared to HADRM’s more efficient scheduling. This is mainly due to the lack of dynamic task mapping in HEFT-LA and the trade-off between VIP and ordinary tasks in PA-LBIMM.
Figure 7 presents the normalized communication cost across all 14 datasets for each algorithm. From the graph, it is clear that HADRM consistently outperforms the other algorithms in minimizing communication cost, especially in large-scale environments. For example, in dataset D 5 , HADRM reduces the communication cost by approximately 2% compared to IDCE-GA, and by over 4% compared to MILP-STUDIO. Additionally, in the dataset D 13 , HADRM achieves a significant reduction of around 5% compared to PA-LBIMM, and a 3% improvement over RA-CNP. These results highlight HADRM’s superior efficiency in managing communication resources, especially when task scale increases.
In contrast, algorithms like IDCE-GA and TS-SVMPP, although they perform well in certain datasets, exhibit a higher communication cost due to the inefficiencies in their communication scheduling mechanisms. Similarly, RA-CNP and VNE-AC struggle with communication overhead, while HEFT-LA and PA-LBIMM, though competitive in some aspects, still lag behind HADRM. This is primarily due to HADRM’s advanced dynamic resource mapping and task scheduling mechanisms, which optimize both task execution and communication cost effectively in large-scale environments.

5.4. Summary

In conclusion, the proposed HADRM method outperformed the comparative algorithms in multiple key performance indicators. HADRM performs particularly well in large-scale task scenarios, with significantly improved resource utilization and significantly better objective function values and task execution times than IDCE-GA, TS-SVMPP, MILP-STUDIO, HEFT-LA, RA-CNP, and VNE-AC. RA-CNP and VNE-AC significantly lag behind. HADRM also consistently outperforms the other algorithms in terms of CC, especially in large-scale environments. Compared to IDCE-GA, TS-SVMPP, and MILP-STUDIO, HADRM achieves a significant reduction in communication cost. It is worth noting that MILP-STUDIO has demonstrated solution accuracy and efficiency on a small task scale that is not inferior to the algorithm proposed in this paper. However, in large-scale models, our method outperforms MILP-STUDIO in terms of resource utilization and execution time. In contrast, IDCE-GA has slower convergence speed and is prone to becoming stuck in local optima due to the limitations of genetic algorithms. When the workload of TS-SVMPP surges, the search space expands rapidly and the computational load increases rapidly. RA-CNP adopts a static network model, which cannot easily adapt to environmental changes. The fixed scheduling mechanism of VNE-AC also limits its response to dynamic changes in task dependencies. Although HEFT-LA and PA-LBIMM perform well in certain evaluation metrics such as resource utilization, HEFT-LA’s local optimal decision-making is enhanced by introducing forward-looking information from task sub-nodes, but still lacks adaptability to dynamic environmental changes. The load balancing mechanism of PA-LBIMM also limits its flexible response to changes in task priority. It is precisely these specific shortcomings that have led to a significant decline in the performance of comparative algorithms in large-scale complex task environments, while HADRM has achieved superior performance due to its real-time dynamic optimization mechanism.

6. Conclusions

In this paper, we solve the problem of low efficiency in complex task decomposition and dynamic resource allocation in multi-satellite collaborative computing. The innovation of this study lies in proposing a dynamic resource mapping method based on task priority using graph theory decomposition, game theory, and deep reinforcement learning, as well as a real-time task scheduling method based on temporal logic.
The experimental results show that in the selected dataset, the objective function value of our method is reduced by at least 32.4% compared to the comparative algorithms (such as IDCE-GA), and the task execution time is significantly reduced, especially in large-scale scenarios where the resource utilization advantage is obvious. However, this method still has shortcomings in terms of parameter tuning and computational complexity control, and further optimization of the algorithm’s convergence and real-time performance is needed in practical applications. In future research, we will explore low-complexity joint optimization strategies and combine digital twin technology to achieve virtual real linkage, in order to further enhance the dynamic adaptability of the algorithm in heterogeneous satellite network collaborative scenarios.

Author Contributions

Conceptualization, S.W.; methodology, S.W. and J.L.; software, C.Z.; validation, S.W. and C.Z.; formal analysis, S.W.; investigation, Z.S.; resources, S.W. and J.L.; data curation, Z.S.; writing—original draft preparation, S.W.; writing—review and editing, S.W. and C.Z.; visualization, Z.S. and C.Z.; supervision, L.L. and J.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant No. U2003208 and No. 61972398), the Science and Technology Plan of Hunan Province (Grant No. 2016TP1003), the Key Technology R&D Program of Hunan Province (Grant No. 2018GK2052), the Foundation for Equipment Procurement (Grant No.6141A020227), and the Theory and Method of Multi-Satellite Collaborative Application Foundation (Grant No. 2019-JCJQ-ZD-342-00).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ouyang, M.; Zhang, R.; Wang, B.; Liu, J.; Huang, T.; Liu, L.; Tong, J.; Xin, N.; Yu, F.R. Network Coding-Based Multipath Transmission for LEO Satellite Networks With Domain Cluster. IEEE Internet Things J. 2024, 11, 21659–21673. [Google Scholar] [CrossRef]
  2. Yin, G.; Liang, Y.; Shi, W.; Zhang, C.; Qi, Y.; Chen, H.; Deng, X.; Li, H. Traffic Offloading Based on Bidirectional Auction in LEO Satellite Networks. In Proceedings of the 2024 4th International Conference on Intelligent Communications and Computing (ICICC), Zhengzhou, China, 18–20 October 2024; pp. 131–134. [Google Scholar] [CrossRef]
  3. Jin, L.; Wang, L.; Jin, X.; Zhu, J.; Duan, K.; Li, Z. Research on the Application of LEO Satellite in IOT. In Proceedings of the 2022 IEEE 2nd International Conference on Electronic Technology, Communication and Information (ICETCI), Online, 27–29 May 2022; pp. 739–741. [Google Scholar] [CrossRef]
  4. Zhang, W.; Huang, H.; Qiu, M. Design and Implementation of Virtual Pottery Space Based on Ceramic Cloud Service Platform. In Proceedings of the 2020 7th IEEE International Conference on Cyber Security and Cloud Computing (CSCloud)/2020 6th IEEE International Conference on Edge Computing and Scalable Cloud (EdgeCom), New York, NY, USA, 1–3 August 2020; pp. 145–150. [Google Scholar] [CrossRef]
  5. Wang, W.; Li, Y.; Tian, H.; Lei, Y.; Li, X.; Jiang, Y. Design and Verification of Autonomous Task Management for Remote Sensing Satellites. In Proceedings of the 2024 IEEE 7th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 15–17 March 2024; Volume 7, pp. 1364–1368. [Google Scholar] [CrossRef]
  6. Hua, Q.; Hongqiang, W.; Jihong, Z.; Yongyue, Y. A Multi-dimensional Resource Allocation Algorithm Based on Task Splitting and Adjustment in Satellite Networks. In Proceedings of the 2022 3rd Information Communication Technologies Conference (ICTC), Nanjing, China, 6–8 May 2022; pp. 66–72. [Google Scholar] [CrossRef]
  7. Wang, Y.; Qi, X.; Lin, X.; Wang, X. Computing Offloading-based Task Scheduling for Space-based Cloud-fog Networks. In Proceedings of the 2021 2nd International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT), Shanghai, China, 15–17 October 2021; pp. 266–270. [Google Scholar] [CrossRef]
  8. Challita, S.; Paraiso, F.; Merle, P. Towards Formal-Based Semantic Interoperability in Multi-Clouds: The FCLOUDS Framework. In Proceedings of the 2017 IEEE 10th International Conference on Cloud Computing (CLOUD), Honolulu, HI, USA, 25–30 June 2017; pp. 710–713. [Google Scholar] [CrossRef]
  9. Tang, X.; Tang, Z.; Cui, S.; Jin, D.; Qiu, J. Dynamic Resource Allocation for Satellite Edge Computing: An Adaptive Reinforcement Learning-based Approach. In Proceedings of the 2023 IEEE International Conference on Satellite Computing (Satellite), Shenzhen, China, 25–26 November 2023; pp. 55–56. [Google Scholar] [CrossRef]
  10. Chen, C.C.; Sun, Z.W.; Zhong, W.C. The satellite optimization design using collaborative optimization method based on normal cloud model. In Proceedings of the Proceedings of 2011 International Conference on Electronic & Mechanical Engineering and Information Technology, Harbin, China, 12–14 August 2011; Volume 7. pp. 3445–3448. [Google Scholar] [CrossRef]
  11. Song, Z.; Hao, Y.; Liu, Y.; Sun, X. Energy-Efficient Multiaccess Edge Computing for Terrestrial-Satellite Internet of Things. IEEE Internet Things J. 2021, 8, 14202–14218. [Google Scholar] [CrossRef]
  12. Hao, Y.; Song, Z.; Zheng, Z.; Zhang, Q.; Miao, Z. Joint Communication, Computing, and Caching Resource Allocation in LEO Satellite MEC Networks. IEEE Access 2023, 11, 6708–6716. [Google Scholar] [CrossRef]
  13. Ren, B.; Liu, J.; Li, Z.; Wu, J.; Yao, Z. Satellite Requirement Preference Driven TT&C Resources Scheduling Algorithm for Time Sensitive Missions. In Proceedings of the 2020 IEEE 3rd International Conference on Electronic Information and Communication Technology (ICEICT), Shenzhen, China, 13–15 November 2020; pp. 15–19. [Google Scholar] [CrossRef]
  14. Yang, X.; Huang, R.; Wu, J.; Li, J.; Wang, Y.; Wu, W.; Liu, J. Intelligent Elastic Scheduling Algorithm for Multi-Resource Low Earth Orbit Satellite Networks. In Proceedings of the 2024 3rd International Conference on Computing, Communication, Perception and Quantum Technology (CCPQT), Zhuhai, China, 25–27 October 2024; pp. 323–327. [Google Scholar] [CrossRef]
  15. Fang, Z.; Liu, Y.; Zhu, K.; Sun, F.; Shi, C.; Xie, R.; Tang, Q.; Huang, T. Dynamic Task Scheduling in LEO Satellites Mega-constellations: A Domain-based Perspective. In Proceedings of the 2023 9th International Conference on Computer and Communications (ICCC), Chengdu, China, 8–11 December 2023; pp. 355–360. [Google Scholar] [CrossRef]
  16. Zhiliang, L.; Xiaojiang, L. Current Status and Prospect of Imaging Satellite Task Dynamic Scheduling Methods. In Proceedings of the 2016 8th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China, 27–28 August 2016; Volume 02, pp. 436–439. [Google Scholar] [CrossRef]
  17. Dai, C.; Zheng, G.; Chen, Q. Satellite constellation design with multi-objective genetic algorithm for regional terrestrial satellite network. China Commun. 2018, 15, 1–10. [Google Scholar] [CrossRef]
  18. Sarkheyli, A.; Ghorbani Vaghei, B.; Bagheri, A. New tabu search heuristic in scheduling earth observation satellites. In Proceedings of the 2010 2nd International Conference on Software Technology and Engineering, San Juan, Puerto Rico, 3–5 October 2010; Volume 2, pp. V2–199–V2–203. [Google Scholar] [CrossRef]
  19. Zhang, Z.; Sun, X.; Hou, L.; Chen, W.; Shi, Y.; Cao, X. A cooperative co-evolutionary multi-agent system for multi-objective layout optimization of satellite module. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; pp. 147–151. [Google Scholar] [CrossRef]
  20. Kieffer, E.; Stathakis, A.; Danoy, G.; Bouvry, P.; Talbi, E.G.; Morelli, G. Multi-objective evolutionary approach for the satellite payload power optimization problem. In Proceedings of the 2014 IEEE Symposium on Computational Intelligence in Multi-Criteria Decision-Making (MCDM), Orlando, FL, USA, 9–12 December 2014; pp. 202–209. [Google Scholar] [CrossRef]
  21. Wang, M.; Pei, C.; Chen, X.; Dai, G.; Song, Z.; Peng, L. A dynamic decomposition optimization framework for multi-satellite scheduling of area target observation. Adv. Space Res. 2025, 65, 123–135. [Google Scholar] [CrossRef]
  22. Li, J.; Yao, F.; Bai, B.; He, R. A Decomposition-Based Algorithm for Imaging Satellites Scheduling Problem. In Proceedings of the 2009 International Conference on Information Engineering and Computer Science, Taiyuan, China, 10–11 July 2009; pp. 1–6. [Google Scholar] [CrossRef]
  23. Chai, F.; Zhang, Q.; Yao, H.; Xin, X.; Gao, R.; Guizani, M. Joint Multi-Task Offloading and Resource Allocation for Mobile Edge Computing Systems in Satellite IoT. IEEE Trans. Veh. Technol. 2023, 72, 7783–7795. [Google Scholar] [CrossRef]
  24. Yao, F.; Li, J.; Chen, Y.; Chu, X.; Zhao, B. Task allocation strategies for cooperative task planning of multi-autonomous satellite constellation. Adv. Space Res. 2019, 63, 1073–1084. [Google Scholar] [CrossRef]
  25. Liu, H.; Wang, J.; Zhang, W.; Geng, Z.; Kuang, Y.; Li, X.; Li, B.; Zhang, Y.; Wu, F. MILP-StuDio: MILP Instance Generation via Block Structure Decomposition. arXiv 2024, arXiv:2410.22806. [Google Scholar] [CrossRef]
  26. Bittencourt, L.F.; Sakellariou, R.; Madeira, E.R. Dag scheduling using a lookahead variant of the heterogeneous earliest finish time algorithm. In Proceedings of the 2010 18th Euromicro Conference on Parallel, Distributed and Network-Based Processing, Pisa, Italy, 17–19 February 2010; IEEE: New York, NY, USA, 2010; pp. 27–34. [Google Scholar] [CrossRef]
  27. Chen, Y.; Xu, M.; Shen, X.; Zhang, G.; Lu, Z.; Xu, J. A multi-objective modeling method of multi-satellite imaging task planning for large regional mapping. Remote Sens. 2020, 12, 344. [Google Scholar] [CrossRef]
  28. Xiaolu, L.; Baocun, B.; Yingwu, C.; Feng, Y. Multi satellites scheduling algorithm based on task merging mechanism. Appl. Math. Comput. 2014, 230, 687–700. [Google Scholar] [CrossRef]
  29. Ikoma, A.; Ohsita, Y.; Murata, M. Resource Allocation Considering Impact of Network on Performance in a Disaggregated Data Center. IEEE Access 2024, 67600–67618. [Google Scholar] [CrossRef]
  30. Li, J.; Yang, X.; Yang, Y.; Liu, X. Cooperative mapping task assignment of heterogeneous multi-UAV using an improved genetic algorithm. Knowl.-Based Syst. 2024, 296, 111830. [Google Scholar] [CrossRef]
  31. Koubàa, M.; Karar, A.S.; Bahloul, F. Optimizing Scheduled Virtual Machine Requests Placement in Cloud Environments: A Tabu Search Approach. Computers 2024, 13, 321. [Google Scholar] [CrossRef]
  32. Chen, H.; Wang, F.; Helian, N.; Akanmu, G. User-priority guided Min-Min scheduling algorithm for load balancing in cloud computing. In Proceedings of the 2013 National Conference on Parallel Computing Technologies (PARCOMPTECH), Karnataka, India, 21–23 February 2013; IEEE: New York, NY, USA, 2013; pp. 1–8. [Google Scholar] [CrossRef]
  33. Fan, H.; Yang, Z.; Zhang, X.; Wu, S.; Long, J.; Liu, L. A novel multi-satellite and multi-task scheduling method based on task network graph aggregation. Expert Syst. Appl. 2022, 205, 117565. [Google Scholar] [CrossRef]
  34. Zhang, H.; Zhao, H.; Liu, R.; Kaushik, A.; Gao, X.; Xu, S. Collaborative Task Offloading Optimization for Satellite Mobile Edge Computing Using Multi-Agent Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2024, 73, 15483–15498. [Google Scholar] [CrossRef]
  35. Yao, F.; Li, J.; Bai, B.; He, R. Earth observation satellites scheduling based on decomposition optimization algorithm. Int. J. Image Graph. Signal Process. 2010, 2, 10–18. [Google Scholar] [CrossRef]
  36. Cai, J.; Liu, W.; Huang, Z.; Yu, F.R. Task Decomposition and Hierarchical Scheduling for Collaborative Cloud-Edge-End Computing. IEEE Trans. Serv. Comput. 2024, 17, 4368–4382. [Google Scholar] [CrossRef]
  37. Zilberstein, I.; Rao, A.; Salis, M.; Chien, S. Decentralized, decomposition-based observation scheduling for a large-scale satellite constellation. J. Artif. Intell. Res. 2025, 82, 169–208. [Google Scholar] [CrossRef]
  38. Deng, B.; Jiang, C.; Kuang, L.; Guo, S.; Lu, J.; Zhao, S. Two-phase task scheduling in data relay satellite systems. IEEE Trans. Veh. Technol. 2017, 67, 1782–1793. [Google Scholar] [CrossRef]
  39. Alandihallaj, M.A.; Emami, M.R. Satellite replacement and task reallocation for multiple-payload fractionated Earth observation mission. Acta Astronaut. 2022, 196, 157–175. [Google Scholar] [CrossRef]
  40. Tang, Q.; Xie, R.; Fang, Z.; Huang, T.; Chen, T.; Zhang, R.; Yu, F.R. Joint Service Deployment and Task Scheduling for Satellite Edge Computing: A Two-Timescale Hierarchical Approach. IEEE J. Sel. Areas Commun. 2024, 42, 1063–1079. [Google Scholar] [CrossRef]
  41. Shi, J.; Chen, X.; Zhang, Y.; Chen, X.; Pan, C. Joint Optimization of Task Offloading and Resource Allocation in Satellite-Assisted IoT Networks. IEEE Internet Things J. 2024, 11, 34337–34348. [Google Scholar] [CrossRef]
  42. Li, J.; Wu, G.; Liao, T.; Fan, M.; Mao, X.; Pedrycz, W. Task Scheduling Under a Novel Framework for Data Relay Satellite Network via Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2023, 72, 6654–6668. [Google Scholar] [CrossRef]
  43. Han, D.; Ye, Q.; Peng, H.; Wu, W.; Wu, H.; Liao, W.; Shen, X. Two-Timescale Learning-Based Task Offloading for Remote IoT in Integrated Satellite–Terrestrial Networks. IEEE Internet Things J. 2023, 10, 10131–10145. [Google Scholar] [CrossRef]
  44. Sinha, P.K.; Dutta, A. Multi-satellite task allocation algorithm for Earth observation. In Proceedings of the 2016 IEEE Region 10 Conference (TENCON), Singapore, 22–25 November 2016; pp. 403–408. [Google Scholar] [CrossRef]
  45. Qi, W.; Yang, W.; Xing, L.; Yao, F. Modeling and Solving for Multi-Satellite Cooperative Task Allocation Problem Based on Genetic Programming Method. Mathematics 2022, 10, 3608. [Google Scholar] [CrossRef]
  46. Xiong, J.; Leus, R.; Yang, Z.; Abbass, H.A. Evolutionary multi-objective resource allocation and scheduling in the Chinese navigation satellite system project. Eur. J. Oper. Res. 2016, 251, 662–675. [Google Scholar] [CrossRef]
  47. Lyu, T.; Xu, Y.; Liu, F.; Xu, H.; Han, Z. Task Offloading and Resource Allocation for Satellite-Terrestrial Integrated Networks. IEEE Internet Things J. 2025, 12, 262–275. [Google Scholar] [CrossRef]
  48. Lv, M.; Peng, X.; Xie, W.; Guan, N. Task Allocation for Real-time Earth Observation Service with LEO Satellites. In Proceedings of the 2022 IEEE Real-Time Systems Symposium (RTSS), Houston, TX, USA, 5–8 December 2022; pp. 14–26. [Google Scholar] [CrossRef]
  49. Li, G.; Li, X.; Li, J.; Chen, J.; Shen, X. PTMB: An online satellite task scheduling framework based on pre-trained Markov decision process for multi-task scenario. Knowl.-Based Syst. 2024, 284, 111339. [Google Scholar] [CrossRef]
  50. Huang, J.; Xing, R.; Ma, X.; Zhou, A.; Wang, S. Profit-Aware Task Allocation in Satellite Computing. In Proceedings of the 2024 IEEE International Conference on Web Services (ICWS), Shenzhen, China, 7–13 July 2024; pp. 696–706. [Google Scholar] [CrossRef]
  51. Ramírez, T.; Mosquera, C.; Alagha, N. Flexible User Mapping for Radio Resource Assignment in Advanced Satellite Payloads. IEEE Trans. Broadcast. 2022, 68, 723–739. [Google Scholar] [CrossRef]
  52. Du, H.; Huang, Y.; Cao, T.; Ji, T.; Lin, J.; Jin, S.; Peng, X.; Chen, M. User Mapping and Radio Resource Management in Demand-based GEO Satellite Systems. In Proceedings of the 2024 IEEE Wireless Communications and Networking Conference (WCNC), Dubai, United Arab Emirates, 21–24 April 2024; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. Schematic of main idea of HADRM method.
Figure 1. Schematic of main idea of HADRM method.
Mathematics 13 01183 g001
Figure 2. A three-dimensional global simulation depicting satellite orbital paths.
Figure 2. A three-dimensional global simulation depicting satellite orbital paths.
Mathematics 13 01183 g002
Figure 3. Objective function values of eight algorithms across 14 datasets.
Figure 3. Objective function values of eight algorithms across 14 datasets.
Mathematics 13 01183 g003
Figure 4. Comparative resource utilization in static and iterative scenarios. (a) Resource utilization. (b) Resource utilization over varying iterations.
Figure 4. Comparative resource utilization in static and iterative scenarios. (a) Resource utilization. (b) Resource utilization over varying iterations.
Mathematics 13 01183 g004
Figure 5. Comparative task execution times in static and iterative scenarios. (a) Task execution times. (b) Task execution times in different iterations.
Figure 5. Comparative task execution times in static and iterative scenarios. (a) Task execution times. (b) Task execution times in different iterations.
Mathematics 13 01183 g005
Figure 6. Task execution rates of eight algorithms across four datasets.
Figure 6. Task execution rates of eight algorithms across four datasets.
Mathematics 13 01183 g006
Figure 7. Communication Cost of eight algorithms across 14 datasets.
Figure 7. Communication Cost of eight algorithms across 14 datasets.
Mathematics 13 01183 g007
Table 1. A detailed description of remote sensing tasks.
Table 1. A detailed description of remote sensing tasks.
Real DatasetTask IDRegionDiscrete Time RangeResource Type
Latitude (deg)Longitude (deg)Start Time (s)End Time (s)All
D r e a l T10−0.2808061086,4501,2,3
089.5788
59.802689.5788
59.8026−0.280806
T2059.81281086,4501,2,3
29.605359.8128
29.6053−30.0468
−0.592105−30.0468
Table 2. Field definitions and value ranges.
Table 2. Field definitions and value ranges.
Satellite Data
Field NameNotationValue Range
Satellite ID s i 1–300
Computation Resource r b ( s i ) 70–150
Storage Resource C i 300–500 MB
Payload Resource r p ( s i ) 1–3
Task ID m i 1–300
Subtask ID T i j
Subtask Typecompression, observation
Priority ω i j 1–5
Computation Requirement r b ( T i j ) 32–100
Storage Requirement D ( T i j ) 173–294
Payload Requirement r p ( T i j ) 0–3
Start Time E k 2024-09-06 T 18:29:44 Z
End Time L k 2024-09-06 T 20:29:44 Z
Dependenciesdependenciesempty or list of Subtask IDs
Table 3. Number of satellites and tasks in selected datasets.
Table 3. Number of satellites and tasks in selected datasets.
Datasets D 1 D 2 D 3 D 4 D 5 D 6 D 7
Satellite Nodes1020406080100120
Task Quantity204060100120140150
Datasets D 8 D 9 D 10 D 11 D 12 D 13 D 14
Satellite Nodes140160180200220240300
Task Quantity200240280300360400300
Table 4. Objective function value results across 14 datasets.
Table 4. Objective function value results across 14 datasets.
DatasetsHEFT-LA [26]PA-LBIMM [32]IDCE-GA [30]TS-SVMPP [31]MILP-STUDIO [25]RA-CNP [29]VNE-AC [36]HADRM
D 1 12,031.1211,530.5712,058.489102.379524.489500.258964.2511,031.12
D 2 24,041.9823,012.3213,548.4812,001.3311,548.4812,314.6612,856.3522,541.99
D 3 76,841.2575,341.26578,412.13520,147.66508,741.13525,478.99553,284.2775,321.26
D 4 107,481.54105,486.1798,521.4889,214.6581,425.4890,214.5595,632.46105,481.55
D 5 265,145.89252,347.36198,745.48194,512.37174,785.48198,523.13204,654.30252,145.90
D 6 425,512.32420,134.12451,478.33412,654.75386,521.13435,254.74426,587.11404,512.33
D 7 765,648.25752,218.74815,478.651,014,520.85964,587.331,023,200.991,045,545.99745,648.25
D 8 1,021,457.151,051,234.281,158,745.211,302,541.661,224,587.661,325,000.551,355,596.651,152,145.15
D 9 1,223,521.251,212,356.981,387,452.151,554,789.541,458,741.211,574,215.651,580,025.241,203,521.25
D 10 1,387,582.361,378,620.541,758,412.661,812,254.151,725,412.651,823,654.551,856,833.001,366,582.37
D 11 1,652,645.291,556,321.271,854,785.211,925,874.661,798,541.151,954,785.551,972,285.361,542,645.30
D 12 1,634,658.481,534,625.722,254,874.212,203,547.122,145,874.212,245,214.992,270,025.691,523,658.48
D 13 1,774,511.541,611,241.282,458,741.662,365,412.152,314,789.652,385,412.662,402,569.551,604,511.55
D 14 2,258,836.492,052,354.123,685,412.333,502,541.653,414,789.333,552,148.663,595,541.002,038,836.49
Table 5. Resource utilization across 14 datasets.
Table 5. Resource utilization across 14 datasets.
DatasetsHEFT-LA [26]PA-LBIMM [32]IDCE-GA [30]TS-SVMPP [31]MILP-STUDIO [25]RA-CNP [29]VNE-AC [36]HADRM
D 1 93.0214%95.1215%89.5412%96.5874%97.8541%96.0415%98.5415%94.5214%
D 2 92.6521%94.4542%86.9874%97.5412%96.9874%97.1245%98.1123%94.1521%
D 3 90.5426%90.5124%87.4578%95.5874%96.5874%95.5874%97.5641%91.5426%
D 4 86.0141%87.5126%74.5478%87.2542%88.5478%87.4523%88.6541%87.2141%
D 5 87.0541%88.5126%72.5874%86.0987%87.4574%86.0145%88.3354%88.6541%
D 6 84.0424%84.9732%75.2145%82.5412%89.5214%81.2542%84.5412%85.1424%
D 7 86.0145%88.7215%76.5874%83.0124%91.0874%82.8124%84.5524%88.2145%
D 8 81.5471%83.7423%72.8541%79.8452%88.2354%80.2531%80.5569%83.5471%
D 9 82.5487%83.6523%69.8574%76.4521%85.0214%76.8421%77.3695%83.5487%
D 10 89.2541%92.4542%68.5478%73.1254%83.2478%72.6589%73.5586%92.2541%
D 11 81.0424%82.0652%64.5874%71.8745%81.0587%71.0548%73.5548%81.5496%
D 12 81.0493%82.0632%61.2478%70.1254%79.0154%69.3214%72.2154%81.5493%
D 13 79.0469%80.7392%57.8541%66.5478%76.5874%65.8742%67.5541%80.2469%
D 14 76.5143%77.9435%53.2478%62.8745%72.5874%61.8541%63.2895%77.5143%
Average85.0246%86.6049%72.2428%79.3857%90.1071%79.3928%88.0500%83.1071%
Table 6. Task execution time across 14 datasets.
Table 6. Task execution time across 14 datasets.
Algorithms D 1 D 2 D 3 D 4 D 5 D 6 D 7
HEFT-LA [26] 320.35810.211070.352600.221800.234350.258650.65
PA-LBIMM [32] 325.46815.211045.222558.751760.354480.778664.25
IDCE-GA [30] 354.21412.48854.212854.333145.484854.339124.21
TS-SVMPP [31] 220.13390.21705.212175.132385.664550.484789.15
MILP-STUDIO [25] 198.13356.48685.211954.332187.483845.334201.21
RA-CNP [29] 245.66410.24710.132204.132405.664900.545250.33
VNE-AC [36] 215.66452.16769.352335.272546.294685.894961.23
HADRM314.25798.211044.352554.221752.234265.258547.65
Algorithms D 8 D 9 D 10 D 11 D 12 D 13 D 14
HEFT-LA [26] 5700.657630.339764.3312,850.3310,250.6611,300.6529,950.15
PA-LBIMM [32] 5634.827637.549684.3312,647.5410,124.7611,117.6530,052.54
IDCE-GA [30] 6541.1512,548.4815,847.3318,457.1321,874.4827,541.2139,874.13
TS-SVMPP [31] 5987.4512,012.4814,985.2119,458.3322,514.4826,847.4539,874.13
MILP-STUDIO [25] 5001.1510,845.4813,487.3317,845.1320,541.4824,847.2137,254.13
RA-CNP [29] 6100.1512,245.6615,234.6619,852.1322,841.6627,041.6640,532.13
VNE-AC [36] 6553.4912,509.3715,660.3320,053.4423,360.2327,654.1241,025.67
HADRM5562.657536.339564.3312,541.3310,032.6611,054.6529,652.15
Table 7. Task execution performance across 4 datasets.
Table 7. Task execution performance across 4 datasets.
Algorithms D 5 D 9 D 13 D 14
EQUQSREQUQSREQUQSREQUQSR
HEFT-LA [26]77396.22%1441689.99%2103087.45%2613987.01%
PA-LBIMM [32]77395.77%1421888.55%2083286.59%2564485.30%
IDCE-GA [30]75593.75%1382286.25%2063485.83%2505083.33%
TS-SVMPP [31]77396.25%1431789.38%2103087.50%2534784.33%
MILP-STUDIO [25]76495.00%1441690.00%2103087.50%2604086.67%
RA-CNP [29]77396.25%1431789.38%2093187.08%2613987.00%
VNE-AC [36]73791.25%1342683.75%1994182.92%2435781.00%
HADRM78297.50%1461491.25%2152589.58%2643688.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, S.; Zhang, C.; Su, Z.; Liu, L.; Long, J. Multi-Satellite Task Parallelism via Priority-Aware Decomposition and Dynamic Resource Mapping. Mathematics 2025, 13, 1183. https://doi.org/10.3390/math13071183

AMA Style

Wang S, Zhang C, Su Z, Liu L, Long J. Multi-Satellite Task Parallelism via Priority-Aware Decomposition and Dynamic Resource Mapping. Mathematics. 2025; 13(7):1183. https://doi.org/10.3390/math13071183

Chicago/Turabian Style

Wang, Shangpeng, Chenyuan Zhang, Zihan Su, Limin Liu, and Jun Long. 2025. "Multi-Satellite Task Parallelism via Priority-Aware Decomposition and Dynamic Resource Mapping" Mathematics 13, no. 7: 1183. https://doi.org/10.3390/math13071183

APA Style

Wang, S., Zhang, C., Su, Z., Liu, L., & Long, J. (2025). Multi-Satellite Task Parallelism via Priority-Aware Decomposition and Dynamic Resource Mapping. Mathematics, 13(7), 1183. https://doi.org/10.3390/math13071183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop