Next Article in Journal
An End-to-End Generation Model for Chinese Calligraphy Characters Based on Dense Blocks and Capsule Network
Previous Article in Journal
A Novel SCA-Based SOCP Method for Time-Varying Beamforming Optimization in RIS-Assisted MU-MISO Downlinks
Previous Article in Special Issue
Intelligent Security Authentication for Connected and Autonomous Vehicles: Attacks and Defenses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computation Offloading Strategy for Detection Task in Railway IoT with Integrated Sensing, Storage, and Computing

1
College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
2
Beijing Key Laboratory of Work Safety Intelligent Monitoring, School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(15), 2982; https://doi.org/10.3390/electronics13152982 (registering DOI)
Submission received: 3 July 2024 / Revised: 26 July 2024 / Accepted: 26 July 2024 / Published: 29 July 2024
(This article belongs to the Special Issue Control Systems Design for Connected and Autonomous Vehicles)

Abstract

:
Online detection devices, powered by artificial intelligence technologies, enable the comprehensive and continuous detection of high-speed railways (HSRs). However, the computation-intensive and latency-sensitive nature of these detection tasks often exceeds local processing capabilities. Mobile Edge Computing (MEC) emerges as a key solution in the railway Internet of Things (IoT) scenario to address these challenges. Nevertheless, the rapidly varying channel conditions in HSR scenarios pose significant challenges for efficient resource allocation. In this paper, a computation offloading system model for detection tasks in the railway IoT scenario is proposed. This system includes direct and relay transmission models, incorporating Non-Orthogonal Multiple Access (NOMA) technology. This paper focuses on the offloading strategy for subcarrier assignment, mode selection, relay power allocation, and computing resource management within this system to minimize the average delay ratio (the ratio of delay to the maximum tolerable delay). However, this optimization problem is a complex Mixed-Integer Non-Linear Programming (MINLP) problem. To address this, we present a low-complexity subcarrier allocation algorithm to reduce the dimensionality of decision-making actions. Furthermore, we propose an improved Deep Deterministic Policy Gradient (DDPG) algorithm that represents discrete variables using selection probabilities to handle the hybrid action space problem. Our results indicate that the proposed system model adapts well to the offloading issues of detection tasks in HSR scenarios, and the improved DDPG algorithm efficiently identifies optimal computation offloading strategies.

1. Introduction

1.1. Background

With the rapid development of communication and artificial intelligence technologies, new horizons for development in the field of rail transit operation and maintenance have emerged. Particularly for high-speed railways (HSRs), due to the fast velocities of high-speed trains, the high safety standards required, the complexity of the operational environment, the substantial passenger load, and the prolonged duration of high-speed travel, there is a significant safety risk [1] associated with the detection of high-speed rail vehicles. The detection process for high-speed trains still relies heavily on manual efforts, with routine checks constituting the bulk of maintenance activities, leading to high safety risks, elevated maintenance costs, and low efficiency. The deployment of online detection devices, big data analytics [2], and machine vision-based non-contact non-destructive testing techniques [3] could allow for the synchronous and continuous detection of the condition of rail transit vehicles and the status of railway lines without stopping, without sacrificing accuracy and comprehensiveness. However, these detection tasks are typically computation-intensive and latency-sensitive. Equipping each detection device with extensive computing power is clearly a high-cost and wasteful approach.
A railway Internet of Things (IoT) [4] system with integrated sensing, storage, and computing, based on mobile edge computing (MEC) [5], can effectively address this issue. Such an integrated system embeds certain edge computing nodes that provide the necessary computing power and memory capacity for detection tasks. Thus, by utilizing MEC, these detection tasks can be offloaded to edge servers for processing. This approach can offer low-latency computational services and life-cycle-long management [6] for detection tasks.
Although some progress has been made in the research on the allocation of communication and computing resources in MEC based on optimization methods [7,8] within traditional IoT environments, the channel conditions in railway scenarios are rapidly time-varying, particularly for HSR. Moreover, given the massive number of detection devices along the entire railway, the computing power and memory capacity [9] of edge servers may not always be sufficient to immediately serve every task. In short, in dynamic and resource-constrained environments, system and environmental parameters are constantly changing. Thus, finding an optimal allocation strategy for communication, computing, and memory resources for detection tasks quickly and effectively is challenging.

1.2. Related Work and Contribution

Numerous noteworthy studies have concentrated on investigating the issues of computation offloading and resource management within the domain of MEC. Optimization-based methods are the primary solutions for addressing this problem. Zhao et al. [10] aimed to minimize the system’s maximum delay by optimizing user beamforming vectors, BS beamforming matrices, and MEC server resource allocation while adhering to communication and computing budget constraints. Xu et al. [8] proposed a novel approach to optimize network throughput in HSR scenarios by admitting as many tasks as possible under delay constraints and multicasting the maximum number of computation results, considering the high mobility of trains and frequent handovers. They introduced a multi-group-shared Group Steiner tree model and an efficient heuristic algorithm to address the multicast routing problem, enhancing the quality of data services for passengers during long journeys. Deep Reinforcement Learning (DRL) is being increasingly applied in resource allocation problems. Jiang et al. [11] investigated the optimization of task execution, focusing on offloading decisions and resource assignments in MEC environments to reduce task latency, and used Q-learning to solve the problem. Chen et al. [12] explored energy efficiency in AR task offloading and resource allocation in single and multiple MEC system environments and proposed a multiagent deep reinforcement learning model for dynamic communication environments. Ale et al. [13] tackled the challenge of task offloading in environments with multiple IoT devices and edge servers, formulating the problem as a Markov decision process (MDP) and introducing a DRL algorithm for hybrid action spaces that involve both task partitioning and computational power allocation. Deng et al. [14] developed an autonomous system for the partial offloading of delay-sensitive tasks in multiuser IIoT MEC systems, employing the Deep Deterministic Policy Gradient (DDPG) to facilitate continuous offloading decisions and optimize system performance.
While research has explored computation offloading and resource allocation in MEC from diverse optimization perspectives, the majority of these research studies focus on quasi-static system settings and do not take into account the requirements for multiple resources by dynamic systems with time-varying conditions in high-speed contexts. In the dynamic sector of MEC for railway IoT, Xu et al. [5] crafted a hybrid DRL approach that synergizes a Double Deep Q-Network (DDQN) and DDPG to minimize computational costs by fine-tuning subcarrier assignment, offloading ratios, and resource distributions. Addressing HSR, Li et al. [15] developed a joint resource allocation and computation offloading scheme based on millimeter waves and matched game theory to curtail latency and bolster user capacity. To maximize throughput, Xu et al. [8] addressed the issue of maximizing network throughput in HSR scenarios by proposing an efficient heuristic algorithm for multicasting computation results in an MEC network. The approach considered the dynamic nature of HSRs, optimizing multicast tree construction and task admission under delay constraints to enhance data services for passengers. Lastly, Tian et al. [16] introduced a blockchain-enhanced, multi-hop transmission framework for MEC in smart rail systems, applying DRL to refine offloading paths and strategies for improved system performance.
With the development of communication technology, many technologies have become increasingly mature and applied in practice, such as millimeter wave and multiple-input multiple-output technologies [17]. The introduction of these technologies has also brought more possibilities to MEC. In particular, the Non-Orthogonal Multiple Access (NOMA) technology has brought about significant performance enhancements. Xu et al. [18] tackled the complex joint task offloading and resource allocation in NOMA-based HetNets for MEC, proposing an iterative solution to the combinatorial problem. Du et al. [19] studied cost-effective offloading in NOMA-enabled vehicular networks, developing a Vehicle Edge Computing (VEC) network model to address the cost minimization challenge through heuristic algorithms and deriving closed-form solutions for cloud-based optimizations. Du et al. [20] introduced a NOMA-assisted ultra-dense edge computing model optimized by DDPG to enhance spectrum utilization and reduce task execution delay and energy consumption in healthcare IoT, outperforming local computation and traditional offloading methods.
Although some of the aforementioned papers have applied MEC to railway scenarios, the impact of high-speed movement on communication is often overlooked. Moreover, detection devices may be deployed on trains or alongside tracks, adding uncertainty and complexity to the computation offloading problem. However, summarizing these issues and considering the specific environment of high-speed railway IoT for train vehicle and track detection, we propose a computation offloading strategy for detection tasks in railway IoT with integrated sensing, storage, and computing. In this paper, we consider the interference caused by high-speed movement. Our system supports two communication modes, direct communication and relay communication, to accommodate the uncertainty in device deployment. Additionally, the differences in device and relay transmission power open up possibilities for the introduction of NOMA technology. We incorporate NOMA technology on top of Orthogonal Frequency Division Multiplexing (OFDM) to enhance channel utilization. By selecting appropriate edge servers, we also mitigate data congestion on each server. Building upon the aforementioned system model, we adopt an improved DDPG algorithm in conjunction with a low-complexity channel allocation algorithm to optimize the configuration of communication and computational resources. This study has threefold specific contributions, as follows:
  • To better accommodate detection devices deployed in different locations and environments in various types of railways, we proposed two transmission models, namely the direct transmission model and the relay transmission model. Furthermore, in the OFDM system, we introduced NOMA technology to increase channel capacity.
  • Taking into account the limitation of computing resources and memory capacity, we allocated computing resources through server selection, combined with the order of arrival, while also ensuring that the task queue pending for each server does not exceed its memory capacity.
  • We proposed an improved DDPG algorithm combined with a low-complexity subcarrier allocation algorithm to solve the optimization problem of communication and computing resource allocation.
The organization of the rest of this paper is as follows. Section 2 describes the model of the MEC system in a railway IoT system for detection tasks and optimization problem formulation. Section 3 proposes the subcarrier allocation algorithm and the improved DDPG algorithm. Section 4 presents and analyzes the simulation results. Section 5 concludes this paper.

2. System Model and Problem Formulation

In this section, we initially outline the network framework of MEC for detection tasks in the railway IoT with integrated sensing, storage, and computing. We then present two transmission models, the NOMA-based transmission model and the limited computing resource and memory capacity allocation model. Finally, we formulate an optimization problem to minimize the average delay ratio.

2.1. System Model

As shown in Figure 1, we consider an MEC system beside HSR to provide task offloading services for multiple detection devices. Within a certain range along the track, a Base Station (BS) is deployed beside the railway, forming a cell. Meanwhile, a communication relay is installed on the train. Near the BS, there is an MEC server installed. The MEC servers along the entire railway are interconnected via fiber optic links to form an edge computing network. To achieve multi-directional intelligent safety detection in HSR, detection devices incorporate onboard detection sensors and ground detection sensors. The detection devices in this system are equipped with the necessary capabilities for detection and communication but lack computing capacity. Thus, the various detection tasks need to be offloaded to MEC servers for processing. The detection devices have two transmission modes. One is direct transmission, in which the detection devices transmit the detection tasks directly to the BS. The other is relay transmission, in which the detection devices first send the tasks to the communication relay on the train, which then forwards them to the BS.
We assume that, in a time slot, there are M onboard devices and N ground devices that have requests for task offloading. Therefore, a total of I tasks need to be offloaded to the MEC, denoted by I = { 1 , 2 , , I } , where I = M + N , { 1 , , M } represents the task from the onboard devices and { M + 1 , , I } represents the task from the ground devices. In a time slot, the i - th task, i I , is represented by a quadruple l i , μ i , T i max , L i , where l i is the data size of the i - th task in bits, μ i is the computation workload of the i - th task in cycles/bit, T i max is the maximum allowable delay for the i - th task, and L i is the loaded data size for the i - th task, including the data size of the neural network model to be loaded and the intermediate data generated during the calculation [21,22].

2.2. Direct Transmission Model

The detection devices directly transmit the tasks to the BS of the cell where the train is located, which is referred to as sensor-to-BS (S2B) data transmission. We assume that the transmission power of each detection device can be monitored before making decisions, and there will not be significant differences among the devices. This implies that the channel gain differences among different tasks during direct communication will not be too large. Thus, we adopt OFDM technology for the communication between ground devices and the BSs. Interference between subcarriers is negligible. At the same time, in the HSR environments, since the distance between trains exceeds 50 km, the interference between cells can be considered negligible.
However, for the detection tasks of onboard devices, namely i { 1 , , M } , due to the high mobility of trains in the HSR environments, the transmitted signal will inevitably experience a Doppler shift, which leads to inter-channel interference (ICI) [23]. This interference can be considered a form of noise. Thus, the received Signal-to-Interference-plus-Noise Ratio (SINR) [24] of the detection i - th task created by the onboard device can be expressed as:
Γ i B = p i | h i B | 2 ϖ p i | h i B | 2 + σ 2 ,
where σ 2 is the noise power, p i is the transmission power of the i - th device, | h i B | 2 is the channel gain between the i - th device and the BS, and ϖ is the ICI factor due to the Doppler effect [24], which is given by:
ϖ = 1 1 1 ( 1 | x | ) J 0 ( 2 π f c v c τ s x ) d x ,
where v and c represent the velocity of the train and the speed of light, respectively. τ s is the signal symbol duration. J 0 ( · ) is the zero-order Bessel function of the first kind.
However, for the ground detection devices, namely i { M + 1 , , M + N } , which are fixed as is the BS, there is no other interference. The received SINR of the i - th task created by the ground device can be expressed as:
Γ i B = p i | h i B | 2 σ 2 .
Assuming there are C subcarriers in this OFDM system, λ i , c denotes the variable of subcarrier allocated, where λ i , c { 0 , 1 } . When λ i , c is 1, the i - th task is offloaded to the BS via the c - th subcarrier. Otherwise, λ i , c is 0. Thus, the transmission rate for the i - th task can be expressed as:
R i B = c = 1 C λ i , c W   log 2 ( 1 + Γ i B ) ,
where W is the subcarrier bandwidth.

2.3. Relay Transmission Model

In addition to direct transmission, the detection devices can also choose the relay transmission model. This is divided into two phases. First, the device transmits the task to the relay located in the middle of the train, namely the sensor-to-relay (S2R) phase. Then, it is forwarded from the relay to the BS, which is the relay-to-BS (R2B) phase. To enhance spectrum utilization, we adopt Co-time Co-frequency Full Duplex (CCFD) [25]. This technology allows a relay to transmit and receive data simultaneously on the same channel. However, the received and transmitted signals on the same subcarrier will inevitably interfere with each other.
When the task is in the S2R phase, the onboard devices and the relay are relatively stationary, whereas the channel between the ground devices and the relay rapidly changes due to the Doppler effect. Thus, the S2R phase is the opposite of the direct communication model. The interference received by ground tasks needs to include ICI, while the onboard tasks do not require it. Furthermore, the signal received by the relay will be interfered with by the transmitted signal, and existing self-interference (SI) cancellation technologies can already eliminate the interference from the transmitted signal to a great extent.
Thus, during the S2R phase, the received SINR [26] of the i - th onboard device connected can be expressed as:
Γ i S 2 R = p i | h i S 2 R | 2 σ 2 + η p i R ,
where | h i S 2 R | 2 is the channel gain between the i - th onboard device and the relay, i { 1 , , M } , and η denotes the SI cancelation level [26].
However, at this time, the received SINR of the i - th ground device connected can be expressed as:
Γ i S 2 R = p i | h i S 2 R | 2 ϖ p i | h i S 2 R | 2 + σ 2 + η p i ,
where i { M + 1 , , M + N } .
Thus, during the S2R phase, the transmission rate of the i - th task can be expressed as:
R i S 2 R = c = 1 C λ i , c W   log 2 ( 1 + Γ i S 2 R ) .
When the task is in the R2B phase, the relay moves at high speed relative to the BS, which can result in ICI caused by the Doppler effect. Thus, the SINR during the R2B phase can be expressed as:
Γ i R 2 B = p i R | h R 2 B | 2 ϖ p i R | h R 2 B | 2 + σ 2 ,
where p i R is the transmission power allocated to the i - th task and | h R 2 B | 2 is the channel gain between the relay and the BS. During the R2B phase, the transmission rate of the i - th task can be expressed as:
R i R 2 B = c = 1 C λ i , c W   log 2 ( 1 + Γ i R 2 B ) .
Since different subcarriers are allocated for the S2R and R2B transmission phases in the OFDM system, the relay can achieve full-duplex communication, which means it can receive and transmit at the same time. Thus, the transmission rate of the relay transmission model is determined by the slower phase and can be represented as:
R i R = min { R i S 2 R , R i R 2 B } .

2.4. NOMA-Based Transmission Model

When a task adopts the relay transmission model, the transmission power allocated by the relay often exceeds the transmission power of the device itself. Moreover, within a time slot, since the starting points of the tasks are all within the length range of the train and the transmission between the relay and the BS is often line-of-sight (LOS), the channel gain between the relay and the BS is always greater than that between the sensing device and the BS. As a result, the signal strength of the signal transmitted via the relay is higher than that of the signal from direct communication. Therefore, the NOMA-based transmission model [27] is viable. In this model, NOMA subcarriers can superimpose a group of signals with varying signal strengths. Signals in this group will interfere with one another. Successive Interference Cancellation (SIC) [28] technology can separate the signals in a group. SIC technology initially decodes the signal with the higher signal strength by treating the lower signal strength signal as interference. Then, it subtracts the decoded signal with high signal strength from the original signal. As a result, the remaining signal will be free of any additional interference.
After the relay transmission uses NOMA technology, the tasks uploaded in the R2B phase are considered signals with high signal strength. Thus, the interference they encounter needs to include the signal strength of the co-subcarrier signals. The transmission rate of the i - th task using NOMA can be expressed as:
R i R 2 B - N O M A = c = 1 C λ i , c N O M A W   log 2 ( 1 + p i R | h R 2 B | 2 ϖ p i R | h R 2 B | 2 + σ 2 + g = 1 A λ g , c p g | h g B | 2 ) ,
where g = 1 A λ g , c p g | h g B | 2 is the signal strength of the co-subcarrier signals. λ i , c N O M A denotes the variable that indicates whether a subcarrier is allocated to the R2B-NOMA channel. It should be noted that the subcarriers allocated to the R2B-NOMA channel can overlap with any subcarriers allocated to direct transmission channels. However, these subcarriers must not overlap with those allocated to the relay transmission channel because their signal strengths are similar.
NOMA technology can increase the relay’s transmission rate on top of the existing OFDM system without altering the other two channels. Therefore, the total transmission rate when tasks choose the relay transmission model can be expressed as:
R i R = min { R i S 2 R , R i R 2 B + R i R 2 B - N O M A } .

2.5. Transmission Delay

For tasks using the direct transmission model, the transmission delay can be expressed as:
T i u B = l i R i B .
For tasks using the relay transmission model, the transmission delay can be expressed as:
T i u R = l i R i R .
For the i - th task, which can choose either of the two transmission models, the transmission delay can be expressed as:
T i u = β i T i u B + ( 1 β i ) T i u R ,
where β i is the model selection variable. When β i is 1, the task will use the direct transmission model. When β i is 0, the task will use the relay transmission model.

2.6. Computing Delay

We assume that the MEC servers communicate with each other via fiber optics, which allows us to neglect the delay. This means that a task can be processed by any server. However, considering the large volume of tasks generated by multiple trains within the entire railway range, deploying multiple servers near every base station along the entire railway is very expensive and wasteful. Thus, we consider the scenario where there is a limited number of MEC servers, meaning that not all tasks uploaded can be processed immediately. It is presumed that each server is limited to processing a single task concurrently. Additionally, the tasks within a time slot significantly outnumber the servers available to handle them.
In this system, there are K MEC servers available in a time slot, denoted by K = { 1 , 2 , , K } . We use a tuple f k , q k max to describe the performance of an MEC server, where f k denotes the computational capability of the k - th server and q k max denotes the memory capacity of the server. Let z i , k { 0 , 1 } denotes the variable of server allocation, where k K . When z i , k is 1, the i - th task is processed on the k - th server. Otherwise, z i , k is 0. For the task processing order on the same server, we adopt the First-Come First-Served (FCFS) principle. The system can control the allocation of communication resources to allow tasks with the smaller maximum allowable delay to reach the server first, thereby prioritizing their processing. Let o i , j { 0 , 1 } denote the arrival order of the i - th task, where j J , J = { 1 , 2 , , J } , J = I . o i , j = 1 indicates that the i - th task is the j - th to arrive. The matrix composed of z i , k is denoted as Z I K . The matrix composed of o i , j is denoted as O I J . Let matrix Ψ K J be the result of the cross product of matrices Z I K T , Φ I I , O I J , denoted as Ψ K J = Z I K T × Φ I I × O I J , where Φ I I = d i a g ( 1 , 2 , , I ) , ψ k , j = i denote that on the k - th server, the i - th task is the j - th in the queue to be processed.
Thus, the total delay of the i - th task can be expressed as:
T ψ k , j = { T ψ k , j u + T ψ k , j c , j = 1 max { T ψ k , j u , T ψ k , j 1 } + T ψ k , j c , j 2 ,
where T ψ k , j 1 denotes the total delay from the start of a time slot to the completion of the task at the j - th position in the queue on the k - th server, T ψ k , j c denotes the computing delay that a server computing task ψ k , j costs, T ψ k , j c = l ψ k . j μ ψ k , j / f k . Specifically, both T 0 u and T 0 c are equal to 0.
In addition, the maximum queue length of tasks being processed on the k - th server can be expressed as:
q k = max j { g = 1 j ( l ψ k , g + L ψ k , g ) g = 1 j ( l ψ k , g + L ψ k , g ) } ,
where g = 1 j ( l ψ k , g + L ψ k , g ) denotes the total length of tasks on the k - th server after the j - th task arrives and g = 1 j ( l ψ k , g + L ψ k , g ) denotes the length of tasks that have been processed so far.

2.7. Problem Formulation

We aim to minimize the average delay ratio through the joint optimization of communication and computational resources. The delay ratio refers to the proportion of the actual delay to the maximum allowable delay. Thus, we take the computation offloading strategy as the decision variable, with communication, computation, and memory resources as constraints, aiming to minimize the average delay ratio as the optimization objective. The problem formulation can be expressed as follows:
min p , λ , β , z 1 I i = 1 I T i T i max s . t . C 1 : i = 1 I ( 1 β i ) p i R p max C 2 : i = 1 I λ i , c 1 C 3 : c = 1 C ( λ i , c N O M A i = 1 I ( 1 β i ) λ i , c ) = 0 C 4 : k = 1 K z i , k = 1 C 5 : T i T i max C 6 : q k q k max ,
where constraint C1 denotes that the total sum of transmit power allocated to tasks using the relay transmission model cannot exceed the relay’s maximum transmission power. Constraint C2 denotes that tasks using OFDM cannot share a subcarrier. Constraint C3 denotes that subcarriers allocated to the NOMA system cannot overlap with those allocated to the R2B channel. Constraint C4 denotes that a task can only be allocated to one server and must be allocated to a server. Constraint C5 denotes that the total delay of a task cannot exceed the maximum tolerable delay. Constraint C6 denotes that the queue of tasks being processed and awaiting processing on the server cannot exceed the memory capacity of the server.

3. DDPG-Based Resource Allocation Algorithm

3.1. Subcarrier Allocation Algorithm

The performance of NOMA systems is affected by the difference in signal strengths between relayed and direct communications. A larger disparity in signal strength can yield better performance gains. Conversely, a minimal difference in signal strength can result in poor system performance and may even preclude the separation of the two signals. Therefore, we propose a low-complexity subcarrier matching strategy. This algorithm aims to reduce the likelihood of a small intensity gap between two signals, preventing the assignment of a relay communication signal with a lower intensity to a subcarrier of a direct communication signal with a higher intensity, which would make it impossible to decode and separate the two. While this algorithm may not be optimal, it allows NOMA technology to be stably compatible with the entire system. Employing this approach, the complex issue of allocating subcarriers in high dimensions can be simplified by distributing quantities of subcarriers. The step-by-step procedure for this method is outlined in the pseudocode as Algorithm 1.
Algorithm 1: Subcarrier Allocation Algorithm
Input:  δ i , δ i N O M A
Output:  λ a , c , λ a , c N O M A
1:
Randomly allocate subcarriers to tasks based on the number of subcarriers δ i assigned to the tasks, and obtain λ a , c .
2:
Sort the subcarriers for tasks that use the direct allocation model in descending order by signal strength.
3:
Sort the tasks that use the relay communication model in descending order by the power allocated to the relay.
4:
Based on the number of subcarriers δ i N O M A allocated to the R2B-NOMA channel, allocate subcarriers with higher signal strength to tasks with higher relay power, and obtain λ a , c N O M A .
5:
Return λ a , c and λ a , c N O M A
Thus, the optimization problem of Equation (18) can be transformed into the following form:
min p , λ , β , z 1 I i = 1 I T i T i max s . t . C 1 : i = 1 I ( 1 β i ) p i R p max C 2 : i = 1 I δ i C C 3 : i = 1 I δ i N O M A i = 1 I ( 1 β i ) δ i C 4 C 6 ,
where constraint C2 denotes that the aggregate sum of subcarriers assigned to all tasks must not surpass the total quantity of subcarriers at disposal, and constraint C3 denotes that the sum of subcarriers allocated to the R2B-NOMA channel cannot surpass the sum of subcarriers allocated to tasks using direct transmission models. All other constraints remain unchanged.

3.2. MDP Model

Since the problems of relay power allocation and server selection are non-convex, securing the optimal solution through conventional optimization techniques proves to be challenging. Thus, an MDP is introduced to effectively learn policies in complex environments. The aforementioned optimization problem can be formulated as an MDP, denoted as M = ( S , A , P , r , γ ) , where components align with the state space, action space, state transition matrix, reward function, and discount factor of the problem, in that order.
Observation: in time slot t , the state space includes the basic information of the detection task and the data transmission rates between the sensors and BS, as well as between the sensors and the relay. In time slot t , the state space S t S can be expressed as follows:
S t = { l t , μ t , T t max , L t , p t , | h t B | 2 , | h t S 2 R | 2 } ,
where l t is the data size in time slot t , μ t is the computation workload in time slot t , T t max is the maximum allowable delay in time slot t , L t represents the data size of the neural network model to be loaded and the intermediate data generated in time slot t , p t is the power of the devices, | h t B | 2 is the data transmission rates between the sensors and BS, and | h t S 2 R | 2 is the data transmission rates between the sensors and the relay.
Action: in time slot t , the action space includes communication and computation resource allocation strategies. In time slot t , the action space A t A can be expressed as follows:
A t = { β t , δ t , δ t N O M A , p t R , z t } ,
where β t { 0 , 1 } is the selected model, δ t is the number of subcarriers allocated to the tasks, δ t N O M A is the number of subcarriers allocated to the R2B-NOMA channel, p t R is the power allocated to the task by the relay, and z t { 1 , 2 , , k } is the server selection strategy.
Reward: the objective of the joint optimization problem of computation offloading and resource allocation is to minimize the average delay ratio. However, the aim of training with reinforcement learning is to optimize the total accumulated reward over time. Concurrently, tasks that surpass the threshold of the maximum allowable delay will trigger a predefined penalty. Thus, the reward for the i - th task in time slot t can be expressed as follows:
r i , t = { T i T i max , T i T i max ε , T i > T i max   o r   Q k > Q k max ( z i , k = 1 ) ,
where ε is the absolute value of the penalty.
Thus, the total reward in time slot t can be expressed as follows:
r t = 1 I i = 1 I r i , t .

3.3. Improved Deep Deterministic Policy Gradient Algorithm

Given that the action space comprises both continuous and discrete variables, the DDPG algorithm demonstrates commendable performance in addressing continuous action control problems.
The DDPG algorithm merges the aspects of actor–critic algorithms with deep learning techniques to enable effective policy learning in high-dimensional, continuous action spaces. The core of DDPG lies in learning a deterministic policy, which specifies the action to be taken in each state, along with a Q-function that estimates the expected returns. The architecture of the DDPG algorithm is shown in Figure 2.
The Q-function, which estimates the expected reward of taking an action A t in state S t , following a policy π , is defined as Q ( s t , a t | θ Q ) , where θ Q are the parameters of the critic network. The expected return from the Q-function is computed as:
Q ( s t , a t ) = E [ R t ( s t , a t ) ] ,
where the return R t is the sum of the immediate reward r and the discounted future rewards, and is given by:
R t ( s t , a t ) = r ( s t , a t ) + γ Q [ s t + 1 , π ( s t + 1 | θ π ) ] ,
where γ is the discount factor that balances immediate and future rewards.
The objective of the critic network is to reduce the mean squared error, focusing on bringing the forecasted Q-values in line with the target Q-values as described by the following function:
L ( θ Q ) = 1 N t [ y t Q ( s t , a t | θ Q ) ] 2 ,
where y is the target value defined as:
y t = r t + γ Q [ s t + 1 , π ( s t + 1 | θ π ) | θ Q ] ,
where Q and π represent the target critic and target actor networks, respectively.
The policy gradient for updating the actor network is derived from the critic network and is calculated using the sampled gradients of the Q-function with respect to the actions [14]:
θ π J 1 N t π Q [ s t , π ( s t ) | θ Q ] θ π π ( s t | θ π ) .
Finally, the actor network is updated through gradient ascent using the policy gradient, adjusting the parameters θ π in the direction that maximizes the Q-function:
θ π θ π + α θ π J ,
where α is the learning rate.
The pseudocode for the training process of the DDPG algorithm is presented in Algorithm 2.
Algorithm 2: The Training Process of DDPG Algorithm
1:
Use random network weights θ π and θ Q initializing actor network and critic network, respectively
2:
Copy the same weights θ π θ π and θ Q θ Q to initialize the target networks, respectively
3:
Initialize replay buffer B
4:
for episode = 1 to M do
5:
  Initialize a random process N for action exploration
6:
  Obtain initial observation state s 1
7:
  for time-step = 1 to T do
8:
   Select action a t = π ( s t | θ π ) + N t using the current policy affected by noise
9:
   Perform action a t and observe reward r t and new state s t + 1
10:
   Store transition ( s t , a t , r t , s t + 1 ) in B
11:
   Sample a random minibatch of N transitions ( s t , a t , r t , s t + 1 ) from B
12:
   Set y t = r t + γ Q [ s t + 1 , π ( s t + 1 | θ π ) | θ Q ] for non-terminal states
13:
   Update critic by minimizing the loss: L ( θ Q ) = 1 N t [ y t Q ( s t , a t | θ Q ) ] 2
14:
   Update the actor policy using the sampled policy gradient:
15:
     θ π J 1 N t π Q [ s t , π ( s t ) | θ Q ] θ π π ( s t | θ π )
16:
   Update the target networks:
17:
     θ π τ θ π + ( 1 τ ) θ π
18:
     θ Q τ θ Q + ( 1 τ ) θ Q
19:
  end for each time-step
20:
end for each episode
However, the DDPG algorithm does not perform well with certain discrete variables, especially those that represent the selection of indices. Therefore, we modified the architecture of the actor network, as shown in Figure 3. Actions β t and Z t in the action space A t are exactly such discrete variables. Although for β t , which is a binary variable, the tanh function can effectively solve the binary classification problem. As for the variable Z t , if the number of servers is more than two, this becomes a multi-class variable, and common activation functions struggle to address this issue. We process z t as follows:
We replace z t in the action space with Z t , where Z t is an I K matrix and z i , k represents the probability of choosing the k - th server for the i - th task. We select the server number k with the highest probability as the choice strategy for the i - th task, which can be expressed as z i = arg max k K z i , k .
The optimization problem represented by Equation (17) can be effectively addressed by improving the DDPG algorithm through the aforementioned approach for discrete variables, in conjunction with the channel allocation algorithm presented in Section 4.1.

4. Simulation Results and Discussion

In this section, we use simulation experiments to evaluate the system model proposed in Section 3 and the algorithm presented in Section 4. The simulation results are obtained using Python 3.7 and Tensorflow 2.11 on a machine powered by an Intel Core i5-13400F CPU with an NVIDIA GeForce RTX 4060Ti GPU.

4.1. Simulation Setting

In real-world train detection scenarios, the generation of detection tasks is cyclical and patterned. To demonstrate the effectiveness of the system model and algorithms, we set the task generation locations to be random yet confined within the length range of the train. This is because ground detection equipment only generates tasks when the train passes by. Furthermore, the parameters describing the tasks are also random. Specifically, the task data length l i follows a uniform distribution of [1, 4] Mbits, the computation workload μ i follows a uniform distribution of [800, 1000] cycles/bit, the maximum allowable delay T i max follows a uniform distribution of [1, 4] seconds, the loaded data size L i follows a uniform distribution of [15, 25] Mbits, and the power of the devices p i follows a uniform distribution of [0.1, 0.5] W [29]. For the entire railway system, the allocation of subcarrier numbers to each train, the number of available servers, and other parameter settings could be considered a queuing theory problem for optimization allocation. However, this is beyond the scope of this paper. Therefore, we assume that these parameters are pre-allocated and fixed. Specifically, the number of servers k is set to 3, with computational capability f k of 7 GHz and maximum memory capacity Q k max of 128 Mbits. In addition, for the algorithm proposed in Section 4, the absolute value of the penalty ε is set to 5. Other determined parameters are provided in Table 1.
We compare the proposed improved DDPG algorithms with standard DDPG as a benchmark algorithm, which was used in [31]. Then, we compare the system model proposed in Section 2 with three benchmark schemes as follows:
Full Direct Communication (FDC): As in the task offloading scheme for HSR described in [5], all detection tasks will be directly offloaded to the edge servers in this scheme.
Full Relay Communication (FRC): In this scheme, all detection tasks choose the relay communication model.
OFDM: In this scheme, no tasks will employ NOMA technology for transmission.

4.2. Result Analysis

In reinforcement learning algorithms, batch size is a critical hyperparameter. Smaller batch sizes can enhance generalization and facilitate the escape from local optima but may increase variance, affecting learning stability. Larger batch sizes provide more accurate gradient estimates, though they may lead to local optima and increased memory requirements. Figure 4 shows the influence of batch size on convergence. The calculation method of the reward is shown in Equation (23). The DDPG algorithm obtains rewards through interaction with the environment and learns to achieve higher rewards. Higher rewards also imply a lower average delay ratio. Despite the initial rapid increase and subsequent stabilization of the reward curves across four batch sizes, it indicates that our improved DDPG algorithm can learn a stable computation offloading policy. Notably, a batch size of 64 yields better rewards. Therefore, we set the batch size to 64 in subsequent simulations.
Additionally, the learning rate is a crucial hyperparameter. Figure 5 shows the impact of different learning rates on training outcomes. It can be observed from the figure that the convergence of the improved DDPG algorithm is highly sensitive to the learning rate. When the learning rate is too low, the training process converges to a local optimum. This is also the reason for the reward curve’s pattern of the initial increase followed by a decline, ultimately making it challenging to achieve satisfactory training outcomes. Conversely, when the learning rate is too high, the parameters fluctuate too violently during training, making it difficult to converge to a desirable solution and resulting in significant volatility. However, it is evident that setting the learning rate to 0.00001 leads to stable convergence and superior performance. Therefore, in subsequent simulations, we will set the learning rate to 0.00001.
Figure 6 shows the average delay ratio and success rate under different absolute values of penalty ε when the number of devices is 20 (high load). In Figure 6a, the average delay ratio increases with the increase in ε . However, when ε 4 , the average delay ratio is not sensitive to changes in ε . Within a 300% variation range of ε , the average delay ratio only fluctuates by 2.27%. Due to the randomness of the simulation environment, these fluctuations are acceptable. In Figure 6b, the success rate also shows an increasing trend as ε increases. When ε 4 , the success rate is also not sensitive to changes in ε . Within a 300% variation range of ε , the average delay ratio only fluctuates by 0.71%. By comparing these two figures, we can observe that when ε 3 , our proposed algorithm sacrifices the success rate to achieve a lower average delay ratio, as the penalty strength is not sufficient. Although a larger ε reduces algorithm errors and leads to a higher success rate, when ε 4 , disregarding the bias in the training effect, we can consider the training effect of the algorithm to be stable and not sensitive to changes in ε . However, when ε is too large, an increase in the number of tasks may cause the Q value to exceed the floating-point limit. Thus, setting ε as 5 is reasonable.
In Figure 7, we demonstrate the performance comparison of our proposed improved DDPG algorithm against the standard DDPG algorithms mentioned in Section 4.1 under both low- and high-load conditions by varying the number of onboard and ground sensors. In Figure 7a, the numbers of both types of sensors are set to 5, whereas in Figure 7b, they are set to 10. It was observed that under low-load conditions, all algorithms show good performance, with the improved DDPG slightly outperforming the others. However, under high-load conditions, the differences between algorithms became more pronounced. Compared to the standard DDPG, our improved DDPG algorithm is able to find more optimal offloading strategies. This indicates its efficacy in handling hybrid action space problems and higher-dimensional action spaces.
Figure 8 illustrates the average delay ratio and completion rate for different schemes. In the HSR scenario, a lower delay means anomalies can be detected more quickly, allowing for more time to respond; a higher success rate ensures that no details are missed, enhancing the completeness and meticulousness of the detection. Overall, these two metrics reflect the contribution of different schemes to safety.
Figure 8a demonstrates the trend of the average delay ratio obtained by different schemes as the number of devices varies. It should be noted that the quantity of the two types of devices is identical here. It is clear that the average delay ratio for each algorithm increases with the number of devices. When compared to the FDC, OFDM, and FRC schemes mentioned in Section 4.1, our scheme proposed in Section 2 consistently achieves lower delay ratios, reducing it by 4.69%, 6.56%, and 11.84%, respectively. This is attributed to the different channel conditions and interference that ground and onboard sensors experience under various transmission modes. Comparing the FDT and FDR schemes, we can demonstrate that our proposed approach better adapts to detection-oriented scenarios in railway IoT. The combination of these two transmission models allows for a more suitable transmission model for each detection task. Furthermore, the comparison with the OFDM scheme demonstrates that the introduction of NOMA technology indeed increases channel capacity, resulting in higher overall transmission rates and lower delay ratios. In summary, our proposed computation offloading model consistently delivers superior performance across different task quantities.
Figure 8b demonstrates the success rate trends of different schemes as the number of devices varies in the simulation. With an increase in device quantity, the success rate of each scheme generally decreases. Compared with the FDC, OFDM, and FRC schemes mentioned in Section 4.1, the scheme we proposed in Section 2 consistently achieves a higher completion rate. Specifically, in high-load environments, the failure rate is reduced by 89.6%, 64.9%, and 96.7%, respectively. This also indicates that our proposed scheme can better adapt to detection tasks in railway IoT. Furthermore, comparing the FDC and OFDM schemes in Figure 7, it is evident that FDC shows a lower delay ratio, while OFDM achieves higher completion rates. This observation reinforces the notion that a faster transmission rate is not always preferable. The server queue is closely linked to the task arrival time, influenced by the transmission rate. Excessive speed may result in server queue backlogs that exceed memory capacity. Thus, our enhanced DDPG algorithm effectively allocates communication and computational resources.

5. Conclusions

This study explores the dynamic resource management problem in the scenario of railway IoT systems with integrated sensing, storage, and computing, specifically focusing on the computation offloading for detection tasks. By optimizing subcarrier allocation, transmission mode selection, relay power allocation, and server selection, the objective is to minimize the average delay ratio, which is the ratio between the delay and the maximum tolerable delay. To address this problem, a low-complexity subcarrier allocation algorithm is proposed to reduce the action space dimension. Furthermore, an improved DDPG algorithm is proposed, which represents discrete variables using selection probabilities to handle the hybrid action space problem. Through simulation experiments, the performance of the improved DDPG algorithm is evaluated against existing benchmark schemes and algorithms, some of which referenced existing literature. The results indicate that the MEC system proposed in this paper can achieve a lower average delay ratio and a higher completion rate, demonstrating significant advantages over other schemes and algorithms, especially under high-load conditions. This also means that HSR will have a more secure guarantee. The findings also highlight the adaptability of the proposed MEC system model for detection tasks in HSR environments.
Through simulation, our proposed MEC system demonstrates lower average latency ratios and higher completion rates compared to other literature. In addition, due to its excellent adaptability, our proposed MEC system could be compatible with various types of rail transportation. However, the ultimate goal is to deploy it in a real HSR environment. However, the deployment of edge servers and online monitoring equipment is a significant undertaking. Future work will focus on the practical deployment of MEC, particularly in dealing with real-world network conditions. We will further collect data and conduct performance tests on the proposed MEC system to analyze its potential applications in various rail transit scenarios. Future work will also include research on potential issues such as energy consumption and security.

Author Contributions

Research, Z.X.; conceptualization, Z.X. and Y.W.; algorithm design, Z.X. and Q.G.; simulation analysis, Q.G. and Z.X.; verification, J.Y. and Y.W.; writing—original draft preparation, Q.G. and Z.X.; writing—review and editing, J.Y. and Y.W.; supervision, J.Y. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All data underlying the results are available as part of the article and no additional source data are required.

Acknowledgments

The authors would like to thank the editors and the reviewers for their valuable time and constructive comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, J. Maintenance Scheduling at High-Speed Train Depots: An Optimization Approach. Reliab. Eng. Syst. Saf. 2024, 243, 109809. [Google Scholar] [CrossRef]
  2. Yang, N.; Chen, M. Design and Application of Big Data Technology Management for the Analysis System of High Speed Railway Operation Safety Rules. In Proceedings of the 2023 IEEE International Conference on Integrated Circuits and Communication Systems (ICICACS), Raichur, India, 24–25 February 2023; pp. 1–6. [Google Scholar]
  3. Sobhy, H.; Zohny, H.N.; Elhabiby, M. Railway Inspection Using Non-Contact Non-Destructive Techniques. Int. J. Eng. Appl. Sci. (IJEAS) 2020, 7. [Google Scholar] [CrossRef]
  4. Singh, P.; Elmi, Z.; Krishna Meriga, V.; Pasha, J.; Dulebenets, M.A. Internet of Things for Sustainable Railway Transportation: Past, Present, and Future. Clean. Logist. Supply Chain 2022, 4, 100065. [Google Scholar] [CrossRef]
  5. Xu, J.; Ai, B.; Chen, L.; Cui, Y.; Wang, N. Deep Reinforcement Learning for Computation and Communication Resource Allocation in Multiaccess MEC Assisted Railway IoT Networks. IEEE Trans. Intell. Transp. Syst. 2022, 23, 23797–23808. [Google Scholar] [CrossRef]
  6. Wang, Y.; Yang, S.; Ren, X.; Zhao, P.; Zhao, C.; Yang, X. IndustEdge: A Time-Sensitive Networking Enabled Edge-Cloud Collaborative Intelligent Platform for Smart Industry. IEEE Trans. Ind. Inform. 2022, 18, 2386–2398. [Google Scholar] [CrossRef]
  7. Li, Y.; Zhu, L.; Wu, J.; Wang, H.; Yu, F.R. Computation Resource Optimization for Large-Scale Intelligent Urban Rail Transit: A Mean-Field Game Approach. IEEE Trans. Veh. Technol. 2023, 72, 9868–9879. [Google Scholar] [CrossRef]
  8. Xu, J.; Wei, Z.; Yuan, X.; Qiao, Y.; Lyu, Z.; Han, J. Throughput Maximization for Result Multicasting by Admitting Delay-Aware Tasks in MEC Networks for High-Speed Railways. IEEE Trans. Veh. Technol. 2024, 73, 8765–8781. [Google Scholar] [CrossRef]
  9. Khan, K.; Pasricha, S.; Kim, R.G. A Survey of Resource Management for Processing-In-Memory and Near-Memory Processing Architectures. J. Low Power Electron. Appl. 2020, 10, 30. [Google Scholar] [CrossRef]
  10. Zhao, C.; Cai, Y.; Liu, A.; Zhao, M.; Hanzo, L. Mobile Edge Computing Meets mmWave Communications: Joint Beamforming and Resource Allocation for System Delay Minimization. IEEE Trans. Wirel. Commun. 2020, 19, 2382–2396. [Google Scholar] [CrossRef]
  11. Jiang, K.; Zhou, H.; Li, D.; Liu, X.; Xu, S. A Q-Learning Based Method for Energy-Efficient Computation Offloading in Mobile Edge Computing. In Proceedings of the 2020 29th International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 3–6 August 2020; pp. 1–7. [Google Scholar]
  12. Chen, X.; Liu, G. Energy-Efficient Task Offloading and Resource Allocation via Deep Reinforcement Learning for Augmented Reality in Mobile Edge Networks. IEEE Internet Things J. 2021, 8, 10843–10856. [Google Scholar] [CrossRef]
  13. Ale, L.; King, S.A.; Zhang, N.; Sattar, A.R.; Skandaraniyam, J. D3PG: Dirichlet DDPG for Task Partitioning and Offloading With Constrained Hybrid Action Space in Mobile-Edge Computing. IEEE Internet Things J. 2022, 9, 19260–19272. [Google Scholar] [CrossRef]
  14. Deng, X.; Yin, J.; Guan, P.; Xiong, N.N.; Zhang, L.; Mumtaz, S. Intelligent Delay-Aware Partial Computing Task Offloading for Multiuser Industrial Internet of Things Through Edge Computing. IEEE Internet Things J. 2023, 10, 2954–2966. [Google Scholar] [CrossRef]
  15. Li, L.; Niu, Y.; Mao, S.; Ai, B.; Zhong, Z.; Wang, N.; Chen, Y. Resource Allocation and Computation Offloading in a Millimeter-Wave Train-Ground Network. IEEE Trans. Veh. Technol. 2022, 71, 10615–10630. [Google Scholar] [CrossRef]
  16. Tian, L.; Li, M.; Si, P.; Yang, R.; Sun, Y.; Wang, Z. Design and Optimization in MEC-Based Intelligent Rail System by Integration of Distributed Multi-Hop Communication and Blockchain. Math. Probl. Eng. 2023, 2023, 8858263. [Google Scholar] [CrossRef]
  17. Xue, Q.; Wei, R.; Li, Z.; Liu, Y.; Xu, Y.; Chen, Q. Beamforming Design for Cooperative Double-RIS Aided mmWave MU-MIMO Communications. IEEE Trans. Green Commun. Netw. 2024; Early Access. [Google Scholar] [CrossRef]
  18. Xu, C.; Zheng, G.; Zhao, X. Energy-Minimization Task Offloading and Resource Allocation for Mobile Edge Computing in NOMA Heterogeneous Networks. IEEE Trans. Veh. Technol. 2020, 69, 16001–16016. [Google Scholar] [CrossRef]
  19. Du, J.; Sun, Y.; Zhang, N.; Xiong, Z.; Sun, A.; Ding, Z. Cost-Effective Task Offloading in NOMA-Enabled Vehicular Mobile Edge Computing. IEEE Syst. J. 2023, 17, 928–939. [Google Scholar] [CrossRef]
  20. Du, R.; Wang, J.; Gao, Y. Computing Offloading and Resource Scheduling Based on DDPG in Ultra-Dense Edge Computing Networks. J Supercomput. 2024, 80, 10275–10300. [Google Scholar] [CrossRef]
  21. Zhou, X.; Tian, Y.; Wang, X. MEC-DA: Memory-Efficient Collaborative Domain Adaptation for Mobile Edge Devices. IEEE Trans. Mob. Comput. 2024, 23, 3923–3937. [Google Scholar] [CrossRef]
  22. Wang, J.; Ge, M.; Ding, B.; Xu, Q.; Chen, S.; Kang, Y. NicePIM: Design Space Exploration for Processing-In-Memory DNN Accelerators With 3-D Stacked-DRAM. IEEE Trans. Comput.-Aided Des. Integr. Circuits Syst. 2024, 43, 1456–1469. [Google Scholar] [CrossRef]
  23. Li, Y.; Cimini, L.J. Bounds on the Interchannel Interference of OFDM in Time-Varying Impairments. IEEE Trans. Commun. 2001, 49, 401–404. [Google Scholar] [CrossRef]
  24. Gao, M.; Ai, B.; Niu, Y.; Wu, W.; Yang, P.; Lyu, F.; Shen, X. Edge Caching and Content Delivery with Minimized Delay for Both High-Speed Train and Local Users. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar]
  25. Jiao, B.; Liu, S.; Lei, Y.; Ma, M. A Networking Solution on Uplink Channel of Co-Frequency and Co-Time System. China Commun. 2016, 13, 183–188. [Google Scholar] [CrossRef]
  26. Wang, Y.; Niu, Y.; Wu, H.; Ai, B.; Zhong, Z.; Wu, D.O.; Juhana, T. Relay Assisted Concurrent Scheduling to Overcome Blockage in Full-Duplex Millimeter Wave Small Cells. IEEE Access 2019, 7, 105755–105767. [Google Scholar] [CrossRef]
  27. Xue, J.; An, Y. Joint Task Offloading and Resource Allocation for Multi-Task Multi-Server NOMA-MEC Networks. IEEE Access 2021, 9, 16152–16163. [Google Scholar] [CrossRef]
  28. Wildemeersch, M.; Quek, T.Q.S.; Kountouris, M.; Rabbachin, A.; Slump, C.H. Successive Interference Cancellation in Heterogeneous Networks. IEEE Trans. Commun. 2014, 62, 4440–4453. [Google Scholar] [CrossRef]
  29. Zhong, A.; Li, Z.; Wu, D.; Tang, T.; Wang, R. Stochastic Peak Age of Information Guarantee for Cooperative Sensing in Internet of Everything. IEEE Internet Things J. 2023, 10, 15186–15196. [Google Scholar] [CrossRef]
  30. Maudet, S.; Andrieux, G.; Chevillon, R.; Diouris, J.-F. Practical Evaluation of Wi-Fi HaLow Performance. Internet Things 2023, 24, 100957. [Google Scholar] [CrossRef]
  31. Zhao, X.; Liu, M.; Li, M. Task Offloading Strategy and Scheduling Optimization for Internet of Vehicles Based on Deep Reinforcement Learning. Ad Hoc Netw. 2023, 147, 103193. [Google Scholar] [CrossRef]
Figure 1. MEC system for detection task in railway IoT.
Figure 1. MEC system for detection task in railway IoT.
Electronics 13 02982 g001
Figure 2. The architecture of the DDPG algorithm.
Figure 2. The architecture of the DDPG algorithm.
Electronics 13 02982 g002
Figure 3. The architecture of actor network.
Figure 3. The architecture of actor network.
Electronics 13 02982 g003
Figure 4. Influence of batch size on convergence.
Figure 4. Influence of batch size on convergence.
Electronics 13 02982 g004
Figure 5. Influence of learning rate on convergence.
Figure 5. Influence of learning rate on convergence.
Electronics 13 02982 g005
Figure 6. Simulation results of penalty. (a) Average delay ratio against absolute value of penalty; (b) success rate against absolute value of penalty.
Figure 6. Simulation results of penalty. (a) Average delay ratio against absolute value of penalty; (b) success rate against absolute value of penalty.
Electronics 13 02982 g006
Figure 7. Simulation results of different algorithms. (a) Comparison of algorithm (number of devices is 10); (b) comparison of algorithm (number of devices is 20).
Figure 7. Simulation results of different algorithms. (a) Comparison of algorithm (number of devices is 10); (b) comparison of algorithm (number of devices is 20).
Electronics 13 02982 g007
Figure 8. Simulation results of different schemes. (a) Average delay ratio against the number of devices; (b) the success rate against the number of devices.
Figure 8. Simulation results of different schemes. (a) Average delay ratio against the number of devices; (b) the success rate against the number of devices.
Electronics 13 02982 g008
Table 1. Simulation parameters.
Table 1. Simulation parameters.
ParameterValue
Number of subcarriers Λ 256
Subcarrier bandwidth W 15 KHz
Carrier frequency f c 3 GHz
Noise power σ 2 −130 dBm
Train speed v 360 km/s
Power of relay p max 20 W
signal symbol duration [30]: τ s 4 µs
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Q.; Xu, Z.; Yuan, J.; Wei, Y. Computation Offloading Strategy for Detection Task in Railway IoT with Integrated Sensing, Storage, and Computing. Electronics 2024, 13, 2982. https://doi.org/10.3390/electronics13152982

AMA Style

Guo Q, Xu Z, Yuan J, Wei Y. Computation Offloading Strategy for Detection Task in Railway IoT with Integrated Sensing, Storage, and Computing. Electronics. 2024; 13(15):2982. https://doi.org/10.3390/electronics13152982

Chicago/Turabian Style

Guo, Qichang, Zhanyue Xu, Jiabin Yuan, and Yifei Wei. 2024. "Computation Offloading Strategy for Detection Task in Railway IoT with Integrated Sensing, Storage, and Computing" Electronics 13, no. 15: 2982. https://doi.org/10.3390/electronics13152982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop