Next Article in Journal
Multimodal Ultrasound Radiomic Technology for Diagnosing Benign and Malignant Thyroid Nodules of Ti-Rads 4-5: A Multicenter Study
Previous Article in Journal
Simultaneous Measurement of Strain and Displacement for Railway Tunnel Lining Safety Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sequence Decision Transformer for Adaptive Traffic Signal Control

1
College of Automotive Engineering, Jilin University, Changchun 130025, China
2
Graduate School of Information and Science Technology, The University of Tokyo, Tokyo 113-8654, Japan
3
National Key Laboratory of Automotive Chassis Integration and Bionics, Jilin University, Changchun 130025, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6202; https://doi.org/10.3390/s24196202
Submission received: 12 August 2024 / Revised: 16 September 2024 / Accepted: 23 September 2024 / Published: 25 September 2024
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Urban traffic congestion poses significant economic and environmental challenges worldwide. To mitigate these issues, Adaptive Traffic Signal Control (ATSC) has emerged as a promising solution. Recent advancements in deep reinforcement learning (DRL) have further enhanced ATSC’s capabilities. This paper introduces a novel DRL-based ATSC approach named the Sequence Decision Transformer (SDT), employing DRL enhanced with attention mechanisms and leveraging the robust capabilities of sequence decision models, akin to those used in advanced natural language processing, adapted here to tackle the complexities of urban traffic management. Firstly, the ATSC problem is modeled as a Markov Decision Process (MDP), with the observation space, action space, and reward function carefully defined. Subsequently, we propose SDT, specifically tailored to solve the MDP problem. The SDT model uses a transformer-based architecture with an encoder and decoder in an actor–critic structure. The encoder processes observations and outputs, both encoded data for the decoder, and value estimates for parameter updates. The decoder, as the policy network, outputs the agent’s actions. Proximal Policy Optimization (PPO) is used to update the policy network based on historical data, enhancing decision-making in ATSC. This approach significantly reduces training times, effectively manages larger observation spaces, captures dynamic changes in traffic conditions more accurately, and enhances traffic throughput. Finally, the SDT model is trained and evaluated in synthetic scenarios by comparing the number of vehicles, average speed, and queue length against three baselines, including PPO, a DQN tailored for ATSC, and FRAP, a state-of-the-art ATSC algorithm. SDT shows improvements of 26.8%, 150%, and 21.7% over traditional ATSC algorithms, and 18%, 30%, and 15.6% over the FRAP. This research underscores the potential of integrating Large Language Models (LLMs) with DRL for traffic management, offering a promising solution to urban congestion.

1. Introduction

Urban traffic congestion has long been a global issue, negatively impacting both the economy and the environment. The energy used during instances of traffic congestion contributes to the emission of greenhouse gases such as carbon dioxide, exacerbating the greenhouse effect. In the United States, traffic congestion in just one year has been estimated to result in economic damages totaling USD 121 million, along with the generation of 25,396 tons of carbon dioxide [1]. Meanwhile, in China, a minimum of 24% extra travel time is necessary for commuting during peak hours in major cities like Beijing, Shijiazhuang, and Chongqing [2]. Curbing vehicle purchases is not a feasible solution, and expanding traffic infrastructure is costly. There is a pressing need to mitigate pollution emissions from traffic congestion by leveraging technological innovations to improve environmental conditions, while enhancing the efficiency traffic signal systems appears to be a more manageable approach [3].
Currently, many intersections operate on a fixed-time traffic signal system that sets the timing of traffic signals based on historical data rather than in response to real-time traffic needs. Conventional traffic signal control techniques typically use predefined rules derived from strategies based on expert experience. These methods lack the capability to dynamically adjust signal phases based on instantaneous traffic flow data. In the context, Adaptive Traffic Signal Control (ATSC) [4,5,6] stands out. ATSC can be used to optimize traffic flow across regional road networks and effectively reduces congestion [7] by dynamically altering signal phases in response to current traffic conditions. The concept of ATSC was introduced many years ago, and key systems such as the Split Cycle Offset Optimization Technique (SCOOT) [8] and Sydney Coordinated Adaptive Traffic System (SCATS) [9] are other notable examples.
With the explosive growth of Deep Learning (DL), an increasing array of DL-based ATSC methodologies are emerging. Sun et al. [10] employed Convolutional Neural Networks (CNNs) to analyze image data from traffic cameras, deriving a method for predicting short-term traffic flow and thereby enabling real-time signal optimization. Similarly, Kong et al. [11] utilized Recurrent Neural Networks (RNNs) to process historical traffic data, facilitating proactive signal adjustments. In addition, traffic signal control is a sequential decision-making process that can be formulated as a Markov Decision Process (MDP). Thus, it is possible to frame this issue as a problem that can be solved using reinforcement learning (RL) [12,13,14]. One of the earliest contributions of RL to ATSC is based on Q-learning [15], and there are some other works applying RL for ATSC in intersection environments [16,17,18], where the focus is on dynamically adjusting traffic signals based on real-time and historical data to enhance the overall efficiency of traffic management systems.
Recently, RL has been enhanced using deep neural networks, giving rise to a subfield known as deep reinforcement learning (DRL). DRL combines the advantages of powerful hierarchical feature extraction and nonlinear approximation abilities, with the interaction between the agent and the environment. This integration has a range of appealing attributes, generating significant academic interest for the application of DRL in ATSC [7,19,20]. Li et al. [21] employed stacked autoencoders in RL for ATSC to ascertain the Q-function, facilitating the efficient compression and storage of agent inputs. Wei et al. [22] contributed an advancement to the field by integrating reward and policy interpretation into a DQN-embedded ATSC system. Nishi et al. [23] modeled the impact of neighboring nodes using a static adjacency matrix within the framework of GCN. Additionally, Wu et al. [24] proposed using CNNs for the improved extraction of state information features relevant to network topology-related issues, specifically addressing the ATSC problem. Wang et al. [25] used a decentralized RL approach with a region-aware strategy, incorporating an actor–critic model and graph attention networks to optimize traffic signal control. Zheng et al. [26] introduced the FRAP model, which optimizes traffic signal control by leveraging phase competition and symmetry invariance. Furthermore, Zhang et al. [27] applied meta-reinforcement learning using model-agnostic meta-learning (MAML) and flow clustering to generalize traffic signal control across diverse environments with support from a WGAN-generated traffic flow dataset. DRL excels over traditional RL by efficiently handling high-dimensional, complex data and automatically extracting nonlinear features while additionally surpassing DL by integrating decision-making and control capabilities, thereby enabling continuous learning and an effective balance between exploration and exploitation in dynamic environments.
However, due to the rapidly changing traffic conditions and the high dimensionality of traffic information, current DRL-based ATSC algorithms struggle to effectively capture the relationships between traffic information across different time sequences and to adequately understand traffic scenarios. The primary deficiencies are manifested as follows. Firstly, their reasoning capabilities are insufficient, resulting in non-stationarity during the training process. Secondly, most ATSC algorithms require extensive data to learn the optimal policy, necessitating greater computational resources and time requirements [28]. In addition, traditional DRL agents are specialized in their training environment, limiting their generalization and transferability to new environments. These considerations restrict the application of DRL to simple traffic scenarios, hindering its applicability in complex real-world traffic situations.
To address these issues, a novel DRL-based ATSC approach named Sequence Decision Transformer (SDT) is proposed, in which DRL is formulated as a sequence decision model. Given that large language models (LLMs) like the GPT series and BERT have demonstrated outstanding performance in natural language processing, computer vision, and reinforcement learning [29], the robust understanding and reasoning capabilities of LLMs are applied to the ATSC problem to accommodate the high dynamics of traffic conditions and the high dimensionality of traffic information. This enables SDT to incrementally learn optimal policy within complex traffic flows. The algorithm employs an encoder–decoder structure, stores historical trajectories in the replay buffer, and utilizes Proximal Policy Optimization (PPO) to update parameters, effectively addressing challenges in ATSC such as the demand for large amounts of data and the low generalization performance.
The main contributions of this paper are summarized as follows:
(1) The ATSC problem is first converted into a DRL formulation using a Markov Decision Process (MDP), where the essential elements, such as the observation space, action space, and reward functions, are defined with the goal of improving traffic efficiency and reducing congestion. This article introduce SDT, an extension of a standard MDP, with an encoder–decoder structure. This framework serves as means for deriving a formulation from the ATSC problem.
(2) Secondly, the SDT is employed to solve the ATSC problem formulated as an MDP. SDT utilizes a transformer to process data in parallel, which can potentially reduce training time and accommodate larger observation and action spaces. The self-attention mechanism in SDT captures dynamic changes in the environment and enhances the representation of general features, addressing issues related to non-stationarity and improving model generalization in new environments. Additionally, PPO is introduced, using a truncated probability ratio to limit the policy update step, thereby preventing large policy changes during training and increasing the stability of the training process.
(3) Thirdly, extensive experiments were carried out across various traffic scenarios. The experimental results demonstrate that the presented ATSC method based on SDT consistently excels at alleviating traffic congestion and enhancing the efficiency of the traffic system. The SDT model, compared to PPO, a DQN for ATSC, and FRAP, shows improvements of 26.8%, 150%, and 21.7% over traditional ATSC algorithms, and 18%, 30%, and 15.6% over the state-of-the-art (SOTA) under the most complex conditions.
The rest of this paper is organized as follows: The ATSC problem and the ATSC framework based on SDT are introduced and defined in Section 2. In Section 3, the ATSC problem is formulated as a Markov decision process. Section 4 details the algorithm, including its structure and update policy. The experiments are presented in Section 5, followed by the conclusions in Section 6. Appendix A details the explanation of notations appeared in this article.

2. Problem Statement

2.1. Problem Definition

The intersection scenario controlled by an ATSC agent is illustrated in Figure 1. This represents a classic multiple-lane cross intersection, with multiple incoming lanes and outgoing lanes in each direction. When vehicles approach the intersection, they are presented with three options: go straight, turn left, or turn right. All vehicles’ passing choices directly affect future traffic conditions. The ATSC agent focuses on traffic flow information that affects traffic efficiency to make and control the traffic signal, specifically the traffic flow information from the incoming lanes.
The ATSC agent periodically obtains information for each lane i ( i N l a n e and i N + :
s i = n i v i
where N l a n e is the maximum number of lanes, n i R is the number of vehicles belonging to lanes i at step t N , and v i R is the average velocity of vehicles in lane i at step t.
ATSC methods are typically embedded within traffic signal control devices, such as traffic lights, in transportation systems. In regular time steps, the ATSC agent, embedded with the SDT algorithm, will determine the traffic signal phase p from A = { p h a s e 1 , , p h a s e N } based on real-time traffic data gathered from external equipment like sensors and cameras across all lanes. The selected phase controls the intersection, and A denotes the set of traffic signal phases. Furthermore, the ATSC methods embedded in the traffic system are adjustable, allowing relevant personnel to modify the adjustable parameters of the ATSC methods to achieve superior operational efficiency.

2.2. Methodological Framework

As shown in Figure 1, the entire framework of SDT is depicted. It utilizes a transformer neural network architecture that includes an encoder and a decoder, forming an actor–critic structure. The encoder is tasked with receiving observations, learning the representations of these observations, and inputting the encoded observations into the decoder. As part of its role as a value network, the encoder also outputs the value of the received observations, which is utilized for the subsequent update of its parameters. The decoder, working as the policy network, is responsible for receiving the encoded data from the dncoder and outputting the action that the agent is required to execute. In addition, PPO is responsible for updating the parameters of the policy network using data from historical trajectories, enabling the ATSC agent to make better decisions.
The framework based on SDT-based ATSC is comprised of two parts: Environment Interaction Collecting (EIC) and Policy Evaluation Optimization (PEO). The EIC is tasked with acquiring updated neural network parameters, then employing these parameters to collect data from a road intersection environment. This component employs MDP to formally express the process through which an ATSC agent collects traffic information, executes actions, and receives rewards within an intersection environment. This procedure subsequently produces discrete time-series trajectory data, encompassing observations, actions, and rewards. These data form the basis for optimizing the neural network. The PEO works in conjunction with the EIC, using the data gathered from the EIC to update the SDT neural network. This component features a PPO method that solves the MDP problem to update the policy for maximizing ATSC performance. It then aligns the updated parameters with the EIS for subsequent rounds of collecting and optimization in a repetitive cycle, continuing until the targeted traffic performance at the intersection is achieved.

3. Representation of ATSC Problem to Markov Decision Process

3.1. Markov Decision Process

The Markov decision process (MDP), widely recognized for formally expressing the process through which a agent traverse the environment, fundamentally falls under the category of a discrete-time decision-making architecture. An MDP is defined by a tuple O , A , P , R , T , where:
  • O is the observation space of the agent, where observations are the pieces of information that an agent receives o t O at each time step t N .
  • A represents the set of action space of the agent in the environment, where the agent performs a t A in the discrete time step t.
  • R : O × A R max , R max is a function that gives rewards; this function maps observation o to a numerical reward.
  • P : O × A × O R represents the transition probability function, which defines the dynamics of the environment.
  • T : O × A O is a function that describes the probability of transitioning from observation o t to observation o t + 1 , given a particular action a t . T ( o , a , o ) = P ( o t + 1 = o o t = o , a t = a ) .
Following the MDP model, the agent interacts with the environment in discrete time steps. At time step t, the agent produces an observation o t by interacting with the environment, and the agent performs an action a t . After the agent has taken the action, it receives a reward R o t | a t . The environment then transits to a new observation o t + 1 based on the probability function P o t + 1 | o t , a t . When the trajectory τ = o 0 , a 0 , o 1 , from an epoch is gathered, the policy π undergoes an update, subsequently enabling the agent to continue interacting with the environment using this newly updated policy. The goal of an MDP is to enable the agent to learn the optimal policy π * that maximizes the total accumulated reward over time through continuous policy updates.

3.2. Converting the ATSC Problem to MDP

In this section, the ATSC problem is transformed into an MDP problem by defining its basic elements, including the observation space, action space, and reward function set. Table 1 shows the observation and action space in this work.
Observation: The observation space is used to describe the observation information during the intersection between the agent and the environment. A reasonable observation space is crucial for efficient training of DRL algorithm. In the ATSC system, the capacity to accurately extract and reconstruct observation information from the complex and dynamic environment of intersections is significant, as it determines the ability to output appropriate actions accurately.
The observation space is defined as the sum of the observation of lanes in the current scenario, i.e., O = i = 1 N l a n e O i . Considering that the traffic conditions at intersections vary according to the dynamic and static traffic information of each lane, the observation subset is set as O i = n i , t i , q i , v i , where n i represents the total number of vehicles in lane i at time t, t i = 1 n i m = 1 n i t m , v m < 0.1 m / s 0 , v m 0.1 m / s represents the average waiting time of vehicles with a velocity of less than 0.1 m / s in lane i at time t; v m and t m are the velocity and the waiting time of vehicle m, respectively; q i = m = 1 n i 1 , v m < 0.1 m / s 0 , v m 0.1 m / s represents the queue length of lane i at time t, which can be denoted by the number of queued vehicles; and v i = 1 n i m = 1 n i v m v max represents the average velocity of vehicles in lane i at time t. The above observation space can accommodate complex and randomly changing traffic intersection scenarios, effectively characterizing the traffic conditions at the intersection.
Action: In this paper, the discrete set of traffic light phases is considered as the action space A = { phase 1 , , phase N } for the agent. Upon receiving an observation o t at time step t, the agent selects and executes an appropriate action p from { 0 , , N p h a 1 } , and N p h a indicates the quantity of phases. By executing p, the agent determines the roads that are allowed and disallowed to pass at time t + 1 and maintains the phase p h a s e j duration.
Reward: The reward function serves as quantitative feedback received by the agent from the environment after executing an action. In this study, the reward function guides the agent in continuously exploring and learning optimal strategies while maximizing rewards. The design of these functions is crucial for the convergence speed of DRL algorithms. If the reward function is overly optimized for a specific scenario, it may lead to overfitting, reducing the trained model’s ability to adapt to new scenarios. Generally, performance metrics should be designed based on the actual objectives intended to be achieved in the environment. To comprehensively describe the environment, the reward function focuses on holistically enhancing the traffic efficiency at intersections, ensuring that the reward system accurately reflects the desired outcomes in the environment, not just how the agent should behave. Such a design helps avoid overtraining in specific environments while enhancing the model’s adaptability and generality in new scenarios. Therefore, considering all aspects, the reward function comprises the following components:
(1) The total number of vehicles ( R n u m ): By focusing on the total number of vehicles, the reward function can directly impact and assess traffic flow at intersections. Reducing the total number of vehicles at intersections aids in alleviating traffic congestion and facilitating smoother traffic. This metric effectively reflects the intersection’s capacity and efficiency in managing traffic flow.
R n u m = i = 1 N l a n e n i
(2) The average speed of vehicles ( R v o l ): Average speed is one of the key metrics used to assess traffic efficiency. By increasing the average speed of vehicles, traffic congestion can be reduced and travel times shortened, thereby enhancing the overall efficiency of the intersection. Higher average speeds generally indicate good traffic conditions, where vehicles do not need to frequently stop or slow down.
R v o l = 1 N l a n e i = 1 N l a n e v i
(3) The average queue lengths of all vehicles ( R q u e ): This metric reflects the queuing situation of vehicles at the intersection. Shorter queue lengths mean that vehicles can pass through the intersection more quickly, reducing waiting and idling times, which helps to enhance the continuity of traffic flow and reduce congestion. Controlling queue length can effectively optimize the distribution of traffic flow and the scheduling of traffic signals at the intersection.
R q u e = 1 N l a n e i = 1 N l a n e q i
(4) The average waiting time of all vehicles ( R t i m e ): The average waiting time is an important metric for measuring the efficiency of an intersection. By reducing the average waiting time of vehicles at the intersection, the smoothness and efficiency of traffic flow can be significantly improved.
R t i m e = 1 N l a n e i = 1 N l a n e t i
In conclusion, the reward function is formulated as follows:
R = w 1 × R n u m + w 2 × R v o l + w 3 × R q u e + w 4 × R t i m e
Here, w 1 , w 2 , w 3 , w 4 represent the weights for the items of the reward function, respectively. The final goal of the proposed algorithm is to optimize a weighted sum of four objectives, which serves as a normalization method. This weighted summation approach was chosen because the scales of these four factors vary significantly in the initial episodes for each environment. Simply adding them as a reward would cause the agent to focus disproportionately on the factors with larger magnitudes, potentially leading to the algorithm’s failure to converge.
Weighting these four metrics ensures that each indicator has an appropriate proportion within the reward function, thereby overcoming issues of differing scales and uneven impact among the indicators. This design helps the algorithm balance the importance of different metrics, avoiding the overoptimization of a single metric while neglecting others that are equally important. Additionally, this weighted approach enhances the stability and convergence of the algorithm, enabling it to effectively learn and adapt in various traffic environments, ultimately achieving the goals of reducing congestion and improving traffic efficiency.

4. Sequence Decision Transformer

This section provides a detailed description of the SDT algorithm used to solve the MDP problem; an advanced framework using an encoder–decoder architecture. The encoder processes environmental observations and evaluates their values, while the decoder determines actions based on the policy updated through PPO to enhance decision-making accuracy and adaptability in dynamic traffic conditions.

4.1. Structure and Update

Figure 2 depicts the conceptual structure of SDT, which is mainly comprised of two components: the encoder and the decoder. The encoder is responsible for receiving observations o t from the environment, learning representations of observations o ^ t , and outputting the value V t of the received observations at time t. The decoder is in charge of outputting the action a t to be executed by the agent at time t. Consequently, after the agent performs a predetermined number of steps, a portion of the historical trajectory τ will be given to PPO for updating the policy π and the encoder–decoder networks.
The transformer model is the key to SDT, specifically the self-attention mechanism, which computes the response at a position in a sequence by attending to all positions and drawing global dependencies between the input and output. The attention function is written as follows:
A t t e n t i o n Q , K , V = s o f t max Q K T d k V
Q = S × W q , K = S × W k , V = S × W v
where S denotes the input sequence, and W q , W k , W v are the weight matrix. Q, K, and V correspond to the vector of queries, keys, and values that can be learned during training, and d k represents the dimension of Q and K. Self-attention refers to cases when Q , K , V share the same set of parameters.
Encoder: The function of the encoder, whose parameters are represented by ϕ , is as follows: Observation o t first passes through an embedding layer o E = E m b ( o t ) , which maps discrete input into vectors of continuous numbers in a lower-dimensional space to enhance computational efficiency. It then passes through the self-attention mechanism and residual connections o A = A t t e n t i o n ( Q o E , K o E , V o E ) + o E to prevent gradient vanishing and network degradation. Finally, the data pass through a multi-layer perception (MLP) to obtain the o ^ t = M L P ( o A ) , which is an input of the decoder. In addition, o ^ t is also utilized to calculate the value V ( o ^ t ) = M L P ( o ^ t ) , facilitating subsequent updates within the network. The role of the encoder is similar to that of a critic, as it aims to approximate the value function. Its objective is to minimize the empirical Bellman error:
L E n c o d e r ϕ = 1 T t = 0 T 1 R o t , a t + γ V ϕ ^ o ^ t + 1 V ϕ o ^ t 2
where ϕ ^ represents the target network’s parameters, which are non-differentiable and updated every few time steps.
Decoder: The embedded action a 0 (an arbitrary symbol indicating the start of decoding) passes into the decoder, whose parameters are denoted by θ . The decoder consists of many decoder blocks, like the encoder. Each block contains two attention mechanisms and one MLP. The first mechanism is responsible for receiving a 0 and generating a A 1 = A t t e n t i o n ( Q a E , K a E , V a E ) + a E , where a E is obtained by passing through an embedding layer that performs a dimensional transformation. The attention scores are calculated in the second attention mechanism a A 2 = A t t e n t i o n ( Q o ^ t , K a A 1 , V a A 1 ) + o ^ t . After passing through two MLP, the action representation a t + 1 = M L P ( M L P ( a A 2 ) + a A 2 ) is output. The second MLP outputs the probability distribution of the agent’s actions, namely the policy π θ ( a t + 1 | o ^ t , a t ) .
Then, PPO is utilized to update θ . PPO is a policy gradient method for reinforcement learning that balances implementation simplicity with sample efficiency. It addresses the limitations of previous strategies by allowing for multiple epochs of minibatch updates while avoiding excessively large policy updates that could destabilize training. This stability is achieved through a clipped objective function, which restricts the update step to remain within a small range around the old policy. The clipped objective function is as follows:
L C L I P ( θ ) = E t min r t ( θ ) A ^ t , c l i p ( r t ( θ ) , 1 ± ϵ ) A ^ t
Here, r t ( · ) is the policy ratio, and ϵ is a small positive number used to define the clipping range. A ^ t is an estimate of the advantage function, which is represented as follows:
A o , a = Q o , a V o
Q o , a = E o 1 : P , a 1 : π R γ | o 0 = o , a 0 = a
V o = E a 0 π , o 1 : P , a 1 : π R γ | o 0 = o
It measures the relative benefit of choosing a particular action a in an observation o over the average action under the current policy. This function highlights the excess expected return of an action compared to the average policy performance, thereby facilitating more informed and effective policy updates and decision-making in learning algorithms.
Based on these theorems, the following clipped PPO objective is minimized to train the decoder:
L D e c o d e r θ = 1 T t = 0 T 1 min r t θ A ^ t , c l i p r t θ , 1 ± ϵ A ^ t
r t θ = π θ a t + 1 | o ^ t , a t π θ o l d a t + 1 | o ^ t , a t

4.2. Algorithm Overview

This section introduces the ATSC algorithm based on SDT, as presented in Algorithm 1. The SDT algorithm initiates by initializing the encoder and decoder networks parameters ϕ and θ , alongside a replay buffer D to store transitions (line 1). The environment is then reset for a initial observation o 0 , a 0 , R 0 , etc., in the EIC (line 2). Within a loop that persists for a preset number of episodes, the algorithm, at each timestep t within an episode, calculates the value of V t , subsequently selecting the most suitable action a t . Following the execution of an action, the agent observes the outcome and receives a corresponding reward R t , which, along with the value and the new observation o t + 1 , is stored in the replay buffer (line 3–8). Upon the completion of each episode, the update of the encoder’s parameters and the decoder’ parameters is triggered. This update utilizes the computed returns derived from the accumulated data, specifically through the value function and PPO policy, as detailed in Equation (9) and Equation (14), respectively. This updating mechanism is systematically repeated at the end of each episode (line 9), ensuring continuous refinement and optimization of the network parameters based on the most recent episodic experiences. This iterative approach is integral to the model’s learning process, promoting enhanced performance and stability. This cyclical process aims to refine the agent’s policy for traffic signal timing, enhancing the management of intersection traffic flow.
Algorithm 1: Sequence Decision Transformer
1:
Initialization: Encoder network ϕ , Decoder network θ , and replay buffer D
2:
Reset environment and get initial observation o 0 , action a 0 , reward R 0 and current policy etc. in the EIC
3:
for episode = 1 to MAX_EPISODES do
4:
  for t = 0 to EPISODE_LENGTH do
5:
    Calculate value V t and select action a t from Encoder-Decoder
6:
    Obtain observation o t + 1 and reward R t after agent takes a t
7:
    Store transition ( o t , V t , o t + 1 , R t ) in D
8:
  end for
9:
  Compute return and update parameters ϕ and θ every episode by
     L E n c o d e r ϕ = 1 T t = 0 T 1 R o t , a t + γ V ϕ ^ o ^ t + 1 V ϕ o ^ t 2 and
     L D e c o d e r θ = 1 T t = 0 T 1 min ( r t θ A ^ t , c l i p r t θ , 1 ± ϵ A ^ t )
10:
end for

5. Experiment

To evaluate the proposed ATSC algorithm based on the Sequence Decision Transformer (SDT), two sets of experiments were conducted. All experiments were performed using Simulation of Urban Mobility (SUMO) [30], a traffic simulation tool. In the experiments, all intersections were set in a synthetic traffic environment. To ensure the practicality of the proposed model, the experiments focused on standard intersections, the most common type of traffic scenario in modern cities, where vehicles can proceed straight, turn left, or turn right. U-turns were not permitted in the designed synthetic scenarios, as most urban intersections do not allow direct U-turns due to the increased risk of traffic accidents. Therefore, the option for vehicles to make U-turns was excluded from our experiments.
Specifically, the first experiment demonstrates the training process of the proposed algorithm, showing the evolution of performance during iterations of policy updates and the final performance of the policy. The second experiment compares the performance of a fixed-time traffic signal, a baseline PPO method, a DQN algorithm specifically designed for ATSC problems [31], and a FRAP [26] algorithm to the SDT algorithm under various traffic scenarios. This section begins with the experimental setup and implementation details, followed by an analysis of the training and test results obtained.

5.1. Experiment Setting

The training scenario is a intersection that has 32 lanes, while each direction has 3 straight lanes, 1 right-hand lane, and 1 left-hand, lane and they are all 750 m long. The traffic flow density is approximately 600–5000 veh/h during the training process.
In SUMO, the shortest simulation time interval, namely, the time-step τ , is 1 s. The traffic flow, which is randomly generated at various moments within 3600 s, will enter incoming lanes, thereby pass by the intersection. The SDT algorithm has been trained 3 million times with an episode length of 500, which means that there are 6000 episodes in this training scenario. In the synthetic scenario, the weights assigned to each item of the discount reward function are { w 1 , w 2 , w 3 , w 4 } = { 0.5 , 0.5 , 2 , 0.5 }. Each decision-making T has an interval of 6 s, meaning the agent will decide whether to maintain the current phase or switch to another phase every 6 s. If T is set too short, such as 1 s, it can cause traffic lights to switch too often, which is impractical and increases the computational load. Conversely, if T is set too long, it will slow the agent’s learning rate. The parameters of the baseline PPO, DQN, and FRAP algorithm are restored to the original settings. Please refer to Table 2 for the detailed parameters of the SUMO simulator, baseline PPO method, DQN, FRAP algorithm, and SDT algorithm.

5.2. Analysis of Performance during Training Process

In this experiment, a high-performing ATSC agent was successfully trained using the SDT algorithm. Figure 3a illustrates the changes in reward values throughout the training process. The curve in the graph represents the variation of the discount rewards as the number of time steps increases. During exploration of the agent and updating of the network parameters, the rewards curve gradually increases, indicating that the policy has become adept at efficiently handling problems, and the agent is continuously adapting to varying scenarios. At around 75,000 steps, the reward curve reaches its peak and subsequently fluctuates with minimal variance for an extended period. This suggests that the agent has learned the optimal policy after approximately 75,000 steps, and the subsequent fluctuations are due to randomness in traffic scenarios and vehicle flows at different times. The rapid convergence and superior policy performance can be attributed to the integration of the SDT framework, which leverages the encoder–decoder architecture for efficient data processing. The self-attention mechanism within the transformer model enables the SDT to capture intricate dependencies and dynamic changes in traffic conditions, thereby enhancing the learning process. Additionally, the application of PPO ensures stable and continuous policy updates, preventing large policy shifts and maintaining consistent performance improvements. This robust methodological framework contributes significantly to the observed rapid convergence and high-quality policy performance during training.
To validate the convergence of the rewards and each component, the changes in three reward components were recorded with respect to the number of steps. Figure 3b,c display the total number of vehicles and the average waiting time of vehicles, respectively. The trends in these two reward items are essentially consistent with the reward curve. The total number of vehicles decreases with the update of the policy, reaching a low point around 100,000 steps and subsequently fluctuating within a small range. The average waiting time reaches its lowest point after around 80,000 steps, as the policy updates and remains relatively stable, with only minor fluctuations thereafter. Figure 3d shows the queue length as the number of time steps increases. During the training process, the queue length initially increases slightly and then gradually decreases. As the policy is updated, the minimum queue length occurs around 160,000 steps and maintains relative stability thereafter with only minor fluctuations. Although the reward items converge at different times, the trends and the final outcomes are the same. This indicates that the convergence of the rewards is accompanied by the convergence of the reward items, and the traffic light at the intersection becomes more efficient in managing traffic flows under the influence of the SDT algorithm, leading to improved traffic conditions at the intersection.

5.3. Performance Comparison

This section compares the performance of the policy trained using the SDT algorithm with the fixed-time method, a PPO method, a DQN tailored for the ATSC method, and a SOTA ATSC algorithm, FRAP. To demonstrate the performance of the SDT algorithm, three typical scenarios were used, which are illustrated in Figure 4. In all scenarios, traffic flow follows a Poisson distribution; namely, traffic flow is generated randomly, and the probability of arrival is the same within equal time intervals during the simulation period. The traffic density for each lane at the intersection remains constant and does not vary under different conditions such as peak hours.
  • Easy scenario: In this scenario, traffic can only flow from north to south and from west to east, with each direction consisting of two lanes. The traffic density is 600 veh/h.
  • Medium scenario: This scenario is similar to the training scenario, with each direction consisting of four lanes. The traffic density is 4000 veh/h.
  • Hard scenario: This intersection has 48 lanes, while each direction has 4 straight lanes, 2 right-turn lane, and 2 left-turn lane. The traffic flow density is approximately 5000 veh/h.
In the scenarios described above, several experiments were also set up to compare these algorithms with the SDT algorithm:
  • FIX6/12/18/24: The figures represent the fixed duration of each green light phase. In the FIXED experiments, the green light phases cycles with a fixed duration, simulating the traffic lights commonly found at most real-life intersections to influence the traffic conditions.
  • Baseline PPO: This implementation is a standard, unmodified version of PPO without any specific design tailored for ATSC. The policy model is MLP, which is used to learn and optimize traffic signal control strategies from the traffic environment. However, due to the lack of specific design adjustments, this PPO algorithm represents a straightforward application in the ATSC problem.
  • DQN [31]: This DQN-based DRL algorithm is designed for ATSC. It leverages a deep convolutional neural network (CNN) to extract useful features from traffic data and utilizes the Q-learning algorithm to identify the optimal traffic signal control strategy.
  • FRAP [26]: FRAP models traffic signal control as a phase competition problem, giving priority to the traffic phase with the highest demand. By leveraging symmetry in traffic flow (flipping and rotation), FRAP reduces the problem’s complexity and state space, thus improving learning efficiency.
Table 3 shows final results of the comparison. To evaluate the performances of the different methods, three key metrics were used: the total number of vehicles at the intersection (vehicle number), the average speed of all vehicles (speed), and the average queue length at the intersection (queue). These metrics provide a straightforward reflection of the congestion level at the intersection, serving as core indicators for assessing the model’s performance, with their calculation formulas corresponding to Equations (2), (3), and (4), respectively. The results demonstrate that, across all scenarios, the SDT algorithm significantly outperforms the FIXED method and the baseline PPO method in all aspects.
In the easy scenario, SDT achieves improvements of 5% and 2% in vehicle number and speed compared to the FRAP method. These relatively modest improvements can be attributed to the lighter traffic conditions in this scenario, where the traditional ATSC algorithm is still sufficiently effective. However, the encoder–decoder architecture of SDT allows it to capture more nuanced dynamics of traffic flow, leading to slight performance gains.
In the medium scenario, the improvements by SDT become much more pronounced, with gains of 23%, 65%, and 13% in vehicle number, speed, and queue, respectively. This significant enhancement is primarily due to the self-attention mechanism within the SDT’s transformer model, which captures complex dependencies and dynamic interactions among traffic flows, enabling SDT to adjust signal control more precisely than the DQN and FRAP method. As traffic fluctuations and complexity increase in medium scenarios, FRAP’s limited capacity to model temporal dependencies and interactions among multiple streams becomes more apparent, leading to inferior performance.
In the hard scenario, SDT’s advantages are even more evident, with improvements of 26%, 24%, and 34% in vehicle number, speed, and queue, respectively. In such high-complexity scenarios, FRAP and the DQN struggle to cope with the high traffic volume and intricate interactions, resulting in significantly lower performances. In contrast, SDT not only leverages its encoder–decoder architecture to effectively process complex data but also benefits from the stable and continuous policy updates of the PPO algorithm, preventing large policy shifts and ensuring consistent performance improvements. SDT’s ability to dynamically adjust according to real-time traffic conditions greatly alleviates congestion and enhances throughput efficiency.
However, it is worth noting that the SDT algorithm relies heavily on computational resources and is sensitive to hyperparameter settings. While these factors contribute to its high performance, they also highlight areas for further optimization. Reducing computational demands and enhancing the algorithm’s robustness to hyperparameter variations could broaden SDT’s applicability.
Overall, as the scale of the intersection and traffic flow increases, the SDT algorithm, with its sophisticated architecture and advanced learning mechanisms, effectively addresses the limitations of traditional methods such as FRAP in terms of dynamic adaptability and handling complex traffic conditions, achieving more flexible and efficient traffic management.

6. Conclusions

The SDT introduced in this paper represents a significant advancement in the field of ATSC using DRL. By adopting a sequence decision-making framework for traffic control, the SDT model effectively addresses the high variability and complexity of urban traffic flows with remarkable efficiency. The integration of a PPO-based algorithm within this framework ensures that the model remains stable and adaptable across various traffic conditions, which is essential for real-world applications. Our comprehensive experimental evaluation demonstrates that the SDT outperforms existing methods, including fixed-time, the baseline PPO method, and advanced FRAP-based systems, by significantly improving traffic throughput and reducing congestion. These improvements are achieved without compromising the system’s adaptability and responsiveness to dynamic traffic patterns, highlighting the potential of DRL in transforming traffic management systems. The success of the SDT model in these experiments indicates that it can serve as a robust framework for future developments in intelligent transportation systems, potentially leading to more sustainable and efficient urban environments. However, further research is needed to address certain limitations. Firstly, the SDT algorithm requires extensive computational resources for training, which may limit its scalability. Secondly, the potential sensitivity to hyperparameter settings necessitates careful tuning to ensure optimal performance across different scenarios. To mitigate these issues, future work will focus on optimizing the computational efficiency and exploring automated hyperparameter tuning techniques, thereby enhancing the overall robustness and applicability of the SDT algorithm.

Author Contributions

Conceptualization, R.Z.; methodology, H.H.; software, H.H.; validation, Y.L.; formal analysis, Y.F.; investigation, Z.G.; resources, F.G.; writing—original draft preparation, H.H.; writing—review and editing, R.Z.; visualization, F.G.; funding acquisition, R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Science Foundation of China under Grants 52202495, 52202494, and 52394261, as well as the Science and Technology Development Project of Jilin Province (No. 202302013).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available from the corresponding author upon request.

Conflicts of Interest

On behalf of all the authors, the corresponding author states that there are no conflicts of interest.

Appendix A

The appendix contains a comprehensive table that lists and explains the notations used throughout the main text, providing a clear and concise reference for better understanding of the content discussed.
Table A1. List of notations.
Table A1. List of notations.
NotationDescription
s i The raw information of lane i
N l a n e The maximum number of lanes
OThe observation space
AThe action space
RThe reward function
PThe transition probability function
TThe probability of transitioning from observation o t to o t + 1
O i The observation space of lane i
R n u m , R v o l , R q u e , R t i m e The items of the reward function
w 1 , w 2 , w 3 , w 4 The weights for items of the reward function
W q , W k , w v The attention weight matrix
Q , K , V The vector of queries, keys, and values
d k The dimension of Q and K
L ( · ) The loss function
Q ( o , a ) The Q function
V ( o ) The value function
A ( o , a ) The advantage function

References

  1. Wu, Q.; Shen, J.; Yong, B.; Wu, J.; Li, F.; Wang, J.; Zhou, Q. Smart fog based workflow for traffic control networks. Future Gener. Comput. Syst. 2019, 97, 825–835. [Google Scholar] [CrossRef]
  2. Bao, Z.; Ng, S.T.; Yu, G.; Zhang, X.; Ou, Y. The effect of the built environment on spatial-temporal pattern of traffic congestion in a satellite city in emerging economies. Dev. Built Environ. 2023, 14, 100173. [Google Scholar] [CrossRef]
  3. Kumar, N.; Rahman, S.S.; Dhakad, N. Fuzzy inference enabled deep reinforcement learning-based traffic light control for intelligent transportation system. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4919–4928. [Google Scholar] [CrossRef]
  4. Wei, H.; Zheng, G.; Gayah, V.; Li, Z. Recent advances in reinforcement learning for traffic signal control: A survey of models and evaluation. ACM SIGKDD Explor. Newsl. 2021, 22, 12–18. [Google Scholar] [CrossRef]
  5. Guo, Q.; Li, L.; Ban, X.J. Urban traffic signal control with connected and automated vehicles: A survey. Transp. Res. Part C Emerg. Technol. 2019, 101, 313–334. [Google Scholar] [CrossRef]
  6. Alegre, L.N.; Ziemke, T.; Bazzan, A.L. Using reinforcement learning to control traffic signals in a real-world scenario: An approach based on linear function approximation. IEEE Trans. Intell. Transp. Syst. 2021, 23, 9126–9135. [Google Scholar] [CrossRef]
  7. Han, G.; Liu, X.; Wang, H.; Dong, C.; Han, Y. An Attention Reinforcement Learning–Based Strategy for Large-Scale Adaptive Traffic Signal Control System. J. Transp. Eng. Part A Syst. 2024, 150, 04024001. [Google Scholar] [CrossRef]
  8. Hunt, P.; Robertson, D.; Bretherton, R.; Royle, M.C. The SCOOT on-line traffic signal optimisation technique. Traffic Eng. Control 1982, 23, 190–192. [Google Scholar]
  9. Luk, J. Two traffic-responsive area traffic control methods: SCAT and SCOOT. Traffic Eng. Control 1984, 25, 14. [Google Scholar]
  10. Sun, Q.W.; Han, S.Y.; Zhou, J.; Chen, Y.H.; Yao, K. Deep Reinforcement-Learning-Based Adaptive Traffic Signal Control with Real-Time Queue Lengths. In Proceedings of the 2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Prague, Czech Republic, 9–12 October 2022; pp. 1760–1765. [Google Scholar]
  11. Kong, A.Y.; Lu, B.X.; Yang, C.Z.; Zhang, D.M. A deep reinforcement learning framework with memory network to coordinate traffic signal control. In Proceedings of the 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China, 8–12 October 2022; pp. 3825–3830. [Google Scholar]
  12. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  13. Noaeen, M.; Naik, A.; Goodman, L.; Crebo, J.; Abrar, T.; Abad, Z.S.H.; Bazzan, A.L.; Far, B. Reinforcement learning in urban network traffic signal control: A systematic literature review. Expert Syst. Appl. 2022, 199, 116830. [Google Scholar] [CrossRef]
  14. Wang, X.; Sanner, S.; Abdulhai, B. A Critical Review of Traffic Signal Control and A Novel Unified View of Reinforcement Learning and Model Predictive Control Approaches for Adaptive Traffic Signal Control. arXiv 2022. [Google Scholar] [CrossRef]
  15. Abdulhai, B.; Pringle, R.; Karakoulas, G.J. Reinforcement learning for true adaptive traffic signal control. J. Transp. Eng. 2003, 129, 278–285. [Google Scholar] [CrossRef]
  16. Raeis, M.; Leon-Garcia, A. A deep reinforcement learning approach for fair traffic signal control. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 2512–2518. [Google Scholar]
  17. Rasheed, F.; Yau, K.L.A.; Noor, R.M.; Wu, C.; Low, Y.C. Deep reinforcement learning for traffic signal control: A review. IEEE Access 2020, 8, 208016–208044. [Google Scholar] [CrossRef]
  18. Cao, M.; Li, V.O.; Shuai, Q. A gain with no pain: Exploring intelligent traffic signal control for emergency vehicles. IEEE Trans. Intell. Transp. Syst. 2022, 23, 17899–17909. [Google Scholar] [CrossRef]
  19. Ma, Z.; Cui, T.; Deng, W.; Jiang, F.; Zhang, L. Adaptive optimization of traffic signal timing via deep reinforcement learning. J. Adv. Transp. 2021, 2021, 6616702. [Google Scholar] [CrossRef]
  20. Zhang, G.; Chang, F.; Jin, J.; Yang, F.; Huang, H. Multi-objective deep reinforcement learning approach for adaptive traffic signal control system with concurrent optimization of safety, efficiency, and decarbonization at intersections. Accid. Anal. Prev. 2024, 199, 107451. [Google Scholar] [CrossRef]
  21. Li, L.; Lv, Y.; Wang, F.Y. Traffic signal timing via deep reinforcement learning. IEEE CAA J. Autom. Sin. 2016, 3, 247–254. [Google Scholar] [CrossRef]
  22. Wei, H.; Zheng, G.; Yao, H.; Li, Z. Intellilight: A reinforcement learning approach for intelligent traffic light control. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 19–23 August 2018; pp. 2496–2505. [Google Scholar]
  23. Nishi, T.; Otaki, K.; Hayakawa, K.; Yoshimura, T. Traffic signal control based on reinforcement learning with graph convolutional neural nets. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 877–883. [Google Scholar]
  24. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  25. Wang, M.; Wu, L.; Li, J.; He, L. Traffic signal control with reinforcement learning based on region-aware cooperative strategy. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6774–6785. [Google Scholar] [CrossRef]
  26. Zheng, G.; Xiong, Y.; Zang, X.; Feng, J.; Wei, H.; Zhang, H.; Li, Y.; Xu, K.; Li, Z. Learning phase competition for traffic signal control. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 1963–1972. [Google Scholar]
  27. Zhang, H.; Liu, C.; Zhang, W.; Zheng, G.; Yu, Y. Generalight: Improving environment generalization of traffic signal control via meta reinforcement learning. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, New York, NY, USA, 19–23 October 2020; pp. 1783–1792. [Google Scholar]
  28. Du, W.; Ye, J.; Gu, J.; Li, J.; Wei, H.; Wang, G. Safelight: A reinforcement learning method toward collision-free traffic signal control. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 14801–14810. [Google Scholar]
  29. Wen, M.; Kuba, J.; Lin, R.; Zhang, W.; Wen, Y.; Wang, J.; Yang, Y. Multi-agent reinforcement learning is a sequence modeling problem. Adv. Neural Inf. Process. Syst. 2022, 35, 16509–16521. [Google Scholar]
  30. Lopez, P.A.; Behrisch, M.; Bieker-Walz, L.; Erdmann, J.; Flötteröd, Y.P.; Hilbrich, R.; Lücken, L.; Rummel, J.; Wagner, P.; Wießner, E. Microscopic traffic simulation using sumo. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 2575–2582. [Google Scholar]
  31. Gao, J.; Shen, Y.; Liu, J.; Ito, M.; Shiratori, N. Adaptive traffic signal control: Deep reinforcement learning algorithm with experience replay and target network. arXiv 2017, arXiv:1705.02755. [Google Scholar]
Figure 1. Conceptual framework of sequence decision transformer.
Figure 1. Conceptual framework of sequence decision transformer.
Sensors 24 06202 g001
Figure 2. The encoder–decoder architecture of SDT. At each time step, the encoder takes in the agent’s observations and encodes them into a sequence of latent representations, which is then passed into the decoder, and the decoder generates the agent’s optimal action.
Figure 2. The encoder–decoder architecture of SDT. At each time step, the encoder takes in the agent’s observations and encodes them into a sequence of latent representations, which is then passed into the decoder, and the decoder generates the agent’s optimal action.
Sensors 24 06202 g002
Figure 3. Performance of SDT during the training process. (a) Rewards curve with time steps. (b) Number of vehicles in the intersection over time. (c) Average waiting time for the intersection over time. (d) Queue length of the intersection over time.
Figure 3. Performance of SDT during the training process. (a) Rewards curve with time steps. (b) Number of vehicles in the intersection over time. (c) Average waiting time for the intersection over time. (d) Queue length of the intersection over time.
Sensors 24 06202 g003
Figure 4. Performance comparison scenario of SDT, FRAP, baseline PPO, and FIXED method. (a) The easy scenario. (b) The medium scenario. (c) The hard scenario.
Figure 4. Performance comparison scenario of SDT, FRAP, baseline PPO, and FIXED method. (a) The easy scenario. (b) The medium scenario. (c) The hard scenario.
Sensors 24 06202 g004
Table 1. Observation space and action space.
Table 1. Observation space and action space.
NotationDescriptionRangeScale
v i Average velocity of lane i [ 0 , V max ] m/s [ 0 , V max ]
t i Waiting time of lane i [ 0 , ] s [ 0 , ]
q i Queue length of lane i [ 0 , ] [ 0 , ]
n i Total number of vehicles in lane i [ 0 , N max ] [ 0 , N max ]
pTraffic signal light phase { p h a s e 1 , , p h a s e N } { 0 , , N p h a 1 }
Table 2. Main parameters of the experiments.
Table 2. Main parameters of the experiments.
ParametersValues
SUMO simulator
Time-step τ 1 s
Traffic flow600–5000 veh/h
Edge length750 m
Baseline PPO
Learning rate α 1 × 10 3
Batch size B 64
Target update interval Δ t target 1024
Policy modelMLP
Exploration rate ϵ 0.1
Discount factor γ 0.95
Exploration rate decay ϵ decay 0.995
Minimum exploration rate ϵ min 0.01
DQN algorithm
Learning rate α 1 × 10 3
Replay buffer size N 50,000
Target update interval Δ t target 500
OptimizerAdam
Update strategy π MLP
Exploration factor ϵ frac 0.05
FRAP algorithm
Number of layers2
Dense dimension20
Exploration rate ϵ 0.1
Exploration rate decay ϵ decay 0.995
Minimum exploration rate ϵ min 0.01
Demand shape1
SDT algorithm
Learning rate α 5 × 10 4
Discount factor γ 0.99
Decision-making T 6 s
PPO clip0.05
OptimizerAdam
Timesteps3 × 10 6
Episode length500
Table 3. Performance comparison.
Table 3. Performance comparison.
ScenariosMetricsFIX6FIX12FIX18FIX24PPODQNFRAPSDT
EasyVehicle number3.433.73.44.11.92.01.9
Average speed0.730.630.460.390.590.880.850.87
Queue length0.1250.20.40.470.150.0250.0250.025
MediumVehicle number14341129.61055.51004.3949.5929.6427.1326.3
Average speed0.0560.10.110.130.070.080.230.38
Queue length34.4833.6634.2234.0354.737.729.725.6
HardVehicle number352.8301280.9426.7629.1143.7154.5113.3
Average speed0.0920.1660.1290.10.1200.0950.1360.168
Queue length6.354.565.157.1534.014.764.823.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, R.; Hu, H.; Li, Y.; Fan, Y.; Gao, F.; Gao, Z. Sequence Decision Transformer for Adaptive Traffic Signal Control. Sensors 2024, 24, 6202. https://doi.org/10.3390/s24196202

AMA Style

Zhao R, Hu H, Li Y, Fan Y, Gao F, Gao Z. Sequence Decision Transformer for Adaptive Traffic Signal Control. Sensors. 2024; 24(19):6202. https://doi.org/10.3390/s24196202

Chicago/Turabian Style

Zhao, Rui, Haofeng Hu, Yun Li, Yuze Fan, Fei Gao, and Zhenhai Gao. 2024. "Sequence Decision Transformer for Adaptive Traffic Signal Control" Sensors 24, no. 19: 6202. https://doi.org/10.3390/s24196202

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop