Next Article in Journal
Exponentially Long Transient Time to Synchronization of Coupled Chaotic Circle Maps in Dense Random Networks
Next Article in Special Issue
Self-Organisation of Prediction Models
Previous Article in Journal
An Agent-Based Statistical Physics Model for Political Polarization: A Monte Carlo Study
Previous Article in Special Issue
Schrödinger’s What is Life?—Complexity, Cognition and the City
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Reinforcement Meta-Learning and Self-Organization in Complex Systems: Applications to Traffic Signal Control

ETH Zurich, Computational Social Science, 8092 Zurich, Switzerland
Entropy 2023, 25(7), 982; https://doi.org/10.3390/e25070982
Submission received: 22 May 2023 / Revised: 12 June 2023 / Accepted: 16 June 2023 / Published: 27 June 2023
(This article belongs to the Special Issue Information and Self-Organization III)

Abstract

:
We studied the ability of deep reinforcement learning and self-organizing approaches to adapt to dynamic complex systems, using the applied example of traffic signal control in a simulated urban environment. We highlight the general limitations of deep learning for control in complex systems, even when employing state-of-the-art meta-learning methods, and contrast it with self-organization-based methods. Accordingly, we argue that complex systems are a good and challenging study environment for developing and improving meta-learning approaches. At the same time, we point to the importance of baselines to which meta-learning methods can be compared and present a self-organizing analytic traffic signal control that outperforms state-of-the-art meta-learning in some scenarios. We also show that meta-learning methods outperform classical learning methods in our simulated environment (around 1.5–2× improvement, in most scenarios). Our conclusions are that, in order to develop effective meta-learning methods that are able to adapt to a variety of conditions, it is necessary to test them in demanding, complex settings (such as, for example, urban traffic control) and compare them against established methods.

1. Introduction

As we learn more about the many systems that make up our world, the more complex it appears to become. The economy, biosphere, society, power grids, cities, and urban traffic are just a few examples of systems beyond our ability to fully understand, predict, or control.
A proposed response to the limitations of our understanding has recently come in the form of black box models such as deep learning. In complex systems, deep reinforcement learning (RL) methods have been especially favored and frequently used in control problems. There are many benefits to using deep learning, including its universality and ease with which it can be applied to a variety of problems. Indeed, in many tasks, deep learning has been shown to perform better than humans [1,2].
Nevertheless, there are also limitations to what deep learning methods can do. Some of the most commonly mentioned limitations include lack of robustness, adaptability, explainability, and fairness [3]. Most of the deep models do not currently contribute to a good understanding of the various systems they might help us control. Generally, we do not understand traffic any better by knowing that a particular deep learning model would control it in a certain way under some particular conditions.
In actuality, a deep learning model is an additional complex system we fail to fully understand. It might appear somehow unfounded to create a complex system, the inner workings of which we do not fully grasp, to control another complex system we do not entirely understand. This also makes it difficult to accurately assess the deep model’s success in an absolute sense, since we do not usually know the optimal behavior of the system we are attempting to control. To understand and interpret the performance and utility of deep learning models, it is useful to compare them against non-deep learning models that are well-studied and -understood.
Moreover, as dynamic complex systems continue to evolve in unpredictable ways, classical deep learning approaches often become inadequate as they cannot function effectively under significant data-shifts. Hence, the more adaptable meta-learning methods have become popular when applying deep learning to complex systems [4].
A real-world example of a complex system is urban traffic. Traffic congestion is an emergent property of most, if not all, sufficiently large modern metropolises. It is mostly considered as a negative emergent and leads to lost time, air and sound pollution, and frustration. Any truly smart city would hope to resolve the issue or at least mitigate it. Currently, the main way of dealing with congestion is through the design of smart traffic signal control approaches.
A great variety of RL multi-agent systems have been recently designed for control of traffic signals in urban environments [5,6,7], including ones that are able to meta-learn [8,9]. However, the majority of these methods have only been compared against previous RL methods (which they are usually shown to outperform). While this shows that deep learning for traffic signal control is continuously getting better, it is not yet clear how well it actually performs with regard to other approaches, such as analytic ones.
In this paper, we study the advantages and limitations of deep learning approaches for the problem of traffic signal control in the complex system of urban traffic. We conducted experiments on traffic signal control in simulated urban networks to compare RL methods with an analytic self-organization approach. We highlight that, while using deep learning in the context of complex systems has much potential, there is a need for strong baselines agains twhich to compare. Lastly, we conclude that, in the context of traffic signal control and at the current state of deep learning development, self-organization appears to be a good baseline, which is able to outperform RL methods in many settings.

2. Motivation

In this section, we discuss the relevant motivation, delineate key concepts, and present some background needed to grasp our experiments.

2.1. Deep Learning

Deep learning refers to any model that uses a Deep Neural Network (DNN). In most cases, the DNN is trained with some data and used as a classifier or decision maker. Deep learning does not require any specialist knowledge about the system for which it is used (but may still incorporate it in certain models [10,11]). Deep learning can be widely applied to a large variety of different systems; for example, the same DNN architecture can be used to learn to control traffic or a power grid.
On the other hand, the training of most deep learning models requires large, representative datasets and is computationally demanding [12]. The DNN is itself a complex structure, often containing thousands of parameters across many layers. It can itself be studied as a complex system [13] (where, from the local interactions of individual nodes, there emerge some global activation patterns). Due to its great number of unexplainable parameters, deep learning has recently been criticized for lacking fairness [14] and being prone to biases induced by training data [15]. Some of the limitations of deep learning have been linked to the strong influence on the field by the practice of engineering [3].
In traffic systems, deep RL has been applied to the problem of traffic signal control [16]. A variety of methods have been proposed, including [5,9].

2.2. Meta-Learning

In the context of learning conditions—where the underlying system might undergo an unpredictable change—or using data—in which distribution might shift—meta-learning has been seen as a promising methodology. Generally, meta-learning methods, unlike traditional deep learning, are able to be generalized to scenarios different from the one they were trained on [17]. In this context, a common method is to use gradient-based model agnostic meta-learning, periodically alternating between a global and individual adaptation [4]. One might also benefit from using mixed experience replays based on different scenarios [8]. Similarly, improved generalization can be gained from using model agnostic meta-learning and flow clustering [9]. Meta-learning can also be used to improve data efficiency, which then allows one to limit the interactions with the environment necessary for convergence [18]. Incorporating short- and long-term information by employing decentralized advantage actor–critic [19] can also lead to better generalization ability.
It is worth noting that, of the meta-learning methods proposed recently, many are highly complex. While showing great promise in dealing with the issues of adaptability to changing conditions, which are especially pertinent in complex systems, the meta-learning methods have mostly been compared to other learning methods [8,9]. It is the interest of this paper to also offer a comparison with non-learning methods that can be used as a simple baseline for future comparisons.

2.3. Complex Systems

A complex system is usually defined as a system composed of many sub-systems that interact with each other [20]. Such a system may be difficult to model, as the relationships between the parts are often non-linear and high-dimensional. For the same reason, such a system is also difficult to control. It may not even be clear how a given local intervention or action will affect the global state of the system. A concept often associated with complex systems is emergence, which is a spontaneous appearance of a global pattern precipitated by the local interactions of the system’s parts [21]. Since complex systems exhibit high levels of unpredictability and change, meta-learning is the best-suited learning paradigm for such settings.
Urban traffic can be seen as an instance of a complex system. It consists of many individual systems: intersections, vehicles, and pedestrians, all interacting with each other non-linearly [22].

2.4. Self-Organization

We follow a working definition of self-organization, wherein it is an adaptive process through which a system can acquire and maintain structure in the absence of external control [21]. Due to its decentralized character, such self-organized order exhibits high resilience to perturbations [23]. In social sciences, self-organization is also known as spontaneous order or invisible hand phenomenon, which characterizes liberal free-market economies [24]. Certain phenomena might also be considered to be self-organizing or not, depending on the perspective of the observer [25]. As such, self-organization is not a control approach, but rather a paradigm that aims to achieve desirable properties of the system, by creating favorable conditions for their emergence. Many definitions of self-organization have been given in the literature (see [21] for an overview); however, in this work, we do not follow a strict definition (only a working one, given above), but, rather, consider self-organization in the particular context of a complex system, where local interactions lead to a certain order being established at a global level. In Table 1, we present a comparison between deep learning and self-organization approaches in the context of complex systems.
An emergent self-organizing property observable in urban traffic would be the synchronization of neighboring signals, resulting in a ‘green wave’. An example of unfavorable self-organization, in contrast, would be the formation of traffic jams. A self-organizing traffic control aiming to overcome such congestion would be based on a set of modified interaction rules that are followed by individual agents in a decentralized manner (as, for example, in [26]).

2.5. Problem Description

In this work, we compare the effectiveness of deep learning and self-organization for controlling the traffic signals in an urban traffic system. The traffic system presents a challenging problem for deep learning algorithms as the traffic conditions can be highly dynamic and difficult to predict. A model which is trained only on morning traffic will likely under-perform on evening traffic. On the other hand, training a model to expect certain temporal patterns (e.g., morning vs. evening) might produce a model, which will not work well when an unexpected accident or roadwork causes the traffic patterns to change unexpectedly. Due to its high sensitivity, dynamic nature, and low predictability, the urban traffic system and its control are a difficult but interesting problem for the domain of deep learning. As we have mentioned, the particular approach that appears highly relevant to such a setting is meta-learning, which offers to produce models that are able to adapt and work well, even in changing conditions.
Generally, the traffic signal control problem consists of individual intersections (agents) taking actions. The actions consist of flows (traffic moving through an intersection from one lane to another) aggregated into phases (a phase represents the set of all flows that are given green, see Figure 1). The goal is to optimize a certain metric (often, the average travel time) to avoid or delay congestion.
The learning agents typically used to control signals at an intersection are based on deep reinforcement learning algorithms. The design of these agents can be defined by the reinforcement learning paradigm they use (e.g., q-learning or policy gradient), the actions, states, and reward. The actions available to such an agent correspond to phases that are possible at the intersection. An example of possible actions for an intersection with eight flows is given in Figure 1. The states can vary in detail (depending on the particular implementation) but correspond to the meaningful features of the traffic situation at the intersection. As such, they usually contain information such as the number of vehicles on the lanes (occupancy), the current phase, and so on. The reward corresponds to the measure of success and usually involves average travel time, delays, or number of stops. A description of the design of the intersection agents used in our experiments can be found in Section 3.

3. Methods

In this section, we introduce the main methods of the study—the Analytic+—and describe the details of the simulation experiments, including the reinforcement learning methods we compare against.

3.1. Analytic+

The Analytic+ algorithm [27] is based on short term anticipation of traffic. It relies on two rules: an optimization rule and a stabilization rule. The optimization rule is designed to minimize the total waiting times at a given intersection, while the stabilization rule maintains the stability and fairness of the service by ensuring that no flow incurs waiting times above a certain maximum. The Analytic+ method is decentralized, meaning the agents do not communicate with each other and make decisions based only on locally available information. This local information does, however, depend on the decisions taken by neighboring intersections; thus, an implicit communication between neighbors does occur. Each agent follows only two explainable rules. Despite the simplicity, from the actions of each agent emerges a global, self-organizing order. The algorithm has carefully designed explainable parameters that define the average and maximum waiting times allowed by the stabilization rule. The pseudocode for the algorithm is presented in Algorithm 1, and the code of the implementation used for this study is available at https://github.com/mbkorecki/meta_traffic (accessed on 26 June 2023).
Algorithm 1: Analytic+ pseudocode.
Entropy 25 00982 i001

3.1.1. Optimization Rule

The optimization rule prioritizes the flows with highest priority scores according to Equation (1):
π i = n ^ i τ i , σ p e n + τ i + g ^ i .
Herein, g ^ i is the required green time needed to clear lane i (this is a variable dependent on the number of cars on i and their maximum legal speed), τ i is the penalty for switching to another traffic flow, τ i , σ p e n is the penalty for switching back, and n ^ i is the number of vehicles that are expected to be served in time period g ^ i + τ i . This formulation ensures that the total waiting times of the vehicles at an intersection are minimized (assuming an optimization horizon that includes only these vehicles, whose waiting time depends on the current traffic signal).

3.1.2. Stabilization Rule

The stabilization rule is based on two parameters T a v g and T m a x , indicating the average waiting time and the maximum waiting time that can be incurred by each flow. These parameters can be specified separately for each intersection or homogeneously throughout the entire system. Using T a v g and T m a x , we define a critical threshold n i c r i t . When the number of vehicles waiting for a given flow surpasses n i c r i t , the flow is added to a priority queue and is served in the next service period. The n i c r i t is fully specified in Equation (2), where Q i indicates the average arrival rate of flow i, and z i is the variable service interval between two successive service processes for i:
n i c r i t = Q i T a v g T m a x z i T m a x T a v g .

3.2. Simulation Experiments

Two main components of the experiments were the analysis of the analytic approach—the effects of its paramterization on performance—and its comparison with state-of-the-art RL methods. The experiments were conducted in Cityflow [28]. The Cityflow simulator was selected for its high-speed performance (as compared to, for example, SUMO). In addition, most of the recent deep learning methods applied to traffic signal control (including the methods we reported in our experiments) have been implemented and tested in this particular simulator.
For the comparison experiments, we were mostly interested in the adaptability of the compared methods. The adaptability was expressed in terms of consistency of performance across a variety of scenarios (increasingly different from the training scenario, in the case of learning methods). Furthermore, the consistency of the performance across a given flow distribution was also of interest. Hence, we report the variation in performance in terms of the maximum and minimum values achieved on a given flow distribution by a given method.

3.3. Scenarios

We ran the experiments on two scenarios representing parts of Atlanta and Hangzhou; the road networks are visualized in Figure 2. Note that Atlanta had a much more irregular network, while Hangzhou had a regular grid. Moreover, intersections in the Atlanta scenario were heterogeneous (connecting four, three, and two roads), while, in the Hangzhou road network, all intersections were the same (connecting four roads). We chose these scenarios because they allowed us to test the methods in both heterogeneous and homogeneous settings. Furthermore, they have been used in many previous publications on these topics, making it easier to compare across a spectrum of publications and methods [5,8,9].
In real world traffic systems, the flows change at different time scales (minutes, hours, days, months). They might also be affected by unpredictable events such as accidents or road work. Clearly, one of the key characteristics of a good traffic signal control method is how well it can adapt to changing traffic conditions. Specifically, RL methods must be tested in conditions that are different from the ones they were trained on. Therefore, to test the adaptability of the compared methods, we generated a variety of traffic flows for each of the two scenarios. These flows can be understood as representing different traffic conditions that may result, for instance, due to disruptions or different times of day.
We followed the methodology from [9], where a Wasserstein generative adversarial network (WGAN) was used to generate traffic flows from distributions away from the training distribution (which is based on real world data). To make sure that the generated traffic flows were realistic enough, some constraints were enforced through the loss function. For example, the variation in the count of vehicles between two adjacent time intervals could not differ too much. The exact details of the implementation can be found in [9].
We generated four traffic flow distributions away from the training distribution by 0.005, 0.01, 0.05, and 0.1, in terms of Wasserstein distance. For each distribution, we generated ten flows, tested on all of them, and reported minimum, maximum, and mean average travel times achieved by each method. In Table 2, we present the details of the scenarios. As the Wasserstein distance from the initial training distribution increased, the median entry time into the system decreased, leading to more vehicles being present in the system at the same time, which increased the complexity of the system and difficulty of controlling it. The number of vehicles also increased with the Wasserstein distance, which again led to more complex traffic dynamics. The increase in the number of vehicles was more pronounced for the Hangzhou scenario.
To study the effects of parameter settings on Analytic+, we ran the method on all scenarios with five different parameter settings and reported the mean average travel times achieved for each of the flow scenarios (ten different flows from each distribution). The parameter settings were selected according to their meaning. The first parameter indicated the average wait time, while the second indicated the maximum wait time. Therefore, they needed to be sufficiently different (with the second being higher) in order to make sense. Generally, the lower the maximum waiting time, the better for the individual drivers (but not necessarily better for the system’s global performance). We searched through a varied range of parameters, from low (better for individuals) to high (potentially better for the system). The results for the RL methods were taken from [9].

3.4. Compared Methods

In our studies, we compared the results of the Analytic+ against eight RL methods. The full results are available in Appendix A, but, for the sake of brevity, we only present the results of the three best performing methods in Section 4. These methods were chosen because they were reported to have achieved state-of-the-art results when published and were compared with many other learning algorithms (but not analytic baselines, which was important for this paper). Moreover, the Generalight and Metalight methods explicitly used a form of meta-learning, which was of particular interest and was highly applicable to our problem statement. The hyperparameters of the learning methods and their implementations have been given in https://github.com/only-changer/GeneraLight, this repository (accessed on 26 June 2023), which follows [9]. The best hyperparameters for each algorithm were assumed to have been worked out in [9] or in each publication introducing the corresponding method, as referenced below. We described each agent in terms of its states, rewards, and additional details. The actions of each agent were the same and correspond to the phases given in Figure 1.
Specifically, the algorithms are listed below:
  • Analytic+: The self-organizing method described in Section 3.1;
  • Generalight: A meta-RL method based on a deep q-learning paradigm, using the MAML framework [4] and clustering [9]. It outperforms other RL methods in scenarios that are different from the training scenarios. States: one-hot encoded vector representing the current phase at the intersection and the numbers of vehicles on each incoming lane and outgoing lane. Reward: the negative of the pressure of the intersection, which is a metric measuring the imbalance between outgoing and incoming lanes [7];
  • Colight: An RL method based on a deep q-learning paradigm, designed to handle multi-intersection environments and using attention mechanism to share information between neighbors. States: one-hot encoded vector representing the current phase at the intersection and the number of vehicles on each incoming lane. Reward: the sum of queue lengths of each incoming lane [5];
  • Metalight: This method transfers knowledge between intersections, designed to generalize to intersections with varied numbers of incoming and outgoing lanes. States: one-hot encoded vector representing the current phase at the intersection and the number of vehicles on each incoming lane. Reward: the average queue length of incoming lanes [8].
It is worth noting that the Analytic+ method can be easily scaled to as large a number of intersection as is needed. Since it does not require training, the cost of deployment is constant. The cost of operation is low since the algorithm does not perform any costly computations. On the other hand, the deployment cost of the learning algorithms is higher as they need to be trained for each specific scenario.

4. Results

In this section, we report the results of the parameter study (Table 3 and Table 4), as well as the performance of the studied methods in the two scenarios (Figure 3 and Figure 4).

4.1. Analytic+ Parameter Study

Table 3 shows that, for the Atlanta scenarios, a good choice of parameters (w.r.t. avg. travel time) was T a v g = 120 , T m a x = 240 . This setting produced the lowest avg. travel times in three out of five flow scenarios. Note that the differences for flows D 0 , D 0.005 , and D 0.01 were minimal; they increased for D 0.05 and were even greater for D 0.1 . T a v g = 60 , T m a x = 120 was by far worse than the other settings, with D 0.01 , D 0.05 , and D 0.1 .
Based on Table 4, the optimal parameters for the first three flow scenarios of the Hangzhou scenarios were T = , T m a x = , and, for the last two, T = 240 , T m a x = 360 . The difference between the two parameter settings for the first three flow scenarios was very small (around 1%). For the last two flow scenarios, they were larger (around 3%). Since we considered the last two scenarios to be more complex (because they were most different from the training scenario), we preferred to use T = 240 , T m a x = 360 as the optimal parameter settings.
Note that the optimal parameter setting for the Hangzhou scenarios is different from the Atlanta scenarios, showcasing that these parameters are dependent on the flows and road networks. In the following, we used T = 120 , T m a x = 240 for Atlanta and T = 240 , T m a x = 360 for Hangzhou.

4.2. Comparison Study

Figure 3 shows the average travel time achieved in each of the compared methods on the Atlanta scenarios. In all five flow scenarios, the Analytic+ achieved better mean performance, in most cases with relatively tight variability intervals. It allows for approximately two times lower travel times in four out of five flow scenarios when compared to Generalight—the next-best method. Generalight and Metalight achieved similar results for the first three scenarios but diverged for the last two scenarios, where Generalight was superior. Moreover, Analytic+ also had the smallest variability intervals in four out of five flow scenarios, meaning it performed more consistently over scenarios sampled from the same distribution than the other methods.
Figure 4 displays the performance of the compared methods for the Hangzhou scenarios in terms of average travel time. Here, Analytic+ performed at the same level as Generalight in all five flow scenarios. Colight was significantly worse than both Generalight and Analytic+. Again, Analytic+ displayed the smallest variability intervals for all flows. Metalight performed worse than the other two RL methods, unlike in the Atlanta scenario, where it performed better than Colight.
Exact numerical results, along with comparison with additional methods, are available in Appendix A.

5. Discussion

The immediate impression from our results is that reinforcement learning methods do not appear to handle heterogeneity well. While Analytic+ and Generalight performed at a comparable level in the homogeneous, grid-like Hangzhou street network, Analytic+ was twice as good as Generalight in the heterogeneous artery of Atlanta. This might occur due to different lengths of incoming lanes in Atlanta scenario, as has been observed in [29]. Moreover, the heterogeneity of intersection (difference in the number of incoming lanes) might negatively affect RL’s ability to generalize [8]. Analytic+ does not appear to be affected—it can deal well with heterogeneity, both in terms of lane lengths and the number of lanes.
Moreover, in both scenarios, Analytic+ had the smallest variation in performance, indicating that it performs consistently well over any flow drawn from a given distribution. The same was not true for RL agents, some of which were strongly inconsistent, even across flows drawn from the same distribution (e.g., Colight). It is also clear (Appendix A) that most of the RL methods, other than Generalight, performed at a sub-optimal level in both scenarios across all distributions.
One of the limitations of this study is that it only compares the performance in terms of one metric—the average travel time. Nevertheless, there are other important metrics that can be taken into account, including waiting times (which can be optimized by the Analytic+ stabilization rule; (as such, Analytic+ is expected to perform very well again). A multi-objective approach to traffic optimization and comparison of methods across more than one metric will be a future direction of research.
Furthermore, a potential future extension is to learn the optimal parameters of the Analytic+ using RL methods. Thus, we could combine the benefits of deep learning with the benefits of the self-organization. Since Analytic+ parameters are explainable, the results learned by an RL agent could be easily validated.
Nevertheless, this traffic signal control problem seems highly relevant to the complexity as well as machine learning community. It provides a real-world example of a complex system which proves highly challenging for traditional deep learning as well as meta-learning algorithms. The need for lifelong adaptation is also clear in the traffic system, where unpredictable events such as accidents, jams, and roadwork can occur daily (the effects of such disruptions on learning algorithms in the context of traffic has been investigated in [30]). In order for a traffic signal control algorithm to be actually useful, the learning algorithm would need to enable continuous learning and be highly adaptable.
Moreover, this study raises a question of the applicability of deep learning to control of complex systems in general. This question should certainly be further studied; the most promising methodologoies appear to be meta-learning approaches. The non-learning baseline that we provide makes it easier to propose and compare new effective algorithms for this particular problem setting.
In conclusion, we have investigated the applicability of deep learning and self-organization methods to complex systems, exemplified by an urban traffic system. We have shown that the self-organization method can achieve better or comparable results to deep learning, while not requiring any training, in addition to being explainable and fair. Apparently, there is more work to be performed if deep learning is to be used in the field of complex systems. Clearly, deep learning has great potential (especially meta-learning, in the context of complex systems), but, in order for it to be fully realized, it needs to be approached critically and challenged against strong, non-deep learning baselines. We have provided an approachable setting in which the performance of learning and non-learning control methods can be studied in a complex system. We have also provided a strong, non-learning baseline for this system. We hope that our contribution will make it more accessible for future researchers to study and design learning approaches that can work well in the challenging domain of complex dynamic systems.

Funding

This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme, grant agreement No 833168.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The author would like to express deep gratitude to the members of Computational Social Science group at ETH Zurich for their support with special thanks to Dirk Helbing, Thomas Asikis, Vaiva Vasiliauskaite, Sachit Mahajan, Javier Argota Sánchez-Vaquerizo, Andrea Musso, Cesare Carissimo and Elisabeth Stockinger for their invaluable help in reviewing and proofreading this manuscript.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A. Additional Results

Here, we present the full results of the comparison experiments. The additional methods are listed here:
  • DQN: Simple deep q-network approach. Uses vehicles’ positions as states [6];
  • PressLight: State-of-the-art RL method. Uses number of vehicles on incoming and outgoing lanes, current phase as state, and pressure as the reward [7];
  • +MAML: Adds the MAML RL framework to a given method [4].
Table A1. Performance of all methods in the Atlanta scenarios, in terms of average travel time (seconds). The best methods are shown in bold.
Table A1. Performance of all methods in the Atlanta scenarios, in terms of average travel time (seconds). The best methods are shown in bold.
ModelAvg. Travel Time(s)
D 0 D 0.005 D 0.01 D 0.05 D 0.1
MaxMinMeanMaxMinMeanMaxMinMeanMaxMinMeanMaxMinMean
DQN301.7255.5280.3297.1241.8266.1577.1349.1399.9690.1375.5441.9991.3578.3755.3
+MAML264.3201.3226.5237.8194.0214.8517.9299.2344.8641.9345.0414.5983.2525.3742.1
PressLight472.0207.2286.9477.3198.0277.6561.0297.0346.2631.6383.3430.4968.0684.4859.7
+MAML223.0189.5210.1230.9195.8218.5370.8266.2306.4435.7361.9394.3994.5551.8669.8
MetaLight206.9176.8196.5258.4177.0209.6463.1279.9338.9583.3386.3429.9991.2616.5819.2
CoLight580.7195.5433.3593.6187.1435.2785.4256.6579.1903.4421.9797.5911.1526.5849.2
+MAML632.6140.9427.3600.0190.2290.4869.4255.1558.7919.4580.5794.0992.9527.1847.3
GeneraLight194.5165.1180.3200.0167.7181.6283.4232.8253.9363.3298.5329.0646.6409.0554.2
Analytic+86.582.984.085.882.083.996.792.593.8532.8230.3307.8263.3196.7232.6
Table A2. Performance of all methods in the Hangzhou scenarios, in terms of average travel time (seconds). The best methods are shown in bold.
Table A2. Performance of all methods in the Hangzhou scenarios, in terms of average travel time (seconds). The best methods are shown in bold.
ModelAvg. Travel Time(s)
D 0 D 0.005 D 0.01 D 0.05 D 0.1
MaxMinMeanMaxMinMeanMaxMinMeanMaxMinMeanMaxMinMean
DQN979.1846.6921.01016935.7986.71069964.11010112710241088128412091249
+MAML530.1394.8448.2605.0462.0529.5707.9504.2617.0855.8715.7781.01147884.41021
PressLight395.9384.6389.9435.8415.7426.6484.2436.2460.3662.0565.7592.9855.1749.2813.4
+MAML468.1395.6424.0597.9468.9520.5712.4569.1634.5909.8665.9801.41100817.01004
MetaLight514.4475.0497.2587.3512.0545.6671.6580.5637.8843.3719.9767.4937.9814.2861.1
CoLight425.6368.7383.9603.0387.7456.4721.5402.7537.8887.9517.1624.11040.4726.2880.2
+MAML402.7390.9397.5490.4424.8452.0531.7458.6482.9643.1513.0570.2785.4638.0718.3
GeneralLight385.1376.3380.0410.2393.6402.0458.7409.4432.2582.0443.5493.5652.2579.8622.5
Analytic+403.2395.4398.4428.3419.3423.6452.0444.2448.8548.5526.4537.5682.3616.3640.9

References

  1. Van den Berg, J.; Miller, S.; Duckworth, D.; Hu, H.; Wan, A.; Fu, X.Y.; Goldberg, K.; Abbeel, P. Superhuman performance of surgical tasks by robots using iterative learning from human-guided demonstrations. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–8 May 2010; pp. 2074–2081. [Google Scholar]
  2. Silver, D.; Hubert, T.; Schrittwieser, J.; Antonoglou, I.; Lai, M.; Guez, A.; Lanctot, M.; Sifre, L.; Kumaran, D.; Graepel, T.; et al. Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm. arXiv 2017, arXiv:1712.01815. [Google Scholar]
  3. Ha, D.; Tang, Y. Collective Intelligence for Deep Learning: A Survey of Recent Developments. Collect. Intell. 2021, 1, 26339137221114874. [Google Scholar] [CrossRef]
  4. Finn, C.; Abbeel, P.; Levine, S. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  5. Wei, H.; Xu, N.; Zhang, H.; Zheng, G.; Zang, X.; Chen, C.; Zhang, W.; Zhu, Y.; Xu, K.; Li, Z. Colight: Learning network-level cooperation for traffic signal control. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 1913–1922. [Google Scholar]
  6. Wei, H.; Zheng, G.; Yao, H.; Li, Z. IntelliLight. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 2496–2505. [Google Scholar]
  7. Wei, H.; Chen, C.; Zheng, G.; Wu, K.; Gayah, V.; Xu, K.; Li, Z. Presslight: Learning Max pressure control to coordinate traffic signals in arterial network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; Volume 1, pp. 1290–1298. [Google Scholar]
  8. Zang, X.; Yao, H.; Zheng, G.; Xu, N.; Xu, K.; Li, Z. MetaLight: Value-Based Meta-Reinforcement Learning for Traffic Signal Control. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1153–1160. [Google Scholar]
  9. Zhang, H.; Liu, C.; Zhang, W.; Zheng, G.; Yu, Y. GeneraLight: Improving Environment Generalization of Traffic Signal Control via Meta Reinforcement Learning. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Online, 19–23 October 2020; pp. 1783–1792. [Google Scholar]
  10. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  11. Böttcher, L.; Antulov-Fantulin, N.; Asikis, T. AI Pontryagin or how artificial neural networks learn to control dynamical systems. Nat. Commun. 2022, 13, 1–9. [Google Scholar] [CrossRef] [PubMed]
  12. Korecki, M.; Helbing, D. Analytically guided reinforcement learning for green it and fluent traffic. IEEE Access 2022, 10, 96348–96358. [Google Scholar] [CrossRef]
  13. Testolin, A.; Piccolini, M.; Suweis, S. Deep learning systems as complex networks. J. Complex Netw. 2018, 8, cnz018. [Google Scholar] [CrossRef]
  14. Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 2021, 54, 1–35. [Google Scholar] [CrossRef]
  15. Tommasi, T.; Patricia, N.; Caputo, B.; Tuytelaars, T. A deeper look at dataset bias. In Domain Adaptation in Computer Vision Applications; Springer: Berlin/Heidelberg, Germany, 2017; pp. 37–55. [Google Scholar]
  16. Mannion, P.; Duggan, J.; Howley, E. An Experimental Review of Reinforcement Learning Algorithms for Adaptive Traffic Signal Control. Autonomic Road Transport Support Systems; Springer: Berlin/Heidelberg, Germany, 2016; pp. 47–66. [Google Scholar]
  17. Hospedales, T.; Antoniou, A.; Micaelli, P.; Storkey, A. Meta-Learning in Neural Networks: A Survey. In IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  18. Huang, X.; Wu, D.; Jenkin, M.; Boulet, B. ModelLight: Model-Based Meta-Reinforcement Learning for Traffic Signal Control. arXiv 2021, arXiv:2111.08067. [Google Scholar]
  19. Yang, S.; Yang, B. A Meta Multi-agent Reinforcement Learning Algorithm for Multi-intersection Traffic Signal Control. In Proceedings of the 2021 IEEE International Symposium on Dependable, Autonomic and Secure Computing (DASC), Online, 25–28 October 2021; pp. 18–25. [Google Scholar]
  20. Ladyman, J.; Lambert, J.; Wiesner, K. What is a complex system? Eur. J. Philos. Sci. 2013, 3, 33–67. [Google Scholar] [CrossRef]
  21. Wolf, T.D.; Holvoet, T. Emergence versus self-organisation: Different concepts but promising when combined. In Proceedings of the International Workshop on Engineering Self-Organising Applications; Springer: Berlin/Heidelberg, Germany, 2004; pp. 1–15. [Google Scholar]
  22. Helbing, D. Traffic and related self-driven many-particle systems. Rev. Mod. Phys. 2001, 73, 1067. [Google Scholar] [CrossRef] [Green Version]
  23. Heylighen, F. The science of self-organization and adaptivity. Encycl. Life Support Syst. 2001, 5, 253–280. [Google Scholar]
  24. Hayek, F.A. Law, Legislation and Liberty, Volume 1: Rules and Order; University of Chicago Press: Chicago, IL, USA, 1973; Volume 1. [Google Scholar]
  25. Gershenson, C.; Heylighen, F. When Can we Call a System Self-organizing? In Proceedings of the Advances in Artificial Life: 7th European Conference, ECAL 2003, Dortmund, Germany, 14–17 September 2003.
  26. Gershenson, C. Self-organizing traffic lights. arXiv 2004, arXiv:nlin/0411066. [Google Scholar]
  27. Lämmer, S.; Helbing, D. Self-control of traffic lights and vehicle flows in urban road networks. J. Stat. Mech. Theory Exp. 2008, 2008, 1–36. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, H.; Ding, Y.; Zhang, W.; Feng, S.; Zhu, Y.; Yu, Y.; Li, Z.; Liu, C.; Zhou, Z.; Jin, H. CityFlow: A multi-agent reinforcement learning environment for large scale city traffic scenario. In The Web Conference 2019—Proceedings of the World Wide Web Conference, WWW 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 3620–3624. [Google Scholar]
  29. Korecki, M. Adaptability and sustainability of machine learning approaches to traffic signal control. Sci. Rep. 2022, 12, 16681. [Google Scholar] [CrossRef] [PubMed]
  30. Korecki, M.; Dailisan, D.; Helbing, D. How Well Do Reinforcement Learning Approaches Cope With Disruptions? The Case of Traffic Signal Control. IEEE Access 2023, 11, 36504–36515. [Google Scholar] [CrossRef]
Figure 1. Left: An intersection with eight flows (not counting the right turns, which are assumed to always be given green in any phase). The green arrows correspond to the flows that would be given green by phase 1 from the right figure, red arrows represent flows that would be given red. Right: The possible actions (phases) available to an intersection agent controlling an intersection with eight flows (indicated by number 1 through 8).
Figure 1. Left: An intersection with eight flows (not counting the right turns, which are assumed to always be given green in any phase). The green arrows correspond to the flows that would be given green by phase 1 from the right figure, red arrows represent flows that would be given red. Right: The possible actions (phases) available to an intersection agent controlling an intersection with eight flows (indicated by number 1 through 8).
Entropy 25 00982 g001
Figure 2. The road networks of the two studied scenarios. The black lines represent roads, the blue dots intersections.
Figure 2. The road networks of the two studied scenarios. The black lines represent roads, the blue dots intersections.
Entropy 25 00982 g002
Figure 3. The performance, in terms of average travel time (seconds), of the compared methods in the Atlanta scenarios. The variability intervals represent the minimum and maximum values achieved for a given set of scenarios; the mean is marked on the line. Lines are guides to the eye only, to indicate trends.
Figure 3. The performance, in terms of average travel time (seconds), of the compared methods in the Atlanta scenarios. The variability intervals represent the minimum and maximum values achieved for a given set of scenarios; the mean is marked on the line. Lines are guides to the eye only, to indicate trends.
Entropy 25 00982 g003
Figure 4. The performance, in terms of average travel time (seconds), of the compared methods in the Hangzhou scenarios. The variability intervals represent the minimum and maximum values achieved for a given set of scenarios; the mean is marked on the line. Lines are guides to the eye only, to indicate trends.
Figure 4. The performance, in terms of average travel time (seconds), of the compared methods in the Hangzhou scenarios. The variability intervals represent the minimum and maximum values achieved for a given set of scenarios; the mean is marked on the line. Lines are guides to the eye only, to indicate trends.
Entropy 25 00982 g004
Table 1. Differences between deep learning and self-organization applied to complex systems.
Table 1. Differences between deep learning and self-organization applied to complex systems.
Deep LearningSelf-Organization
Requires training on large and representative datasetsIdeally no training and no or little data requirements
Might need retraining to adapt if the system evolves in a way such that the training data are not representativeCan adapt to an evolution that the system might undergo and needs no retraining
Needs an extremely large set of parameters, which are not explainableTypically uses a limited number of explainable parameters
Might be lacking fairnessFairness can be promoted/guaranteed
Requires no specialist knowledge or limited knowledge of the systemMight require specialist knowledge about the system to design effective rules
Can be universally applied to a variety of different systemsApplication is limited to specific kinds of systems which are able to self-organize
Table 2. The number of vehicles and their median entry time (s) into the system for all studied scenarios. Median entry time represents the time at which half of all the vehicles have entered the system.
Table 2. The number of vehicles and their median entry time (s) into the system for all studied scenarios. Median entry time represents the time at which half of all the vehicles have entered the system.
Number of VehiclesMedian Entry Time (s)
Scenario D 0 D 0.005 D 0.01 D 0.05 D 0.1 D 0 D 0.005 D 0.01 D 0.05 D 0.1
Atlanta2530257030603150312019281938185618461676
Hangzhou5397644774347875810619411923189418381680
Table 3. Mean of the average travel times in seconds achieved by the Analytic+ algorithm in the Atlanta scenarios for different parameter settings. D . indicates the traffic flow type of the scenario in terms of the Wasserstein distance from the D 0 scenario. The best (lowest) average travel times are shown in bold. Parameters set to correspond to the stabilization rule not being applied.
Table 3. Mean of the average travel times in seconds achieved by the Analytic+ algorithm in the Atlanta scenarios for different parameter settings. D . indicates the traffic flow type of the scenario in terms of the Wasserstein distance from the D 0 scenario. The best (lowest) average travel times are shown in bold. Parameters set to correspond to the stabilization rule not being applied.
Avg. Travel Time(s)
Parameter Settings D 0 D 0.005 D 0.01 D 0.05 D 0.1
T a v g = 60 , T m a x = 120 85.586.0393.1664.21113.5
T a v g = 120 , T m a x = 240 84.083.993.8307.8232.6
T a v g = 240 , T m a x = 360 83.884.293.9289.4315.1
T a v g = 360 , T m a x = 480 83.984.093.9276.1353.0
T a v g = , T m a x = 83.984.093.9292.3363.7
Table 4. Mean of the average travel time in seconds achieved by the Analytic+ algorithm in the Hangzhou scenarios for different parameter settings. D . indicates the traffic flow type of the scenario in terms of the Wasserstein distance from the D 0 scenario. The best (lowest) average travel times are shown in bold. Parameters set to correspond to the stabilization rule not being applied.
Table 4. Mean of the average travel time in seconds achieved by the Analytic+ algorithm in the Hangzhou scenarios for different parameter settings. D . indicates the traffic flow type of the scenario in terms of the Wasserstein distance from the D 0 scenario. The best (lowest) average travel times are shown in bold. Parameters set to correspond to the stabilization rule not being applied.
Avg. Travel Time(s)
Parameter Settings D 0 D 0.005 D 0.01 D 0.05 D 0.1
T a v g = 60 , T m a x = 120 542.2645.5763.1851.1995.8
T a v g = 120 , T m a x = 240 438.0504.4625.9735.4842.5
T a v g = 240 , T m a x = 360 398.4423.6448.8537.5640.9
T a v g = 360 , T m a x = 480 400.8425.3445.0554.1648.9
T a v g = , T m a x = 396.7417.0442.8557.3660.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Korecki, M. Deep Reinforcement Meta-Learning and Self-Organization in Complex Systems: Applications to Traffic Signal Control. Entropy 2023, 25, 982. https://doi.org/10.3390/e25070982

AMA Style

Korecki M. Deep Reinforcement Meta-Learning and Self-Organization in Complex Systems: Applications to Traffic Signal Control. Entropy. 2023; 25(7):982. https://doi.org/10.3390/e25070982

Chicago/Turabian Style

Korecki, Marcin. 2023. "Deep Reinforcement Meta-Learning and Self-Organization in Complex Systems: Applications to Traffic Signal Control" Entropy 25, no. 7: 982. https://doi.org/10.3390/e25070982

APA Style

Korecki, M. (2023). Deep Reinforcement Meta-Learning and Self-Organization in Complex Systems: Applications to Traffic Signal Control. Entropy, 25(7), 982. https://doi.org/10.3390/e25070982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop