Next Article in Journal
Green Supplier Selection Using Fuzzy AHP, Fuzzy TOSIS, and Fuzzy WASPAS: A Case Study of the Moroccan Automotive Industry
Previous Article in Journal
Perspectives on Artificial Intelligence Adoption for European Union Elderly in the Context of Digital Skills Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Multi-Vehicle Collaborative Planning Technology under Automatic Driving

1
College of Aviation, Inner Mongolia University of Technology, Hohhot 010051, China
2
Inner Mongolia High-Grade Highway Construction and Development Company, Hohhot 010051, China
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(11), 4578; https://doi.org/10.3390/su16114578
Submission received: 6 March 2024 / Revised: 9 May 2024 / Accepted: 21 May 2024 / Published: 28 May 2024
(This article belongs to the Section Sustainable Transportation)

Abstract

:
Autonomous vehicles hold the potential to significantly improve traffic efficiency and advance the development of intelligent transportation systems. With the progression of autonomous driving technology, collaborative planning among multiple vehicles in autonomous driving scenarios has emerged as a pivotal challenge in realizing intelligent transportation systems. Serving as the cornerstone of unmanned mission decision-making, collaborative motion planning algorithms have garnered increasing attention in both theoretical exploration and practical application. These methods often follow a similar paradigm: the system initially discerns the driving intentions of each vehicle, subsequently assesses the surrounding environment, engages in path-planning, and formulates specific behavioral decisions. The paper discusses trajectory prediction, game theory, following behavior, and lane merging issues within the paradigm mentioned above. After briefly introducing the background of multi-vehicle autonomous driving, it provides a detailed description of the technological prerequisites for implementing these techniques. It reviews the main algorithms in motion planning, their functionalities, and applications in road environments, as well as current and future challenges and unresolved issues.

1. Introduction

As urban ground traffic congestion becomes increasingly severe, it has evolved into a significant challenge hindering urban development. Congestion in large urban areas results in increased time and fuel costs for residents, leading to economic losses amounting to USD 121 billion [1]. The indirect losses caused by traffic congestion are also rising annually. The air pollution generated by urban traffic can lead to health damages ranging from GBP 4.5 billion to GBP 10 billion [2]. Therefore, implementing effective measures to alleviate traffic congestion is urgently needed.
One important way to handle problems like traffic congestion, energy usage, and environmental pollution is through multi-vehicle collaboration technology [3]. It efficiently lowers energy usage, exhaust pollutants, and traffic congestion by optimizing vehicle driving routes and speeds. In contrast to single vehicles, multi-vehicle collaboration demonstrates heightened efficacy in various tasks such as transportation, emergency response, surveying, mapping, and mining, often displaying superior efficiency and operational capabilities [4,5]. Advancements in internet of vehicles technology and vehicle-road collaboration technology facilitate the further development of communication technology, thereby presenting opportunities for coordinated control of multiple vehicles. The efficiency of multi-agent systems can be enhanced without compromising individual vehicle safety or efficiency. Consequently, this technology contributes to the enhancement of environmental quality and facilitates the promotion of sustainable transportation development [6].
In autonomous vehicle systems, the autonomous driving decision-making process comprises four hierarchical levels: path-planning, behavioral planning, motion planning, and local feedback control [7] (Figure 1). Path-planning involves the acquisition of real-time traffic data and road information, where network parameters are defined and an optimal path is computed through the utilization of path search algorithms.
At the behavioral planning stage, the vehicle’s onboard sensors detect other traffic participants and consider traffic regulations and additional information to formulate local path-planning like lane changes and vehicle following. Challenges encountered at this level include planning and tracking trajectories amidst uncertainty and limited information [8].
The subsequent phase in the short-term strategy, behavioral planning, takes into account factors such as vehicle dynamics to devise a rational path for the vehicle. At the lowest tier of local feedback control, adjustments and corrections to the vehicle’s movements are executed in real-time based on the vehicle’s sensor data and status information. This ensures that the vehicle adheres to the anticipated trajectory and behavior during motion.
Currently, technologies based on multi-vehicle collaboration integrate various research achievements in perception, communication, decision-making, and planning, forming relatively comprehensive but still rapidly developing research direction. Despite some reviews focusing on autonomous vehicle planning, they predominantly center around single-vehicle intelligence. There is currently no comprehensive literature review on the perception, communication, and planning aspects of multi-vehicle collaborative technologies. We categorize multi-vehicle collaborative motion planning techniques into rule-based collaborative methods and non-regular motion planning approaches. This paper aims to address this gap by comprehensively analyzing and summarizing the relevant research progress, with the hope of providing insights for the development of this field.
This article presents a comprehensive examination of multi-vehicle collaborative planning technology within the context of autonomous driving. Section 1 provides an overview of the motivation behind the study, reviews related work in the field, and outlines the contributions of the article. Section 2 discusses the development of rule-based multi-vehicle collaborative planning algorithms, addressing subtasks and game-theoretic issues in autonomous driving. Section 3 focuses on trajectory planning in multi-vehicle coordination, covering both vehicle platoons and non-platoon formations. This section provides a detailed overview of perception-based trajectory prediction, a crucial foundation for collaborative planning, and elaborates on the application characteristics of reinforcement learning in this field. Section 4 examines the challenges and solutions related to vehicle interactions in road traffic environments. Finally, Section 5 presents forward-looking recommendations for future research endeavors.

2. Rule-Based Multi-Vehicle Collaboration Planning

Much akin to human-driven vehicles, autonomous vehicles are bound by a plethora of regulations governing their operation. These regulations impose constraints on motion planning, necessitating that collaborative planning among vehicles consistently adheres to these constraints [9]. Rule-based multi-vehicle collaborative planning algorithms are crafted upon pre-defined rules and logic, enabling autonomous vehicles to execute safe and efficient path-planning. This algorithmic paradigm facilitates cooperative vehicle navigation by adhering to a shared set of rules, obviating the necessity for global optimization or centralized control. Consequently, such approaches enhance traffic efficiency, mitigate accident risks, and ensure the stable operation of vehicles within complex and dynamic traffic environments.

2.1. Behavior Selection in Multi-Vehicle Driving

Rule-based multi-vehicle collaboration algorithms typically derive from heuristic principles, encompassing strategies such as first in first out (FIFO), shortest path first (SPF), earliest arrival first (EAF), priority first, and minimum conflict first (MCF), among others. For instance, Ref. [10] employs FIFO rules to address the challenge of multi-vehicle cooperative merging at highway entrances. Building upon this, Ref. [11] proposes a rule-based adjustment algorithm, striking a favorable balance between computational complexity and traffic efficiency. This approach yields a near-optimal merging sequence, demonstrating applicability beyond highway entrances, including unsignalized intersections as highlighted by [12].
In current driving scenarios, scholars segment the research process into distinct subtasks, encompassing activities such as starting, lane keeping, and parking, among others, and formulate corresponding rules. Vehicles can then select behaviors by adhering to predefined switching rules [13]. Addressing the imperative of minimizing collision risks within structured road scenarios such as highways, Ref. [14] proposes a vehicle path-planning algorithm.
Fuzzy logic and expert systems are typical rule-based decision-making methods [15]. Fuzzy logic is a method for handling uncertainty and vagueness, allowing reasoning and decision-making in ambiguous environments. Ref. [16] utilizes fuzzy logic to process fuzzy input data between vehicles, considering variables such as distance from oncoming vehicles, own speed, and oncoming vehicle speed to determine the optimal lane-changing behavior. Ref. [17] employs fuzzy decision-making to facilitate the multi-vehicle merging process. Expert systems utilize domain knowledge and rules to coordinate actions between multiple vehicles. However, in practical operations, a large amount of accurate and detailed empirical knowledge is often required for coordination. The tendency to rely on precise empirical knowledge makes it ineffective in dealing with uncertainty in real-world scenarios.
The maturity of communication technology has facilitated advancements in rule-based multi-vehicle collaborative methods and enriched the completion of subtasks in driving environments. In recent years, road vehicle network technology has experienced rapid development. Dedicated short range communication (DSRC) technology operates in frequency bands ranging from 75 MHz to 5.9 GHz [18]. The technology of cellular vehicle-to-everything (C-V2X) communication, based on LTE-V networks, is characterized by large communication range, low latency, and frequent data exchange rates. Dynamic spectrum access (DSA) meets the growing bandwidth demands. These road communication systems can fulfill various communication tasks [19,20].
Based on V2V communication, a rule-based vehicle lane-changing control strategy proposed by [16] can smoothly navigate around sudden blockages on highways. Ref. [21] introduces a cooperative lane-changing model for scenarios with multiple congested lanes at intersections. Some following vehicles on the target lane decelerate to create enough space for coordinated maneuvers. However, the current state of communication is still immature, with instances of communication delays and interruptions. Ref. [22] presents an algorithm catering to both autonomous and non-autonomous vehicles. This algorithm, designed for mixed traffic scenarios, obviates the need for inter-vehicle communication and models various maneuvers utilizing empirical formulas.

2.2. Game Theory

In traditional algorithms, path-planning typically aims to devise an optimal or nearly optimal route from a set of origins to destinations. However, when multiple vehicles share the same origin and destination concurrently, employing this method may lead to congestion if all vehicles opt for the same recommended path [23]. Consequently, the calculated optimal solution ceases to be the most advantageous path in practice. This scenario is commonly referred to as a congestion game [24]. In order to solve this problem, Ref. [23] models the vehicle routing problem at the intersection as a population game and proposed a distributed cooperative routing algorithm (DCR) based on evolutionary games to achieve Nash equilibrium.
In traffic control, intersections are divided into signal-controlled intersections and uncontrolled intersections [25]. Coordinating multiple vehicles at uncontrolled intersections presents a more intricate challenge compared to signal-controlled intersections. Opting for overly conservative decisions may lead to vehicles becoming jammed and unable to traverse the intersection, while excessively aggressive decisions can increase the risk of collisions.
Leveraging vehicle-to-vehicle (V2V) communication, vehicle-to-infrastructure (V2I) communication, and a multi-vehicle collaborative planning strategy can significantly enhance the safety and efficiency of intersection traffic (Figure 2).
Drawing upon the concept of a zero-sum game, Ref. [26] investigates two collision-free control strategies for vehicles. Ref. [27] introduces a framework centered on a novel dynamic game formulation and progressive view optimization to capture vehicle interaction behavior at uncontrolled intersections. This framework explicitly incorporates the dynamic behavior of vehicles and adheres to common traffic rules. In addition to intersection scenarios, several researchers have explored multi-vehicle collaboration using game theory principles in highway settings [28,29,30,31]. For instance, Ref. [32] introduces a method based on game theory techniques for highways, proposing a novel collaborative driving prediction and planning framework tailored for highly automated driving in dynamic environments. This approach facilitates model-based interactive sensory motion prediction for all vehicles within the scene, enabling the planning of maneuver sequences over extended time scales. Furthermore, in the context of highway entrance ramp merging scenarios, an Intention-integrated Prediction- and Cost Function-based Algorithm (IPCB) framework is introduced to facilitate autonomous vehicles in executing multi-vehicle collaborative behaviors [33]. Additionally, Ref. [34] delves into the interaction and decision-making processes of autonomous vehicles with diverse driving characteristics, presenting a cooperative decision-making framework grounded in joint game theory to address the multi-lane merging challenges encountered by autonomous vehicles in multi-lane merging areas.
V2V and V2I communications undoubtedly provide great convenience for multi-vehicle collaborative driving. However, due to the high cost of the required infrastructure, network scalability issues and related security issues, there are still certain difficulties in the marketization of multi-vehicle collaborative driving. Moreover, self-driving vehicles must effectively collaborate and interact with vehicles lacking inter-vehicle communication capabilities. Challenges arise due to limitations and uncertainties in information sharing and communication among multiple vehicles, resulting in delays and gaps in vehicle-acquired information. The absence of accurate and real-time data may compromise the effectiveness of game theory methods.
Nevertheless, rule-based methods often exhibit limited applicability. They are tailored to specific scenarios and predefined traffic situations, thereby struggling to fully adapt to the intricacies of complex and dynamic traffic environments [35,36,37]. Consequently, in certain instances, the integration of other intelligent algorithms becomes imperative, such as deep reinforcement learning. In this regard, Ref. [38] merges rule-based algorithms with reinforcement learning, presenting an integrated decision-making framework (IDF) for autonomous vehicles (Figure 3). In situations where risks are identified within the decision-making network, caution rules are invoked to ensure driving safety. Similarly, manually encoding rules necessitate a substantial corpus of driving data to encompass all conceivable scenarios [39,40]. Addressing this limitation, Ref. [41] proposes a method that circumvents dependency on extensive driving data by employing a hierarchical reinforcement learning approach with labeled driving data. As such, the subsequent section delves into related research endeavors aimed at achieving greater flexibility and intelligence in multi-vehicle collaboration.
In decentralized multi-vehicle games, vehicles do not need to rely on infrastructure for communication; instead, they achieve their goals through autonomous decision-making and collaboration. Compared to centralized communication networks, decentralized multi-vehicle games not only reduce communication costs but also enhance system reliability and security. Recent work has applied decentralized methods [42,43]. On one hand, concurrent experience replay trajectory methods are used to share experiences among multiple agents and perform concurrent replays to address the issue of low learning efficiency. On the other hand, adjusting reward mechanisms and replay strategies enhances agents’ understanding of the environment and improves their behavior.
Additionally, in multi-vehicle collaborative planning, considerations such as overall traffic flow efficiency, congestion levels, and resource utilization require attention. Game theory methods often prioritize individual participants’ optimization goals, neglecting the optimization and equilibrium of the entire system. The pareto optimization theory provides a framework for optimizing resource allocation and utility distribution among multiple vehicles. It ensures the maximization of overall system welfare while considering the diverse goals and constraints of individual agents [44]. Ref. [45] introducing the Pareto optimal method addresses cooperative control problems in autonomous driving vehicles. Future advancements should integrate game theory approaches with system optimization methods to achieve a more comprehensive and cohesive framework for multi-vehicle collaborative planning. By incorporating system-level considerations into the decision-making process, such as optimizing traffic flow efficiency and resource allocation, the resulting methods can better address the complex dynamics of real-world traffic scenarios.

3. Non-Regular Multi-Vehicle Collaboration Planning

Accurate and efficient trajectory prediction for multi-vehicle autonomous driving is crucial in complex traffic scenarios. To accomplish driving environment subtasks, intelligent vehicles need to infer the precise future trajectories of neighboring vehicles. Therefore, trajectory prediction is a fundamental prerequisite for non-regular multi-vehicle collaboration planning [46]. Non-regular multi-vehicle collaboration is divided into scenarios involving coordinated motion between vehicle formations and those without vehicle formations. Cooperative motion between vehicle formations requires consideration of the mutual influence and interaction between vehicles to achieve coordinated motion [47,48]. Both scenarios necessitate predicting surrounding information in the environment to assess risks associated with vehicle behavior and subsequently take appropriate action [49,50,51].

3.1. Multi-Vehicle Trajectory Prediction Method

Compared to single autonomous vehicles, trajectory prediction in multi-vehicle scenarios is more complex. Particularly in urban environments, road information and interactions with traffic participants can influence the predictive information of the fleet (Shown in Figure 4). Road information encompasses both inherent road infrastructure details in maps (lanes, traffic lights) and corresponding traffic regulations (yielding to pedestrians, traffic light rules). Furthermore, interactions between vehicles and other traffic participants further increase the complexity.
Therefore, the accuracy of perception in the traffic environment forms the foundation of multi-vehicle collaborative planning. Currently, autonomous vehicles are equipped with heterogeneous sensor systems to provide ample environmental information, reducing issues related to unstable or missing information due to external factors. For instance, in environments where GNSS modules fail, optical flow modules can assist vehicles in position estimation. Radar, with its short-wave and high-penetration characteristics, exhibits strong robustness in rainy or snowy weather conditions. Additionally, it accurately perceives the surrounding environment in darkness, compensating for the shortcomings of camera perception. Hence, a multi-sensor fusion approach effectively complements perception information, acquiring as much valid information as possible for autonomous driving [52,53]. This supports trajectory prediction of surrounding vehicles (illustrated in Figure 5).
In multi-vehicle collaborative planning, the planning algorithm predicts the trajectories of other vehicles independently of its prediction range [54]. In complex traffic scenarios, various traffic participants take different traffic actions, causing the traffic environment to be dynamically changing. If an autonomous vehicle encounters an emergency, it needs to take proactive actions, such as slowing down to allow surrounding vehicles to move in, accelerating to change lanes to overtake, to ensure driving safety and efficiency. Therefore, the vehicle needs to accurately predict the location of adjacent vehicles. Future trajectories to allow risk assessment of vehicle behavior and further appropriate actions to be taken.
In the early days, it was assumed that the vehicle motion state was deterministic. The most primitive trajectory prediction method based on kinematics for multi-vehicle trajectory prediction was the single trajectory simulation method [55,56]. This method is usually based on a fixed prediction scenario or input condition, and cannot handle the uncertainties in the actual traffic environment, such as emergencies, or behavioral changes of other traffic participants. This may cause the planned trajectory to not adapt to the actual situation. Therefore, researchers research on trajectory prediction methods for vehicle motion uncertainty, such as the trajectory prediction method based on the Kalman filter [57], the Monte Carlo method [58], and Gaussian noise simulation. However, this method has poor performance in predicting vehicle motion trajectories in the vehicle operating environment.
In recent years, due to the development of positioning technology and sensor technology, the amount of vehicle natural driving trajectory data collection has been greatly improved compared to the past. Neural networks can automatically learn and capture complex features and patterns during vehicle driving by training on a large number of data samples. Trajectory prediction methods based on neural networks have been used in multi-vehicle collaborative planning [59,60,61,62,63]. As a variant of recurrent neural network (RNN), long-short term memory (LSTM) has a longer memory period and is suitable for long-term training of other vehicles. Tracking and trajectory prediction algorithms, which are favored by scholars, are used to track positions based on ranging sensors [64], identify driver intentions based on trajectory data [65], and predict future vehicle positions [66].
The essence of vehicle trajectory prediction lies in time series data regression [63]. Ref. [67] uses LSTM to predict trajectories for highway scenes. Ref. [66] expands the practicality of the former work and estimated the future positions of surrounding vehicles at a predetermined time. In order to improve the accuracy of prediction, Ref. [68] uses a priori trajectory position sequences to predict vehicle trajectories, which significantly improves the prediction performance and reduces computational losses.
While these methods exhibit noticeable efficacy in accomplishing their respective tasks, they nonetheless encounter challenges such as incomplete trajectory sequences or difficulties in establishing a simple probabilistic framework. Ref. [69] proposes a new vehicle trajectory analysis and prediction technology based on the LSTM encoder-decoder architecture. This method is able to generate future trajectories of surrounding vehicles based on the latest sensor measurement sequences.
However, simple LSTM models have two drawbacks: the inability to simultaneously model the spatial interaction between vehicles and trajectory sequences, and the possibility of encountering the vanishing gradient problem when modeling long-term sequences. Ref. [70] combines the characteristics of high-level spatio-temporal graphs with the sequence learning of RNN, and proposes structural-RNN, which constructs temporal relationships and spatial interactions into different time series to analyze the relationship between vehicles and trajectory sequences. Ref. [71] integrates the spatial expansion characteristics of convolutional neural network (CNNs) with the temporal expansion features of LSTM to forecast the trajectories of surrounding vehicles. The proposed method optimizes the model’s hyperparameters using the grid search algorithm to fulfill the dual precision prediction demands concerning space and time. Concerning the vanishing gradient issue encountered during modeling, Ref. [72] a trajectory prediction model based on spatio-temporal (ST-LSTM). This model incorporates shortcut connections between the input and output of each LSTM layer to directly convey historical trajectory information to subsequent layers. Such a structure aims to mitigate the vanishing gradient problem during backpropagation.
To acquire the aforementioned information and enhance the accuracy of multi-vehicle collaborative trajectory prediction models, existing research investigates various stages including the encoding-decoding phase [73,74], model optimization phase, and trajectory evaluation phase. The encoding phase captures interaction information between agents, laying the foundation for the model to understand the current state and interactions of agents. The model optimization phase improves prediction capability by refining the model structure and utilizing advanced training algorithms [75,76]. The trajectory evaluation phase focuses on assessing the output trajectories. However, there is limited research on trajectory prediction evaluation, with most studies modeling traffic agents as points without imposing constraints. Such unconstrained predictions may introduce significant uncertainty in predicting future states.

3.2. Cooperative Planning Method of Multi-Vehicle Formation

In multi-vehicle collaborative planning, through formation control, vehicles can move in specific geometric shapes or formations to achieve preset collective objectives. Formation control ensures that vehicles maintain reasonable relative positions and spacing, thus enabling efficient and coordinated collective actions. This is a specialized method deployed among vehicles with limited sensing and communication ranges [77]. In recent years, with the continuous iterative optimization of reinforcement learning algorithm models, deployment in the field of autonomous driving has also gradually matured.
Reinforcement learning does not rely on specific rules and models. Through the interaction between the agents and the environment, it can learn and optimize the behavior strategy of the vehicle to achieve the coordinated movement of multiple vehicles under autonomous driving. The Markov decision process describes of the relationship between the agent and the environment. A mathematical framework for the environment interaction process. In this process, the agent is in different states and selects an action to interact with the environment based on the current state. The environment transitions to the next state based on the agent’s action and the current state, subsequently providing the agent with a reward or punishment signal. This process is a sequential decision-making process, and the agent’s goal is to maximize the cumulative reward by selecting the optimal action sequence(Figure 6).
In autonomous vehicle path-planning using reinforcement learning, the Q-learning algorithm has been widely employed due to its ability to address stochastic transformations and reward dynamics without requiring adaptation to environmental changes. The Q-learning algorithm also has a certain degree of parallelism and scalability, and can be used in multiple environments. Collaborative learning and path-planning are performed between individual agents to achieve path-planning and decision-making in more complex traffic scenarios. Refs. [78,79] developed a multi-layered architecture for multi-vehicle formation control, utilizing a decentralized control framework for multi-agent systems and employing traditional Q-learning for vehicle path-planning. However, traditional Q-learning exhibits deficiencies such as inadequate exploration-exploitation balance and sluggish convergence rates in multi-vehicle collaboration scenarios [79]. To mitigate these shortcomings, Ref. [80] adjusts dynamic parameters through trial and error, enhancing the Q-learning algorithm by incorporating an action deletion mechanism to better reconcile exploration and exploitation objectives. Such approaches face challenges in complex environments where anticipated outcomes remain elusive. Subsequently, Ref. [81] introduces a novel update method for the Q-function, proposing an enhanced Q-learning algorithm tailored for multi-vehicle cooperative obstacle avoidance in static environments. This refined algorithm facilitates the identification of optimal paths within the same iteration, resulting in shortened final path trajectories.
In recent years, researchers have shown significant interest in the task of path-planning for autonomous multi-vehicle “platoons”. Ref. [82] introduces a novel approach known as distributed receding horizon control (DRHC) architecture, which integrates nonlinear platoon dynamics with a deep reinforcement learning (DRL) scheme. This innovative framework addresses the path-planning challenges encountered by autonomous multi-vehicle platoons navigating urban road networks. By leveraging DRL for route decision-making, the DRHC architecture ensures the safe navigation of the platoon towards its target destination while preventing collisions between adjacent vehicles. Similarly, Ref. [83] recognizes the presence of non-autonomous vehicles within traffic flow and proposes a motion decision-making algorithm based on Nash-Q learning. This algorithm is specifically designed to account for the interactive dynamics between autonomous and non-autonomous vehicles. The decision-making framework is tailored for the study of multi-agent systems, with Q-learning serving as the foundational motion decision-making model, capable of determining optimal state-action sequences for autonomous vehicles.
As the number of vehicles increases and traffic flows become more complex, the importance of information exchange and cooperation among multiple vehicles becomes paramount. Through information exchange, vehicles can share real-time traffic conditions, road conditions, and expected actions, enabling more efficient path-planning and decision-making. This collaborative mechanism can assist vehicles in avoiding traffic jams, reducing traffic accidents, and optimizing overall traffic mobility. Ref. [84] employs a deep Q network (DQN) approach to tackle the hierarchical collaborative multi-vehicle tracking problem. By making decisions grounded in encoded states, DQN dynamically adjusts individual decisions based on real-time traffic information, thus deriving an optimal dataset of collective actions. Building upon this methodology, Ref. [85] introduces a collision avoidance algorithm tailored for multi-vehicle systems, effectively addressing scenarios characterized by high vehicle density and unrestricted driving directions. This approach strategically plans routes, thereby mitigating detour inconveniences and equitably distributing traffic flow.

3.3. Multi-Vehicle Collaborative Planning without Formation

As the development of self-driving cars advances, the emergence of hybrid autonomous driving environments, characterized by the coexistence of autonomous vehicles and human-driven counterparts, becomes increasingly likely. Given the unpredictable behavior of human drivers regarding cooperation with autonomous vehicles, the path-planning strategies of autonomous vehicles require meticulous consideration of the dynamics and roles of surrounding vehicles. In multi-vehicle collaboration scenarios, vehicles often contend for scarce resources such as the right of way or road access. Failure to accurately ascertain priority may precipitate contention conflicts, thereby impeding the seamless progression of vehicle operations.
In multi-agent reinforcement learning, the training cycle changes dynamically, and classic Q-learning and other methods cannot solve the problem of unstable agent environment in the system. Thus, the introduction of policy optimization methods becomes imperative to enhance the efficacy of large-scale multi-agent learning, particularly in environments characterized by non-stationarity. Ref. [86] employs deep deterministic policy gradients (DDPG) to facilitate the acquisition of cooperative overtaking strategies among multiple vehicles within simulated highway scenarios. Notably, the reward function is designed to leverage original sensor data from the current time step to inform decision-making processes. By driving agents to explore their environment and circumvent local optima, the RND-DDPG algorithm expedites algorithmic convergence, thereby enhancing overall learning efficiency.
The actor-critic algorithm, predicated on the fusion of value function and strategy optimization, introduces two distinct neural networks: the action network and the critic network. Ref. [87] delves into the mechanisms through which a decentralized reward structure fosters altruistic behaviors among vehicles. To incentivize vehicles to consider the welfare of both autonomous and human-driven counterparts, they introduces a synchronous Advantage Actor-Critic algorithm. Figure 7 illustrates diverse simulation outcomes of autonomous vehicles employing self-serving versus altruistic strategies during lane merging maneuvers.
With the rapid advancements witnessed in artificial intelligence technology, swarm intelligence algorithms have found widespread application in path-planning tasks for unmanned vehicles. Prominent among these intelligent bionic algorithms are particle swarm optimization (PSO) [88,89,90,91], ant colony algorithms [92], genetic algorithms [93,94], and artificial neural network algorithms. Leveraging their biomimetic attributes, these methods exhibit enhanced intelligence and efficiency. Swarm intelligence algorithms, inspired by natural processes, replicate cooperative behaviors observed in decentralized self-organizing systems to achieve purposeful activities. These algorithms excel in addressing complex optimization challenges that are difficult for conventional mathematical methods, particularly those characterized by computational complexities or uncertainties from stochastic processes. However, they are susceptible to encountering issues in path-planning applications, including but not limited to, the pitfalls of local optimal solutions and slow convergence rates.
In a concerted effort to address path-planning and coordination intricacies inherent in multiple autonomous driving scenarios, Ref. [95] introduces a probability-based continuous ant colony optimization method, coupled with a random walk strategy and adaptive waypoint rectification mechanism. This approach meticulously optimizes the trajectory of each unmanned vehicle. Subsequently, the conundrum of obstacle avoidance within a multi-agent coordination framework is deftly resolved through the application of a speed transformation optimization algorithm. Furthermore, recognizing the paramount importance of mitigating energy consumption in multi-vehicle collaborative endeavors, Ref. [96] integrates the ant-based algorithm with a predictive model for power/energy consumption. This innovative fusion introduces the concept of green ant (G-Ant), offering a tailored solution for multi-vehicle collaboration. It facilitates collision-free, energy-efficient shortest path-planning strategies.
Compared to alternative algorithms, the particle swarm optimization (PSO) algorithm demonstrates a significant advantage in swiftly converging towards solutions in task allocation scenarios. In the context of autonomous driving challenges, multi-vehicle task allocation stands out as a pivotal concern. Ref. [97] formulates a comprehensive mathematical model encompassing pertinent constraints and functions relevant to both vehicles and targets. Moreover, this study introduces an innovative particle encoding and decoding methodology aimed at enhancing the performance of the particle swarm algorithm, thus facilitating the derivation of feasible multi-task allocation solutions. Nevertheless, a noteworthy drawback of the PSO algorithm lies in its propensity to prematurely converge towards local optima. Despite this limitation, recent endeavors have seen select researchers endeavoring to mitigate this shortcoming through the integration of PSO into multi-vehicle task distribution frameworks.

4. Future Development and Discussion

Existing multi-vehicle autonomous driving cooperative technology has laid the foundation for enhancing the autonomy of future smart city traffic. However, many key technologies still need further development, and there is still a long way to go for autonomous driving vehicles to form a transportation system. The preceding text has conducted a detailed analysis and discussion of the intelligent key technologies of autonomous driving vehicles. This section will expound on future development trends and the obstacles faced, and propose corresponding potential solutions based on experience.

4.1. Environmental Uncertainty

In multi-vehicle collaborative planning, the interaction and collaboration among vehicles are significantly more complex compared to single-vehicle scenarios. The intentions and dynamic behaviors of other vehicles serve as crucial elements in this interaction, contributing to increased complexity and uncertainty in the planning process. Additionally, factors such as traffic flow dynamics, variations in road conditions, and changes in the state of traffic signals, among other environmental variables, undergo temporal and spatial alterations. These dynamic transformations profoundly impact parameters such as vehicular velocities, trajectory selection, and overall traffic conduct, thereby posing formidable challenges to the effectiveness of multi-vehicle collaborative planning efforts.
In uncertain environments, the uncertainty surrounding the future trajectories of other road users can be attributed to several factors. Firstly, the intentions of other vehicles and traffic participants are often unknown, contributing to the unpredictability of their actions. Secondly, the lack of precise longitudinal predictions regarding the future movements of other vehicles further compounds this uncertainty. Additionally, the probabilistic nature of interactions between other vehicles and the target vehicle introduces another layer of unpredictability. Moreover, limitations inherent in sensor technology utilized for data collection and perception also contribute to the uncertainty surrounding future trajectories [54].
Addressing the challenge of uncertainty pertaining to the future trajectories of other road users, Ref. [98] leverages a partially observable Markov decision process to estimate the behavior of vehicles or pedestrians based on data gathered from noise sensors. Integration of this information into path-planning methodologies enables the realization of safe navigation in uncertain environments.
Multi-vehicle collaboration involves the states and behaviors of multiple vehicles, causing the state space to grow exponentially. Vehicles must navigate high-dimensional state spaces to seek optimal solutions, imposing greater challenges on search algorithms and computational efficiency requirements. The decisions and actions of each vehicle will affect the status and behavior of other vehicles, causing uncertainty to spread and diffuse in the fleet. The interaction between vehicles amplifies the complexity and instability of the entire system. Vehicles must take into account the behaviors and potential decisions of other vehicles to mitigate potential conflicts and collisions.
The vehicle’s own state is also a crucial factor influencing safe driving. The dynamic state of autonomous vehicles varies with changes in the target trajectory, vehicle speed, and front wheel steering angle. Trajectory-tracking control is essential for ensuring the safe and accurate tracking of the target trajectory by intelligent vehicles [99,100]. Tracking control for vehicle platooning is achieved by dispersing the longitudinal dynamic model of individual vehicles [101]. However, there is currently limited research on multi-vehicle trajectory tracking control. Most studies still focus on single-vehicle control, overlooking the interactive effects between vehicles.
In addition, pedestrians play a significant role in road traffic scenarios. Unlike vehicles, pedestrians have unpredictable movement patterns and behaviors, making them challenging for autonomous vehicles to safely navigate around. Typically, autonomous vehicles perceive pedestrians as dynamic obstacles and plan their trajectories to avoid potential collisions within the pedestrians’ movement areas. For instance, understanding pedestrian intent and behavior can help vehicles anticipate their movements and plan safer trajectories accordingly. Moreover, incorporating pedestrian awareness into the decision-making process of autonomous vehicles can lead to more efficient navigation strategies, reducing the risk of accidents and improving overall traffic flow [102].

4.2. Vehicle Collision

In complex urban traffic environments, vehicle conflicts pose a significant challenge. Multi-vehicle conflicts commonly arise in specific road segments or intersections, particularly in congested sections, intersections, highways, parking lots, and other densely trafficked areas (Figure 8). These conflicts, which may not necessarily culminate in collisions, often lead to traffic congestion, reduced efficiency, or undesirable traffic behaviors.
At intersections, conflicts can arise when multiple vehicles arrive simultaneously and attempt to traverse the intersection without well-defined coordination and priority rules. For example, if vehicles such as car A and car B intend to make conflicting maneuvers, such as left and right turns, respectively, within the limited green light duration, simultaneous entry into the intersection can lead to conflicts. Similarly, on highway merge lanes, multiple vehicles attempting to merge onto the main thoroughfare without coordination may result in frequent lane changes, decelerations, and accelerations, leading to traffic congestion and accidents. Human factors, excessive traffic volumes, road conditions, and weather represent key contributors to vehicle conflict instances, exerting an impact on the efficacy of multi-vehicle cooperation efforts.
In highly constrained spaces like parking lots, vehicles are particularly susceptible to model errors. Furthermore, these spatial environments are non-convex, and relying on geometric approximations can quickly deplete the available free space. Ref. [103] conceptualizes the self-driving multi-vehicle parking problem as a symmetric mixed integer linear optimization, with a focus on optimizing intentions and driving styles. Furthermore, tackling the inherent nonlinear and non-convex nature of this optimal control problem, conflict resolution is modeled in [104] as a multi-agent partially observable Markov decision process. This framework explores effective conflict resolution strategies within constrained spaces, generating collision-free and kinematically feasible paths. While reinforcement learning has attracted considerable attention in the field of autonomous driving, it still faces numerous challenges in practical applications. The black-box nature of reinforcement learning may lead vehicles to adopt unsafe measures in unforeseen circumstances. Although many studies have employed rule-based constraints to prevent reinforcement learning from executing unsafe strategies, achieving a balance between learned optimal strategies and safety remains a significant challenge.

4.3. Sustainable Development

Intelligent transportation systems have the potential to enhance traffic efficiency, alleviate congestion, and mitigate emissions, consequently reducing adverse environmental effects. Achieving intelligent transportation systems relies heavily on the advancement of autonomous systems, which contribute significantly to the development of autonomous driving and multi-agent control strategies for establishing sustainable and coordinated traffic management. Autonomous vehicles offer notable advantages in various key parameters, including reduced energy intensity, fuel consumption, emissions, travel times, and congestion levels.
Decision-making conflicts and competition between vehicles may lead to an increase in energy consumption. Energy conservation and management require a large amount of vehicle operating data and energy consumption information for modeling and optimization. However, obtaining real-time and accurate vehicle energy consumption data may face difficulties, especially when it comes to data sharing between different models and manufacturers. The lack of sufficient data may limit energy consumption research and practical applications. In multi-vehicle collaboration, conflicts can arise between optimizing energy consumption and minimizing travel time. Finding a balance between energy and time considerations is a crucial challenge to address. One approach to tackle this issue is by leveraging the Markov chain model to implement road energy consumption modeling. This involves considering various factors such as different driving modes, the agent’s surrounding environment, road conditions, and applicable limitations. By incorporating these factors and employing suitable learning strategies, the energy demand of the driving path can be estimated based on input parameters and model-based predictions [105].
Interactions between vehicles and pedestrians are inevitable in future scenarios. Ethical dilemmas faced by autonomous decision-makers will also constrain the development of smart cities. Incorporating ethical considerations into multi-vehicle decision-making frameworks, such as setting objective functions and costs within ethical systems, is a thought-provoking issue that requires careful consideration. In addition, public acceptance of autonomous vehicles is also a significant factor influencing their sustainable development. According to [106], technological, normative, and ethical problems all affect public acceptability. Relying solely on technological progress to address public concerns is insufficient. Addressing societal perceptions begins with ensuring transparency in the development and deployment of autonomous systems. Additionally, normative issues such as legal frameworks and regulatory standards must be established to govern the use of autonomous vehicles. Under the premise of ensuring safety, autonomous vehicles can be allowed to operate under common driving conditions with behavior patterns similar to human-driven vehicles [107].

5. Conclusions

This collaborative planning paradigm plays a crucial role in aiding vehicles to circumvent collisions in intricate traffic environments, optimize traffic flow, and enhance driving efficiency and safety. With the vigorous development of vehicle technology, the vision of smart cities led by intelligent autonomous vehicles may soon become a reality. Considering the complexity of multi-vehicle cooperative systems, future research must emphasize breakthroughs in intelligent and information technologies, in addition to highlighting planning studies among multiple vehicles. Our literature review has revealed a wealth of technology in multi-vehicle collaborative planning. We aim to showcase the diversity and potential of multi-vehicle cooperative planning. Through these planning efforts, we hope readers can further explore research in related fields.

Author Contributions

R.M. and J.G. were responsible for the finalization and adjustment of the overall framework, and provided guidance on the challenges facing the paper in the future. S.R. was responsible for some literature research and article writing. P.C. and Z.Q. were responsible for some literature research and mapping. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Research on Key Technologies for Big Data Intelligent Analysis and Decision Making in 5G Smart Expressway Operation (2022YFSJ0040) and Research on Big Data Intelligent Analysis Technology for Road Conditions in the Hu-Bao Section of theG6 Expressway (JY20220170).

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Authors Pengfei Cui and Zhi Qiao were employed by the company Inner Mongolia High-Grade Highway Construction and Development Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Ahangari, H.; Outlaw, J.; Atkinson-Palombo, C.; Garrick, N.W. Investigation into impact of fluctuations in gasoline prices and macroeconomic conditions on road safety in developed countries. Transp. Res. Rec. 2014, 2465, 48–56. [Google Scholar] [CrossRef]
  2. Hitchcock, G.; Conlan, B.; Kay, D.; Brannigan, C.; Newman, D. Air Quality and Road Transport. London: Impacts and Solutions; RAC Foundation: London, UK, 2014. [Google Scholar]
  3. Shi, L.; Chen, H. A Collaborative Control Scheme for Smart Vehicles based on Multi-agent Deep Reinforcement Learning. IEEE Access 2023, 11, 96221–96234. [Google Scholar] [CrossRef]
  4. Alonso-Mora, J.; Baker, S.; Rus, D. Multi-robot formation control and object transport in dynamic environments via constrained optimization. Int. J. Robot. Res. 2017, 36, 1000–1021. [Google Scholar] [CrossRef]
  5. Liu, S.; Mohta, K.; Shen, S.; Kumar, V. Towards collaborative mapping and exploration using multiple micro aerial robots. In Experimental Robotics: The 14th International Symposium on Experimental Robotics; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 865–878. [Google Scholar]
  6. Kolat, M.; Kővári, B.; Bécsi, T.; Aradi, S. Multi-agent reinforcement learning for traffic signal control: A cooperative approach. Sustainability 2023, 15, 3479. [Google Scholar] [CrossRef]
  7. Aradi, S. Survey of deep reinforcement learning for motion planning of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2020, 23, 740–759. [Google Scholar] [CrossRef]
  8. Lian, F.L.; Murray, R. Real-time trajectory generation for the cooperative path planning of multi-vehicle systems. In Proceedings of the 41st IEEE Conference on Decision and Control, 2002, Las Vegas, NV, USA, 10–13 December 2002; Volume 4, pp. 3766–3769. [Google Scholar]
  9. Schwarting, W.; Alonso-Mora, J.; Rus, D. Planning and decision-making for autonomous vehicles. Annu. Rev. Control. Robot. Auton. Syst. 2018, 1, 187–210. [Google Scholar] [CrossRef]
  10. Rios-Torres, J.; Malikopoulos, A.A. Automated and cooperative vehicle merging at highway on-ramps. IEEE Trans. Intell. Transp. Syst. 2016, 18, 780–789. [Google Scholar] [CrossRef]
  11. Ding, J.; Li, L.; Peng, H.; Zhang, Y. A rule-based cooperative merging strategy for connected and automated vehicles. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3436–3446. [Google Scholar] [CrossRef]
  12. Malikopoulos, A.A.; Cassandras, C.G.; Zhang, Y.J. A decentralized energy-optimal control framework for connected automated vehicles at signal-free intersections. Automatica 2018, 93, 244–256. [Google Scholar] [CrossRef]
  13. Wang, Q.; Weiskircher, T.; Ayalew, B. Hierarchical hybrid predictive control of an autonomous road vehicle. Dynamic Systems and Control Conference. Am. Soc. Mech. Eng. 2015, 57267, V003T50A006. [Google Scholar]
  14. Glaser, S.; Vanholme, B.; Mammar, S.; Gruyer, D.; Nouveliere, L. Maneuver-based trajectory planning for highly autonomous vehicles on real road with traffic and driver interaction. IEEE Trans. Intell. Transp. Syst. 2010, 11, 589–606. [Google Scholar] [CrossRef]
  15. Chen, S.; Hu, X.; Zhao, J.; Wang, R.; Qiao, M. A Review of Decision-Making and Planning for Autonomous Vehicles in Intersection Environments. World Electr. Veh. J. 2024, 15, 99. [Google Scholar] [CrossRef]
  16. Asano, S.; Ishihara, S. Safe, Smooth, and Fair Rule-Based Cooperative Lane Change Control for Sudden Obstacle Avoidance on a Multi-Lane Road. Appl. Sci. 2022, 12, 8528. [Google Scholar] [CrossRef]
  17. Guo, P.; Hua, Y.; Wang, T.; Gong, J. Multi-vehicle mixed formation driving method based on fuzzy decision. J. Phys. Conf. Ser. 2021, 1802, 042011. [Google Scholar] [CrossRef]
  18. Kenney, J.B. Dedicated short-range communications (DSRC) standards in the United States. Proc. IEEE 2011, 99, 1162–1182. [Google Scholar] [CrossRef]
  19. Ahmad, I.; Noor, R.; Z’aba, M.R. LTE efficiency when used in traffic information systems: A stable interest aware clustering. Int. J. Commun. Syst. 2019, 32, e3853. [Google Scholar] [CrossRef]
  20. Zhang, X.; Huang, J.; Huang, Y.; Huang, K.; Yang, L.; Han, Y.; Wang, L.; Liu, H.; Luo, J.; Li, J. Intelligent amphibious ground-aerial vehicles: State of the art technology for future transportation. IEEE Trans. Intell. Veh. 2022, 8, 970–987. [Google Scholar] [CrossRef]
  21. Backfrieder, C.; Ostermayer, G.; Lindorfer, M.; Mecklenbräuker, C.F. Cooperative lane-change and longitudinal behaviour model extension for TraffSim. In Proceedings of the Smart Cities: First International Conference, Smart-CT 2016, Málaga, Spain, 15–17 June 2016; Proceedings 1. Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  22. Kala, R.; Warwick, K. Motion planning of autonomous vehicles in a non-autonomous vehicle environment without speed lanes. Eng. Appl. Artif. Intell. 2013, 26, 1588–1601. [Google Scholar] [CrossRef]
  23. Lu, J.; Li, J.; Yuan, Q.; Chen, B. A multi-vehicle cooperative routing method based on evolutionary game theory. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 987–994. [Google Scholar]
  24. Kok, A.L.; Hans, E.W.; Schutten, J.M.J. Vehicle routing under time-dependent travel times: The impact of congestion avoidance. Comput. Oper. Res. 2012, 39, 910–918. [Google Scholar] [CrossRef]
  25. McGee, H.W.; Soika, J.; Fiedler, R.; Albee, M.; Holzem, A.; Eccles, K.; Quinones, J. Unsignalized Intersection Improvement Guide. Available online: https://toolkits.ite.org/uiig/ (accessed on 20 May 2024).
  26. Xing, Y.; Lv, C.; Mo, X.; Hu, Z.; Huang, C.; Hang, P. Toward safe and smart mobility: Energy-aware deep learning for driving behavior analysis and prediction of connected vehicles. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4267–4280. [Google Scholar] [CrossRef]
  27. Li, N.; Yao, Y.; Kolmanovsky, I.; Atkins, E.; Girard, A.R. Game-theoretic modeling of multi-vehicle interactions at uncontrolled intersections. IEEE Trans. Intell. Transp. Syst. 2020, 23, 1428–1442. [Google Scholar] [CrossRef]
  28. Wu, B.; Qian, L.; Lu, M.; Qiu, D.; Liang, H. Optimal control problem of multi-vehicle cooperative autonomous parking trajectory planning in a connected vehicle environment. IET Intell. Transp. Syst. 2019, 13, 1677–1685. [Google Scholar] [CrossRef]
  29. Oyler, D.W.; Yildiz, Y.; Girard, A.R.; Li, N.I.; Kolmanovsky, I.V. A game theoretical model of traffic with multiple interacting drivers for use in autonomous vehicle development. In Proceedings of the 2016 American Control Conference (ACC), Boston, MA, USA, 6–8 July 2016; pp. 1705–1710. [Google Scholar]
  30. Li, N.; Oyler, D.; Zhang, M.; Yildiz, Y.; Girard, A.; Kolmanovsky, I. Hierarchical reasoning game theory based approach for evaluation and testing of autonomous vehicle control systems. In Proceedings of the 2016 IEEE 55th Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 727–733. [Google Scholar]
  31. Li, N.; Oyler, D.W.; Zhang, M.; Yildiz, Y.; Kolmanovsky, I.; Girard, A.R. Game theoretic modeling of driver and vehicle interactions for verification and validation of autonomous vehicle control systems. IEEE Trans. Control. Syst. Technol. 2017, 26, 1782–1797. [Google Scholar] [CrossRef]
  32. Bahram, M.; Lawitzky, A.; Friedrichs, J.; Aeberhard, M.; Wollherr, D. A game-theoretic approach to replanning-aware interactive scene prediction and planning. IEEE Trans. Veh. Technol. 2015, 65, 3981–3992. [Google Scholar] [CrossRef]
  33. Wei, J.; Dolan, J.M.; Litkouhi, B. Autonomous vehicle social behavior for highway entrance ramp management. In Proceedings of the 2013 IEEE Intelligent Vehicles Symposium (IV), Gold Coast, Australia, 23–26 June 2013; pp. 201–207. [Google Scholar]
  34. Hang, P.; Lv, C.; Huang, C.; Xing, Y.; Hu, Z. Cooperative decision making of connected automated vehicles at multi-lane merging zone: A coalitional game approach. IEEE Trans. Intell. Transp. Syst. 2021, 23, 3829–3841. [Google Scholar] [CrossRef]
  35. Li, G.; Yang, Y.; Qu, X. Deep learning approaches on pedestrian detection in hazy weather. IEEE Trans. Ind. Electron. 2019, 67, 8889–8899. [Google Scholar] [CrossRef]
  36. Duan, J.; Li, R.; Hou, L.; Wang, W.; Li, G.; Li, S.E.; Cheng, B.; Gao, H. Driver braking behavior analysis to improve autonomous emergency braking systems in typical Chinese vehicle-bicycle conflicts. Accid. Anal. Prev. 2017, 108, 74–82. [Google Scholar] [CrossRef]
  37. Hou, L.; Duan, J.; Wang, W.; Li, R.; Li, G.; Cheng, B. Drivers’ braking behaviors in different motion patterns of vehicle-bicycle conflicts. J. Adv. Transp. 2019, 2019, 4023970. [Google Scholar] [CrossRef]
  38. Xu, C.; Zhao, W.; Liu, J.; Wang, C.; Lv, C. An integrated decision-making framework for highway autonomous driving using combined learning and rule-based algorithm. IEEE Trans. Veh. Technol. 2022, 71, 3621–3632. [Google Scholar] [CrossRef]
  39. Paden, B.; Cap, M.; Yong, S.Z.; Yershov, D.; Frazzoli, E. A survey of motion planning and control techniques for self-driving urban vehicles. IEEE Trans. Intell. Veh. 2016, 1, 33–55. [Google Scholar] [CrossRef]
  40. Bojarski, M.; Yeres, P.; Choromanska, A.; Choromanski, K.; Firner, B.; Jackel, L.; Muller, U. Explaining how a deep neural network trained with end-to-end learning steers a car. arXiv 2017, arXiv:1704.07911. [Google Scholar]
  41. Duan, J.; Li, S.E.; Guan, Y.; Sun, Q.; Cheng, B. Hierarchical reinforcement learning for self-driving decision-making without reliance on labelled driving data. IET Intell. Transp. Syst. 2020, 14, 297–305. [Google Scholar] [CrossRef]
  42. Yadav, P.; Mishra, A.; Kim, S. A comprehensive survey on multi-agent reinforcement learning for connected and automated vehicles. Sensors 2023, 23, 4710. [Google Scholar] [CrossRef]
  43. Omidshafiei, S.; Pazis, J.; Amato, C.; How, J.P.; Vian, J. Deep decentralized multi-task multi-agent reinforcement learning under partial observability. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2681–2690. [Google Scholar]
  44. Yan, Y.; Du, H.; He, D.; Li, W. Pareto optimal information flow topology for control of connected autonomous vehicles. IEEE Trans. Intell. Veh. 2022, 8, 330–343. [Google Scholar] [CrossRef]
  45. Liang, J.; Li, Y.; Yin, G.; Xu, L.; Lu, Y.; Feng, J.; Shen, T.; Cai, G. A MAS-based hierarchical architecture for the cooperation control of connected and automated vehicles. IEEE Trans. Veh. Technol. 2022, 72, 1559–1573. [Google Scholar] [CrossRef]
  46. Wang, Y.; Zhao, S.; Zhang, R.; Cheng, X.; Yang, L. Multi-vehicle collaborative learning for trajectory prediction with spatio-temporal tensor fusion. IEEE Trans. Intell. Transp. Syst. 2020, 23, 236–248. [Google Scholar] [CrossRef]
  47. Yuanchang, L.; Bucknall, R. A survey of formation control and motion planning of multiple unmanned vehicles. Robotica 2018, 36, 1019–1047. [Google Scholar]
  48. Duan, X.; Sun, C.; Tian, D.; Zhou, J.; Cao, D. Cooperative lane-change motion planning for connected and automated vehicle platoons in multi-lane scenarios. IEEE Trans. Intell. Transp. Syst. 2023, 24, 7073–7091. [Google Scholar] [CrossRef]
  49. Kuang, J.; Tan, G.; Guo, X.; Pei, X.; Peng, D. Research of obstacle vehicles avoidance for automated heavy vehicle platoon by switching the formation. IET Intell. Transp. Syst. 2024, 18, 630–644. [Google Scholar] [CrossRef]
  50. Xu, B.; Wang, G.; Yang, Z.; Bian, Y.; Wang, X.; Hu, M. Multi-Vehicle Collaborative Trajectory Planning in Unstructured Conflict Areas Based on V-Hybrid A. IEEE Trans. Intell. Transp. Syst. 2024, 1–14. [Google Scholar] [CrossRef]
  51. Liu, Y.; Long, K.; Wu, W.; Liu, W. Multi-Vehicle Collaborative Trajectory Planning for Emergency Vehicle Priority at Autonomous Intersections. Transp. Res. Rec. 2024, 03611981241235229. [Google Scholar] [CrossRef]
  52. Cesetti, A.; Frontoni, E.; Mancini, A.; Zingaretti, P.; Longhi, S. A vision-based guidance system for UAV navigation and safe landing using natural landmarks. J. Intell. Robot. Syst. 2010, 57, 233–257. [Google Scholar] [CrossRef]
  53. Ho, H.W.; de Croon, G.C.; Van Kampen, E.; Chu, Q.P.; Mulder, M. Adaptive gain control strategy for constant optical flow divergence landing. IEEE Trans. Robot. 2018, 34, 508–516. [Google Scholar] [CrossRef]
  54. Hubmann, C.; Schulz, J.; Becker, M.; Althoff, D.; Stiller, C. Automated driving in uncertain environments: Planning with interaction and uncertain maneuver prediction. IEEE Trans. Intell. Veh. 2018, 3, 5–17. [Google Scholar] [CrossRef]
  55. Schubert, R.; Richter, E.; Wanielik, G. Comparison and evaluation of advanced motion models for vehicle tracking. In Proceedings of the 2008 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–6. [Google Scholar]
  56. Schubert, R.; Adam, C.; Obst, M.; Mattern, N.; Leonhardt, V.; Wanielik, G. Empirical evaluation of vehicular models for ego motion estimation. In Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), Baden-Baden, Germany, 5–9 June 2011; pp. 534–539. [Google Scholar]
  57. Veeraraghavan, H.; Papanikolopoulos, N.; Schrater, P. Deterministic sampling-based switching kalman filtering for vehicle tracking. In Proceedings of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 1340–1345. [Google Scholar]
  58. Broadhurst, A.; Baker, S.; Kanade, T. Monte Carlo road safety reasoning. In Proceedings of the IEEE Proceedings, Intelligent Vehicles Symposium, 2005, Las Vegas, NV, USA, 6–8 June 2005; pp. 319–324. [Google Scholar]
  59. Zaman, M.; Zohrabi, N.; Abdelwahed, S. A CNN-based path trajectory prediction approach with safety constraints. In Proceedings of the 2020 IEEE Transportation Electrification Conference & Expo (ITEC), Chicago, IL, USA, 23–26 June 2020; pp. 267–272. [Google Scholar]
  60. Ding, W.; Chen, J.; Shen, S. Predicting vehicle behaviors over an extended horizon using behavior interaction network. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 8634–8640. [Google Scholar]
  61. Cui, H.; Radosavljevic, V.; Chou, F.C.; Lin, T.H.; Nguyen, T.; Huang, T.K.; Schneider, J.; Djuric, N. Multimodal trajectory predictions for autonomous driving using deep convolutional networks. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 2090–2096. [Google Scholar]
  62. Phillips, D.J.; Wheeler, T.A.; Kochenderfer, M.J. Generalizable intention prediction of human drivers at intersections. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 1665–1670. [Google Scholar]
  63. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  64. Ondruska, P.; Posner, I. Deep tracking: Seeing beyond seeing using recurrent neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  65. Khosroshahi, A.; Ohn-Bar, E.; Trivedi, M.M. Surround vehicles trajectory analysis with recurrent neural networks. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 2267–2272. [Google Scholar]
  66. Kim, B.; Kang, C.M.; Kim, J.; Lee, S.H.; Chung, C.C.; Choi, J.W. Probabilistic vehicle trajectory prediction over occupancy grid map via recurrent neural network. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 399–404. [Google Scholar]
  67. Altché, F.; de La Fortelle, A. An LSTM network for highway trajectory prediction. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 353–359. [Google Scholar]
  68. IIp, A.; Irio, L.; Oliveira, R. Vehicle trajectory prediction based on LSTM recurrent neural networks. In Proceedings of the 2021 IEEE 93rd Vehicular Technology Conference (VTC2021-Spring), Helsinki, Finland, 25–28 April 2021; pp. 1–5. [Google Scholar]
  69. Park, S.H.; Kim, B.; Kang, C.M.; Chung, C.C.; Choi, J.W. Sequence-to-sequence prediction of vehicle trajectory via LSTM encoder-decoder architecture. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1672–1678. [Google Scholar]
  70. Jain, A.; Zamir, A.R.; Savarese, S.; Saxena, A. Structural-rnn: Deep learning on spatio-temporal graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5308–5317. [Google Scholar]
  71. Xie, G.; Shangguan, A.; Fei, R.; Ji, W.; Ma, W.; Hei, X. Motion trajectory prediction based on a CNN-LSTM sequential model. Sci. China Inf. Sci. 2020, 63, 212207. [Google Scholar] [CrossRef]
  72. Dai, S.; Li, L.; Li, Z. Modeling vehicle interactions via modified LSTM models for trajectory prediction. IEEE Access 2019, 7, 38287–38296. [Google Scholar] [CrossRef]
  73. Zeng, J.; Ren, Y.; Wang, K.; Hu, X.; Li, J. Spatio-Temporal-Attention-Based Vehicle Trajectory Prediction Considering Multi-Vehicle Interaction in Mixed Traffic Flow. Appl. Sci. 2023, 14, 161. [Google Scholar] [CrossRef]
  74. Luo, W.; Park, C.; Cornman, A.; Sapp, B.; Anguelov, D. Jfp: Joint future prediction with interactive multi-agent modeling for autonomous driving. In Proceedings of the Conference on Robot Learning, Atlanta, GA, USA, 6 November 2023; pp. 1457–1467. [Google Scholar]
  75. Bhattacharyya, P.; Huang, C.; Czarnecki, K. Ssl-lanes: Self-supervised learning for motion forecasting in autonomous driving. In Proceedings of the Conference on Robot Learning, Atlanta, GA, USA, 6 November 2023; pp. 1793–1805. [Google Scholar]
  76. Wu, X.; Chandra, R.; Guan, T.; Bedi, A.; Manocha, D. Intent-Aware Planning in Heterogeneous Traffic via Distributed Multi-Agent Reinforcement Learning. In Proceedings of the Conference on Robot Learning, Atlanta, GA, USA, 6 November 2023; pp. 446–477. [Google Scholar]
  77. Dunbar, W.B.; Murray, R.M. Model predictive control of coordinated multi-vehicle formations. In Proceedings of the 41st IEEE Conference on Decision and Control, 2002, Las Vegas, NV, USA, 10–13 December 2002; Volume 4, pp. 4631–4636. [Google Scholar]
  78. Prasad, B.K.S.; Manjunath, A.G.; Ramasangu, H. Multi-agent Polygon Formation using Reinforcement Learning. In Proceedings of the 9th International Conference on Agents and Artificial Intelligence (ICAART 2017), Porto, Portugal, 24–26 February 2017; pp. 159–165. [Google Scholar]
  79. Afifi, A.M.; Alhosainy, O.H.; Elias, C.M.; Shehata, O.M.; Morgan, E.I. Deep policy-gradient based path planning and reinforcement cooperative q-learning behavior of multi-vehicle systems. In Proceedings of the 2019 IEEE International Conference on Vehicular Electronics and Safety (ICVES), Cairo, Egypt, 4–6 September 2019; pp. 1–7. [Google Scholar]
  80. Li, C.; Li, M.J.; Du, J. An improved method of reinforcement learning action strategy ε-greedy. Intell. Comput. Technol. Autom. 2019, 38, 141–145. [Google Scholar]
  81. Cao, Y.; Fang, X. Optimized-Weighted-Speedy Q-Learning Algorithm for Multi-UGV in Static Environment Path Planning under Anti-Collision Cooperation Mechanism. Mathematics 2023, 11, 2476. [Google Scholar] [CrossRef]
  82. Giannini, F.; Fortino, G.; Franzè, G.; Pupo, F. A Deep Q Learning-Model Predictive Control Approach to vehicle routing and control with platoon constraints. In Proceedings of the 2022 IEEE 18th International Conference on Automation Science and Engineering (CASE), Mexico City, Mexico, 20–24 August 2022; pp. 563–568. [Google Scholar]
  83. Xu, C.; Zhao, W.; Li, L.; Chen, Q.; Kuang, D.; Zhou, J. A nash Q-learning based motion decision algorithm with considering interaction to traffic participants. IEEE Trans. Veh. Technol. 2020, 69, 12621–12634. [Google Scholar] [CrossRef]
  84. Yang, Y.; Li, X.; Yuan, Z.; Wang, Q.; Xu, C.; Zhang, L. Graded-Q Reinforcement Learning with Information-Enhanced State Encoder for Hierarchical Collaborative Multi-Vehicle Pursuit. In Proceedings of the 2022 18th International Conference on Mobility, Sensing and Networking (MSN), Guangzhou, China, 14–16 December 2022; pp. 534–541. [Google Scholar]
  85. Wang, Q.; Phillips, C. Cooperative path-planning for multi-vehicle systems. Electronics 2014, 3, 636–660. [Google Scholar] [CrossRef]
  86. Kaushik, M.; Prasad, V.; Krishna, K.M.; Ravindran, B. Overtaking maneuvers in simulated highway driving using deep reinforcement learning. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 1885–1890. [Google Scholar]
  87. Toghi, B.; Valiente, R.; Sadigh, D.; Pedarsani, R.; Fallah, Y.P. Altruistic maneuver planning for cooperative autonomous vehicles using multi-agent advantage actor-critic. arXiv 2021, arXiv:2107.05664. [Google Scholar]
  88. Biswas, S.; Anavatti, S.G.; Garratt, M.A. Obstacle avoidance for multi-agent path planning based on vectorized particle swarm optimization. In Intelligent and Evolutionary Systems: The 20th Asia Pacific Symposium Proceedings, Proceedings of the IES 2016, Canberra, Australia, 9 November 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 61–74. [Google Scholar]
  89. Ma, Q.; Lei, X.; Zhang, Q. Mobile robot path planning with complex constraints based on the second-order oscillating particle swarm optimization algorithm. In Proceedings of the 2009 WRI World Congress on Computer Science and Information Engineering, Los Angeles, CA, USA, 31 March–2 April 2009; Volume 5, pp. 244–248. [Google Scholar]
  90. Qiaorong, Z.; Guochang, G. Path planning based on improved binary particle swarm optimization algorithm. In Proceedings of the 2008 IEEE Conference on Robotics, Automation and Mechatronics, Chengdu, China, 21–24 September 2008; pp. 462–466. [Google Scholar]
  91. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  92. Eberhart, R.C.; Shi, Y. Comparison between genetic algorithms and particle swarm optimization. In International Conference on Evolutionary Programming; Springer: Berlin/Heidelberg, Germany, 1998; pp. 611–616. [Google Scholar]
  93. Liu, C.; Liu, H.; Yang, J. A path planning method based on adaptive genetic algorithm for mobile robot. J. Inf. Comput. Sci. 2011, 8, 808–814. [Google Scholar]
  94. Nearchou, A.C. Path planning of a mobile robot using genetic heuristics. Robotica 1998, 16, 575–588. [Google Scholar] [CrossRef]
  95. Liu, J.; Anavatti, S.; Garratt, M.; Abbass, H.A. Modified continuous ant colony optimisation for multiple unmanned ground vehicle path planning. Expert Syst. Appl. 2022, 196, 116605. [Google Scholar] [CrossRef]
  96. Jabbarpour, M.R.; Zarrabi, H.; Jung, J.J.; Kim, P. A green ant-based method for path planning of unmanned ground vehicles. IEEE Access 2017, 5, 1820–1832. [Google Scholar] [CrossRef]
  97. Zou, R.; Liu, J.; Shi, H. Multi-vehicle Task Allocation to Attack Targets Based on Modified Particle Swarm Optimization Algorithm. In Proceedings of the 2020 IEEE 6th International Conference on Computer and Communications (ICCC), Chengdu, China, 11–14 December 2020; pp. 1612–1616. [Google Scholar]
  98. Bai, H.; Cai, S.; Ye, N.; Hsu, D.; Lee, W.S. Intention-aware online POMDP planning for autonomous driving in a crowd. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 454–460. [Google Scholar]
  99. Liang, J.; Tian, Q.; Feng, J.; Pi, D.; Yin, G. A polytopic model-based robust predictive control scheme for path tracking of autonomous vehicles. IEEE Trans. Intell. Veh. 2023, 9, 3928–3939. [Google Scholar] [CrossRef]
  100. Zhang, S.; Zhao, X.; Zhu, G.; Shi, P.; Hao, Y.; Kong, L. Adaptive trajectory tracking control strategy of intelligent vehicle. Int. J. Distrib. Sens. Netw. 2020, 16, 1550147720916988. [Google Scholar] [CrossRef]
  101. Shalaby, M.K.; Farag, A.; AbdelAziz, O.M.; Mahfouz, D.M.; Shehata, O.M.; Morgan, E.I. Design of various dynamical-based trajectory tracking control strategies for multi-vehicle platooning problem. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 1631–1637. [Google Scholar]
  102. Qiu, Q.; Yao, S.; Wang, J.; Ma, J.; Chen, G.; Ji, J. Learning to socially navigate in pedestrian-rich environments with interaction capacity. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 279–285. [Google Scholar]
  103. Kessler, T.; Knoll, A. Multi vehicle trajectory coordination for automated parking. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 661–666. [Google Scholar]
  104. Shen, X.; Borrelli, F. Multi-vehicle Conflict Resolution in Highly Constrained Spaces by Merging Optimal Control and Reinforcement Learning. arXiv 2022, arXiv:2211.01487. [Google Scholar] [CrossRef]
  105. Aljohani, T.M.; Ebrahim, A.; Mohammed, O. Real-Time metadata-driven routing optimization for electric vehicle energy consumption minimization using deep reinforcement learning and Markov chain model. Electr. Power Syst. Res. 2021, 192, 106962. [Google Scholar] [CrossRef]
  106. Cartenì, A. The acceptability value of autonomous vehicles: A quantitative analysis of the willingness to pay for shared autonomous vehicles (SAVs) mobility services. Transp. Res. Interdiscip. Perspect. 2020, 8, 100224. [Google Scholar] [CrossRef]
  107. Cascetta, E.; Carteni, A.; Di Francesco, L. Do autonomous vehicles drive like humans? A Turing approach and an application to SAE automation Level 2 cars. Transp. Res. Part C Emerg. Technol. 2022, 134, 103499. [Google Scholar] [CrossRef]
Figure 1. Hierarchical division of planning under automatic driving.
Figure 1. Hierarchical division of planning under automatic driving.
Sustainability 16 04578 g001
Figure 2. Intelligent connected vehicle communication diagram.
Figure 2. Intelligent connected vehicle communication diagram.
Sustainability 16 04578 g002
Figure 3. Integrated decision framework of motion decision system.
Figure 3. Integrated decision framework of motion decision system.
Sustainability 16 04578 g003
Figure 4. Multiple vehicles exchange constraint information in a road scenario.
Figure 4. Multiple vehicles exchange constraint information in a road scenario.
Sustainability 16 04578 g004
Figure 5. The heterogeneous multi-sensor configuration in autonomous vehicles.
Figure 5. The heterogeneous multi-sensor configuration in autonomous vehicles.
Sustainability 16 04578 g005
Figure 6. The MDP model for reinforcement learning based multi-vehicle autonomous driving.
Figure 6. The MDP model for reinforcement learning based multi-vehicle autonomous driving.
Sustainability 16 04578 g006
Figure 7. Vehicle confluence area.
Figure 7. Vehicle confluence area.
Sustainability 16 04578 g007
Figure 8. Dense traffic area scenario. (a) Heavy traffic lanes. (b) Urban intersection. (c) Parking lot.
Figure 8. Dense traffic area scenario. (a) Heavy traffic lanes. (b) Urban intersection. (c) Parking lot.
Sustainability 16 04578 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rong, S.; Meng, R.; Guo, J.; Cui, P.; Qiao, Z. Multi-Vehicle Collaborative Planning Technology under Automatic Driving. Sustainability 2024, 16, 4578. https://doi.org/10.3390/su16114578

AMA Style

Rong S, Meng R, Guo J, Cui P, Qiao Z. Multi-Vehicle Collaborative Planning Technology under Automatic Driving. Sustainability. 2024; 16(11):4578. https://doi.org/10.3390/su16114578

Chicago/Turabian Style

Rong, Songsong, Ruifeng Meng, Junhong Guo, Pengfei Cui, and Zhi Qiao. 2024. "Multi-Vehicle Collaborative Planning Technology under Automatic Driving" Sustainability 16, no. 11: 4578. https://doi.org/10.3390/su16114578

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop