Next Article in Journal
Homogeneous Sensor Fusion Optimization for Low-Cost Inertial Sensors
Previous Article in Journal
A Framework for Representing, Building and Reusing Novel State-of-the-Art Three-Dimensional Object Detection Models in Point Clouds Targeting Self-Driving Applications
Previous Article in Special Issue
A Path Planning Method with a Bidirectional Potential Field Probabilistic Step Size RRT for a Dual Manipulator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Heuristics and Learning Models for Dubins MinMax Traveling Salesman Problem

Mechanical Engineering, Texas A&M University, College Station, TX 77843, USA
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6432; https://doi.org/10.3390/s23146432
Submission received: 30 May 2023 / Revised: 5 July 2023 / Accepted: 11 July 2023 / Published: 15 July 2023
(This article belongs to the Special Issue Aerial Robotics: Navigation and Path Planning)

Abstract

:
This paper addresses a MinMax variant of the Dubins multiple traveling salesman problem (mTSP). This routing problem arises naturally in mission planning applications involving fixed-wing unmanned vehicles and ground robots. We first formulate the routing problem, referred to as the one-in-a-set Dubins mTSP problem (MD-GmTSP), as a mixed-integer linear program (MILP). We then develop heuristic-based search methods for the MD-GmTSP using tour construction algorithms to generate initial feasible solutions relatively fast and then improve on these solutions using variants of the variable neighborhood search (VNS) metaheuristic. Finally, we also explore a graph neural network to implicitly learn policies for the MD-GmTSP using a learning-based approach; specifically, we employ an S-sample batch reinforcement learning method on a shared graph neural network architecture and distributed policy networks to solve the MD-GMTSP. All the proposed algorithms are implemented on modified TSPLIB instances, and the performance of all the proposed algorithms is corroborated. The results show that learning based approaches work well for smaller sized instances, while the VNS based heuristics find the best solutions for larger instances.

1. Introduction

The multiple traveling salesman problem (mTSP) is a challenging optimization that naturally arises in various real-world routing and scheduling applications. It is a generalization of the well-known traveling salesman problem (TSP) [1,2], where given a set of n targets and m vehicles, the objective is to find a path for each vehicle, such that each target is visited at least once by some vehicle, and an objective which depends on the cost of the paths is minimized. The mTSP has many practical application, including but not limited to industrial robotics [3], transportation and delivery [4], monitoring and surveillance [5], disaster management [6,7], precision agriculture [8], search and rescue missions [9,10], multi-robot task allocation and scheduling [11], print press scheduling [12], satellite surveying networks [13], transportation planning [14], mission planning [15], co-operative planning for autonomous robots [16], and unmanned aerial vehicle planning [17].
The standard mTSP aims to find paths for multiple vehicles without taking into account the dynamics of the vehicles. However, some vehicles, like fixed-wing unmanned aerial vehicles and ground robots, have kinematic constraints that need to be considered while planning their paths. Specifically, these vehicles have yaw rate or turning radius constraints. To address this issue, L.E. Dubins [18] extensively studied the paths for curvature-constrained vehicles moving between two configurations. Dubins considered the model of a vehicle that moves at a constant speed v 0 on a 2D plane that cannot reverse and has a minimum turning radius ρ ; this vehicle is also called the Dubins vehicle in the literature. The configuration of a Dubins vehicle is defined by its position and heading ( x , y , θ ) , where x and y are the vehicle’s position coordinates on the plane, and θ is its heading angle. The motion of a Dubins vehicle can be fully defined using the following set of equations:
x ˙ = v 0 cos θ , y ˙ = v 0 sin θ , θ ˙ = v 0 ρ u , 1 u 1
where u is the input to the vehicle and ρ is the minimum turning radius of the vehicle. Dubins proposed that in an obstacle-free environment, the shortest path between any two configurations on a 2D plane for a Dubins vehicle must be either the C C C type or the C S C type or a combination of sub-paths of these two types; here, C represents an arc of a circle with a radius of ρ , and S represents a straight line segment. These paths can be expressed as L S L , R S R , R S L , L S R , R L R , or L R L , where L and R indicate clockwise and counterclockwise turns of a C segment, respectively, as illustrated in Figure 1. When the vehicle is also capable of reversing, the shortest path problem was solved by Reeds and Shepp in [19]. Optimal solutions to these problems were also derived by Sussmann and Tang (1991) [20] as well as Boissonnat et al. (1994) [21], utilizing the Pontryagin maximum principle [22]. Tang and Özgüner (2005) [23] and Rathinam et al. (2007) [24] have further expanded upon Dubins’ research by applying it to routing problems involving car-like mobile robots or fixed-wing aerial vehicles that maintain a constant forward speed and adhere to curvature constraints while making turns.
The vehicle routing problem for a Dubins vehicle is called the Dubins traveling salesman problem (DTSP) [25]. In the DTSP, the challenge is not only to find the optimal order of the targets to visit but also to determine suitable heading angles for the vehicle when visiting the targets. A feasible solution to the DTSP involves a curved path where the radius of curvature at any point along the path is at least equal to the turning radius ρ 0 , and each target is visited at least once. Since the problem involves both discrete and continuous decision variables, with the target order being discrete and the heading angles being continuous, finding an optimal solution to the DTSP is an NP-hard problem, as shown by Jerome Le Ny et al. [26]. Currently, no exact algorithm exists in the literature that can find an optimal solution to the DTSP. However, over the last two decades, several heuristics and approximation algorithms have been developed to find feasible solutions for the DTSP [23,24,27,28,29,30,31,32,33,34].
One popular approach to solving the DTSP, as discussed in several studies [33,35,36,37,38], involves sampling the interval [ 0 , 2 π ] into k candidate headings at each target and posing the resulting problem as a generalized traveling salesman problem (GTSP) formulation. In GTSP (also known as the one-in-a-set TSP), given n nodes partitioned into k mutually exclusive sets, the goal is to compute the shortest path that contains exactly one node from each of the k sets. GTSP is an NP-hard problem since it reduces to the TSP when each cluster contains only one node. DTSP can be reduced to a one-in-a-TSP by forming n mutually exclusive sets consisting of k vehicle configurations corresponding to k heading angles at each of the n targets. Increasing the angle discretizations at each target for the one-in-a-set TSP variant results in solutions that get close to the DTSP optimum. This approach provides a natural way to find a good feasible solution to the DTSP problem, as noted in previous studies [36]. Figure 2 illustrates the discretization of heading angles at a target for the DTSP.
In this work, we explore the MinMax variant of the mTSP problem for a Dubins-like vehicle by formulating it as a one-in-a-set TSP, called the MinMax Dubins generalized multiple traveling salesman problem (MD-GmTSP) and present three different approaches to solve it. The MinMax variant is an important problem to study as it directly relates the objective to the mission completion time for visiting all the targets by a team of vehicles. This variant is relevant for real-world applications as it accounts for the worst-case scenario in situations where m homogeneous vehicles are assigned travel paths by accounting for limitations on load/delivery capacities or refueling constraints.

Contributions of This Article

  • We formulate a mixed-integer linear program for the MD-GmTSP (Section 3).
  • We develop heuristics to generate feasible solutions for the MD-GmTSP in a relatively short amount of time, followed by using a combination of variable neighborhood search (VNS) metaheuristic [39] for improving solution quality (Section 5).
  • We explore learning-based methods to solve the MD-mTSP. An architecture consisting of a shared graph neural network and distributed policy networks is used to define a learning policy for MD-GmTSP. Reinforcement learning is used to learn the allocation of agents to vertices to generate feasible solutions for the MD-GmTSP, thus eliminating the need for accurate ground truth (Section 6).
  • Finally, we implement all the algorithms on a set of modified instances from the TSPLIB library [40] using CPLEX and present simulation results to corroborate the performance of the proposed approaches (Section 7).

2. Literature Review

The traveling salesman problem (TSP) is one of the most well-known optimization problems in the literature [41]. Several methods, ranging from exact techniques (branch and bound/cut/price) [41] to fast heuristics and approximation algorithms [42,43], have been developed to solve the TSP. Heuristics generally aim to trade off optimality for computational efficiency. The mTSP deals with a generalization of the TSP with multiple vehicles. mTSPs are much harder to solve to optimality compared to the single TSP because targets have to be allocated to vehicles in addition to finding a tour for each of the vehicles. Numerous variants and generalizations of the mTSP have been addressed in the literature. For example, mTSP with motion constraints are addressed in [24,44]; fuel constraints are addressed in [45]; capacity and other resource constraints are addressed in [46,47]. Many recent methods have focused on the application of metaheuristics like genetic algorithms (GA) [48,49], simulated annealing (SA) [50], memetic search [51], tabu search [52], swarm optimization [53], and other methods [54,55] to solve the mTSP problem for the single-depot and multi-depot case.
The mTSP problem can be classified into two main types based on its objectives: the MinSum TSP and the MinMax TSP. In MinSum mTSP, the goal is to minimize the sum of the path costs for visiting all targets by the m vehicles, while in the MinMax mTSP [56,57,58], the objective is to minimize the maximum among the m path costs. While there are several approximation algorithms for the MinSum mTSP [24], there are very few theoretical results for the MinMax mTSP [59,60,61]. Even further, the MinMax mTSP for vehicles with turning radius constraint (Dubins-type vehicle) has received limited or no attention in the literature primarily due to the computational complexity involved in solving the problem.
For the case when all the vehicles are homogeneous without motion constraints, there are several heuristics [62,63,64,65,66] in the literature for solving MinMax mTSPs and related vehicle routing problems. In [62], the authors present four heuristics and found that a linear program (LP)-based heuristic in combination with load balancing and region partitioning ideas performed the best. In [63], the authors present an ant-colony-based metaheuristic to solve the min-max problem with limited computational experiments. A transformation-based method by decoupling the task partitioning problem amongst the vehicles and the routing problems is presented in [65]. In [66], a new heuristic denoted as MD is presented for a min-max vehicle routing problem. Through extensive computational results, the MD algorithm is shown to be the best amongst the considered min-max algorithms in [66].
Recently, MinMax mTSPs with other constraints have also been considered in surveillance applications. In [67], a mixed-integer linear program (MILP) based approach is presented along with market based heuristics for a min-max problem with additional revisit constraints to the targets. In [68], a min-max routing problem is considered where the vehicles are functionally heterogeneous, i.e., there are vehicle-target assignment constraints that specify the sub-set of targets that each vehicle can visit. Both approximation algorithms and heuristics were presented in [68]. In [69], insertion based heuristics are presented for a team of underwater autonomous vehicles with functional heterogeneity. An energy-aware VNS is presented for a vehicle routing problem in [70] where there is a fuel or energy constraint for each of the vehicles. There is also recent work [71] on generalized min-max routing problems where the targets are partitioned into sets, and a vehicle is only required to visit exactly one target from each set.

3. Problem Statement

Let T = { T 1 , T 2 , T t } be the set of t targets on a 2D plane. Each target is associated with h candidate heading angles. For simplicity, lets assume that these heading angles are obtained by sampling the interval [ 0 , 2 π ] uniformly, i.e., Φ = { 0 , 2 π h , , c 2 π h , , 2 π } , where c h . Let { T 0 } denote the initial target (depot). Let ( x i , y i ) denote the location of target T i , and θ i Φ denote the arrival angle of a vehicle at target i. The angle of arrival and angle of departure of a vehicle from a target is assumed to be the same. To simplify the presentation, we also refer to ( x i , y i ) as target T i . The configuration of the vehicle at a target T i at an angle of θ i is denoted by ( x i , y i , θ i ) or simply ( T i , θ i ) .
Let ν = { 1 , 2 , m } be the set of m homogeneous vehicles with a minimum turning radius ρ . Let τ k T be the set of targets toured by vehicle k visiting each target exactly once, starting and ending at depot T 0 . Let d k ( T i , θ i , T j , θ j ) denote the shortest Dubins path traversed by vehicle k ν , from ( T i , θ i ) to ( T j , θ j ) . Since the vehicles are homogeneous, d k ( T i , θ i , T j , θ j ) remains the same for all k ν . Let the heading angle at the depot for all the vehicles be the same and be given. The objective of MD-GmTSP is to find a tour for each vehicle and the heading angle at each target in T such that
  • Each target in T is visited once by some vehicle at a specified heading angle;
  • The tour for each vehicle starts and ends at the depot;
  • The length of the longest Dubins tour among the m vehicles is minimized.

4. Mixed-Integer Linear Programming Formulation

To model the MD-GmTSP as a mixed-integer linear programming problem, we define the related sets, decision variables, and parameters as follows:

4.1. Notations

  • V : = { 0 , 1 , 2 , , n } , the set of all possible Dubins vehicle configurations. Let V be partitioned into mutually exclusive and exhaustive non-empty sub-sets where V 0 : = { 0 } corresponds to the configuration at the depot, and V 1 , , V t corresponds to the t target clusters.
  • A : = ( i , j ) : i , j V , i j , the set of arcs representing the Dubins path between configurations.
  • ν : = { 1 , 2 , , m } , the set of homogeneous (uniform) vehicles that serve these customers.
  • c k i j , the cost of traveling from configuration i to configuration j for vehicle k, where k ν , i j , i V p , j V r , p r , p , r T .
  • u k p : , the rank order (visit order) of cluster V p on the tour of vehicle k ν , p T .
  • Q : = t m + 1 , the maximum number of clusters that a vehicle can visit.

4.2. Decision Variables

Let us define the following binary variables:
x k i j = 1 if arc ( i , j ) is included in the tour of vehicle k ν , i V p , j V r , p , r T 0 otherwise
y k p r = 1 if path exists from target cluster V p to cluster V r in tour k , p , r T , k ν 0 otherwise

4.3. Cluster Degree Constraints

For each target cluster V p (excluding V 0 ), there can only be a single outgoing arc to any other target cluster belonging to the tour of any of the m vehicles. This condition is imposed by the following constraints:
k ν i V p j V / V p x k i j = 1 , p 0 , p T .
There can only be a single incoming (entering) arc to a target cluster from any other vertex belonging to other target clusters, excluding V 0 for a given vehicle. This condition is imposed by the following constraints:
k ν i V / V p j V p x k i j = 1 , p 0 , p T .
For each vehicle, there should be a single leaving arc from and a single entering arc to the depot, which are imposed by
i = 1 n x k 0 i = m , k ν .
i = 1 n x k i 0 = m , k ν .

4.4. Cluster Connectivity Constraints

The entering and leaving nodes should be the same for a given vehicle in each cluster, which is satisfied by  
i V / V p x k i j = i V / V p x k j i j V p , p T , k ν .
Flows from target cluster V p to target cluster V r for vehicle k are defined by y k p r . Thus, y k p r should be equal to the sum of x k i j ’s from V p to V r . Hence, we write
y k p r = i V p j V r x k i j , p r , p , r T , k ν .

4.5. Sub-Tour Elimination Constraints

The sub-tour elimination constraints and capacity bounding constraints are proposed using the auxiliary variables u k p and Q as
u k p + r = 1 , r p y r p 1 p 0 , p T , k ν .
u k p + ( Q 1 ) y k p r Q , p 0 , p T , k ν .
u k p u k r + Q y k p r + ( Q 2 ) y k r p Q 1 , p r 0 , p , r T , k ν .

4.6. Objective

The integer linear programming formulation for M D G m T S P is given by
( MD GmTSP ) : minimize max i = 0 n j = 0 n c k i j x k i j , k ν
subject to (2)–(10).

4.7. Solving the MILP

To obtain a globally optimal solution for a problem expressed as an MILP, it is common to use solvers such as Gurobi or CPLEX. Although these solvers are effective in solving smaller instances of the Euclidean mTSP (for example, an instance with 100 nodes and 3–5 robots takes up to 3 h to solve using a MILP formulation [72]), they struggle to scale to larger instances that involve multiple robots, especially for difficult problems such as the MD-GmTSP. In the following section, we introduce a variable-neighborhood-search-based heuristic that leverages the problem’s structure to generate fast and high-quality solutions.

5. Heuristics for the MD-GmTSP

In this section, we present a heuristic based on variable neighborhood search (VNS) to solve the MD-GmTSP. VNS is a metaheuristic proposed by Hansen and Mladenovic [39,73] that employs the idea of systematic exploration of neighborhoods of an initial solution and improves it using local search methods. The neighborhood exploration involves descent to a local minimum in the neighborhood of the incumbent solution or an escape from the valleys containing them to obtain better solutions (global minimum).
The heuristic approach can be divided into two broad phases: (1) construction phase—initial solution and (2) improvement phase—neighborhood search. In the construction phase, a tour construction heuristic is used to generate an initial feasible solution for the given problem. Based on this initial solution, the neighborhood search phase aims to obtain better solutions by making incremental improvements. The tour improvement phase starts by defining a set of neighborhood structures in the solution space. These neighborhood search structures, denoted as N 1 , N 2 , , N κ m a x , can be chosen either arbitrarily or based on a sequence of neighborhood changes with increasing cardinality. Once the neighborhood search structures are fixed, different neighborhoods are explored in deterministic or stochastic ways, depending on the type of VNS metaheuristic used. If a better solution is found in the κ -th neighborhood, then the search is re-centered around the new incumbent. Otherwise, the search moves to the next neighborhood, κ + 1 . Depending on the chosen neighborhood structures, different neighborhoods of the solution space are reached. VNS does not follow a predefined trajectory but explores increasingly distant neighborhoods of the current solution. A jump to a new solution is made only if an improvement is achieved, often retaining favorable characteristics of the incumbent solution to obtain promising neighboring solutions.

5.1. Initial Solution: Construction Phase

Construction heuristics are commonly used to address vehicle routing problems, and they can be broadly categorized into three classes: insertion heuristics, savings heuristics, or clustering heuristics. In this work, we focus on insertion-based heuristics for the MD-GmTSP, where the algorithm starts with a closed sub-tour comprising a few targets and progressively inserts new targets into the tour until all the targets are visited. Specifically, we explore two types of insertion heuristics for the MD-GmTSP: (1) greedy k-insertion for MD-GmTSP and (2) cheapest k-insertion for MD-GmTSP.
The key differentiator between these algorithms is the order of target selection and the position at which each target is inserted. The heuristics begin with a tour starting at a depot (initial location). A new target is inserted into the tour such that the increase in tour length is minimized. Greedy insertion algorithms, namely Nearest Insertion and Farthest Insertion, select an unvisited target whose distance from any target in the current tour is minimum and maximum, respectively, for insertion. In contrast, the cheapest-insertion algorithm selects an unvisited target whose insertion causes the lowest increase in tour length. The main steps involved in these construction heuristics for the MD-GmTSP are discussed below.
Consider a set of t targets and their associated heading angles, V = { ( T 1 , θ 1 ) , , ( T i , θ i ) , , ( T t , θ t ) } to be visited by a m Dubins vehicles, and let ( T 0 , θ 0 ) correspond to the depot target. To simplify the presentation, we also refer to any target configuration ( T i , θ i ) as a target. Let τ v = { ( T 0 , θ 0 ) , , ( T i , θ i ) , , ( T 0 , θ 0 ) } , i [ 1 , p ] denote a Dubins tour for vehicle v ν , through p ( t ) cities, starting and ending at the depot. Let d v ( T i , T i + 1 ) denote the length of the shortest Dubins path from ( T i , θ i ) to ( T i + 1 , θ i + 1 ) . Consider a target ( T j , θ j ) from the set of unvisited targets. The increase in tour length by inserting the T j between targets T i and T i + 1 in tour τ v , without changing the associated heading angles, would be Δ D v ( T j ) = d v ( T i , T j ) + d v ( T j , T i + 1 ) d v ( T i , T i + 1 ) . We wish to insert the new unvisited target T j in a tour τ v at a position such that the increase in length of tour v ( Δ D v ( T j ) ) results in a minimum increase in length of the largest of the m tours. One way of achieving this is by greedily inserting the new target T j in the smallest of the m tours without affecting the rest of the tour(s).
A relaxed version of the above approach is to allow modifications to the heading angles of k ( 0 ) targets adjacent to the newly inserted target such that the increase in tour length is minimum. Given a Dubins tour τ v = { ( T 0 , θ 0 ) , , ( T p , θ p ) , , ( T 0 , θ 0 ) } , let D v ( T i k , T ( i + 1 ) + k ) be the length of the shortest possible Dubins sub-tour through targets T i k , , T i , T i + 1 , , T ( i + 1 ) + k with predefined headings θ i k , , θ ( i + 1 ) + k . Then, Δ D v ( T j ) = D ( T i k , T j ) + D ( T j , T ( i + 1 + k ) D ( T i k , T ( i + 1 ) + k ) describes the increase in the tour length of vehicle v by inserting target T j . This version allows for the heading angles of 2 k targets between T i k and T ( i + 1 ) + k to be optimized when the target T j in inserted into the tour. Algorithm 1 presents a formal description of the k greedy MD-GmTSP tour construction heuristic.
Algorithm 1 Greedy k-insertion for MD-GmTSP.
Initialization: Start with m initial Dubins tours τ v , v [ 1 , m ] starting and ending at the depot ( T 0 , θ 0 ) .
Repeat: While the set of unvisited targets N u is non-empty
1.
Identify the tour with the shortest Dubins tour length τ ν m i n min ( D v ) , v [ 1 , m ]
2.
Choose a target T j from the list of unvisited targets N u according to one of the following greedy strategy.
(a)
Random: Choose a random target T j N u
(b)
Nearest: Find a target T j N u such that the Dubins distance d ( T j , T j ) is minimum T j τ ν m i n
(c)
Farthest: Find a target T j N u such that the Dubins distance d ( T j , T j ) is maximum T j τ ν m i n
3.
Determine a position i in the tour τ ν m i n such that the increase in tour length Δ D ν min ( T j ) = D ( T i k , T j ) + D ( T j , T ( i + 1 + k ) D ( T i k , T ( i + 1 ) + k ) is minimum; insert T j in τ ν m i n .
4.
Remove T j from the set of unvisited targets N u
In the cheapest k-insertion method (Algorithm 2), during each iteration, the algorithm finds the most promising target (from the set of unvisited targets) and the position for insertion in the current tour simultaneously. In this algorithm, we construct new tours by inserting new targets at positions that lead to the smallest increase in tour lengths.
Algorithm 2 Cheapest k-insertion for MD-GmTSP
Initialization: Start with m initial Dubins tours τ v , v [ 1 , m ] starting and ending at the depot ( T 0 , θ 0 ) .
Repeat: While the set of unvisited targets N u is non-empty
1.
Identify the tour with the shortest Dubins tour length τ ν m i n min ( D v ) , v [ 1 , m ]
2.
Identify the target T j from the list of unvisited targets N u , and a position i in the tour τ ν m i n such that the increase in tour length Δ D ν m i n ( T j ) = D ( T i k , T j ) + D ( T j , T ( i + 1 + k ) D ( T i k , T ( i + 1 ) + k ) is minimum; insert T j in τ ν m i n .
3.
Remove T j from the set of unvisited targets N u

5.2. Neighborhood Search: Improvement Phase

In the improvement phase, we use VNS-based local search heuristics to enhance the incumbent solution from the construction phase. These heuristics modify the current solution by performing a sequence of operations to produce a new feasible solution that improves the objective function. The algorithm evaluates the effect of modifying the current solution systematically and replaces it with a new solution if it is better. The algorithm may also be allowed to make changes that lead to poorer solutions with the hope of finding a better solution later.
Let I denote a problem instance of the MD-GmTSP. Let X represent the set of feasible solutions for this instance. For any x X , let f ( x ) denote a function that calculates the objective cost, which, in the case of the MD-GmTSP, corresponds to the Dubins tour length of the largest tour within x. It is important to note that the solution set X is finite but the size of the set can exponentially increase with the size (number of targets) of an instance of MD-GmTSP. Our goal is to find a solution x such that f ( x ) f ( x ) , x X . To facilitate this search, we define a neighborhood N ( x ) X for each solution x X , where N is a function that maps from a solution x to a set of solutions in its neighborhood. A solution x is defined to be locally optimal or is said to be a local optimum with respect to a neighborhood N if f ( x ) f ( x ) , x N ( x ) . Improvement heuristics aim to find a locally optimal solution before the termination condition is met. In this paper, we investigate three improvement heuristics based on VNS for solving the MD-GmTSP, namely (1) basic variable neighborhood search (BVNS), (2) variable neighborhood descent (VND), and (3) general variable neighborhood search (GVNS).

5.2.1. Basic Variable Neighborhood Search (BVNS)

BVNS [39] uses a stochastic approach to reach the global minimum. In this approach, the neighborhood search is continued in a random direction, away from the initial solution. The steps of BVNS are explained in Algorithm 3.
Algorithm 3 Steps of the basic VNS by Hansen and Mladenovic  [73]
Initialization: Choose a set of neighborhood structures denoted by N k ( κ = 1 , , κ m a x ) for the search process; find an initial solution x; select a termination/stopping condition and repeat the following steps until the termination condition is met:
1.
Set κ 1 .
2.
Repeat the following steps until κ = κ max :
(a)
Shaking: Generate a feasible solution x at random from the κ th neighborhood of x ( x N k ( x ) ) ;
(b)
Local search: Apply a local search method with x as an initial solution; denote the obtained local optimum as x ;
(c)
Move or not: If x is better than the incumbent x, set x x and continue the search process with N 1 ( κ 1 ) ; otherwise, set κ κ + 1 ;

Neighborhood Structures

In this work, we use k exchange operators to explore neighborhoods of MD-GmTSP using BVNS. A solution χ is considered to be in the k exchange neighborhood of solution χ if there is exactly one vehicle V i such that:
(1)
χ and χ differ only in the order of target visits in the tour of vehicle V i .
(2)
The tour for vehicle V i in χ differs from the tour of vehicle V i in χ by at most k edges.

Shake

‘Shaking’ is the randomization part of the BVNS heuristic. Shaking perturbs the initial solution while ensuring that the new neighborhood still retains certain aspects of the initial incumbent. In this step, a random feasible solution x in the current neighborhood of x is found. This random selection enables the algorithm to avoid stopping at the local optima in the local search procedure. In this work, we investigate two different Shake neighborhoods: Shake and Shake m o d . Shake refers to picking two random targets T j , T j from tours of two random vehicles v and v , respectively, and swapping the targets between tours. The position of insertion for the targets is also selected at random. For the MD-GmTSP, in addition to Shake, a modified shake Shake m o d is also evaluated. In Shake m o d , a random target T j is selected from a random tour τ v , and is transferred to another tour τ v chosen at random. The insertion of the targets into the tours can be random or greedy (position leading to a minimal increase in tour length). The different shake neighborhoods are illustrated in Figure 3.

Local Search

After obtaining the solution x from the Shake procedure, we proceed to enhance it through local search. During this step, all feasible neighbors of x within the current neighborhood are explored to identify a potential solution x that has a lower cost. It is worth noting that there is a possibility for the current neighborhood of x to be empty. In that case, we set x to be the same as x , and the search continues in the next neighborhood. Exploration of neighborhoods can be of two types: (a) steepest descent and (b) first descent. The steepest descent and first descent heuristics are detailed in Algorithms 4 and 5, respectively.
For a strong NP-hard problem like the MD-GmTSP, the steepest descent heuristic is a highly time-consuming process. In this work, a first improvement version of the 2-opt and 3-opt algorithms explores the different neighborhood structures of the MD-GmTSP.
Algorithm 4 SteepestDescent Heuristic by Hansen and Mladenovic [73]
Initialization: Find an initial solution x; Choose the neighborhood structures N ( x ) , and the objective function f ( x ) that calculates the length of the longest tour in x,
Repeat:
1.
Find x = arg min x N ( x ) f ( x )
2.
If f ( x ) < f ( x ) set x x and iterate; otherwise, stop
Algorithm 5 FirstImprovement heuristic by Hansen and Mladenovic [73]
Initialization: Find an initial solution x; Choose the neighborhood structures N ( x ) , and the objective function f ( x ) that calculates the length of the longest tour in x,
Repeat:
1.
Find first solution x N ( x )
2.
If f ( x ) > f ( x ) find next solution x N ( x ) ; set x x and iterate (2); otherwise set x x and iterate (1);
3.
If all solutions of N ( x ) have been considered, stop.

2-opt

Given a solution, the 2-opt algorithm, introduced by Croes in 1958 [74], is a local search heuristic that explores all possible 2-Exchange neighborhoods of the solution. Its objective is to gradually enhance a feasible tour by iteratively swapping two edges in the current tour with two new edges, aiming to reach a local optimum where no further improvements are possible. During each improving step, the 2-opt algorithm selects two distinct edges that connect ( i , i ) and ( j , j ) from the tour. It then replaces these edges with the edges connecting ( i , j ) and ( i , j ) , provided that this change reduces the tour length. This process continues until a termination condition is encountered (See Figure 4).

3-opt

The 3-opt algorithm [75] works similarly to the 2-opt but by reconnecting three edges instead of two, as seen in Figure 5. There are seven different ways of reconnecting the edges. Each new tour formed by reconnecting the edges is analyzed to find the optimum one. This process repeats until all possible three-edge combinations in the network are checked for improvement. A 3-optimal tour is also a 2-optimal one. A 3-opt exchange yields better solutions, but it is much slower ( O ( n 3 ) complexity) compared to the 2-opt ( O ( n 2 ) complexity).
On completion of the local search, the cost of x is compared to the cost of x. If the cost of x is less than the cost of x, then x is set to be x. The current neighborhood of x is set as the first neighborhood ( κ 1 ), and the algorithm continues from the shaking step. On the contrary, if the cost of x x , then the solution x is discarded and search continues in the next neighborhood κ + 1 . The algorithm terminates when there are no new neighborhoods to be explored. The steps of neighborhood change are detailed in Algorithm 6.
Algorithm 6 Steps for neighborhood change by Hansen and Mladenovic [73]
  • function neighborhoodChange(x, x k)
  •     if  f ( x ) < f ( x )  then
  •          x x ; k 1                              ▹ Make a move
  •     else
  •          k k + 1                             ▹ Next neighborhood

5.2.2. Variable Neighborhood Descent (VND)

The main difference between the BVNS and the VND approach is that in the VND, the neighborhood changes are deterministic. Algorithm 7 explains the VND metaheuristic in detail.
Depending on the problem, multiple local search methods can be nested to optimize the VND search. The objective of the local search heuristics is to find the local minimum among all the l m a x neighborhoods. As a result, the likelihood of reaching a global minimum is higher when using VND with a larger l m a x than when using a single-neighborhood structure.
Algorithm 7 Steps of the basic VND by Hansen and Mladenovic  [73]
Initialization: Choose a set of neighborhood structures denoted by N l ( l = 1 , , l m a x ) that will be used in the descent process; find an initial solution x; select a termination/stopping condition and repeat the following steps until the termination condition is met:
1.
Set l 1 .
2.
Repeat the following steps until l = l max :
(a)
Exploration of neighborhood: Find the best neighbor x of x ( x N l ( x ) );
(b)
Move or not: If x is better than the incumbent x, set x x and continue the search process with N 1 ( l 1 ) ; otherwise set l l + 1 ;

Exploration of Neighborhoods for MD-GmTSP

We use a combination of four different neighborhood structures to explore the neighborhoods for MD-GmTSP. An x-point move is nested with a k-opt search (2-opt or 3-opt) to explore distant neighborhoods during local search coupled with greedy and random insertion techniques. The neighborhood structures are discussed in detail below.
  • One-point move: Given a solution x, a one-point move transfers a target T j from tour τ v to a new feasible position in another tour τ v ( τ v ) . The target to be moved is chosen from the tour having the largest tour length and is relocated to a tour having the smallest tour length. The computational complexity of these local search operators is O ( n 2 ) for a given solution x .
  • Two-point move: A two-point move swaps a pair of nodes rather than transferring a node between tours as in a one-point move. A target T j belonging to the tour τ v having the largest tour length is swapped with a target T j belonging to another tour τ v ( τ v ) . After performing two-point moves from the solution x , the best solution x is returned depending on the search strategy employed (FirstImprovement or SteepestDescent).

5.2.3. General Variable Neighborhood Search (GVNS)

The general variable neighborhood search (GVNS) [73] merges the techniques used in both VND (deterministic descent to a local optimum) and BVNS (stochastic search) to reach increasingly distant neighborhoods in the search space. In GVNS, a sequential VND search ( V N D s e q ) improves the initial solution as compared to a local search heuristic (as in the case of BVNS). Algorithm 8 details the steps of GVNS as proposed by Hansen and Mladenovic.
Algorithm 8 Steps of the GVNS by Hansen and Mladenovic [73]
Initialization: Choose the set of neighborhood structures denoted by N κ ( κ = 1 , , κ m a x ) that will be used in the shaking phase and the set of neighborhood structures N l ( l = 1 , , l m a x ) that will be used in the local search process; find an initial solution x; choose a termination/stopping condition.
Repeat the following steps until the termination condition is met:
1.
Set κ 1 .
2.
Repeat the following steps until κ = κ max :
(a)
Shaking: Generate a feasible solution x at random from the κ th neighborhood of x ( x N k ( x ) ) ;
(b)
Local search by VND:
  • Set l 1
  • Repeat the following steps until l = l m a x
    -
    Exploration of neighborhood. Find the best neighbor x of x in N l ( x )
    -
    Move or not. If f ( x ) < f ( x ) , set x x and l 1 ; otherwise set l l + 1
(c)
Move or not: If the local optimum x is better than the incumbent, move there ( x x ) and continue the search with N 1 ( κ 1 ) ; otherwise set κ κ + 1 ;

Neighborhood Search Structures for MD-GmTSP

To ensure the effectiveness of GVNS, it is crucial to employ suitable local search strategies during the execution of Local Search by VND ( V N D s e q ). The performance of V N D s e q is influenced by the choice of neighborhood structures and the order in which these neighborhoods are explored. A combination of x point and k edge move neighborhoods are used in this paper to solve the MD-GmTSP. The details of these neighborhoods are provided below and illustrated in Figure 6.
1.
One-point move: In GVNS, we use the same One-point move operator as in BVNS.
2.
Two-point move: In GVNS, we use the same Two-point move operator as in BVNS.
3.
Or-opt2 move: An or-opt2 move selects a string of two adjacent nodes belonging to the tour having the maximum length and transfers it into a new tour. After performing the or-opt2 move for all strings of nodes x , the best solution x is returned by the operator depending on the search strategy employed. The or-opt2 operator generates x N 2 ( x ) as compared to generic or-optk move operator generates x N k ( x ) . For our use case, no improvement was observed for k > 2 , which can be attributed to the complexity in the MD-GmTSP problem.
4.
Three-point move: The three-point move operation involves selecting a pair of adjacent nodes from the tour with the maximum length and exchanging them with a node from another tour. By repeatedly applying three-point moves starting from the initial solution x , the operator generates a new solution x . Depending on the chosen search strategy, the best solution x N 3 ( x ) is returned by the operator.
5.
2-opt move: We use a 2-opt move operator to improve the resulting tours x obtained from the rest of the operators by performing intra-tour local optimization.

Local Search by VND for MD-GmTSP

A nested local search strategy is employed to solve the MD-GmTSP using GVNS. The V N D s e q as specified in Algorithm 9 explores the neighbors of a given solution according to a particular sequence. This sequence is determined based on the non-decreasing order of neighborhood size. Based on the local search in each neighborhood, if the obtained solution ( x ) is better than the current best ( x ), the 2-opt search is carried out for possible intra-tour improvements. The resulting solution replaces the current best ( x x ), and the search restarts at the first neighborhood. Otherwise, the search continues with a larger neighborhood search operator.
Algorithm 9  V N D s e q
1:
procedure V N D s e q ( x , l m a x )
2:
     l 1
3:
    while  l < l m a x  do
4:
        if  l = 1  then
5:
            x onepoint ( x )
6:
        if  l = 2  then
7:
            x twopoint ( x )
8:
        if  l = 3  then
9:
            x or_opt2 ( x )
10:
        if  l = 4  then
11:
            x threepoint ( x )
12:
        if  f ( x ) < f ( x )  then
13:
            x 2 _ o p t ( x )
14:
            x x ; l 1
15:
        else
16:
            l l + 1
17:
    return  x
Using heuristics to solve NP-hard problems provides a distinct advantage over traditional solvers as it can produce feasible solutions quickly. The quality of the solution can be further improved by employing improvement heuristics, but this often comes at the cost of computational power and time. Heuristics are rule-based approaches that prioritize improving computational efficiency over finding the optimal solution. Recently, there has been significant interest in learning-based approaches from the combinatorial optimization research community. These approaches utilize models that can solve large problem instances in real-time. The concept behind this approach is that heuristics can be viewed as decision-making policies which can be parameterized using neural networks. The policies can then be trained to solve different combinatorial optimization problems. In the next section, we introduce a reinforcement learning method to develop policies for solving the MD-GmTSP. This method holds great promise for achieving high-quality solutions in a shorter amount of time than traditional heuristics.

6. Learning-Based Approach for the MD-GmTSP

Several machine learning techniques, including neural networks [76], pointer networks [77], attention networks [78,79] and reinforcement learning [80,81] have been explored to solve the TSP problem. Wouter et al. [79] extended a single-agent TSP model to learn strong heuristics for the vehicle routing problem (VRP) and its variants. However, supervised learning, which is successful for most machine learning problems, is not applicable for combinatorial optimization problems due to the unavailability of optimal solutions for large instances, especially for co-operative combinatorial optimization problems like multiple agent TSP, asymmetric mTSP, and multi-agent task assignment and scheduling problems. These problems pose several challenges, such as an explosion of the search space with the increase in the number of agents and cities, a lack of data with ground truth solutions, and difficulty in developing an architecture that can capture interactive behaviors among agents. Reinforcement learning (RL) is a commonly explored paradigm to solve problems with the above difficulties, whose solution quality can be verified and provided as a reward to a learning algorithm. Recent studies [82,83] have investigated using RL-based approaches to optimize the Euclidean version of MinMax TSP, but no significant study has been conducted for the Dubins MinMax TSP. The main advantage of using the RL-based method is that the majority of computational time and resources are spent on training. Once a trained model is available, a feasible solution can be inferred almost in real-time.
In this paper, we utilize the RL framework developed by Hu et al. [82] and expand its application to the Dubins MTSP. The model consists of a shared graph neural network and distributed policy networks that collectively learn a universal policy representation for generating nearly optimal solutions for the Dubins MTSP. To address the challenges of optimization, the extensive search space of the problem is effectively partitioned into two tasks: assigning cities to agents with candidate heading angles and performing small-scale Dubins TSP planning for each agent. As proposed by Hu et al. [82], an S-sample batch reinforcement learning technique is used to address the lack of training data, reducing the gradient approximation variance and significantly enhancing the convergence speed and performance of the RL model.

6.1. Framework Architecture

In this work, the graph isomorphism network (GIN) [84] policy is used to summarize the state of the instance graph and assigns nodes to each agent for visitation, effectively transforming the MD-GmTSP into a single-agent DTSP problem. This approach facilitates the use of algorithms [85,86], learning-based methods [77,79,80], and solvers [87,88,89] to quickly obtain near-optimal solutions. Additionally, we utilize a group of distributed policy networks to address the node-to-agent assignment problem.

6.1.1. Graph Embedding

For each node v V , the graph embedding network GIN [84] computes a p-dimensional feature embedding f v by information sharing from the neighboring connected nodes according to the graph structure. GIN uses the following update process to parameterize the graph neural network:
f v ( k ) = M L P ( k ) { ( 1 + ϵ ( k ) ) . h v ( k 1 ) + u N ( v ) h u ( k 1 ) }
where f v k is the feature vector of node v V at the k th iteration/layer, N v denotes all the neighbors of node v, u is neighboring node of v (i.e., u N v ), M L P represents a multi-layer perceptron, and ϵ is a learnable parameter.

6.1.2. Distributed Policy Networks

We propose a two-stage distributed policy network-based approach for designing a graph neural network with the attention mechanism [90]. In the first stage, each agent autonomously generates its agent embedding by leveraging both the global information and the node embeddings present in the graph. In the second stage, each node selects an agent to associate with itself based on its own embedding and all the prior agent embeddings.

Calculation of Agent Embedding

Graph embeddings, denoted as g f , are computed by max pooling from the set of node features f = { f 1 , f 2 , f n } , i.e.,
g f = max { f 1 , f 2 , , f n } .
These graph embeddings are then used as inputs to construct an agent embedding, where f i R p , n is the number of nodes, p is the dimension of node embedding, and g f R p . To construct the agent embedding, an attention mechanism is employed to generate attention coefficients that signify the significance of a particular node’s feature in creating its embedding.
  • Graph Context embedding: A graph context embedding is used to ensure that every city, except the depot, is visited by only one agent and the depot is visited by all agents. By setting the depot as the first node in the graph ( f d e p o t = f 1 ), we concatenate the depot features with the global embedding to create the graph context embedding, represented as f c I R 2 p . The concatenation operation is denoted by [ . ; . ] and is shown in Equation (12):
    f c = [ g f ; f 1 ] .
  • Attention Mechanism: The attention mechanism [90] is used to convey the importance of a node to an agent a. The node feature set obtains the keys and values, and the graph context embedding computes the query for agent a, which is standard for all agents.
    q a = θ a g d k × 2 p f c 2 p ,
    k a i = θ a k d k × p f i p , i = { 2 , 3 , , n } ,
    v a i = θ a v d v × p f i p , i = { 2 , 3 , , n } ,
    where d k and d v are the dimension of key and values. A SoftMax is used to compute the attention weights w a i [ 0 , 1 ] , which the agent puts on node i, using:
    w a i = e u a i j e u a j , i = 2 , 3 , , n , j = 2 , 3 , , n ,
    where u a i is the compatibility of the query to be associated with the agent and all its nodes given by
    u a i = q a T k a i d k , i = 2 , 3 , , n .
  • Agent embedding: From the attention weights, we construct the agent embedding using
    h a = j w a j v a j .
In this section, we outline the process for determining the policy that assigns an agent to a specific node. To evaluate the importance of each agent to the node, we utilize the following set of equations. These equations are applied to each agent individually as follows:
k a i = θ a k d k × p f i p , i = 2 , 3 , , n ,
q a = θ a q d k × d v h a d v ,
u a i = q a T k a i d k ,
i m p a i = C tanh ( u a i ) , i = 2 , 3 , , n .
Here, d k is the dimension of new keys; θ a k and θ a q are parameters of neural networks to project the embeddings back to d k dimensions. The importance i m p a i is computed by obtaining u a i for a new round of attention and clipping the result within [ 10 , 10 ] using tanh [80].
Since each agent is limited to visiting only one city, the agent’s importance factor determines which city it visits. A SoftMax function evaluates the probability of an agents likelihood to visit a node:
p a i = e i m p a i a e i m p a i
where p a i is the probability of agent a visiting node i; i m p a i is the importance of node i for agent a; a 1 , 2 , , m ; and i 1 , 2 , , n . The SoftMax function is used for creating a probabilistic policy that can be optimized using gradient-based techniques. In the context of a policy network, the parameters of the SoftMax function are typically learned from data that reflects the characteristics of the agents being modeled. As such, the policy network may need to be retrained when those characteristics change, such as when different turning radius constraints or a different number of agents are introduced.

S-Samples Batch Reinforcement Learning

The parameter θ for our model is estimated by maximizing the expected reward of the policy, i.e.,
θ = arg max θ L R ( θ ) .
L R ( θ ) = E ( G , m , n ) D E λ π ( θ ) R ( λ ) = E ( G , m , n ) D λ π θ ( λ | G , m , n ) R ( λ )
where D is the training set, λ is an assignment of cities representing which agent visits it, R ( λ ) is the reward of the assignment, and λ , π θ is the distribution of the assignments over θ , i.e., π θ ( λ ) = i { 1 , , n } p a i . However, the inner sum over all assignments in Equation (20) is intractable to compute. REINFORCE [91] is used as the estimator for the gradient of the expected reward. An S-sample batch approximation decreases variances in the gradient estimate and speeds up convergence.
L R ( θ ) E ( G , m , n ) D s = 1 S π θ ( λ s | G , m , n ) R ( λ s ) .
The variance during training is decreased by introducing an advantage function A ( λ s ) :
A ( λ s ) R ( λ s ) 1 S z = 1 S R ( λ z ) .
This results in a new optimization function:
θ = arg max θ L R ( θ ) arg max θ E D s = 1 S π θ ( λ s ) A ( λ s ) ,
θ L R ( θ ) E D s = 1 S i = 1 n θ log p a i A ( λ s ) .
The pyDubins [92] code computes the rewards for every assignment, which calculates the DTSP tour lengths and returns the negative of the maximum tour length over all agents as the reward for the assignment.

7. Computational Results

Algorithms were implemented using an Intel NUC Kit (NUC8i7HVK) equipped with an Intel Core i7 processor and 32 GB of RAM. The effectiveness of the proposed methods was evaluated by solving instances from the TSPLIB library [40], which contains sample instances for the TSP and related problems with varying sizes ranging from 16 to 176 cities. To adapt the TSPLIB instances for the MD-GmTSP, each city was associated with a set of h = 8 candidate heading angles. The minimum turning radius ( ρ ) was calculated for each instance based on the layout of the cities. Specifically, ρ was computed as follows:
ρ = 0.1 × max i , j I ( y j y i ) 2 + ( x j x i ) 2
where ( x i , y i ) and ( x j , y j ) are the 2D coordinates of city i and city j, respectively, for all i , j I .

7.1. MILP Results

Each instance was solved using the MILP formulation for MD-GmTSP described in Section 3. The CPLEX 12.10 [87] solver was used with a computational time limit of 10,800 s (3 h) to generate optimal results. Table 1 presents the results of the MD-GmTSP for m = 3 homogeneous vehicles and h = 8 heading angle discretizations. CPLEX was able to generate feasible solutions within the time limit only for instances with fewer than 52 cities and did not find optimal solutions for any of the instances. The optimality gap for the best result, obtained on the Ulysses-22 instance with 22 cities, was 14.09%.

7.2. VNS-Based Heuristics Results

The algorithms discussed in Section 5, including the construction heuristics and the variable neighborhood search (VNS)-based improvement heuristics, were implemented in Python3 and tested on the same set of TSPLIB instances for m = 3 vehicles and h = 8 (discrete heading angles).

7.2.1. Quality of Initial Feasible Solutions

Table 2 presents a comparison of the performance of different construction heuristics ( k = 0 relaxation) on the MD-GmTSP. The construction phase allowed angle optimization only for new unvisited targets to be inserted into tours while keeping the rest of the tour unchanged ( k = 0 ). The initial feasible solutions were generated without imposing any time limit. The results demonstrate that the cheapest construction method produced min-max tours with an average cost that was 11.81% higher than the nearest construction method. However, the nearest and farthest construction methods produced tours with identical min-max costs. Additionally, the cheapest-insertion heuristics generated solutions approximately 2.5 times faster than the nearest-insertion method.

7.2.2. Analysis of VNS Improvements for MD-GmTSP

The impact of different neighborhood structures on improving the MD-GmTSP tours using BVNS, VND, and GVNS heuristics is shown in Table 3, Table 4 and Table 5, respectively. These improvement heuristics were executed with a computational time limit of t m a x = 600  s (10 min) as the stopping criterion, with FirstImprovement used to switch between neighborhoods. The presented results in each row of Table 3, Table 4 and Table 5 represent the average MD-GmTSP solutions for all TSPLIB instances considered. The analysis reveals that the improvement heuristics can achieve solution quality improvements of up to 30.78%. Furthermore, the average time required to solve these instances ranged from 898.6 s to 1439.96 s. Refer to Figure 7 for sample solutions with respect to the GVNS heuristic.
Table 6 presents the best-performing combination of construction and improvement heuristics for each instance. The most effective results are achieved using cheapest insertion and nearest insertion as the construction heuristics, in addition to GVNS with Shake m o d and VND with one-point move as the improvement heuristics. Refer to Figure 8 for sample solutions corresponding to the att48 instance. It should be noted that solutions generated using the cheapest-insert construction with VNS have, on average, costs that are 14.96% higher than the solutions produced by the nearest-insert construction with VNS. On the other hand, the GVNS + Shake m o d scheme resulted in an average improvement of 3.46% over the initial tours, while the VND + one-point move scheme generated an average improvement of 3.32% over the initial solution. A comparison is made between the best solutions obtained from VNS-based heuristics and MILP formulations for each instance in Table 7. It is observed that VNS heuristics generally produce superior solutions in significantly less time than the MILP solutions obtained from CPLEX.

7.3. Learning-Based Approach

In the learning-based approach, training and testing datasets were generated separately for m Dubins agents and n cities. Each city coordinate was uniformly generated at random in the interval ( 0 , 1 ) 2 and paired with h = 8 heading angles in the range [ 0 , 2 π ] . During training, the data were randomly generated at every iteration. To overcome performance limitations, separate models were trained for different numbers of cities and radius of curvature ( ρ ) specific to each instance. A total of 1000 instances were generated for each type and utilized to train the RL model described in Section 6. The RL model was trained using the Adam optimization method with the following hyperparameters: a learning rate of = 1 e 5 , a clipping gradients norm of 3, and S set of 10. Figure 9 illustrates the convergence of a model trained on the pr76 dataset after approximately 2000 iterations during the training process. The dataset consists of 1000 instances, each comprising 76 cities randomly distributed within the original boundaries of the pr76 instances. The tours for these instances are calculated using a Dubins turning radius ( ρ = 980 ).
In order to assess the effectiveness of the RL model, we compare its solutions to those generated by a heuristics-based approach for the MD-GmTSP. The trained RL model is utilized to evaluate the min-max tours for the original pr76 instance, ensuring a valid basis for comparison between the two approaches. Table 8 compares the solutions obtained from the RL model to the best solutions obtained from the MILP and VNS-based formulation. It is important to note that due to the absence of pre-existing baselines for the Dubins MTSP, it was not possible to determine the optimality gap for the solutions obtained from the RL model.
It can be observed that the RL model outperformed both the MILP formulation and the VNS-based heuristics on smaller instances of the problem. Specifically, the RL-based solutions performed better in terms of solution quality for smaller instances. However, as the problem instances became larger, the VNS-based heuristics generated superior solutions compared to the RL model. This indicates that the RL model is especially effective for solving MD-GmTSP instances with a relatively small number of cities and Dubins agents. This effectiveness can be attributed to the fact that the quality of Dubins tours used during the training phase is higher for smaller instances, while the training data quality decreases with the increase in instance size. The RL model’s ability to learn and generalize patterns from the training data enables it to achieve high-quality solutions in these instances. There is scope for additional research and experimentation to explore the potential for enhancing the RL model performance and scalability for larger instances of the MD-GmTSP by leveraging advanced RL techniques and alternative network architectures.

8. Conclusions

In this paper, we explore different methods to solve MinMax routing problems for a team of Dubins vehicles. We formulate the routing problem as a one-in-a-set mTSP with Dubins paths and refer to it as the MD-GmTSP. We then develop a MILP formulation for the MD-SmTSP and solve it on 16 TSPLIB instances (ranging from 16 to 137 cities) using CPLEX for finding optimal solutions. We show that this method does not scale well for large instances involving higher numbers of nodes or robots. As an alternate approach, we develop fast heuristics based on insertion techniques to obtain good, feasible solutions for the MD-GmTSP. The results show that, on average, the nearest-insertion algorithm generates solutions with tour costs 11.81 % lower than the solutions constructed using the cheapest-insertion algorithm. Subsequently, a combination of VNS-based heuristics with several neighborhood search structures is explored to improve the quality of initial feasible solution within a 600 s time limit. The neighborhood structures we studied include 2-opt, 3-opt, k-point, Or-opt, Shake, and the Shake m o d . We solved the TSPLIB instances to check the performance of different combinations of heuristics and identified the best combination to solve MD-GmTSP. Overall, GVNS-based heuristics produced the most promising improvements over the initial solutions. The best-performing combination to solve the MD-GmTSP was to use the nearest-insertion heuristic to construct the initial tour followed by a GVNS-type heuristic to improve the solution quality using neighborhood search structures. Finally, we also explore a learning-based method to generate solutions for the MD-GmTSP. Our architecture consists of a shared graph neural network, with distributed policy networks that extract a common policy for the Dubins multiple traveling salesmen. An S-sample batch reinforcement learning method trains the model, achieving significant improvements in convergence speed and performance. The resulting RL model is used to generate fast feasible paths to the MD-GmTSP, and a comprehensive comparison is presented on the solution quality obtained from each of the other approaches. Overall, the learning based methods work well for smaller instances, while the GVNS based approaches perform better for large instances. Future work can address factors related to planning paths in the presence of obstacles and uncertainties in the position of targets.

Author Contributions

Methodology, A.N. and S.R.; Software, A.N.; Investigation, S.R.; Writing—original draft, A.N.; Writing—review & editing, S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are openly available at http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/. Refer to [40].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lawler, E.L.; Lenstra, J.K.; Rinnooy Kan, A.H.; Shmoys, D.B. Erratum: The traveling salesman problem: A guided tour of combinatorial optimization. J. Oper. Res. Soc. 1986, 37, 655. [Google Scholar] [CrossRef]
  2. Laporte, G. The traveling salesman problem: An overview of exact and approximate algorithms. Eur. J. Oper. Res. 1992, 59, 231–247. [Google Scholar] [CrossRef]
  3. Boscariol, P.; Gasparetto, A.; Scalera, L. Path Planning for Special Robotic Operations. In Robot Design: From Theory to Service Applications; Carbone, G., Laribi, M.A., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 69–95. [Google Scholar] [CrossRef]
  4. Ham, A.M. Integrated scheduling of m-truck, m-drone, and m-depot constrained by time-window, drop-pickup, and m-visit using constraint programming. Transp. Res. Part C Emerg. Technol. 2018, 91, 1–14. [Google Scholar] [CrossRef]
  5. Venkatachalam, S.; Sundar, K.; Rathinam, S. A two-stage approach for routing multiple unmanned aerial vehicles with stochastic fuel consumption. Sensors 2018, 18, 3756. [Google Scholar] [CrossRef] [Green Version]
  6. Cheikhrouhou, O.; Koubâa, A.; Zarrad, A. A cloud based disaster management system. J. Sens. Actuator Netw. 2020, 9, 6. [Google Scholar] [CrossRef] [Green Version]
  7. Bähnemann, R.; Lawrance, N.; Chung, J.J.; Pantic, M.; Siegwart, R.; Nieto, J. Revisiting Boustrophedon Coverage Path Planning as a Generalized Traveling Salesman Problem. In Field and Service Robotics; Ishigami, G., Yoshida, K., Eds.; Springer: Singapore, 2021; pp. 277–290. [Google Scholar]
  8. Conesa-Muñoz, J.; Pajares, G.; Ribeiro, A. Mix-opt: A new route operator for optimal coverage path planning for a fleet in an agricultural environment. Expert Syst. Appl. 2016, 54, 364–378. [Google Scholar] [CrossRef]
  9. Zhao, W.; Meng, Q.; Chung, P.W. A heuristic distributed task allocation method for multivehicle multitask problems and its application to search and rescue scenario. IEEE Trans. Cybern. 2015, 46, 902–915. [Google Scholar] [CrossRef] [Green Version]
  10. Xie, J.; Carrillo, L.R.G.; Jin, L. An Integrated Traveling Salesman and Coverage Path Planning Problem for Unmanned Aircraft Systems. IEEE Control Syst. Lett. 2019, 3, 67–72. [Google Scholar] [CrossRef]
  11. Hari, S.K.K.; Nayak, A.; Rathinam, S. An Approximation Algorithm for a Task Allocation, Sequencing and Scheduling Problem Involving a Human-Robot Team. IEEE Robot. Autom. Lett. 2020, 5, 2146–2153. [Google Scholar] [CrossRef]
  12. Gorenstein, S. Printing press scheduling for multi-edition periodicals. Manag. Sci. 1970, 16, B-373. [Google Scholar] [CrossRef]
  13. Saleh, H.A.; Chelouah, R. The design of the global navigation satellite system surveying networks using genetic algorithms. Eng. Appl. Artif. Intell. 2004, 17, 111–122. [Google Scholar] [CrossRef]
  14. Angel, R.; Caudle, W.; Noonan, R.; Whinston, A. Computer-assisted school bus scheduling. Manag. Sci. 1972, 18, B-279. [Google Scholar] [CrossRef]
  15. Brumitt, B.L.; Stentz, A. Dynamic mission planning for multiple mobile robots. In Proceedings of the IEEE International Conference on Robotics and Automation, Minneapolis, MN, USA, 22–28 April 1996; Volume 3, pp. 2396–2401. [Google Scholar]
  16. Yu, Z.; Jinhai, L.; Guochang, G.; Rubo, Z.; Haiyan, Y. An implementation of evolutionary computation for path planning of cooperative mobile robots. In Proceedings of the 4th World Congress on Intelligent Control and Automation (Cat. No. 02EX527), Shanghai, China, 10–14 June 2002; Volume 3, pp. 1798–1802. [Google Scholar]
  17. Ryan, J.L.; Bailey, T.G.; Moore, J.T.; Carlton, W.B. Reactive tabu search in unmanned aerial reconnaissance simulations. In Proceedings of the 1998 Winter Simulation Conference. Proceedings (Cat. No. 98CH36274), Washington, DC, USA, 13–16 December 1998; Volume 1, pp. 873–879. [Google Scholar]
  18. Dubins, L.E. On curves of minimal length with a constraint on average curvature, and with prescribed initial and terminal positions and tangents. Am. J. Math. 1957, 79, 497–516. [Google Scholar] [CrossRef]
  19. Reeds, J.; Shepp, L. Optimal paths for a car that goes both forwards and backwards. Pac. J. Math. 1990, 145, 367–393. [Google Scholar] [CrossRef] [Green Version]
  20. Sussmann, H.J.; Tang, G. Shortest paths for the Reeds-Shepp car: A worked out example of the use of geometric techniques in nonlinear optimal control. Rutgers Cent. Syst. Control Tech. Rep. 1991, 10, 1–71. [Google Scholar]
  21. Boissonnat, J.D.; Cérézo, A.; Leblond, J. Shortest paths of bounded curvature in the plane. J. Intell. Robot. Syst. 1994, 11, 5–20. [Google Scholar] [CrossRef]
  22. Kolmogorov, A.N.; Mishchenko, Y.F.; Pontryagin, L.S. A Probability Problem of Optimal Control; Technical Report; Joint Publications Research Service: Arlington, VA, USA, 1962. [Google Scholar]
  23. Tang, Z.; Ozguner, U. Motion planning for multitarget surveillance with mobile sensor agents. IEEE Trans. Robot. 2005, 21, 898–908. [Google Scholar] [CrossRef]
  24. Rathinam, S.; Sengupta, R.; Darbha, S. A Resource Allocation Algorithm for Multivehicle Systems with Nonholonomic Constraints. IEEE Trans. Autom. Sci. Eng. 2007, 4, 98–104. [Google Scholar] [CrossRef]
  25. LaValle, S.M. Planning Algorithms; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
  26. Ny, J.; Feron, E.; Frazzoli, E. On the Dubins traveling salesman problem. IEEE Trans. Autom. Control 2011, 57, 265–270. [Google Scholar]
  27. Manyam, S.G.; Rathinam, S.; Darbha, S.; Obermeyer, K.J. Lower bounds for a vehicle routing problem with motion constraints. Int. J. Robot. Autom 2015, 30. [Google Scholar] [CrossRef]
  28. Ma, X.; Castanon, D.A. Receding horizon planning for Dubins traveling salesman problems. In Proceedings of the 45th IEEE Conference on Decision and Control, San Diego, CA, USA, 13–15 December 2006; pp. 5453–5458. [Google Scholar]
  29. Savla, K.; Frazzoli, E.; Bullo, F. Traveling salesperson problems for the Dubins vehicle. IEEE Trans. Autom. Control 2008, 53, 1378–1391. [Google Scholar] [CrossRef]
  30. Yadlapalli, S.; Malik, W.; Darbha, S.; Pachter, M. A Lagrangian-based algorithm for a multiple depot, multiple traveling salesmen problem. Nonlinear Anal. Real World Appl. 2009, 10, 1990–1999. [Google Scholar] [CrossRef]
  31. Macharet, D.G.; Campos, M.F. An orientation assignment heuristic to the Dubins traveling salesman problem. In Proceedings of the Ibero-American Conference on Artificial Intelligence, Santiago de, Chile, Chile, 24–27 November 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 457–468. [Google Scholar]
  32. Sujit, P.; Hudzietz, B.; Saripalli, S. Route planning for angle constrained terrain mapping using an unmanned aerial vehicle. J. Intell. Robot. Syst. 2013, 69, 273–283. [Google Scholar] [CrossRef]
  33. Isaiah, P.; Shima, T. Motion planning algorithms for the Dubins travelling salesperson problem. Automatica 2015, 53, 247–255. [Google Scholar] [CrossRef]
  34. Babel, L. New heuristic algorithms for the Dubins traveling salesman problem. J. Heuristics 2020, 26, 503–530. [Google Scholar] [CrossRef] [Green Version]
  35. Manyam, S.G.; Rathinam, S. On tightly bounding the dubins traveling salesman’s optimum. J. Dyn. Syst. Meas. Control 2018, vol. 140, 071013. [Google Scholar] [CrossRef] [Green Version]
  36. Manyam, S.G.; Rathinam, S.; Darbha, S. Computation of lower bounds for a multiple depot, multiple vehicle routing problem with motion constraints. J. Dyn. Syst. Meas. Control 2015, 137, 094501. [Google Scholar] [CrossRef]
  37. Cohen, I.; Epstein, C.; Shima, T. On the discretized dubins traveling salesman problem. IISE Trans. 2017, 49, 238–254. [Google Scholar] [CrossRef]
  38. Oberlin, P.; Rathinam, S.; Darbha, S. Today’s traveling salesman problem. IEEE Robot. Autom. Mag. 2010, 17, 70–77. [Google Scholar] [CrossRef]
  39. Hansen, P.; Mladenović, N. Variable neighborhood search: Principles and applications. Eur. J. Oper. Res. 2001, 130, 449–467. [Google Scholar] [CrossRef]
  40. Reinhelt, G. {TSPLIB}: A Library of Sample Instances for the TSP (and Related Problems) from Various Sources and of Various Types. 2014. Available online: http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/ (accessed on 1 January 2023).
  41. Applegate, D.L.; Bixby, R.E.; Chvatal, V.; Cook, W.J. The Traveling Salesman Problem: A Computational Study (Princeton Series in Applied Mathematics); Princeton University Press: Princeton, NJ, USA, 2007. [Google Scholar]
  42. Vazirani, V.V. Approximation Algorithms; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  43. Rosenkrantz, D.J.; Stearns, R.E.; Lewis, P.M. An analysis of several heuristics for the traveling salesman problem. In Fundamental Problems in Computing: Essays in Honor of Professor Daniel J. Rosenkrantz; Ravi, S.S., Shukla, S.K., Eds.; Springer: Dordrecht, The Netherlands, 2009; pp. 45–69. [Google Scholar] [CrossRef]
  44. Manyam, S.; Rathinam, S.; Casbeer, D. Dubins paths through a sequence of points: Lower and upper bounds. In Proceedings of the 2016 International Conference on Unmanned Aircraft Systems (ICUAS), Arlington, VA, USA, 7–10 June 2016; pp. 284–291. [Google Scholar]
  45. Sundar, K.; Rathinam, S. Algorithms for routing an unmanned aerial vehicle in the presence of refueling depots. IEEE Trans. Autom. Sci. Eng. 2014, 11, 287–294. [Google Scholar] [CrossRef] [Green Version]
  46. Sundar, K.; Rathinam, S. An exact algorithm for a heterogeneous, multiple depot, multiple traveling salesman problem. In Proceedings of the 2015 International Conference on Unmanned Aircraft Systems, Denver, CO, USA, 9–12 June 2015; pp. 366–371. [Google Scholar]
  47. Sundar, K.; Venkatachalam, S.; Rathinam, S. An Exact Algorithm for a Fuel-Constrained Autonomous Vehicle Path Planning Problem. arXiv 2016, arXiv:1604.08464. [Google Scholar]
  48. Lo, K.M.; Yi, W.Y.; Wong, P.K.; Leung, K.S.; Leung, Y.; Mak, S.T. A genetic algorithm with new local operators for multiple traveling salesman problems. Int. J. Comput. Intell. Syst. 2018, 11, 692–705. [Google Scholar] [CrossRef] [Green Version]
  49. Bao, X.; Wang, G.; Xu, L.; Wang, Z. Solving the Min-Max Clustered Traveling Salesmen Problem Based on Genetic Algorithm. Biomimetics 2023, 8, 238. [Google Scholar] [CrossRef]
  50. Zhang, Y.; Han, X.; Dong, Y.; Xie, J.; Xie, G.; Xu, X. A novel state transition simulated annealing algorithm for the multiple traveling salesmen problem. J. Supercomput. 2021, 77, 11827–11852. [Google Scholar] [CrossRef]
  51. He, P.; Hao, J.K. Memetic search for the minmax multiple traveling salesman problem with single and multiple depots. Eur. J. Oper. Res. 2023, 307, 1055–1070. [Google Scholar] [CrossRef]
  52. He, P.; Hao, J.K. Hybrid search with neighborhood reduction for the multiple traveling salesman problem. Comput. Oper. Res. 2022, 142, 105726. [Google Scholar] [CrossRef]
  53. Venkatesh, P.; Singh, A. Two metaheuristic approaches for the multiple traveling salesperson problem. Appl. Soft Comput. 2015, 26, 74–89. [Google Scholar] [CrossRef]
  54. Hamza, A.; Darwish, A.H.; Rihawi, O. A New Local Search for the Bees Algorithm to Optimize Multiple Traveling Salesman Problem. Intell. Syst. Appl. 2023, 18, 200242. [Google Scholar] [CrossRef]
  55. Rathinam, S.; Rajagopal, H. Optimizing Mission Times for Multiple Unmanned Vehicles with Vehicle-Target Assignment Constraints. In Proceedings of the AIAA SCITECH 2022 Forum, San Diego, CA, USA, 3–7 January 2022; p. 2527. [Google Scholar]
  56. Patil, A.; Bae, J.; Park, M. An algorithm for task allocation and planning for a heterogeneous multi-robot system to minimize the last task completion time. Sensors 2022, 22, 5637. [Google Scholar] [CrossRef]
  57. Dedeurwaerder, B.; Louis, S.J. A Meta Heuristic Genetic Algorithm for Multi-Depot Routing in Autonomous Bridge Inspection. In Proceedings of the 2022 IEEE Symposium Series on Computational Intelligence (SSCI), Singapore, 4–7 December 2022; pp. 1194–1201. [Google Scholar]
  58. Park, J.; Kwon, C.; Park, J. Learn to Solve the Min-max Multiple Traveling Salesmen Problem with Reinforcement Learning. In Proceedings of the 2023 International Conference on Autonomous Agents and Multiagent Systems, London, UK, 29 May–2 June 2023; pp. 878–886. [Google Scholar]
  59. Frederickson, G.N.; Hecht, M.S.; Kim, C.E. Approximation algorithms for some routing problems. In Proceedings of the 17th Annual Symposium on Foundations of Computer Science (sfcs 1976), Houston, TX, USA, 25–27 October 1976; pp. 216–227. [Google Scholar] [CrossRef]
  60. Yadlapalli, S.; Rathinam, S.; Darbha, S. 3-Approximation algorithm for a two depot, heterogeneous traveling salesman problem. Optim. Lett. 2010, 6, 141–152. [Google Scholar] [CrossRef]
  61. Chour, K.; Rathinam, S.; Ravi, R. S*: A Heuristic Information-Based Approximation Framework for Multi-Goal Path Finding. In Proceedings of the International Conference on Automated Planning and Scheduling, Guangzhou, China, 7–12 June 2021; Volume 31, pp. 85–93. [Google Scholar]
  62. Carlsson, J.G.; Ge, D.; Subramaniam, A.; Wu, A. Solving Min-Max Multi-Depot Vehicle Routing Problem. In Proceedings of the Lectures on Global Optimization (Volume 55 in the Series Fields Institute Communications); American Mathematical Society: Providence, RI, USA, 2007. [Google Scholar]
  63. Venkata Narasimha, K.; Kivelevitch, E.; Sharma, B.; Kumar, M. An ant colony optimization technique for solving Min-max Multi-Depot Vehicle Routing Problem. Swarm Evol. Comput. 2013, 13, 63–73. [Google Scholar] [CrossRef]
  64. Lu, L.C.; Yue, T.W. Mission-oriented ant-team ACO for min-max MTSP. Appl. Soft Comput. 2019, 76, 436–444. [Google Scholar] [CrossRef]
  65. Liu, J.; Zhang, Y.; Wang, X.; Xu, C.; Ma, X. Min-max Path Planning of Multiple UAVs for Autonomous Inspection. In Proceedings of the 2020 International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 21–23 October 2020; pp. 1058–1063. [Google Scholar] [CrossRef]
  66. Wang, X.; Golden, B.H.; Wasil, E.A. The min-max multi-depot vehicle routing problem: Heuristics and computational results. J. Oper. Res. Soc. 2015, 66, 1430–1441. [Google Scholar] [CrossRef]
  67. Scott, D.; Manyam, S.G.; Casbeer, D.W.; Kumar, M. Market Approach to Length Constrained Min-Max Multiple Depot Multiple Traveling Salesman Problem. In Proceedings of the 2020 American Control Conference (ACC), Denver, CO, USA, 1–3 July 2020; pp. 138–143. [Google Scholar] [CrossRef]
  68. Prasad, A.; Sundaram, S.; Choi, H.L. Min-Max Tours for Task Allocation to Heterogeneous Agents. In Proceedings of the 2018 IEEE Conference on Decision and Control (CDC), Miami Beach, FL, USA, 17–19 December 2018; pp. 1706–1711. [Google Scholar] [CrossRef] [Green Version]
  69. Banik, S.; Rathinam, S.; Sujit, P. Min-Max Path Planning Algorithms for Heterogeneous, Autonomous Underwater Vehicles. In Proceedings of the 2018 IEEE/OES Autonomous Underwater Vehicle Workshop (AUV), Porto, Portugal, 6–9 November 2018; pp. 1–6. [Google Scholar] [CrossRef]
  70. Ding, L.; Zhao, D.; Ma, H.; Wang, H.; Liu, L. Energy-Efficient Min-Max Planning of Heterogeneous Tasks with Multiple UAVs. In Proceedings of the 2018 IEEE 24th International Conference on Parallel and Distributed Systems (ICPADS), Singapore, 11–13 December 2018; pp. 339–346. [Google Scholar] [CrossRef]
  71. Deng, L.; Xu, W.; Liang, W.; Peng, J.; Zhou, Y.; Duan, L.; Das, S.K. Approximation Algorithms for the Min-Max Cycle Cover Problem With Neighborhoods. IEEE/ACM Trans. Netw. 2020, 28, 1845–1858. [Google Scholar] [CrossRef]
  72. Kara, I.; Bektas, T. Integer linear programming formulations of multiple salesman problems and its variations. Eur. J. Oper. Res. 2006, 174, 1449–1458. [Google Scholar] [CrossRef]
  73. Hansen, P.; Mladenović, N.; Pérez, J.A.M. Variable neighbourhood search: Methods and applications. Ann. Oper. Res. 2010, 175, 367–407. [Google Scholar] [CrossRef]
  74. Croes, G.A. A method for solving traveling-salesman problems. Oper. Res. 1958, 6, 791–812. [Google Scholar] [CrossRef]
  75. Lin, S. Computer solutions of the traveling salesman problem. Bell Syst. Tech. J. 1965, 44, 2245–2269. [Google Scholar] [CrossRef]
  76. Hopfield, J.J.; Tank, D.W. “Neural” computation of decisions in optimization problems. Biol. Cybern. 1985, 52, 141–152. [Google Scholar] [CrossRef]
  77. Vinyals, O.; Fortunato, M.; Jaitly, N. Pointer networks. arXiv 2015, arXiv:1506.03134. [Google Scholar]
  78. Deudon, M.; Cournut, P.; Lacoste, A.; Adulyasak, Y.; Rousseau, L.M. Learning heuristics for the tsp by policy gradient. In Proceedings of the International Conference on the Integration of Constraint Programming, Artificial Intelligence, and Operations Research, Delft, The Netherlands, 26–29 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 170–181. [Google Scholar]
  79. Kool, W.; Van Hoof, H.; Welling, M. Attention, learn to solve routing problems! arXiv 2018, arXiv:1803.08475. [Google Scholar]
  80. Bello, I.; Pham, H.; Le, Q.V.; Norouzi, M.; Bengio, S. Neural combinatorial optimization with reinforcement learning. arXiv 2016, arXiv:1611.09940. [Google Scholar]
  81. Nazari, M.; Oroojlooy, A.; Snyder, L.; Takác, M. Reinforcement learning for solving the vehicle routing problem. In Proceedings of the 32nd Annual Conference on Neural Information Processing Systems (NIPS 2018), Montreal, QC, Canada, 2–8 December 2018. [Google Scholar]
  82. Hu, Y.; Yao, Y.; Lee, W.S. A reinforcement learning approach for optimizing multiple traveling salesman problems over graphs. Knowl.-Based Syst. 2020, 204, 106244. [Google Scholar] [CrossRef]
  83. Park, J.; Bakhtiyar, S.; Park, J. ScheduleNet: Learn to solve multi-agent scheduling problems with reinforcement learning. arXiv 2021, arXiv:2106.03051. [Google Scholar]
  84. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How powerful are graph neural networks? arXiv 2018, arXiv:1810.00826. [Google Scholar]
  85. Rosenkrantz, D.J.; Stearns, R.E.; Lewis, P.M., II. An analysis of several heuristics for the traveling salesman problem. SIAM J. Comput. 1977, 6, 563–581. [Google Scholar] [CrossRef]
  86. Helsgaun, K. An effective implementation of the Lin–Kernighan traveling salesman heuristic. Eur. J. Oper. Res. 2000, 126, 106–130. [Google Scholar] [CrossRef] [Green Version]
  87. IBM ILOG CPLEX Optimizer. En Ligne. 2012. Available online: https://www.ibm.com/products/ilog-cplex-optimization-studio/cplex-optimizer (accessed on 1 January 2023).
  88. Applegate, D.; Bixby, R.; Chvatal, V.; Cook, W. Concorde TSP Solver. 2006. Available online: https://www.math.uwaterloo.ca/tsp/concorde (accessed on 1 January 2023).
  89. Helsgaun, K. An Extension of the Lin-Kernighan-Helsgaun TSP Solver for Constrained Traveling Salesman and Vehicle Routing Problems; Roskilde University: Roskilde, Denmark, 2017; pp. 24–50. [Google Scholar]
  90. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017.
  91. Williams, R.J. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 1992, 8, 229–256. [Google Scholar] [CrossRef] [Green Version]
  92. Walker, A. pyDubins. 2020. Available online: https://github.com/AndrewWalker/pydubins (accessed on 1 January 2023).
Figure 1. Examples of Dubins Paths.
Figure 1. Examples of Dubins Paths.
Sensors 23 06432 g001
Figure 2. Dubins path between targets with discretized heading angles.
Figure 2. Dubins path between targets with discretized heading angles.
Sensors 23 06432 g002
Figure 3. Different Shake neighborhoods. The black and the red paths indicate the tours of the vehicles involved in the shake process.
Figure 3. Different Shake neighborhoods. The black and the red paths indicate the tours of the vehicles involved in the shake process.
Sensors 23 06432 g003
Figure 4. A sample 2-opt move for a given tour.
Figure 4. A sample 2-opt move for a given tour.
Sensors 23 06432 g004
Figure 5. All possible combinations of 3-Opt moves for a given tour.
Figure 5. All possible combinations of 3-Opt moves for a given tour.
Sensors 23 06432 g005
Figure 6. Local search operators. The black and the red paths denote the tours of the vehicles involved in each move. The blue arrow in the Or-opt2 move shows the string of two adjacent nodes that is transferred between tours.
Figure 6. Local search operators. The black and the red paths denote the tours of the vehicles involved in each move. The blue arrow in the Or-opt2 move shows the string of two adjacent nodes that is transferred between tours.
Sensors 23 06432 g006
Figure 7. Feasible MD-GmTSP paths for the Ulysses22 instance for m = 3 vehicles and h = 8 angle discretizations obtained from different tour construction and GVNS heuristics. The vehicle paths are differentiated by the colors of the paths.
Figure 7. Feasible MD-GmTSP paths for the Ulysses22 instance for m = 3 vehicles and h = 8 angle discretizations obtained from different tour construction and GVNS heuristics. The vehicle paths are differentiated by the colors of the paths.
Sensors 23 06432 g007aSensors 23 06432 g007b
Figure 8. Feasible MD-GmTSP paths for the att48 instance for m = 3 vehicles and h = 8 angle discretizations obtained from tour construction and improvement heuristics. The vehicle paths are differentiated by the colors of the paths.
Figure 8. Feasible MD-GmTSP paths for the att48 instance for m = 3 vehicles and h = 8 angle discretizations obtained from tour construction and improvement heuristics. The vehicle paths are differentiated by the colors of the paths.
Sensors 23 06432 g008
Figure 9. Learning curve for Dubins pr 76.
Figure 9. Learning curve for Dubins pr 76.
Sensors 23 06432 g009
Table 1. MD-GmTSP results for m = 3 vehicles and h = 8 discretizations on TSPLIB instances using CPLEX 12.10 with a runtime of 3 h (10,800 s). Instances that CPLEX could not solve are indicated by *.
Table 1. MD-GmTSP results for m = 3 vehicles and h = 8 discretizations on TSPLIB instances using CPLEX 12.10 with a runtime of 3 h (10,800 s). Instances that CPLEX could not solve are indicated by *.
InstanceTargetsTurning Radius ( ρ )Minmax CostOptimality Gap (%)
att484739428,315.348642.23
berlin5251555433.47443.54
ch130 *12939--
eil101 *1004--
eil51504333.728546.46
eil76 *754--
gr137 *1364--
gr96 *954--
kroA100 *99200--
pr136 *135516--
pr76 *75980--
rat99 *985--
rd100 *9957--
st70 *696--
ulysses1615257.748719.27
ulysses2221258.038614.09
Table 2. Comparison of MinMax costs of feasible solutions obtained from insertion heuristics for the MD-GmTSP for m = 3 vehicles, C = 8 angle discretizations, and k = 0 relaxation. Times are specified in secs.
Table 2. Comparison of MinMax costs of feasible solutions obtained from insertion heuristics for the MD-GmTSP for m = 3 vehicles, C = 8 angle discretizations, and k = 0 relaxation. Times are specified in secs.
InstanceRadiusCheapest InsertionFarthest InsertionNearest Insertion c c c n (%)
Cost ( c c )Time ( t c )Cost ( c f )Time ( t f )Cost ( c n )Time ( t n )
att4839432,336.3943.3828,197.11108.7228,197.11107.47114.68
berlin52556025.9954.055453.38138.375453.38138.63110.50
ch130396472.87751.585998.741904.175998.741944.28107.90
eil1014554.77361.14450.26898.30450.26900.88123.21
eil514308.3050.70277.73126.92277.73128.76111.01
eil764416.51157.05345.66394.27345.66402.35120.50
gr1374847.06850.68701.392154.25701.392217.90120.77
gr964551.00311.42488.93770.56488.93780.32112.69
kroA10020027,236.35355.9922,592.95872.5322,592.95879.44120.55
pr13651698,637.39842.1582,162.292130.3082,162.292191.75120.05
pr7698097,434.78156.9499,023.43398.9899,023.43400.0198.40
rat995798.94337.41819.30841.77819.30850.6597.51
rd100578679.56343.936833.40870.016833.40869.33127.02
st706593.76125.90550.76310.80550.76318.95107.81
ulysses16266.402.5776.345.5076.345.9486.97
ulysses22288.005.2680.4912.7880.4912.94109.33
Table 3. Influence of different neighborhood structures in the BVNS heuristic on improving the MD-GmTSP solutions with t m a x = 600 s time limit.
Table 3. Influence of different neighborhood structures in the BVNS heuristic on improving the MD-GmTSP solutions with t m a x = 600 s time limit.
Construction
Heuristic
N 1 N 2 Average
Improvement
(%)
Max
Improvement
(%)
Standard
Deviation
Average
Computation
Time (s)
MoveInsertionLocal Search
CheapestShakeGreedy2-opt3.20121.6276.225900.413
CheapestShakeGreedy3-opt2.75019.4565.648909.045
CheapestShakeRandom2-opt1.98717.3995.309899.961
CheapestShakeRandom3-opt2.09117.0825.244907.608
CheapestShake m o d Greedy2-opt2.80215.7555.392900.571
CheapestShake m o d Greedy3-opt2.42313.5084.017916.210
CheapestShake m o d Random2-opt1.82314.4844.640900.069
CheapestShake m o d Random3-opt1.93814.4843.921919.457
FarthestShakeGreedy2-opt2.78524.7367.6091354.735
FarthestShakeGreedy3-opt2.58525.3426.9691381.107
FarthestShakeRandom2-opt1.70318.9395.0431359.844
FarthestShakeRandom3-opt1.65020.9885.3311371.328
FarthestShake m o d Greedy2-opt2.52626.5896.9311366.066
FarthestShake m o d Greedy3-opt1.98626.5676.5981374.865
FarthestShake m o d Random2-opt1.47021.8695.4551362.314
FarthestShake m o d Random3-opt1.13518.1584.5401373.770
NearestShakeGreedy2-opt2.75025.1547.0151355.997
NearestShakeGreedy3-opt2.55025.0256.7811376.476
NearestShakeRandom2-opt1.74520.9885.4151363.908
NearestShakeRandom3-opt1.89620.9885.6001368.852
NearestShake m o d Greedy2-opt2.62726.5896.9321370.339
NearestShake m o d Greedy3-opt1.63526.1426.5351375.723
NearestShake m o d Random2-opt1.61720.3515.1731359.805
NearestShake m o d Random3-opt1.47716.3644.3351378.464
Table 4. Influence of different neighborhood structures in VND heuristic on improving the MD-GmTSP tours with t m a x = 600 s time limit.
Table 4. Influence of different neighborhood structures in VND heuristic on improving the MD-GmTSP tours with t m a x = 600 s time limit.
Construction
Heuristic
N 1 N 2 Average
Improvement
(%)
Max
Improvement
(%)
Standard
Deviation
Average
Computation
Time (s)
MoveInsertionLocal
Search
Cheapest2-moveGreedy2-opt4.46126.6707.283899.159
Cheapest2-moveGreedy3-opt3.37721.6756.248912.802
Cheapest2-moveRandom2-opt2.19118.3155.487898.602
Cheapest2-moveRandom3-opt1.91216.7865.201916.977
Cheapest1-moveGreedy2-opt5.64429.8948.035899.059
Cheapest1-moveGreedy3-opt4.60326.1057.240914.916
Cheapest1-moveRandom2-opt2.54719.8815.932899.179
Cheapest1-moveRandom3-opt2.59617.5575.607913.728
Farthest2-moveGreedy2-opt3.48325.3427.8021372.255
Farthest2-moveGreedy3-opt2.99725.2026.8681373.696
Farthest2-moveRandom2-opt1.92220.9885.6331365.133
Farthest2-moveRandom3-opt1.95820.9885.5141381.819
Farthest1-moveGreedy2-opt3.72925.6228.0501356.863
Farthest1-moveGreedy3-opt3.69125.6227.4861374.722
Farthest1-moveRandom2-opt2.20023.3506.1111358.016
Farthest1-moveRandom3-opt2.18023.9136.2251376.558
Nearest2-moveGreedy2-opt3.46225.3427.5981358.205
Nearest2-moveGreedy3-opt3.18125.3427.0291374.512
Nearest2-moveRandom2-opt1.77520.9885.4471362.595
Nearest2-moveRandom3-opt1.82220.9885.5021379.830
Nearest1-moveGreedy2-opt4.49426.5898.4301360.322
Nearest1-moveGreedy3-opt4.16725.6227.5351372.429
Nearest1-moveRandom2-opt2.06823.9906.2321367.803
Nearest1-moveRandom3-opt1.85523.3505.9111375.266
Table 5. Influence of different neighborhood structures in GVNS on improving the MD-GmTSP tours with t m a x = 600 s time limit.
Table 5. Influence of different neighborhood structures in GVNS on improving the MD-GmTSP tours with t m a x = 600 s time limit.
Construction
Heuristic
N 1 N 2 Average
Improvement
(%)
Max
Improvement
(%)
Standard
Deviation
Average
Computation
Time (s)
MoveInsertionLocal
Search
CheapestShakeGreedyVND s e q 5.86222.8406.842915.206
CheapestShakeRandomVND s e q 2.24121.1216.233969.554
CheapestShake m o d GreedyVND s e q 6.16523.1417.486919.908
CheapestShake m o d RandomVND s e q 2.90221.6666.005974.873
FarthestShakeGreedyVND s e q 2.59526.5897.2081370.046
FarthestShakeRandomVND s e q 2.39926.5897.0891406.246
FarthestShake m o d GreedyVND s e q 2.99626.5897.2261368.846
FarthestShake m o d RandomVND s e q 2.81325.5487.0521409.714
NearestShakeGreedyVND s e q 3.55524.3637.1831368.566
NearestShakeRandomVND s e q 2.17623.5066.3491412.890
NearestShake m o d GreedyVND s e q 4.22225.7997.5351391.104
NearestShake m o d RandomVND s e q 2.07823.1706.1611433.516
Table 6. Best-performing combination of heuristics for the MD-GmTSP. Time is specified in secs.
Table 6. Best-performing combination of heuristics for the MD-GmTSP. Time is specified in secs.
InstanceCitiesRadiusConstruction PhaseImprovement PhaseImprovement (%)
HeuristicCostTimeScheme N 1 N 2 CostTime
MoveInsertionMove
att4848394Cheapest32,336.3943.90GVNSShake m o d GreedyVND S e q 25,775.29646.8320.29
berlin525255Nearest5453.38139.50GVNSShakeGreedyVND S e q 5201.58744.594.62
ch13013039Nearest5998.741931.72VND1-pointGreedy2-opt5839.332533.202.66
eil1011014Nearest450.26901.52VND2-pointGreedy2-opt410.151502.138.91
eil51514Nearest277.73131.30VND1-pointGreedy2-opt274.60731.931.13
eil76764Nearest345.66397.13VND1-pointGreedy3-opt344.291016.680.40
gr1371374Nearest701.392196.77GVNSShake m o d GreedyVND S e q 668.912939.974.63
gr96964Nearest488.93782.38VND1-pointGreedy3-opt457.931401.156.34
kroA100100200Nearest22,592.95904.13GVNSShake m o d GreedyVND S e q 22,321.631536.181.20
pr136136516Farthest82,162.292180.81GVNSShake m o d GreedyVND S e q 81,379.622786.210.95
pr7676980Nearest99,023.43393.95GVNSShake m o d GreedyVND S e q 90,809.31997.128.30
rat99995Cheapest798.94334.75VND1-pointGreedy3-opt787.09952.941.48
rd10010057Nearest6833.40867.27GVNSShake m o d GreedyVND S e q 6827.481472.390.09
st70706Farthest550.76318.94VND1-pointGreedy3-opt550.75951.810.00
ulysses16162Cheapest66.402.24GVNSShake m o d GreedyVND S e q 56.04614.0915.60
ulysses22222Nearest80.4912.98VND1-pointGreedy2-opt60.31613.1125.08
Table 7. Comparison for best solutions obtained from MILP formulation ( t m a x = 10,800 s) against the VNS-based heuristics ( t m a x = 600 s).
Table 7. Comparison for best solutions obtained from MILP formulation ( t m a x = 10,800 s) against the VNS-based heuristics ( t m a x = 600 s).
InstanceCitiesRadiusVNS SolutionMILP Solution C m C h C h 100
Cost ( C h )Time ( T h )Cost ( C m )Optimality Gap
att484839425,775.29690.7228,315.3542.239.85
berlin5252555201.58884.095433.4743.544.46
ch130130395839.334464.92---
eil1011014410.152403.65---
eil51514274.60863.23333.7346.4621.53
eil76764344.291413.81---
gr1371374668.915136.74---
gr96964457.932183.53---
kroA10010020022,321.632440.30---
pr13613651681,379.624967.02---
pr767698090,809.311391.07---
rat99995787.091287.69---
rd100100576827.482339.65---
st70706550.751270.75---
ulysses1616256.04616.3357.7519.273.05
ulysses2222260.31626.0958.0414.09−3.76
Table 8. Comparison for best solutions obtained from MILP formulation ( t m a x = 10,800 s) against the VNS-based heuristics ( t m a x = 600 s).
Table 8. Comparison for best solutions obtained from MILP formulation ( t m a x = 10,800 s) against the VNS-based heuristics ( t m a x = 600 s).
InstanceCitiesRadiusVNS SolutionMILP SolutionRL Solution
Cost ( C h )Cost ( C m ) C m C h Cost ( C R L ) C R L C h
att484839425,775.2928,315.351.1022,846.380.89
eil51514274.60333.731.22252.640.92
pr767698090,809.31--98,174.431.08
rd100100576827.48--7549.841.11
st70706550.75--564.641.03
ulysses1616256.0457.751.0354.860.98
ulysses2222260.3158.040.9657.980.96
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nayak, A.; Rathinam, S. Heuristics and Learning Models for Dubins MinMax Traveling Salesman Problem. Sensors 2023, 23, 6432. https://doi.org/10.3390/s23146432

AMA Style

Nayak A, Rathinam S. Heuristics and Learning Models for Dubins MinMax Traveling Salesman Problem. Sensors. 2023; 23(14):6432. https://doi.org/10.3390/s23146432

Chicago/Turabian Style

Nayak, Abhishek, and Sivakumar Rathinam. 2023. "Heuristics and Learning Models for Dubins MinMax Traveling Salesman Problem" Sensors 23, no. 14: 6432. https://doi.org/10.3390/s23146432

APA Style

Nayak, A., & Rathinam, S. (2023). Heuristics and Learning Models for Dubins MinMax Traveling Salesman Problem. Sensors, 23(14), 6432. https://doi.org/10.3390/s23146432

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop