Next Article in Journal
Optimizing the Electrode Geometry of an In-Plane Unimorph Piezoelectric Microactuator for Maximum Deflection
Previous Article in Journal
A Spatiotemporal Locomotive Axle Temperature Prediction Approach Based on Ensemble Graph Convolutional Recurrent Unit Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AutoRL-Sim: Automated Reinforcement Learning Simulator for Combinatorial Optimization Problems

by
Gleice Kelly Barbosa Souza
1 and
André Luiz Carvalho Ottoni
2,*
1
Technologic and Exact Center, Federal University of Recôncavo da Bahia, R. Rui Barbosa, Cruz das Almas 44380-000, Bahia, Brazil
2
Department of Computing, Federal University of Ouro Preto, Campus Morro do Cruzeiro, Ouro Preto 35400-000, Minas Gerais, Brazil
*
Author to whom correspondence should be addressed.
Modelling 2024, 5(3), 1056-1083; https://doi.org/10.3390/modelling5030055
Submission received: 21 April 2024 / Revised: 1 August 2024 / Accepted: 22 August 2024 / Published: 26 August 2024

Abstract

:
Reinforcement learning is a crucial area of machine learning, with a wide range of applications. To conduct experiments in this research field, it is necessary to define the algorithms and parameters to be applied. However, this task can be complex because of the variety of possible configurations. In this sense, the adoption of AutoRL systems can automate the selection of these configurations, simplifying the experimental process. In this context, this work aims to propose a simulation environment for combinatorial optimization problems using AutoRL. The AutoRL-Sim includes several experimentation modules that cover studies on the symmetric traveling salesman problem, the asymmetric traveling salesman problem, and the sequential ordering problem. Furthermore, parameter optimization is performed using response surface models. The AutoRL-Sim simulator allows users to conduct experiments in a more practical way, without the need to worry about implementation. Additionally, they have the ability to analyze post-experiment data or save them for future analysis.

1. Introduction

One of the most critical points of machine learning (ML) is selecting the appropriate algorithm and parameter tuning for the process [1,2,3,4,5]. In this sense, for each dataset and application, there may be an ideal combination of algorithm and parameters, which can make manual evaluation a costly task [6,7,8,9,10]. In the literature, there are several approaches that can be employed to optimize these ML parameters, such as response surface methodology (RSM) [11], variable neighborhood search (VNS) [12], reactive search [13], projective simulation [7], empirical methods [14], random search [15], grid search [16], Bayesian optimization [17], and Hyperband [18].
It should also be noted that defining the initial conditions of the experiment may involve multiple steps and verifications, requiring considerable time from the experimenter. However, even with all this effort, there is still the possibility that the chosen configuration is not the most appropriate for the problem at hand. Therefore, one way to reduce the time spent starting the experiment and help the artificial intelligence user is to automate the ML process. In this context, automated machine learning (AutoML) emerges, which can be used to optimize [2,19,20] and recommend [1,9,21] parameters and algorithms. In AutoML, the objective is to minimize human work by automatically determining the optimal algorithms and/or parameters for the tasks under analysis [1,2,3,6,22].
When AutoML is applied to reinforcement learning (RL), it is called AutoRL (automated reinforcement learning) [23,24]. RL is a field of artificial intelligence and machine learning that covers areas such as statistics and computer science. In this type of learning, an agent is immersed in a dynamic environment without previously knowing which actions should be selected [25,26]. Thus, the agent must learn about the environment through its interactions, using trial and error to determine which actions yield better results [13,25,26]. In AutoRL, in the same way as in conventional RL, the agent has no prior knowledge and learns from the reinforcements received while interacting with the environment. It is worth mentioning that, while in conventional RL, the learning conditions are defined by the experimenter, in AutoRL, the experimental process is automated [7,27,28]. However, the literature still lacks dedicated AutoRL frameworks for combinatorial optimization problems.
Reinforcement learning has relevant applications in combinatorial optimization, such as the vehicle routing problem (VRP) [29], the symmetric traveling salesman problem (TSP) [30,31], the asymmetric traveling salesman problem (ATSP) [30], minimum vertex cover (MVC) [31,32], maximum cut (Max-Cut) [31,32], and the bin-packing problem (BPP) [32]. On the other hand, it can be challenging for users, especially beginners, to develop algorithms and configure initial RL experiment settings for combinatorial optimization. In this sense, several studies in the literature have already addressed the challenge of tuning RL parameters for combinatorial optimization problems. The paper by [11] proposed the application of response surface methodology to model the behavior of learning rate and discount factor parameters in RL performance to solve the TSP. Following this line, the study by [33] investigated the influence of RL parameters in resolving instances of the sequential ordering problem. In another work [34], the effects of RL parameters on a variation of TSP with refueling on routes were evaluated. Faced with this challenge, one way to overcome this problem would be to develop AutoRL simulation environments.
The development of simulation environments and frameworks in order to facilitate experiments for users is a frequent topic in several studies [35,36,37,38,39,40,41]. In the context of reinforcement learning, some work has already been proposed in this area, including applications in combinatorial optimization [39], mobile robotics [42], environmental modeling [43], scheduling problems [44], and automation in test selection for continuous integration [45]. The literature also already has some proposals for AutoRL simulators. For example, the authors of [38] develop a simulator for pH control. In [41], the authors present a framework for automated circuit sizing. In another paper, the authors of [24] develop a framework to automate the creation of pipelines for AutoRL. However, there is still a lack of studies that address the development of simulation environments aimed at tuning RL parameters for combinatorial optimization problems using AutoRL.
In this context, the objective of this paper is to propose an AutoRL simulation environment for carrying out experiments in combinatorial optimization. The proposed framework has applications in three types of problems: the symmetric traveling salesman problem, the asymmetric traveling salesman problem, and the sequential ordering problem (SOP). Furthermore, the developed simulator is free and was developed in the R language, widely covered in research in the ML field. Therefore, in summary, the main contributions of this paper are:
  • New AutoRL simulator for combinatorial optimization problems.
  • Modules for simulations with three traditional problems from the literature: TSP, ATSP, and SOP.
  • Modules for simulations with AutoML for automated parameter tuning.
  • Free environment developed using the R language and available in a GitHub repository.
  • Case studies with three combinatorial optimization problems using AutoRL.
  • Optimization of RL parameters using response surface models for the three applications: TSP, ATSP, and SOP.
This paper is divided into six sections. In Section 2, the theoretical foundation is presented. In Section 3, the automated reinforcement learning simulator (AutoRL-Sim) for combinatorial optimization problems is described in detail. In Section 4, case studies that were carried out are presented. Finally, Section 5 presents a comparison with other literature studies, and Section 6 presents conclusions.

2. Background

2.1. Reinforcement Learning

Reinforcement learning is an artificial intelligence method that is based on the trial-and-error process [25]. This technique is based on Markov decision processes, which are characterized by states (S), actions (A), rewards (R), and a transition function (T). In summary, the probabilistic function T describes state transitions, and R is the reward function that provides rewards for actions performed in specific states [25,46,47].
In RL, an agent makes decisions by taking actions on states based on its perceptions. Subsequently, the agent receives rewards for successes or penalties for failures [25,48]. Through interactions with the environment, the agent acquires knowledge about the learning task. Thus, the learning agent seeks to make decisions that result in optimized rewards [25,26,49].

2.1.1. Q-Learning

Q-learning is a reinforcement learning algorithm based on the temporal difference method proposed by Watkins [50] and is suitable for model-free learning situations [26,51]. Through this method, the agent seeks to learn a value function called Q(s,a), where a represents the action performed in the state s [49].
The update of the learning matrix Q in the Q-learning algorithm is expressed by Equation (1) [26,51]:
Q t + 1 ( s t , a t ) = Q t ( s t , a t ) + α [ r ( s t , a t ) + γ m a x a Q ( s , a ) Q t ( s t , a t ) ] ,
in which Q is the learning matrix with dimensions ns × na, where ns is the number of states and na is the number of actions permitted to the agent; α is the learning rate; γ is the discount factor; st is the current state; at is the action performed in the current state; r(s,a) is the reinforcement received for performing the action at in state st; s′ is the future state; and a′ is the action that will be performed in the future state.
Algorithm 1 presents the Q-learning [26] procedural form:
Algorithm 1: Q-learning Algorithm
1 
Set the parameters: α ,   γ   e   ϵ
2 
In each s,a do Q(s,a) = 0
3 
Observe the state s
4 
do
5 
  Select action a using policy ϵ-greedy
6 
  Run the action a
7 
  Receive the immediate reward r(s,a)
8 
  Observe the new state s
9 
   Q ( s t , a t ) = Q ( s t , a t ) + α [ r ( s t , a t ) + γ Q ( s , a ) Q ( s t , a t ) ]
10 
  s = s
11 
while the stop criterion is satisfied;

2.1.2. SARSA

The SARSA algorithm is a variation of Q-learning. This method was named by [52] and is based on the sequence State → Action → Reward→ State → Action [26,49]. The update of the SARSA algorithm is expressed according to Equation (2):
Q t + 1 ( s t , a t ) = Q t ( s t , a t ) + α [ r ( s t , a t ) + γ Q ( s , a ) Q t ( s t , a t ) ] ,
in which Q is the learning matrix; st is the current state; at is the action performed in the current state; r(s,a) is the reinforcement received for performing the action at in the state st; α is the learning rate; γ is the discount factor; s′ is the future state; and a′ is the action that will be performed in the future state.
Algorithm 2 shows SARSA [26]:
Algorithm 2: SARSA Algorithm
1 
Set the parameters: α ,   γ   e   ϵ
2 
In each s,a do Q(s,a) = 0
3 
Observe the state s
4 
Select action a using policy ϵ -greedy
5 
do
6 
  Run the action a
7 
  Receive the immediate reward r(s,a)
8 
  Observe the new state s
9 
  Select action a′ using policy ϵ-greedy
10 
   Q ( s t , a t ) = Q ( s t , a t ) + α [ r ( s t , a t ) + γ Q ( s , a ) Q ( s t , a t ) ]
11 
  s = s′
12 
  a = a′
13 
while the stop criterion is satisfied;

2.2. Automated Reinforcement Learning

RL is a constantly evolving research area with applications across a wide range of domains, including games [53,54,55] and robotics [56,57,58]. However, conducting studies in RL can be a challenging task, as experimenters need to define a variety of settings, such as parameters and algorithms [23].
Adjusting the optimal configurations for machine learning methods is crucial, as these can have a significant impact on the results obtained [10,59]. Defining these characteristics can be costly and susceptible to error given the vast number of combinations of possible parameter values, which increases the probability of not finding the best configuration for the problem in question [23].
RL methods present several characteristics that can be optimized, such as algorithms, parameters, and reward functions. One approach to automating these learning decisions is the use of AutoML. Thus, when applied to RL, AutoML is called AutoRL [23,59].
In the literature, there are several studies that address the use of AutoRL. For example, the authors of [60] use AutoRL to optimize the reward function applied to the locomotion of quadruped robots. In [61], the authors apply AutoRL for optimizations during processes carried out with microgrids. In the work of [62], the authors propose a system that uses AutoRL to train an agent capable of controlling traffic lights. The authors of [63] present an automated approach to defining network architectures. Finally, in the work presented by [64], the authors present an automated system for determining appropriate configurations in dynamic pricing problems.

2.3. Combinatorial Optimization Problems

Optimization problem applications seek to minimize or maximize an objective function. Thus, the goal is to find an optimal solution for a specific situation. In summary, an optimization problem can be characterized by the following properties [32,65]:
  • Objective function: Contains the criteria that can be adjusted during the search for the optimal solution. All combinatorial optimization problems have at least one objective function.
  • Restrictions: Applied to ensure the viability of the solution found. An optimization problem can contain several constraints or none.
  • Decision variables: Represent the criteria that will be adjusted during the search process for the optimal solution.
Some examples of combinatorial optimization problems can be found in the literature, such as [32,33,66,67,68]: the symmetric traveling salesman problem (TSP), the asymmetric traveling salesman problem (ATSP), the sequential ordering problem (SOP), and the jobshop scheduling problem (JSP).

3. AutoRL-Sim

In this section, an automated reinforcement learning simulator for combinatorial optimization problems (AutoRL-Sim) is proposed. The primary objective of this tool is to facilitate the execution of experiments, providing users with a simplified approach and allowing them to focus their efforts on analyzing the results. Figure 1 illustrates an overview of how AutoRL-Sim works.
Algorithms 3 and 4 show how the tool works using procedural algorithms. Algorithm 3 shows how the modules that do not use AutoML work. Algorithm 4 shows how the modules that use AutoML work. It should be noted that the entire description presented in Algorithms 3 and 4 also applies to the free modules, differing only in the way the data are entered before the experiment begins.
Algorithm 3: Workflow of Modules without AutoML
Step 1 (on the interface)
1 
Select the TSP/ATSP/SOP instance to be executed
2 
Enter the value of the learning rate, the discount factor, the e-greedy policy, and the number of episodes, or choose the option to generate these values randomly
3 
Click on the start experiment button
Step 2 (on the server)
4 
The values for α , γ , and ϵ and number of episodes used are those defined by the user
5 
The number of epochs is set to 1
6 
Run SARSA or Q-learning
7 
Returns the distance result obtained and the other necessary data
Step 3 (on the interface)
8 
Displays the graphs and the report to the user
In Algorithm 3, which represents how the modules work without AutoML, the user must initially enter the information needed to carry out the experiment via the interface. Then, on the server, the selected instance will be executed with the parameters (learning rate, discount factor, e-greedy policy, and number of episodes) defined by the user. Finally, the results obtained are presented to the user on the interface via graphs and a report.
Algorithm 4: AutoML Module Workflow
Step 1 (on the interface)
1 
Select the TSP/ATSP/SOP instance to be executed
2 
Click on the start experiment button
Step 2 (on the server)
3 
The α and γ parameters are defined
4 
The ϵ parameter is set to 0.01
5 
The number of epochs is set to 5
6 
The number of episodes is set to 1000
7 
foreach  α t α do
8 
foreach  γ t γ do
9 
  for epoch to numberEpochs do
10 
   Run SARSA or Q-learning
11 
   Stores the distance obtained and the parameters used in the current run
12 
  end
13 
end
14 
end
15 
Fits a regression model to the data
16 
Generates the model’s statistical data
17 
Generates the response surface using RSM
18 
Extracts the stationary points from the RSM model
19 
The new α and γ values are defined from the model’s stationary points
20 
if (normality > 0.05 & significance < 0.05 & ( 0.01 < α < 1 ) & ( 0 < γ < 1 )) then
21 
α and γ values are maintained
22 
else
23 
 The α and γ values are updated to the values that gave the best distance result
   when running the combinations
24 
end
25 
The number of episodes is set to 10,000
26 
for epoch to numberEpochs do
27 
 Run SARSA or Q-learning
28 
end
29 
Returns the best distance result obtained from the 5 epochs and the other necessary data
Step 3 (on the interface)
30 
Displays the graphs and the report to the user
Algorithm 4 shows how modules with AutoML work. Here, the user only has to select the instance to be executed (and the other necessary data, if it is an experiment in the free module). On the server, the selected problem will initially be run with various combinations of parameters (the α and γ parameters are initialized with the values [0.01, 0.15, 0.30, 0.45, 0.60, 0.75, 0.90, 0.99], while ϵ is constantly kept at 0.01), and the results will be stored. After that, RSM will be applied to generate the regression model of the problem, and some validations will be carried out. If the model meets the requirements, the learning rate and discount factor parameters are set as the model’s stationary points. Otherwise, the parameters (learning rate and discount factor) are set to the values of alpha and gamma that gave the best result during the combination stage. Next, the problem is run again, this time with only the new parameters and over 5 epochs. Finally, the best result from these 5 epochs and the rest of the information from the experiment are shown to the user in the form of graphs and a report.
AutoRL-Sim is available for download from the GitHub repository (https://github.com/KellyBarbosa/autorl_sim, accessed on 21 August 2024). In the following subsections, details about the operation and structure of AutoRL-Sim are presented.

3.1. Methodology of AutoRL-Sim Development

The methodology proposed for the development of AutoRL-Sim is structured in the following steps:
  • Definition of software resources and tools.
  • Proposal of graphical interface prototypes.
  • Selection of combinatorial optimization problems.
  • Definition of a dataset for combinatorial optimization problem experiments.
  • Development of reinforcement learning framework with model, algorithms, and parameters.
  • Proposal of AutoRL algorithms.
  • Simulation of experiments and analysis of results with case studies.
The six initial stages of the proposed methodology are described in the following subsections. Subsequently, in Section 4, the results of the simulations for the proposed case studies are presented.

3.2. Software Tools

AutoRL-Sim uses several software tools. Initially, it is worth highlighting that the R programming language was selected for this project. The R language was chosen based on the following factors: (i) extensively used in machine learning research; (ii) includes various functions and libraries for the use of statistical methods, such as ANOVA and RSM; (iii) development platforms with public access; and (iv) an environment for developing graphical interfaces. Other programming languages were considered, such as Python and MATLAB; however, the R environment is the one that most accurately covers all the aspects analyzed in the selection.
Table 1 lists the software resources and libraries used, as well as their respective versions and the type of technology they belong to. Furthermore, it presents some of the main R language functions used, specifying which library they belong to.
The Shiny software package was used to develop the interface, together with HTML, CSS, and JavaScript. Furthermore, LaTeX was used to create the reports. The R software with the stats and rsm packages were adopted to develop the system, analyze the model, and adjust the parameters, respectively.

3.3. Interface

The basis for the development of the AutoRL-Sim environment was the Shiny library [69] together with the R language [70]. In addition, resources from HTML, CSS, JavaScript, and LaTeX were also used. Using this set of tools, eight modules were developed, covering the TSP, ATSP, and SOP problems.
For each problem covered by the interface, there is a module without AutoML and another with AutoML. It is also possible to use the “Free module”, which has options with AutoML/without AutoML and accepts problems such as TSP, ATSP, and SOP. The “Free module” adds flexibility to the use of AutoRL-Sim because, while in the other modules the problems are already pre-defined, in the “Free module”, the user is responsible for entering all the data for the problem they wish to analyze.
Figure 2 shows the AutoRL-Sim home page, where the tool logo is displayed. On this same screen, there is a navigation menu containing the options: Home, TSP, ATSP, SOP, TSP-AutoML, ATSP-AutoML, SOP-AutoML, Free module, and More. Below is a brief description of each interface tab.
  • Home: Shows the AutoRL-Sim home page.
  • TSP: In this module, users can perform TSP experiments using pre-defined instances in the tool.
  • ATSP: In this module, users can perform ATSP experiments using pre-defined instances in the tool.
  • SOP: In this module, users can perform SOP experiments using pre-defined instances in the tool.
  • TSP-AutoML: In this module, users can perform TSP experiments using pre-defined instances in the tool using AutoML.
  • ATSP-AutoML: In this module, users can perform ATSP experiments using pre-defined instances in the tool using AutoML.
  • SOP-AutoML: In this module, users can perform SOP experiments using pre-defined instances in the tool using AutoML.
  • Free module: In this module, the user is responsible for providing all information about the instance, making the tool more flexible in relation to the variety of experiments that can be performed. The user will have access to the “Without AutoML” and “With AutoML” options.
  • More: By clicking on this option, the user will have access to the “Additional Information” and “About” options.
    Additional Information: There is an explanation of some of the main topics covered in the tool.
    About: The user will find a brief description of AutoRL-Sim and information about its developers.
In Figure 3, an example of an experiment that can be carried out in AutoRL-Sim is shown. In this case, the experiment was conducted with the “ft70.2” instance of the SOP type. Therefore, to carry out the experiment, the user must select one of the instances already registered in the tool and define the learning rate, the discount factor, the ϵ-greedy policy, and the number of episodes. After completing this information, the user must click on the “START EXPERIMENTS” button. At the end of the experiment, the user will have access to the distance graph and a summary of the results. The user can choose to download the results for later analysis.
Figure 4 shows a simulation example with AutoML. In the image, the experiment was carried out with the “ftv33” instance of the ATSP type. Initially, the user must select one of the instances already registered in the tool and click on the “START EXPERIMENTS” button. After the simulation, the distance graph, contour graph, surface graph, and experiment results are available. This information is displayed in separate windows with their respective names. It should be noted that the user has the possibility of downloading the results from the corresponding tab for future analysis.
In Figure 5, an example of an experiment using the Free module (with AutoML) is shown. In the image, the simulation was carried out with the “swiss42” instance of the TSP type. To do this, the user must fill in the name of the instance, its type (TSP, ATSP, or SOP), the data format (2D Euclidean or matrix), the dimension of the instance, and whether there is already a known optimal value (if so, the user must enter it). Additionally, the user must upload a “.txt” file containing the instance data. After adding this information, the user must click on the “START EXPERIMENTS” button. After the simulation, the user will have access to the distance graph, the contour graph, the surface graph, and the experiment results. These results are displayed in separate windows with their respective names, and the user can download them for later analysis.

3.4. Combinatorial Optimization Problems

AutoRL-Sim provides simulations with three types of combinatorial optimization problems based on the traveling salesman problem. The traveling salesman problem is one of the classic problems of NP-complete combinatorial optimization, where the objective is to determine a Hamiltonian cycle of minimum cost. In this problem, given a set of cities (nodes), the salesman must visit each city exactly once, returning to the starting node at the end of the route [71,72,73,74].
Figure 6 illustrates an example of the route taken by the traveling salesman. This is a problem with 51 cities (represented by nodes) generated during an experiment conducted in AutoRL-Sim.
In the literature, there are several variations of the traveling salesman problem [66,73,75], and three of them are addressed by AutoRL-Sim:
  • TSP: In TSP (symmetric), which is the simplest and most general case of the problem, the distance between nodes does not depend on the direction of displacement [71,74].
  • ATSP: In ATSP (or asymmetric TSP), the distance between nodes can vary according to the direction of displacement adopted [66,75].
  • SOP: This is one of the variations of ATSP. In SOP, as in ATSP, distances between cities may vary depending on the direction of travel adopted. Furthermore, this problem includes the additional order of precedence constraint [73,75].

3.5. Dataset

AutoRL-Sim uses a dataset from the TSPLIB library (http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/, accessed on 7 July 2023) [76,77]. TSPLIB is an open repository that provides data and known optimal values for combinatorial optimization problems and is widely addressed in the literature. Instances of the TSP, ATSP, and SOP types were selected to create the AutoRL-Sim dataset, according to Table 2, Table 3, and Table 4, respectively.
In Figure 7, it is possible to check how TSPLIB instances can be accessed. For this, the user must select one of the problems available in “TSPLIB instances”.
In modules where problems are not predefined (Free module), the user must provide the necessary information about the problem data. An example of this situation can be seen in Figure 8.

3.6. Reinforcement Learning

In this section, some reinforcement learning features adopted in the AutoRL-Sim framework are presented. Initially, it is highlighted that the Q-learning and SARSA algorithms were used, as discussed in Section 2. In the following subsections, the RL modeling adopted and the parameters adjusted by AutoRL-Sim are described.

3.6.1. Reinforcement Learning Model

The modeling adopted for RL was based on works available in the literature and involves states, actions, and reinforcements [11,78,79]. Therefore, the RL structure adopted in AutoRL-Sim was defined as follows:
  • States: All the cities (nodes) in the studied problem.
  • Actions: Possible actions are the cities that have not yet been visited by the agent, that is, the nodes available on the route.
  • Reinforcements: The reinforcement received is equivalent to the distance between the cities (nodes i and j) multiplied by 1 , according to Equation (3):
R i j = D i j .

3.6.2. Parameters

ML algorithms have parameters (or hyperparameters) that directly influence the learning process. When analyzing other types of ML algorithms, it is common to differentiate between the terms “parameters” and “hyperparameters”. For example, in artificial neural networks, the parameters are weights adjusted by the optimization method (e.g., gradient descent), whereas hyperparameters are values defined by the user, such as the number of layers or neurons [80]. In this paper, based on recent studies on RL for combinatorial optimization problems [11,33], the convention that the term “RL parameters” is synonymous with “RL hyperparameters” will be adopted.
The RL algorithms presented in Section 2 have three parameters:
  • Learning Rate: The learning rate is represented by the symbol α, and its value can vary between 0 and 1 [81]. The α parameter controls the impact of updates on the learning matrix [82]. When α = 0, no learning occurs, as the learning matrix update equation reduces to Q t + 1 ( s t , a t ) = Q t ( s t , a t ) .
  • Discount Factor: The discount factor is represented by the symbol γ and can be defined between 0 and 1. The value attributed to γ reflects the degree of importance of future rewards for the learning agent. As γ approaches 0, future rewards are considered more insignificant. On the other hand, future rewards gain high relevance when γ approaches 1 [49,83,84,85].
  • ϵ -greedy: The ϵ-greedy policy has the ϵ parameter, whose value can vary between 0 and 1. This policy is adopted in the selection of actions, and this parameter value determines the degree of randomness in a decision. Through the ϵ-greedy policy, the learning agent can alternate between exploration, which involves searching for new experiences in the environment, and exploitation, which is based on previous experience and accumulated knowledge. The ϵ-greedy policy follows the update rule shown in Equation (4) [13,83,84]:
    π ( s ) = a ,   with   probability   1 ϵ a a ,   with   probability   ϵ ,
    where π(s) is the policy for the current state, a* is the highest scoring action in the learning matrix, and aa is a random action among those available.

3.7. AutoRL Using RSM

In this study, automated tuning of RL parameters was performed using response surface models. To achieve this, the AutoRL structure was based on recent research [11,86], where the RSM approach for parameter recommendation applied to combinatorial optimization problems was proposed and validated.
RSM is a widely recognized mathematical and statistical technique for analyzing and modeling the relationship between input and output variables. In RSM, modeling occurs by fitting a multiple linear regression. This allows you to model equations for problems with one or more independent variables [87,88,89]. During the RSM application process, the values of the independent variables are manipulated and analyzed to find the mathematical equation that best describes the problem along with its respective response (dependent variable). This method makes it possible to optimize the response variable, identifying the ideal values to maximize or minimize the desired output [88,89].
Generally, the most common RSM models are first- or second-order [90]. First-order models can be represented by Equation (5) [87]:
y = β 0 + i = 1 k β i x i + ε .
Second-order models can be represented by Equation (6) [87]:
y = β 0 + i = 1 k β i x i + i = 1 k β i i x i 2 + 1 i j k β i j x i x j + ε .
In this work, the AutoRL-Sim framework uses second-order RSM models to adjust two RL parameters: α and γ. For this, the model variables were adopted as x 1 = α and x 2 = γ, and the dependent variable (y) was the adjustment of the final distance covered by the traveling salesman. Thus, the final RSM equation used to automate this reinforcement learning process is defined according to Equation (7) [11,86]:
y = β 0 + β 1 α + β 2 γ + β 3 α 2 + β 4 γ 2 + β 5 α γ + ε .
Figure 9 represents how RSM is used in the proposed simulator. Initially, the selected instance is executed with different combinations of values for the parameters α and γ. Then, RSM is applied. Information about the fitted model is extracted using functions available in the R software. Some of the functions used are:
  • rsm: response surface model fitting [91];
  • lm: fitting of linear models [70];
  • anova: extracts information about analysis of variance models [70];
  • canonical: determines the stationary points of the RSM model [91];
  • ks.test: performs the Kolmogorov–Smirnov residual normality test [70];
  • summary: shows a summary of the adjusted model results [70].
After applying the RSM, some criteria were implemented to determine whether the parameters (α and γ) generated were adequate:
  • Is the value of α between 0 and 1?
  • Is the value of γ between 0 and 1?
  • Are the residuals normal (significance criterion of 5%)?
  • Is there statistical significance (significance criterion of 5%)?
If all four criteria are satisfied, the parameters (α and γ) generated by the RSM are considered the most appropriate for the problem studied. Otherwise, the AutoRL-Sim system considers the combination of parameters that resulted in the best final route distance as the most appropriate for the analyzed instance. Finally, the selected instance is executed again, this time with the best values found for α and γ.

4. Case Studies

In this section, four case studies of simulations with AutoRL-Sim are presented:
  • Case study 1: module without AutoML.
  • Case study 2: module with AutoML.
  • Case study 3: Free module.
  • Case study 4: comparison between modules with and without AutoML.

4.1. Case Study 1: Module without AutoML

In the first case study, the experiment was carried out with the “ft53” instance (ATSP). For this, the learning rate and discount factor values presented by [86] were used, in which the authors present these parameters as the best combination found for the “ft53” problem. After the simulation, the user has the generated distance graph available (Figure 10). It is observed that, as the episodes are executed, the distance values gradually decrease until they reach a relatively constant stabilization point.
Table 5 presents the report containing a summary of the settings used and the results obtained. In this aspect, it is noted that the minimum distance value obtained (8182) presented a percentage error of only 18.49% in relation to the optimal value provided by TSPLIB (6905). Furthermore, it is worth highlighting the fact that the minimum distance value obtained coincides with the value presented by [86], which reinforces the effectiveness of AutoRL-Sim.

4.2. Case Study 2: Module with AutoML

In this second case study, the “eil51” instance with the AutoML module was used. To carry out the experiment in this module, it is only necessary to define which instance will be evaluated. After the experiment, the user will be able to analyze the four graphs: route, distance, response surface, and contour lines. Furthermore, the AutoRL-Sim user will also be able to view the report containing a summary of results.
Figure 11 and Figure 12 present the surface and contour graphs, respectively. These graphs allow you to check the value ranges of the learning rate and discount factor parameters that tend to minimize the distance to the final route. In this aspect, the regions corresponding to these ranges of values are represented by the reddest tones in both graphs. It can be seen in Figure 11 and Figure 12 that the parameters that tend to minimize the final distance are approximately 0.15     γ   0.55 and 0.55   α   0.85 .
The other graphs generated in the second case study are presented in Figure 13 and Figure 14. Furthermore, the simulation report is shown in Table 6. When analyzing the results of this case study, it is important to highlight the similarity between the results achieved and those presented by [86]. In [86], the adjusted parameters were γ = 0.352 and α = 0.693, with a final route distance of 475. In this work, the parameters obtained in the experiment were γ = 0.34 and α = 0.69, also generating a final route distance of 475. Thus, the results obtained in the second case study reinforce the adequate functioning of AutoRL-Sim.

4.3. Case Study 3: Free Module

In the third case study, the Free module and the “ESC78” instance (SOP) were adopted, with a known optimal value of 18,230 (TSPLIB). In this simulation, the parameter values (learning rate, discount factor, and e-greedy policy) and the number of episodes were randomly defined using the functionality available in the AutoRL-Sim interface called “GENERATE RANDOM VALUES”.
The results of this experiment can be seen in Figure 15 and Table 7. From the simulation, it is possible to observe the learning process of the “ESC78” instance, in which the distance decreases throughout the episodes. Furthermore, it is noteworthy that the minimum distance (19,910) achieved was considerably closer to the optimal value provided by TSPLIB (18,230).

4.4. Case Study 4: Comparison between Modules with and without AutoML

For the fourth case study, some instances of TSP, ATSP, and SOP were selected. This study aims to compare the results obtained when using modules without AutoML and modules with AutoML. It also aims to compare the results obtained with those provided by TSPLIB and other works presented in the literature. The data are shown in Table 8. Finally, a comparison will also be made between the execution time of the experiments in the modules with and without AutoML, the results of which are shown in Table 9.
Table 8 shows the route values for some selected cases. It can be seen that the results obtained with the application of AutoML are, in general, close to the results presented in the literature by other authors and to the value considered optimal by TSPLIB. An important point to note is that the experiments without the application of AutoML were carried out with parameterizations defined at random (using the “GENERATE ALEATORY VALUES” feature available in the tool) in order to simulate an inexperienced experimenter.
Table 9 shows the computational time required to run the experiments described in Table 8. It shows that, as the complexity and size of the problem increase, the time required for the experiment also increases.
Only a few sample cases were selected, not all the problems presented in Table 2, Table 3 and Table 4, due to the high computational cost required to carry out all the experiments. This is evidenced by the case of instance “ft70.1” in Table 9.
It is also important to note that, unlike the other instances, instance “ft70.1” showed a result with a noticeably greater difference compared to the other cases tested. This discrepancy can be explained by the complexity of the problem. As this is a more complex problem, this instance would benefit from more epochs and episodes in addition to those already programmed internally in AutoRL-Sim. However, the computational cost would tend to be much higher.

5. Comparison with Other Studies

This section presents a comparison between AutoRL-Sim and other frameworks present in the literature that simulate reinforcement learning. In this sense, three papers were selected: I [39], II [42], and III [36]. These studies were considered to have relevant similarities with definitions in the present paper. The main factor is that the three papers present new graphical interfaces for reinforcement learning. Furthermore, the papers also adopt the SARSA and Q-learning methods as learning algorithms. It is also important to highlight that these simulators considered for comparison allow adjustments and visualization of graphs/reports, proving to be tools with relevant features for the RL user.
For this, we compared the development environment used, the RL algorithms used, the advanced techniques used, the problems addressed, and the features available. Table 10 presents a summary of the analysis performed.
Table 10 shows that the proposed system (AutoRL-Sim) presents important advances in relation to previous studies. First, the proposed environment was developed with the R language, while the other works used MATLAB ([39,42]) and Visual Studio ([36]). Furthermore, AutoRL-Sim stands out as a dedicated tool for experiments with automated machine learning. In AutoRL-Sim, this technique is implemented with RSM for the purpose of parameter optimization. Furthermore, the Free module allows the user to enter new data and also perform simulations with AutoML.
Regarding the problems addressed, AutoRL-Sim is focused on three classic combinatorial optimization problems: TSP, ATSP, and SOP. In general, the other studies analyzed developed tools for other applications, such as agent navigation [39,42], MKP [39], and reservoir systems [36].
As for the functionalities available, all proposed systems offer the option of generating graphs and reports, but only AutoRL-Sim offers parameter optimization with AutoML. Furthermore, only AutoRL-Sim and the Ottoni et al. [39] proposal allow the selection of the instance to be studied. Another point to highlight is that only AutoRL-Sim offers an additional module in which the user can insert the data of the problem to be analyzed, providing greater freedom and flexibility in creating their own datasets for experimentation. A similarity between the studies is the use of classic RL algorithms (Q-learning and SARSA) in all works.

6. Conclusions

This paper proposed a simulation environment aimed at combinatorial optimization problems using automated reinforcement learning. Thus, AutoRL-Sim addresses three classic combinatorial optimization problems: TSP, ATSP, and SOP. It is noteworthy that, when using AutoRL-Sim, users can carry out various experiments and can also choose whether or not they wish to use AutoRL to optimize parameters. Two other modules allow the user to add data to the problems. Furthermore, AutoRL-Sim makes it possible to analyze simulation results using graphs and reports.
In future work, advancements are expected in two specific research areas using AutoRL-Sim. The first source of future developments comprises the adoption of the proposed simulator as a teaching tool in teaching reinforcement learning in undergraduate and postgraduate courses. In this sense, exploring the simulator in the proposal in practical evaluation activities in artificial intelligence training courses is suggested. The second research direction refers to expanding the scope of the simulator by implementing more modules and incorporating new features. In this regard, it is expected to implement other traditional advanced techniques, such as transfer learning [68], and more methods for hyperparameter optimization [2].

Author Contributions

Conceptualization, G.K.B.S. and A.L.C.O.; methodology, G.K.B.S. and A.L.C.O.; software, G.K.B.S. and A.L.C.O.; validation, A.L.C.O.; formal analysis, A.L.C.O.; writing—original draft preparation, G.K.B.S.; writing—review and editing, A.L.C.O.; supervision, A.L.C.O. All authors have read and agreed to the published version of the manuscript.

Funding

The author André Luiz Carvalho Ottoni is grateful for the funding received under Edital PROPPI 18/2024 - UFOP.

Data Availability Statement

The dataset analyzed during the current study is available in the TSPLIB repository, http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/ (accessed on 7 July 2023). The simulation environment developed is available at https://github.com/KellyBarbosa/autorl_sim (accessed on 21 August 2024).

Acknowledgments

The authors are grateful to UFRB and UFOP (Edital PROPPI 18/2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
αlearning rate
ϵϵ-greedy
γdiscount factor
π(s) policy for the current state
afuture action
abest action available
aarandom action
atcurrent action
Dijcost of moving from i to j
nanumber of actions
nsnumber of states
Rijreinforcement received for moving from i to j
sfuture state
stcurrent state
a actions
ATSP asymmetric traveling salesman problem
AutoML automated machine learning
AutoRL automated reinforcement learning
AutoRL-Sim automated reinforcement learning simulator
BPP bin-packing problem
CSS Cascading Style Sheets
HTML HyperText Markup Language
JSP job-shop scheduling problem
Max-Cut maximum cut
ML machine learning
MVC minimum vertex cover
Q learning matrix
QAP quadratic assignment problem
r rewards
RL reinforcement learning
RSM response surface methodology
s states
SARSA state–action–reward–state–action
SOP sequential ordering problem
T transition function
t time
TSP symmetric traveling salesman problem
TSPLIB traveling salesman problem library
VNS variable neighborhood search
VRP vehicle routing problem

References

  1. Brazdil, P.; Carrier, C.G.; Soares, C.; Vilalta, R. Metalearning: Applications to Data Mining; Springer Science & Business Media: Berline/Heidelberg, Germany, 2008. [Google Scholar]
  2. Hutter, F.; Kotthoff, L.; Vanschoren, J. (Eds.) Automated Machine Learning: Methods, Systems, Challenges; Springer: Berlin/Heidelberg, Germany, 2018; in press; Available online: http://automl.org/book (accessed on 21 August 2024).
  3. Tuggener, L.; Amirian, M.; Rombach, K.; Lorwald, S.; Varlet, A.; Westermann, C.; Stadelmann, T. Automated Machine Learning in Practice: State of the Art and Recent Results. In Proceedings of the 2019 6th Swiss Conference on Data Science (SDS), Bern, Switzerland, 14 June 2019; pp. 31–36. [Google Scholar] [CrossRef]
  4. Vidnerová, P.; Neruda, R. Air Pollution Modelling by Machine Learning Methods. Modelling 2021, 2, 659–674. [Google Scholar] [CrossRef]
  5. Ottoni, A.L.C.; Souza, A.M.; Novo, M.S. Automated hyperparameter tuning for crack image classification with deep learning. Soft Comput. 2023, 27, 18383–18402. [Google Scholar] [CrossRef]
  6. Feurer, M.; Klein, A.; Eggensperger, K.; Springenberg, J.; Blum, M.; Hutter, F. Efficient and Robust Automated Machine Learning. In Advances in Neural Information Processing Systems 28; Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2015; pp. 2962–2970. [Google Scholar]
  7. Makmal, A.; Melnikov, A.A.; Dunjko, V.; Briegel, H.J. Meta-learning within projective simulation. IEEE Access 2016, 4, 2110–2122. [Google Scholar] [CrossRef]
  8. Mantovani, R.G.; Rossi, A.L.; Alcobaça, E.; Vanschoren, J.; de Carvalho, A.C. A meta-learning recommender system for hyperparameter tuning: Predicting when tuning improves SVM classifiers. Inf. Sci. 2019, 501, 193–221. [Google Scholar] [CrossRef]
  9. Cai, H.; Lin, J.; Lin, Y.; Liu, Z.; Wang, K.; Wang, T.; Zhu, L.; Han, S. AutoML for Architecting Efficient and Specialized Neural Networks. IEEE Micro 2020, 40, 75–82. [Google Scholar] [CrossRef]
  10. Ottoni, L.T.C.; Ottoni, A.L.C.; Cerqueira, J.d.J.F. A Deep Learning Approach for Speech Emotion Recognition Optimization Using Meta-Learning. Electronics 2023, 12, 4859. [Google Scholar] [CrossRef]
  11. Ottoni, A.L.C.; Nepomuceno, E.G.; de Oliveira, M.S. A Response Surface Model Approach to Parameter Estimation of Reinforcement Learning for the Travelling Salesman Problem. J. Control Autom. Electr. Syst. 2018, 29, 350–359. [Google Scholar] [CrossRef]
  12. Shahrabi, J.; Adibi, M.A.; Mahootchi, M. A reinforcement learning approach to parameter estimation in dynamic job shop scheduling. Comput. Ind. Eng. 2017, 110, 75–82. [Google Scholar] [CrossRef]
  13. Santos, J.P.Q.d.; de Melo, J.D.; Neto, A.D.D.; Aloise, D. Reactive search strategies using reinforcement learning, local search algorithms and variable neighborhood search. Expert Syst. Appl. 2014, 41, 4939–4949. [Google Scholar] [CrossRef]
  14. Gershman, S.J. Empirical priors for reinforcement learning models. J. Math. Psychol. 2016, 71, 1–6. [Google Scholar] [CrossRef]
  15. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  16. Syarif, I.; Prugel-Bennett, A.; Wills, G. SVM parameter optimization using grid search and genetic algorithm to improve classification performance. TELKOMNIKA (Telecommun. Comput. Electron. Control) 2016, 14, 1502–1509. [Google Scholar] [CrossRef]
  17. Victoria, A.H.; Maragatham, G. Automatic tuning of hyperparameters using Bayesian optimization. Evol. Syst. 2021, 12, 217–223. [Google Scholar] [CrossRef]
  18. Moussa, C.; Patel, Y.J.; Dunjko, V.; Bäck, T.; van Rijn, J.N. Hyperparameter importance and optimization of quantum neural networks across small datasets. Mach. Learn. 2024, 113, 1941–1966. [Google Scholar] [CrossRef]
  19. Tsiakmaki, M.; Kostopoulos, G.; Kotsiantis, S.; Ragos, O. Implementing AutoML in Educational Data Mining for Prediction Tasks. Appl. Sci. 2019, 10, 90. [Google Scholar] [CrossRef]
  20. Stamoulis, D.; Ding, R.; Wang, D.; Lymberopoulos, D.; Priyantha, B.; Liu, J.; Marculescu, D. Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization. IEEE J. Sel. Top. Signal Process. 2020, 14, 609622. [Google Scholar] [CrossRef]
  21. Mantovani, R.G.; Rossi, A.L.; Vanschoren, J.; Bischl, B.; Carvalho, A.C. To tune or not to tune: Recommending when to adjust SVM hyper-parameters via meta-learning. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–8. [Google Scholar]
  22. Mahdavian, A.; Shojaei, A.; Salem, M.; Laman, H.; Yuan, J.S.; Oloufa, A. Automated Machine Learning Pipeline for Traffic Count Prediction. Modelling 2021, 2, 482–513. [Google Scholar] [CrossRef]
  23. Parker-Holder, J.; Rajan, R.; Song, X.; Biedenkapp, A.; Miao, Y.; Eimer, T.; Zhang, B.; Nguyen, V.; Calandra, R.; Faust, A.; et al. Automated Reinforcement Learning (AutoRL): A Survey and Open Problems. J. Artif. Intell. Res. 2022, 74, 517–568. [Google Scholar] [CrossRef]
  24. Mussi, M.; Lombarda, D.; Metelli, A.M.; Trovó, F.; Restelli, M. ARLO: A framework for Automated Reinforcement Learning. Expert Syst. Appl. 2023, 224, 119883. [Google Scholar] [CrossRef]
  25. Kaelbling, L.P.; Littman, M.L.; Moore, A.W. Reinforcement learning: A survey. J. Artif. Intell. Res. 1996, 4, 237–285. [Google Scholar] [CrossRef]
  26. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  27. Ali, A.R.; Budka, M.; Gabrys, B. A Meta-Reinforcement Learning Approach to Optimize Parameters and Hyper-parameters Simultaneously. In Proceedings of the PRICAI 2019: Trends in Artificial Intelligence, Cuvu, Yanuca Island, Fiji, 26–30 August 2019; Nayak, A.C., Sharma, A., Eds.; Springer: Cham, Switzerland, 2019; pp. 93–106. [Google Scholar]
  28. Chien, J.T.; Lieow, W.X. Meta Learning for Hyperparameter Optimization in Dialogue System. Proc. Interspeech 2019, 2019, 839–843. [Google Scholar] [CrossRef]
  29. Nazari, M.; Oroojlooy, A.; Takáč, M.; Snyder, L.V. Reinforcement learning for solving the vehicle routing problem. Neural Inf. Process. Syst. Found. 2018, 2018, 9839–9849. [Google Scholar]
  30. Gambardella, L.M.; Dorigo, M. Ant-Q: A Reinforcement Learning Approach to the Traveling Salesman Problem; Morgan Kaufmann Publishers, Inc.: Burlington, MA, USA, 1995; pp. 252–260. [Google Scholar]
  31. Dai, H.; Khalil, E.B.; Zhang, Y.; Dilkina, B.; Song, L. Learning combinatorial optimization algorithms over graphs. Neural Inf. Process. Syst. Found. 2017, 2017, 6349–6359. [Google Scholar]
  32. Mazyavkina, N.; Sviridov, S.; Ivanov, S.; Burnaev, E. Reinforcement learning for combinatorial optimization: A survey. Comput. Oper. Res. 2021, 134, 105400. [Google Scholar] [CrossRef]
  33. Ottoni, A.L.C.; Nepomuceno, E.G.; de Oliveira, M.S.; de Oliveira, D.C.R. Tuning of Reinforcement Learning Parameters Applied to SOP Using the Scott–Knott Method. Soft Comput. 2020, 24, 4441–4453. [Google Scholar] [CrossRef]
  34. Ottoni, A.L.; Nepomuceno, E.G.; Oliveira, M.S.d.; Oliveira, D.C.d. Reinforcement learning for the traveling salesman problem with refueling. Complex Intell. Syst. 2022, 8, 2001–2015. [Google Scholar] [CrossRef]
  35. Gould, S. DARWIN: A Framework for Machine Learning and Computer Vision Research and Development. J. Mach. Learn. Res. 2012, 13, 3533–3537. [Google Scholar]
  36. Rieker, J.D.; Labadie, J.W. An intelligent agent for optimal river-reservoir system management. Water Resour. Res. 2012, 48. [Google Scholar] [CrossRef]
  37. Depaoli, S.; Winter, S.D.; Visser, M. The Importance of Prior Sensitivity Analysis in Bayesian Statistics: Demonstrations Using an Interactive Shiny App. Front. Psychol. 2020, 11, 608045. [Google Scholar] [CrossRef]
  38. Alves Goulart, D.; Dutra Pereira, R. Autonomous pH control by reinforcement learning for electroplating industry wastewater. Comput. Chem. Eng. 2020, 140, 106909. [Google Scholar] [CrossRef]
  39. Ottoni, A.L.C.; Nepomuceno, E.G.; Oliveira, M.S.d. Development of a Pedagogical Graphical Interface for the Reinforcement Learning. IEEE Lat. Am. Trans. 2020, 18, 92–101. [Google Scholar] [CrossRef]
  40. Jak, S.; Jorgensen, T.D.; Verdam, M.G.E.; Oort, F.J.; Elffers, L. Analytical power calculations for structural equation modeling: A tutorial and Shiny app. Behav. Res. Methods 2021, 53, 1385–1406. [Google Scholar] [CrossRef]
  41. Settaluri, K.; Liu, Z.; Khurana, R.; Mirhaj, A.; Jain, R.; Nikolic, B. Automated Design of Analog Circuits Using Reinforcement Learning. IEEE Trans.-Comput.-Aided Des. Integr. Circuits Syst. 2022, 41, 2794–2807. [Google Scholar] [CrossRef]
  42. Altuntaş, N.; İMal, E.; Emanet, N.; Öztürk, C.N. Reinforcement learning-based mobile robot navigation. Turk. J. Electr. Eng. Comput. Sci. 2016, 24, 1747–1767. [Google Scholar] [CrossRef]
  43. Bamford, C.; Jiang, M.; Samvelyan, M.; Rocktäschel, T. GriddlyJS: A Web IDE for Reinforcement Learning. In Proceedings of the 36th Conference on Neural Information Processing Systems, NeurIPS 2022, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  44. Steinbacher, L.M.; Ait-Alla, A.; Rippel, D.; Düe, T.; Freitag, M. Modelling Framework for Reinforcement Learning based Scheduling Applications. IFAC-PapersOnLine 2022, 55, 67–72. [Google Scholar] [CrossRef]
  45. Spieker, H.; Gotlieb, A.; Marijan, D.; Mossige, M. Reinforcement learning for automatic test case prioritization and selection in continuous integration. In Proceedings of the 26th ACM SIGSOFT International Symposium on Software Testing and Analysis, ISSTA 2017, Santa Barbara, CA, USA, 10–14 July 2017. [Google Scholar] [CrossRef]
  46. Chen, Z.; Lai, J.; Li, P.; Awad, O.I.; Zhu, Y. Prediction Horizon-Varying Model Predictive Control (MPC) for Autonomous Vehicle Control. Electronics 2024, 13, 1442. [Google Scholar] [CrossRef]
  47. Jia, C.; Zhang, F.; Xu, T.; Pang, J.C.; Zhang, Z.; Yu, Y. Model gradient: Unified model and policy learning in model-based reinforcement learning. Front. Comput. Sci. 2024, 18, 184339. [Google Scholar] [CrossRef]
  48. Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997. [Google Scholar]
  49. Russell, S.J.; Norvig, P. Artificial Intelligence; Pearson: Hoboken, NJ, USA, 2013. [Google Scholar]
  50. Watkins, C.J.C.H. Learning from Delayed Rewards. Ph.D. Thesis, University of Cambridge, Cambridge, UK, 1989. [Google Scholar]
  51. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  52. Sutton, R.S. Generalization in reinforcement learning: Successful examples using sparse coarse coding. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 27 November–2 December 1995; pp. 1038–1044. [Google Scholar]
  53. Szita, I. Reinforcement learning in games. In Reinforcement Learning: State-of-the-Art; Springer: Berlin/Heidelberg, Germany, 2012; pp. 539–577. [Google Scholar]
  54. Lample, G.; Chaplot, D.S. Playing FPS Games with Deep Reinforcement Learning. Proc. AAAI Conf. Artif. Intell. 2017, 31. [Google Scholar] [CrossRef]
  55. Samsuden, M.A.; Diah, N.M.; Rahman, N.A. A Review Paper on Implementing Reinforcement Learning Technique in Optimising Games Performance. In Proceedings of the 2019 IEEE 9th International Conference on System Engineering and Technology (ICSET), Shah Alam, Malaysia, 7 October 2019; pp. 258–263. [Google Scholar] [CrossRef]
  56. Kober, J.; Bagnell, J.A.; Peters, J. Reinforcement learning in robotics: A survey. Int. J. Robot. Res. 2013, 32, 1238–1274. [Google Scholar] [CrossRef]
  57. Kormushev, P.; Calinon, S.; Caldwell, D.G. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics 2013, 2, 122–148. [Google Scholar] [CrossRef]
  58. Akalin, N.; Loutfi, A. Reinforcement Learning Approaches in Social Robotics. Sensors 2021, 21, 1292. [Google Scholar] [CrossRef] [PubMed]
  59. Afshar, R.R.; Zhang, Y.; Vanschoren, J.; Kaymak, U. Automated reinforcement learning: An overview. arXiv 2022, arXiv:2201.05000. [Google Scholar]
  60. Kim, M.; Kim, J.S.; Park, J.H. Automated Hyperparameter Tuning in Reinforcement Learning for Quadrupedal Robot Locomotion. Electronics 2024, 13, 116. [Google Scholar] [CrossRef]
  61. Li, Y.; Wang, R.; Yang, Z. Optimal Scheduling of Isolated Microgrids Using Automated Reinforcement Learning-Based Multi-Period Forecasting. IEEE Trans. Sustain. Energy 2022, 13, 159–169. [Google Scholar] [CrossRef]
  62. Xu, Y.; Wang, Y.; Liu, C. Training a Reinforcement Learning Agent with AutoRL for Traffic Signal Control; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2022; pp. 51–55. [Google Scholar] [CrossRef]
  63. Wang, Z.; Zhang, J.; Li, Y.; Gong, Q.; Luo, W.; Zhao, J. Automated Reinforcement Learning Based on Parameter Sharing Network Architecture Search; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2021; pp. 358–363. [Google Scholar] [CrossRef]
  64. Afshar, R.R.; Rhuggenaath, J.; Zhang, Y.; Kaymak, U. An Automated Deep Reinforcement Learning Pipeline for Dynamic Pricing. IEEE Trans. Artif. Intell. 2023, 4, 428–437. [Google Scholar] [CrossRef]
  65. Timofieva, N.K. Artificial Intelligence Problems and Combinatorial Optimization. Cybern. Syst. Anal. 2023, 59, 511–518. [Google Scholar] [CrossRef]
  66. Dorigo, M.; Maniezzo, V.; Colorni, A. Ant system: Optimization by a colony of cooperating agents. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 1996, 26, 29–41. [Google Scholar] [CrossRef]
  67. Mittelmann, H.D. Combinatorial Optimization Problems in Engineering Applications. In Proceedings of the 4th International Conference on Numerical Analysis and Optimization, NAO-IV 2017, Muscat, Oman, 2–5 January 2017; Springer: Cham, Switzerland, 2018; pp. 193–208. [Google Scholar]
  68. Souza, G.K.B.; Santos, S.O.S.; Ottoni, A.L.C.; Oliveira, M.S.; Oliveira, D.C.R.; Nepomuceno, E.G. Transfer Reinforcement Learning for Combinatorial Optimization Problems. Algorithms 2024, 17, 87. [Google Scholar] [CrossRef]
  69. Chang, W.; Cheng, J.; Allaire, J.; Sievert, C.; Schloerke, B.; Xie, Y.; Allen, J.; McPherson, J.; Dipert, A.; Borges, B. Shiny: Web Application Framework for R, R Package Version 1.8.0. 2023. Available online: https://github.com/rstudio/shiny (accessed on 7 July 2023).
  70. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2024. [Google Scholar]
  71. Hopfield, J.; Tank, D. “Neural” computation of decisions in optimization problems. Biol. Cybern. 1985, 52, 141–152. [Google Scholar] [CrossRef]
  72. Geem, Z.W.; Kim, J.H.; Loganathan, G. A New Heuristic Optimization Algorithm: Harmony Search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  73. Castelino, K.; D’Souza, R.; Wright, P.K. Toolpath optimization for minimizing airtime during machining. J. Manuf. Syst. 2003, 22, 173–180. [Google Scholar] [CrossRef]
  74. Xu, R.; Wunsch II, D. Survey of clustering algorithms. IEEE Trans. Neural Netw. 2005, 16, 645–678. [Google Scholar] [CrossRef] [PubMed]
  75. Escudero, L. An inexact algorithm for the sequential ordering problem. Eur. J. Oper. Res. 1988, 37, 236–249. [Google Scholar] [CrossRef]
  76. Reinelt, G. TSPLIB—A traveling salesman problem library. ORSA J. Comput. 1991, 3, 376–384. [Google Scholar] [CrossRef]
  77. Reinelt, G. Tsplib95; University Heidelberg: Heidelberg, Germany, 1995. [Google Scholar]
  78. Bianchi, R.A.C.; Ribeiro, C.H.C.; Costa, A.H.R. On the relation between Ant Colony Optimization and Heuristically Accelerated Reinforcement Learning. In Proceedings of the 1st International Workshop on Hybrid Control of Autonomous System, Pasadena, CA, USA, 13 July 2009; pp. 49–55. [Google Scholar]
  79. Júnior, F.C.D.L.; Neto, A.D.D.; De Melo, J.D. Hybrid metaheuristics using reinforcement learning applied to salesman traveling problem. In Traveling Salesman Problem, Theory and Applications; IntechOpen: London, UK, 2010. [Google Scholar]
  80. Ottoni, A.L.C.; Novo, M.S.; Oliveira, M.S. A Statistical Approach to Hyperparameter Tuning of Deep Learning for Construction Machine Classification. Arab. J. Sci. Eng. 2024, 49, 5117–5128. [Google Scholar] [CrossRef]
  81. Mounjid, O.; Lehalle, C.A. Improving reinforcement learning algorithms: Towards optimal learning rate policies. Math. Financ. 2019, 34, 588–621. [Google Scholar] [CrossRef]
  82. Even-Dar, E.; Mansour, Y.; Bartlett, P. Learning Rates for Q-learning. J. Mach. Learn. Res. 2003, 5, 1–25. [Google Scholar]
  83. Bashir, S.A.; Khursheed, F.; Abdoulahi, I. Adaptive-Greedy Exploration for Finite Systems. Gedrag Organ. Rev. 2021, 34, 417–431. [Google Scholar]
  84. Song, L.; Li, Y.; Xu, J. Dynamic Job-Shop Scheduling Based on Transformer and Deep Reinforcement Learning. Processes 2023, 11, 3434. [Google Scholar] [CrossRef]
  85. Chen, L.; Wang, Q.; Deng, C.; Xie, B.; Tuo, X.; Jiang, G. Improved Double Deep Q-Network Algorithm Applied to Multi-Dimensional Environment Path Planning of Hexapod Robots. Sensors 2024, 24, 2061. [Google Scholar] [CrossRef]
  86. Souza, G.K.B.; Ottoni, A.L.C. AutoRL-TSP-RSM: Sistema de aprendizado por reforço automatizado com metodologia de superfície de resposta para o problema do caixeiro viajante. Rev. Bras. Comput. Apl. 2021, 13, 86–100. [Google Scholar] [CrossRef]
  87. Bezerra, M.A.; Santelli, R.E.; Oliveira, E.P.; Villar, L.S.; Escaleira, L.A. Response surface methodology (RSM) as a tool for optimization in analytical chemistry. Talanta 2008, 76, 965–977. [Google Scholar] [CrossRef] [PubMed]
  88. Hemmati, A.; Asadollahzadeh, M.; Torkaman, R. Assessment of metal extraction from e-waste using supported IL membrane with reliable comparison between RSM regression and ANN framework. Sci. Rep. 2024, 14, 3882. [Google Scholar] [CrossRef] [PubMed]
  89. Kulkarni, T.; Toksha, B.; Autee, A. Optimizing nanoparticle attributes for enhanced anti-wear performance in nano-lubricants. J. Eng. Appl. Sci. 2024, 71, 30. [Google Scholar] [CrossRef]
  90. Myers, R.H.; Montgomery, D.C.; Anderson-Cook, C.M. Response Surface Methodology: Process and Product Optimization Using Designed Experiments; John Wiley & Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  91. Lenth, R. Response-Surface Methods in R, Using rsm. J. Stat. Softw. 2009, 32, 1–17. [Google Scholar] [CrossRef]
  92. Deng, W.; Chen, R.; He, B.; Liu, Y.; Yin, L.; Guo, J. A novel two-stage hybrid swarm intelligence optimization algorithm and application. Soft Comput. 2012, 16, 1707–1722. [Google Scholar] [CrossRef]
  93. Paletta, G.; Triki, C. Solving the asymmetric traveling salesman problem with periodic constraints. Networks 2004, 44, 31–37. [Google Scholar] [CrossRef]
  94. Ottoni, A.L.C.; Nepomuceno, E.G.; de Oliveira, M.S. Aprendizado por Reforço na solução do Problema do Caixeiro Viajante Assimétrico: Uma comparação entre os algoritmos Q-learning e SARSA. In Proceedings of the Simpósio de Mecânica Computacional, Diamantina, Brazil, 23–25 May 2016. [Google Scholar]
  95. Anghinolfi, D.; Montemanni, R.; Paolucci, M.; Maria Gambardella, L. A hybrid particle swarm optimization approach for the sequential ordering problem. Comput. Oper. Res. 2011, 38, 1076–1085. [Google Scholar] [CrossRef]
Figure 1. Flowchart with an overview of AutoRL-Sim.
Figure 1. Flowchart with an overview of AutoRL-Sim.
Modelling 05 00055 g001
Figure 2. AutoRL-Sim home page.
Figure 2. AutoRL-Sim home page.
Modelling 05 00055 g002
Figure 3. AutoRL-Sim: Experiment using the SOP module (without AutoML).
Figure 3. AutoRL-Sim: Experiment using the SOP module (without AutoML).
Modelling 05 00055 g003
Figure 4. AutoRL-Sim: Experiment using the ATSP-AutoML module.
Figure 4. AutoRL-Sim: Experiment using the ATSP-AutoML module.
Modelling 05 00055 g004
Figure 5. AutoRL-Sim: Experiment using the Free module (with AutoML).
Figure 5. AutoRL-Sim: Experiment using the Free module (with AutoML).
Modelling 05 00055 g005
Figure 6. Sample route graph for a problem with 51 cities (eil51).
Figure 6. Sample route graph for a problem with 51 cities (eil51).
Modelling 05 00055 g006
Figure 7. AutoRL-Sim: TSPLIB instance selection.
Figure 7. AutoRL-Sim: TSPLIB instance selection.
Modelling 05 00055 g007
Figure 8. AutoRL-Sim: Data registration in the Free module.
Figure 8. AutoRL-Sim: Data registration in the Free module.
Modelling 05 00055 g008
Figure 9. AutoRL-Sim: AutoRL using RSM.
Figure 9. AutoRL-Sim: AutoRL using RSM.
Modelling 05 00055 g009
Figure 10. Case study 1: Distance graph for the “ft53” instance.
Figure 10. Case study 1: Distance graph for the “ft53” instance.
Modelling 05 00055 g010
Figure 11. Case study 2: Response surface graph for instance “eil51”.
Figure 11. Case study 2: Response surface graph for instance “eil51”.
Modelling 05 00055 g011
Figure 12. Case study 2: Graph of contour lines for instance “eil51”.
Figure 12. Case study 2: Graph of contour lines for instance “eil51”.
Modelling 05 00055 g012
Figure 13. Case study 2: Distance graph for the “eil51” instance.
Figure 13. Case study 2: Distance graph for the “eil51” instance.
Modelling 05 00055 g013
Figure 14. Case study 2: Route graph for the “eil51” instance.
Figure 14. Case study 2: Route graph for the “eil51” instance.
Modelling 05 00055 g014
Figure 15. Case study 3: Distance graph for the “ESC78” instance.
Figure 15. Case study 3: Distance graph for the “ESC78” instance.
Modelling 05 00055 g015
Table 1. Software, libraries, and functions used.
Table 1. Software, libraries, and functions used.
TechnologyVersionType of TechnologyFunction
R4.3.3Programming Language-
Shiny1.8.0Framework-
HTMLHTML5Markup Language-
CSSCSS3Style Sheet Language-
JavaScriptES2023Programming Language-
LaTeXLaTeX2eTypesetting System-
stats4.3.3R packagelm()
anova()
ks.test()
summary()
rsm2.10.4R packagecanonical()
rsm()
Table 2. TSP instances from the TSPLIB dataset.
Table 2. TSP instances from the TSPLIB dataset.
ProblemNodesBest Known Solution
eil5151426
berlin52527542
st7070675
eil7676538
pr7676108,159
rat99991211
kroA10010021,282
eil101101629
bier127127118,282
ch1301306110
ch1501506528
a2802802579
lin31831842,029
d1655165562,128
Table 3. ATSP instances from the TSPLIB dataset.
Table 3. ATSP instances from the TSPLIB dataset.
ProblemNodesBest Known Solution
ftv33341286
p43435620
ftv44451613
ftv47481776
ry48p4814,422
ft53536905
ftv64651839
ft707038,673
Table 4. SOP instances from the TSPLIB dataset.
Table 4. SOP instances from the TSPLIB dataset.
ProblemNodesBest Known Solution
br17.101855
br17.121855
p43.14428,140
p43.24428,480
p43.34428,835
p43.44483,005
ry48p.14915,805
ry48p.24916,074
ry48p.34919,490
ry48p.44931,446
ft53.1547531
ft53.2548026
ft53.35410,262
ft53.45414,425
ft70.17139,313
ft70.27140,101
ft70.37142,535
ft70.47153,530
Table 5. Case study 1: Summary and results of experiment with the “ft53” instance.
Table 5. Case study 1: Summary and results of experiment with the “ft53” instance.
Problem typeATSP
Problemft53
Average distance9326.63
Minimum distance8182
Episode of minimum distance2877
Optimal distance presented by TSPLIB6905
Percentage relative error18.49%
Discount factor0.4
Learning rate0.707
E-greedy policy0.01
Number of episodes10,000
Runtime25.4 s
Table 6. Case study 2: Summary and results of experiment with the “eil51” instance.
Table 6. Case study 2: Summary and results of experiment with the “eil51” instance.
Problem typeTSP
Problemeil51
Average distance562.4
Minimum distance475
Episode of minimum distance4670
Epoch of minimum distance4
Optimal distance presented by TSPLIB426
Percentage relative error11.5%
Discount factor0.34
Learning rate0.69
E-greedy policy0.01
Number of episodes10,000
Number of epochs5
Runtime7.28 min
Table 7. Case study 3: Summary and results of experiment with the “ESC78” instance.
Table 7. Case study 3: Summary and results of experiment with the “ESC78” instance.
Problem typeSOP
ProblemESC78
Average distance23,097.72
Minimum distance19,910
Episode of minimum distance7486
Optimal distance18,230
Percentage relative error9.22%
Discount factor0.415
Learning rate0.68
E-greedy policy0.13
Number of episodes8275
Runtime33.88 min
Table 8. Comparison between the results obtained with AutoML, without AutoML, with TSPLIB, and other results in the literature.
Table 8. Comparison between the results obtained with AutoML, without AutoML, with TSPLIB, and other results in the literature.
ProblemTypeTSPLIBWithout AutoMLWith AutoMLOther Authors
berlin52TSP7542904478717871 [86]
eil76TSP538603565545.39 [92]
p43ATSP5620573156275621.40 [93]
ftv64ATSP1839301021002140 [94]
br17.10SOP55575555 [33]
ft70.1SOP39,31364,22556,64239,313 [95]
Table 9. Comparison of experiment time between modules with AutoML and without AutoML.
Table 9. Comparison of experiment time between modules with AutoML and without AutoML.
ProblemWithout AutoML (s)With AutoML (s)
berlin525.2472.2
eil7617.3813.6
p4312.31644.4
ftv6420.181150.8
br17.1012.172073.6
ft70.134.6335,004.0
Table 10. Comparison between AutoRL-Sim and other reinforcement learning frameworks: I [39], II [42], and III [36].
Table 10. Comparison between AutoRL-Sim and other reinforcement learning frameworks: I [39], II [42], and III [36].
AutoRL-SimIIIIII
EnvironmentR---
MATLAB--
Visual Studio---
AlgorithmsSARSA
Q-learning
Advanced techniquesAutoML---
Problems addressedTSP--
ATSP---
SOP---
MKP---
Navigation--
Reservoir system---
FeaturesParameter optimization---
Selecting instances--
Free module---
Graphics
Reports
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Souza, G.K.B.; Ottoni, A.L.C. AutoRL-Sim: Automated Reinforcement Learning Simulator for Combinatorial Optimization Problems. Modelling 2024, 5, 1056-1083. https://doi.org/10.3390/modelling5030055

AMA Style

Souza GKB, Ottoni ALC. AutoRL-Sim: Automated Reinforcement Learning Simulator for Combinatorial Optimization Problems. Modelling. 2024; 5(3):1056-1083. https://doi.org/10.3390/modelling5030055

Chicago/Turabian Style

Souza, Gleice Kelly Barbosa, and André Luiz Carvalho Ottoni. 2024. "AutoRL-Sim: Automated Reinforcement Learning Simulator for Combinatorial Optimization Problems" Modelling 5, no. 3: 1056-1083. https://doi.org/10.3390/modelling5030055

Article Metrics

Back to TopTop