1. Introduction
Metaheuristics can be classified into various categories based on their natural inspiration [
1]. One prominent category is swarm intelligence-based algorithms, which draw inspiration from the collective behavior of social insect colonies, bird flocks, or animal herds. Swarm intelligence (SI) algorithms simulate the cooperative and self-organizing behavior observed in natural swarms to solve complex optimization problems [
2,
3].
SI, inspired by the collective behavior of social insect colonies, encompasses a diverse range of algorithms that facilitate efficient problem-solving through cooperation and self-organization. These algorithms simulate the collaboration and information exchange observed in natural swarms, enabling them to achieve global optimization. By harnessing the collective intelligence exhibited by swarm systems, SI metaheuristics offer promising avenues for optimizing neural network training and enhancing the identification and control capabilities of robotic systems.
To conduct a comprehensive analysis, we selected four state-of-the-art SI algorithms known for their unique characteristics and optimization strategies. Ant Lion Optimization (ALO), drawing inspiration from the hunting behavior of ant lions; employs a powerful search mechanism to explore and exploit the solution space efficiently. Bat Algorithm (BA) mimics the echolocation behavior of bats, utilizing frequency tuning and pulse emission concepts to achieve effective optimization. Grey Wolf Optimizer (GWO) emulates the social hierarchy and hunting dynamics of grey wolves, employing three fundamental types of wolf-inspired operators to strike a balance between exploration and exploitation. Moth Flame Optimization (MFO), inspired by the moth’s phototaxis behavior toward flames, incorporates attraction and repulsion mechanisms to guide the optimization process effectively. Finally, we compare these algorithms against the well-established Particle Swarm Optimization (PSO), which draws inspiration from the social behavior of bird flocking, enabling particles to adaptively search the solution space based on individual and swarm experience.
Overall, SI algorithms offer powerful optimization techniques that leverage the collective intelligence and self-organization observed in natural swarms. Their robustness, global exploration capabilities, self-adaptation, parallelism, scalability, and bio-inspired concepts make them well-suited for addressing a wide range of optimization problems in various domains. By mimicking the behavior of swarms, these algorithms provide effective solutions and insights for solving complex optimization challenges [
4,
5].
Furthermore, the nature-inspired characteristics of SI optimization methods introduce robustness and adaptability to different problem domains. They can be readily applied to robotic systems, including those with complex dynamics and uncertain environments [
6]. By employing metaheuristic optimization techniques, the Unscented Kalman Filter (UKF) initialization parameters can be tailored to specific robotic platforms and tasks, leading to an improved estimation and control performance.
By integrating SI algorithms with UKF-based neural training, we aim to improve the accuracy of identification and control in a two-degrees-of-freedom (DOF) robot manipulator.
To elaborate further, let us delve into the distinguishing features and underlying principles of ALO, BA, GWO, and MFO algorithms. ALO utilizes a population of artificial ant lions to mimic hunting behaviors, where each ant lion represents a potential solution in the search space [
7]. The algorithm employs pride update and position update mechanisms to perform efficient exploration and exploitation. BA, on the other hand, emulates the echolocation behavior of bats to optimize solutions. Bats navigate through a combination of random flight, frequency tuning, and pulse emission, allowing them to find optimal solutions in dynamic environments [
8]. GWO, inspired by the cooperative hunting dynamics of grey wolves, utilizes three types of wolf operators (alpha, beta, and delta) to balance exploration and exploitation. The alpha wolf coordinates exploration, while the beta and delta wolves perform local exploitation and global exploration, respectively [
9]. MFO draws inspiration from the attraction of moths to flame, employing attraction and repulsion mechanisms to guide the optimization process effectively. Moths are attracted to the light source but are also repelled by other moths, leading to a balanced exploration–exploitation trade-off [
10].
To compare these algorithms against the widely used PSO, we consider PSO ability to adaptively search the solution space based on individual and swarm experience. PSO employs velocity updates and position adjustments to explore and exploit the search space efficiently [
1,
11,
12]. By contrasting the unique characteristics and optimization strategies of ALO, BA, GWO, and MFO with the PSO algorithm, we can gain valuable insights into their relative strengths and weaknesses by applying them to our proposed identification and control scheme. We selected these algorithms mainly because of the following advantages: their small numbers of tuning parameters, low CPU time costs, the ability to maintain joint torque limits, and a better overall performance than other SI algorithms, such as Artificial Bee Colony (ABC), Ant Colony Optimization (ACO), Cuckoo Search (CS), Accelerated Particle Swarm Optimization (APSO), and Whale Optimization Algorithm (WOA).
The structure of this work is as follows. In
Section 2, we mentioned some main characteristics of the SI algorithms employed.
Section 3 describes the methodology of neural identification and control scheme. The simulation and real-time results for trajectory tracking are presented in
Section 4. Discussions of the results are reflected in
Section 5. Finally, concluding remarks are given in
Section 6.
2. Swarm Intelligence Algorithms
Metaheuristics are a family of optimization algorithms designed to find suitable solutions for complicated optimization problems. In contrast to traditional optimization methods, which aim to find the global optimal, metaheuristic algorithms obtain acceptable results quickly, even in the presence of multiple local optima.
In summary, metaheuristics are fantastic tools for finding good solutions to a wide variety of optimization problems. They are especially useful in situations where traditional methods are not effective, such as problems with high dimensionality, non-convex, noise, or incomplete data.
Bio-inspired algorithms are unique metaheuristic methods inspired by natural processes, phenomena, concepts, and systems mechanisms. Each has features and strengths that provide interpretability and inspiration for solving real-world problems in diverse fields, such as engineering, computer science, economics, and biology. These algorithms mimic the behavior of systems in nature, such as evolutionary computation and swarm behavior [
1].
SI is a subfield of bio-inspired algorithms that draws inspiration from collective behavior in nature and focuses on the emergent behavior of decentralized populations through local interactions and self-organization.
Figure 1 shows the flow chart of the proposed methodology, which is described as follows: firstly, the UKF parameters to be optimized are selected; then, each SI optimization algorithm is executed within an off-line simulation. Once the UKF initialization parameters
,
, and
are obtained, they are applied in real-time in the decentralized neural block control (DNBC) scheme for the trajectory tracking task of a 2-DOF robot manipulator. Finally, the results are compared according to the objective function evaluation.
Although there is an endless number of SI algorithms, which have multiple modifications, in the proposed approach, we use algorithms in their original versions, of which a brief description is presented below.
2.1. Particle Swarm Optimization (PSO)
Particle Swarm Optimization (PSO) is a widely used metaheuristic algorithm inspired by the collective behavior of bird flocks or fish schools. It has demonstrated remarkable success in solving various optimization problems [
11]. PSO operates on the principle of iteratively adjusting the positions and velocities of particles in a multidimensional search space. Algorithm 1 shows the implementation of PSO [
12].
Algorithm 1 PSO Pseudocode |
Require: |
1: n: number of particles |
2: d: dimension of the search space |
3: : maximum number of iterations |
4: w: inertia weight |
5: : cognitive acceleration coefficient |
6: : social acceleration coefficient |
7: : position of particle i |
8: : velocity of particle i |
9: : best position of particle i |
10: : global best position |
Ensure: |
11: : optimal solution |
12: : optimal fitness |
13: Function PSO() |
14: for to n do |
15: Initialize and randomly |
16: end for |
17: while do |
18: for to n do |
19: Update velocity: |
20: Update position: |
21: end for |
22: for to n do |
23: if then |
24: |
25: end if |
26: if then |
27: |
28: end if |
29: end for |
30: |
31: end while |
32: |
33: |
34: end Function |
2.2. Ant Lion Optimizer (ALO)
The Ant Lion Optimizer (ALO) is a potent metaheuristic algorithm inspired by the predatory behavior of ant lions. It has gained significant attention in optimization engineering due to its ability to deal with complex problems effectively. ALO displays unique characteristics that distinguish it from other metaheuristics, such as PSO [
7].
ALO emulates the hunting strategy employed by ant lions to capture their prey, which consists of building conical pits in sandy areas. For optimization problems, ALO capitalizes on this natural behavior to explore and exploit the solution space efficiently. Algorithm 2 shows the implementation of ALO [
13].
Algorithm 2 ALO Pseudocode |
Require: |
1: p: number of antlions |
2: u: upper bounds of variables |
3: l: lower bounds of variables |
4: : evaporation rate |
5: : attractiveness rate |
6: : maximum number of iterations |
7: : position of antlion i |
8: : fitness of antlion i |
9: : best position of antlion i |
10: : best performance of antlion i |
Ensure: |
11: : optimal solution |
12: : optimal fitness |
13: Function ALO() |
14: for to p do |
15: Initialize randomly |
16: end for |
17: while do |
18: for to p do |
19: Generate a new position |
20: Calculate the fitness of |
21: if then |
22: |
23: end if |
24: end for |
25: for to p do |
26: Evaporation: |
27: Attractiveness: |
28: end for |
29: |
30: end while |
31: |
32: |
33: end Function |
2.3. Bat Algorithm (BA)
The Bat Algorithm (BA) is an SI algorithm inspired by the echolocation behavior of bats. The BA demonstrates unique characteristics that set it apart from other metaheuristic algorithms [
8].
The algorithm begins by initializing a population of bats, where each bat represents a potential solution to the optimization problem. Bats fly through the search space, continuously adjusting their positions and velocities based on their knowledge. BA implementation is illustrated in Algorithm 3 [
1].
Algorithm 3 BA Pseudocode |
Require: |
1: n: number of bats |
2: d: dimension of the search space |
3: : maximum number of iterations |
4: A: loudness |
5: r: pulse rate |
6: : cooling factor |
7: : wavelength |
8: : position of bat i |
9: : fitness of bat i |
10: : best position of bat i |
11: : best fitness of bat i |
Ensure: |
12: : optimal solution |
13: Function BA() |
14: for to n do |
15: Initialize randomly |
16: end for |
17: while do |
18: for to n do |
19: Generate a new position |
20: Calculate the fitness of |
21: if then |
22: |
23: end if |
24: Update loudness: |
25: Update pulse rate: |
26: end for |
27: for to n do |
28: Probability of loudness: |
29: Probability of pulse rate: |
30: if then |
31: |
32: else |
33: |
34: end if |
35: end for |
36: |
37: end while |
38: |
39: |
40: end Function |
2.4. Grey Wolf Optimizer (GWO)
The Grey Wolf Optimizer (GWO) is an SI algorithm inspired by the hunting behavior of grey wolves in nature. The GWO imitates the social hierarchy and cooperative hunting strategies observed in wolf packs to guide the search for optimal solutions [
9].
In the GWO, a population of candidate solutions, represented as grey wolves, explores the search space by adjusting their positions and mimicking the hunting behaviors of alpha, beta, and delta wolves. GWO implementation is shown in Algorithm 4 [
14].
Algorithm 4 GWO Pseudocode |
Require: |
1: n: number of wolves |
2: d: dimension of the search space |
3: : maximum number of iterations |
4: a: alpha coefficient |
5: b: beta coefficient |
6: c: delta coefficient |
7: : position of wolf i |
8: : fitness of wolf i |
9: : best position of wolf i |
10: : best fitness of wolf i |
Ensure: |
11: : optimal solution |
12: Function GWO() |
13: for to n do |
14: Initialize randomly |
15: end for |
16: while do |
17: for to n do |
18: Calculate a, b, and c |
19: Update position: |
20: end for |
21: for to n do |
22: if then |
23: |
24: end if |
25: end for |
26: |
27: end while |
28: |
29: |
30: end Function |
2.5. Moth Flame Optimization (MFO)
The Moth Flame Optimization (MFO) is an SI algorithm inspired by the navigation behavior of moths in nature. The MFO mimics the attraction of moths toward artificial light sources to guide the search for optimal solutions [
10]. Algorithm 5 displays MFO implementation [
15].
Algorithm 5 MFO Pseudocode |
Require: |
1: n: number of moths |
2: d: dimension of the search space |
3: : maximum number of iterations |
4: a: absorption coefficient |
5: r: random number |
6: : position of moth i |
7: : fitness of moth i |
8: : best position of moth i |
9: : best fitness of moth i |
Ensure: |
10: : optimal solution |
11: Function MFO() |
12: for to n do |
13: Initialize randomly |
14: end for |
15: while do |
16: for to n do |
17: Generate a new position |
18: Calculate the fitness of |
19: if then |
20: |
21: end if |
22: Absorption: |
23: Random walk: |
24: end for |
25: |
26: end while |
27: |
28: |
29: end Function |
3. Decentralized Neural Block Control (DNBC-UKF)
This section shows the proposed SI optimization approach for UKF learning of decentralized neural block control (DNBC-UKF) [
16] applied to a 2-DOF robot manipulator.
For this purpose, we take the system to the following form
where
,
,
.
N is the number of subsystems and
is the input vector.
,
, and
are assumed smooth and bounded functions, with
, and
; in addition, the structures of the subsystems are expressed by
. On the other hand, the interconnection terms
are described by reflecting the relation between the
-th subsystem and the other ones.
The following RHONN structure is used in order to identify the behavior of system (
1)
where
is the
j-th block neuron state with
and
;
are fixed parameters with
.
is the activation function and
represents the input vector.
The NN training task consists of finding values of
that minimize the identification error. For this reason, we propose to use a learning method using only the identification error information, such as the UKF described in
Figure 2.
The UKF is a powerful estimation and control tool: with wide applications in control theory; for the identification of nonlinear systems; and for the training of neural networks. Its versatility lies in its ability to handle nonlinear dynamics, non-Gaussian distributions, and uncertainties associated with real-world systems.
In the context of control theory, the UKF serves as an efficient means for state estimation in nonlinear systems. It allows for the real-time estimation of the system’s internal states, which are often unobservable or difficult to measure directly [
17]. By incorporating nonlinear models and the measured system outputs, the UKF provides accurate and reliable estimates of the system’s states, enabling effective control strategies to be devised. The estimated states obtained from the UKF can then be utilized for feedback control, trajectory tracking, and system stabilization in a wide range of dynamic systems [
18,
19].
The UKF also plays a significant role in the identification of nonlinear systems. Identification refers to the process of determining the mathematical models or parameters that represent the underlying dynamics of a system based on observed input–output data. Nonlinear systems pose significant challenges in identification due to their complex dynamics. The UKF addresses these challenges by iteratively updating the system model parameters, enabling an accurate estimation of the nonlinear system’s behavior. By leveraging the filtering and estimation capabilities of the UKF, researchers and engineers can effectively identify the dynamics, parameters, and structure of complex nonlinear systems, leading to an improved understanding and control of such systems [
20].
Additionally, the UKF is employed in the training of neural networks, specifically in the context of Recurrent Neural Networks (RNNs). RNNs are powerful architectures for modeling sequential data and time series. It is possible to use the UKF in the training process to optimize the internal states, weights, and biases of the network in order to improve its learning capability and prediction accuracy [
21]. By incorporating the UKF within the training process, the neural network can effectively capture and model the complex nonlinear dependencies present in the data, leading to an improved performance in tasks such as time series forecasting, speech recognition, and natural language processing [
22,
23].
The initialization of a UKF involves determining the initial state estimate, the covariance matrix and the process noise covariance matrix, which can make the selection of these initialization parameters a complex task, especially when using heuristic techniques [
18]. The challenge arises because these parameters significantly impact the filter’s performance and are often problem-specific, requiring domain expertise and careful tuning. The manual selection of these parameters can be time-consuming, and may not guarantee optimal performance.
In such scenarios, employing metaheuristic optimization methods proves to be a promising approach for selecting these initialization parameters. Metaheuristic optimization methods offer several advantages when applied to the selection of UKF initialization parameters. Firstly, these methods provide a systematic and automated approach to parameter tuning, relieving the burden of manual parameter selection. They can efficiently explore the vast parameter space, searching for the optimal combination that minimizes the error or maximizes a performance metric. By leveraging the search mechanisms inherent in metaheuristic algorithms, such as exploration and exploitation, the initialization parameters can be fine-tuned to enhance the convergence and accuracy of the UKF.
Secondly, metaheuristic optimization methods can handle nonlinearity, multimodality, and non-convexity in the optimization landscape, which are common challenges in parameter selection for UKF initialization. These algorithms possess the flexibility to adapt and explore diverse regions of the parameter space, avoiding local optima and finding near-optimal or globally optimal solutions.
3.1. SI Optimization for UKF Learning
Using the UKF to estimate the NN weights, and correcting for identification errors, the filter is updated at each step. Usually, , , and are initialized as diagonal matrices with entries , , and , respectively. Given that, typically, these entries are defined heuristically, we propose employing SI methods to improve the UKF training algorithm.
According to the optimal control theory [
24], it is common to use error-based performance measures such as those described in
Table 1. On the other hand, in (
3), the Bolza form [
25] is described and used as an objective function to evaluate the overall performance, including information from the control input of the system.
where
is the initial iteration;
is the final iteration;
T is the sampling time;
is the error vector;
is the control input vector; and
L,
Q, and
R are gain matrices with appropriate dimensions.
Although the Bolza form is a good performance criterion, it presents a serious disadvantage for this work. The final value of the error vector is not very useful for our methodology because it does not significantly represent the system identification and trajectory tracking; in other words, we need to know how it behaves throughout the simulation. For this reason, we propose an objective function, based on that found in [
26], for the SI algorithms as follows:
where MSE represents the mean square error;
t is the total time of the simulation;
represents the identification error;
is the tracking error;
represents the input control; and
,
, and
are scaling factors to bring all the terms of the objective function to a similar order.
3.2. DNBC-UKF Controller Design
Once the RHONN training has been defined, we design a controller based on the tracking error
as follows:
where
is the desired trajectory signal and
is the NN state [
27].
The new value is obtained as:
Then, system (
2) should be expressed as a function of variables
as:
When a sliding mode control strategy is implemented, the control input must be limited by
as:
The sliding surface is designed as
; then, system (
7) is rewritten as follows:
The proper selection of the sliding manifold [
28] presents the possibility of finding a bounded control law by
; the control
is composed as
where
is calculated from
as
Figure 3 illustrates the block diagram of the proposed SI optimization approach.
4. Results
The performance of the proposed approach was analyzed and compared using the following SI algorithms: ABC, ACO, ALO, BA, CS, GWO, MFO, PSO, APSO, and WOA. The comparisons were performed to find the algorithm that best minimizes the trajectory tracking error without exceeding the limits of the input torques.
This last condition of not surpassing the torque bounds is necessary for algorithm selection in real-time experiments. The experiments were performed on a 2-DOF vertical direct-drive robot manipulator, which is located at the Tecnologico Nacional de Mexico/Instituto Tecnologico de La Laguna, Mexico.
4.1. Prototype Description
To illustrate the implementation of the proposed scheme, we used the robot manipulator shown in
Figure 4, which consists of two rigid links articulated by high-torque brushless direct-drive servos that present a reduced backlash and a significantly lower joint friction to drive the joints. The robot actuators act as torque sources and receive analog voltage as a torque reference signal. Joint positions are obtained using incremental encoders that send information to a DAQ [
16].
The numerical values for the 2-DOF robot manipulator parameters alongside the dynamic model can be found in [
29].
In order to prove the proposed approach, the discrete-time trajectories [
27] were chosen as
where
,
,
, and
are used for the first joint, while
,
,
, and
are used for the second joint.
4.2. Simulation Results
For the simulations, the parameter settings of the SI algorithms considered are summarized as follows: starting with the common parameters, the number of iterations, which was 15; the population size was 30; the variables were 54; and the simulation time was 10 s. The particular parameter settings are given in
Table 2.
SI algorithms presented in
Table 2 were considered for comparison purposes because they have been employed previously in the state-of-the-art for neural network training, as reported in [
39,
40,
41,
42,
43,
44,
45,
46,
47,
48].
In all simulations, the specifications of the test machine were an AMD Ryzen 9 4900HS® (AMD Ryzen is a registered trademark of Advanced Micro Devices, Inc., Santa Clara, CA, USA ) CPU 3.0 GHz and 16 GB of RAM. Moreover, the experiments were performed in the MATLAB R2015a environment® (MATLAB is a registered trademark of MathWorks, Inc., Natick, MA, USA).
For comparative purposes, we tested each SI algorithm 50 times independently. To qualify the results, we calculated statistical data of the mean, standard deviation (SD), and the best and worst results for different performance indices and the proposed objective function. The performance of the algorithms is reflected in a small mean value with a low standard distribution, implying a small difference between the best and worst results.
Table 3 shows the performance measures.
4.3. Experimental Results
The selection of the ALO, BA, GWO, MFO, and PSO algorithms for real-time experiments was because they show a balanced performance with low computational costs. However, the main reason was that during the entire simulation, none of them exceeded the torque limits.
Real-time experiments were implemented using Ansi C on WinMechLab, a real-time platform running on an Intel Pentium 4 PC with real-time Windows XP, with a 2.5 ms sampling period, and using a MultiQ-PCI data acquisition board from Quanser Consulting Inc., Markham, ON, CAN [
49].
Figure 5 and
Figure 6 show the tracking trajectories for each link obtained in the simulation using the SI algorithms compared to the non-optimized UKF, we include in
Table 4 and
Figure 7 the values of the performance of the algorithms by evaluating the objective function and
-norm, described by
where
T is the sampling time and
t is the total time of the simulation, which, for this case is 20 s. In addition,
Table 5 includes the RMS of the joint input torques comparing each algorithm with the UKF with non-optimized parameters. Moreover, the input pairs for each link are shown in
Figure 8a to
Figure 9e.
5. Discussion
For the tracking control problem of a 2-DOF robot manipulator, we proposed a DNBC controller, which does not require any knowledge of the system since it uses a UKF-trained neural identifier, whose added value presents an optimization tuning via SI algorithms. Everyone tuned the initial filter parameters in an off-line simulation subsequently used in real-time experiments.
The selection of SI algorithms employed in this work is due to the following criteria:
The computational cost is the factor that we consider to be most important when carrying out the proposed optimization task, at least for the off-line simulation. Since the stochastic nature of both the UKF filter and the algorithms must be taken into account, in addition to the added cost of the controller, this translates into an increase in the number of attempts to achieve acceptable results. For this reason,
Table 3 shows the average CPU time for each SI algorithm utilized.
Interpreting the results obtained in the off-line simulation, we can highlight the following:
According to
Table 2 and
Table 3, all algorithms meet the selection criteria, have few parameters, acceptable performance, and relatively low CPU time.
All performance indices show similar average CPU time for each algorithm. The algorithm with the lowest average CPU time was BA, while the highest was ABC.
Based on the statistical mean and standard deviation, the best-performing algorithm overall was ALO, followed by BA, GWO, and MFO, in that order.
In the case of real-time experiments, the principal selection criterion is to maintain joint torque limits. Only the Bolza form (
3) and the objective function (
4) of all performance measures consider the joint torques. Although
Table 3 shows that all algorithms perform well, input torque signals in the simulation of the ABC, ACO, CS, APSO, and WOA algorithms exceeded the joint torque limits. For this reason, they were discarded from being used in real-time experiments.
Interpreting the results obtained in real-time, we can reach the following final remarks:
According to
Figure 5,
Figure 6 and
Figure 7, all algorithms perform better than the UKF without optimization. This demonstrates the advantages of using the proposed methodology.
Table 4 shows the values of the objective function (
4) evaluated in all used algorithms; the performance of these algorithms is reflected by minimizing the value of the objective function since it is described in terms of tracking and identification errors. As we can observe, the one with the best performance is the ALO, which we can also contrast in
Figure 7. The controller performance shows a notorious improvement concerning a previous heuristic tuning.
Concerning
Figure 5, the performance of the GWO for the first link is not up to par. However, according to
Figure 6, this algorithm on the second link performs better than the other algorithms.
Figure 9c exhibits this in the noise of the input torque signal, which is reflected in the RMS value in
Table 5.
6. Conclusions
In this work, we have presented the implementation of SI-inspired algorithms in the selection of UKF initialization parameters and their real-time application in a discrete-time decentralized neural block control scheme. We proposed a new objective function that effectively utilizes information from trajectory tracking and identification errors paired with the slopes of the input torques. This function allows us to meet minimizing tracking errors without overshooting the bounds on the control input signals. We performed in simulation a comparative experimental study of the performance of the following SI algorithms: ABC, ACO, ALO, BA, CS, GWO, MFO, PSO, APSO, and WOA. For this analysis, we used five performance indices in addition to our proposed objective function. The real-time experiments were carried out on a 2-DOF robot manipulator, showing ALO, BA, GWO, MFO, and PSO performance, which were the best in our comparative study.
In summary, the main contribution of this work is the implementation of the use of SI-inspired algorithms in the selection of UKF initialization parameters and its real-time implementation in a discrete-time decentralized neural block control; moreover, an experimental comparative study of performance was carried out between the ALO, BA, GWO, MFO, and PSO.
Author Contributions
Conceptualization, J.F.G., R.G.-H., M.A.L. and V.S.; methodology, J.F.G., R.G.-H., M.A.L. and V.S.; software, J.F.G. and V.S.; validation, J.F.G., R.G.-H., M.A.L. and V.S.; formal analysis, J.F.G., R.G.-H., M.A.L. and V.S.; investigation, J.F.G., R.G.-H., M.A.L. and V.S.; resources, J.F.G., R.G.-H., M.A.L. and V.S.; writing—original draft preparation, J.F.G. and R.G.-H.; writing—review and editing, J.F.G., R.G.-H., M.A.L. and V.S.; visualization, J.F.G., R.G.-H., M.A.L. and V.S.; supervision, R.G.-H., M.A.L. and V.S.; project administration, R.G.-H.; funding acquisition, R.G.-H., M.A.L. and V.S. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by financial support of the research projects of the Tecnológico Nacional de México/I. T. La Laguna and CONACYT.
Data Availability Statement
Data sharing not applicable.
Acknowledgments
The authors would like to thank to all the staff in División de Estudios de Posgrado e Investigación del Tecnológico Nacional de México/I. T. de La Laguna.
Conflicts of Interest
The authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
ABC | Artificial Bee Colony |
ACO | Ant Colony Optimization |
ALO | Ant Lion Optimizer |
APSO | Accelerated Particle Swarm Optimization |
BA | Bat Algorithm |
CS | Cuckoo Search |
DOF | Degrees of Freedom |
EKF | Extended Kalman filter |
GWO | Grey Wolf Optimizer |
IAE | Integral Absolute Error |
ISE | Integral Squared Error |
ITAE | Integral Time-weighted Absolute Error |
ITSE | Integral Time-weighted Squared Error |
MFO | Moth Flame Optimization |
MSE | Mean Square Error |
NBC | Nonlinear Block Control |
NN | Neural network |
PSO | Particle Swarm Optimization |
RHONN | Recurrent High Order Neural Networks |
RMS | Root Mean Square |
SD | Standard Deviation |
SI | Swarm Intelligence |
UKF | Unscented Kalman filter |
WOA | Whale Optimization Algorithm |
References
- Yang, X.S.; Karamanoglu, M. Nature-inspired computation and swarm intelligence: A state-of-the-art overview. In Nature-Inspired Computation and Swarm Intelligence; Elsevier: Amsterdam, The Netherlands, 2020; pp. 3–18. [Google Scholar]
- Tang, J.; Liu, G.; Pan, Q. A review on representative swarm intelligence algorithms for solving optimization problems: Applications and trends. IEEE/CAA J. Autom. Sin. 2021, 8, 1627–1643. [Google Scholar] [CrossRef]
- Alanis, A.Y. Bioinspired Intelligent Algorithms for Optimization, Modeling and Control: Theory and Applications. Mathematics 2022, 10, 2334. [Google Scholar] [CrossRef]
- Degas, A.; Islam, M.R.; Hurter, C.; Barua, S.; Rahman, H.; Poudel, M.; Ruscio, D.; Ahmed, M.U.; Begum, S.; Rahman, M.A.; et al. A survey on artificial intelligence (AI) and eXplainable AI in air traffic management: Current trends and development with future research trajectory. Appl. Sci. 2022, 12, 1295. [Google Scholar] [CrossRef]
- Cao, X.; Yan, H.; Huang, Z.; Ai, S.; Xu, Y.; Fu, R.; Zou, X. A multi-objective particle swarm optimization for trajectory planning of fruit picking manipulator. Agronomy 2021, 11, 2286. [Google Scholar] [CrossRef]
- Malik, A.; Henderson, T.; Prazenica, R. Multi-objective swarm intelligence trajectory generation for a 7 degree of freedom robotic manipulator. Robotics 2021, 10, 127. [Google Scholar] [CrossRef]
- Mirjalili, S. The ant lion optimizer. Adv. Eng. Softw. 2015, 83, 80–98. [Google Scholar] [CrossRef]
- Yang, X.S.; He, X. Bat algorithm: Literature review and applications. Int. J. Bio-Inspired Comput. 2013, 5, 141–149. [Google Scholar] [CrossRef]
- Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
- Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
- Eberhart, R.; Kennedy, J. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; IEEE: New York, NY, USA, 1995; Volume 4, pp. 1942–1948. [Google Scholar]
- Rao, S.S. Engineering Optimization: Theory and Practice; John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar]
- Oliveira, J.; Oliveira, P.M.; Boaventura-Cunha, J.; Pinho, T. Evaluation of hunting-based optimizers for a quadrotor sliding mode flight controller. Robotics 2020, 9, 22. [Google Scholar] [CrossRef]
- Panda, M.; Das, B. Grey wolf optimizer and its applications: A survey. In Proceedings of the Third International Conference on Microelectronics, Computing and Communication Systems: MCCS 2018; Springer: Singapore, 2019; pp. 179–194. [Google Scholar]
- William, M.V.A.; Ramesh, S.; Cep, R.; Kumar, M.S.; Elangovan, M. MFO Tunned SVR Models for Analyzing Dimensional Characteristics of Cracks Developed on Steam Generator Tubes. Appl. Sci. 2022, 12, 12375. [Google Scholar] [CrossRef]
- Guerra, J.F.; Garcia-Hernandez, R.; Llama, M.A.; Santibañez, V. UKF-Based Neural Training for Nonlinear Systems Identification and Control Improvement. IEEE Access 2022, 10, 114501–114513. [Google Scholar] [CrossRef]
- Wan, E.A.; Van Der Merwe, R. The unscented Kalman filter for nonlinear estimation. In Proceedings of the Adaptive Systems for Signal Processing, Communications, and Control Symposium 2000, AS-SPCC, the IEEE 2000, Lake Louise, AB, Canada, 4 October 2000; IEEE: New York, NY, USA, 2000; pp. 153–158. [Google Scholar]
- Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 401–422. [Google Scholar] [CrossRef]
- Simon, D. Optimal State Estimation: Kalman, H Infinity, and Nonlinear Approaches; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
- Särkkä, S. Bayesian Filtering and Smoothing; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
- Zhang, G.; Qi, Y. Neural network-based nonlinear dynamic modeling for process control. Control Eng. Pract. 2005, 13, 185–192. [Google Scholar]
- Zhang, Z. Neural networks for nonlinear dynamic system modeling and identification. In Proceedings of the Advances in Neural Networks; Springer: Berlin/Heidelberg, Germany, 2008; pp. 268–272. [Google Scholar]
- Alanis, A.Y.; Arana-Daniel, N.; Lopez-Franco, C. Bio-Inspired Algorithms for Engineering; Butterworth-Heinemann: Oxford, UK, 2018. [Google Scholar]
- Kirk, D.E. Optimal Control Theory: An Introduction; Courier Corporation: North Chelmsford, MA, USA, 2004. [Google Scholar]
- Sethi, S.P.; Sethi, S.P. What Is Optimal Control Theory? Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
- Llama, M.; Flores, A.; Garcia-Hernandez, R.; Santibañez, V. Heuristic global optimization of an adaptive fuzzy controller for the inverted pendulum system: Experimental comparison. Appl. Sci. 2020, 10, 6158. [Google Scholar] [CrossRef]
- Garcia-Hernandez, R.; Lopez-Franco, M.; Sanchez, E.N.; Alanis, A.Y.; Ruz-Hernandez, J.A. Decentralized Neural Control: Application to Robotics; Studies in Systems, Decision and Control; Springer: Cham, Switzerland, 2017; Volume 96. [Google Scholar]
- Utkin, V.; Guldner, J.; Shi, J. Sliding Mode Control in Electro-Mechanical Systems; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
- Kelly, R.; Davila, V.S.; Perez, J.A.L. Control of Robot Manipulators in Joint Space; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
- Camarena, O.; Cuevas, E.; Pérez-Cisneros, M.; Fausto, F.; González, A.; Valdivia, A. Ls-II: An improved locust search algorithm for solving optimization problems. Math. Probl. Eng. 2018, 2018, 1–15. [Google Scholar] [CrossRef]
- Deif, D.S.; Gadallah, Y. An ant colony optimization approach for the deployment of reliable wireless sensor networks. IEEE Access 2017, 5, 10744–10756. [Google Scholar] [CrossRef]
- Abualigah, L.; Shehab, M.; Alshinwan, M.; Mirjalili, S.; Elaziz, M.A. Ant lion optimizer: A comprehensive survey of its variants and applications. Arch. Comput. Methods Eng. 2021, 28, 1397–1416. [Google Scholar] [CrossRef]
- Chakri, A.; Ragueb, H.; Yang, X.S. Bat algorithm and directional bat algorithm with case studies. In Nature-Inspired Algorithms and Applied Optimization; Springer: Cham, Switzerland, 2018; pp. 189–216. [Google Scholar]
- Hernandez-Barragan, J.; Martinez-Soltero, G.; Rios, J.D.; Lopez-Franco, C.; Alanis, A.Y. A Metaheuristic Optimization Approach to Solve Inverse Kinematics of Mobile Dual-Arm Robots. Mathematics 2022, 10, 4135. [Google Scholar] [CrossRef]
- Mirjalili, S.; Gandomi, A.H. Comprehensive Metaheuristics: Algorithms and Applications; Elsevier: Amsterdam, The Netherlands, 2023. [Google Scholar]
- Mirjalili, S. Handbook of Moth-Flame Optimization Algorithm: Variants, Hybrids, Improvements, and Applications; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar]
- Mirjalili, S. Evolutionary algorithms and neural networks. In Studies in Computational Intelligence; Springer: Cham, Switzerland, 2019; Volume 780. [Google Scholar]
- Mirjalili, S.; Dong, J.S.; Lewis, A. Nature-inspired optimizers. Stud. Comput. Intell. 2020, 811, 7–20. [Google Scholar]
- Abdolrasol, M.G.; Hussain, S.S.; Ustun, T.S.; Sarker, M.R.; Hannan, M.A.; Mohamed, R.; Ali, J.A.; Mekhilef, S.; Milad, A. Artificial neural networks based optimization techniques: A review. Electronics 2021, 10, 2689. [Google Scholar] [CrossRef]
- Na, Q.; Yin, G.; Liu, A. A novel heuristic artificial neural network model for urban computing. IEEE Access 2019, 7, 183751–183760. [Google Scholar] [CrossRef]
- Heidari, A.A.; Faris, H.; Mirjalili, S.; Aljarah, I.; Mafarja, M. Ant lion optimizer: Theory, literature review, and application in multi-layer perceptron neural networks. In Nature-Inspired Optimizers: Theories, Literature Reviews and Applications; Springer: Cham, Switzerland, 2020; pp. 23–46. [Google Scholar]
- Bangyal, W.H.; Ahmad, J.; Rauf, H.T. Optimization of neural network using improved bat algorithm for data classification. J. Med. Imaging Health Inform. 2019, 9, 670–681. [Google Scholar] [CrossRef]
- Tran-Ngoc, H.; Khatir, S.; De Roeck, G.; Bui-Tien, T.; Wahab, M.A. An efficient artificial neural network for damage detection in bridges and beam-like structures by improving training parameters using cuckoo search algorithm. Eng. Struct. 2019, 199, 109637. [Google Scholar] [CrossRef]
- Zhang, X.; Hou, J.; Wang, Z.; Jiang, Y. Joint SOH-SOC estimation model for lithium-ion batteries based on GWO-BP neural network. Energies 2022, 16, 132. [Google Scholar] [CrossRef]
- Pham, V.D.; Nguyen, Q.H.; Nguyen, H.D.; Pham, V.M.; Bui, Q.T. Convolutional neural network—Optimized moth flame algorithm for shallow landslide susceptible analysis. IEEE Access 2020, 8, 32727–32736. [Google Scholar] [CrossRef]
- Liu, X.h.; Zhang, D.; Zhang, J.; Zhang, T.; Zhu, H. A path planning method based on the particle swarm optimization trained fuzzy neural network algorithm. Clust. Comput. 2021, 24, 1901–1915. [Google Scholar] [CrossRef]
- Khan, A.; Bukhari, J.; Bangash, J.I.; Khan, A.; Imran, M.; Asim, M.; Ishaq, M.; Khan, A. Optimizing connection weights of functional link neural network using APSO algorithm for medical data classification. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 2551–2561. [Google Scholar] [CrossRef]
- Brodzicki, A.; Piekarski, M.; Jaworek-Korjakowska, J. The whale optimization algorithm approach for deep neural networks. Sensors 2021, 21, 8003. [Google Scholar] [CrossRef]
- Pizarro-Lerma, A.; Santibañez, V.; Garcia-Hernandez, R.; Villalobos-Chin, J. Sectorial fuzzy controller plus feedforward for the trajectory tracking of robotic arms in joint space. Mathematics 2021, 9, 616. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).