Next Article in Journal
CAD-Integrated Automatic Gearbox Design with Evolutionary Algorithm Gear-Pair Dimensioning and Multi-Objective Genetic Algorithm-Driven Bearing Selection
Previous Article in Journal
Transient Analysis of Vortex-Induced Pressure Pulsations in a Vertical Axial Pump with Bidirectional Flow Passages Under Stall Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Memetic Algorithm for Energy Optimization in Point-to-Point Robotized Operations

1
Faculty of Informatics, Juraj Dobrila University of Pula, 52100 Pula, Croatia
2
Department of Computer Science and Informatics, University North, 48000 Koprivnica, Croatia
3
Department of Computer Engineering, Faculty of Engineering, University of Rijeka, 51000 Rijeka, Croatia
*
Author to whom correspondence should be addressed.
Machines 2026, 14(1), 35; https://doi.org/10.3390/machines14010035
Submission received: 18 November 2025 / Revised: 12 December 2025 / Accepted: 24 December 2025 / Published: 25 December 2025
(This article belongs to the Section Robotics, Mechatronics and Intelligent Machines)

Abstract

This paper presents a memetic algorithm (MA) for energy cost estimation of a robot path. The developed algorithm uses a random recombination genetic algorithm (GA) as the basis for the first stage of the algorithm and performs a local search based on feature importances determined from the data in the second stage. To allow for the faster determination of the solution quality, the algorithm uses an ML-driven fitness function, based on MLP, for the determination of path energy. The performed tests show that not only does the GA itself optimize the point-to-point paths well, but the usage of MA can lower the energy use by 58% on average (N = 100) when compared to a linear path between the same two points.

1. Introduction

The number of industrial robots (IRs) present in manufacturing plants is ever growing [1]. A common use of IRs is in the so-called point-to-point operation mode, where the movement between two points does not have to follow a specific path; instead, it can move freely between two points in space—with most operations applying a simple movement in between those two points, such as a linear or semi-circular motion. Due to the number of robots and the large amount of working hours applied, the energy required to operate these machines is not insignificant. Assuming a 5 kW power per robot and 1000 work hours a year, the consumption in the European Union area exceeds 15 TWh. It is not surprising that a large number of papers cover the need for energy optimization during IR operations. For this reason, there are many papers that explore the path tuning of IRs with the goal of lowering the energy use. Vysockyy et al. [2] applied a particle swarm optimization (PSO) algorithm for point-to-point robot path optimization, focusing on non-technological movements—segments where the robot is not directly involved in a production task. The optimized Bezier-curve-based trajectories were verified on a UR3 robotic arm, demonstrating energy savings between 10 and 40%. The authors suggest that similar approaches could be extended to technological segments when process constraints allow. Garriz and Domingo [3] optimized sealant application in the automotive industry using a Kalman-based trajectory planner, achieving up to 20% reduction in energy use under realistic constraints, though at the cost of increased cycle time. The authors note that this trade-off may be undesirable in assembly-line environments where timing affects subsequent tasks, indicating that approaches minimizing the energy without prolonging the operation time would be more generally applicable. Shrivastava and Dalla [4] utilized a classical genetic algorithm (GA) to optimize the energy consumption of a multi-axis manipulator based on bond-graph-simulated kinematic and dynamic data. Using joint positions as optimization parameters, their method achieved a 38% improvement compared to direct path planning while maintaining compliance with physical constraints. Similarly, Nonoyama et al. [5] used the K-ROSET simulation environment to apply GA and PSO algorithms to tune the parameters of a PID controller in a SCARA robot. Their results showed that the GA yielded a lower computational cost and improved energy efficiency by 18% relative to the untuned trajectory, though the method’s applicability to other robot types may be limited by the difficulty of adjusting the controller parameters. Luneckas et al. [6] extended heuristic optimization to hexapod robots, applying a Red Fox algorithm to enhance the gait-switching efficiency, resulting in energy savings of up to 21%. These findings indicate that more advanced heuristic approaches can outperform classical algorithms such as GA and PSO in trajectory optimization. Lu et al. [7] further demonstrated the use of a memetic algorithm (MA) combining GA and variable neighborhood search (VNS) for collaborative welding applications, where a local neighborhood search after global optimization improved the results by up to 10%. Despite its success, the study did not explore alternative algorithm combinations for the two stages, even though prior work suggests that other evolutionary algorithms, such as differential evolution (DE), may yield superior results [8]. The main findings of these works, including the highest reported improvements in energy efficiency, are summarized in Table 1. The reported improvements are given as a percentage of the improvement from the initial state. It is important to note that different studies express the values of the fitness function in different ways, resulting in different baselines and physical quantities. The assumption used for the presented value is that with the given optimization strategy a similar enough fitness function would yield results close to what was presented in Table 1 [9]. The main findings of these works, including the highest reported improvements in energy efficiency, are summarized in Table 1.
The research review shows that there is a clear gap in the application of memetic algorithms (MA) in the process of the energy path optimization of IRs. Still, another issue exists in the application of these algorithms, which is the need for the precise and fast evaluation of the actual energy use. ML algorithms have been shown to have a wide array of applications in both robotics and energy use prediction. Gadaleta et al. [10] present an experimental framework for studying and optimizing the energy consumption of IRs. Their setup, based on a KUKA KR210 R2700 Prime robot performing various operations under different loads, serves as a foundation for building data sets that enable analysis of factors influencing energy use. The results indicate that even basic operational adjustments, such as improved temperature control, can reduce energy consumption by up to 50%, while higher-level strategies like trajectory optimization could yield further improvements. This study highlights the importance of creating comprehensive experimental data sets to identify variables with the largest effect on robotic energy efficiency. Zhang and Yan [11] propose a data-driven approach using a back-propagation artificial neural network (ANN) to predict the energy consumption of an Epson C4 IR operating at different speeds and accelerations. Their model achieves an R-value of 0.9699, although the lack of cross-validation and limited data set diversity suggest possible overfitting. Gao et al. [12] employ a Long Short-Term Memory (LSTM) neural network for the torque-based energy prediction of a planar parallel manipulator (PPM), using torque values derived from Newton–Euler dynamics. The model achieves a root mean square error (RMSE) of 0.000041 and a mean absolute percentage error (MAPE) of 2.5%. Lin et al. [13] extend this approach with a batch-normalized (BN) LSTM, trained on 64,600 samples from the Yaskawa GP7 data set, and report an RMSE of 3.67—representing a 23% improvement over previous results. Similarly, Jiang et al. [14] apply an LSTM-based deep neural network to predict the energy consumption of a KUKA KR60-3 IRM, achieving an MAPE of 4.21%. Although LSTM architectures have shown excellent performance in modeling temporal energy trends, they are limited in real-time adaptability. Jaramillo-Morales et al. [15] demonstrate an alternative numeric model for the direct prediction of instantaneous power consumption in a differential-drive robot, the Nomad Super Scout II. While this approach enables use in dynamic environments, its accuracy drops to 81.25% on curved paths, remaining inferior to ML-based models. Overall, these works underline the growing effectiveness of machine learning methods, particularly ANN and LSTM architectures, for predicting robotic energy use, though further refinement is needed for real-time and generalized applications. The summarized results of the reviewed research are presented in Table 2.
Based on this, the following hypotheses can be set:
  • An ML-based model can be used for regressing the value of energy use in the given moment, using the information about IR speed and position,
  • An MA algorithm can be developed to lower the energy use of IR in point-to-point-based operations.
The paper will first demonstrate the build of the algorithm—from the data collection and analysis to the modeling approach, feature importance estimation, and construction of the MA algorithm. Then, the results of each of these steps will be provided and commented on prior to drawing conclusions.

2. Materials and Methods

2.1. Determining the Energy Use of the Robot to Be Used as a Fitness Function

The energy use of the IR can be defined as the total sum of the products of torques and velocities in each individual joint [16]:
E total = i = 0 i = 6 E i = i = 0 i = 6 τ i ω i Δ t ,
where:
  • E total —total energy consumed by the robot during the movement, J;
  • E i —energy consumed by the i-th joint, J;
  • τ i —torque of the i-th joint, N ,   m ;
  • ω i —angular velocity of the i-th joint, rad / s ;
  • Δ t —time step between two measurements, s .
While obtaining the velocity values is simple and direct, the torque of each individual joint is complex—it requires the use of an iterative algorithm such as Newton–Euler or Lagrange–Euler, and the resulting model is often computationally very complex [17]. The complexity issue is non-trivial here, as the applied evolutionary algorithms require the calculation of the total energy a large number of times. If a computationally complex function is used, the time required for optimization could increase significantly [18]. Due to this, the authors used an ML approach, that, while more complex to train, allows for a much faster inference time [19]. The data for training are obtained from the ABB IRB 120 IR. The data are collected from measurements given by the IR control unit, with the code given in Appendix A. The data were collected on the movement between 1000 randomly selected data points. This approach was selected in order to include a higher variety of movement between these data points. As the frequency of collection was 40 Hz—40 data points per second, the final data set consists of 75,790 data points. The torque measurements were obtained from the IR firmware (so-called RobotWare) as provided by the manufacturer, through the manufacturer’s software for control of robots, ABB RobotStudio (version 2024.2) [20]. The values are derived from motor current internally measured by the IR control unit and the motor model (e.g., the torque constants of the motors and filters) [21]. To ensure the results of the models are correct, the data collection experiment was collected to obtain the test set. The test set was generated on 100 points in space, which returned 7298 data points. The models that were trained on the original data in the cross-validation scheme were then re-evaluated on this data set. The collected values in each of these points are as follows:
  • Tool center position (TCP) given as the x, y, and z coordinates of the end of the IR relative to the origin placed in the base of the IR;
  • Quarternions ( q 1 , q 2 , q 3 , and q 4 ) and Euler angles ( ϕ , θ , ψ —encoded as X, Y, and Z, respectively) that define the rotation of the TCP;
  • Speed of each individual joint;
  • Torque of each individual joint.
In addition to this, the following values were calculated for each collected data point:
  • Δ t —the measurement time between two data points,
  • Acceleration of each axis,
  • The change in the linear position of TCP since the last measurement,
  • The change in each of the Euler’s angles,
  • The linear speeds and accelerations of TCP;
  • The angular speeds and accelerations,
  • The total energy that the robot used in the movement per Equation (1).
The measurement was performed according to the diagram given in Figure 1. The torque data were primarily collected by the IRC5, which measures the current sent to the ABB IRB 120 motors through the A05 connection. Other used information, such as the robot position and speed was collected and measured by the Serial Measurement Board (SMB) located in the base of the IRB 120. The SMB sends the collected data back to the IRC5 control unit through an A06 connection. Both of these sets of values are then sent to the RobotStudio instance on a PC over an ethernet connection, where the data can be logged for later use in modeling.
These values were calculated in order to allow for simpler regression of the target output. The training data set is provided in Supplementary Materials Table S1, while the test data set is provided in Supplementary Materials Table S2. All methods have been trained using five-fold cross-validation with grid search for hyperparameter determination. Four ML methods have been tested. The first is multilayer perceptron (MLP). MLP is one of the most fundamental neural network architectures. It consists of an input layer, one or more hidden layers, and an output layer, where each neuron applies a nonlinear activation function to a weighted sum of its inputs [22]. The training process involves adjusting the connection weights through backpropagation to minimize a defined loss function, often using stochastic gradient descent algorithm or its variants [23]. The main advantage of the MLP is its ability to approximate complex nonlinear mappings, making it suitable for regression and classification tasks. However, due to its dense structure, it may require large data sets and regularization to avoid overfitting. The hyperparameters tested within the grid search, for the MLP, are given in Table 3.
Support Vector Regression (SVR) extends the principles of the Support Vector Machine to continuous-valued prediction tasks [24]. It constructs a regression function that deviates from the true targets by at most a predefined margin, while maintaining the model simplicity by maximizing the margin between data points and the regression boundary [25]. The optimization process relies on solving a convex quadratic programming problem that balances the model flatness and tolerance to error. By using kernel functions, the SVR can capture nonlinear dependencies between variables, making it an effective tool for high-dimensional and complex regression problems. The hyperparameter search space for this method is given in Table 4.
The passive-aggressive regressor (PAR) belongs to a family of online learning algorithms that updates the model parameters incrementally as new data become available [26]. Unlike batch learners, it does not retrain on the entire data set but instead performs minimal corrections to the model after each sample. The “passive” aspect reflects that no update is made when the prediction error is within a defined tolerance, whereas the “aggressive” aspect refers to the immediate correction applied when the error exceeds this threshold [27]. This makes the method well-suited for real-time or streaming regression tasks, where adaptability and computational efficiency are essential. The tuned hyperparameters and their possible values are presented in Table 5.
Gradient Boosted Trees (XGB) are an ensemble learning method that combines the outputs of multiple weak learners, typically decision trees, into a strong predictive model [28]. The approach incrementally builds the model in a stage-wise fashion, where each new tree is trained to minimize the residual errors of the previous ensemble using the gradient descent in function space [29]. This allows GBTs to achieve high predictive accuracy and robustness, particularly on structured and tabular data. The trade-off lies in the relatively high computational cost of training, which scales with the number of boosting stages and the depth of the individual trees. The hyperparameters of this method are given in Table 6.

2.2. Feature Importance

Feature importance can be determined through several complementary approaches, each quantifying the contribution of individual predictors to the model output in a different manner. The Mean Decrease in Impurity (MDI) evaluates the importance in tree-based models by averaging the reduction in the impurity of nodes, such as the Gini index or entropy, weighted by the number of samples reaching each node [30]. It can be expressed as follows.
I j = t T j p ( t ) , Δ i ( t ) ,
where p ( t ) is the proportion of samples at node t, and Δ i ( t ) is the decrease in impurities. The feature permutation method estimates importance by measuring the change in model performance when the values of a given feature are randomly permuted, effectively breaking its relationship with the target variable [31]. The resulting importance is computed as
I j = M baseline M perm ( j ) ,
where M denotes the performance metric chosen. Statistical correlation-based methods provide model-agnostic alternatives. The Pearson correlation coefficient measures linear relationships as follows:
r x y = i ( x i x ¯ ) ( y i y ¯ ) i ( x i x ¯ ) 2 i ( y i y ¯ ) 2 ,
while the Spearman rank correlation uses ranked variables to capture monotonic associations, given by
ρ x y = 1 6 i d i 2 n ( n 2 1 ) ,
where d i is the rank difference for sample i [32]. Finally, the Kendall correlation coefficient quantifies the concordance between pairs of observations as
τ = n c n d 1 2 n ( n 1 ) ,
where n c and n d are the numbers of concordant and discordant pairs, respectively [33]. These methods collectively offer a thorough insight into the importance of variables, integrating both model-specific and statistical viewpoints.

2.3. Evolutionary Optimization Algorithms

A genetic algorithm (GA) is a search heuristic and optimization technique inspired by the principles of natural selection and genetics. It is used to solve complex optimization and search problems. Genetic algorithms belong to the larger class of evolutionary algorithms, which generate solutions to optimization problems using techniques inspired by natural evolution, such as inheritance, mutation, selection, and crossover.
GA is an iterative process that starts with the initialization of multiple randomized solutions. For each of these solutions, the quality of the solution is calculated. In the presented research, the solution is the trajectory of the robot in the joint space, defined with six values per each joint:
a 1 b 1 c 1 d 1 e 1 f 1 a 2 b 2 c 2 d 2 e 2 f 2 a 3 b 3 c 3 d 3 e 3 f 3 a 4 b 4 c 4 d 4 e 4 f 4 a 5 b 5 c 5 d 5 e 5 f 5 a 6 b 6 c 6 d 6 e 6 f 6 .
This allows for the joint path (in the joint space) to be defined as
q ( t ) = a 6 + b 5 + c 4 + d 3 + e 2 + f .
After the generation, the iterative part begins. Each of the solutions gets their fitness—the quality of the solution estimated. In the presented case, the function for this is obtained as a model from the previously defined ML algorithm. Then, iteratively, solutions are selected randomly with a bias towards the higher-quality solutions [34,35]. These solutions have the evolutionary operations applied to them: crossover and mutation. Crossover (also refered to as recombination) selects chromosomes and pairs them to produce an offspring solution [36]. This involves swapping portions of their (individual values within the previously shown matrix) structure to create new chromosomes. In mutation, the offspring chromosomes may undergo random changes (mutations) in some of their elements to maintain genetic diversity within the population and to explore new areas of the search space. Finally, the solutions newly generated from this process are used to replace the old solutions. This process is repeated for either a fixed number of generations or until a satisfactory solution is reached.
The GA is a relatively simple algorithm that allows for optimization through the selection of higher quality solutions. There exist many algorithms that use this basic principle and expand upon it, two of which were tested in this research—differential evolution (DE) and memetic algorithm (MA).
DE works in the same manner as the GA, with the simple difference of the recombination and mutation process being replaced by the differential calculation [37]:
n i = a i , 1 + F · ( a i , 2 a i , 3 ) ,
where n i is the newly calculated solution, based on three randomly selected solutions a i , x . F is the differential weight that controls the influence of the calculated difference—in other words, it controls the distance from the original solution chosen, a i , 1 . DE can be thought of as a special case of GA with a different type of recombination. The mutation is not included, as the randomness of the search for the solution space is introduced by differentiation between a i , 2 and a i , 3 .
The MA is also an extension of the GA. Instead of changing the way the genes are recombined, the MA adds an additional step—local search. The main idea behind the MA is that the process GA uses for the search does not guarantee finding the optimal solution [38], and the solutions found can often be assumed to be near optimal [39]. Then, by using the domain-specific knowledge related to the problem that is being solved, an additional search is constructed, in order to find possible better solutions near the ones found by the GA. In this sense, the MA is a two stage process, in which the first stage consists of the classical GA application, while the second stage deals with the specific search of the local solution space surrounding the local solution [40]. In the presented research, this is performed by adjusting the parameters connected to those joints that had the highest influence on the total energy—information obtained from the analysis using the methods described in Section 2.2 The trajectory parameters, as given in Equations (2) and (3), are lowered with the idea of lowering the movement, speeds, and accelerations of that particular joint.
The GA functions by generating an initial population of candidate solutions—in the presented work, the paths are defined by their interpolation values. The fitness is then evaluated using the selected method for each individual solution, which starts the loop. While the loop can be exited in different ways, such as reaching the predetermined value or determining that there was no significant change in past iterations, this study utilizes a fixed number of iterations (called “generations”), iterating for 300 times across all of the elements in the population. If the number of generations has not been reached, the selection is performed. The selection, using roulette wheel selection, has a higher likelihood of selecting better-fit individuals (paths that use lower energy. Then, with different probabilities, either the mutation, crossover, or reproduction is performed, creating a new member of the population, which is inserted into the population—generating the new population set for the given generation. The difference between the GA and MA is illustrated in the Figure 2. The added step (indicated with the red background) allows for the second search stage to be performed. While given generally here, this usually describes a localized deterministic search adjusted to the search space. The need for this second point arises from the nature of the GA. Combining the solutions using crossover allows for the wide search of the space, and while GA converges to a solution, it does not guarantee that the best solution will be found, as the stochastic nature of the search can skip over solutions [35]. The benefit of the added second stage in the MA is that it is relatively quick (the simplest search would simply include observing the immediate surrounding values in the search space). Due to this secondary search having constant or linear complexity ( O ( 1 ) or O ( n ) ), it allows for a quick check of the local solution space (especially since it can be performed only for the best solution in the generation, if additional speed is necessary). Since the optimization of the GA is comparatively relatively long, adding the MA does not significantly lengthen the process, meaning that there is not a significant loss of performance even if the local search introduced by the MA does not determine a better solution  [35].

3. Results

3.1. Regression Model Results

The results of the models during the cross-validation process are given in Table 7.
The comparison of the regression models presented in Table 7 provides an overview of their predictive performance and the optimized hyperparameters obtained through the validation procedure. The evaluation was based on the mean coefficient of determination ( R 2 ¯ ), its standard deviation ( σ R 2 ¯ ), the mean absolute error ( M A E ¯ ), and the standard deviation of the absolute error ( σ M A E ¯ ). These metrics jointly quantify the average predictive accuracy, stability, and generalization capacity of each model across the validation folds. The MLP achieved the highest performance, with  R 2 ¯ = 0.999 and σ R 2 ¯ = 0.001 , indicating that the model consistently reproduced the reference values with negligible deviation. The corresponding M A E ¯ of 0.022 and σ M A E ¯ of 0.010 confirm the model’s low prediction error and robustness. The optimized hyperparameters consisted of two hidden layers of ten neurons each (HLS = (10, 10)), an identity activation function, and a constant learning rate ( α t y p e = constant) with an initial value of α i n i t . = 0.1 . The optimization was performed using the LBFGS solver, which is a quasi-Newton approach known for its high convergence efficiency on small data sets. The obtained configuration suggests that a relatively simple and linearly activated network was sufficient to fully model the target variable. The PAR exhibited nearly identical performance, also reaching R 2 ¯ = 0.999 and σ R 2 ¯ = 0.001 . Although the M A E ¯ of 0.044 was slightly higher than that of the MLP, the result still indicates a strong predictive capability. The optimal hyperparameters were C = 0.1 and ε = 0.1 , with both the fit intercept and ε -insensitive loss options enabled. The tolerance of the solver was set to 1 · 10 5 , ensuring stable convergence. This configuration indicates that the model maintained a balance between aggressive updates and regularization, allowing for high precision while preserving robustness to data variations. The SVR demonstrated equally high performance, with  R 2 ¯ = 0.999 and σ R 2 ¯ = 0.001 . The mean absolute error of 0.040 ( σ M A E ¯ = 0.006 ) reflects a comparable level of predictive precision to the previously discussed models. The optimal hyperparameters included C = 10 , a polynomial degree of 2, ε = 0.1 , kernel coefficient γ = scale, and a linear kernel. The selected configuration suggests that a linear mapping of the feature space was sufficient for achieving near-perfect prediction accuracy, further confirming the linearity of the underlying relationships in the data set. In contrast, the XGB model achieved a lower performance, with  R 2 ¯ = 0.928 and σ R 2 ¯ = 0.013 , indicating reduced consistency across folds. The  M A E ¯ of 8.050 and σ M A E ¯ of 1.663 demonstrate a considerably higher average error and variability. The optimal parameters were defined by a learning rate X L = 1 , one estimator ( X N = 1 ), tree depth X T = 1 , α = 0.1 , maximum tree depth of 6, 100 boosting rounds ( n = 100 ), L 1 = 0 , L 2 = 0.5 , and subsampling ratio of 0.5. This configuration represents a conservative setup, with strong regularization and shallow trees that likely limited the model’s expressive power. The combination of a high learning rate and a minimal number of estimators resulted in a model prone to underfitting, which explains its comparatively weaker performance. Overall, the MLP, PAR, and SVR models demonstrated near-ideal regression behavior, characterized by minimal error and near-complete variance explanation, while the XGB model underperformed due to its limited complexity. The observed results imply that the data set exhibits an approximately linear dependency structure, effectively captured by linear and shallow nonlinear models, while more complex ensemble approaches provided no additional benefit under the given parameter constraints.
To confirm the model performance on the validation data, the models which were determined to have best performance have been tested on the previously collected test data, completely separate from the data that the models were trained on. The goal of this is to determine the performance of the model obtained with a given method and hyperparameters, as given in Table 7. As such, Figure 3 demonstrates the performance of these models, as evaluated with R 2 , Figure 3a, and MAE, Figure 3b, metrics on the test set—evaluating the performance of the models as close as possible to what their performance should be on real data. The results from Figure 3 clearly indicate that the SVR, PAR, and MLP models reached near-perfect performance, each achieving an R 2 value of 1.00 with corresponding mean absolute errors of 0.05, 0.04, and 0.03, respectively. The results from Figure 3 suggest that these models were capable of fully reproducing the target values with numerical precision. In contrast, the XGB model achieved a lower R 2 value of 0.93 and a significantly higher MAE of 7.71, indicating reduced accuracy and a considerably larger prediction deviation. The difference in performance between the boosting model and the other approaches implies that the underlying dependency in the data set is predominantly linear, favoring regression methods based on linear mappings or shallow network architectures. The obtained comparison confirms the observations reported in Table 3, where the MLP, PAR, and SVR models exhibited nearly ideal regression behavior, while the XGB configuration remained limited by underfitting.
Comparing the scores to scores obtained by other researchers who developed Energy-use prediction models, as provided in Table 2, the scores obtained by the best performing models were comparable or better. The study using R 2 as the metric achieved a score of 0.97 [11], while the best performing models presented in this study achieve scores of 0.99. As for the error scores, the normalized MAE is in the similar range to the models evaluated by error metrics and obtained [3,13,14], but it must be noted that all of those models use historical time-series data, which are not used in the presented research, which uses momentary values of physical quantities related to the robot arm movement and position, allowing for a wider model application.

3.2. Feature Importance Results

The comparison of feature importance values obtained for the variable E is presented in Figure 4. Figure 4 provides a logarithmic-scale visualization of the relative importance of all the considered input features across multiple evaluation methods, including the Mean Decrease in Impurity ( I M D I ), permutation-based importance ( I p e r m ), and three correlation-based measures: Pearson’s ( ρ P ), Spearman’s ( ρ S ), and Kendall’s ( ρ K ) coefficients. Such a combined representation enables a comprehensive assessment of both model-driven and statistical dependencies between the predictors and the target variable.
The results indicate that the positional and angular variables—particularly x, y, and  ψ —exhibit consistently higher importance values across all evaluation metrics, suggesting that they contribute most significantly to the prediction of E. These features maintain high importance levels in both impurity-based and correlation-based analyses, implying that the target is primarily influenced by geometric and orientation-related parameters. The joint variables q 1 through q 6 also demonstrate non-negligible contributions, although with noticeable variation across methods, reflecting their indirect or secondary influence on the predicted output.
Among the correlation-based measures, Pearson’s and Spearman’s coefficients show the strongest alignment, indicating that the relationship between the features and the output is predominantly monotonic and approximately linear. Kendall’s ρ K values follow similar trends but remain slightly lower in magnitude, which is expected due to its stricter concordance-based formulation. The permutation importance values I p e r m largely confirm the same pattern observed in I M D I , emphasizing the robustness of the identified key predictors.
Overall, the presented comparison highlights that the predictive behavior of the models is primarily governed by translational and rotational components, while the joint-space variables contribute less directly. The consistency between multiple importance evaluation methods further reinforces the interpretability of the regression models and the reliability of the observed feature relevance structure.
What is important to note is that the joints, whose parameters show the highest influence on energy, are joints q 1 and q 2 (visible near the bottom of the Figure 4). This indicates that, by controlling the speeds of those individual joints by specifically lowering the parameters controlling their trajectories in joint space, the largest influence should be achievable. For this reason, the implementation of the MA algorithm that will be used in the presented research will aim to lower the parameters connected to those two joints.

3.3. Comparison of Evolutionary Algorithms

The results presented in Table 8 summarize the performance of the GA under different recombination strategies and parameter configurations. Three recombination types were examined: random, average, and differential. For each configuration, the probability of crossover P , mutation P ( M ) , and, in the case of differential recombination, the scaling factor F, were varied to observe their influence on optimization efficiency. The Improvement represents the mean percentage gain achieved across N = 100 independent runs, while σ denotes the corresponding standard deviation, indicating the consistency of results.
From the data, it can be seen that the random and average recombination approaches generally yielded higher improvements, exceeding 40% in several configurations, whereas the differential recombination achieved more moderate yet consistent results, with improvements in the range of 35–38%. The highest recorded improvement was 50.15% for the random recombination with P = 0.95 and P ( M ) = 0.05 , while the lowest was 33.01% for random recombination with lower mutation probability. The standard deviation values remain relatively low (0.20–0.27), suggesting that the algorithm maintained stable convergence across repeated runs. Overall, the results indicate that a higher mutation rate tends to improve the performance for random and average recombination, while the differential approach demonstrates robustness to parameter variations.
Based on these results (as shown in Table 8), the GA with randomized recombination at P = 0.95 and P ( M ) = 0.05 was selected as the basis of the MA algorithm. The results shown in Figure 5 illustrate the convergence dynamics of the MA when an informed local search strategy is employed. Each gray line corresponds to a single optimization path out of the 100 randomly initialized runs, while the red curve shows the mean performance across all runs. The overall trend demonstrates a steep initial reduction in fitness, occurring during the first 20–40 generations, which suggests that the algorithm quickly identifies promising regions of the search space. After this phase, the rate of improvement slows, indicating convergence toward local optima and a stabilization of the solution quality.
The presence of overlapping fitness trajectories in the early generations reflects the stochastic nature of the population-based search, while the convergence of lines in later generations implies the strong exploitation of high-performing individuals. The low variance observed after approximately 150 generations confirms that the informed local search effectively enhances the convergence stability, minimizing performance fluctuations between independent runs. These findings confirm that incorporating a local search within the evolutionary process improves the convergence efficiency while maintaining the robust exploration of the solution space.
Table 9 presents the optimization results of a few randomly selected paths, comparing the energy consumption that a linear path between the same two points in the space would take, against the energy consumption of the interpolated path whose parameters were obtained via the MA. The optimization process was tested 100 times, repeating the optimization between two randomly selected points in the space (using the same methodology as the random selection of paths for obtaining the training data set), with the average improvement being 58.16%, with a standard deviation of 0.22%.
Figure 6 shows the improvement between the best performing GA, as given in Table 8 and the MA, which applies the local search. Both algorithms were tested on 100 randomly generated runs. As the image clearly shows, the addition of the local search to the GA shows a clear improvement of approximately 8% on average. Considering the simplicity and low computational complexity of the local search, this improvement is definitely notable and justifies the usage of the MA over the simple GA for the given problem. Comparing the presented approach with the scores obtained by other researchers on similar problems (provided in Table 1), the results obtained by both the single-stage methodology and the two-stage MA are shown to be at least competitive if not better than the results obtained with the methodologies presented by other researchers in the literature review. The best score presented was a 40% improvement compared to the initial path, obtained by [2], using a combination of GA, PSO, and Bezier curves for path optimization. The scores obtained in this study were 50.15% for the GA and 58.16% for the MA, both of which are notably higher than the presented state-of-the-art score.

4. Conclusions

The presented work has shown the process in which an MA algorithm, consisting of an MLP network for the energy cost estimation of a robot path, is developed. The developed algorithm uses the GA with random recombination as the basis for the first stage and performs a local search based on feature importance determined from the data in the second stage.
The performed tests show that not only does the GA itself optimize the point-to-point paths well, but the usage of the MA can lower the energy use by 58% on average) when compared to a linear path between the same two points.
It is important to note that the specified technique has two main limitations. First, it is only applicable to point-to-point paths, where movement in between two points is free. Second, the fitness function based on the MLP model would probably require re-training for application on different IRs, but this has not been tested in the current work.
It must be noted that the different scenarios and environment changes could require adjustments to the proposed method. Further testing, for example in areas with obstacles, leveraging more advanced IRs with higher sensing capabilities, is planned as a part of future work.   

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/machines14010035/s1, Table S1: PROCESSED_real; Table S2: TEST.

Author Contributions

Conceptualization, S.B.Š. and N.A.; methodology, S.B.Š.; software, I.L.; validation, I.L., D.F. and N.A.; formal analysis, D.F. and N.A.; investigation, S.B.Š.; resources, D.F.; data curation, S.B.Š.; writing—original draft preparation, S.B.Š. and N.A.; writing—review and editing, I.L. and D.F.; visualization, I.L.; supervision, D.F.; project administration, I.L.; funding acquisition, D.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the following Projects: SPIN Projects IP.1.03.1020, IP.1.03.0158, and IP.1.1.03.0039; UNIRI projects UNIRI-IZ-25-220 and UNIRI-IZ-25-6; University North Project UNIN-TEH-25-1-8; the EU NextGeneration under the Juraj Dobrila University of Pula Institutional Research Project number IIP_010144 and IIP_010136; and the EC Digital Europe Programme EDIH Adria 101083838.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Acknowledgments

During the preparation of this manuscript/study, the authors used ChatGPT5 and Github Copilot GPT-5 mini for the purposes of generating and editing parts of the text and code generation. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial Neural Network
BNBatch Normalization
DEDifferential Evolution
GAGenetic Algorithm
GA+VNSGenetic Algorithm combined with Variable Neighborhood Search
IMDIImpurity-Based Importance (variant of MDI)
IRIndustrial Robot
IRMIndustrial Robotic Manipulator
LBFGSLimited-memory Broyden–Fletcher–Goldfarb–Shanno
LSTM       Long Short-Term Memory
MAMemetic Algorithm
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
MDIMean Decrease in Impurity
MLMachine Learning
MLPMultilayer Perceptron
PARPassive-Aggressive Regressor
PPMPlanar Parallel Manipulator
PSOParticle Swarm Optimization
R2Coefficient of Determination
RMSERoot Mean Square Error
SMBSerial Measurement Board
SVRSupport Vector Regression
SVMSupport Vector Machine
TCPTool Center Position
VNSVariable Neighborhood Search
XGBExtreme Gradient Boosting
σ Standard Deviation

Appendix A

Appendix A.1

The code below, given in RAPID programming language, was used during data collection in order to move the IR during measurement.
Listing A1. RAPID module used for generating random robot joint configurations and executing motion commands.
Machines 14 00035 i001aMachines 14 00035 i001b

Appendix A.2

The code below was used in parallel with the code above, in order to note the actual values as measured by the IR controller and save them to a file for later analysis and use.
Listing A2. RAPID module for real-time measurement and logging of joint and TCP data during execution.
Machines 14 00035 i002aMachines 14 00035 i002b

References

  1. International Federation of Robotics. World Robotics 2024—Press Conference Presentation; Technical Report; Presentation Slides; International Federation of Robotics (IFR): Frankfurt, Germany, 2024. [Google Scholar]
  2. Vysockỳ, A.; Papřok, R.; Šafařík, J.; Kot, T.; Bobovskỳ, Z.; Novák, P.; Snášel, V. Reduction in robotic arm energy consumption by particle swarm optimization. Appl. Sci. 2020, 10, 8241. [Google Scholar] [CrossRef]
  3. Garriz, A.; Domingo, C. Trajectory Optimization in Terms of Energy and Performance of an Industrial Robot in the Manufacturing Industry. Sensors 2022, 22, 7538. [Google Scholar] [CrossRef]
  4. Shrivastava, A.; Dalla, P. Failure control and energy optimization of multi-axes space manipulator through genetic algorithm approach. J. Braz. Soc. Mech. Sci. Eng. 2021, 43, 445. [Google Scholar] [CrossRef]
  5. Nonoyama, Y.; Matsuda, S.; Takahashi, R. Energy-Efficient Robot Configuration and Motion Planning Using Genetic Algorithm and Particle Swarm Optimization. Energies 2022, 15, 2074. [Google Scholar] [CrossRef]
  6. Luneckas, M.; Bacevičius, L.; Damaševičius, R. Hexapod Robot Gait Switching for Energy Consumption and Cost of Transport Management Using Heuristic Algorithms. Appl. Sci. 2021, 11, 1139. [Google Scholar] [CrossRef]
  7. Lu, Y.; Wang, Y.; Chen, X. Human–robot collaborative scheduling in energy-efficient welding shop. IEEE Trans. on Ind. Inf. 2023, 20, 936–971. [Google Scholar] [CrossRef]
  8. Baressi Šegota, S.; Horvat, T.; Grbac, I. Path optimization of robotic manipulators using differential evolution for improved energy efficiency. Int. J. Adv. Robot. Syst. 2020, 17, 1–12. [Google Scholar] [CrossRef]
  9. Jin, Y.; Olhofer, M.; Sendhoff, B. A framework for evolutionary optimization with approximate fitness functions. IEEE Trans. Evol. Comput. 2002, 6, 481–494. [Google Scholar] [CrossRef]
  10. Gadaleta, M.; Berselli, G.; Pellicciari, M.; Grassia, F. Extensive experimental investigation for the optimization of the energy consumption of a high payload industrial robot with open research dataset. Robot. Comput.-Integr. Manuf. 2021, 68, 102046. [Google Scholar] [CrossRef]
  11. Zhang, M.; Yan, W. A data-driven method for optimizing the energy consumption of industrial robots. Energy Rep. 2021, 7, 8747–8756. [Google Scholar] [CrossRef]
  12. Gao, Y.; Li, J.; Sun, B. Energy optimization of a planar parallel manipulator using long short-term memory network. IEEE Access 2020, 8, 175964–175972. [Google Scholar]
  13. Lin, X.; Zhang, K.; Luo, W. BN-LSTM-based energy consumption modeling approach for an industrial robot manipulator. Robot. Comput.-Integr. Manuf. 2024, 85, 102629. [Google Scholar] [CrossRef]
  14. Zhang, M.; Jiang, H.; Wang, X.; Chen, J. An integrated deep-learning-based approach for energy consumption prediction of machining systems. Sustainability 2023, 15, 5781. [Google Scholar] [CrossRef]
  15. Jaramillo-Morales, J.; López, C.; Suárez, J. Energy estimation for differential drive mobile robots on straight and rotational trajectories. Int. J. Adv. Rob. Sys. 2020, 17, 880–888. [Google Scholar] [CrossRef]
  16. Baressi Šegota, S.; Anđelić, N.; Šercer, M.; Meštrić, H. Dynamics modeling of industrial robotic manipulators: A machine learning approach based on synthetic data. Mathematics 2022, 10, 1174. [Google Scholar] [CrossRef]
  17. Cardona, M.; Serrano, F.E. Lagrange-Euler Approach. In Rehabilitation Robotics: Kinematics, Dynamics, and Control Techniques; Springer: Berlin/Heidelberg, Germany, 2025; pp. 77–89. [Google Scholar]
  18. Adam, M.S.; Nordin, R.; Abdullah, N.F.; Abu-Samah, A.; Amodu, O.A.; Alsharif, M.H. Optimizing disaster response through efficient path planning of mobile aerial base station with genetic algorithm. Drones 2024, 8, 272. [Google Scholar] [CrossRef]
  19. Dax, M.; Green, S.R.; Gair, J.; Gupte, N.; Pürrer, M.; Raymond, V.; Wildberger, J.; Macke, J.H.; Buonanno, A.; Schölkopf, B. Real-time inference for binary neutron star mergers using machine learning. Nature 2025, 639, 49–53. [Google Scholar] [CrossRef] [PubMed]
  20. ABB Robotics. RobotStudio Operating Manual; Operating Manual for ABB RobotStudio, Document ID 3HAC032104-001, Revision V; ABB Group: Zurich, Switzerland, 2017. [Google Scholar]
  21. ABB Robotics. Technical Reference Manual: RAPID Instructions, Functions and Data Types; Technical Reference for the RAPID Programming Language, Document ID 3HAC050917-001, Revision F, RobotWare 6.06; ABB Group: Västerås, Sweden, 2017. [Google Scholar]
  22. Ostovar, A.; Davari, D.D.; Dzikuć, M. Determinants of design with multilayer perceptron neural networks: A comparison with logistic regression. Sustainability 2025, 17, 2611. [Google Scholar] [CrossRef]
  23. Mekala, S.; Dasari, K.; Katta, D. DNS DDoS Amplification Attack Detection Using Multi-Layer Perceptron Classification Algorithm. In Proceedings of the 2024 IEEE 3rd World Conference on Applied Intelligence and Computing (AIC), Gwalior, India, 27–28 July 2024; pp. 1355–1360. [Google Scholar]
  24. Ren, X.; Wang, N.; Pan, J.; Bu, L. Combining style generative adversarial networks with particle swarm optimisation-support vector regression to design affective social robot for public health intervention. J. Eng. Des. 2025, 36, 160–190. [Google Scholar] [CrossRef]
  25. Wang, Y.; Hong, M.; Chai, H.; Zhang, Y.; Wang, G.; Wu, C.; Guo, M. Research on Foot Contact State Detection Technology of Wheel-Legged Robot. Sensors 2024, 24, 3956. [Google Scholar] [CrossRef]
  26. Alqarni, M.; Ashour, A.A.; Shafie, A.; Alqarni, A.; Felemban, M.F.; Shukr, B.S.; Alzubaidi, M.A.; Algahtani, F.S. Intelligence computational analysis of letrozole solubility in supercritical solvent via machine learning models. Sci. Rep. 2024, 14, 21677. [Google Scholar] [CrossRef]
  27. Rahaman, M.; Cajes, N.; Gupta, B.B.; Nedjah, N.; Chui, K.T. Real-Time Machine Learning Techniques in Robotics for Detecting Phishing Attacks. In Proceedings of the 2025 IEEE International Conference on Real-time Computing and Robotics (RCAR), Toyama, Japan, 1–6 June 2025; pp. 1124–1129. [Google Scholar]
  28. Khajezadeh, H.; Pour, M.S.; Manthouri, M. Effective Detection and Classification of Robot Execution Failures using Extreme Gradient Boosting (XGboost) Algorithm. In Proceedings of the IEEE 2024 10th International Conference on Control, Instrumentation and Automation (ICCIA), Kashan, Iran, 5–7 November 2024; pp. 1–6. [Google Scholar]
  29. Kaushal, A.; Sharma, A.K.; Gupta, K. Decentralised coordination in swarm robots through xgboost-enhanced colour light communication. Arab. J. Sci. Eng. 2024, 49, 16253–16269. [Google Scholar] [CrossRef]
  30. He, Y.X.; Lyu, S.H.; Jiang, Y. Interpreting deep forest through feature contribution and mdi feature importance. ACM Trans. Knowl. Discov. Data 2024, 20, 15. [Google Scholar] [CrossRef]
  31. Geeitha, S.; Ravishankar, K.; Cho, J.; Easwaramoorthy, S.V. Integrating cat boost algorithm with triangulating feature importance to predict survival outcome in recurrent cervical cancer. Sci. Rep. 2024, 14, 19828. [Google Scholar] [CrossRef]
  32. O’brien, P.C. Comparing two samples: Extensions of the t, rank-sum, and log-rank tests. J. Am. Stat. Assoc. 1988, 83, 52–61. [Google Scholar]
  33. Ragano, A.; Martinez, H.B.; Hines, A. Beyond Correlation: Evaluating Multimedia Quality Models with the Constrained Concordance Index. IEEE Trans. Multimed. 2025, 27, 5604–5616. [Google Scholar] [CrossRef]
  34. Alhijawi, B.; Awajan, A. Genetic algorithms: Theory, genetic operators, solutions, and applications. Evol. Intell. 2024, 17, 1245–1256. [Google Scholar] [CrossRef]
  35. Baressi Šegota, S. Determining the Energy-Optimal Path of Six-Axis Industrial Robotic Manipulators Using Machine Learning and Memetic Algorithms. Ph.D. Thesis, Faculty of Engineering, University of Rijeka, Rijeka, Croatia, 2025. [Google Scholar]
  36. Londe, M.A.; Pessoa, L.S.; Andrade, C.E.; Resende, M.G. Biased random-key genetic algorithms: A review. Eur. J. Oper. Res. 2025, 321, 1–22. [Google Scholar] [CrossRef]
  37. Pesaramelli, R.S.; Sujatha, B. Principle correlated feature extraction using differential evolution for improved classification. Aip Conf. Proc. 2024, 2919, 070001. [Google Scholar]
  38. Alexakis, K.; Benekis, V.; Kokkinakos, P.; Askounis, D. Genetic algorithm-based multi-objective optimisation for energy-efficient building retrofitting: A systematic review. Energy Build. 2025, 328, 115216. [Google Scholar] [CrossRef]
  39. Mahmud, T.; Datta, N.; Chakma, R.; Das, U.K.; Aziz, M.T.; Islam, M.; Salimullah, A.H.M.; Hossain, M.S.; Andersson, K. An approach for crop prediction in agriculture: Integrating genetic algorithms and machine learning. IEEE Access 2024, 12, 173583–173598. [Google Scholar] [CrossRef]
  40. Ab Wahab, M.N.; Nazir, A.; Khalil, A.; Ho, W.J.; Akbar, M.F.; Noor, M.H.M.; Mohamed, A.S.A. Improved genetic algorithm for mobile robot path planning in static environments. Expert Syst. Appl. 2024, 249, 123762. [Google Scholar] [CrossRef]
Figure 1. Diagram of the data collection process.
Figure 1. Diagram of the data collection process.
Machines 14 00035 g001
Figure 2. Scheme of the memetic algorithm used in the presented research.
Figure 2. Scheme of the memetic algorithm used in the presented research.
Machines 14 00035 g002
Figure 3. Comparison of the best-performing regression models on the separate test data set.
Figure 3. Comparison of the best-performing regression models on the separate test data set.
Machines 14 00035 g003
Figure 4. The stacked feature importance of each individual variable.
Figure 4. The stacked feature importance of each individual variable.
Machines 14 00035 g004
Figure 5. Average improvement across generations for the memetic algorithm (MA) with an informed local search, calculated over 100 randomly selected evolutionary paths. The gray lines represent individual fitness trajectories, while the red line denotes the average fitness value across generations. A rapid decrease in fitness is observed within the initial iterations, followed by gradual convergence toward a stable minimum. This behavior indicates that the algorithm effectively exploits the search space early on, with subsequent refinements leading to diminishing returns in improvement.
Figure 5. Average improvement across generations for the memetic algorithm (MA) with an informed local search, calculated over 100 randomly selected evolutionary paths. The gray lines represent individual fitness trajectories, while the red line denotes the average fitness value across generations. A rapid decrease in fitness is observed within the initial iterations, followed by gradual convergence toward a stable minimum. This behavior indicates that the algorithm effectively exploits the search space early on, with subsequent refinements leading to diminishing returns in improvement.
Machines 14 00035 g005
Figure 6. Comparison of different stages based on improvement metrics, with the first stage representing the average improvement of 100 test runs for GA with random recombination, and the second stage representing the average improvement of 100 test runs for MA based on the same GA, but extended with a localized search.
Figure 6. Comparison of different stages based on improvement metrics, with the first stage representing the average improvement of 100 test runs for GA with random recombination, and the second stage representing the average improvement of 100 test runs for MA based on the same GA, but extended with a localized search.
Machines 14 00035 g006
Table 1. The best results from the reviewed research focused on application of evolutionary computing algorithms on robot energy efficiency improvement. Improvement between the unoptimized and optimized paths expressed as a percentage and rounded to the closest full value.
Table 1. The best results from the reviewed research focused on application of evolutionary computing algorithms on robot energy efficiency improvement. Improvement between the unoptimized and optimized paths expressed as a percentage and rounded to the closest full value.
ReferenceApproachImprovement [%]
Vysocky et al. [2]GA, PSO, Bezier curves40
Garriz and Domingo [3]Kalman20
Shrivastava and Dalla [4]GA38
Nonoyama et al. [5]K-ROSET, GA, PSO18
Luneckas et al. [6]Red Fox21
Lu et al. [7]MA − GA + VNS10
Table 2. The best results from the reviewed research focused on the energy use modeling of robots. RMSE—root mean square error, MAPE—mean absolute percentage error.
Table 2. The best results from the reviewed research focused on the energy use modeling of robots. RMSE—root mean square error, MAPE—mean absolute percentage error.
ReferenceApproachScore
Zhang & Yan [11]ANNR-score = 0.97
Gao et al. [12]LSTMMAPE = 2.5%
Lin et al. [13]BN-LSTMRMSE = 3.67
Jiang et al. [14]LSTMMAPE = 4.24%
Jaramillo et al. [15]NumericAccuracy = 82.15%
Table 3. Hyperparameter values tested in the GS process for MLP.
Table 3. Hyperparameter values tested in the GS process for MLP.
HyperparameterValues
Hidden layer sizes(10), (50), (100), (10, 10), (50, 50), (100, 100),
(10, 10, 10), (50, 50, 50), (100, 100, 100),
(10, 10, 10, 10), (50, 50, 50, 50), (100, 100, 100, 100)
Activation functionidentity, logistic, tanh, relu
Solverlbfgs, adam
Alpha0.0001, 0.001, 0.01, 0.1
Learning rate typeconstant, adaptive, invscaling
Initial learning rate0.01, 0.1, 1
Table 4. Hyperparameters and their values for SVM regressor.
Table 4. Hyperparameters and their values for SVM regressor.
HyperparameterValues
Kernellinear, poly, rbf, sigmoid
Degree2, 3, 4
γ scale, auto
C0.1, 0.5, 1, 10
ε 0.1, 0.2, 0.5, 1
Table 5. Hyperparameters and their values for PAR.
Table 5. Hyperparameters and their values for PAR.
HyperparameterValues
C0.1, 0.5, 1, 10
fit_interceptTrue, False
Tolerance 10 3 , 10 4 , 10 5
Lossepsilon_insensitive, squared_epsilon_insensitive
ε 0.1, 0.2, 0.5, 1
Table 6. Hyperparameters and their values for XGB.
Table 6. Hyperparameters and their values for XGB.
HyperparameterValues
n_estimators10, 50, 100
max_depth3, 4, 5, 6, 7
learning_rate0.01, 0.1, 1
subsample0.5, 0.75, 1
colsample_bytree0.5, 0.75, 1
colsample_bylevel0.5, 0.75, 1
colsample_bynode0.5, 0.75, 1
reg_alpha0, 0.1, 0.5, 1
reg_lambda0, 0.1, 0.5, 1
Table 7. Comparison of model performance metrics and optimized hyperparameters for each regression method. The table reports the mean coefficient of determination R 2 ¯ , its standard deviation σ R 2 ¯ , the mean absolute error MAE ¯ , and its standard deviation σ MAE ¯ across validation folds. For each method, the optimal hyperparameters are shown: MLP (multilayer perceptron)—HLS: hidden layer sizes; ϕ : activation function; α type : learning rate type; α init . : initial learning rate; Solver: optimization solver. PAR (passive-aggressive regressor)—C: regularization parameter; ε : tolerance margin; F.I.: fit intercept flag; ε ins : ε -insensitive loss flag; Tol.: solver tolerance. SVR (support vector regressor)—C: penalty parameter; Degree: polynomial kernel degree; ε : epsilon-tube within which no penalty is applied; γ : kernel coefficient; Kernel: kernel type. XGB (extreme gradient boosting)— X L : learning rate; X N : number of estimators; X T : tree depth; α : learning rate; max(d): maximum tree depth; n: number of boosting rounds; L 1 : L1 regularization weight; L 2 : L2 regularization weight; Subs.: subsample ratio.
Table 7. Comparison of model performance metrics and optimized hyperparameters for each regression method. The table reports the mean coefficient of determination R 2 ¯ , its standard deviation σ R 2 ¯ , the mean absolute error MAE ¯ , and its standard deviation σ MAE ¯ across validation folds. For each method, the optimal hyperparameters are shown: MLP (multilayer perceptron)—HLS: hidden layer sizes; ϕ : activation function; α type : learning rate type; α init . : initial learning rate; Solver: optimization solver. PAR (passive-aggressive regressor)—C: regularization parameter; ε : tolerance margin; F.I.: fit intercept flag; ε ins : ε -insensitive loss flag; Tol.: solver tolerance. SVR (support vector regressor)—C: penalty parameter; Degree: polynomial kernel degree; ε : epsilon-tube within which no penalty is applied; γ : kernel coefficient; Kernel: kernel type. XGB (extreme gradient boosting)— X L : learning rate; X N : number of estimators; X T : tree depth; α : learning rate; max(d): maximum tree depth; n: number of boosting rounds; L 1 : L1 regularization weight; L 2 : L2 regularization weight; Subs.: subsample ratio.
MethodOptimized Hyperparameters R 2 ¯ σ R 2 ¯ MAE ¯ σ MAE ¯
MLPHLS ϕ α t y p e α i n i t . Solver0.9990.0010.0220.010
(10, 10)identityconst.0.1LBFGS
PARC ε F.I. ε i n s Tol.0.9990.0010.0440.011
0.10.1TrueTrue 1 · 10 5
SVRCDegree ε γ Kernel0.9990.0010.0400.006
1020.1scalelinear
XGB X L X N X T α max(d)n L 1 L 2 Subs.0.9280.0138.0501.663
1110.1610000.50.5
Table 8. The scores of the GA algorithms for given parameters. Recombination—the type of recombination used (random, average or differential). P —the probability of recombination happening for a given iteration over the population, P ( M ) —the probability of a mutation happening for a given iteration over the population, Improvement—the average improvement across N = 100 runs, σ —standard deviation.
Table 8. The scores of the GA algorithms for given parameters. Recombination—the type of recombination used (random, average or differential). P —the probability of recombination happening for a given iteration over the population, P ( M ) —the probability of a mutation happening for a given iteration over the population, Improvement—the average improvement across N = 100 runs, σ —standard deviation.
RecombinationPP(M)FImprovement [%] σ
Random0.950.05-50.150.23
Random0.950.01-33.610.24
Random0.900.05-38.030.27
Random0.900.01-33.010.22
Average0.950.05-44.420.22
Average0.950.01-33.190.23
Average0.900.05-48.850.22
Average0.900.01-33.780.20
Differential0.95-0.0537.840.27
Differential0.95-0.1536.780.26
Differential0.90-0.0535.910.25
Differential0.90-0.1536.180.26
Table 9. Comparison of few randomly selected paths, between the energy use of a linear point-to-point movement and the optimized path. Gene—the path parameters obtained via optimization, q S —starting positions in joint space, q E —ending positions in joint space, E L —energy of the linear path, E optimized —energy use of the optimized path, Δ —improvement with optimization.
Table 9. Comparison of few randomly selected paths, between the energy use of a linear point-to-point movement and the optimized path. Gene—the path parameters obtained via optimization, q S —starting positions in joint space, q E —ending positions in joint space, E L —energy of the linear path, E optimized —energy use of the optimized path, Δ —improvement with optimization.
Gene q S q E E L E optimized Δ [%]
0.00, 0.62, −0.12, −0.12, −1.18, 0.00,−14.32120.3275.2124.5267.39
−0.00, −0.10, −0.08, 0.08, 0.00, −0.14,−22.92131.78
9.45, 1.60, 5.85, 1.25, 7.85, −4.00,5.73123.19
3.30, 7.00, −4.70, −9.20, −3.85, −5.05,40.11140.37
8.15, 1.30, 7.60, 3.65, −7.50, −7.65,48.70114.59
1.00, 6.75, 4.30, 7.45, −5.55, 6.40−54.43166.16
1.12, 1.42, 0.04, 0.74, −0.74, −0.08,28.65166.1637.5719.4448.25
−0.16, 0.65, −0.12, −1.06, −0.02, 0.13,−28.65117.46
−1.25, 8.75, 5.20, 1.40, 2.35, 5.25,37.24117.46
−9.40, 0.25, −8.30, 0.25, −2.95, 3.20,20.05163.29
3.05, −5.80, 1.35, −4.35, −6.25, 3.60,51.57166.16
8.25, −7.40, 0.00, −8.95, 4.90, −1.8537.24114.59
−0.24, −0.00, 0.98, 0.00, −1.20, 0.43,22.91114.5992.2025.1672.71
0.03, −0.00, −0.85, 1.15, −1.02, 0.38,−40.11120.32
2.20, 3.20, 6.60, 2.00, −6.45, 5.95,22.92137.51
−3.25, −3.70, 3.90, −1.05, −1.35, −0.05,−20.05126.05
4.30, 8.05, 2.95, 1.05, −3.55, −5.20,−25.78128.92
−9.70, −2.15, −1.25, 8.25, −1.70, 7.45−20.05114.59
−0.01, 0.94, 0.00, 0.00, −0.08, 0.15,−42.97143.2438.7618.0053.56
0.01, 0.00, 0.00, 0.13, 0.18, 0.00,28.65146.10
6.70, −8.45, −1.30, 0.20, −4.10, −8.00,31.51154.69
4.50, 0.35, −8.25, 7.00, −8.50, 8.65,51.57117.46
1.85, 4.60, −1.90, 4.45, −3.55, 7.45,−40.11123.19
6.15, 5.95, −6.50, −7.20, −9.70, 3.9022.92143.24
−0.00, −0.02, 0.00, −0.97, 0.10, −0.19,−40.11143.2438.3929.2623.78
0.00, 1.00, −0.01, 0.08, 0.00, 0.00,−22.92154.69
−7.95, 5.90, 4.70, −9.25, −7.45, −7.05,51.57140.37
8.80, −8.80, −1.35, 7.15, 0.25, 1.45,8.59137.51
−4.40, −0.40, −0.15, 5.60, 4.65, −3.55,25.78134.65
−3.15, −8.65, −9.05, 6.00, −1.55, −1.7020.06131.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Baressi Šegota, S.; Frank, D.; Lorencin, I.; Anđelić, N. Memetic Algorithm for Energy Optimization in Point-to-Point Robotized Operations. Machines 2026, 14, 35. https://doi.org/10.3390/machines14010035

AMA Style

Baressi Šegota S, Frank D, Lorencin I, Anđelić N. Memetic Algorithm for Energy Optimization in Point-to-Point Robotized Operations. Machines. 2026; 14(1):35. https://doi.org/10.3390/machines14010035

Chicago/Turabian Style

Baressi Šegota, Sandi, Domagoj Frank, Ivan Lorencin, and Nikola Anđelić. 2026. "Memetic Algorithm for Energy Optimization in Point-to-Point Robotized Operations" Machines 14, no. 1: 35. https://doi.org/10.3390/machines14010035

APA Style

Baressi Šegota, S., Frank, D., Lorencin, I., & Anđelić, N. (2026). Memetic Algorithm for Energy Optimization in Point-to-Point Robotized Operations. Machines, 14(1), 35. https://doi.org/10.3390/machines14010035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop