1. Introduction
The control problem of artillery stabilization systems has long been a research focus. Early stabilization systems primarily relied on mechanical stabilizers. Traditional control strategies such as PID controllers have been widely applied in artillery stabilization systems. However, with the continuous advancement of artillery technology, control strategies for artillery stabilization have evolved to meet increasingly complex system demands. Due to the complexity of modern artillery systems, particularly those driven by electro-hydraulic actuators, control performance is influenced not only by mechanical design but also by nonlinear factors and external disturbances [
1]. These factors significantly impact the stability accuracy of artillery in real combat environments. Moreover, the presence of uncertainties such as mechanical collisions, structural deformations, and friction forces that are difficult to model accurately together with mechanical wear over time make PID parameter tuning more challenging [
2]. Therefore, determining appropriate control parameters that can adapt to complex and dynamically changing environments remains an urgent problem to be addressed in artillery stabilization systems.
Various control approaches have been proposed to address the challenges posed by strong nonlinearities and time-varying parameters in artillery stabilization systems, including neural network control, adaptive control, and fuzzy inference control [
3,
4,
5,
6,
7]. Wang et al. [
8] introduced a variable-structure wavelet neural network optimized using an adaptive differential evolution algorithm, improving both system accuracy and response speed. Ma et al. [
9] proposed an adaptive robust feedback control strategy and verified its effectiveness through co-simulation. Rafiei et al. [
10] incorporated fuzzy logic into a radial basis function (RBF) neural network for online optimization of control parameters. These studies primarily focus on model-based online optimization of control parameters to ensure basic accuracy under complex operating conditions.
Online tuning strategies have proven effective in artillery control systems, but often rely on real-time computation and high-fidelity models, which can increase implementation costs and hardware complexity. This challenge has led to the exploration of offline parameter tuning approaches [
11,
12], which utilize precalibrated optimization to minimize computational burdens during operation. Offline optimization has been widely explored in areas such as robotics and power systems, where it has shown the potential to enhance system performance and efficiency. To improve control accuracy while tuning parameters in a model-free environment, some researchers have turned to intelligent optimization algorithms for offline PID parameter optimization. For instance, Kong et al. [
13] proposed an improved Dung Beetle Optimizer (DBO) for self-coupling PID parameter tuning. Rodrigues et al. [
14] developed an enhanced Particle Swarm Optimization (PSO) algorithm to optimize PID parameters in automatic voltage regulators, while Azeez et al. [
15] introduced an Artificial Bee Colony (ABC) algorithm with an adaptive learning strategy for optimizing control parameters in robotic arms.
The Manta Ray Foraging Optimization (MRFO) algorithm [
16], first proposed in 2020, is an innovative bio-inspired swarm intelligence algorithm. This algorithm draws inspiration from the unique foraging behavior of manta rays, which utilize chain foraging, spiral foraging, and somersault foraging strategies to enhance their search capabilities. The manta ray’s flattened pectoral fins and broad dorsal fins offer exceptional maneuverability, enabling rapid directional changes and adjustments in posture. These foraging behaviors are effectively emulated by the MRFO algorithm, which is distinguished by its simplicity, minimal parameter requirements, and robust abilities in both global and local search. It has been widely applied to solve various optimization problems, including power systems, mechanical design, image segmentation, and path planning [
17,
18,
19,
20]. For instance, Ma et al. [
21] proposed a two-strategy enhanced MRFO for image segmentation, demonstrating the algorithm’s strong adaptability and competitiveness in visual data analysis. Similarly, Adamu et al. [
22] employed MRFO for hyperparameter optimization in skin cancer classification, revealing its potential in complex biomedical signal processing tasks. These applications reflect the potential and applicability of the MRFO algorithm in addressing diverse optimization challenges across multiple fields. Despite its advantages, the standard MRFO algorithm faces challenges, particularly in high-dimensional function optimization problems, where it tends to converge to local optima and exhibits slow convergence rates. These limitations hinder the algorithm’s performance in complex optimization tasks such as parameter tuning in systems with intricate dynamics.
To further enhance the performance of the artillery stabilization control system in a model-free setting, this paper proposes an Improved Manta Ray Foraging Optimization (IMRFO) algorithm that integrates multiple optimization strategies for parameter tuning. First, circle mapping, Lévy flight, and the sigmoid function are introduced to enhance the global search capability of the traditional MRFO algorithm. Then, the effectiveness of the proposed optimization method is validated using five benchmark test functions. Finally, the algorithm’s effectiveness and feasibility are demonstrated through co-simulation and physical system verification.
This paper makes three significant contributions. First, an Improved Manta Ray Foraging Optimization (IMRFO) algorithm incorporating circle chaotic mapping, Lévy flight, and a sigmoid function-based dynamic adjustment strategy is proposed to enhance the global search capability and convergence speed of the traditional MRFO algorithm. Second, a PID control parameter optimization method based on the IMRFO algorithm is designed for artillery stabilization systems. This method effectively addresses the challenges posed by strong nonlinearities and time-varying parameters, significantly improving the system’s stability and response performance. Third, the effectiveness and feasibility of the proposed method are validated through simulations and physical system experiments, demonstrating its potential as a reliable and efficient solution for complex control systems.
3. Improved MRFO Algorithm
3.1. Circle Chaotic Mapping for Population Initialization
In artillery stabilization systems, the complexity of simulations and high time cost of physical experiments often limit the feasible population size in optimization algorithms. Under such constraints, the standard MRFO algorithm’s reliance on random initialization can lead to insufficient population diversity, hindering comprehensive exploration of the solution space. To address this, we introduce circle chaotic mapping for population initialization. This method generates initial solutions with better uniformity and ergodicity compared to random initialization, thereby enhancing the algorithm’s ability to explore the search space comprehensively and avoid premature convergence. By improving the diversity of the initial population, the chaotic mapping strategy effectively mitigates the limitations of standard MRFO during the complex high-dimensional optimization tasks inherent in artillery stabilization systems.
Chaos mapping is a method for generating random behavior based on deterministic systems. It is capable of producing sequences with ergodicity and non-repetition within a finite range [
23]. In this paper, circle chaos mapping is used for population initialization. Circle chaos mapping is a chaotic mapping method based on trigonometric functions and modular arithmetic, and its mathematical expression is as follows:
where the value range of the circle chaos mapping sequence is [0, 1] and where
a and
b are control parameters. To ensure the randomness and uniformity of the chaos mapping,
a and
b are generally set to values around 0.5 and 0.2, respectively. The steps for population initialization using circle chaotic mapping are as follows:
- Step 1:
Set the population size N and dimension D, initializing the first individual in each dimension. The initial values of i and j are 1.
- Step 2:
Initialize the control parameters as follows: , .
- Step 3:
With
, generate
according to Equation (
8).
- Step 4:
If i is greater than N, proceed to Step 5; otherwise, return to Step 3.
- Step 5:
With
, check whether
j is greater than
D. If no, then return to Step 2 and set
; if yes, then the initialization is complete and the initial population positions are output according to Equation (
9).
3.2. Sigmoid Function-Based Strategy Selection Factor
In the MRFO algorithm, each iteration alternates between chain and spiral foraging based on a strategy selection factor , which is typically fixed in standard implementations. However, this approach may not fully address the complexities of artillery stabilization systems, where the dynamics are highly nonlinear and influenced by numerous factors such as barrel vibrations, projectile characteristics, and environmental conditions. In such systems, an adaptive strategy is crucial; early iterations should emphasize global exploration in order to thoroughly search the solution space, while later iterations should focus on local exploitation in order to fine-tune the solution near the optimal region. To balance global search and local exploitation, this paper proposes a strategy selection factor based on the sigmoid function, allowing the algorithm to adaptively adjust in different phases.
The sigmoid function is a commonly used nonlinear function, and its mathematical expression is as follows:
where
x is the input variable,
k is the parameter that controls the slope of the function, and
is the center point of the function. In the improved MRFO algorithm, the sigmoid function is used to adjust the strategy selection factor. Its mathematical expression is
where
is the strategy selection factor for the t-th iteration,
and
are the minimum and maximum values of the strategy selection factor, respectively, and
T is the maximum number of iterations. By setting
,
, the algorithm tends to favor spiral foraging during the early iterations, which helps it to quickly explore the solution space and avoid becoming trapped in local optima. In the later stages of the iteration, the focus shifts to chain foraging, which enables fine-tuning around the optimal solution and improves convergence accuracy.
3.3. Lévy Flight-Integrated Somersault Foraging
The MRFO algorithm also employs somersault foraging, in which individuals oscillate around the current optimal solution to enhance local exploration. However, using a fixed oscillation range can result in premature convergence to local optima. To address this limitation, the proposed algorithm incorporates Lévy flight step sizes during the somersault phase, introducing random perturbations that expand the search range and improve the algorithm’s global exploration capability. This enhancement is particularly beneficial for optimizing complex systems such as artillery stabilization, where the search space is vast and traditional optimization methods may struggle to escape local minima. By integrating Lévy flight, the ability of the MRFO algorithm to explore diverse solutions is enhanced, leading to more robust and effective optimization outcomes in challenging engineering applications.
Lévy flight is a stochastic walk strategy based on a heavy-tailed distribution, where most step sizes are short but long steps occasionally occur with a small probability [
24]. This characteristic allows Lévy flight to balance local exploitation and global exploration. Its mathematical expression is
where
is the Gamma function,
is the Lévy flight step length, and
is the Lévy exponent, which determines the probability of large step lengths. Usually,
is set and both
and
follow a normal distribution such that
To enhance the search efficiency of the somersault foraging strategy, the Lévy flight step length is introduced into the position update process and Equation (
7) is updated as follows:
where
k is the Lévy flight step length control factor, which helps to avoid excessive influence of short step lengths on local fine-tuning.
Figure 4 shows the position distribution of three individuals after 200 somersaults.
Figure 4 shows the somersault foraging strategy in the standard MRFO algorithm, where the individuals are concentrated within the rolling region after somersaulting. In contrast, using the improved somersault foraging strategy with the random step length characteristic of Lévy flight expands the search space. Most individuals remain concentrated in the rolling region near the current optimal solution for fine-grained search, while a small portion of individuals jump out of the current region through long steps to explore other solution spaces. This approach retains the efficiency of local exploitation while enhancing global exploration capability.
Our proposed improved MRFO algorithm is subsequently referred to as the IMRFO algorithm. The pseudocode for the IMRFO algorithm is shown in Algorithm 1.
Algorithm 1 IMRFO Algorithm |
- 1:
Initialization: Population size N, number of iterations T, upper bound , lower bound - 2:
Generate initial population positions based on Equation ( 8), calculate fitness values of all individuals, and find the best solution - 3:
repeat - 4:
- 5:
if then - 6:
Spiral foraging - 7:
if then - 8:
Update positions of all individuals based on Equation ( 5) - 9:
else - 10:
Update positions of all individuals based on Equation ( 3) - 11:
end if - 12:
else - 13:
Chain foraging - 14:
Update positions of all individuals based on Equation ( 1) - 15:
end if - 16:
Update the best fitness value - 17:
for N do - 18:
somersault foraging - 19:
Update positions of all individuals based on Equation ( 16) - 20:
Update the best fitness value again - 21:
end for - 22:
until is no longer true - Output:
Best fitness value and position of the optimal individual
|
3.4. Algorithm Validation
To evaluate the performance of the proposed algorithm, five benchmark functions are used for assessment;
and
are unimodal test functions, while
,
, and
are multimodal test functions. The unimodal test functions are primarily used to assess the convergence speed of the algorithm, whereas the multimodal test functions are chosen based on those areas where the standard IMRFO algorithm exhibits relatively weaker performance. This selection aims to further validate the proposed algorithm’s global optimization capability in complex problems. The benchmark functions are shown in
Table 1.
We evaluated the IMRFO algorithm against five commonly used optimization algorithms, including the Manta Ray Foraging Optimization (MRFO) algorithm [
25], Particle Swarm Optimization (PSO) algorithm [
26], Grey Wolf Optimizer (GWO) algorithm [
27], Whale Optimization Algorithm (WOA) [
28], Ant Lion Optimizer (ALO) algorithm [
29], and Covariance Matrix Adaptation Evolution Strategy (CMA-ES) [
30]. For all algorithms, the population size was set to 100 with 2000 iterations, and each algorithm was run ten times.
The parameter settings for the comparison algorithms are shown in
Table 2. In Particle Swarm Optimization (PSO), the inertia weight was decreased linearly from
to
, gradually shifting the search focus from global to local exploration. The cognitive and social coefficients were both set to
, ensuring a balanced influence between individual experience and social sharing. For the Grey Wolf Optimizer (GWO) and Whale Optimization Algorithm (WOA), the convergence control parameter
a decreased linearly from 2 to 0, allowing for a smooth transition between the exploration and exploitation phases. In the Ant Lion Optimizer (ALO) algorithm, the random walk boundaries were adaptively adjusted based on iteration count and ant lion selection was performed via roulette wheel selection to maintain diversity. In the Covariance Matrix Adaptation Evolution Strategy (CMA-ES), the initial step size
was set to one-third of the search range to provide a balanced initial exploration. The parent number
was set to half of the population size to ensure sufficient diversity. The step control parameter
and covariance matrix update rate
were computed based on the problem dimension and effective parent number to maintain stability and efficient adaptation of the covariance matrix.
The optimal and average values of each algorithm on the benchmark functions are presented in
Table 3, while the convergence curves are shown in
Figure 5. It can be observed that the convergence speed of IMRFO is slightly slower than that of MRFO for the unimodal functions
and
, which is due to the updates in the somersault model. However, both algorithms quickly converge to the optimal solution, and the optimal solution and the mean optimal solution are significantly better than those of other algorithms. For multimodal function
, IMRFO outperforms MRFO in both convergence speed and accuracy, indicating that the improves the strategy selection factor enhances the global optimization capability in the early iterations and focuses more on local exploration in the later stages. For multimodal function
, IMRFO achieves the best optimal solution and mean value, with a mean that is significantly better than the other algorithms. This suggests that the improved somersault foraging mechanism can better escape local optima and enhance the stability of the algorithm. For multimodal function
, the MRFO algorithm was only able to escape local optima in two out of the ten optimization runs, while the proposed IMRFO algorithm was able to find the optimal solution multiple times, further proving the global optimization capability of the improved somersault foraging strategy.
The Wilcoxon Signed-Rank Test (WSRT) is used to assess statistical differences between two optimizers. The results for the ten runs across the various benchmark functions are presented in
Table 4 and
Table 5. The
p-value indicates statistical significance, with values below 0.05 showing a significant difference; “W” represents the Wilcoxon test statistic, calculated as the smaller of the sums of ranks for positive and negative differences between paired observations; the “Result” column shows “+” for better performance by the first algorithm, “−” for the second, and “=” for no significant difference. The results indicate that IMRFO outperforms MRFO with significant differences (
p-value < 0.05) in most cases, particularly for multimodal functions
and
. IMRFO also shows better performance than the other algorithms on most functions. However, for function
, IMRFO shows no significant advantage over CMA-ES or PSO. Overall, the test results demonstrate that IMRFO generally outperforms the other algorithms, confirming its effectiveness.
4. Parameter Optimization of Artillery Stabilization Control System Based on IMRFO
4.1. Development of Co-Simulation Platform
To verify the effectiveness of the IMRFO algorithm in the PID controller of an artillery stabilization control system, a system model was established using a complex system modeling and simulation platform. Based on the system structural parameters shown in
Table 6, the dynamic model of the artillery stabilization control system was built in a multibody dynamics simulation platform. The control system was designed in a control simulation environment using a PID controller, and the IMRFO algorithm was implemented in an intelligent optimization module to optimize the PID controller parameters.
To achieve seamless coordination between the control system and the mechanical system, a co-simulation platform was established by integrating a multibody dynamics simulation platform with a numerical control system as illustrated in
Figure 6. This setup enables real-time data exchange and computation between the control model and the artillery dynamics model. The control system processes the barrel elevation displacement state parameters obtained from the multibody dynamics model and computes the corresponding output torque, which is then fed back to drive the mechanical system. This approach ensures accurate and efficient interaction between the control strategy and the dynamic behavior of the artillery system.
4.2. Simulation Parameter Settings
Let the control parameters P, I, and D be , , and , respectively. Through multiple simulation verification runs, it was found that the control system tends to diverge when the controller parameters are too large or too small. To prevent ineffective searches by the algorithm, the search ranges for the control parameters were set to , , and .
In the IMRFO algorithm, the population size was set to N and the maximum number of iterations was T. Larger values of N and T improve convergence accuracy, but significantly increase training time. Considering these factors, N and T were chosen as and .
Considering the control requirements of high tracking accuracy and fast response speed, the objective function is defined as
where
is the simulation time,
and
are the weight coefficients for response speed and tracking accuracy, respectively, and
represents the difference between the actual system output and the desired output, with
,
, and
.
To verify the effectiveness and accuracy of the proposed IMRFO algorithm in the artillery stabilization control system, a comparative analysis was performed between the IMRFO and WOA algorithms, both of which performed well in benchmark function tests. Sinusoidal were used to validate the dynamic control performance of the artillery elevation stabilization system, while step inputs were used to validate its static performance. The desired elevation angle motion trajectories under both conditions were as follows:
In addition, to better evaluate the control performance of the controller, we used the Integral Absolute Error (IAE) and Integral Time-weighted Absolute Error (ITAE) as our performance metrics. Among these, the IAE reflects the cumulative error of the system during the entire operation, while the ITAE focuses more on the error performance over a longer period of time. The specific expressions are as follows:
4.3. Simulation Results
The simulation results are shown in
Figure 7,
Figure 8,
Figure 9 and
Figure 10. In these figures, the two graphs in
Figure 7 represent the fitness value convergence curves of the IMRFO algorithm under sinusoidal and step signals. From the simulation results, it can be seen that the IMRFO algorithm demonstrates superior performance over the WOA algorithm in both scenarios. Additionally, the IMRFO algorithm achieves relatively good results in the early stages of training due thanks to its incorporation of chaotic mapping-based population initialization.
The three graphs in
Figure 8 present the iteration curves of the control parameters. After multiple iterations, the global optimal solution of the IMRFO algorithm under sinusoidal signal input is
,
,
, while the global optimal solution of the WOA is
,
,
. Under step signal input, the global optimal solution of the IMRFO algorithm is
,
,
, while the global optimal solution of the WOA is
,
,
.
Figure 9 and
Figure 10 compare the control system’s input–output and error, respectively. The controller performance is shown in
Figure 11. Under sinusoidal input, the IMRFO and GWO algorithms perform similarly, with IMRFO slightly outperforming GWO in terms of IAE and ITAE values. However, under step output, IMRFO significantly outperforms GWO, with GWO’s IAE value being 51% higher than that of IMRFO and its ITAE value being 11% higher. Additionally, IMRFO converges within 0.3 s without significant oscillation; although the GWO algorithm has a lower steady-state error than the IMRFO algorithm, it has a slower response time and excessive overshoot.
4.4. Physical Validation
To validate the control performance of the PID controller parameters optimized by the IMRFO algorithm in an actual artillery stabilization system, a physical verification test was conducted. The primary objective of the test was to evaluate whether the optimized control parameters could outperform the empirical parameters in real-world applications and to further verify the practicality and reliability of the algorithm.
The physical verification was conducted on a semi-physical co-simulation platform integrating mechanical structure and electrical control. The test platform was provided by the China North Vehicle Research Institute. It was configured to focus on the vertical stabilization subsystem, with the performance of a key electromechanical actuator evaluated under different parameter settings. The mechanical subsystem includes key components such as a cradle, barrel, electric actuator, and adjustable hinges. The electrical control subsystem supports closed-loop control, comprising a host computer and a real-time processor from the mainstream DSP series. The system operates within a local network environment, allowing for real-time transmission of control commands and feedback signals.
Sensors such as a tilt sensor, rotary transformer, and eddy current probe were used to monitor the absolute and relative angular positions as well as the structural clearances. The control algorithm was developed in a mainstream embedded development environment and deployed to the DSP. Simulink-based models were compiled and downloaded to the real-time hardware, with experiment management handled via dSPACE ControlDesk. During the experiment, the system was powered up in a fixed sequence and real-time execution was initiated by switching to animation mode. The entire setup ensured that control strategies could be validated under realistic electromechanical coupling conditions, supporting effective evaluation of algorithm performance.
The experimental platform was configured to investigate the optimization of PI parameters within the velocity loop of the vertical stabilization system. The empirically determined parameters of and were used as the benchmark for comparison. To prevent system oscillation and divergence caused by improper control parameters, the optimization search range was set to , .
During the experiment, the reference signals included a sinusoidal signal (frequency 0.3 Hz, amplitude approximately 1°/s) and a step signal (output value of approximately 1°/s). The IMRFO algorithm was used to optimize the PI parameters with the objective of minimizing the tracking error of the control system for the sinusoidal signal, thereby enhancing the system’s dynamic response and stability.
Based on the training results of the IMRFO algorithm under sinusoidal signals, the optimal solution was determined as
,
.
Figure 12 shows the convergence curve of the fitness value, while
Figure 13a,b presents the system outputs under sinusoidal and step signal inputs using the optimized control parameters.
As shown in
Figure 14, the optimized PI parameters generally outperform the empirically determined values under both sinusoidal and step signals. Under sinusoidal inputs, the optimized control system achieves lower IAE and ITAE values, resulting in a smoother response and smaller error. Under step inputs, the optimized control parameters again surpass the empirical values. Although the maximum overshoot increases slightly, both IAE and ITAE decrease significantly, demonstrating a notable improvement in error control.