Next Article in Journal
Multi-Strategy Boosted Fick’s Law Algorithm for Engineering Optimization Problems and Parameter Estimation
Next Article in Special Issue
Dynamic Random Walk and Dynamic Opposition Learning for Improving Aquila Optimizer: Solving Constrained Engineering Design Problems
Previous Article in Journal
Nonlinear Growth Dynamics of Neuronal Cells Cultured on Directional Surfaces
Previous Article in Special Issue
BGOA-TVG: Binary Grasshopper Optimization Algorithm with Time-Varying Gaussian Transfer Functions for Feature Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Random Particle Swarm Optimization Based on Cosine Similarity for Global Optimization and Classification Problems

1
School of Intelligent Manufacturing Engineering, Jiangxi College of Application Science and Technology, Nanchang 330000, China
2
School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China
3
School of Computing and Artificial Intelligence, Southwestern University of Finance and Economics, Chengdu 611130, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2024, 9(4), 204; https://doi.org/10.3390/biomimetics9040204
Submission received: 26 February 2024 / Revised: 23 March 2024 / Accepted: 25 March 2024 / Published: 28 March 2024
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2024)

Abstract

:
In today’s fast-paced and ever-changing environment, the need for algorithms with enhanced global optimization capability has become increasingly crucial due to the emergence of a wide range of optimization problems. To tackle this issue, we present a new algorithm called Random Particle Swarm Optimization (RPSO) based on cosine similarity. RPSO is evaluated using both the IEEE Congress on Evolutionary Computation (CEC) 2022 test dataset and Convolutional Neural Network (CNN) classification experiments. The RPSO algorithm builds upon the traditional PSO algorithm by incorporating several key enhancements. Firstly, the parameter selection is adapted and a mechanism called Random Contrastive Interaction (RCI) is introduced. This mechanism fosters information exchange among particles, thereby improving the ability of the algorithm to explore the search space more effectively. Secondly, quadratic interpolation (QI) is incorporated to boost the local search efficiency of the algorithm. RPSO utilizes cosine similarity for the selection of both QI and RCI, dynamically updating population information to steer the algorithm towards optimal solutions. In the evaluation using the CEC 2022 test dataset, RPSO is compared with recent variations of Particle Swarm Optimization (PSO) and top algorithms in the CEC community. The results highlight the strong competitiveness and advantages of RPSO, validating its effectiveness in tackling global optimization tasks. Additionally, in the classification experiments with optimizing CNNs for medical images, RPSO demonstrated stability and accuracy comparable to other algorithms and variants. This further confirms the value and utility of RPSO in improving the performance of CNN classification tasks.

1. Introduction

Scientific and technological development generate a significant amount of data, including medical data, transaction data, house-price data, and more. Researchers can utilize this data to enhance their understanding of convenient technology. However, dealing with diverse types of data requires the construction of complex models, which are processed through machine learning. The diversity of data poses limitations on the machine learning approach, as the parameters in machine learning need to be updated differently depending on the data being processed. Consequently, optimizing machine learning becomes a major issue.
As an optimization technique, meta-heuristic algorithms are driven by natural phenomena and behaviors [1]. These algorithms possess the benefit of not requiring specialized understanding regarding the issue at hand. Instead, they iteratively explore and solve problems in a certain space, providing robustness and universality [2]. Particle Swarm Optimization (PSO) [3], a type of meta-heuristic algorithm, is representative and widely used. PSO is an optimization technique that mimics the natural hunting behavior of birds. It is characterized by its simplicity, requiring fewer parameters and being easy to implement compared to other methods. PSO exhibits good learning ability through the interaction of individual information during the evolutionary process. Consequently, it effectively addresses numerous real-world issues, including energy systems [4], path planning [5], image processing [6], DNA computing [7], and engineering optimization [8]. Consequently, there is significant research interest in combining PSO with machine learning techniques. In summary, meta-heuristic algorithms, such as PSO, offer a powerful approach for solving optimization problems. Their ability to adapt and explore solution spaces without requiring explicit problem knowledge makes them valuable tools for a wide range of applications.
PSO has gained significant attention in recent research for its ability to solve challenging problems in machine learning. Several studies have applied PSO to optimize various machine learning models and algorithms. For instance, Yong Zhu improved the LeNet-5 model by employing PSO to optimize its parameters [9]. Wu Deng optimized the parameters of the Least Squares Support Vector Machine (LS-SVM) for fault classification using an enhanced PSO algorithm [10]. Rajesh K. Yadav and Anubhav introduced a hybrid training algorithm combining PSO and the genetic algorithm (GA) alongside Adam Optimization to enhance the training process of artificial neural networks [11]. Kun Zhang utilized a PSO-optimized Least Squares Support Vector Machine (LSSVM) to establish a fault diagnosis model [12]. J. Beschi Raja and S. Chenthur Pandian employed PSO to optimize the Fuzzy Clustering Mean (FCM) for disease prediction [13]. Mohammed Nasser Al-Andoli integrated PSO with the backpropagation (BP) algorithm for malware detection [14]. Chao Huang used PSO to optimize Deep Convolutional Neural Networks (DCNNs) for automatic defect detection and classification [15]. Thomas George employed PSO for feature selection in electroencephalogram (EEG) signal classification [16]. Mohamed Issa combined PSO with the bat optimization algorithm for improved detection techniques [17]. Abhishek Dixit utilized a mix of algorithms, primarily based on the Differential Evolutionary Algorithm (DE) and PSO, for optimizing feature selection and classification [18]. Muhammad Sharif employed PSO for image segmentation and the GA for feature selection [19]. Hanuman optimized the FCM clustering algorithm for medical image segmentation using PSO [20]. Debendra Muduli leveraged Multi-Objective Discrete Particle Swarm Optimization (MODPSO) to enhance the performance of feedforward neural networks in breast cancer detection by optimizing hidden node parameters [21]. However, traditional PSO approaches face challenges in complex optimization models, as they can easily become caught in local optima, resulting in premature convergence. Additionally, the stability of the optimization capability of PSO leaves room for improvement [22].
To further boost the optimization capability of PSO, this article proposes a random particle swarm optimization algorithm based on cosine similarity, referred to as RPSO. RPSO adjusts the internal parameters of PSO and introduces the Random Contrastive Interaction mechanism and quadratic interpolation (QI). It employs cosine similarity to choose between the two mechanisms, which efficiently and dynamically updates population information, enhancing the global and local optimization capabilities of the algorithm. The performance tests were carried out based on two dimensions of CEC 2022 and compared with the PSO variant and the top algorithm of CEC to verify the competitiveness and effectiveness of RPSO. Finally, RPSO was optimized for the hyperparameters of the CNN, and the optimized model was confirmed to possess superior diagnostic abilities in both datasets. The specific work performed and innovations produced in this study are as follows:
  • Introduction of the Random Contrastive Interaction mechanism.
  • Introduction of the Quadratic interpolation mechanism.
  • Design of a judgment condition based on cosine similarity.
  • Comparison of PSO variants and the top algorithm in two dimensions of CEC 2022.
  • Optimization of CNN for the diagnosis of two diseases.
This paper is structured as follows: Section 2 introduces the theory of PSO and cosine similarity; Section 3 provides the introduction and analysis of the proposed algorithm; Section 5 presents the testing of the proposed algorithm based on CEC 2022; Section 4 highlights the improvement provided by the suggested algorithm when applied to CNN for classification experiments; and the last section concludes the work of the paper with a summary and discussion.

2. Related Theories

2.1. Particle Swarm Algorithms

In the PSO algorithm optimization process, the position information vector x is symbolized by a matrix of N × d and x i = ( x i 1 ,   x i 2 , , x i d ) , and the generated velocity vector v i = ( v i 1 , v i 2 , , v i d ) . The velocity and position are continuously revised through the following equation [12].
v i ( t + 1 ) = w v i ( t ) + z 1 r 1 ( x p b , i ( t ) x i ( t ) ) + z 2 r 2 ( x g b , i ( t ) x i ( t ) ) x i ( t + 1 ) = x i ( t ) + v i ( t + 1 )
In Equation (1), v i ( t + 1 ) and x i ( t + 1 ) represent individual i’s velocity and position during generation t, z1 and z2 denote the two acceleration coefficients, r1 and r2 denote the uniform random numbers [0, 1], and w is the inertia weight. x p b ( t ) and x g b ( t ) denote the global optimum and the historical optimal position of an individual at generation t, respectively.

2.2. Cosine Similarity

The cosine similarity measure serves to assess how similar two vectors are by finding the cosine of their intersection angle [23]. This value ranges between −1 and 1, where a cosine similarity value of 1 indicates that the vectors share the same direction, 0 indicates orthogonality (a 90-degree angle between the vectors), and −1 indicates that the vectors share opposite directions.
Unlike other similarity measures, cosine similarity disregards vector lengths and solely examines their pointing directions. As a result, it can determine whether two vectors are essentially heading in the same direction based solely on the cosine of the angle between them. Cosine similarity is typically utilized for positive spaces and yields values within the range [−1, 1]. This is illustrated in Figure 1, which represents the most common 2D space.
Figure 1a demonstrates that as the angle between two vectors approaches 0°, the cosine value approaches 1, indicating a high similarity in direction. In Figure 1b, a cosine value close to 0 indicates that the vectors are orthogonal, meaning they are independent of each other. Figure 1c illustrates that a cosine value closer to −1 implies that the vectors are oriented in opposite directions.
It is important to note that cosine similarity is independent of vector length and only considers the direction of the vectors. The formula for calculating cosine similarity is as follows [24]:
cos ( θ ) = i = 1 n ( A i × B i ) i = 1 n ( A i ) 2 × i = 1 n ( B i ) 2
where A i and B i are the i-th components of A and B, respectively.

3. The Proposed Algorithm

This subsection provides a description of the proposed algorithm and its principles.

3.1. Related Research from PSO

In recent years, researchers have made significant advancements in improving the PSO, which can be broadly divided into four primary categories: hybrid algorithms, multiple populations or multi-sampling approaches, adaptive learning mechanisms, and other miscellaneous techniques.
Hybrid algorithms have gained popularity in enhancing the performance of PSO by combining different optimization algorithms. For instance, Zeng [25] incorporated the DE algorithm into PSO to address the issue of premature convergence. Zaman [26] fused the Backtracking Search Optimization Algorithm (BSA) with PSO by modifying the mutation and crossover operators using neighborhood information, resulting in an improved convergence rate. Xu [27] proposed a strategy learning framework that utilized an adaptive DE algorithm to learn an optimal combination of strategies. Taking inspiration from the bee-foraging search mechanism of the Artificial Bee Colony (ABC) algorithm, Chen [28] developed a hybrid algorithm to tackle complex global optimization problems (GOPs). Khan [29] combined the local search ability of the Gravitational Search Algorithm (GSA) with the social thinking capability (gbest) of PSO. Another approach was taken by Molaei [30], who combined PSO with the GA. Ahmed [31] proposed a novel Hybrid Particle Swarm Optimization and Genetic Algorithm (HPSOGA) to minimize the energy function of molecules. HPSOGA balances exploration and exploitation, enhances diversity, and avoids premature convergence. It outperforms standard PSO and benchmark algorithms in solving large-scale global optimization problems and molecular energy functions.
In the field of multiple population or multi-sample techniques, several approaches have been proposed to enhance Particle Swarm Optimization (PSO). Xia [32] developed a Triple Archives PSO (TAPSO) that incorporates three different types of archives: elite particles, profiteer particles, and outstanding exemplars. These archives guide future merit-seeking mechanisms through genetic properties. Zhang [33] introduced the Particle Swarm Optimization-Alternating Least Squares (PSO-ALS) algorithm, which adaptively groups the population into sub-swarms and further divides particles into ordinary particles and the locally best particle within each sub-swarm. Two different learning strategies are designed to update the particles. Li [34] divides the particle swarm into elite and normal groups based on a ranking system and defines different types of neighborhoods accordingly. Sheng [35] proposed a PSO variant with multi-level population sampling and dynamic p-learning mechanisms. Lu [36] proposed an Enhanced Multi-Swarm Cooperative Particle Swarm Optimizer (EMCPSO) that divides the entire population into four identical sub-swarms. This algorithm incorporates a delayed-activation (DA) mechanism to detect sub-swarm stagnation and introduces a repulsive mechanism to prevent premature clustering of the entire population.
In terms of adaptive learning mechanisms, Liu [37] developed a novel PSO algorithm that incorporates a sigmoid-function-based weighting strategy. This strategy adjusts the acceleration coefficients by considering the distance of particles to the global best position and their personal best position. Liu [38] also proposed several other adaptive mechanisms, including a chaos-based non-linear inertia weight, stochastic and mainstream learning strategies, an adaptive position updating strategy, and a terminal replacement mechanism. Li [39] introduced a landscape-based adaptive operator selection mechanism. This mechanism quantifies the landscape modality, selects the most suitable evolutionary operator based on the population’s needs at different evolutionary stages, and adopts the mirrored boundary handling method to alleviate trapping in local optima. Zhang [40] proposed a particle swarm optimization algorithm with an empirical balance strategy (EBPSO), which selects a better search strategy from two equations using an adaptive adjustment mechanism. The algorithm dynamically adjusts the influence weight of the search equations and introduces a dynamic random search mechanism. Wang [41] proposed a hybrid PSO called Adaptive Strategy PSO (ASPSO), which incorporates a chaotic map, a position updating strategy, elite and dimensional learning strategies, and a competitive substitution mechanism. Li [42] proposed a novel PSO variant with a state-based adaptive velocity limit strategy (PSO-SAVL). In this strategy, the velocity limit is adaptively adjusted according to the estimated evolutionary state (ESE), with high values representing the global searching state and low values representing the local searching state.
In addition to the aforementioned categories, there are several other methods for improving PSO. Wang [43] presented a reinforcement learning strategy to adaptively adjust the optimization mechanism of PSO. Tian [44] proposed Variable Surrogate Model-based Particle Swarm Optimization (VSMPSO), which utilizes a single surrogate model constructed through simple random sampling and a variable model management strategy. Liu [45] incorporated Evolutionary Game Theory (EGT) to control the population state and proposed a selection mechanism and a mutation mechanism inspired by EGT concepts. Liu [46] employed Gaussian white noise with adjustable intensity to randomly perturb the acceleration coefficients, aiming to explore the problem space more extensively. Machado [47] introduced the concepts of complex-order derivatives (CD) and conjugate-order differentials.
Despite achieving certain optimization results, the aforementioned research approaches still have limitations. For example, hybrid algorithms may struggle to guarantee the complementary advantages of the two algorithms. Multiple swarms may not ensure effective local development. Adaptive mechanisms may possess irreversible search mechanisms. Additionally, it can be challenging to ensure the learning and adaptive ability of the algorithm in the existing research. To solve these limitations, this paper proposes a judgment mechanism based on cosine similarity and introduces the Random Contrastive Interaction (RCI) and QI mechanisms to jointly enhance the effective search capability of the algorithm.

3.2. Random Contrastive Interaction

In order to cope with complex and variable optimization environments, it is difficult for traditional particle swarms to maintain a good search ability to find high quality information, especially when the formulas use information about historical optimal positions. This can lead to a situation where multiple historical optimal positions cannot be updated with the global optimal position, resulting in a lack of diversity in the algorithm. For this reason, we introduce Random Contrastive Interaction [48], which provides better evolutionary information by means of the states within the population adaptively learning from each other. A specific schematic is shown in Figure 2.
As shown in Figure 2, P individuals are randomly selected from the population to form a random topology. Within this small population, there will be a local best solution, qbest, and a worst solution, qworst. The role of RCI is to enable the individuals within the small population to learn from qbest and qworst, and to interact with each other to obtain information, meaning that dynamically updating an individual’s position enhances the flexibility of the algorithm. The details are as follows [48]:
Z i = Q 1 · V i + Q 2 · ( q b e s t x P i ) + φ · Q 3 · ( q w o r s t x P i ) S i = x P i + Z i
where, Q 1 , Q 2 , Q 3 are uniform random numbers in the range of [0, 1] and φ is the learning parameter controlling qworst, which is 0.3. An optimal setting of ϕ for RPSO falls within a moderate range, neither too large (such as φ = 0.5) nor too small (such as φ < 0.3), as extremes hinder RPSO from achieving optimal performance. A φ that is too small diminishes the impact of the worst position on each updated particle, potentially causing rapid movement of particles towards the location of the best position and resulting in premature convergence or being trapped in local optima. On the other hand, a φ that is too large amplifies the influence of the worst position excessively, disrupting swarm convergence and leading to suboptimal performance of RPSO. Z i is the velocity of the newly generated individual and S i is the position of the newly generated individual. x P i denotes the position of the ith individual in the selected population P. It can be seen that the selection of individuals makes it possible to continuously move from the locally optimal individual to the locally worst position, increasing the exchange of information in the population and also improving the global exploration capacity.
Which the number of randomly selected individuals is constantly changing, in the early stage of the algorithm, in the stage of global exploration needs global search capability, at this time a larger number of individuals need to RCI operation, and in the later stage, a portion of the individuals need to carry out a certain amount of exploration to increase the diversity of the population. In the literature [48], the variation in the size of the selected populations is shown in Figure 3. TS is the population size for each selection. As shown in the Figure 3, the randomly selected population size presents an increasing manner, which will lead to a problem that the local exploitation ability is very poor at the later stage, and the feasible solution cannot be mined effectively. Therefore, this paper designs a new population design scheme, which is formulated as follows:
T Q = ( r a n d ( 1 ) + 1 ) × ( 1 t / T ) × N / 2 T S t = r o u n d ( m a x ( T Q , N / 10 ) )
The use of a max function is needed to ensure that there are at least N/10 individuals in the population. This ensures that the individuals need to carry out a certain exploration, in which the value of TQ is constantly changing. The overall trend is gradually decreasing, but this is an oscillating change. Taking the population number of 50 as an example, the specific value changes as shown in Figure 4; it can be seen that the range of changes in the early stage is larger, and in the later stage, this value gradually becomes smaller, and finally the structure of N/10 is unanimously adopted. Such a screening mechanism not only ensures the diversity of the population but also enables a detailed search of a certain area to find a higher quality of feasible method. On the other hand, the oscillation characteristics are beneficial, mainly reflected in the reversible changes in the population size. This can diversify the search process and compensate for any vulnerabilities in the algorithm’s search. The specific pseudo-code is shown in Algorithm 1.
Algorithm 1. Random Contrastive Interaction (RCI)
Input:
t: global optimum position
T: Matrix of population information
N: population size
fit: Matrix of objective function valuesx: Position of the population
Output:
X*: Matrix of newly generated stock information
The number of individuals TS to be selected for this iteration is calculated according to Equation (4);
1.      rs = randperm(N, TS); %Generate non-repeating individual subscripts based on TS.
2.      [af, bf] = sort(fit); % af is the result of the ascending order of the fitness values and bf is the corresponding subscript position.
3.      db = find(bf == rs(1));Find the location of the optimal individual in rs.
4.      dw = find(bf == rs(TS));Find the worst individual position in rs.
5.      dbest = pos(bbb(db),:);
6.      dworst = pos(bbb(dw),:);
7.      For i = 1:size(rs,2)
8.          For j = 1:size(x,2)
9.              Calculate the corresponding new position in rs according to Equation (3).
10.          End
11.      End
Return X*

3.3. Quadratic Interpolation (QI)

QI is a better local search technique that employs a parabola to adapt the form of a quadratic function to locate the extreme points of a curve [49]. The quadratic interpolation technique is used to further accelerate the information exchange of the population and improve the local search ability. The specific formula is as follows [50]:
x n e w = 0.5 · ( ( x 1 2 x 2 2 ) · f ( x 3 ) + ( x 3 2 x 1 2 ) · f ( x 1 ) + ( x 1 2 x 3 2 ) · f ( x 2 ) ( x 1 x 2 ) · f ( x 3 ) + ( x 2 x 3 ) · f ( x 1 ) + ( x 1 x 3 ) · f ( x 2 ) )
where x1, x2, and x3 are three individuals randomly selected from the population. In this paper, x1 represents the location of the globally optimal individual, while x2 and x3 represent two randomly selected individuals from the population. Algorithm 2 provides the specific pseudo-code. By facilitating information interaction between the globally optimal individual and the two individuals from the population, the algorithm guides the population to converge towards the vicinity of the optimal solution, thereby enhancing the algorithm’s local search capability.
Algorithm 2. Quadratic interpolation (QI)
Input:
gbest: global optimum position
x: Matrix of population information
Output:
x*: Matrix of newly generated stock information
1.       For i = 1:size(x)
2.       k = randperm(N, 2)%Two non-repeating individuals were randomly generated.
3.       The positional information of the new individual is calculated by substituting the three individuals k(1), k(2), and gbest into Equation (5).
4.       End
Return x*

3.4. Setting of Parameters

In the particle swarm algorithm, with fixed parameters, there is a fixed search amplitude and it is difficult to increase the diversity. Many scholars have proposed using the following formula for the setting of z1 and z2 with weights w [51].
w = w m a x ( w m a x w m i n ) × t / T z 1 = 2.5 2 × t / T z 2 = 0.5 + 2 × t / T
where w m a x is 0.9 and w m i n is 0.2, w decreases with the number of iterations, and the population’s diversity levels are gradually reducing. z 1 and z 2 changes are set by biasing the historical optimum and the global optimum; in the early stage of the algorithm, a larger z 1 and a smaller z 2 cause the historical optimum position to have a larger bias, which can cause the global exploration ability of the algorithm to increase, and a smaller z 1 and a larger z 2 can cause the search of the algorithm to bias towards the global optimum position and enhance the precision of the solution. This analysis shows that the combined effect of these three factors causes the algorithm to tend away from global exploration to local exploitation, and also balances the search scope of the algorithm.

3.5. Selection Mechanisms

In the process of algorithm optimization, establishing how to judge the population state more accurately is always an important direction of research. Making a clear plan according to the population state is the only way to ensure that the algorithm is constantly close to the theoretical optimal solution. Qiang Zhang used cosine similarity to dynamically update the mutation operation of DE [52], Moutaz Alazab used CS to convert continuous problems into binary problems [53], and Wei Li et al. used CS to upgrade algorithm speed [54]. In this paper, cosine similarity is introduced to determine the distance of an individual from the current optimal position, the number of individual similarity angles greater than 90 are recorded separately, and the choice of mechanism is made based on the number of recorded CN. When the angle is greater than 90 the cosine similarity is greater than 0, and vice versa. The specific selection mechanism is as follows:
R C I         i f     C N < N / 2 Q I                       o t h e r w i s e    
where i f   C N < N / 2 indicates that the similarity angle between most of the individuals within the current population and the current optimal position is less than 90 degrees. This reflects the fact that most of the individuals have high degrees of similarity with each other, and there is low population diversity. There also exists the probability of falling into the local optimum. This occurs when the RCI is used to improve the learning degree of the worst position and enhance the diversity of the population.
Meanwhile, if the number of individuals within the population with an angle greater than ninety degrees from the current optimal position is high, then this reflects the fact that the individuals within the population are not similar to the current optimal position that gives better global exploration ability. Under these circumstances, the QI mechanism is used to enhance the local search of the global optimal position, balancing the local and global search ability of the algorithm. This mechanism of selection ensures that each iteration does not completely lose the local exploitation or global exploration capacity, and it dynamically regulates the learning ability of the algorithm.

3.6. Flow of the Algorithm

To improve the judgment and learning ability of PSO, this paper proposes a stochastic learning PSO algorithm based on cosine similarity. Firstly, the algorithm is dynamically set up, and secondly, the selection of RCI and QI is carried out according to the characteristics of cosine similarity, jointly ensuring the capacity of the algorithm to globally explore and locally develop and improving the accuracy of the solution. Using the minimization problem as a case study, the specific pseudo-code of the Algorithm 3 is as follows.
Algorithm 3. RPSO
Input:
N: population size
T: Matrix of population information
Maxfes: Maximum number of iterations
Dim: Dimension of the problem
Lb: Lower bound of the problem
Ub: Upper bound of the problem
f: objective function
Output:
xgbest: Matrix of newly generated stock information
Fmin: optimal solution
1.      The position matrix x is obtained by random initialization based on the set parameters.
2.      Calculate the objective function value based on f to obtain the optimal position xgbest, the optimal solution fmin
3.       Fes = N;
4.       T = 1;
5.       While (t <= T)&&(Fes <= Maxfes)
6.           Calculate the new population position x according to Equations (1) and (7)
7.       Calculate the objective function value according to f to get the current optimal position xl its subscripts q, update the optimal position xgbest, the optimal solution fmin
8.       Fes = Fes + N;
9.           Nk = 0
10.         For i = 1:N
11.       If i ≠ q
12.       The cosine similarity between the individual and xl was calculated to obtain nc according to Equation (2)
13.       End
14.       If nc < 0
15.       Nk = Nk + 1;
16.       end
17.       end
18.         If Nk > N/2
19.       Perform according to Algorithm 2.
20.       Calculate the objective function value based on f to obtain the optimal position xgbest, the optimal solution fmin
21.       Fes = Fes + TS;
22.       else
23.       Perform according to Algorithm 1.
24.       Calculate the objective function value based on f to obtain the optimal position xgbest, the optimal solution fmin
25.       Fes = Fes + N;
26.         End
27.       t = t + 1;
28.       End
Return xgbest, fmin

4. Performance Tests

This section validates the proposed algorithm in the CEC 2022 test set, compares it with the variant of PSO and the top algorithm in the CEC competition, and conducts ablation experiments.

4.1. Comparison with PSO Variants

To demonstrate the optimized performance of RPSO, a comparison is conducted with recent variants of PSO algorithms to highlight the novelty and feasibility of RPSO. The compared variants include Adaptive multi-strategy ensemble particle swarm optimization (AMEPSO) [55], Elite archives-driven particle swarm optimization (EAPSO) [56], Improved Phasor Particle Swarm Optimization with Fitness Distance Balance (FDBPSO) [57], Velocity pausing particle swarm optimization (VPPSO) [58], a PSO variant for single-objective numerical optimization (PSOsono) [59], modified particle swarm optimization (MPSO), and Pyramid particle swarm optimization (PPSO) [60]. The internal parameter settings for each algorithm are based on the literature. The population size is set to 100, and the algorithms are tested on 10 and 20 dimensions using a maximum number of evaluations of 200,000 and 1,000,000. The experiments are performed utilizing MATLAB 2019a on an 11th Gen Intel(R) Core(TM) i5-11500 @ 2.70GHz processor. It is worth noting that in the program, all functions of CEC 2022 are minimization functions. We subtract the theoretical optimal value from the optimized result, so the theoretical optimal value becomes 0. The closer the optimization results of each algorithm are to 0, the better the optimization performance. Each algorithm is independently run 30 times, and the optimal value, worst value, median, mean, and standard deviation are recorded. These metrics are used to measure the stability and optimization capability of the algorithm. In simple terms, the smaller the value, the better the optimization effect of the algorithm. The Wilcoxon rank-sum test is used to analyze the differences between RPSO and other algorithms, with a significance level of 0.05. The symbol “+” indicates that RPSO performs better than a certain algorithm in a specific function, “−” indicates the opposite, and “=” shows that the optimization efficiency of the two algorithms is equal. Additionally, the Friedman test is used to measure the overall differences among algorithms and provide a final ranking based on the entire CEC 2022 test set [61]. The detailed optimization results are presented in Table 1 and Table 2. To visualize the distribution of optimization performance for each algorithm, Figure 5 and Figure 6 display the distribution of optimal solutions for each algorithm across different functions.
As shown in Table 1 and Figure 4, in the case of 10 dimensions the RPSO algorithm performs better than the other variants of the PSO algorithm in most cases. Specifically, the optimization standard deviation of the RPSO algorithm is the lowest value in the functions F1(X), F2(X), F5(X), F6(X), F7(X), F10(X), and F12(X). The theoretical optimum is found every time in F1(X), F5(X), and F11(X). In other cases, from small to large, the variance ranks second or third, indicating strong search stability.
The best and worst values of the PSO algorithm variant are contrasted with the best and worst values of the RPSO algorithm, and the number of times that the best and worst values of the PSO algorithm variant are greater than those of the RPSO algorithm in 12 functions is counted. The RPSO algorithm beats the EAPSO algorithm and the FDBPSO algorithm 11 times, the VPPSO algorithm 10 times, the AMSEPSO algorithm 7 times, the MSPO algorithm 5 times, the PPSO and PSOsono algorithms 4 times regarding the optimal value. Regarding the worst value, the RPSO algorithm beats the AMSEPSO, EAPSO, and PDBPSO algorithms 12 times, beats the PSOsono algorithms 11 times, beats the VPPSO algorithm 10 times, and beats the MPSO and PPSO algorithms 9 times and 8 times, respectively. The above information reveals that the RPSO algorithm outperforms the AMSEPSO, EAPSO, FDBPSO, and VPPSO algorithms multiple times in terms of both the optimal value and the worst value. The RPSO algorithm is obviously superior to the above four algorithms. Although the advantages of the MPSO, PPSO and PSOsono algorithms are not obvious in the optimal value, they show great advantages in the worst value. Therefore, the superiority of the proposed algorithm over the variants of the PSO algorithm is verified.
As shown in Table 2 and Figure 5, in the case of 20 dimensions, the RPSO algorithm performs better than other variants of the PSO algorithm in most cases. Specifically, the optimization standard deviation of the RPSO algorithm is the lowest value in functions F1(X), F2(X), F6(X), F10(X), F11(X), and F12(X). The theoretical optimum is found every time in F1(X) and F5(X). In other cases, from small to large, the variance ranks second or third, indicating strong search stability.
The best and worst values of the PSO algorithm variant are compared with the best and worst values of the RPSO algorithm, and the number of times that the best and worst values of the PSO algorithm variant is greater than those of the RPSO algorithm in 12 functions is counted. The RPSO algorithm beats the FDBPSO algorithm 12 times, the EAPSO algorithm 9 times, the VPPSO and AMSEPSO algorithms 8 times, the PPSO algorithm 5 times, and the MPSO and PSOsono algorithms 4 times regarding the optimal value. Regarding the worst value, the RPSO algorithm beats the FDBPSO algorithm 12 times, the EAPSO and VPPSO algorithms 11 times, the AMSEPSO and PPSO algorithms 10 times, and the MSPO and PSOsono algorithms 9 times and 8 times, respectively. The above information shows that the RPSO algorithm beats the AMSEPSO, EAPSO, FDBPSO and VPPSO algorithms many times in terms of the optimal value and the worst value. The RPSO algorithm is clearly superior to the four algorithms mentioned above. While the advantages of the MPSO, PPSO, and PSOsono algorithms may not be obvious in terms of optimal value, they demonstrate significant advantages in terms of worst value. Thus, the proposed algorithm’s superiority over the variants of the PSO algorithm is confirmed. The comparison values mentioned above are similar to the comparison values between the RPSO algorithm and the variants of the PSO algorithm in the case of ten dimensions, indicating the strong robustness of the RPSO algorithm.

4.2. Convergence Analysis

The convergence effect is also an important assessment index for measuring the optimization ability of the algorithm, and the results of the iterative convergence can show the responsiveness of the algorithm in different optimization-seeking environments, which further reflects the learning and adaptability of the algorithm. In this section, the convergence process for the above 30 times is averaged to obtain values, and the average convergence of each algorithm in each function is shown in Figure 7 and Figure 8.
As presented in Figure 7 and Figure 8, the convergence effect of RPSO in both dimensions has some advantages; furthermore, in ten dimensions, the convergence effect of RPSO on F(1–2), F(5–6), F8, and F(10–11) displays a reliable ability to seek optima compared with other algorithms, especially F(2), F(5–6), F(8), and F(11), where the convergence speed of RPSO is more significant. In 20 dimensions, the convergence of RPSO is better in the functions F(1), F(5–7), and F(9–11), particularly in the functions F(5–7), F(10–11), where the convergence speeds of RPSO display a significant advantage. The optimization search effect of RPSO still demonstrates better convergence performance in different dimensions, which verifies the learning and adaptability of RPSO.

4.3. Comparison with the Top Algorithm

This subsection aims to reveal the competitive edge of RPSO by comparing it with top algorithms from different years of the CEC competition, including the gaining-sharing knowledge-based algorithm with adaptive parameters hybrid with IMODE (APGSK-IMODE) [62], Eigen crossover in cooperative model of evolutionary algorithms (EA4eig) [63], Improved multi-operator differential evolution (IMODE) [64], Population’s variance-based adaptive differential evolution (PVADE) [65], and the adaptive gaining-sharing knowledge-based algorithm (AGSK) [66]. EA4eig was the champion algorithm in CEC 2022, APGSK-IMODE ranked fourth in CEC 2021, IMODE was the champion in CEC 2020, AGSK was the runner-up in CEC 2020, and PVADE was a recognized algorithm in the competition. The experimental parameters for each algorithm are set consistently as mentioned earlier, and the optimization results are presented in Table 3 and Table 4.
Analyzing Table 3, when the RPSO algorithm was compared with the five most advanced algorithms in ten dimensions, it can be observed that RPSO does not always achieve the best results across the 12 functions when compared with these algorithms. However, the optimization standard deviation of RPSO is the lowest in F(2), F(5), F(9), F(10), and F(11), indicating a strong search stability. Moreover, RPSO outperforms the EA4eig algorithm on the F(10) function, the IMODE algorithm on the F(4) function, the PVADE algorithm on the F(8) and F(10) functions, and the AGSK algorithm on the F(4) and F(10) functions in terms of the optimal value. RPSO also outperforms the APGSK-IMODE algorithm on the F(10) function; the EA4eig algorithm on the F(2) function; the IMODE algorithm on the F(4), F(5), F(10), and F(11) functions; and the PVADE algorithm on the F(2), F(8), and F(11) functions. Additionally, RPSO outperforms the AGSK algorithm on the F(2), F(4), F(5), and F(11) functions.
Examining Table 4, when comparing the RPSO algorithm with the top algorithms in 20 dimensions, it can be observed that RPSO does not always achieve the best results across the 12 functions when compared with these algorithms. However, RPSO exhibits better performance compared to the 10-dimensional case. The optimization standard deviation is the lowest in F(1), F(5), and F(7), indicating strong search stability. Additionally, RPSO outperforms the APGSK-IMODE algorithm on the F(1), F(5), and F(11) functions; the EA4eig algorithm on the F(11) function; and the IMODE algorithm on the F(3), F(4), F(5), F(7), F(8), F(10), and F(11) functions in terms of the optimal value. RPSO also outperforms the AGSK algorithm on the F(4), F(5), and F(11) functions. In terms of the worst value, RPSO outperforms the APGSK-IMODE algorithm on the F(1) and F(4) functions; the EA4eig algorithm on the F(10) and F(11) functions; and the IMODE algorithm on the F(4), F(5), F(7), F(10), and F(11) functions. RPSO also outperforms the PVADE algorithm on the F(9) and F(10) functions; and the AGSK algorithm on the F(1), F(4), F(5), and F(11) functions.

4.4. Ablation Experiments

In order to further confirm the effectiveness of strategy fusion, this subsection conducts ablation experiments one by one on the mechanisms proposed in this paper. The PSO that only incorporates random exchange learning is set as PSO-1, the PSO algorithm that only incorporates QI is set as PSO-2, and the PSO algorithm that fuses the two mechanisms and does not use the selection mechanism is named PSO-3. It is worth mentioning that the parameter settings of RPSO are derived from previous research by other scholars, and thus are not included in the scope of validation. The other experimental parameters are consistent with those described, and the specific optimization results are shown in Table 5.
As can be seen from the Table 5, except for F12, RPSO has optimal values in all other functions, especially in the functions F1–3, F5, F7–11, where the indexes of RPSO are optimal in these functions. The other variants are all able to find the theoretical optimal values in the functions F1–3, F5, and F11. This verifies that the incorporation of the various mechanisms has a certain degree of rationality. Taken together, RPSO has certain advantages in the arrangement and combination of integrated mechanisms, verifying the feasibility and effectiveness of RPSO.

5. Classification Experiments

In this section, we optimize the hyperparameters of a CNN for enhancing the diagnostic rate through RPSO and verify the feasibility of this model using two datasets.

5.1. Means

In this study, the optimization of hyperparameters was performed using RPSO. The accuracy and convergence of CNNs are heavily determined by hyperparameters [67]. Properly choosing hyperparameters is crucial, as they are determined by the particular use of the CNN. Some commonly employed hyperparameters for training CNNs include the learning rate, the number of epochs, the momentum, and the regularization factor. The learning rate governs the step size of the gradient descent algorithm, dictating how quickly or slowly the network parameters are updated. An increased learning rate facilitates faster convergence yet may result in overshooting the optimal solution. On the other hand, a decreased learning rate may take more iterations to converge, but it can yield more accurate results. The number of epochs specifies the frequency of updating the network parameters based on the training dataset. Striking a balance between not training enough epochs and overtraining, which could lead to underfitting and overfitting, is crucial. Momentum is a parameter that governs the influence of previous weight updates on the current weight update. It helps to accelerate convergence and overcome local minima. Regularization is a technique employed to prevent overfitting in the network. It accomplishes this by adding a regularization term to the loss function, penalizing complex models and promoting simpler models that can be generalized to unseen data more effectively. Optimizing hyperparameters is imperative to achieve the optimal performance of the CNN. Techniques such as RPSO can be utilized to explore the hyperparameter space and identify the most suitable values for the specific problem at hand.

5.2. Training

In this section, two datasets, CT (https://github.com/UCSD-AI4H/COVID-CT, accessed on 20 March 2023) and XY (https://www.kaggle.com/datasets/paultimothymooney/chest-xray-pneumonia, accessed on 20 March 2023), were selected to validate the algorithm’s ability to optimize the CNN. Overall, 70% of the data were utilized for training, while the remaining 30% were allocated for network testing. The proposed architecture aims to distinguish between normal and abnormal images. The dataset was randomly divided into training and testing sets. Through data augmentation techniques, all images were resized to 224 × 224 × 3 and converted into color images. A six convolutional layer structure was employed for image classification, with Stochastic Gradient Descent (SGD) training used to adjust parameters and RPSO for hyperparameter optimization during the training process.

5.3. Testing

Firstly, the proposed network was augmented with data and trained by resizing the image to 224 × 224 × 3. Then, the test image was fed into the trained CNN that had already optimized all the parameters of Convolutional Layers (CLs) and Fully Connected Layers (FCL). The CNN starts by extracting image characteristics and then categorizes them into appropriate groups through FCL and soft-max classifiers.

5.4. Experiments

To show the diagnostic performance of RPSO, the diagnostic effectiveness of RPSO was compared with that of PSO, the whale optimization algorithm (WOA) [68], the grey wolf optimizer (GWO), the hybrid differential evolution algorithm and particle swarm optimization (DEPSO) [69], and the advanced squirrel search optimization algorithm (ASSOA) [70]. Each algorithm was run independently for ten epochs; the best value, mean value, and worst value of every algorithm were recorded; and the confusion matrix of the optimal value, Accuracy, Precision, Recall, and F1 Score [71] was produced. The specific results are shown in Table 6 and Figure 9 and Figure 10.
According to Table 6 and Figure 9 and Figure 10, in the XY data set the algorithm accuracy is sorted from low to high. The results show that the accuracy rate of the PSO algorithm in the diagnosis of pneumonia is 98.21%, the number of normal correct diagnoses times is 85, and the number of correct diagnoses of COVID-19 is 84. The accuracy rate of the WOA algorithm in the diagnosis of pneumonia is 98.81%, the number of normal and correct diagnoses times is 81, and the number of times with correct diagnosis of COVID-19 is 85. The accuracy of the DEPSO algorithm in the diagnosis of pneumonia is 98.81%, the number of normal correct diagnoses is 81, and the number of correct diagnoses of COVID-19 is 85. The accuracy of the ASSOS algorithm in the diagnosis of pneumonia is 98.81%, the number of normal correct diagnoses is 82, and the number of correct diagnoses of COVID-19 is 84. The accuracy of the GWO algorithm in the diagnosis of pneumonia is 99.40%, and the number of normal correct diagnoses is 82, while the number of times with the correct diagnosis of COVID-19 is 85. The accuracy rate of the RPSO algorithm in diagnosing pneumonia is 100%. The number of correct diagnoses for normal cases is 82, while the number of correct diagnoses for COVID-19 cases is 86.
In the CT dataset, the algorithm accuracy is ranked from low to high. The results indicate that the ASSOA algorithm has an accuracy of 85.71% in diagnosing pneumonia, the number of normal correct diagnoses is 104, and the number of correct diagnoses of COVID-19 cases is 88. The DEPSO algorithm has an accuracy rate of 83.92% in diagnosing pneumonia. The number of correct diagnoses for normal cases is 104, and the number of correct diagnoses for COVID-19 cases is 84. The GWO algorithm has an accuracy of 86.16% in diagnosing pneumonia. The number of correct diagnoses for normal cases is 103, and the number of correct diagnoses for COVID-19 cases is 90. The WOA algorithm has an accuracy of 86.61% in diagnosing pneumonia. The number of correct diagnoses for normal cases is 102, and the number of correct diagnoses for COVID-19 cases is 92. The accuracy of the PSO algorithm in diagnosing pneumonia is 87.05%. The number of correct diagnoses for normal cases is 111, and for COVID-19 cases it is 84. The accuracy rate of the RPSO algorithm in diagnosing pneumonia is also 87.05%. The number of correct diagnoses for normal cases is 107, and for COVID-19 cases it is 88.
It can be seen that, compared with other algorithms, the RPSO algorithm has the highest accuracy in both data sets, surpassing all of the compared algorithms. Although the accuracy of the PSO algorithm in the CT data set is equal to that of RPSO, with both ranking first, it has the lowest accuracy in the XY data set and is highly unstable. It is concluded that RPSO has obvious superiority and good robustness.

6. Conclusions

To further enhance the global optimization ability of the particle swarm algorithm and its applicability in threshold segmentation problems, this paper proposes an RPSO based on cosine similarity. This algorithm introduces the RCI mechanism to enhance the mechanism and the global search capability of the algorithm. Utilizing cosine similarity for RCI and QI selection, the population information is effectively and dynamically updated, thereby improving the global and local optimization capabilities of the algorithm. RPSO is evaluated based on the CEC 2022 test set and compared with other PSO variants and top algorithms in the CEC community. In comparison with other variants of PSO, RPSO is ranked top at 7.9167 and 2.5833 in 10 and 20 dimensions, respectively. When compared with the top algorithms, RPSO achieves optimal values in F(1), F(5), and F(11), performing on par with the leading algorithms. The results confirm that RPSO exhibits strong competitive and global optimization capabilities. In the optimization of CNN classification experiments, RPSO demonstrates the highest classification accuracy of 100% on the XY dataset, with an average accuracy exceeding 98%. On the CT dataset, the highest accuracy reached is over 87%, with an average accuracy of above 85%. RPSO demonstrates excellent diagnostic performance based on two disease datasets compared to other basic algorithms and algorithm variants, highlighting its practical value.
Despite the promising optimization results achieved by RPSO based on the CEC 2022 test set, some limitations still exist. For instance, in functions F(3) and F(4), RPSO exhibits some shortcomings in search accuracy, indicating the need to improve its local search capability while maintaining its global search ability, as RPSO is not capable of making intelligent choices based on the optimization-seeking environment. Additionally, optimizing the hyperparameters of the CNN to significantly enhance the classification accuracy is a challenging task, as optimization is a major problem that takes an inordinate amount of time to ensure. Therefore, the next step of this research focuses on three main areas:
  • Designing smarter judgment mechanisms for rational global and local searching.
  • Incorporating efficient parameter estimation schemes to reduce the time consumption of the optimization process.
  • Optimizing the architecture of the neural network to enable automatic architectural tuning for different datasets.
  • Applying the algorithm to practical scenarios in high demand today, such as 6G base station deployment, optimization of new energy systems, and structural design optimization.

Author Contributions

Conceptualization, Y.Z. (Yuan Zeng). and R.L.; methodology, G.H.; software, X.Z.; validation Y.Z. (Yuemai Zhang). and T.L.; formal analysis, W.L.; data curation, D.Z.; writing—original draft preparation, Y.L.; writing—review and editing, G.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Research Project of Jiangxi Provincial Education Department (Grant No. GJJ2203314).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Eltamaly, A.M. A novel musical chairs algorithm applied for MPPT of PV systems. Renew. Sustain. Energy Rev. 2021, 146, 111135. [Google Scholar] [CrossRef]
  2. Zhu, D.; Wang, S.; Zhou, C.; Yan, S.; Xue, J. Human memory optimization algorithm: A memory-inspired optimizer for global optimization problems. Expert Syst. Appl. 2024, 237, 121597. [Google Scholar] [CrossRef]
  3. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; IEEE: New York, NY, USA, 1995; pp. 1942–1948. [Google Scholar]
  4. Eltamaly, A.M. A novel particle swarm optimization optimal control parameter determination strategy for maximum power point trackers of partially shaded photovoltaic systems. Eng. Optim. 2022, 54, 634–650. [Google Scholar] [CrossRef]
  5. Zhu, D.; Wang, S.; Shen, J.; Zhou, C.; Li, T.; Yan, S. A multi-strategy particle swarm algorithm with exponential noise and fitness-distance balance method for low-altitude penetration in secure space. J. Comput. Sci. 2023, 74, 102149. [Google Scholar] [CrossRef]
  6. Kocak, O.; Erkan, U.; Toktas, A.; Gao, S. PSO-based image encryption scheme using modular integrated logistic exponential map. Expert Syst. Appl. 2024, 237, 121452. [Google Scholar] [CrossRef]
  7. Xie, L.; Wang, S.; Zhu, D.; Hu, G.; Zhou, C. DNA Sequence Optimization Design of Arithmetic Optimization Algorithm Based on Billiard Hitting Strategy. Interdisciplinary Sciences. Comput. Life Sci. 2023, 15, 231–248. [Google Scholar]
  8. Guedria, N.B. Improved accelerated PSO algorithm for mechanical engineering optimization problems. Appl. Soft Comput. 2016, 40, 455–467. [Google Scholar] [CrossRef]
  9. Zhu, Y.; Li, G.; Wang, R.; Tang, S.; Su, H.; Cao, K. Intelligent fault diagnosis of hydraulic piston pump combining improved LeNet-5 and PSO hyperparameter optimization. Appl. Acoust. 2021, 183, 108336. [Google Scholar] [CrossRef]
  10. Deng, W.; Yao, R.; Zhao, H.; Yang, X.; Li, G. A novel intelligent diagnosis method using optimal LS-SVM with improved PSO algorithm. Soft Comput. 2019, 23, 2445–2462. [Google Scholar] [CrossRef]
  11. Yadav, R.K. PSO-GA based hybrid with Adam Optimization for ANN training with application in Medical Diagnosis. Cogn. Syst. Res. 2020, 64, 191–199. [Google Scholar] [CrossRef]
  12. Zhang, K.; Su, J.; Sun, S.; Liu, Z.; Wang, J.; Du, M.; Zhang, Q. Compressor fault diagnosis system based on PCA-PSO-LSSVM algorithm. Sci. Prog. 2021, 104, 368504211026110. [Google Scholar] [CrossRef] [PubMed]
  13. Raja, J.B.; Pandian, S.C. PSO-FCM based data mining model to predict diabetic disease. Comput. Methods Programs Biomed. 2020, 196, 105659. [Google Scholar] [CrossRef]
  14. Al-Andoli, M.N.; Tan, S.C.; Sim, K.S.; Lim, C.P.; Goh, P.Y. Parallel Deep Learning with a hybrid BP-PSO framework for feature extraction and malware classification. Appl. Soft Comput. 2022, 131, 109756. [Google Scholar] [CrossRef]
  15. Huang, C.; Zhang, Z.; Wang, L. PSOPruner: PSO-based deep convolutional neural network pruning method for PV module defects classification. IEEE J. Photovolt. 2022, 12, 1550–1558. [Google Scholar] [CrossRef]
  16. George, S.T.; Subathra, M.S.P.; Sairamya, N.J.; Susmitha, L.; Premkumar, M.J. Classification of epileptic EEG signals using PSO based artificial neural network and tunable-Q wavelet transform. Biocybern. Biomed. Eng. 2020, 40, 709–728. [Google Scholar] [CrossRef]
  17. Issa, M.; Helmi, A.M.; Elsheikh, A.H.; Abd Elaziz, M. A biological sub-sequences detection using integrated BA-PSO based on infection propagation mechanism: Case study COVID-19. Expert Syst. Appl. 2022, 189, 116063. [Google Scholar] [CrossRef] [PubMed]
  18. Dixit, A.; Mani, A.; Bansal, R. CoV2-Detect-Net: Design of COVID-19 prediction model based on hybrid DE-PSO with SVM using chest X-ray images. Inf. Sci. 2021, 571, 676–692. [Google Scholar] [CrossRef]
  19. Sharif, M.; Amin, J.; Raza, M.; Yasmin, M.; Satapathy, S.C. An integrated design of particle swarm optimization (PSO) with fusion of features for detection of brain tumor. Pattern Recognit. Lett. 2020, 129, 150–157. [Google Scholar] [CrossRef]
  20. Verma, H.; Verma, D.; Tiwari, P.K. A population based hybrid FCM-PSO algorithm for clustering analysis and segmentation of brain image. Expert Syst. Appl. 2021, 167, 114121. [Google Scholar] [CrossRef]
  21. Muduli, D.; Dash, R.; Majhi, B. Fast discrete curvelet transform and modified PSO based improved evolutionary extreme learning machine for breast cancer detection. Biomed. Signal Process. Control 2021, 70, 102919. [Google Scholar] [CrossRef]
  22. Luo, X.; Chen, J.; Yuan, Y.; Wang, Z. Pseudo Gradient-Adjusted Particle Swarm Optimization for Accurate Adaptive Latent Factor Analysis. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 2213–2226. [Google Scholar] [CrossRef]
  23. Gupta, A.; Nahar, P. Sandpiper optimization algorithm with cosine similarity based cross-layer routing protocol for smart agriculture in wireless sensor network assisted internet of things systems. Int. J. Commun. Syst. 2023, 36, e5514. [Google Scholar] [CrossRef]
  24. Kwak, B.I.; Han, M.L.; Kim, H.K. Cosine similarity based anomaly detection methodology for the CAN bus. Expert Syst. Appl. 2021, 166, 114066. [Google Scholar] [CrossRef]
  25. Zeng, N.; Wang, Z.; Liu, W.; Zhang, H.; Hone, K.; Liu, X. A dynamic neighborhood-based switching particle swarm optimization algorithm. IEEE Trans. Cybern. 2020, 52, 9290–9301. [Google Scholar] [CrossRef]
  26. Zaman, H.R.R.; Gharehchopogh, F.S. An improved particle swarm optimization with backtracking search optimization algorithm for solving continuous optimization problems. Eng. Comput. 2022, 38 (Suppl. S4), 2797–2831. [Google Scholar] [CrossRef]
  27. Xu, H.Q.; Gu, S.; Fan, Y.C.; Li, X.S.; Zhao, Y.F.; Zhao, J.; Wang, J.J. A strategy learning framework for particle swarm optimization algorithm. Inf. Sci. 2023, 619, 126–152. [Google Scholar] [CrossRef]
  28. Chen, X.; Tianfield, H.; Du, W. Bee-foraging learning particle swarm optimization. Appl. Soft Comput. 2021, 102, 107134. [Google Scholar] [CrossRef]
  29. Khan, T.A.; Ling, S.H. A novel hybrid gravitational search particle swarm optimization algorithm. Eng. Appl. Artif. Intell. 2021, 102, 104263. [Google Scholar] [CrossRef]
  30. Molaei, S.; Moazen, H.; Najjar-Ghabel, S.; Farzinvash, L. Particle swarm optimization with an enhanced learning strategy and crossover operator. Knowl.-Based Syst. 2021, 215, 106768. [Google Scholar] [CrossRef]
  31. Ali, A.F.; Tawhid, M.A. A hybrid particle swarm optimization and genetic algorithm with population partitioning for large scale optimization problems. Ain Shams Eng. J. 2017, 8, 191–206. [Google Scholar] [CrossRef]
  32. Xia, X.; Gui, L.; Yu, F.; Wu, H.; Wei, B.; Zhang, Y.L.; Zhan, Z.H. Triple archives particle swarm optimization. IEEE Trans. Cybern. 2019, 50, 4862–4875. [Google Scholar] [CrossRef] [PubMed]
  33. Zhang, Y.; Liu, X.; Bao, F.; Chi, J.; Zhang, C.; Liu, P. Particle swarm optimization with adaptive learning strategy. Knowl.-Based Syst. 2020, 196, 105789. [Google Scholar] [CrossRef]
  34. Li, H.; Li, J.; Wu, P.; You, Y.; Zeng, N. A ranking-system-based switching particle swarm optimizer with dynamic learning strategies. Neurocomputing 2022, 494, 356–367. [Google Scholar] [CrossRef]
  35. Sheng, M.; Wang, Z.; Liu, W.; Wang, X.; Chen, S.; Liu, X. A particle swarm optimizer with multi-level population sampling and dynamic p-learning mechanisms for large-scale optimization. Knowl.-Based Syst. 2022, 242, 108382. [Google Scholar] [CrossRef]
  36. Lu, J.; Zhang, J.; Sheng, J. Enhanced multi-swarm cooperative particle swarm optimizer. Swarm Evol. Comput. 2022, 69, 100989. [Google Scholar] [CrossRef]
  37. Liu, W.; Wang, Z.; Yuan, Y.; Zeng, N.; Hone, K.; Liu, X. A novel sigmoid-function-based adaptive weighted particle swarm optimizer. IEEE Trans. Cybern. 2019, 51, 1085–1093. [Google Scholar] [CrossRef]
  38. Liu, H.; Zhang, X.W.; Tu, L.P. A modified particle swarm optimization using adaptive strategy. Expert Syst. Appl. 2020, 152, 113353. [Google Scholar] [CrossRef]
  39. Li, Y.; Yu, K.; Liang, J.; Yue, C.; Qiao, K. A landscape-aware particle swarm optimization for parameter identification of photovoltaic models. Appl. Soft Comput. 2022, 131, 109793. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Kong, X. A particle swarm optimization algorithm with empirical balance strategy. Chaos Solitons Fractals X 2023, 10, 100089. [Google Scholar] [CrossRef]
  41. Wang, R.; Hao, K.; Chen, L.; Wang, T.; Jiang, C. A novel hybrid particle swarm optimization using adaptive strategy. Inf. Sci. 2021, 579, 231–250. [Google Scholar] [CrossRef]
  42. Li, X.; Mao, K.; Lin, F.; Zhang, X. Particle swarm optimization with state-based adaptive velocity limit strategy. Neurocomputing 2021, 447, 64–79. [Google Scholar] [CrossRef]
  43. Wang, F.; Wang, X.; Sun, S. A reinforcement learning level-based particle swarm optimization algorithm for large-scale optimization. Inf. Sci. 2022, 602, 298–312. [Google Scholar] [CrossRef]
  44. Tian, J.; Hou, M.; Bian, H.; Li, J. Variable surrogate model-based particle swarm optimization for high-dimensional expensive problems. Complex Intell. Syst. 2023, 9, 3887–3935. [Google Scholar] [CrossRef]
  45. Liu, Z.; Nishi, T. Strategy dynamics particle swarm optimizer. Inf. Sci. 2022, 582, 665–703. [Google Scholar] [CrossRef]
  46. Liu, W.; Wang, Z.; Zeng, N.; Yuan, Y.; Alsaadi, F.E.; Liu, X. A novel randomised particle swarm optimizer. Int. J. Mach. Learn. Cybern. 2021, 12, 529–540. [Google Scholar] [CrossRef]
  47. Machado, J.T.; Pahnehkolaei, S.M.A.; Alfi, A. Complex-order particle swarm optimization. Commun. Nonlinear Sci. Numer. Simul. 2021, 92, 105448. [Google Scholar] [CrossRef]
  48. Yang, Q.; Song, G.W.; Chen, W.N.; Jia, Y.H.; Gao, X.D.; Lu, Z.Y.; Zhang, J. Random Contrastive Interaction for Particle Swarm Optimization in High-Dimensional Environment. IEEE Trans. Evol. Comput. 2023; in press. [Google Scholar]
  49. Ye, J.; Xie, L.; Wang, H. A water cycle algorithm based on quadratic interpolation for high-dimensional global optimization problems. Appl. Intell. 2023, 53, 2825–2849. [Google Scholar] [CrossRef]
  50. Qaraad, M.; Amjad, S.; Hussein, N.K.; Farag, M.A.; Mirjalili, S.; Elhosseini, M.A. Quadratic interpolation and a new local search approach to improve particle swarm optimization: Solar photovoltaic parameter estimation. Expert Syst. Appl. 2024, 236, 121417. [Google Scholar] [CrossRef]
  51. Ratnaweera, A.; Halgamuge, S.K.; Watson, H.C. Self-organizing hierarchical particle swarm optimizer with time-varying acceleration coefficients. IEEE Trans. Evol. Comput. 2004, 8, 240–255. [Google Scholar] [CrossRef]
  52. Zhang, Q.; Zou, D.; Duan, N. An improved differential evolution using self-adaptable cosine similarity for economic emission dispatch. Energy 2023, 283, 128421. [Google Scholar] [CrossRef]
  53. Alazab, M.; Khurma, R.A.; Awajan, A.; Camacho, D. A new intrusion detection system based on Moth–Flame Optimizer algorithm. Expert Syst. Appl. 2022, 210, 118439. [Google Scholar] [CrossRef]
  54. Li, W.; Liang, P.; Sun, B.; Sun, Y.; Huang, Y. Reinforcement learning-based particle swarm optimization with neighborhood differential mutation strategy. Swarm Evol. Comput. 2023, 78, 101274. [Google Scholar] [CrossRef]
  55. Yang, J.; Yu, J.; Huang, C. Adaptive multistrategy ensemble particle swarm optimization with Signal-to-Noise ratio distance metric. Inf. Sci. 2022, 612, 1066–1094. [Google Scholar] [CrossRef]
  56. Zhang, Y. Elite archives-driven particle swarm optimization for large scale numerical optimization and its engineering applications. Swarm Evol. Comput. 2023, 76, 101212. [Google Scholar] [CrossRef]
  57. Duman, S.; Kahraman, H.T.; Korkmaz, B.; Bakir, H.; Guvenc, U.; Yilmaz, C. Improved Phasor Particle Swarm Optimization with Fitness Distance Balance for Optimal Power Flow Problem of Hybrid AC/DC Power Grids. In Proceedings of the International Conference on Artificial Intelligence and Applied Mathematics in Engineering, Belek, Turkey, 1–3 October 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 307–336. [Google Scholar]
  58. Shami, T.M.; Mirjalili, S.; Al-Eryani, Y.; Daoudi, K.; Izadi, S.; Abualigah, L. Velocity pausing particle swarm optimization: A novel variant for global optimization. Neural Comput. Appl. 2023, 35, 9193–9223. [Google Scholar] [CrossRef]
  59. Meng, Z.; Zhong, Y.; Mao, G.; Liang, Y. PSO-sono: A novel PSO variant for single-objective numerical optimization. Inf. Sci. 2022, 586, 176–191. [Google Scholar] [CrossRef]
  60. Li, T.; Shi, J.; Deng, W.; Hu, Z. Pyramid particle swarm optimization with novel strategies of competition and cooperation. Appl. Soft Comput. 2022, 121, 108731. [Google Scholar] [CrossRef]
  61. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  62. Mohamed, A.W.; Hadi, A.A.; Agrawal, P.; Sallam, K.M.; Mohamed, A.K. Gaining-sharing knowledge based algorithm with adaptive parameters hybrid with IMODE algorithm for solving CEC 2021 benchmark problems. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation (CEC), Kraków, Poland, 28 June–1 July 2021; IEEE: New York, NY, USA, 2021; pp. 841–848. [Google Scholar]
  63. Bujok, P.; Kolenovsky, P. Eigen crossover in cooperative model of evolutionary algorithms applied to CEC 2022 single objective numerical optimization. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; IEEE: New York, NY, USA, 2022; pp. 1–8. [Google Scholar]
  64. Sallam, K.M.; Elsayed, S.M.; Chakrabortty, R.K.; Ryan, M.J. Improved multi-operator differential evolution algorithm for solving unconstrained problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2020; IEEE: New York, NY, USA, 2021; pp. 1–8. [Google Scholar]
  65. dos Santos Coelho, L.; Ayala, H.V.; Freire, R.Z. Population’s variance-based adaptive differential evolution for real parameter optimization. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; IEEE: New York, NY, USA, 2013; pp. 1672–1677. [Google Scholar]
  66. Mohamed, A.W.; Hadi, A.A.; Mohamed, A.K.; Awad, N.H. Evaluating the performance of adaptive gaining sharing knowledge based algorithm on CEC 2020 benchmark problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 1–8. [Google Scholar]
  67. Da Silva, G.L.F.; Valente, T.L.A.; Silva, A.C.; De Paiva, A.C.; Gattass, M. Convolutional neural network-based PSO for lung nodule false positive reduction on CT images. Comput. Methods Programs Biomed. 2018, 162, 109–118. [Google Scholar] [CrossRef]
  68. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  69. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  70. El-Kenawy, E.S.M.; Mirjalili, S.; Ibrahim, A.; Alrahmawy, M.; El-Said, M.; Zaki, R.M.; Eid, M.M. Advanced meta-heuristics, convolutional neural networks, and feature selectors for efficient COVID-19 X-ray chest image classification. IEEE Access 2021, 9, 36019–36037. [Google Scholar] [CrossRef] [PubMed]
  71. Goel, T.; Murugan, R.; Mirjalili, S.; Chakrabartty, D.K. OptCoNet: An optimized convolutional neural network for an automatic diagnosis of COVID-19. Appl. Intell. 2021, 51, 1351–1366. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Angles in 2D space.
Figure 1. Angles in 2D space.
Biomimetics 09 00204 g001
Figure 2. Schematic diagram of RCI.
Figure 2. Schematic diagram of RCI.
Biomimetics 09 00204 g002
Figure 3. TS in literature [48].
Figure 3. TS in literature [48].
Biomimetics 09 00204 g003
Figure 4. Proposed TS.
Figure 4. Proposed TS.
Biomimetics 09 00204 g004
Figure 5. Distribution of solutions for each algorithm in 10 dimensions.
Figure 5. Distribution of solutions for each algorithm in 10 dimensions.
Biomimetics 09 00204 g005aBiomimetics 09 00204 g005b
Figure 6. Distribution of solutions for each algorithm in 20 dimensions.
Figure 6. Distribution of solutions for each algorithm in 20 dimensions.
Biomimetics 09 00204 g006aBiomimetics 09 00204 g006b
Figure 7. Plot of average convergence effect of each algorithm (dim = 10).
Figure 7. Plot of average convergence effect of each algorithm (dim = 10).
Biomimetics 09 00204 g007aBiomimetics 09 00204 g007b
Figure 8. Plot of average convergence effect of each algorithm (dim = 20).
Figure 8. Plot of average convergence effect of each algorithm (dim = 20).
Biomimetics 09 00204 g008aBiomimetics 09 00204 g008b
Figure 9. Confusion matrix for each algorithm with the XY dataset.
Figure 9. Confusion matrix for each algorithm with the XY dataset.
Biomimetics 09 00204 g009aBiomimetics 09 00204 g009b
Figure 10. Confusion matrix for each algorithm with the XY dataset.
Figure 10. Confusion matrix for each algorithm with the XY dataset.
Biomimetics 09 00204 g010
Table 1. Ten dimensions with PSO algorithm variants.
Table 1. Ten dimensions with PSO algorithm variants.
FIndexAMSE
PSO
EAPSOFDBPSOMPSOPPSOPSOsonoVPPSORPSO
F1(x)Best0.00 × 10001.13 × 10031.12 × 10030.00 × 10000.00 × 10000.00 × 10001.93 × 10−90.00 × 1000
Worst1.26 × 10021.76 × 10041.29 × 10042.25 × 10−070.00 × 10000.00 × 10007.62 × 10−90.00 × 1000
Median2.49 × 10016.75 × 10035.09 × 10031.38 × 10−100.00 × 10000.00 × 10004.34 × 10−90.00 × 1000
Mean3.25 × 10016.90 × 10036.04 × 10031.40 × 10−80.00 × 10000.00 × 10004.39 × 10−90.00 × 1000
Standard3.56 × 10014.25 × 10032.94 × 10034.57 × 10−80.00 × 10000.00 × 10001.28 × 10−90.00 × 1000
Contest1.94 × 10−9 (+)1.21 × 10−12 (+)1.21 × 10−12 (+)1.21 × 10−12 (+)N/A (=)N/A (=)1.21 × 10−12 (+)
F2(x)Best9.47 × 10−034.54 × 10001.01 × 10018.20 × 10−82.45 × 10−040.00 × 10001.45 × 10−021.71 × 10−07
Worst1.63 × 10019.26 × 10019.65 × 10017.08 × 10018.92 × 10001.28 × 10019.28 × 10007.32 × 10−02
Median1.62 × 10001.21 × 10013.11 × 10017.32 × 10008.92 × 10003.99 × 10006.70 × 10005.91 × 10−04
Mean2.83 × 10002.17 × 10014.00 × 10019.12 × 10005.44 × 10004.65 × 10004.81 × 10009.12 × 10−03
Standard3.58 × 10002.16 × 10012.28 × 10011.73 × 10014.06 × 10003.69 × 10004.33 × 10001.80 × 10−02
Contest9.92 × 10−11 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)3.59 × 10−07 (+)7.41 × 10−10 (+)7.13 × 10−05 (+)2.15 × 10−10 (+)
F3(x)Best0.00 × 10003.31 × 10005.74 × 10002.09 × 10−110.00 × 10000.00 × 10001.44 × 10−040.00 × 1000
Worst1.45 × 10−118.85 × 10003.02 × 10012.33 × 10−051.57 × 10−051.22 × 10−118.73 × 10008.79 × 10−05
Median5.53 × 10−035.38 × 10001.39 × 10013.06 × 10−80.00 × 10002.56 × 10−065.67 × 10−111.42 × 10−06
Mean2.40 × 10−025.75 × 10001.47 × 10018.73 × 10−071.30 × 10−065.53 × 10−031.74 × 10007.62 × 10−06
Standard3.41 × 10−021.55 × 10005.78 × 10004.25 × 10−063.55 × 10−062.24 × 10−022.65 × 10001.90 × 10−05
Contest6.48 × 10−11 (=)2.95 × 10−11 (+)2.95 × 10−11 (+)9.43 × 10−03 (−)6.38 × 10−05 (−)4.67 × 10−11 (=)2.95 × 10−11 (+)
F4(X)Best1.55 × 10002.40 × 10016.81 × 10001.99 × 10000.00 × 10002.98 × 10005.97 × 10002.98 × 1000
Worst1.60 × 10016.64 × 10013.70 × 10011.69 × 10016.96 × 10001.69 × 10012.59 × 10019.95 × 1000
Median8.46 × 10004.51 × 10011.88 × 10019.45 × 10001.99 × 10008.95 × 10001.69 × 10018.46 × 1000
Mean8.82 × 10004.44 × 10011.86 × 10019.72 × 10002.32 × 10009.09 × 10001.64 × 10017.63 × 1000
Standard3.45 × 10001.03 × 10016.25 × 10003.82 × 10001.60 × 10003.07 × 10005.12 × 10002.08 × 1000
Contest1.58 × 10−11 (=)2.93 × 10−11 (+)3.72 × 10−10 (+)1.94 × 10−03 (+)4.00 × 10−10 (−)1.64 × 10−11 (=)5.96 × 10−10 (+)
F5(X)Best0.00 × 10003.75 × 10004.85 × 10000.00 × 10000.00 × 10000.00 × 10008.43 × 10−100.00 × 1000
Worst2.69 × 10−039.28 × 10016.91 × 10024.54 × 10−114.54 × 10−118.95 × 10−022.82 × 10000.00 × 1000
Median0.00 × 10001.77 × 10011.97 × 10020.00 × 10000.00 × 10000.00 × 10001.53 × 10−90.00 × 1000
Mean2.77 × 10−042.58 × 10012.42 × 10025.44 × 10−022.41 × 10−022.98 × 10−033.99 × 10−110.00 × 1000
Standard7.17 × 10−042.10 × 10011.77 × 10021.38 × 10−118.57 × 10−021.63 × 10−026.90 × 10−110.00 × 1000
Contest2.16 × 10−04 (+)1.72 × 10−12 (+)1.72 × 10−12 (+)1.26 × 10−9 (+)1.31 × 10−03 (+)1.00 × 1000 (=)1.72 × 10−12 (+)
F6(X)Best5.43 × 10011.84 × 10031.39 × 10042.46 × 10019.30 × 10004.72 × 10001.22 × 10022.70 × 1000
Worst4.98 × 10027.29 × 10−146.85 × 10−141.96 × 10034.18 × 10031.37 × 10036.25 × 10031.96 × 1002
Median1.26 × 10027.69 × 10041.27 × 10−141.57 × 10025.23 × 10025.01 × 10011.82 × 10036.32 × 1001
Mean1.64 × 10021.24 × 10−141.81 × 10−143.70 × 10021.13 × 10031.47 × 10022.38 × 10037.83 × 1001
Standard1.11 × 10021.44 × 10−141.67 × 10055.25 × 10021.36 × 10032.79 × 10021.92 × 10036.02 × 1001
Contest3.01 × 10−04 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)4.23 × 10−03 (+)5.61 × 10−05 (+)6.10 × 10−11 (=)2.37 × 10−10 (+)
F7(X)Best6.35 × 10−112.07 × 10012.74 × 10016.44 × 10−052.28 × 10−022.81 × 10−111.30 × 10011.96 × 10−11
Worst2.73 × 10017.21 × 10017.86 × 10012.25 × 10012.79 × 10012.48 × 10015.04 × 10012.26 × 1001
Median1.83 × 10014.48 × 10015.22 × 10012.62 × 10007.69 × 10002.05 × 10013.00 × 10015.20 × 1000
Mean1.43 × 10014.73 × 10015.29 × 10018.33 × 10001.23 × 10011.43 × 10012.96 × 10019.88 × 1000
Standard1.10 × 10011.17 × 10011.27 × 10019.35 × 10001.04 × 10019.93 × 10008.52 × 10009.20 × 1000
Contest5.94 × 10−02 (=)6.70 × 10−11 (+)3.02 × 10−11 (+)5.69 × 10−11 (=)1.71 × 10−11 (=)2.34 × 10−11 (=)3.50 × 10−9 (+)
F8(X)Best3.80 × 10002.83 × 10012.63 × 10019.43 × 10−116.80 × 10001.63 × 10−113.10 × 10002.21 × 10−11
Worst2.61 × 10014.35 × 10014.45 × 10011.43 × 10022.15 × 10012.32 × 10012.78 × 10012.10 × 1001
Median1.98 × 10013.36 × 10013.23 × 10012.09 × 10012.04 × 10012.03 × 10012.33 × 10014.36 × 1000
Mean1.70 × 10013.47 × 10013.33 × 10012.23 × 10011.99 × 10011.43 × 10012.28 × 10016.25 × 1000
Standard8.23 × 10004.10 × 10004.77 × 10002.40 × 10012.69 × 10009.38 × 10004.99 × 10006.40 × 1000
Contest1.17 × 10−05 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)4.42 × 10−06 (+)2.57 × 10−07 (+)3.64 × 10−02 (+)5.07 × 10−10 (+)
F9(X)Best2.31 × 10022.31 × 10022.47 × 10022.29 × 10022.29 × 10022.30 × 10022.29 × 10022.29 × 1002
Worst2.35 × 10022.52 × 10025.27 × 10022.29 × 10022.29 × 10022.31 × 10022.29 × 10022.29 × 1002
Median2.32 × 10022.34 × 10023.77 × 10022.29 × 10022.29 × 10022.31 × 10022.29 × 10022.29 × 1002
Mean2.32 × 10022.35 × 10023.60 × 10022.29 × 10022.29 × 10022.31 × 10022.29 × 10022.29 × 1002
Standard1.07 × 10004.26 × 10006.49 × 10010.00 × 10000.00 × 10003.41 × 10−111.58 × 10−060.00 × 1000
Contest1.21 × 10−12 (+)1.21 × 10−12 (+)1.21 × 10−12 (+)N/A (=)N/A (=)1.21 × 10−12 (+)1.21 × 10−12 (+)
F10(X)Best1.00 × 10021.01 × 10021.01 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 1002
Worst2.18 × 10022.60 × 10023.15 × 10022.17 × 10022.13 × 10022.19 × 10022.32 × 10021.00 × 1002
Median1.02 × 10022.39 × 10021.03 × 10021.00 × 10021.01 × 10021.00 × 10021.00 × 10021.00 × 1002
Mean1.06 × 10022.26 × 10021.40 × 10021.29 × 10021.33 × 10021.22 × 10021.05 × 10021.00 × 1002
Standard2.13 × 10014.33 × 10017.33 × 10014.92 × 10015.04 × 10014.46 × 10012.41 × 10015.59 × 10−02
Contest5.53 × 10−8 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)1.69 × 10−9 (+)4.98E-1 (+)13.26 × 10−07 (+)2.20 × 10−07 (+)0.00 × 1000
F11(X)Best0.00 × 10006.70 × 10011.55 × 10025.78 × 10−110.00 × 10000.00 × 10004.29 × 10−040.00 × 1000
Worst1.89 × 10021.52 × 10025.27 × 10023.00 × 10023.00 × 10024.00 × 10024.00 × 10020.00 × 1000
Median0.00 × 10001.09 × 10021.98 × 10021.87 × 10001.50 × 10020.00 × 10005.48 × 10−040.00 × 1000
Mean2.41 × 10011.13 × 10022.26 × 10026.99 × 10011.26 × 10021.13 × 10026.39 × 10010.00 × 1000
Standard5.06 × 10013.00 × 10018.51 × 10019.28 × 10011.05 × 10021.51 × 10021.25 × 10020.00 × 1000
Contest5.70 × 10−03 (+)1.68 × 10−11 (+)1.68 × 10−11 (+)1.68 × 10−11 (+)7.30 × 10−04 (+)3.48 × 10−11 (=)1.68 × 10−11 (+)0.00 × 1000
F12(X)Best1.65 × 10021.61 × 10021.64 × 10021.61 × 10021.61 × 10021.63 × 10021.59 × 10021.65 × 1002
Worst1.70 × 10021.69 × 10022.26 × 10021.70 × 10021.74 × 10022.20 × 10021.65 × 10021.68 × 1002
Median1.67 × 10021.66 × 10021.71 × 10021.64 × 10021.66 × 10021.65 × 10021.61 × 10021.66 × 1002
Mean1.67 × 10021.66 × 10021.77 × 10021.64 × 10021.66 × 10021.67 × 10021.62 × 10021.66 × 1002
Standard1.07 × 10001.27 × 10001.54 × 10012.40 × 10002.38 × 10009.96 × 10001.58 × 10008.15 × 10−11
Contest4.22 × 10−04 (+)5.49 × 10−11 (=)5.46 × 10−9 (+)1.44 × 10−03 (−)4.12 × 10−11 (=)4.08 × 10−05 (+)4.98 × 10−11 (−)
+/=/−9/3/011/1/012/0/02008/2/22006/4/25/7/011/0/1
Rank3.836.837.673.833.833.674.421.92
Table 2. Twenty dimensions with PSO algorithm variants.
Table 2. Twenty dimensions with PSO algorithm variants.
FIndexAMSE
PSO
EAPSOFDBPSOMPSOPPSOPSOsonoVPPSORPSO
F1(x)Best0.00 × 10002.34 × 10037.81 × 10030.00 × 10000.00 × 10000.00 × 10001.02 × 10−80.00 × 1000
Worst1.87 × 10021.78 × 10043.39 × 10040.00 × 10007.59 × 10−110.00 × 10002.35 × 10−80.00 × 1000
Median4.17 × 10019.44 × 10031.88 × 10040.00 × 10000.00 × 10000.00 × 10001.59 × 10−80.00 × 1000
Mean5.48 × 10018.82 × 10031.88 × 10040.00 × 10002.53 × 10−020.00 × 10001.57 × 10−80.00 × 1000
Standard5.89 × 10014.09 × 10036.63 × 10030.00 × 10001.38 × 10−110.00 × 10002.94 × 10−90.00 × 1000
Contest3.98 × 10−8 (+)7.57 × 10−12 (+)7.57 × 10−12 (+)2.63 × 10−8 (+)7.46 × 10−12 (+)3.43 × 10−02 (+)7.57 × 10−12 (+)
F2(x)Best2.28 × 10014.91 × 10018.97 × 10012.66 × 10−112.39 × 10−82.32 × 10−042.47 × 10−026.67 × 1000
Worst7.85 × 10017.43 × 10014.98 × 10026.73 × 10016.66 × 10017.15 × 10017.07 × 10014.91 × 1001
Median5.94 × 10014.91 × 10011.62 × 10024.91 × 10014.91 × 10015.14 × 10014.70 × 10014.91 × 1001
Mean5.86 × 10015.68 × 10011.96 × 10024.45 × 10014.15 × 10015.09 × 10013.12 × 10014.54 × 1001
Standard1.07 × 10011.11 × 10011.05 × 10021.71 × 10011.88 × 10011.50 × 10012.55 × 10018.82 × 1000
Contest5.04 × 10−10 (+)3.00 × 10−11 (+)3.00 × 10−11 (+)8.19 × 10−11 (=)4.97 × 10−03 (−)8.44 × 10−9 (+)7.51 × 10−11 (=)
F3(x)Best0.00 × 10009.09 × 10−021.02 × 10010.00 × 10003.19 × 10−070.00 × 10006.42 × 10−117.22 × 10−04
Worst7.91 × 10−024.40 × 10−113.43 × 10019.10 × 10−025.53 × 10−035.94 × 10−111.85 × 10018.08 × 10−11
Median1.14 × 10−131.82 × 10−111.75 × 10010.00 × 10001.69 × 10−058.24 × 10−034.70 × 10006.66 × 10−02
Mean1.21 × 10−022.02 × 10−111.83 × 10013.03 × 10−034.48 × 10−041.02 × 10−117.11 × 10001.28 × 10−11
Standard2.06 × 10−028.29 × 10−026.23 × 10001.66 × 10−021.21 × 10−031.68 × 10−116.15 × 10001.71 × 10−11
Contest1.17 × 10−06 (−)9.52 × 10−04 (+)3.02 × 10−11 (+)1.37 × 10−11 (−)5.46 × 10−11 (−)2.71 × 10−02 (−)4.08 × 10−11 (+)
F4(X)Best8.33 × 10001.81 × 10013.08 × 10011.69 × 10015.97 × 10005.97 × 10003.28 × 10016.96 × 1000
Worst6.07 × 10011.14 × 10026.69 × 10015.27 × 10011.49 × 10014.38 × 10019.15 × 10011.99 × 1001
Median3.89 × 10014.33 × 10014.98 × 10013.34 × 10018.95 × 10001.99 × 10014.97 × 10011.69 × 1001
Mean4.01 × 10014.71 × 10015.00 × 10013.20 × 10019.42 × 10002.07 × 10015.38 × 10011.62 × 1001
Standard1.08 × 10011.75 × 10019.60 × 10001.08 × 10012.66 × 10008.06 × 10001.54 × 10013.27 × 1000
Contest4.48 × 10−10 (+)5.31 × 10−11 (+)2.92 × 10−11 (+)5.55 × 10−10 (+)1.91 × 10−8 (−)2.78 × 10−02 (+)2.92 × 10−11 (+)
F5(X)Best1.77 × 10−072.37 × 10−025.38 × 10020.00 × 10000.00 × 10000.00 × 10003.51 × 10−90.00 × 1000
Worst1.56 × 10002.64 × 10002.93 × 10033.34 × 10013.97 × 10001.29 × 10018.87 × 10020.00 × 1000
Median5.08 × 10−112.00 × 10−111.33 × 10032.24 × 10004.48 × 10−021.14 × 10−139.09 × 10−110.00 × 1000
Mean6.13 × 10−114.13 × 10−111.41 × 10035.77 × 10004.27 × 10−111.11 × 10003.98 × 10010.00 × 1000
Standard4.49 × 10−115.54 × 10−115.85 × 10028.84 × 10008.98 × 10−112.89 × 10001.62 × 10020.00 × 1000
Contest1.40 × 10−11 (+)1.40 × 10−11 (+)1.40 × 10−11 (+)7.34 × 10−11 (+)4.94 × 10−11 (=)8.70 × 10−11 (=)1.40 × 10−11 (+)
F6(X)Best1.86 × 10022.53 × 10022.78 × 10−143.82 × 10016.54 × 10012.16 × 10011.44 × 10025.55 × 1001
Worst3.29 × 10032.17 × 10046.58E+084.52 × 10036.74 × 10037.07 × 10032.30 × 10042.47 × 1002
Median8.08 × 10023.82 × 10032.79E+067.34 × 10021.49 × 10033.14 × 10021.95 × 10031.26 × 1002
Mean1.03 × 10035.20 × 10034.94E+071.29 × 10032.01 × 10031.65 × 10033.46 × 10031.30 × 1002
Standard8.26 × 10025.68 × 10031.41E+081.21 × 10031.76 × 10032.31 × 10035.12 × 10035.08 × 1001
Contest6.70 × 10−11 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)9.26 × 10−9 (+)7.77 × 10−9 (+)2.07 × 10−02 (+)1.17 × 10−9 (+)
F7(X)Best2.13 × 10011.82 × 10015.92 × 10012.19 × 10012.31 × 10012.63 × 10012.20 × 10012.11 × 1001
Worst5.24 × 10019.03 × 10012.05 × 10026.74 × 10017.53 × 10015.35 × 10018.82 × 10013.00 × 1001
Median3.82 × 10014.03 × 10011.14 × 10023.70 × 10013.87 × 10013.28 × 10013.74 × 10012.66 × 1001
Mean3.67 × 10014.49 × 10011.20 × 10023.81 × 10014.16 × 10013.58 × 10014.16 × 10012.58 × 1001
Standard7.47 × 10001.79 × 10013.97 × 10011.24 × 10011.31 × 10017.73 × 10001.51 × 10012.75 × 1000
Contest3.01 × 10−07 (−)6.01 × 10−8 (+)3.02 × 10−11 (+)1.53 × 10−05 (+)5.00 × 10−9 (+)1.31 × 10−8 (+)2.60 × 10−8 (+)
F8(X)Best2.16 × 10012.18 × 10012.96 × 10012.04 × 10012.04 × 10012.06 × 10012.34 × 10011.90 × 1001
Worst3.17 × 10012.62 × 10023.25 × 10021.21 × 10021.42 × 10022.51 × 10014.13 × 10012.91 × 1001
Median2.70 × 10011.44 × 10023.88 × 10012.12 × 10012.13 × 10012.18 × 10012.62 × 10012.40 × 1001
Mean2.70 × 10011.07 × 10027.49 × 10012.47 × 10012.92 × 10012.21 × 10012.69 × 10012.46 × 1001
Standard2.36 × 10007.85 × 10018.19 × 10011.82 × 10013.04 × 10011.08 × 10003.31 × 10002.60 × 1000
Contest7.70 × 10−04 (+)1.29 × 10−06 (+)3.02 × 10−11 (+)6.05 × 10−07 (+)3.83 × 10−06 (+)4.08 × 10−05 (−)2.89 × 10−03 (+)
F9(X)Best1.83 × 10021.81 × 10022.35 × 10021.81 × 10021.81 × 10021.83 × 10021.81 × 10021.82 × 1002
Worst1.84 × 10021.81 × 10025.39 × 10021.81 × 10021.81 × 10021.89 × 10021.81 × 10021.85 × 1002
Median1.84 × 10021.81 × 10023.90 × 10021.81 × 10021.81 × 10021.85 × 10021.81 × 10021.84 × 1002
Mean1.84 × 10021.81 × 10023.83 × 10021.81 × 10021.81 × 10021.85 × 10021.81 × 10021.84 × 1002
Standard4.88 × 10−111.60 × 10−027.32 × 10010.00 × 10004.51 × 10−051.67 × 10002.48 × 10−057.37 × 10−11
Contest4.20 × 10−11 (=)3.02 × 10−11 (−)3.02 × 10−11 (+)2.36 × 10−12 (−)3.01 × 10−11 (−)5.32 × 10−03 (+)3.02 × 10−11 (−)
F10(X)Best6.60 × 10012.16 × 10021.01 × 10021.27 × 10011.00 × 10021.00 × 10021.00 × 10021.00 × 1002
Worst1.24 × 10022.61 × 10022.80 × 10034.98 × 10028.76 × 10022.78 × 10021.71 × 10031.01 × 1002
Median1.06 × 10022.39 × 10021.13 × 10021.01 × 10021.01 × 10021.00 × 10021.00 × 10021.00 × 1002
Mean1.06 × 10022.39 × 10024.53 × 10021.80 × 10021.87 × 10021.35 × 10021.85 × 10021.00 × 1002
Standard1.01 × 10011.10 × 10017.57 × 10021.21 × 10021.76 × 10026.39 × 10012.97 × 10028.67 × 10−02
Contest4.80 × 10−07 (+)3.02 × 10−11 (+)3.02 × 10−11 (+)2.43 × 10−05 (+)5.00 × 10−9 (+)1.05 × 10−11 (=)7.06 × 10−11 (=)
F11(X)Best5.60 × 10−113.01 × 10028.19 × 10023.00 × 10023.00 × 10020.00 × 10002.67 × 10−030.00 × 1000
Worst5.43 × 10024.02 × 10024.10 × 10039.06 × 10024.00 × 10024.00 × 10021.76 × 10033.00 × 1002
Median3.00 × 10023.02 × 10021.60 × 10033.00 × 10023.00 × 10023.00 × 10023.00 × 10023.00 × 1002
Mean3.21 × 10023.09 × 10021.91 × 10034.29 × 10023.10 × 10023.13 × 10023.86 × 10022.80 × 1002
Standard1.23 × 10022.52 × 10019.73 × 10022.09 × 10023.05 × 10017.30 × 10013.32 × 10027.61 × 1001
Contest5.21 × 10−04 (+)4.11 × 10−12 (+)4.11 × 10−12 (+)4.08 × 10−12 (+)2.38 × 10−11 (=)5.64 × 10−02 (=)1.28 × 10−9 (+)
F12(X)Best2.42 × 10022.36 × 10022.65 × 10022.34 × 10022.37 × 10022.47 × 10022.36 × 10022.53 × 1002
Worst3.00 × 10022.73 × 10025.30 × 10022.95 × 10022.76 × 10023.50 × 10025.03 × 10022.70 × 1002
Median2.69 × 10022.44 × 10023.19 × 10022.50 × 10022.53 × 10022.75 × 10022.52 × 10022.62 × 1002
Mean2.69 × 10022.48 × 10023.39 × 10022.53 × 10022.55 × 10022.77 × 10022.67 × 10022.62 × 1002
Standard1.22 × 10019.10 × 10006.34 × 10011.35 × 10011.10 × 10012.22 × 10014.92 × 10013.94 × 1000
Contest3.85 × 10−03 (+)3.65 × 10−8 (−)7.39 × 10−11 (+)7.70 × 10−04 (−)1.52 × 10−03 (−)4.98 × 10−04 (+)1.70 × 10−02 (+)
+/=/−2009/1/210/0/212/0/02008/1/32005/2/52007/3/22009/2/1
Rank4.425.127.833.503.583.925.002.58
Table 3. Ten dimensions with other algorithms.
Table 3. Ten dimensions with other algorithms.
FIndexAPGSK−IMODEEA4eigIMODEPVADEAGSKRPSO
F1(x)Best0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Worst0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Median0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Mean0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Standard0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
ContestN/A (=)N/A (=)N/A (=)N/A (=)N/A (=)
F2(x)Best0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10001.71 × 10−07
Worst0.00 × 10003.99 × 10000.00 × 10008.92 × 10008.92 × 10007.32 × 10−02
Median0.00 × 10000.00 × 10000.00 × 10008.92 × 10003.99 × 10005.91 × 10−04
Mean0.00 × 10006.64 × 10−110.00 × 10005.55 × 10003.09 × 10009.12 × 10−03
Standard0.00 × 10001.51 × 10000.00 × 10003.91 × 10002.09 × 10001.80 × 10−02
Contest1.21 × 10−12 (−)4.70 × 10−06 (+)1.21 × 10−12 (−)1.74 × 10−03 (+)1.53 × 10−03 (+)
F3(x)Best0.00 × 10000.00 × 10002.69 × 10−070.00 × 10000.00 × 10000.00 × 1000
Worst0.00 × 10000.00 × 10004.86 × 10−054.85 × 10−052.08 × 10−058.79 × 10−05
Median0.00 × 10000.00 × 10001.73 × 10−060.00 × 10001.42 × 10−061.42 × 10−06
Mean0.00 × 10000.00 × 10005.93 × 10−061.93 × 10−062.75 × 10−067.62 × 10−06
Standard0.00 × 10000.00 × 10009.55 × 10−068.84 × 10−064.75 × 10−061.90 × 10−05
Contest9.65 × 10−11 (−)5.63 × 10−11 (−)8.75 × 10−02 (=)1.98 × 10−06 (−)1.62 × 10−11 (=)
F4(X)Best9.96 × 10−110.00 × 10004.97 × 10002.72 × 10003.02 × 10002.98 × 1000
Worst6.96 × 10002.98 × 10001.29 × 10018.68 × 10001.13 × 10019.95 × 1000
Median4.97 × 10009.95 × 10−118.95 × 10005.48 × 10007.97 × 10008.46 × 1000
Mean4.61 × 10009.62 × 10−119.02 × 10005.45 × 10007.63 × 10007.63 × 1000
Standard1.21 × 10009.23 × 10−112.20 × 10001.65 × 10002.14 × 10002.08 × 1000
Contest7.86 × 10−06 (−)2.39 × 10−11 (−)2.06 × 10−02 (+)9.69 × 10−05 (−)9.06 × 10−11 (=)
F5(X)Best0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Worst0.00 × 10000.00 × 10001.68 × 10010.00 × 10001.14 × 10−130.00 × 1000
Median0.00 × 10000.00 × 10006.33 × 10−110.00 × 10000.00 × 10000.00 × 1000
Mean0.00 × 10000.00 × 10001.83 × 10000.00 × 10007.58 × 10−150.00 × 1000
Standard0.00 × 10000.00 × 10003.20 × 10000.00 × 10002.88 × 10−140.00 × 1000
ContestN/A (=)N/A (=)6.93 × 10−12 (+)N/A (=)5.70 × 10−11 (=)
F6(X)Best6.55 × 10−021.63 × 10−039.90 × 10−022.67 × 10−025.02 × 10−032.70 × 1000
Worst4.49 × 10−114.28 × 10−114.99 × 10001.70 × 10014.75 × 10−111.96 × 1002
Median1.86 × 10−111.85 × 10−021.18 × 10003.72 × 10−112.47 × 10−116.32 × 1001
Mean2.30 × 10−115.70 × 10−021.65 × 10001.48 × 10002.50 × 10−117.83 × 1001
Standard1.17 × 10−119.46 × 10−021.29 × 10003.57 × 10001.32 × 10−116.02 × 1001
Contest3.01 × 10−11 (−)3.02 × 10−11 (−)4.98 × 10−11 (−)6.70 × 10−11 (−)3.02 × 10−11 (−)
F7(X)Best0.00 × 10000.00 × 10002.65 × 10−050.00 × 10000.00 × 10001.96 × 10−11
Worst1.94 × 10−073.76 × 10−91.24 × 10−032.10 × 10012.09 × 10−102.26 × 1001
Median0.00 × 10005.68 × 10−131.71 × 10−043.12 × 10−114.55 × 10−135.20 × 1000
Mean6.47 × 10−92.77 × 10−103.05 × 10−043.76 × 10001.83 × 10−119.88 × 1000
Standard3.54 × 10−88.53 × 10−103.52 × 10−047.67 × 10005.06 × 10−119.20 × 1000
Contest1.65 × 10−11 (−)2.71 × 10−11 (−)3.02 × 10−11 (−)7.49 × 10−06 (−)2.70 × 10−11 (−)
F8(X)Best2.17 × 10−021.39 × 10−034.30 × 10−027.44 × 10−111.19 × 10−112.21 × 10−11
Worst2.64 × 10008.21 × 10−118.09 × 10002.16 × 10011.61 × 10002.10 × 1001
Median1.99 × 10−113.37 × 10−111.42 × 10001.17 × 10006.47 × 10−114.36 × 1000
Mean3.68 × 10−113.24 × 10−112.72 × 10004.35 × 10007.84 × 10−116.25 × 1000
Standard5.11 × 10−112.21 × 10−112.34 × 10007.15 × 10004.37 × 10−116.40 × 1000
Contest2.66 × 10−9 (−)2.44 × 10−9 (−)2.32 × 10−02 (−)4.06 × 10−02 (−)1.03 × 10−06 (−)
F9(X)Best2.29 × 10021.86 × 10021.37 × 10−062.29 × 10022.29 × 10022.29 × 1002
Worst2.29 × 10021.86 × 10022.29 × 10022.29 × 10022.29 × 10022.29 × 1002
Median2.29 × 10021.86 × 10022.29 × 10022.29 × 10022.29 × 10022.29 × 1002
Mean2.29 × 10021.86 × 10022.14 × 10022.29 × 10022.29 × 10022.29 × 1002
Standard0.00 × 10004.22 × 10−135.82 × 10010.00 × 10008.67 × 10−140.00 × 1000
ContestN/A (=)4.16 × 10−14 (−)7.64 × 10−9 (−)1.69 × 10−14 (−)N/A (=)
F10(X)Best6.72 × 10011.00 × 10023.75 × 10001.00 × 10021.00 × 10021.00 × 1002
Worst1.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 1002
Median1.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 1002
Mean9.91 × 10011.00 × 10028.35 × 10011.00 × 10021.00 × 10021.00 × 1002
Standard6.02 × 10003.69 × 10−023.45 × 10013.82 × 10−023.49 × 10−025.59 × 10−02
Contest9.47 × 10−11 (=)1.24 × 10−03 (−)1.63 × 10−02 (−)6.10 × 10−03 (−)4.20 × 10−11 (=)
F11(X)Best0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Worst4.55 × 10−130.00 × 10002.61 × 10−073.00 × 10024.55 × 10−130.00 × 1000
Median0.00 × 10000.00 × 10002.96 × 10−80.00 × 10004.55 × 10−130.00 × 1000
Mean1.36 × 10−130.00 × 10005.19 × 10−81.00 × 10013.49 × 10−130.00 × 1000
Standard2.12 × 10−130.00 × 10006.11 × 10−85.48 × 10011.96 × 10−130.00 × 1000
Contest5.19 × 10−02 (=)N/A (=)6.92 × 10−8 (+)4.76 × 10−05 (+)1.56 × 10−11 (=)
F12(X)Best1.59 × 10021.45 × 10021.63 × 10021.59 × 10021.59 × 10021.65 × 1002
Worst1.63 × 10021.59 × 10021.65 × 10021.65 × 10021.61 × 10021.68 × 1002
Median1.62 × 10021.46 × 10021.64 × 10021.63 × 10021.60 × 10021.66 × 1002
Mean1.61 × 10021.47 × 10021.64 × 10021.63 × 10021.60 × 10021.66 × 1002
Standard1.54 × 10003.55 × 10006.54 × 10−112.08 × 10009.91 × 10−118.15 × 10−11
Contest2.52 × 10−11 (−)3.02 × 10−11 (−)4.08 × 10−11 (−)9.36 × 10−11 (−)2.56 × 10−11 (−)
+/=/−2000/5/72001/3/82003/2/72002/2/82001/7/4
Rank2.541.964.004.123.834.50
Table 4. Twenty dimensions with other algorithms.
Table 4. Twenty dimensions with other algorithms.
FIndexAPGSKIMODEEA4eigIMODEPVADEAGSKRPSO
F1(x)Best0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Worst5.68 × 10−140.00 × 10000.00 × 10000.00 × 10005.68 × 10−140.00 × 1000
Median0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
Mean7.58 × 10−150.00 × 10000.00 × 10000.00 × 10001.89 × 10−140.00 × 1000
Standard1.97 × 10−140.00 × 10000.00 × 10000.00 × 10002.73 × 10−140.00 × 1000
ContestN/A (=)N/A (=)N/A (=)N/A (=)N/A (=)
F2(x)Best3.63 × 10−060.00 × 10000.00 × 10000.00 × 10000.00 × 10006.67 × 1000
Worst4.91 × 10013.99 × 10004.91 × 10018.92 × 10004.91 × 10014.91 × 1001
Median4.91 × 10010.00 × 10004.91 × 10013.99 × 10004.91 × 10014.91 × 1001
Mean4.47 × 10017.97 × 10−113.45 × 10015.36 × 10002.92 × 10014.54 × 1001
Standard1.18 × 10011.62 × 10002.27 × 10013.69 × 10002.42 × 10018.82 × 1000
Contest1.20 × 10−03 (−)6.27 × 10−12 (−)3.14 × 10−11 (=)8.20 × 10−11 (−)2.95 × 10−05 (−)
F3(x)Best0.00 × 10000.00 × 10002.11 × 10−110.00 × 10000.00 × 10003.49 × 10−03
Worst1.14 × 10−131.14 × 10−133.49 × 10001.42 × 10−068.36 × 10−114.56 × 1000
Median1.14 × 10−131.14 × 10−131.03 × 10000.00 × 10003.41 × 10−131.61 × 1000
Mean1.06 × 10−137.96 × 10−141.08 × 10007.11 × 10−83.56 × 10−121.75 × 1000
Standard2.88 × 10−145.30 × 10−146.96 × 10−112.70 × 10−071.52 × 10−111.30 × 1000
Contest2.36 × 10−12 (−)1.01 × 10−11 (−)2.37 × 10−10 (−)3.16 × 10−12 (−)2.89 × 10−11 (−)
F4(X)Best1.29 × 10013.98 × 10003.68 × 10010.00 × 10001.98 × 10016.96 × 1000
Worst2.98 × 10011.69 × 10018.36 × 10015.97 × 10004.58 × 10011.99 × 1001
Median2.19 × 10018.95 × 10005.62 × 10011.99 × 10003.72 × 10011.69 × 1001
Mean2.23 × 10019.22 × 10005.86 × 10012.46 × 10003.63 × 10011.62 × 1001
Standard4.51 × 10003.49 × 10001.23 × 10011.32 × 10006.85 × 10003.27 × 1000
Contest2.58 × 10−06 (+)4.59 × 10−8 (−)2.92 × 10−11 (+)2.58 × 10−11 (−)3.56 × 10−11 (+)
F5(X)Best0.00 × 10000.00 × 10002.61 × 10020.00 × 10001.14 × 10−130.00 × 1000
Worst0.00 × 10000.00 × 10001.03 × 10030.00 × 10005.44 × 10−110.00 × 1000
Median0.00 × 10000.00 × 10006.72 × 10020.00 × 10008.95 × 10−020.00 × 1000
Mean0.00 × 10000.00 × 10006.82 × 10020.00 × 10007.48 × 10−020.00 × 1000
Standard0.00 × 10000.00 × 10002.08 × 10020.00 × 10001.11 × 10−110.00 × 1000
ContestN/A (=)N/A (=)1.40 × 10−11 (+)N/A (=)1.53 × 10−9 (+)
F6(X)Best1.04 × 10002.18 × 10−029.17 × 10002.20 × 10−022.18 × 10−115.55 × 1001
Worst2.48 × 10012.08 × 10004.54 × 10011.12 × 10001.90 × 10002.47 × 1002
Median6.91 × 10001.20 × 10−112.19 × 10011.25 × 10−113.95 × 10−111.26 × 1002
Mean9.11 × 10002.13 × 10−112.36 × 10013.23 × 10−116.36 × 10−111.30 × 1002
Standard6.77 × 10003.69 × 10−118.37 × 10003.73 × 10−115.27 × 10−115.08 × 1001
Contest3.01 × 10−11 (−)3.02 × 10−11 (−)3.02 × 10−11 (−)3.02 × 10−11 (−)3.02 × 10−11 (−)
F7(X)Best2.51 × 10−020.00 × 10002.51 × 10010.00 × 10003.12 × 10−112.11 × 1001
Worst2.10 × 10012.10 × 10014.48 × 10012.10 × 10012.25 × 10013.00 × 1001
Median3.86 × 10001.80 × 10003.36 × 10010.00 × 10002.11 × 10012.66 × 1001
Mean7.00 × 10004.23 × 10003.36 × 10013.68 × 10001.82 × 10012.58 × 1001
Standard7.08 × 10006.18 × 10005.16 × 10007.62 × 10006.09 × 10002.75 × 1000
Contest3.01 × 10−11 (−)3.00 × 10−11 (−)2.39 × 10−8 (−)1.78 × 10−11 (−)8.89 × 10−10 (−)
F8(X)Best1.41 × 10012.69 × 10−112.08 × 10015.64 × 10−051.61 × 10011.90 × 1001
Worst2.09 × 10012.07 × 10012.33 × 10012.04 × 10012.29 × 10012.91 × 1001
Median2.04 × 10012.02 × 10012.18 × 10011.06 × 10−112.21 × 10012.40 × 1001
Mean2.01 × 10011.77 × 10012.19 × 10011.69 × 10002.20 × 10012.46 × 1001
Standard1.44 × 10006.14 × 10006.63 × 10−115.16 × 10001.16 × 10002.60 × 1000
Contest4.61 × 10−10 (−)3.16 × 10−10 (−)3.83 × 10−06 (−)3.69 × 10−11 (−)1.02 × 10−05 (−)
F9(X)Best1.81 × 10021.65 × 10021.81 × 10022.29 × 10021.81 × 10021.82 × 1002
Worst1.81 × 10021.65 × 10021.81 × 10022.29 × 10021.81 × 10021.85 × 1002
Median1.81 × 10021.65 × 10021.81 × 10022.29 × 10021.81 × 10021.84 × 1002
Mean1.81 × 10021.65 × 10021.81 × 10022.29 × 10021.81 × 10021.84 × 1002
Standard8.67 × 10−140.00 × 10001.18 × 10−070.00 × 10008.67 × 10−147.37 × 10−11
Contest1.21 × 10−12 (−)1.21 × 10−12 (−)3.02 × 10−11 (−)1.21 × 10−12 (+)1.21 × 10−12 (−)
F10(X)Best1.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 10021.00 × 1002
Worst1.00 × 10022.24 × 10021.01 × 10022.00 × 10021.00 × 10021.01 × 1002
Median1.00 × 10021.00 × 10021.01 × 10021.00 × 10021.00 × 10021.00 × 1002
Mean1.00 × 10021.08 × 10021.01 × 10021.03 × 10021.00 × 10021.00 × 1002
Standard3.76 × 10−023.06 × 10017.66 × 10−021.82 × 10012.85 × 10−028.67 × 10−02
Contest9.21 × 10−05 (−)1.01 × 10−8 (−)6.53 × 10−8 (+)5.57 × 10−10 (−)8.99 × 10−11 (−)
F11(X)Best0.00 × 10003.00 × 10023.35 × 10−040.00 × 10003.00 × 10020.00 × 1000
Worst3.00 × 10024.00 × 10023.22 × 10021.50 × 10024.00 × 10023.00 × 1002
Median3.00 × 10023.00 × 10023.00 × 10020.00 × 10004.00 × 10023.00 × 1002
Mean2.50 × 10023.23 × 10022.81 × 10025.01 × 10003.87 × 10022.80 × 1002
Standard1.14 × 10024.30 × 10017.64 × 10012.75 × 10013.46 × 10017.61 × 1001
Contest1.08 × 10−11 (=)2.18 × 10−02 (+)1.28 × 10−9 (+)6.71 × 10−13 (−)8.54 × 10−12 (+)
F12(X)Best2.31 × 10021.89 × 10022.40 × 10021.60 × 10022.31 × 10022.53 × 1002
Worst2.37 × 10022.00 × 10022.67 × 10021.65 × 10022.39 × 10022.70 × 1002
Median2.33 × 10022.00 × 10022.53 × 10021.64 × 10022.34 × 10022.62 × 1002
Mean2.34 × 10022.00 × 10022.53 × 10021.63 × 10022.34 × 10022.62 × 1002
Standard1.52 × 10002.05 × 10007.63 × 10001.51 × 10002.07 × 10003.94 × 1000
Contest3.02 × 10−11 (−)3.02 × 10−11 (−)7.74 × 10−06 (−)2.89 × 10−11 (−)2.97 × 10−11 (−)
+/=/−2001/3/82001/2/92004/2/62001/2/92003/1/8
Rank3.212.334.542.424.004.50
Table 5. Data from ablation experiments.
Table 5. Data from ablation experiments.
FIndexBestWorstMedianMeanStandard
F1(x)PSO10.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
PSO20.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
PSO30.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
RPSO0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
F2(x)PSO10.00 × 10001.08 × 10013.99 × 10005.15 × 10004.29 × 1000
PSO20.00 × 10001.31 × 10016.33 × 10005.97 × 10004.09 × 1000
PSO30.00 × 10001.04 × 10013.99 × 10003.46 × 10003.53 × 1000
RPSO1.71 × 10−077.32 × 10−025.91 × 10−049.12 × 10−031.80 × 10−02
F3(x)PSO10.00 × 10002.31 × 10−021.26 × 10−042.80 × 10−036.28 × 10−03
PSO21.06E107.47 × 10−026.67 × 10−049.25 × 10−031.89 × 10−02
PSO30.00 × 10001.91 × 10−029.10 × 10−053.05 × 10−035.55 × 10−03
RPSO0.00 × 10008.79 × 10−051.42 × 10−067.62 × 10−061.90 × 10−05
F4(X)PSO12.98 × 10002.09 × 10016.96 × 10008.76 × 10004.74 × 1000
PSO29.95 × 10−011.49 × 10015.97 × 10007.23 × 10003.74 × 1000
PSO39.95 × 10−011.69 × 10017.46 × 10008.13 × 10004.27 × 1000
RPSO2.98 × 10009.95 × 10008.46 × 10007.63 × 10002.08 × 1000
F5(X)PSO10.00 × 10004.54 × 10−010.00 × 10002.11 × 10−028.49 × 10−02
PSO20.00 × 10008.95 × 10−020.00 × 10002.98 × 10−031.63 × 10−02
PSO30.00 × 10001.79 × 10−010.00 × 10002.09 × 10−024.51 × 10−02
RPSO0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
F6(X)PSO12.95 × 10001.72 × 10−034.73 × 10012.26 × 10024.03 × 1002
PSO21.28 × 10002.55 × 10−033.93 × 10012.10 × 10024.89 × 1002
PSO31.52 × 10001.45 × 10−034.53 × 10011.90 × 10023.35 × 1002
RPSO2.70 × 10001.96 × 10026.32 × 10017.83 × 10016.02 × 1001
F7(X)PSO12.69 × 10−082.55 × 10011.28 × 10011.20 × 10019.76 × 1000
PSO26.61 × 10−052.30 × 10015.23 × 10001.05 × 10019.26 × 1000
PSO38.44E092.56 × 10011.28 × 10011.19 × 10011.02 × 1001
RPSO1.96 × 10−012.26 × 10015.20 × 10009.88 × 10009.20 × 1000
F8(X)PSO11.76 × 10−012.47 × 10012.06 × 10011.43 × 10019.86 × 1000
PSO28.83 × 10−012.29 × 10012.02 × 10011.29 × 10019.29 × 1000
PSO32.84 × 10−012.42 × 10012.10 × 10011.36 × 10019.57 × 1000
RPSO2.21 × 10−012.10 × 10014.36 × 10006.25 × 10006.40 × 1000
F9(X)PSO12.30 × 10023.76 × 10022.32 × 10022.36 × 10022.64 × 1001
PSO22.30 × 10022.32 × 10022.31 × 10022.31 × 10025.53 × 10−01
PSO32.30 × 10022.32 × 10022.31 × 10022.31 × 10024.77 × 10−01
RPSO2.29 × 10022.29 × 10022.29 × 10022.29 × 10020.00 × 1000
F10(X)PSO11.00 × 10021.22 × 10021.00 × 10021.04 × 10026.10 × 1000
PSO21.00 × 10022.20 × 10021.00 × 10021.49 × 10025.70 × 1001
PSO31.00 × 10022.20 × 10021.00 × 10021.45 × 10025.55 × 1001
RPSO1.00 × 10021.00 × 10021.00 × 10021.00 × 10025.59 × 10−02
F11(X)PSO10.00 × 10003.00 × 10020.00 × 10004.01 × 10018.76 × 1001
PSO20.00 × 10004.00 × 10020.00 × 10008.22 × 10011.20 × 1002
PSO30.00 × 10004.00 × 10020.00 × 10004.34 × 10019.08 × 1001
RPSO0.00 × 10000.00 × 10000.00 × 10000.00 × 10000.00 × 1000
F12(X)PSO11.65 × 10021.67 × 10021.66 × 10021.66 × 10026.60 × 10−01
PSO21.64 × 10021.67 × 10021.65 × 10021.66 × 10027.50 × 10−01
PSO31.65 × 10021.68 × 10021.65 × 10021.66 × 10028.05 × 10−01
RPSO1.65 × 10021.68 × 10021.66 × 10021.66 × 10028.15 × 10−01
Table 6. Classification results for each algorithm.
Table 6. Classification results for each algorithm.
MethodDatasetAccuracyPrecisionRecallF1 ScoreBestMeanWorst
PSOXY98.21%98.82%97.67%98.25%98.21%97.86%97.02%
CT87.05%91.30%80.00%85.28%87.05%84.38%83.04%
GWOXY99.40%100.00%98.84%99.42%99.40%98.10%96.43%
CT86.16%84.91%85.71%85.31%86.16%84.91%83.93%
WOAXY98.81%98.84%98.84%98.84%98.81%97.62%95.83%
CT86.61%84.40%87.62%85.98%86.61%84.91%83.48%
DEPSOXY98.81%98.84%98.84%98.84%98.81%98.45%97.62%
CT83.93%84.85%80.00%82.35%83.93%82.59%79.91%
ASSOAXY98.81%100.00%97.67%98.82%98.81%97.86%96.43%
CT85.71%85.44%83.81%84.62%85.71%84.73%83.93%
RPSOXY100%100%100%100%100%98.69%97.02%
CT87.05%88.00%83.81%85.85%87.05%85.71%84.82%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Zeng, Y.; Li, R.; Zhu, X.; Zhang, Y.; Li, W.; Li, T.; Zhu, D.; Hu, G. A Random Particle Swarm Optimization Based on Cosine Similarity for Global Optimization and Classification Problems. Biomimetics 2024, 9, 204. https://doi.org/10.3390/biomimetics9040204

AMA Style

Liu Y, Zeng Y, Li R, Zhu X, Zhang Y, Li W, Li T, Zhu D, Hu G. A Random Particle Swarm Optimization Based on Cosine Similarity for Global Optimization and Classification Problems. Biomimetics. 2024; 9(4):204. https://doi.org/10.3390/biomimetics9040204

Chicago/Turabian Style

Liu, Yujia, Yuan Zeng, Rui Li, Xingyun Zhu, Yuemai Zhang, Weijie Li, Taiyong Li, Donglin Zhu, and Gangqiang Hu. 2024. "A Random Particle Swarm Optimization Based on Cosine Similarity for Global Optimization and Classification Problems" Biomimetics 9, no. 4: 204. https://doi.org/10.3390/biomimetics9040204

Article Metrics

Back to TopTop