Next Article in Journal
Enhancing Security and Authenticity in Immersive Environments
Previous Article in Journal
From Misinformation to Insight: Machine Learning Strategies for Fake News Detection
Previous Article in Special Issue
Leveraging Machine Learning to Analyze Semantic User Interactions in Visual Analytics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bridging the Gap Between Theory and Practice: Fitness Landscape Analysis of Real-World Problems with Nearest-Better Network

1
School of Automation, China University of Geosciences, Wuhan 430074, China
2
The Hubei Key Laboratory of Advanced Control and Intelligent Automation for Complex Systems, Wuhan 430074, China
3
The Engineering Research Center of Intelligent Technology for Geo-Exploration, Ministry of Education, Wuhan 430074, China
4
School of Artificial Intelligence, Anhui University of Science & Technology, Hefei 232001, China
5
School of Mechanical Engineering and Electronic Information, China University of Geosciences, Wuhan 430074, China
6
School of Computer Science and Informatics, De Montfort University, Leicester LE1 9BH, UK
*
Author to whom correspondence should be addressed.
Information 2025, 16(3), 190; https://doi.org/10.3390/info16030190
Submission received: 14 January 2025 / Revised: 1 February 2025 / Accepted: 3 February 2025 / Published: 1 March 2025
(This article belongs to the Special Issue Information Visualization Theory and Applications)

Abstract

:
For a long time, there has been a gap between theoretical optimization research and real-world applications. A key challenge is that many real-world problems are black-box problems, making it difficult to identify their characteristics and, consequently, select the most effective algorithms to solve them. Fortunately, the Nearest-Better Network has emerged as an effective tool for analyzing the characteristics of problems, regardless of dimensionality. In this paper, we conduct an in-depth experimental analysis of real-world functions from the CEC 2022 and CEC 2011 competitions using the NBN. Our experiments reveal that real-world problems often exhibit characteristics such as unclear global structure, multiple attraction basins, vast neutral regions around the global optimum, and high levels of ill conditioning.

1. Introduction

For a long time, there has been a gap between theoretical research and research oriented towards the real world [1]. Sometimes, a winning algorithm may perform poorly in real-world applications. Each year, numerous optimization algorithms are proposed, and these algorithms can outperform others on certain problems. This can perhaps be explained by the No Free Lunch (NFL) theorem [2]. This theorem essentially states that there is no single metaheuristic that universally outperforms all others across all types of problems. It seems that we need to design different algorithms for different optimization problems. As Marti [3] points out, this theorem emphasizes the significance of understanding problem-specific characteristics and selecting or designing appropriate metaheuristics based on these characteristics. There are numerous optimization problems worldwide, and it is nearly impossible to design an algorithm for each one. What we truly need to focus on is analyzing the characteristics and difficulties of the problems.
Yet, most real-world problems are high-dimensional black-box problems. The characteristics of these problems can be very complex and difficult to analyze. We lack an effective method to analyze the characteristics of problems and the behavior of algorithms to assist us in choosing or designing efficient algorithms. Fortunately, our recent work, the Nearest-Better Network (NBN) [4,5], has been proven to be an effective tool for analyzing the characteristics of problems of any dimensionality. Diao et al. [5] have verified that the characteristics of the fitness landscape can be captured by the NBN visualization, including asymmetry, ill conditioning, neutrality, ruggedness, size of the BoAs, and the number of BoAs.
The main contribution of this paper is to bridge the gap between theory and practice. Due to the black-box nature of real-world problems, it is difficult to uncover their inherent characteristics. Our recent work, NBN, has proven to be an effective tool for analyzing the characteristics of problems across various dimensionalities. For the first time, we visualize all the functions from the CEC 2022 competitions and CEC 2011 real-world problems and conduct a thorough analysis of several problems with distinct characteristics. Several interesting findings are made in this study:
  • In some real-world problems, the global structural features are not clearly defined. Algorithms that learn direction based on population distribution, such as CMA-ES, may easily exceed the problem’s boundaries.
  • Some real-world problems contain a large number of attraction basins. In such cases, the diversity maintenance mechanism may negatively impact the algorithm’s performance.
  • In certain problems, a vast neutral region exists around the global optimum, making it easy for the algorithm to become trapped in this region.
  • Some problems are highly ill-conditioned, causing even the best algorithms to fail to solve them.

2. Related Work

2.1. Fitness Landscape

A fitness landscape [6] is a mapping from solutions in the search space to fitness values with a neighborhood relationship. The solution space X is a set of potential solutions to the problem. The fitness of a solution indicates how good the solution is (the larger the value, the better the solution). The neighborhood relationship can be defined as the distance or accessibility between solutions.
There are many problem characteristics that can affect the algorithm’s performance [7,8]. Some of the characteristics are listed below:
  • Modality [9]: Multimodal problems have more than one global or local optimum. Basin of attraction (BoA) is an important concept for multimodal problems, such as B ( x * ) = x X x * = local - search ( x ) , where the BoA B ( x * ) of a local optimum x * is the set of solutions B ( x * ) that approaches x * by utilizing a local search strategy among the decision variable space X [10].
  • Ruggedness [11]: Ruggedness is usually manifested as steep ascents and descents in the fitness landscape with the existence of many local optima.
  • Neutrality [12]: Neutrality is usually manifested as a flat area of the fitness landscape. In this flat area, the optimization algorithm may have difficulty finding a better solution.
  • Ill condition [13]: An ill-conditioned problem indicates that it is extremely sensitive to slight changes. During the optimization process, small perturbations may lead to significant changes in the solution, making it difficult for the algorithm to converge to the global optima or resulting in a very slow convergence speed.
Generally speaking, there are two ways to bridge the gap between real-world problems and benchmarks: (1) Design problems similar to real-world problems. This type of method mainly focuses on benchmark design. (2) Analyze the characteristics of the given real-world problem. This type of method relies on fitness landscape analysis methods.

2.2. Benchmark Design

There have been many attempts to design problems similar to real-world problems. In the early stage, designers would combine the known characteristics existing in real-world problems, such as multimodality, non-convexity, noise, and constraints, to construct benchmark functions [14]. As people’s understanding of real-world problems deepens, these benchmarks are continuously updated, from CEC 2013 [15] to CEC 2022 [16].
Due to the complex and diverse characteristics of real-world problems, many real-world problems were quite different from the existing benchmarks. Then, Li [17] proposed a customizable benchmark framework based on space segmentation, freepeak. Under this framework, people can design subspaces with different characteristics and then combine them to form a complex problem to simulate any real-world problems. However, the design of the benchmark is limited by people’s understanding of real-world problems. If we do not know the characteristics of a real-world problem, we cannot design a benchmark function similar to this problem. Some benchmarks use functions that are very similar to real-world problems [18] or that even directly include several real-world problems [19].
Whether it is designing benchmarks similar to real-world problems or directly including several real-world problems, the characteristics of real-world problems are quite complex and diverse and it is almost impossible for benchmarks to cover all the characteristics of real-world problems. Therefore, directly analyzing the characteristics of real-world problems using the fitness landscape analysis method is a better choice. The fitness landscape analysis method can be divided into numerical methods and visualization methods according to their ability to visualize data.

2.2.1. Numerical Methods

These methods propose a series of metrics to describe specific characteristics of problems or algorithms. Lip uses the correlation length to evaluate ruggedness: [20]. Davidor employs epistasis variance to assess epistasis [21]. Reidys and Stadler utilized a neutral walk to evaluate neutrality [12]. Lunacek proposed the dispersion metric to evaluate global topology or the presence of funnels [22]. Morgan proposed the length scale to analyze gradient information [23]. Bosman attempted to visualize basins of attraction along with the associated stationary points via gradient-based stochastic sampling [24]. Some work has been carried out to analyze multi-objective landscape features [25] and constrained landscape features [26]. However, real-world problems are quite complex. Different regions in real-world problems may have different characteristics. It is very difficult to mine all the information solely through numerical methods.

2.2.2. Visualization Methods

Theoretically, a good visualization method can help us observe the fitness landscape comprehensively, as well as the search behavior of an algorithm, thereby understanding the problem structure and the algorithm’s working mechanism, as well as helping to design efficient algorithms.
Visual FLAs for continuous problems are very few, with only Local Optima Networks [27], Search Trajectory Networks [28]. and Nearest-Better Networks available. Local Optima Networks (LONs) visualize local optima of the fitness landscape in the form of a graph where nodes are local optima and edges represent possible transitions between optima with a given search operator. The Search Trajectory Network (STN) is defined as a graph whose nodes are locations in a search trajectory, which are representative solutions in different subspaces, and the edges represent the connections between solutions in a search. The Nearest-Better Network is a network where the nodes are solutions and the edges are the nearest-better relationship between solutions.
Previous experiments [5] have proven that the NBN can display many characteristics of the landscape in its visualization, while much characteristic information is lost in the LON and STN visualization. We attempt to use the NBN to visualize the fitness landscape of real-world problems and try to uncover the unknown characteristics of these problems.

3. Nearest-Better Network and Experimental Setup

In the experiment, the Nearest-Better Network is employed to analyze the characteristics of the selected problems. Additionally, sampling by algorithms is utilized to collect data. In this section, we will introduce the definition of the Nearest-Better Network, the selected problems and algorithms, and the sampling method.

3.1. Nearest-Better Network

The Nearest-Better Network [4] (NBN) is a directed graph with the sampled solutions as the vertices and the nearest-better relationships for each solution as the edges. The nearest-better relationship is defined as b ( x ) = arg min y y y X N , f ( y ) > f ( x ) y x , where b ( x ) is the nearest-better solution for the solution x . The distance to the nearest-better solution for the solution x , x , b ( x ) , is also known as its Nearest-Better Distance (NBD). In optimization algorithms, the search process is guided by fitness, and there is a higher probability of exploring the neighborhood of a given solution. Specifically, an algorithm is likelier to find the nearest-better solution from a given solution x . By retaining only the nearest-better relationships, the NBN simplifies the original fitness landscape to make it easier to analyze.
Figure 1a shows the original fitness landscape of CEC 2022 f 9 [16], which will be analyzed in detail in the following experiments. Figure 1b shows the structure of NBN visualization as proposed in reference [4]. The NBN visualization is a tree-like structure where each node is a solution that connects its nearest-better solution. The distance of the edge is the NBD. NBN visualization only preserves the nearest-better relationship and the NBD, so that it can display the NBN structure of problems of any dimensionality. From the original fitness landscape, it can be seen that f 9 is a smooth multimodal problem, and the NBN visualization also shows a similar structure. In the experiments [5], it has been verified that the NBN visualization can retain characteristics such as asymmetry, ill conditioning, neutrality, ruggedness, and size of the basins of attraction (BoAs), and the number of BoAs can be captured by the NBN visualization.

3.2. Problems for Analysis

For the artificially designed benchmark, we select the CEC 2022 benchmark [16]. This benchmark is relatively new and it also contains rich problem characteristics, including multimodal, ill conditioning, ruggedness, neutrality, separateness, etc. At the same time, there are many works [29,30] that have conducted experiments and discussions on this benchmark.
For real-world problems, we choose the CEC 2011 problem [31]. It was released in 2010 and has a citation count of 701. A great many algorithms have been tested using these functions. However, as of now, the characteristics of the functions are still unknown. Table 1 presents the basic information about the CEC 2011 problem. In this paper, we mainly focus on continuous problems. Therefore, we select the functions with continuous coding for analysis. Since one-dimensional and two-dimensional continuous problems are relatively simple, the fitness landscape and the search behavior of the algorithm can be directly visualized. Therefore, we select functions h 1 , h 2 , h 5 , h 6 , h 7 , h 9 , h 10 , h 11 , h 12 , and h 13 for visualization and analysis.

3.3. Selected Algorithms for Sampling

In this paper, we use sampling by algorithm. In high-dimensional problems, data from uniform sampling are extremely sparse relative to the entire solution space and it is difficult to capture the problem characteristics. However, many real-world problems are precisely high-dimensional problems. Although algorithm-based sampling is non-uniform sampling and some characteristics may not be reflected in the algorithm’s search data, this indicates that these characteristics have no impact on the algorithm’s search process. What we focus on are precisely the problem characteristics that have an impact on the algorithm’s search trajectory. Moreover, the NBN method used in this paper has no requirement for the uniformity of sampled data. It can analyze and visualize data from any source.
In this paper, the chosen sampling algorithms can be divided into two categories: global optimization algorithms and multimodal optimization algorithms.
For global optimization algorithms, four champion algorithms of CEC 2022 are selected:
  • EA4 [32]: This algorithm uses a cooperative model which contains four algorithms, including CMEAS [33].
  • NL-LBC [32]: This algorithm uses non-linear population size reduction success-history adaptive differential evolution with linear bias change.
  • NL-MID [34]: This algorithm uses non-linear population size reduction success-history adaptive differential evolution with midpoint.
  • S-DP [35]: This algorithm uses a differential evolution with a dynamic perturbation mechanism for population diversity management.
For multimodal optimization algorithms, four multimodal optimization algorithms with different diversity maintenance mechanisms are selected:
  • ANDE [36]: This algorithm uses an adaptive multi-population mechanism. The number of populations can be adaptively adjusted during the optimization process, and nearest neighbors are used in the elimination mechanism.
  • DHNDE [37]: This algorithm uses a dynamic hybrid niching method to maintain diversity.
  • HillVall [38]: This algorithm uses clustering to divide the solution space based on randomly initialized solutions. Then, a valley detection mechanism is employed to detect whether the divided population covers a peak. Subsequently, the divided population is used as the initial population for evolution. When all the divided populations have evolved and converged, HillVall combines the best solutions of each evolved divided population and some new randomly initialized solutions for re-clustering and evolution.
  • RS-CMSA [39]: This algorithm uses taboo points to repel the subpopulation to prevent convergence to the same basin.

3.4. Sampling Method

The NBN is constructed using all the data from 30 independent runs of each of the eight chosen algorithms mentioned above. However, there is still a problem with this sampling method in that the amount of data is too large. Two schemes to filter the data are used: uniform selection and optimal selection.
Let the set of all of the algorithm search data be S = { s 1 , s 2 , s 3 , , s m } , where each sample s i has a corresponding fitness value f ( s i ) . First, sort the samples in S according to their fitness values in descending order. For any i , j { 1 , 2 , , m } , when f ( s i ) f ( s j ) , s i comes before s j in the sorted result. Denote the sorted set as S = { s 1 , s 2 , s 3 , , s m } , where f ( s 1 ) f ( s 2 ) f ( s m ) .
Then, use one of the following two schemes to filter the data:
  • Uniform selection: Evenly select N solutions. Let n k = k m N for k { 1 , 2 , , N } . The set of selected solutions is S even = { s n 1 , s n 2 , s n 3 , , s n N } .
  • Optimal selection: Select the N solutions with the best fitness values:
    S best = { s 1 , s 2 , s 3 , , s N } .
Since the z-coordinate of the NBN visualization is the fitness value, the NBN constructed from data based on optimal selection pays more attention to the local structure of the top of the NBN, while the NBN constructed from uniformly selected data focuses on the global structure of the NBN. In this paper, the number of filtered solutions, N, is set to 1 × 106.
To show the local structure of the problem, this paper converts the original problem into one with a smaller boundary around the global optimum o = [ o 1 , o 2 , o D ] . The boundary of this problem is defined as [ o i L i * r , o i + R i * r ] , for k { 1 , 2 , , D } , where [ L i , R i ] are the original lower and upper bounds of the i-th dimension of the problem, and r is the reduction ratio of the problem’s search range. D is the dimensionality of the problem.

4. Experimental Analysis

Figure 2 and Figure 3 show the original fitness landscape of the two-dimensional CEC 2022 functions. Figure 4, Figure 5, and Figure 6 are the NBNs of all the functions constructed from uniformly selected data. Figure 7, Figure 8 and Figure 9 are the NBNs based on optimal selection. Table 2 shows the comparison results of the selected algorithms on these functions. From the NBN visualization, it can be observed that the problem characteristics of real-world problems are very diverse. For example, there is a neutral place around the global optima in h 6 . h 2 has a large neutral place near the global optimum. h 7 has many basins of attraction. h 12 and h 13 are highly ill-conditioned. In the following experiment, we will conduct an in-depth analysis and discussion on the functions with unique characteristics.

4.1. Comparison of High-Dimension and Low-Dimension Problems

For a long time, it has been widely believed that there is a certain degree of similarity between high-dimensional and low-dimensional problems. Based on this principle, when artificially designing benchmarks or algorithms, the fitness landscape of two-dimensional problems is observed to infer the characteristics of high-dimensional problems and the behavior of algorithms. However, there are also significant differences between them. The difference can lead to incorrect inferences about the behavior of algorithms on high-dimensional problems if we only rely on observations of the two-dimensional problems.
From the NBN visualization, we found that there is indeed some similarity between high-dimensional problems and low-dimensional problems. For example, CEC 2022 f 2 is ill-conditioned in two-dimensional space, as shown in Figure 2, and this characteristic is retained in the high-dimensional function, as shown in Figure 7.
However, the differences between the two-dimensional and ten-dimensional functions of f 9 and f 12 are relatively large. As shown in Table 2, the performances of algorithms on these two problems with low and high dimensions are quite different. In the two-dimensional problem, all algorithms can find the global optima. In the ten-dimensional problem, only HillVall finds the global optima, and the performance of the other four champion algorithms is not good. If we only look at the fitness landscape of two-dimensional f 9 and f 12 , it is difficult to find the reason. Especially for the two-dimensional f 9 problem, as shown in Figure 2, it seems to be a simple multimodal problem: the fitness landscape is relatively smooth, without any very difficult characteristics. The number of modalities is not very large and the size of the BoA of the global optima is relatively large. However, from Figure 10, we found that in high-dimensional f 9 and f 12 , the BoA of the global optimal solution becomes very small. The size of the BoA, | B o | , of f 9 is only 7.25 × 10 4 and | B o | of f 12 is only 1.58 × 10 4 . This makes the problems very difficult to solve for the four champion algorithms, which are global optimization algorithms.

4.2. Global Structure

In the experiments, we found that EA4 based on CMEAS sometimes exceeds the boundary of the problems, h 9 and f 11 . Table 2 also shows that the results of EA4 on h 9 and h 11 are poor. CMEAS calculates the evolution direction of the next-generation population based on the distribution of the current population. If the algorithm exceeds the boundary, it means that the algorithm cannot learn the direction of the global optima based on the current population. It also indicates that global structures of h 9 and f 11 are not clear.
Figures and Figure 6 demonstrate that h 9 and h 11 indeed possess a global structure. Figure 11 shows the population distribution from which EA4 cannot learn the direction of the global optimum. Owing to the loss of some information on the original high-dimensional problems, it is hard to obtain effective information from the NBN visualization. Therefore, some experiments need to be carried out for a further analysis of the global structure of h 9 and h 11 .
First, we reduce the search range of the problem to observe the search success rate of EA4 to verify whether the global structure of the problem becomes clearer as the search range of the problem shrinks. As shown in Figure 12, the search success rate of EA4 does increase as the search range of the problem shrinks. However, even when the initial search range is quite small, r = 0.1 , the success rate of EA4 does not exceed 0.5. This indicates that the global structure of h 9 and h 11 is not particularly clear even in the local areas.
However, we also need to take into account the influence of dimensionality. As indicated in Table 1, the dimensionality of h 9 and h 11 is relatively high. h 9 is 126-dimensional and h 11 is 120-dimensional. Is it possible that the global structure of high-dimensional problems is not as clear as that of the low-dimensional problems? In response to this doubt, we generated 10-dimensional, 120-dimensional, and 126-dimensional CEC 2022 problems, respectively, to verify the search success rate of EA4. As shown in Table 3, the success rate of EA4 in these functions is uniformly 100%. This indicates that the unclear global structure is a unique characteristic of h 9 and h 11 , which leads to a decline in the performance of EA4, which is based on global structure learning.

4.3. Modality

Multimodality is an important characteristic of the problems. For different types of multimodal characteristics, the behavior of algorithms is quite different. Later, we will analyze algorithm behavior in combination with different types of multimodal characteristics. We noted that when observing the characteristics of different algorithms, the data of the NBN in the figures are different. Due to different data sources, the NBN structures of the same problem are different.

4.3.1. The Size of the BoAs of Global Optima Is Very Small

In Section 4.1, we saw that the BoA of the global optima of CEC 2022 f 9 and f 12 are very small. However, it is still unknown why HillVall performs best among all the other algorithms in these two problems. What mechanism inside HillVall plays a role? Since the problem characteristics of CEC 2022 f 9 and f 12 are similar, only f 9 is discussed in this subsection.
Figure 13 shows that all algorithms except HillVall failed to find the global optima. In particular, ANDE, an adaptive multi-population algorithm, only maintains one population at the end of evolution. This indicates that its multi-population mechanism fails in the case where the BoA is very small. We can infer the reason why HillVall finds the global optima with this very small BoA from Figure 14. As shown in the clustering result, HillVall first divides the solution space into many subspaces, with a total of 42 subspaces. In the evolution that converges to the global optimum, although the clustering result is not very accurate, many solutions in the initial population belong to the BoA of the global optima, making it more likely for the algorithm trajectory based on this initial population to converge to the global optimum.

4.3.2. The Problem Has Many BoAs

In real-world optimization problems, there can be a situation where the number of BoAs is extremely large, such as CEC 2011 h 7 , which has 349 BoAs as shown in Figure 15. It is a difficult problem. In this problem, the behaviors of the algorithms with different mechanisms are very inconsistent with our expectations.
  • The multi-population mechanism does not perform well
    Ideally, the multi-population mechanism can achieve the state where one population covers one peak in multimodal problems. However, the performance is not satisfactory on this problem. From Figure 15, it can be observed that in ANDE, which is an adaptive multi-population algorithm, the number of populations is significantly smaller than the number of BoAs. There are even several subpopulations conducting searches in the same BoA. For example, the green subpopulation and orange subpopulation evolved in the same BoA. This indicates that the multi-population mechanism is unable to divide the BoAs accurately in this problem.
  • The mechanism for maintaining diversity has a negative effect.
    DHNDE and RS-CMSA possess mechanisms for maintaining diversity. From the algorithm trajectories in Figure 15, it can be observed that the diversity of these two algorithms is maintained well with solutions in each BoA. However, there are almost no solutions around the global optimum. This indicates that these two algorithms cannot converge to the global optimal solution.
    Similarly, among global optimization algorithms, S-DP also has a diversity maintenance mechanism, and it even cannot converge in this problem. The color represents the iteration, and the individuals at the last iteration (in red) are scattered at different BoAs. The results in Table 2 also verify its behavior. As one of the four champion algorithms of CEC 2022, its performance on this problem is the worst among all algorithms. This indicates that in the case of a very large number of BoAs, the mechanism for maintaining diversity even plays a negative effect on the algorithm performance.
  • The space segmentation mechanism can reduce the difficulty of the problem.
    HillVall is the algorithm with the best performance on this problem. Why can it outperform other algorithms? Is it similar to the case where the basin of attraction is very small? In fact, in the case where the BoA is very small and the case where the problem has many BoAs, HillVall’s behaviors are quite different.
    As shown in Figure 16, from the figure depicting the distribution of all the best solutions of each evolved population of HillVall, due to the unreasonable positions of some initialized subpopulations, the algorithm converges to slopes many times. This indicates that the clustering mechanism is not very effective.
    In the successful evolution shown in Figure 17, its initialized population and its converged position are not in the same BoA. Although the position of the initialized population is not accurate, evolving based on a population in a small area can still converge to a global optimum. This indicates that although the clustering mechanism is not effective, the space segmentation works. By dividing the solution space into smaller regions, it can reduce the multimodal difficulty of the problem.

4.3.3. Multimodal Optimization Algorithms Do Not Perform Better than Global Optimization Algorithms in Finding Multiple Global Optima

Compared to global optimization algorithms, multimodal optimization algorithms are more focused on finding more global optima. Generally, we would think that in terms of the number of global optimal solutions found, multimodal optimization algorithms can outperform global optimization algorithms. However, this is not the case in real-world problems.
CEC 2011 h 1 is a relatively simple multimodal problem with only eight global optima. The objective of the global optima of CEC 2011 h 1 is 0, as stated in [31], and many algorithms can find the global optimum as shown in Table 2. This indicates that the search operator can solve this problem well. Thus, we can exclude the influence of the search operator’s solving ability and observe the algorithm’s ability to find multiple global optima with this problem.
Figure 18 shows the distribution of algorithms near the global optimal solution on CEC 2011 h 1 . The black rectangular boxes are the global optima. It can be observed that multimodal optimization algorithms do not perform better than global optimization algorithms in finding multiple global optima. For example, HillVall and DHNDE only find two global optima, while EA4 without a diversity mechanism can find three global optima. This indicates that population-based algorithms, with multiple individuals evolving simultaneously, can naturally maintain diversity. And the algorithm that finds the most optima is a global optimization algorithm, NL-MID, which finds five global optima. This algorithm does not have a very complex diversity maintenance mechanism. It uses the midpoint mechanism, that is, it makes use of the geometric center of the population to maintain diversity.

4.4. Neutrality

The default dimensionality of h 2 is D = 30 , and the objective value of its global optima is −37.967600. As shown in Table 2, no algorithm finds the global optimal solution in any run. Since the NBN is generated from algorithm trajectory data if the algorithm does not find the global optima, we cannot observe the structure near the global optima without data.
Based on the principle that high-dimensional problems and low-dimensional problems have similarities, we try to visualize h 2 in low dimensionality. Fortunately, the dimensionality of h 2 can be adjusted. Although it cannot be set as a two-dimensional problem to directly observe the original fitness landscape of the problem, we can set a relatively low dimension to reduce the difficulty of the problem so that the algorithm can find the global optima and facilitate the observation of the structure near the global optima. Therefore, we designed h 2 with D = 6 . For this problem, there are some algorithms that can find the global optimal solution.
As shown in Figure 19, although the problem is a uni-modal problem, there is a neutral region near the global optimal solution, which is also the main difficulty in solving this problem. Table 4 shows the number of times the algorithm becomes stuck in a neutral place. It can be seen that these algorithms easily become stuck in the neutral place on this problem.
Very interestingly, although none of these algorithms find the global optima in h 2 ( D = 30 ) , the NBNs of h 2 ( D = 6 ) and h 2 ( D = 30 ) are very similar. As shown in Figure 5 and Figure 8, h 2 ( D = 30 ) is also a uni-modal problem, and there is a neutral region near the global optimal solution. This indicates that although the algorithm did not find the global optima, the neighborhood structure of the global optima is found by the algorithms.

4.5. Ill Conditioning

CEC 2011 h 12 is a highly ill-conditioned problem and the experiments also show that h 12 is a very difficult problem. Figure 20a shows that the objective value of the best-found solution decreases as the number of independent runs increases. This indicates that in more than 2000 independent runs, the best solution is found in only one run, and this solution may not necessarily be the global optima. This indicates that even the best algorithms fail to solve this complex problem, h 12 .
From Figure 6 and Figure 9, we find that h 12 is a uni-modal but highly ill-conditioned problem, with some solutions with long convergence trajectories. One characteristic of being ill-conditioned is that the algorithm has a slow convergence speed. That is, there are long convergence trajectories in the NBN. In the NBN, for any solution x S , there is a trajectory that converges to the global optimum o , P ˜ ( x , o ) = [ p 1 , p 2 , , p k ] , where p 1 = x , p k = o , and k represents the number of nodes along the path. It can be seen that the longer this trajectory is, the more difficult it is for the current solution to converge to the global optima. Based on this, we use the indicator of the longest convergence trajectory P ˜ max = max x S P ˜ ( x , o ) as the indicator for ill conditioning. As shown in Table 5, compared to other problems, the value of P ˜ max of h 12 is the highest, which indicates that CEC 2011 h 12 is the most ill-conditioned problem.
But one may wonder whether there are many very small BoAs in the problem that lead to the ill-conditioning characteristic of h 12 ? To answer this question, we visualize the local NBN of h 12 as shown in Figure 21. It can be observed that from r = 0.9 to r = 0.05 (r is the reduction ratio of the search range), h 12 is a uni-modal problem. As the problem search range increases, the ill-conditioned characteristic of the problem becomes more obvious. Figure 20b also verifies the uni-modal characteristic of h 12 . In previous work [5], the NBD is used as an indicator to identify local optimal solutions. If its NBD exceeds the threshold, the current solution is considered the local optima. Then, m a x NBD = max x S { o } NBD ( x ) , which is the maximum NBD in the NBN except for the global optimal solution o , can work as an indicator for the multimodal characteristic. The smaller this value is, the less multimodal the problem is. Figure 20b shows the relationship between the reduction ratio r and max NBD. As the problem search range becomes smaller, the max NBD also becomes smaller. This indicates that there is only one BoA at any microscopic scale of the problem. The difficulty of the problem lies only in being ill-conditioned.

5. Discussion

From the experimental results, the NBN is quite useful. It can uncover many problem characteristics, such as multimodality, neutrality, and ill conditioning. It can help us observe the problem characteristics and the behavior of algorithms and it is also a very good tool to assist people in designing algorithms. However, it has some drawbacks. In NBN visualization, some information about high-dimensional problems is inevitably lost. For example, for CEC 2011 h 9 and h 10 , EA4 based on global structure learning exceeds the boundary multiple times, but the NBN cannot observe the structure of the distribution of the population. This is due to the loss of some spatial structure information of high-dimensional problems in NBN visualization. Moreover, the NBN is data-dependent. If the data cannot provide effective information, the NBN cannot observe the relevant characteristics either. For instance, if an algorithm fails to find the global optima, the NBN cannot observe the local structure of the global optima either.
From the experiments, we can find that there is a significant gap between real-world problems and benchmarks. Although the CEC 2022 benchmark contains many characteristics, such as multimodality, neutrality, and ill-conditioning, these characteristics do not have much impact on algorithm behaviors. The NBN structures generated based on algorithm data for f 1 ( D = 10 ) , f 4 ( D = 10 ) , f 5 ( D = 10 ) , f 6 ( D = 10 ) , and f 11 ( D = 10 ) are quite similar. The characteristics of these problems have little impact on algorithm behavior, resulting in these characteristics being ignored in algorithm trajectories. The only one that significantly impacts algorithm behavior is the small size of the BoA of global optima, which is observed in f 9 ( D = 10 ) and f 10 ( D = 10 ) . The characteristics of real-world problems are extremely rich and unique. h 7 has 349 basins of attraction. h 2 has a large neutral space around the global optimum. h 12 is a highly ill-conditioned problem. Although CEC 2022 considered the difficulty of ill conditioning and designed an ill-condition function, f 2 , the degree of the ill conditioning of h 12 is much higher than that of f 2 , as shown in the result of Table 5. This indicates that our understanding of real-world problems is insufficient, which also leads to a large gap between the benchmarks and real-world problems. This can also explain to some extent why the champion algorithms in competitions sometimes perform poorly on real-world problems.
At the same time, the performance of an algorithm is closely related to problem characteristics. For a long time, we have believed that the ability to maintain diversity is important for algorithms and, many algorithms emphasize their ability to maintain diversity [40]. However, in some cases, the diversity maintenance mechanism may reduce the performance of an algorithm. For example, h 7 has a very large number of BoAs (a total of 349). For this problem, algorithms with a diversity maintenance mechanism all have very poor performance and may even become stuck in stagnation, such as S-DP. The characteristics of real-world problems are extremely diverse. If the algorithms blindly use their mechanisms, they cannot adapt to these diverse real-world problems. Future algorithms should have the ability to learn problem characteristics and adaptively adjust their mechanisms.
In this paper, we visualize the real-world CEC 2011 and CEC 2022 benchmarks. While using standardized benchmarks like CEC functions helps facilitate comparisons, they may not fully capture the diversity of real-world optimization challenges. We strongly encourage researchers to apply the NBN to analyze the real-world problems they face, as these may contain many unknown and challenging characteristics. We look forward to researchers sharing their findings with us. A link to the code used in this paper is provided at the end.
The time complexity of the NBN computation consists of two parts: the neighborhood relationship calculation and the nearest-better relationship calculation: (1) The problems analyzed in this paper are continuous, high-dimensional problems. To compute the neighborhood relationships, we use kd-trees [41], which have a time complexity of O ( N ln N ) . (2) The nearest-better relationship calculation algorithm, proposed in [5], has a time complexity of O ( w N D ln N ) , where N is the number of sampled solutions, D is the problem’s dimensionality, and w is a parameter related to the number of peaks and their shapes. Although the time complexity is log-linear, it remains computationally expensive. For instance, computing a single NBN instance with 1000,000,000 samples on a DELL V3670-i7(8700)-GTX1050Ti-48G (manufacturer: Dell Technologies, Wuhan, China) system takes approximately 12 h. Further optimization of the algorithm is still necessary. A parallel implementation is expected to enhance its efficiency, and this is also one of our future projects.

6. Conclusions

A more in-depth perspective on observing real-world problems with the NBN is provided in this paper to enhance people’s understanding of real-world problems and help people design more efficient algorithms. In this paper, we visualize all the functions in CEC 2022 and CEC 2011 and select some of these functions with special characteristics for in-depth analysis with the NBN. The code of this paper can be found on OFEC. There are many interesting findings, such as the following:
  • The global structures of some real-world problems are clear, and on these problems, algorithms based on population distribution learning such as EA4 are prone to exceed the boundary.
  • In real-world problems, there exist some problems that contain a very large number of BoAs. CEC 2011 f 7 has a total of 349 BoAs. In this problem, the diversity maintenance mechanism has a negative impact on the algorithm performance.
  • For some real-world problems, such as CEC 2011 h 2 , there is a large neutral area near the global optimal solution, which makes the algorithm easily stuck in the neutral place.
  • There are some highly ill-conditioned problems that are difficult to solve, such as CEC 2011 f 12 . This problem is uni-modal and highly ill-conditioned and it is characterized by long convergence trajectories. The experimental results show that none of the current best algorithms can solve this problem efficiently.
In our future work, we will make good use of the structure of the NBN to design an intelligent algorithm with the ability to learn problem characteristics and adaptively adjust the mechanisms to adapt to diverse real-world problems.

Author Contributions

Conceptualization, C.L., S.Z., and S.Y.; Formal analysis, Y.D.; Funding acquisition, C.L.; Investigation, Y.D.; Methodology, Y.D. and C.L.; Project administration, C.L.; Resources, C.L.; Software, Y.D. and J.W.; Supervision, C.L., S.Z., and S.Y.; Validation, Y.D.; Visualization, Y.D.; Writing—original draft, Y.D.; Writing—review and editing, Y.D. and C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62476006, in part by the Hubei Provincial Natural Science Foundation of China under Grant 2023AFA049, and in part by the Fundamental Research Funds of the AUST under Grant 2024JBZD0007.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NBNNearest-Better Network
NFLNo Free Lunch
BoABasin of Attraction
LONLocal Optima Network
STNSearch Trajectory Network
NBDNearest-Better Distance

References

  1. Osaba, E.; Villar-Rodriguez, E.; Del Ser, J.; Nebro, A.J.; Molina, D.; LaTorre, A.; Suganthan, P.N.; Coello Coello, C.A.; Herrera, F. A Tutorial On the design, experimentation and application of metaheuristic algorithms to real-World optimization problems. Swarm Evol. Comput. 2021, 64, 100888. [Google Scholar] [CrossRef]
  2. Wolpert, D.; Macready, W. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  3. Martí, R.; Sevaux, M.; Sörensen, K. Fifty years of metaheuristics. Eur. J. Oper. Res. 2024, 321, 345–362. [Google Scholar] [CrossRef]
  4. Diao, Y.; Li, C.; Zeng, S.; Yang, S. Nearest Better Network for Visualization of the Fitness Landscape. In Proceedings of the GECCO’23 Companion: Companion Conference on Genetic and Evolutionary Computation, Lisbon, Portugal, 15–19 July 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 815–818. [Google Scholar]
  5. Diao, Y.; Li, C.; Zeng, S.; Yang, S.; Coello, C.A.C. Nearest-Better Network for Fitness Landscape Analysis of Continuous Optimization Problems. IEEE Trans. Evol. Comput. 2024; published online. [Google Scholar] [CrossRef]
  6. Wright, S. The roles of mutation, inbreeding, crossbreeding and selection in evolution. In Proceedings of the Sixth International Congress of Genetics, Ithaca, NY, USA, 24–31 August 1932; pp. 356–366. [Google Scholar]
  7. Malan, K.M.; Engelbrecht, A.P. A survey of techniques for characterising fitness landscapes and some possible ways forward. Inf. Sci. 2013, 241, 148–163. [Google Scholar] [CrossRef]
  8. Malan, K.M. A Survey of Advances in Landscape Analysis for Optimisation. Algorithms 2021, 14, 40. [Google Scholar] [CrossRef]
  9. Horn, J.; Goldberg, D.E. Genetic Algorithm Difficulty and the Modality of Fitness Landscapes. In Foundations of Genetic Algorithms; Whitley, L.D., Vose, M.D., Eds.; Elsevier: Amsterdam, The Netherlands, 1995; Volume 3, pp. 243–269. [Google Scholar]
  10. Zou, F.; Chen, D.; Liu, H.; Cao, S.; Ji, X.; Zhang, Y. A survey of fitness landscape analysis for optimization. Neurocomputing 2022, 503, 129–139. [Google Scholar] [CrossRef]
  11. Eiben, A.E.; Smith, J.E. Introduction to Evolutionary Computing; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  12. Reidys, C.M.; Stadler, P.F. Neutrality in fitness landscapes. Appl. Math. Comput. 2001, 117, 321–350. [Google Scholar] [CrossRef]
  13. Wright, S.J. Numerical Optimization; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  14. Auger, A.; Hansen, N.; Schoenauer, M. Benchmarking of Continuous Black Box Optimization Algorithms. Evol. Comput. 2012, 20, 481. [Google Scholar] [CrossRef] [PubMed]
  15. Liang, J.J.; Qu, B.; Suganthan, P.N.; Hernández-Díaz, A.G. Problem Definitions and Evaluation Criteria for the CEC 2013 Special Session on Real-Parameter Optimization; Technical Report 201212; Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China; Nanyang 574 Technological University: Singapore, 2013; pp. 281–295. [Google Scholar]
  16. Kumar, A.; Price, K.V.; Mohamed, A.W.; Hadi, A.A.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the 2022 Special Session and Competition on Single Objective Bound Constrained Numerical Optimization; Technical Report; Nanyang Technological University: Singapore, 2021. [Google Scholar]
  17. Li, C.; Nguyen, T.T.; Zeng, S.; Yang, M.; Wu, M. An open framework for constructing continuous optimization problems. IEEE Trans. Cybern. 2018, 49, 2316–2330. [Google Scholar] [CrossRef] [PubMed]
  18. Sergeyev, Y.D.; Kvasov, D.E.; Mukhametzhanov, M.S. A Generator of Multiextremal Test Classes With Known Solutions for Black-Box-Constrained Global Optimization. IEEE Trans. Evol. Comput. 2022, 26, 1261–1270. [Google Scholar] [CrossRef]
  19. Yuriy, R.; Viatcheslav, L.; Borys, B. A Real-World Benchmark Problem for Global Optimization. Cybern. Inf. Technol. 2023, 23, 23–39. [Google Scholar] [CrossRef]
  20. Lipsitch, M. Adaptation on Rugged Landscapes Generated by Iterated Local Interactions of Neighboring Genes. In Proceedings of the 4th International Conference on Genetic Algorithms, San Diego, CA, USA, 13–16 July 1991; Belew, R.K., Booker, L.B., Eds.; Morgan Kaufmann: Burlington, MA, USA, 1991; pp. 128–135. [Google Scholar]
  21. Davidor, Y. Epistasis Variance: A Viewpoint on GA-Hardness. In Foundations of Genetic Algorithms; Rawlins, G.J., Ed.; Elsevier: Amsterdam, The Netherlands, 1991; Volume 1, pp. 23–35. [Google Scholar]
  22. Lunacek, M.; Whitley, D. The dispersion metric and the CMA evolution strategy. In Proceedings of the GECCO’06: 8th Annual Conference on Genetic and Evolutionary Computation, Seattle, WA, USA, 8–12 July 2006; Association for Computing Machinery: New York, NY, USA, 2006; pp. 477–484. [Google Scholar]
  23. Morgan, R.; Gallagher, M. Analysing and characterising optimization problems using length scale. Soft Comput. 2017, 21, 1735–1752. [Google Scholar] [CrossRef]
  24. Bosman, A.S.; Engelbrecht, A.P.; Helbig, M. Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 2020, 400, 113–136. [Google Scholar] [CrossRef]
  25. Liefooghe, A.; Daolio, F.; Verel, S.; Derbel, B.; Aguirre, H.; Tanaka, K. Landscape-aware performance prediction for evolutionary multiobjective optimization. IEEE Trans. Evol. Comput. 2019, 24, 1063–1077. [Google Scholar] [CrossRef]
  26. Malan, K.M.; Oberholzer, J.F.; Engelbrecht, A.P. Characterising constrained continuous optimisation problems. In Proceedings of the 2015 IEEE Congress on Evolutionary Computation (CEC), Sendai, Japan, 25–28 May 2015; pp. 1351–1358. [Google Scholar]
  27. Ochoa, G.; Verel, S.; Daolio, F.; Tomassini, M. Local Optima Networks: A New Model of Combinatorial Fitness Landscapes. In Recent Advances in the Theory and Application of Fitness Landscapes; Springer: Berlin/Heidelberg, Germany, 2014; pp. 233–262. [Google Scholar]
  28. Ochoa, G.; Malan, K.M.; Blum, C. Search trajectory networks: A tool for analysing and visualising the behaviour of metaheuristics. Appl. Soft Comput. 2021, 109, 107492. [Google Scholar] [CrossRef]
  29. Biedrzycki, R. Revisiting CEC 2022 ranking: A new ranking method and influence of parameter tuning. Swarm Evol. Comput. 2024, 89, 101623. [Google Scholar] [CrossRef]
  30. Price, K.V.; Kumar, A.; Suganthan, P. Trial-based dominance for comparing both the speed and accuracy of stochastic optimizers with standard non-parametric tests. Swarm Evol. Comput. 2023, 78, 101287. [Google Scholar] [CrossRef]
  31. Das, S.; Suganthan, P.N. Problem Definitions and EVALUATION criteria for CEC 2011 Competition on Testing Evolutionary Algorithms on Real World Optimization Problems; Jadavpur University: Kolkata, India; Nanyang Technological University: Singapore, 2010; pp. 341–359. [Google Scholar]
  32. Bujok, P.; Kolenovsky, P. Eigen crossover in cooperative model of evolutionary algorithms applied to CEC 2022 single objective numerical optimisation. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  33. Hansen, N.; Ostermeier, A. Adapting arbitrary normal mutation distributions in evolution strategies: The covariance matrix adaptation. In Proceedings of the IEEE International Conference on Evolutionary Computation, Nagoya, Japan, 20–22 May 1996; pp. 312–317. [Google Scholar]
  34. Biedrzycki, R.; Arabas, J.; Warchulski, E. A version of NL-SHADE-RSP algorithm with midpoint for CEC 2022 single objective bound constrained problems. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  35. Van Cuong, L.; Bao, N.N.; Phuong, N.K.; Binh, H.T.T. Dynamic perturbation for population diversity management in differential evolution. In Proceedings of the Genetic and Evolutionary Computation Conference Companion, Boston, MA, USA, 9–13 July 2022; pp. 391–394. [Google Scholar]
  36. Wang, Z.J.; Zhan, Z.H.; Lin, Y.; Yu, W.J.; Wang, H.; Kwong, S.; Zhang, J. Automatic niching differential evolution with contour prediction approach for multimodal optimization problems. IEEE Trans. Evol. Comput. 2019, 24, 114–128. [Google Scholar] [CrossRef]
  37. Wang, K.; Gong, W.; Deng, L.; Wang, L. Multimodal optimization via dynamically hybrid niching differential evolution. Knowl.-Based Syst. 2022, 238, 107972. [Google Scholar] [CrossRef]
  38. Maree, S.C.; Alderliesten, T.; Thierens, D.; Bosman, P.A. Real-valued evolutionary multi-modal optimization driven by hill-valley clustering. In Proceedings of the Genetic and Evolutionary Computation Conference, Kyoto, Japan, 15–19 July 2018; pp. 857–864. [Google Scholar]
  39. Ahrari, A.; Deb, K.; Preuss, M. Multimodal optimization by covariance matrix self-adaptation evolution strategy with repelling subpopulations. Evol. Comput. 2017, 25, 439–471. [Google Scholar] [CrossRef] [PubMed]
  40. Segura, C.; Hernández-Aguirre, A.; Luna, F.; Alba, E. Improving Diversity in Evolutionary Algorithms: New Best Solutions for Frequency Assignment. IEEE Trans. Evol. Comput. 2017, 21, 539–553. [Google Scholar] [CrossRef]
  41. Bentley, J.L. Multidimensional binary search trees used for associative searching. Commun. ACM 1975, 18, 509–517. [Google Scholar] [CrossRef]
Figure 1. Transformation from the original fitness landscape of the CEC 2022 f 9 to the NBN visualization with 2500 samples [4].
Figure 1. Transformation from the original fitness landscape of the CEC 2022 f 9 to the NBN visualization with 2500 samples [4].
Information 16 00190 g001
Figure 2. CEC 2022 functions. D = 2 .
Figure 2. CEC 2022 functions. D = 2 .
Information 16 00190 g002
Figure 3. CEC 2022 functions. D = 2 .
Figure 3. CEC 2022 functions. D = 2 .
Information 16 00190 g003
Figure 4. Nearest-Better Network of the functions with S even .
Figure 4. Nearest-Better Network of the functions with S even .
Information 16 00190 g004aInformation 16 00190 g004b
Figure 5. Nearest-Better Network of the functions with S even .
Figure 5. Nearest-Better Network of the functions with S even .
Information 16 00190 g005aInformation 16 00190 g005b
Figure 6. Nearest-Better Network of the functions with S even .
Figure 6. Nearest-Better Network of the functions with S even .
Information 16 00190 g006
Figure 7. Nearest-Better Network of the functions with S best .
Figure 7. Nearest-Better Network of the functions with S best .
Information 16 00190 g007
Figure 8. Nearest-Better Network of the functions with S best .
Figure 8. Nearest-Better Network of the functions with S best .
Information 16 00190 g008
Figure 9. Nearest-Better Network of the functions with S best .
Figure 9. Nearest-Better Network of the functions with S best .
Information 16 00190 g009
Figure 10. NBN of S even of CEC 2022 ( D = 10 ) f 9 and f 12 : The black area represents the BoA of the global optima and | B o | is the size of this BoA. The calculation of | B o | is described in [5].
Figure 10. NBN of S even of CEC 2022 ( D = 10 ) f 9 and f 12 : The black area represents the BoA of the global optima and | B o | is the size of this BoA. The calculation of | B o | is described in [5].
Information 16 00190 g010
Figure 11. The colored individual points are different individuals in the population P . Evolution is carried out based on the current population and the next-generation population of the EA4 will exceed the problem boundary.
Figure 11. The colored individual points are different individuals in the population P . Evolution is carried out based on the current population and the next-generation population of the EA4 will exceed the problem boundary.
Information 16 00190 g011
Figure 12. The success rate of EA4 over 30 independent runs in the subproblems of CEC 2011 h 9 and h 11 , where r is the reduction ratio of the search area as defined previously.
Figure 12. The success rate of EA4 over 30 independent runs in the subproblems of CEC 2011 h 9 and h 11 , where r is the reduction ratio of the search area as defined previously.
Information 16 00190 g012
Figure 13. Algorithm behaviors on CEC 2022 f 9 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated.
Figure 13. Algorithm behaviors on CEC 2022 f 9 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated.
Information 16 00190 g013
Figure 14. HillVall’s behaviors on CEC 2022 f 9 ( D = 10 ): In the clustering result figure, the colored points are the solutions initialized in the restart stage and the color represents the subspaces to which the solution belongs. The initialized population figure shows the initialized population for the algorithm trait. The algorithm trait figure shows all data of the evolution based on the above initialized population and the color indicates the iteration at which the solutions are generated.
Figure 14. HillVall’s behaviors on CEC 2022 f 9 ( D = 10 ): In the clustering result figure, the colored points are the solutions initialized in the restart stage and the color represents the subspaces to which the solution belongs. The initialized population figure shows the initialized population for the algorithm trait. The algorithm trait figure shows all data of the evolution based on the above initialized population and the color indicates the iteration at which the solutions are generated.
Information 16 00190 g014
Figure 15. The BoAs of h 7 , where the color indicates the BoA to which each solution belongs.
Figure 15. The BoAs of h 7 , where the color indicates the BoA to which each solution belongs.
Information 16 00190 g015
Figure 16. Algorithm behaviors on CEC 2022 h 7 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated.
Figure 16. Algorithm behaviors on CEC 2022 h 7 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated.
Information 16 00190 g016
Figure 17. HillVall’s behaviors on CEC 2022 h 7 ( D = 10 ): In the clustering result figure, the colored points are the solutions initialized in the restart stage and the color represents the subspaces to which the solution belongs. The initialized population figure shows the initialized population for the algorithm trait. The algorithm trait figure shows all data of the evolution based on the above initialized population and the color indicates the iteration at which the solutions are generated.
Figure 17. HillVall’s behaviors on CEC 2022 h 7 ( D = 10 ): In the clustering result figure, the colored points are the solutions initialized in the restart stage and the color represents the subspaces to which the solution belongs. The initialized population figure shows the initialized population for the algorithm trait. The algorithm trait figure shows all data of the evolution based on the above initialized population and the color indicates the iteration at which the solutions are generated.
Information 16 00190 g017
Figure 18. Algorithm behaviors on CEC 2022 h 1 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated. The black rectangular box represents the global optima. F is the number of global optima found by the algorithm.
Figure 18. Algorithm behaviors on CEC 2022 h 1 ( D = 10 ): The data of ANDE are the union of S best and all solutions of each population at the last stage of the evolution. The color represents the subpopulation to which the solution belongs. The data of Hillvall are the union of S best and the best solutions of all evolved populations at each restart time. The color indicates the restart time of each solution. For other algorithms, the data represent the union of S best and all data from one algorithm trait. The color indicates the iteration at which the solutions are generated. The black rectangular box represents the global optima. F is the number of global optima found by the algorithm.
Information 16 00190 g018aInformation 16 00190 g018b
Figure 19. The NBN of CEC 2011 h 2 ( D = 6 ) .
Figure 19. The NBN of CEC 2011 h 2 ( D = 6 ) .
Information 16 00190 g019
Figure 20. (a) Relationship between the objective of the best solution of all data and the number of independent runs. (b) Relationship between the reduction ratio r and max NBD.
Figure 20. (a) Relationship between the objective of the best solution of all data and the number of independent runs. (b) Relationship between the reduction ratio r and max NBD.
Information 16 00190 g020
Figure 21. Local structure of h 12 where r is the reduction ratio of the problem’s search range as introduced above.
Figure 21. Local structure of h 12 where r is the reduction ratio of the problem’s search range as introduced above.
Information 16 00190 g021
Table 1. CEC 2011 competition on real-world optimization problems, where D is the dimensionality of the problems.
Table 1. CEC 2011 competition on real-world optimization problems, where D is the dimensionality of the problems.
NameVariable EncodingD
h 1 Parameter Estimation for Frequency-Modulated Sound Wavescont.6
h 2 Lennard-Jones Potential Problemcont.30
h 3 Bifunctional Catalyst Blend Optimal Control Problemcont.1
h 4 Optimal Control of a Non-Linear Stirred Tank Reactorcont.1
h 5 Tersoff Potential for Model Si (B)cont.30
h 6 Tersoff Potential for Model Si (C)cont.30
h 7 Spread Spectrum Radar Polly Phase Code Designcont.20
h 8 Transmission Network Expansion Planning Problemcomb.7
h 9 Large-Scale Transmission Pricing Problemcont.126
h 10 Circular Antenna Array Design Problemcont.12
h 11 ELD Problems: DED Instance 1cont.120
h 12 Messenger: Spacecraft Trajectory Optimization Problemcont.26
h 13 Cassini 2: Spacecraft Trajectory Optimization Problemcont.22
Table 2. Results of the algorithms on different functions, where D is the dimensionality of the problem, R is each algorithm’s rank, and “mean” is the mean value of the algorithm result among 30 runs.
Table 2. Results of the algorithms on different functions, where D is the dimensionality of the problem, R is each algorithm’s rank, and “mean” is the mean value of the algorithm result among 30 runs.
CEC 2022 D = 2
f 1 f 2 f 3 f 4 f 5 f 6 f 7 f 8 f 9 f 10 f 11 f 12
NameRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmean
EA41−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.004−2433.331−26001−2700
NL-LBC1−3001−4001−6001−8001−9001−18001−20001−2201.722−2306.675−2460.041−26001−2700
NL-MID1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.003−2423.341−26001−2700
S-DP1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.001−2400.001−26001−2700
ANDE1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.002−2422.931−26001−2700
DHNDE1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.001−2400.001−26001−2700
HillVall1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.001−2400.001−26001−2700
RS-CMSA1−3001−4001−6001−8001−9001−18001−20001−2201.721−2300.001−2400.001−26001−2700
CEC 2022 D = 10
f 1 f 2 f 3 f 4 f 5 f 6 f 7 f 8 f 9 f 10 f 11 f 12
NameRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmean
EA41−3003−400.401−600.003−802.891−900.001−1800.001−2000.002−2200.033−2529.287−2500.071−2600.007−2864.60
NL-LBC1−3002−400.131−600.001−800.631−900.003−1800.161−2000.001−2200.003−2529.288−2500.11−2600.008−2864.92
NL-MID1−3001−400.001−600.007−804.153−900.092−1800.071−2000.004−2200.112−2502.552−2403.911−2600.006−2863.57
S-DP1−3001−400.001−600.006−804.021−900.004−1800.301−2000.003−2200.103−2529.281−24001−2600.003−2861.31
ANDE1−3004−400.403−600.218−815.422−900.016−1806.663−2005.186−2207.983−2529.284−2486.992−2625.074−2861.43
DHNDE1−3005−402.161−600.004−803.751−900.005−1801.744−2006.907−2208.893−2529.286−2500.061−2600.002−2859.99
HillVall1−3001−400.002−600.005−803.981−900.008−1818.205−2009.488−2212.741−2379.195−2493.941−2600.001−2841.43
RS-CMSA1−3001−400.001−600.002−801.531−900.007−1808.062−2001.715−2202.724−2529.283−2468.981−2600.005−2863.43
CEC 2011
h 1 h 2 h 5 h 6 h 7 h 9 h 10 h 11 h 12 h 13
NameRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmeanRmean
EA410.000231.70335.696328.9842−0.5838−2.164 × 106522.7468−1.027 × 1072−11.3674−15.602
NL-LBC5−0.365530.72534.299527.9146−0.8763−5.969 × 102621.5192−4.973 × 1045−14.1797−19.386
NL-MID7−0.624132.31136.721129.1664−0.7301−2.085 × 102132.2001−4.952 × 1041−11.1731−12.834
S-DP30.000630.39236.483229.1668−0.9552−2.482 × 102231.6653−5.143 × 1043−12.8332−13.145
ANDE8−3.353726.78734.020724.8137−0.9465−9.017 × 103814.7446−6.131 × 1048−15.8648−20.153
DHNDE20.000331.14634.024625.8695−0.7374−2.517 × 103427.3725−5.261 × 1047−15.1453−15.173
HillVall6−0.553430.78833.844821.1431−0.5237−1.583 × 106721.4374−5.248 × 1046−15.0455−18.542
RS-CMSA4−0.35980.00435.687428.1533−0.6166−2.810 × 105327.5007−6.858 × 1044−13.3966−18.598
Table 3. The success rate of EA4, where success rates lower than 100% are in bold.
Table 3. The success rate of EA4, where success rates lower than 100% are in bold.
CEC 2022
D f 1 f 2 f 3 f 4 f 5 f 6 f 7 f 8 f 9 f 10 f 11 f 12
1030/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/30
12030/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/30
12630/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/3030/30
CEC 2011
D h 1 h 2 h 5 h 6 h 7 h 9 h 10 h 11 h 12 h 13
-30/3030/3030/3030/3030/304/3030/302/3030/3030/30
Table 4. Frequency of becoming stuck in neutral regions of algorithms on CEC 2011 h 2 ( D = 6 ) .
Table 4. Frequency of becoming stuck in neutral regions of algorithms on CEC 2011 h 2 ( D = 6 ) .
EA4NL-LBCNL-MIDS-DPANDEDHNDEHillVallRS-CMSA
30/3030/3030/3028/3030/3030/3030/3011/30
Table 5. P ˜ max of different functions, where the largest values are in bold.
Table 5. P ˜ max of different functions, where the largest values are in bold.
f 1 ( D = 10 ) f 2 ( D = 10 ) f 3 ( D = 10 ) f 4 ( D = 10 ) f 5 ( D = 10 ) f 6 ( D = 10 )
even1.83 × 10−44.60 × 10−41.04 × 10−46.52 × 10−55.67 × 10−52.71 × 10−4
best4.2 × 10−51.25 × 10−43.5 × 10−53.6 × 10−53.8 × 10−52.23 × 10−4
f 7 ( D = 10 ) f 8 ( D = 10 ) f 9 ( D = 10 ) f 10 ( D = 10 ) f 11 ( D = 10 ) f 12 ( D = 10 )
even1.00 × 10−42.28 × 10−43.26 × 10−41.50 × 10−43.66 × 10−43.13 × 10−4
best8.1 × 10−51.09 × 10−49.6 × 10−56.8 × 10−54.8 × 10−51.20 × 10−4
h 1 h 2 h 5 h 6 h 7 h 9
even1.32 × 10−49.83125 × 10−56.72 × 10−46.30 × 10−43.58 × 10−46.50 × 10−4
best1.80 × 10−41.04 × 10−41.23 × 10−41.23 × 10−42.22 × 10−41.85 × 10−4
h 10 h 11 h 12 h 13
even2.92 × 10−46.23 × 10−41.27 × 10−38.04 × 10−4
best1.29 × 10−43.19 × 10−41.76 × 10−37.05 × 10−4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Diao, Y.; Li, C.; Wang, J.; Zeng, S.; Yang, S. Bridging the Gap Between Theory and Practice: Fitness Landscape Analysis of Real-World Problems with Nearest-Better Network. Information 2025, 16, 190. https://doi.org/10.3390/info16030190

AMA Style

Diao Y, Li C, Wang J, Zeng S, Yang S. Bridging the Gap Between Theory and Practice: Fitness Landscape Analysis of Real-World Problems with Nearest-Better Network. Information. 2025; 16(3):190. https://doi.org/10.3390/info16030190

Chicago/Turabian Style

Diao, Yiya, Changhe Li, Junchen Wang, Sanyou Zeng, and Shengxiang Yang. 2025. "Bridging the Gap Between Theory and Practice: Fitness Landscape Analysis of Real-World Problems with Nearest-Better Network" Information 16, no. 3: 190. https://doi.org/10.3390/info16030190

APA Style

Diao, Y., Li, C., Wang, J., Zeng, S., & Yang, S. (2025). Bridging the Gap Between Theory and Practice: Fitness Landscape Analysis of Real-World Problems with Nearest-Better Network. Information, 16(3), 190. https://doi.org/10.3390/info16030190

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop