Next Article in Journal
Nonlinear Robust Control of Vehicle Stabilization System with Uncertainty Based on Neural Network
Previous Article in Journal
The Past, Present, and Future of the Internet: A Statistical, Technical, and Functional Comparison of Wired/Wireless Fixed/Mobile Internet
Previous Article in Special Issue
A Causality-Aware Perspective on Domain Generalization via Domain Intervention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Objective Automatic Clustering Algorithm Based on Evolutionary Multi-Tasking Optimization

1
Air Traffic Control and Navigation School, Air Force Engineering University, Xi’an 710038, China
2
Key Laboratory of Collaborative Intelligence Systems, School of Electronic Engineering, Ministry of Education, Xidian University, Xi’an 710071, China
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(10), 1987; https://doi.org/10.3390/electronics13101987
Submission received: 14 April 2024 / Revised: 7 May 2024 / Accepted: 15 May 2024 / Published: 19 May 2024

Abstract

:
Data mining technology is the process of extracting hidden knowledge and potentially useful information from a large number of incomplete, noisy, and random practical application data. The clustering algorithm based on multi-objective evolution has obvious advantages compared with the traditional single-objective method. In order to further improve the performance of evolutionary multi-objective clustering algorithms, this paper proposes a multi-objective automatic clustering model based on evolutionary multi-task optimization. Based on the multi-objective clustering algorithm that automatically determines the value of k, evolutionary multi-task optimization is introduced to deal with multiple clustering tasks simultaneously. A set of non-dominated solutions for clustering results is obtained by concurrently optimizing the overall deviation and connectivity index. Multi-task adjacency coding based on a locus adjacency graph was designed to encode the clustered data. Additionally, an evolutionary operator based on relevance learning was designed to facilitate the evolution of individuals within the population. It also facilitates information transfer between individuals with different tasks, effectively avoiding negative transfer. Finally, the proposed algorithm was applied to both artificial datasets and UCI datasets for testing. It was then compared with traditional clustering algorithms and other multi-objective clustering algorithms. The results verify the advantages of the proposed algorithm in clustering accuracy and algorithm convergence.

1. Introduction

Clustering is an indispensable technique in data mining [1]. It has attracted widespread interest in image segmentation [2], social network analysis [3], text analysis [4], and so on. Clustering is an important unsupervised classification technique [5], which is suitable for revealing the underlying distribution of data. It can divide vectors in multidimensional space into different categories. Clustering techniques can be roughly divided into several categorization methods, such as partition-based, density-based, grid-based, hierarchical, and model-based clustering [6].
Among the partition-based clustering algorithms, the representative algorithms are k-means (KM) [7], mixed density clustering [8], fuzzy clustering represented by fuzzy C-means (FCM) [9], and graph clustering algorithms [10]. Hierarchical clustering is usually represented as a hierarchical tree structure [11]. According to the direction of constructing the tree structure, clustering can be divided into bottom-up [12] and top-down [13] construction methods. Traditional clustering algorithms are often limited to a criterion evaluation function. Only by selecting appropriate clustering evaluation criteria can an effective clustering result be obtained [14]. In the single-objective clustering algorithm [15], the evaluation function is often designed according to the dataset and the actual problem to be solved, and the number of clusters needs to be specified in advance. However, neither the cluster center nor the number of clusters can achieve adaptive discrimination [16,17]. Since clustering is often an unsupervised task with a general lack of prior knowledge [18], coupled with the explosion of data dimensions and sparse high-dimensional datasets, it is difficult for traditional single-objective clustering algorithms to obtain efficient clustering solutions [19].
The clustering algorithm is simple, fast, and efficient [20], but it is easy to fall into local optimum, the number of clustering classes needs to be specified in advance, and the robustness is not strong [21]. With the increasing scale and dimension of data in the real world, single-objective algorithms can no longer meet the modeling of more and more complex problems. Therefore, it is necessary to build a multi-objective algorithm with stronger problem description ability and seek a fast and efficient multi-objective optimization method [22]. In [23], Handl et al. proposed the multi-objective clustering algorithm, MOCK, and compared it with three single-objective clustering algorithms. Finally, it was found that the clustering algorithm can benefit from multiple objectives. Gong et al. [24] proposed an improved multi-objective clustering framework, IMCPSO, using particle swarm optimization. This framework designs a novel particle representation of clustering problems to help PSO search for clustering solutions in continuous space, and designs a leader selection strategy according to the distribution of Pareto sets, so that the algorithm can avoid falling into local optimum. The experimental part proves the advantages of multi-objective clustering algorithm compared with the single-objective clustering algorithm through a large number of experiments. In [25], Wang et al. considered the number of clusters and the square sum of the distance between the data point and its cluster centroid as the objective, and established a bi-objective clustering model, EMO-KC, to ensure that all clustering results with different k values are effectively obtained in a single run. Different from the single objective optimization problem, the multi-objective problem contains multiple conflicting optimization objectives [26], and a single solution cannot achieve the optimization of multiple objectives at the same time. Under the condition of satisfying the constraints, effectively balancing multiple conflicting objective values to solve multi-objective optimization problems (MOP) is critical [27]. From the perspective of optimization, the clustering problem can be regarded as an NP-hard problem, and traditional optimization methods cannot be directly applied to solve multi-objective problems [28]. As a meta-heuristic algorithm, the evolutionary algorithm is often used to solve NP-hard problems, so multi-objective optimization methods for clustering problems become very meaningful [29]. Handl et al. [23] first proposed a multi-objective adaptive clustering algorithm, MOCK, using PESA-II as an optimization tool. The algorithm takes the intra-class homogeneity and inter-class connectivity of clustering as optimization objectives, and initializes the population by minimum spanning tree and k-means. In the clustering phase, MOCK can adaptively adjust the number of clusters without specifying the number of clusters in advance. Finally, the method of gap statistics was used to select the best result from the non-dominated solutions generated by the multi-objective optimization algorithm. MOCK does not require a priori knowledge of the samples, and can adaptively determine the number of clusters, which is better than the single-objective clustering method. Hruschka et al. [30] introduced the general view of evolutionary multi-objective clustering (EMOC) in their review of evolutionary clustering algorithms. In 2012, Bong et al. [31] proposed a multi-objective clustering method applied to image segmentation. In 2013, Mukhopadhyay et al. [32] conducted a survey on multi-objective evolutionary methods for data mining, in which the authors introduced the general characteristics of EMOC and proposed some algorithms.
At present, most of the multi-objective clustering algorithms are often only for datasets with simple data distribution, and can only optimize one clustering task [33]. For multi-domain datasets with multiple related clustering tasks, because of the limitations in the expression of genotypes, the design of crossover operators, and population search strategies, multi-objective clustering algorithms cannot have a good performance in multi-domain datasets [34]. Simply applying the multi-objective clustering algorithm to complex clustering tasks will only increase the degree of negative transfer of genetic material between individuals and reduce the ability of the population to find the global optimal solution [35]. Therefore, this paper focuses on the positive transfer effect of evolutionary multi-task algorithms [36,37,38,39] between multiple clustering tasks. Gupta et al. [40] proposed a multifactorial evolutionary algorithm, MFEA, or an evolutionary multi-task algorithm, which uses a single evolutionary population to solve multiple optimization problems simultaneously by exploiting the implicit parallelism of population search.
In order to solve the above difficulties, this paper proposes a multi-task multi-objective automatic clustering model based on the MFEA framework and multi-objective clustering algorithm. The algorithm has multi-task solving ability to solve the problem of multi-domain datasets. The overall deviation and connectivity index are used as the two objective functions of clustering to obtain a set of non-dominated solutions with respect to the clustering parameters. Moreover, a location-based adjacency graph (LAG) encoding method was designed, and a mutation and crossover operator based on relevance learning between tasks was designed for information transmission between populations to reduce the impact of negative transfer on population convergence [41]. The specific contributions are as follows:
1. Firstly, the idea of evolutionary multi-task learning is integrated into the framework of multi-objective clustering, and the model of a multi-task multi-objective automatic clustering algorithm (MFMOCK) is established. Through the exchange of genetic material between individuals in the population, the relevant knowledge between different tasks is implicitly transferred, so as to optimize multiple tasks at the same time;
2. Based on the locus-based adjacency graph (LAG) encoding method, an individual representation suitable for evolutionary multi-tasks was designed, which provides a basis for the exchange of genetic information between tasks;
3. The crossover operator between individuals with different skill factors (solving different clustering tasks) was designed to limit the search direction of the algorithm, reduce the adverse evolution caused by random crossover, and accelerate the convergence speed of each task at the same time.
The remainder of this paper is organized as follows. Section 2 introduces the technical basis of the algorithm framework, including multi-objective optimization, MOCK, the multi-tasking evolutionary algorithm, and model correlation learning. Section 3 introduces the specific steps and implementation details of the multi-tasking and multi-objective clustering algorithm, including the innovation and improvement of the coding method, local search strategy, and individual crossover and mutation strategy. Section 4 records the setup and results of the comparison test, and analyzes the experimental results. Section 5 is the conclusion of this paper, which summarizes and describes the algorithms and experimental results proposed in this paper.

2. Background

In this section, the MOCK clustering model is firstly introduced in detail, and then the task correlation learning model is derived and its application in multi-task clustering is introduced.

2.1. Multi-Objective Optimization

Without loss of generality, the multi-objective optimization problems (MOPs) [42] discussed in this paper can be set to the minimum value. A MOP can be expressed as follows:
min F ( x ) = ( f 1 ( x ) , , f m ( x ) ) T subject to x Ω
where Ω is the feasible space, x is the solution of the MOP, R m is the target space, F : Ω R m consists of m real-valued objective functions. In most cases, the objectives in the MOP are contradictory, which means that it is impossible for any point in the feasible space to minimize all objectives at the same time. Therefore, the design of multi-objective optimization is to simultaneously find the best trade-off between them. For minimization, if and only if one solution x u is better than another solution x v
i = 1 , 2 , , m f i ( x u ) f i ( x v ) j = 1 , 2 , , m f j ( x u ) < f j ( x v )
If there is no solution in Ω such that F ( x ) dominates F ( x * ) , we call x * the Pareto-optimal solution. We make a series of F ( x * ) composed of x * become the Pareto-optimal vector. The objectives in the Pareto-optimal vector has such a relationship that the reduction of one objective will lead to the increase of other objective. All Pareto-optimal points constitute a set, which is called the Pareto-optimal set, and the corresponding Pareto-optimal objective vector is called the Pareto-optimal frontier (PF) [43].
Evolutionary algorithms are usually used as an important tool for solving multi-objective problems. Because of the inherent parallelism of population iteration in multi-objective evolutionary algorithms, a near-optimal solution set covering the entire Pareto front can be generated in one iteration [44]. Deb et al. proposed the NSGA-II algorithm using a fast non-dominated sorting strategy [45]. The MOPSO algorithm proposed by Coello and Pulido et al. adopts the adaptive selection strategy based on a grid, and applies the new mutation strategy to the population evolution to ensure the diversity of solutions [46]. Zhang et al. proposed the MOEA/D algorithm, which is a multi-objective evolutionary algorithm based on decomposition [47]. The algorithm takes the Chebyshev method, weight method, and boundary intersection method as examples, decomposes the multi-objective problem into multiple single-objective problems, and, at the same time, through the iterative optimization of sub-problems, it alleviates the issue of large search space and slow iterative evolution in multi-objective optimization, and reduces the time complexity of the algorithm.

2.2. MOCK Clustering Model

MOCK is a multi-objective clustering algorithm that can automatically determine the number of clusters. The algorithm consists of two main stages: the clustering stage and model selection stage.
Clustering stage: The clustering phase of MOCK is based on the multi-objective evolutionary algorithm: the Pareto envelope-based selection algorithm version 2 (PESA-II), which optimizes the two clustering objectives of overall deviation and connectivity. In this stage, the minimum spanning tree and k-means algorithm are used to generate the initial solution set to obtain different numbers and shapes of clusters. The overall deviation is calculated as follows:
D e v ( C ) = C k C i C k D ( i , μ k )
where C is the set of all clusters, μ k is the centroid of the cluster, and D ( · , · ) is the chosen distance function.
The connectivity calculation is as follows:
C o n n ( C ) = i = 1 N j = 1 L x i , n n i j , w h e r e x r , s = 1 j , i f C k : r C k a n d s C K 0 , o t h e r w i s e
where n n i j is the j-th nearest neighbor of datum i, N is the size of the clustered dataset, and L is a parameter determining the number of neighbors that contribute to the connectivity measure.
In MOCK, each individual is encoded using locus-based adjacency representation, where each data point is represented by a gene, and the allele value of the gene defines the connection between the data points. The offspring are generated by combining the genetic information of the two parents using the uniform crossover method. In addition, based on the concept of nearest neighbor, the neighborhood-based mutation is designed to change the connection relationship of data points.
Model Selection Stage: The model selection phase uses gap statistics to determine the number of clusters in the dataset. In addition, the Poisson model is used to generate control data in the feature space to obtain the expected connectivity and the overall deviation value of the unstructured data. Then, the solution and the control front are aligned and normalized, and the score of each point on the solution front is calculated. By analyzing the relationship between the score and the number of clusters, the local optimal solutions are identified, which are considered to be promising solutions. The global maximum may be considered as the best solution for estimation.
The specific implementation of MOCK can be found in [23].

2.3. Task Correlation Learning Models

Multi-task clustering improves the clustering performance of each task by transferring useful knowledge between related tasks [48]. In many practical applications, only part of the latent classes can be shared between tasks. If the correlation between tasks is not considered, it may lead to the problem of negative transfer. Zhang et al. proposed multi-task clustering with model relation correlation learning [49]. Intra-task clustering introduces symmetric non-negative matrix factorization through a linear regression model, and then clusters each task. Among tasks, the parameters of the linear regression model in each task are updated by learning the correlation of model parameters [50]. The block diagram of the algorithm for task relevance learning is as follows:
Given m tasks, each of which has a set of datasets, i.e., X t = x 1 t , x 2 t , , x n t t i d × n t , t = 1 , , m ; where n t is the amount of data for the tth task and d is the dimension of the feature vector. The similarity matrix of the tth task is given by M t i n t × n t . The data X t for each task are divided into h t clusters, i.e., C t = C 1 t , C 2 t , , C h t t . The cluster label matrix of the tth task is given by Y t i n t × h t .
The objective function of multi-tasking model relational relevance learning is formulated as follows:
min Y t , W t , G t s J = t = 1 T 1 2 M t Y t Y t T F 2 + t = 1 T λ Y t X t T W t F 2 + μ W t F 2 + α t = 1 T s t T i = 1 k t j = 1 k s W i t W j s 2 2 G i j t s + β G t s F 2 s . t . Y t 0 , i = 1 k t j = 1 k s G i j t s = 1 , 0 G i j t s 1
This formulation is mainly composed of two parts: intra-task clustering and inter-task correlation learning. The first part of intra-task clustering is mainly used to cluster within each task and update the clustering clusters through the knowledge transferred from other tasks. The formula for this part is as follows:
min Y t W t J i n t = 1 2 M t Y t Y t T F 2 +   λ Y t X t T W t F 2 + μ W t F 2 s . t . Y t 0 .
The first term of the formula is symmetric non-negative matrix factorization (SNMF), and M t is the similarity matrix of task t. The second and third terms of the formula are linear regression models, which are used to fit the clusters in each task, and the features of the input vectors are used to fit and predict the output vectors. However, W t is the regularization term, which can prevent overfitting.
The second part is inter-task correlation learning, which learns cross-task correlation by computing the correlation of model parameters between multiple clusters in each pair of tasks. The relationship of the same cluster across different tasks is explored by learning the correlation of the model parameters corresponding to each cluster, and then the model parameters of each task can be updated by the model parameters of other tasks.
min G ( t , s ) J cross t s = i = 1 h t j = 1 h s W i t W j s 2 2 G i j ( t , s ) + β G ( t , s ) F 2 , s . t . i = 1 h t j = 1 h s G i j ( t , s ) = 1 , 0 G i j ( t , s ) 1 ( t s ) ,
The first term of this formula can ensure that the models with a closer similarity distance have higher corresponding similarity, while the second term can prevent the situation that the similarity between similar models is 1, while the similarity between other models is 0. W j s is the linear regression parameter for the jth cluster of the task s, and G t s is the cluster correlation coefficient matrix between task t and task s, where G i j t s is the correlation coefficient between W j t and W j s .

3. Multi-Task Multi-Objective Automatic Clustering Algorithm

In order to make full use of the parallelism implied by a population search, and extend multi-objective optimization to multi-tasking multi-objective optimization to optimize multiple similar clustering tasks at the same time, this section proposes a multi-objective clustering learning framework based on evolutionary multi-tasking optimization, which is also a multi-tasking multi-objective clustering framework. It tries to improve the clustering performance of each task while accelerating the iteration speed of each task. This section introduces the specific MFMOCK framework according to the process and technical points, and presents the complexity analysis of the algorithm. Table 1 and Table 2 are the abbreviation declaration table and the nomenclature declaration table, respectively.

3.1. Framework of MFMOCK

This part gives the basic framework and specific algorithm steps of the multi-task multi-objective automatic clustering algorithm (MFMOCK). NSGA-II [45] is used as the basic framework of MOEA, and the algorithm extends the single-task multi-objective optimization algorithm into a multi-tasking evolutionary algorithm. The individual coding method, local search strategy, genetic information transmission method between tasks, and crossover and mutation operators are mainly modified for the evolutionary multi-tasking framework. The framework of the MFMOCK is shown in Algorithm 1.
The m clustering tasks are optimized simultaneously, and the jth task is denoted as T j , whose objective function is defined as f j : X j i , where X j is the search space. Then, the evolutionary multi-tasking clustering optimization proposed in this paper can be expressed as x 1 , , x m = argmin f 1 ( x ) , . . , f m ( x ) ; in order to solve this problem by using evolutionary algorithm, in the evolutionary multi-tasking framework, each individual p i , i { 1 , 2 , , | P | } in population P has a set of attribute characteristics, and p i in the m clustering optimization problems can be decoded as x 1 i , , x m i , where x 1 i X 1 , , x m i X m . Taking individual p i as an example, its attribute characteristics are as follows:
1. Individual evaluation value: The individual evaluation value of p i in task T j is Ψ i j = f i j , and the corresponding ranking position is r j i ;
2. Individual fitness: The individual fitness of P i is defined as φ i = 1 / min j { 1 , , m } r j i ;
3. Skill factor: The skill factor of P i is defined as τ i = arg min i r i j , j { 1 , , m } , that is, the number of tasks for which individual performance is optimal.
Algorithm 1 Pseudocode for the MFMOCK algorithm.
Input: evolution iterations g e n s m a x , Population size p o p , Crossover and mutation  probabilities p c and p m , the dataset containing m tasks.
 1: Initialization
 2: Initialize the population P.
 3: Assign skill factors τ and evaluate individuals.
 4: Initialize LearningPool.
 5: while g e n g e n m a x do
 6:     Individuals are selected to enter LearningPool based on the evaluation results
 7:     The parents are selected through a binary tournament
 8:     Generation of offspring population C, refer to Algorithm 2.
 9:     Individuals are evaluated according to the skill factor.
10:    Merging populations P C .
11:    Update the fitness φ and skill factor τ .
12:    The next generation of individuals is selected according to the φ and τ .
13: end while
14: A solution evaluation strategy based on controlled frontier distance is used to select the final solution.
Firstly, based on MOCK, MFMOCK uses the idea of evolutionary multi-tasking, in which each task contributes an additional factor that affects the evolution of the population. Moreover, the skill factor of an individual represents the task number that the individual is best at, and individuals with different skill factors contribute evolutionary factors to the population through chromosome crossover. Due to the unified individual coding method, the data and corresponding features in different tasks are mapped in the same search space, so that in the process of population evolution, each task also provides information for the optimization of other tasks. The unified search space means that the solutions of different tasks are contained in a unified genetic information base, which enables MFMOCK to exploit the potential genetic complementarity between multiple clustering tasks, so as to effectively discover useful genetic information and implicitly transfer it from one task to another. Secondly, because the amount of data in different clustering tasks may be inconsistent, the genotype length of individuals in different tasks obtained by the classical LAG encoding method is inconsistent, which makes it impossible for individuals in the population to cross. The multi-task adjacency representation method is used to solve this problem.
The algorithm also adopted the elite retention strategy of LearningPool, which set a fixed size of LearningPool to maintain a large number of different non-dominated solutions with different skill factors found in the search process. Using the histogram method based on density estimation, the solutions cover the whole objective space instead of gathering in the same region. At the end of each generation, the non-dominated solutions in the LearningPool are updated, new non-dominated solutions are added, and the dominated solutions are eliminated. Multi-objective clustering does not generate a single solution, but a set of clustered solution sets, with different solutions corresponding to different trade-offs between the two objectives. The model selection strategy of the algorithm adopts the solution selection strategy based on the control front (CF). According to the Poisson model in the feature space of the clustered data, the control data are obtained. Specifically, the principal component analysis method is applied to the covariance matrix of the original data to obtain the eigenvectors and eigenvalues, and then the control data are generated. Thus, an estimate of the connectivity and overall bias of unstructured data, which do not contain the original data but are located in the same data space, is obtained. For the non-dominated solution on the actual Pareto front (PF), the distance between the solutions and the point on the control front is calculated, and the non-dominated solution with the largest shortest distance is considered to be the optimal solution. The calculation formula is as follows:
BestSolution = argmax i P F min j C F ( distance ( i , j ) )
In addition, the crossover operator of multi-tasking correlation learning is combined, which not only enables the individuals to exchange genetic information between tasks, but also uses the correlation learning model between tasks to quadratic program the linear regression model parameters between different tasks in the process of individual crossover to extract the correlation between task models. The correlation coefficient is used to minimize the difference between the parameters of different regression models, so that the related tasks have similar clustering results, and the knowledge related to the transfer task is realized, which is ultimately reflected in the change of individual genotypes. Moreover, as a population search strategy, the crossover operator avoids meaningless crossover between individuals with different skill factors, and accelerates the search speed of the population individuals for the global optimal solution in the evolution process.

3.2. Multi-Task Adjacency Coding Based on Locus Adjacency Graph

In order to apply the evolutionary multi-tasking framework to solve multi-objective clustering problems, it is necessary to modify and design the population coding method, the crossover operator between individuals with different skill factors in the population, and the evolution strategy. The above contents are introduced in details in this subsection.
In the coding method of the population, the coding method adopted in this algorithm is the multi-tasking adjacency coding method based on the trajectory adjacency graph (TAG). In the TAG, the encoding of each individual can be represented as a data connection graph, and can be decoded as a clustering result. Among them, each data point in the task is represented as a node in the graph, the connected data are divided into the same cluster, and each individual corresponds to k clusters after the algorithm runs. For example, individual p i contains n alleles p i j , j { 1 , , n } , if the gene at position f is denoted as l, then the TAG of this gene is denoted as p i f p i l , and the data points f and l are connected and belong to the same cluster. The main advantage of the trace-based adjacency coding scheme for solving the clustering problem is that the number of clusters does not need to be fixed in advance, since it is determined automatically in the decoding step.
In Figure 1, taking the clustering task as an example, the task has nine pieces of data, each data point represents a node i { 1 , 2 , , 9 } , the genotype of the individual represented in the figure is 2 , 4 , 2 , 1 , 5 , 5 , 9 , 7 , 8 , the solid line represents the connection track between the samples, the dotted line represents the corresponding cluster, and all the data are divided into three categories. Each bit of the gene in the individual genotype represents the sequence number of the node that is pointed to from the current node. In this representation, starting from a node, only a connecting line can lead to another node.
In a single multi-objective clustering task, since the number of samples is fixed, the genotype length of each individual in the population is consistent, and the gene length problem does not need to be considered for interindividual crossover. However, in the multi-tasking environment, the sample number of each task may not be the same, so when the same individual is decoded into different tasks, the genotype length required is different. Therefore, a unified encoding method, namely, the multi-tasking adjacency representation method, needs to be designed. As shown in Figure 2, the genotypes of individuals are all non-negative real numbers, so the non-negative real numbers need to be mapped into [0, 1]. When optimizing m tasks simultaneously, assume that the dimension of the jth task is D j ; thus, in the unified search space, its dimension D m u l t i t a s k = m a x ( D j ) . In the population initialization step, each individual is a vector of random variables of dimension, and each gene is within a fixed range [0, 1]. If the skill factor τ i = j of individual i, the first D j genes of individual i need to be taken for individual crossover, mutation, coding, and decoding operations.
In the initialization process of the algorithm, the initialization strategy based on MST and k-means is mainly used, because different single-objective clustering algorithms have good clustering effects in different regions of the Pareto front, and the algorithm based on connectivity often generates the optimal solution in those regions of the Pareto front with low connectivity. However, compactness-based algorithms perform well in regions with low overall bias. However, MST and k-means are clustering algorithms based on connectivity and compactness, respectively, so the initial solution generation strategies of MST and k-means are used.
Since different clustering objectives can reflect completely different optimization criteria for clustering solutions, two complementary objective functions are selected in this algorithm: one is based on compactness and the other is based on cluster connectivity. Cluster compactness is expressed by calculating the overall deviation (Dev) of different cluster partitions, which can be calculated as the total distance between a data point and its corresponding cluster center:
Dev ( C ) = C k C i C k δ i , u k
where C is the set of clusters, u k is the cluster center, and δ ( , ) is the distance function (we will choose Euclidean distance here). As the objective function, the overall bias should be minimized. Another objective function that reflects the clustering connectivity criterion is the connectivity index (Con), which evaluates the degree to which neighboring data points are classified into the same cluster, and is formulated as follows:
Con ( C ) = i = 1 N j = 1 L x i , n i j x r , s = 1 j , if C k : r C k s C k 0 , otherwise
where n i j is the jth nearest neighbor of data i, N is the size of the clustered dataset, and L is a parameter that measures the number of neighbors that contribute to the connectivity. As an objective function, the connectivity index should be minimized.

3.3. Evolutionary Operators Based on Correlation Learning

According to the design principle of MFEA, two randomly selected parents must meet certain conditions before they can cross. During crossover, parents with the same cultural background (skill factor) prefer the exchange of genetic material between them. Therefore, if two randomly selected parents have the same skill factor τ , they can freely cross. Crossover is performed according to random mating probability ( r m p ), otherwise mutation occurs. The pseudocode for the algorithm is as follows (Algorithm 2):
Algorithm 2 Mating selection among individuals
 1: Two parent individuals P a and P b are randomly picked from the contemporary population P.
 2: Generate a random number Rand between 0 and 1.
 3: if  τ a = τ b  then
 4:     Parent individuals P a and P b perform intra-task crossover to produce offspring individuals C a and C b .
 5: else if rand < rmp then
 6:     Parent individuals P a and P b perform inter-task crossover according to Algorithm 3 to produce. offspring individuals C a and C b .
 7: else
 8:     Mutation of parent individual P a produces offspring individual C a .
 9:     Mutation of parent individual P b produces offspring individual C b .
10: end if
In the proposed algorithm, individuals with the same skill factor correspond to the same clustering task. If the skill factor of two parents is the same, the two individuals correspond to the same clustering task, and the number of samples, the individual coding length, and the data distribution corresponding to the same clustering task are the same, so the crossover between individuals can be carried out according to the simulated binary crossover (SBX). In addition, the transfer of learning between tasks is also crucial because there is related knowledge between tasks. Although the evolutionary multi-task framework has great advantages in solving multi-tasking problems by using evolutionary algorithms, the exchange of genetic material between tasks of the original MFEA algorithm cannot accelerate the convergence of the clustering task, and the crossover between individuals with different skill factors may introduce the learning effect of the face. So, it is necessary to design new crossover operators between individuals with different skill factors. The new crossover operator based on multi-tasking introduces the idea of model parameter learning between tasks, so that when individuals with different skill factors exchange genetic material, the direction of individual change can be restricted to the direction of clustering learning convergence. The pseudocode for the algorithm is as follows (Algorithm 3):
Algorithm 3 Evolutionary operators based on correlation learning
 1: Step1: Decoding.
 2: The genotypes of parent individuals P a and P b are decoded according to the number of samples in the task, whose skill factors are τ a and τ b , respectively, and the corresponding clustering results r a and r b are obtained by decoding.
 3: Step2: Initialization.
 4: The cluster labels h t of the inter-task correlation learning model are initialized according to the clustering results r a and r b .
 5: Initialize the similarity matrix M t based on the similarity measure of the dataset (in this case, Euclidean distance).
 6: The category indicator label Y t of the task is initialized according to the k-means algorithm.
 7: Two parent individuals P a and P b are randomly picked from the contemporary population P.
 8: W t will be initialized to an all-unity matrix of n t × h t size.
 9: Step3: Learning inter-task parameter relevance.
10: for  t = 1 m  do
11:     for  s = 1 m  do
12:         if  s t  then
13:             The task relevance parameters G ( t , s ) are calculated.
14:         end if
15:     end for
16:     Calculating model parameters W t .
17:     Compute the cluster label matrix Y t .
18: end for
19: Get the task τ a and task τ b clustering results C i and C j .
20: The genotypes of P a and P b are updated according to the clustering results C i and C j to obtain the offspring individuals O a and O b .
When the individuals with different skill factors are crossed, the sample numbers n a , n b of the corresponding task are obtained according to the skill factors of the parents P a and P b , the first n genes of the individual are decoded, the genes in [0, 1] are linearly mapped to [1, n], and the corresponding clustering results r a and r b of the parents are obtained. The cluster label of each sample corresponding to the task is initialized according to the clustering result, and other related matrices are initialized according to the initialization strategy in the algorithm block diagram. In step 3, inter-task correlation learning, the clustering results will be updated through intra-task linear model learning and inter-task model parameter correlation learning. In this step, not only the genetic information between individuals with different task preferences is exchanged, but also the convergence of different tasks corresponding to individuals is accelerated. In addition, in the design of the crossover operator, it is not necessary to optimize the task correlation learning model to complete convergence, but to iterate a certain number of times. Therefore, the evolution strategy of local search is implied in the process of individual crossover, which not only limits the random evolution of the population individuals, but also reduces the time complexity of the algorithm. After the local search strategy, new clustering results C i and C j will be obtained, where, C i is a clustering result of the ith task, and C j is a clustering result of the jth task. It is necessary to encode C i and C j as new offspring individuals O a and O b . In the encoding process, for example, the connection in P a that is not the same as the clustering result indicated by C i is disconnected, the nearest sample belonging to the same cluster is found in the nearest neighbor matrix of the disconnected sample, and the trajectory indicates the new sample to form a new offspring individual. The specific operation is shown in Figure 3. Figure 3a is the genotype representation of the parent individual P a . In the figure, sample point 3 belongs to the same cluster as sample point 2, and there is a trajectory pointing to sample point 2. Figure 3b shows the clustering results obtained after learning the inter-task model relationship in the crossover operator. In Figure 3b, sample point 3 is divided into another cluster, so the adjacent trajectory between sample point 3 and 2 needs to be disconnected, and the adjacent trajectory of sample point 3 needs to be found in the new cluster. Since sample point 9 is the closest point to sample point 3 in the new cluster, the gene O a at the location of offspring genotype 3 is changed to 9. Finally, the genes in [1, n] are linearly mapped to [0, 1] to complete the process of crossover between individuals and the generation of offspring individuals.
When the parent individuals with the same skill factor are crossed, the uniform crossover mode is selected. Because there is no bias in gene ranking for individuals with the same task, the uniform crossover is used to generate any combination of alleles from the two parent individuals. When an individual carries out genotype mutation, the mutation operator based on finite neighbors is used. For the dataset of size N, the search space during mutation is N N . Using the mutation operator based on finite neighbors, each gene can only connect to one of the L nearest neighbors when it mutates, where L N ; then, the search space will become L N . This not only effectively reduces the amount of calculation during the mutation, but also reduces the generation of a lot of meaningless connections. Since it is impossible for each individual generated in MFEA to perform well in all tasks, the individual can only be evaluated according to the task with the best performance. Therefore, the changes of individual genetic factors during the whole evolution process refer to the concept of vertical cultural transmission, based on the vertical cultural transmission algorithm through selective imitation. The skill factor of an individual can mimic the skill factor of any of its parents.

3.4. Analysis of Algorithm Complexity

Given a dataset with m tasks, assuming that the data dimension of each task is n d , the algorithm complexity of the task correlation learning step is O ( n 2 d ) , and the time complexity of fast non-dominated sorting is O ( p o p 2 ) , where p o p is the size of the subpopulation. In the initialization process, the time complexity of generating the initial solution using MST is O ( m n 2 ) , and the time complexity of generating the initial solution using k-means is O ( n ( k + 2 ) m / d ) . In addition, the time complexity of each individual function evaluation is O ( m a x { n 2 , k n d } ) , where k { 1 , , K m a x } is the number of clusters. The time complexity of the MFMOCK algorithm is given by O ( g e n m a x { p o p n , p o p k n d , p o p n 2 d , p o p 2 } ) .

4. Experiment Setup and Results

In this section, firstly, three clustering datasets are introduced, secondly, the comparison algorithms and evaluation indicators are introduced, and then the parameters are analyzed and set. Finally, the clustering results and algorithm convergence are compared and analyzed, respectively. All programs ran on the CPU (13th, Gen Intel(R) Core(TM) i9-13900-k, 3.00 GHz), RAM was 32 GB, and the program was realized by the MATLAB R2023a.

4.1. Datasets

In order to evaluate the proposed algorithm model, this experiment used three different sets of experimental data for testing. The characteristics of all datasets are described in Table 3, Table 4 and Table 5. The first set of datasets was a manually generated two-dimensional dataset, which was used to verify the robustness of the algorithm to the overlap between clusters, and the size and the shape of clusters. The data in these datasets are two-dimensional normally distributed data with a fixed cluster size, mean vector, and standard deviation vector. The instance distribution of the manual dataset is shown in Figure 4.
The second set of datasets was a standard cluster model generated by a random generator and conforming to a multivariate normal distribution. In the first random generator, randomly generated low-dimensional datasets have random directions, whereas in higher dimensions, the clusters are more spherical in shape. In the second generator, the data distribution is more biased towards arbitrarily distributed ellipsoids. The generator and the dataset used in this article are available at https://github.com/garzafabre/Delta-MOCK (accessed on 14 May 2024). The third set of datasets was UCI datasets, all available on the UCI website https://archive.ics.uci.edu/ (accessed on 14 May 2024).

4.2. Comparison Algorithm and Evaluation Index

The comparison algorithms in the experiment are as follows: K-means (KM) is a traditional clustering method based on Euclidean distance; single linkage (SL) is a hierarchical clustering algorithm that defines cluster proximity as the distance between the two closest points of two different clusters; average linkage (AL) is a hierarchical clustering algorithm that defines cluster proximity as the average of the proximity of pairs of points taken from two different clusters; spectral clustering (SC) is a clustering algorithm based on spectral graph theory that transforms the clustering problem into an optimal graph partition problem [51]; fuzzy C-means (FCM) is a clustering algorithm based on fuzzy C-means; MOCK is a multi-objective clustering algorithm that automatically determines the number of clusters and Multiple Information Exchange Multi-objective Clustering Algorithm Based on MOCK (MIE-MOCK) is a MOCK variant clustering algorithm with a random crossover operator pool and random mutation operator pool [52].
To accurately evaluate the clustering results, this paper uses performance metrics such as accuracy, mutual information, and adjusted Rand index, which are standard metrics widely used for clustering. In addition, the introduction of the adjusted Rand index performance index is as follows:
Clustering accuracy (ACC): Clustering accuracy discovers the one-to-one relationship between clustering labels and actual class labels, and measures the extent to which each cluster contains data points from the actual classes. The clustering accuracy is defined as follows:
A c c = i = 1 n δ ( m a p ( r i ) , l i ) n
where r i represents the cluster label of data x i , l i represents the actual class label of data x i , n represents the total amount of data, δ ( x , y ) is the Delta function, equal to 1 if x = y , equal to 0 otherwise, and m a p ( r i ) is the mapping function, which maps each cluster label r i to an equivalent label in the dataset.
Normalized mutual information (NMI): The second metric is normalized mutual information (NMI), which is used to determine the quality of the clusters. Given the clustering results, NMI is defined as follows:
N M I = i = 1 c j = 1 c n i , j log n i , j n i n ^ j i = 1 c n i log n i n j = 1 c n ^ j log n ^ j n
where n i denotes the number of data contained in cluster C i ( 1 i c ) , n ^ j denotes the number of data contained in cluster L j ( 1 j c ) , and n i , j denotes the intersection of the true class L j and cluster C i . A larger NMI indicates a better clustering result.
Adjusted Rand index (ARI): This computes the probability that two instances of two clusters belong to the same cluster or different clusters, defined as follows:
A R I 1 = i = 1 k a j = 1 k b n i j 2 i = 1 k a n i 2 j = 1 k j n j 2 / n 2 A R I 2 = 1 2 · i = 1 k a n i 2 + j = 1 k b n j 2 i = 1 k a n i 2 · j = 1 k b n j 2 / n 2 A R I = A R I 1 / A R I 1
where n i j is the number of identical instances in the c i cluster in solution x a and the c j cluster in solution x b , n i is the number of instances in the c i cluster in solution x b , n j is the number of instances in the c j cluster in solution x b , and k a and k b are the number of clusters in solution x a and solution x b , respectively.
Hypervolume [53] is an index conforming to the Pareto dominance concept and the value of hypervolume represents the volume of hypercube surrounded by individuals in the solution set and reference points in the target space. Therefore, the hypervolume can be used as the evaluation index of the convergence and distribution of the solution set. A higher hypervolume value represents a better convergence and distribution.

4.3. Parameter Analysis and Setup

The experiments in this chapter were carried out on the basis of the datasets proposed in Section 4.1. The characteristics of each dataset are shown in Table 3, Table 4 and Table 5. The number of clusters in all experiments was set to the number of real classes of the data. In the experiments of this chapter, each algorithm was run independently 10 times, and the average of the final results was calculated. The proportion of correctly classified samples was taken as the clustering accuracy, and the accuracy of the algorithm was reflected by the accuracy and mutual information. Take the Sizes dataset and 50d-20c as an example, Figure 5 is the parameter change diagram of MFMOCK, showing the change relationship between the random mutation probability r m p and population size p o p , where the population evolution generation g e n was set to 100. Under different random mutation probabilities, the clustering accuracy was low when the population size p o p was small, and when pop was not less than 50, the clustering accuracy of MFMOCK gradually became stable. Because the data distribution of the Sizes dataset was relatively simple, the number of classes was small, the dimension was low, and it was difficult to observe the effect of r m p on the clustering results. However, can be clearly seen for the clustering results of MFMOCK on the 50d-20c dataset, under the same population size, when r m p was 0.9, the clustering accuracy of the algorithm was higher. Combined with parameter experiments, considering the stability and time complexity of the MFMOCK algorithm, p o p , g e n and r m p were set to 100, 100, and 0.9, respectively.

4.4. Comparison Experiment of Clustering Results

In this clustering performance test experiment, the comparison algorithms were single-task clustering algorithms, that is, there was only one clustering task in each clustering, and the dataset under this task conformed to the assumption of independent and identical distribution. The MFMOCK algorithm is a clustering algorithm based on evolutionary multi-tasking. In order to verify that MFMOCK algorithm effectively shared data features between multiple tasks and learns clustering tasks, it was necessary to design multi-task clustering data for each test dataset for MFMOCK. According to the distribution characteristics of the dataset, a multi-tasking auxiliary test dataset containing relevant knowledge (the data distribution of some categories was consistent) was generated. In addition, the number of categories and samples in the multi-tasking auxiliary test dataset may not be consistent with the original dataset. Moreover, the random sample points were added to the auxiliary test set to test the multi-tasking clustering learning performance of the algorithm.
In this experiment, each algorithm was run 10 times, and the average accuracy of the clustering results under the optimal parameters was selected. This section compares common single-objective and multi-objective clustering algorithms such as: KM, SL, AL, SC, FCM, MOCK, MIE-MOCK, etc. The algorithm results of this experiment are shown in the following table, where the performance indicators Acc and NMI given in the table are in percentage units, and the numerical range of ARI index is [−1, 1].
Table 6 shows the clustering results of all algorithms on manually generated datasets. The test was mainly to detect the clustering performance of each algorithm in datasets with different shapes and distributions. As shown in the table, different classical algorithms had different preferences for clustering datasets with different shapes. The KM algorithm showed excellent performance in clustering the Square and Sizes datasets, and the single join algorithm performed well in the Smile and Long datasets. Based on the objective function of overall deviation and connection index, the MFMOCK algorithm did not make any assumptions about the shape of different clusters, so it could detect clusters of arbitrary shapes. However, it had poor performance for datasets with overlapping data of different categories. In addition, the MFMOCK algorithm had a higher quality of clustering results for the Sizes and Triangle datasets, which was better than the other algorithms. For the Square dataset, because the dataset is relatively simple, the amount of data and data categories are small; thus, the accuracy of the classical clustering algorithm is already high, and the accuracy improvement effect of the MOCK-like algorithm based on the evolutionary algorithm was not obvious. The clustering results of the whole dataset showed that the MFMOCK algorithm had a partial improvement in clustering performance compared with the traditional single-objective clustering algorithm.
Table 7 shows the random datasets with specific dimensions and the number of cluster categories generated by the random generator to test the clustering performance of various clustering algorithms under different combinations of dimensions and number of categories. In the manually generated dataset, the data distribution presented a lot of geometric shapes that rarely appear in practical applications, and the individual variables were uncorrelated with each other, so the second randomly generated dataset was used. Compared with the manual dataset, the random dataset contained test data with higher dimensions and more categories, its dimensions varied from 2 to 100 dimensions, and the number of cluster categories varied from 4 to 40 categories. The randomly generated dataset was used for testing. This was also to test whether the clustering advantages of MFMOCK in manual datasets could be extended to more realistic, general, high-dimensional large-scale datasets. The experimental results showed that the classical single-objective clustering algorithms (e.g., KM and SC) had a better clustering effect when the data sample had a low dimension number and few categories. When the clustering dimension was increased to 10 dimensions and the number of classes was 10 or 20, the performance of classical single-objective clustering was significantly decreased compared with the evolutionary multi-objective clustering algorithm. As the sample dimension and the number of categories continued to increase, MFMOCK showed excellent clustering performance compared with single-task evolutionary multi-objective clustering algorithms such as MOCK and MIEMOCK. Compared with low-dimensional and low-class datasets, MFMOCK had a more obvious performance improvement when solving high-dimensional clustering problems with a large number of classes. The introduction of the idea of evolutionary multi-tasking made the evolutionary clustering algorithm exhibit better performance in the face of complex clustering datasets.
Table 8 shows the comparison of the experimental results of the comparison algorithms on seven UCI datasets in terms of accuracy, mutual information, and adjusted Rand index, where Append is short for the Appendicitis dataset and Aggregate is short for the Aggregation dataset. From the perspective of various indicators, MFMOCK had the best results in the four datasets among the comparison algorithms, which were Thyroid, Aggregate, Cancer, and Wine. For the dataset Zoo, MFMOCK was about 0.02 worse than the best AL in terms of clustering results, and it was optimal for the RAI index. For the dataset Jain, the clustering results of MFMOCK were about 0.05 worse than the best MOCK, and the other indicators were lower than the MOCK algorithm. For NMI and RAI, MFMOCK was higher than other algorithms in most datasets, and in the datasets with suboptimal clustering performance, the interpolation between MFMOCK and the optimal algorithm was small, i.e., within 0.07 .

4.5. Comparison Experiment of Algorithm Convergence

In order to verify that MFMOCK effectively exchanges genetic factors among multiple clustering tasks, which promotes the convergence of the problem and the improvement of clustering performance, we selected the comparison experiment between MOCK and MFMOCK. MOCK only clusters for a set of datasets, while MFMOCK simultaneously clusters the dataset and the auxiliary clustering test set. The auxiliary clustering test set contains some categories with the same data distribution as the clustering test set, and the number of categories and samples contained in the auxiliary test set may not be the same as the clustering dataset.
In order to judge the output of MFMOCK based on the quality of the solution and the robustness of the algorithm, the following parameters should be considered:
1. Average (Avg): The average function value of MFMOCK over all evolution rounds;
2. Coefficient of variation (CV): The ratio of the standard deviation of a function to its mean;
3. Average gap (AG): The difference between the mean and the true value. Expressed as a percentage of the true value.
In all of the following experiments, the population consisted of 100 individuals and evolved for 200 generations. Both MFMOCK and MOCK used the LAG-based chromosome representation method, and corresponding genetic operators, the SBX crossover operator, and Gaussian mutation operator were used to search the solution space; Moreover, r m p in MFMOCK was set to 0.7. Table 9 shows the test data in this comparison experiment. In order to test whether MFMOCK could optimize multiple similar clustering tasks at the same time and improve the performance of each clustering task, three groups of multi-tasking test sets were constructed, and different tasks had different test focuses on MFMOCK. MOCK only performed the clustering problem of the dataset corresponding to task 1. The specific differences are shown in Table 9.
The two tasks in dataset 1, Square and Sizes, had the same number of samples, feature dimensions, and classes. They both had spherical Gaussian clusters in their sample distribution, but they were manually generated, so they had different distributions. Figure 6a shows the convergence change trend of the hypervolume function values of MFMOCK and MOCK on dataset 1. In the initial evolution stage, the function curve of MFMOCK was better than the performance of MOCK, showing better overall convergence performance. Since MOCK and MFMOCK are completely consistent in the encoding way of population individuals and the design of crossover and mutation operators, the performance improvement can be attributed to the ability to transfer the relevant clustering knowledge between tasks by using implicit genetic, and the ability to mine similar data features by multi-tasking correlation learning. The hypervolume value of MFMOCK was better than that of MOCK in most cases, so the Pareto solution set of MFMOCK was more evenly distributed. In clustering problems, this area usually reflects the distribution and separation degree of different clusters, which may mean higher separation between clusters. In addition, hypervolume also considers the convergence and diversity of the solution set. In clustering analysis, this means that the algorithm can find clusters of different shapes and sizes while keeping the center points of these clusters as close as possible to the real situation. These advantages led MFMOCK to achieve better performance on clustering problems. In terms of the quality of the solutions generated during multi-tasking evolution, the results in Table 10 show that the MFMOCK algorithm performed well when dealing with the multi-tasking clustering of the manual dataset, the coefficient of variation of the overall bias and connection index in the two tasks were within 0.02 , and the final clustering solution was within 0.03 of the objective function value of the real classification cluster. This illustrates the excellent performance of MFMOCK in clustering handcrafted datasets.
The two tasks in dataset 2 were the 100d-10c and 100d-40c datasets, which only had the same feature dimension and were inconsistent in terms of the number of samples and classes. As shown in Figure 6b, the test environment of multi-domain data was simulated for the case of different number of categories and inconsistent data distribution. Compared with MOCK, the function value convergence trend in the figure shows the superior convergence of MFMOCK. Table 11 shows the performance of the MFMOCK algorithm in dataset 2, where the algorithm had a good clustering effect for task 2, the coefficient of variation of the target value was within 0.025 , and the deviation between the true value and the final value was about 0.02 . Due to the complexity of data and categories, the clustering performance of the algorithm was worse in task 1 than in task 2.
As shown in Table 12, the two tasks in dataset 3 were the 50d-20c-1 and 50d-20c-2 datasets, which differed only in the number of samples and had the same feature dimensions and number of classes, as well as a similar data distribution because they were generated by the same random generator. As shown in Figure 6c, the test environment of multi-domain data was simulated for the case of the same number of categories but with an inconsistent data distribution. Compared with MOCK, MFMOCK showed good convergence speed and clustering accuracy. In MFEA, the adopted multi-tasking was completely oblivious to multi-tasking, and it is considered that evolutionary multi-tasking does not necessarily guarantee that the performance of each task is improved, because not all genetic transfers are always useful. While certain tasks are positively affected by the implicit genetic transfer available during multi-tasking, certain other tasks may be negatively affected. In the MFMOCK algorithm proposed in this chapter, for multiple related multi-objective clustering problems, the strategy of random crossover between different individuals was abandoned, and multi-tasking correlation learning was introduced into the crossover operator. Only knowledge was transferred between tasks with a strong correlation, which reduced the possibility of negative transfer. The experimental results show that the algorithm is effective in transferring useful genetic information and accelerating the convergence of the algorithm.

5. Conclusions

In this paper, a multi-objective clustering framework based on evolutionary multi-tasking is proposed. By introducing MFEA into the evolutionary multi-objective optimization algorithm, evolutionary multi-objective clustering is extended from single-task learning to multi-tasking learning, and similar knowledge in different clustering tasks is shared by using the implicit search parallelism in the population. Firstly, by building an evolutionary multi-tasking framework, the algorithm improves the coding method of population individuals on the basis of MOCK, which provides a basis for individuals with different skill factors in the population to exchange genetic material. In addition, a new crossover operator between individuals was designed, and the idea of correlation learning between tasks was introduced into the crossover operator. By controlling the number of iterations of the learning algorithm, the crossover operator plays the role of a local search, so that individuals representing different clustering tasks can transfer the relevant knowledge between tasks when exchanging genetic material, and perform evolution towards the convergence direction of the clustering algorithm. By limiting the search direction of individual evolution, the evolution speed of the population is accelerated. In this paper, comparative experiments were carried out on manual, randomly generated and UCI datasets, and the multi-tasking learning ability and the final clustering effect of the model were verified. Through comparative experiments, the feasibility and effectiveness of evolutionary multi-tasking in accelerating multi-objective clustering learning were verified. In the future, we will study more objective functions and verify the proposed framework on more complex datasets. Moreover, we plan to improve the performance of MFMOCK to solve more diverse clustering problems in the real world. In addition, we will focus on studying more efficient individual coding methods to improve the optimization efficiency of MFMOCK.

Author Contributions

Conceptualization, Y.W., R.Y. and M.G.; software, K.D., L.L. and H.L.; formal analysis, K.D.; investigation, Y.W. and K.D.; resources, Y.W. and K.D.; data curation, Y.W. and K.D.; writing—original draft preparation, Y.W., K.D. and L.L.; writing—review and editing, R.Y. and M.G.; visualization, K.D. and L.L.; supervision, Y.W. and H.L.; project administration, R.Y. and M.G.; funding acquisition, H.L. and M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62036006, and in part by the Fundamental Research Funds for the Central Universities.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pande, S.; Sambare, S.; Thakre, V. Data clustering using data mining techniques. Int. J. Adv. Res. Comput. Commun. Eng. 2012, 1, 494–499. [Google Scholar]
  2. Mocanu, A.A.; Iftene, A. PuzzleNN: A neural network for image segmentation based on clustering. In Proceedings of the 2023 International Conference on Innovations in Intelligent Systems and Applications (INISTA), Hammamet, Tunisia, 20–23 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
  3. Camacho, D.; Panizo-LLedot, A.; Bello-Orgaz, G.; Gonzalez-Pardo, A.; Cambria, E. The four dimensions of social network analysis: An overview of research methods, applications, and software tools. Inf. Fusion 2020, 63, 88–120. [Google Scholar] [CrossRef]
  4. Antons, D.; Grünwald, E.; Cichy, P.; Salge, T.O. The application of text mining methods in innovation research: Current state, evolution patterns, and development priorities. R&D Manag. 2020, 50, 329–351. [Google Scholar]
  5. Sinaga, K.P.; Yang, M.S. Unsupervised K-means clustering algorithm. IEEE Access 2020, 8, 80716–80727. [Google Scholar] [CrossRef]
  6. Mehta, V.; Bawa, S.; Singh, J. Analytical review of clustering techniques and proximity measures. Artif. Intell. Rev. 2020, 53, 5995–6023. [Google Scholar] [CrossRef]
  7. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A k-means clustering algorithm. J. R. Stat. Soc. Ser. Appl. Stat. 1979, 28, 100–108. [Google Scholar] [CrossRef]
  8. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the KDD, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  9. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  10. Schaeffer, S.E. Graph clustering. Comput. Sci. Rev. 2007, 1, 27–64. [Google Scholar] [CrossRef]
  11. Nielsen, F.; Nielsen, F. Hierarchical clustering. In Introduction to HPC with MPI for Data Science; Springer: Berlin/Heidelberg, Germany, 2016; pp. 195–211. [Google Scholar]
  12. Bouguettaya, A.; Yu, Q.; Liu, X.; Zhou, X.; Song, A. Efficient agglomerative hierarchical clustering. Expert Syst. Appl. 2015, 42, 2785–2797. [Google Scholar] [CrossRef]
  13. Xiong, T.; Wang, S.; Mayers, A.; Monga, E. DHCC: Divisive hierarchical clustering of categorical data. Data Min. Knowl. Discov. 2012, 24, 103–135. [Google Scholar] [CrossRef]
  14. Hancer, E.; Xue, B.; Zhang, M. A survey on feature selection approaches for clustering. Artif. Intell. Rev. 2020, 53, 4519–4545. [Google Scholar] [CrossRef]
  15. Yan, D.; Cao, H.; Yu, Y.; Wang, Y.; Yu, X. Single-objective/multiobjective cat swarm optimization clustering analysis for data partition. IEEE Trans. Autom. Sci. Eng. 2020, 17, 1633–1646. [Google Scholar] [CrossRef]
  16. Mok, P.Y.; Huang, H.; Kwok, Y.; Au, J. A robust adaptive clustering analysis method for automatic identification of clusters. Pattern Recognit. 2012, 45, 3017–3033. [Google Scholar] [CrossRef]
  17. Bhatia, S.K. Adaptive K-means clustering. In Proceedings of the FLAIRS, Miami Beach, FL, USA, 17–19 May 2004; pp. 695–699. [Google Scholar]
  18. Wang, S.; Zhao, D.; Zhang, C.; Guo, Y.; Zang, Q.; Gu, Y.; Li, Y.; Jiao, L. Cluster alignment with target knowledge mining for unsupervised domain adaptation semantic segmentation. IEEE Trans. Image Process. 2022, 31, 7403–7418. [Google Scholar] [CrossRef] [PubMed]
  19. Abu Khurma, R.; Aljarah, I. A review of multiobjective evolutionary algorithms for data clustering problems. In Evolutionary Data Clustering: Algorithms and Applications; Springer: Singapore, 2021; pp. 177–199. [Google Scholar]
  20. Ikotun, A.M.; Ezugwu, A.E.; Abualigah, L.; Abuhaija, B.; Heming, J. K-means clustering algorithms: A comprehensive review, variants analysis, and advances in the era of big data. Inf. Sci. 2023, 622, 178–210. [Google Scholar] [CrossRef]
  21. Ahmad, M.F.; Isa, N.A.M.; Lim, W.H.; Ang, K.M. Differential evolution: A recent review based on state-of-the-art works. Alex. Eng. J. 2022, 61, 3831–3872. [Google Scholar] [CrossRef]
  22. Kwon, K.; Seo, M.; Min, S. Efficient multi-objective optimization of gear ratios and motor torque distribution for electric vehicles with two-motor and two-speed powertrain system. Appl. Energy 2020, 259, 114190. [Google Scholar] [CrossRef]
  23. Handl, J.; Knowles, J. An evolutionary approach to multiobjective clustering. IEEE Trans. Evol. Comput. 2007, 11, 56–76. [Google Scholar] [CrossRef]
  24. Gong, C.; Chen, H.; He, W.; Zhang, Z. Improved multi-objective clustering algorithm using particle swarm optimization. PLoS ONE 2017, 12, e0188815. [Google Scholar] [CrossRef]
  25. Wang, R.; Lai, S.; Wu, G.; Xing, L.; Wang, L.; Ishibuchi, H. Multi-clustering via evolutionary multi-objective optimization. Inf. Sci. 2018, 450, 128–140. [Google Scholar] [CrossRef]
  26. Tanabe, R.; Ishibuchi, H. An easy-to-use real-world multi-objective optimization problem suite. Appl. Soft Comput. 2020, 89, 106078. [Google Scholar] [CrossRef]
  27. Verma, S.; Pant, M.; Snasel, V. A comprehensive review on NSGA-II for multi-objective combinatorial optimization problems. IEEE Access 2021, 9, 57757–57791. [Google Scholar] [CrossRef]
  28. Qian, S.Y.; Jia, Z.H.; Li, K. A multi-objective evolutionary algorithm based on adaptive clustering for energy-aware batch scheduling problem. Future Gener. Comput. Syst. 2020, 113, 441–453. [Google Scholar] [CrossRef]
  29. Zhou, Y.; Kang, J.; Kwong, S.; Wang, X.; Zhang, Q. An evolutionary multi-objective optimization framework of discretization-based feature selection for classification. Swarm Evol. Comput. 2021, 60, 100770. [Google Scholar] [CrossRef]
  30. Hruschka, E.R.; Campello, R.J.; Freitas, A.A. A survey of evolutionary algorithms for clustering. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2009, 39, 133–155. [Google Scholar] [CrossRef]
  31. Bong, C.; Rajeswari, M. Multiobjective clustering with metaheuristic: Current trends and methods in image segmentation. IET Image Process. 2012, 6, 1–10. [Google Scholar] [CrossRef]
  32. Mukhopadhyay, A.; Maulik, U.; Bandyopadhyay, S.; Coello, C.A.C. Survey of multiobjective evolutionary algorithms for data mining: Part II. IEEE Trans. Evol. Comput. 2013, 18, 20–35. [Google Scholar] [CrossRef]
  33. Wang, L.; Cui, G.; Zhou, Q.; Li, K. A multi-clustering method based on evolutionary multiobjective optimization with grid decomposition. Swarm Evol. Comput. 2020, 55, 100691. [Google Scholar] [CrossRef]
  34. He, R.; Liu, S.; He, S.; Tang, K. Multi-domain active learning: Literature review and comparative study. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 7, 791–804. [Google Scholar] [CrossRef]
  35. Orouskhani, M.; Shi, D.; Orouskhani, Y. Multi-objective evolutionary clustering with complex networks. Expert Syst. Appl. 2021, 165, 113916. [Google Scholar] [CrossRef]
  36. Li, H.; Wan, F.; Gong, M.; Qin, A.; Wu, Y.; Xing, L. Privacy-enhanced multitasking particle swarm optimization based on homomorphic encryption. IEEE Trans. Evol. Comput. 2023; early access. [Google Scholar]
  37. Li, H.; Li, D.; Gong, M.; Li, J.; Qin, A.; Xing, L.; Xie, F. Sparse Hyperspectral Unmixing With Preference-Based Evolutionary Multiobjective Multitasking Optimization. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 1922–1937. [Google Scholar] [CrossRef]
  38. Li, H.; Luo, T.; Liu, L.; Gong, M.; Qiao, W.; Xie, F.; Qin, A. Selective Transfer Based Evolutionary Multitasking Optimization for Change Detection. IEEE Trans. Emerg. Top. Comput. Intell. 2024; early access. [Google Scholar]
  39. Li, H.; Ong, Y.S.; Gong, M.; Wang, Z. Evolutionary multitasking sparse reconstruction: Framework and case study. IEEE Trans. Evol. Comput. 2018, 23, 733–747. [Google Scholar] [CrossRef]
  40. Gupta, A.; Ong, Y.S.; Feng, L. Multifactorial evolution: Toward evolutionary multitasking. IEEE Trans. Evol. Comput. 2015, 20, 343–357. [Google Scholar] [CrossRef]
  41. Ma, J.; Zhao, Z.; Yi, X.; Chen, J.; Hong, L.; Chi, E.H. Modeling task relationships in multi-task learning with multi-gate mixture-of-experts. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1930–1939. [Google Scholar]
  42. Li, H.; Xiong, P.; Gong, M.; Qin, A.; Wu, Y.; Xing, L. Fast Heterogeneous Multi-Problem Surrogates for Transfer Evolutionary Multiobjective Optimization. IEEE Trans. Evol. Comput. 2024; early access. [Google Scholar]
  43. Tian, Y.; Si, L.; Zhang, X.; Tan, K.C.; Jin, Y. Local model-based Pareto front estimation for multiobjective optimization. IEEE Trans. Syst. Man Cybern. Syst. 2022, 53, 623–634. [Google Scholar] [CrossRef]
  44. Ouelmokhtar, H.; Benmoussa, Y.; Diguet, J.P.; Benazzouz, D.; Lemarchand, L. Near-optimal covering solution for USV coastal monitoring using PAES. J. Intell. Robot. Syst. 2022, 106, 24. [Google Scholar] [CrossRef]
  45. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  46. Coello, C.A.C.; Pulido, G.T.; Lechuga, M.S. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  47. Zhang, Q.; Li, H. MOEA/D: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  48. Zhong, G.; Pun, C.M. Local learning-based multi-task clustering. Knowl.-Based Syst. 2022, 255, 109798. [Google Scholar] [CrossRef]
  49. Zhang, X.; Zhang, X.; Liu, H.; Luo, J. Multi-task clustering with model relation learning. In Proceedings of the IJCAI, Stockholm, Sweden, 13–19 July 2018; pp. 3132–3140. [Google Scholar]
  50. Vandenhende, S.; Georgoulis, S.; Van Gansbeke, W.; Proesmans, M.; Dai, D.; Van Gool, L. Multi-task learning for dense prediction tasks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3614–3633. [Google Scholar] [CrossRef] [PubMed]
  51. Nie, F.; Zeng, Z.; Tsang, I.W.; Xu, D.; Zhang, C. Spectral embedded clustering: A framework for in-sample and out-of-sample spectral clustering. IEEE Trans. Neural Netw. 2011, 22, 1796–1808. [Google Scholar]
  52. Tsai, C.W.; Chen, W.L.; Chiang, M.C. A modified multiobjective EA-based clustering algorithm with automatic determination of the number of clusters. In Proceedings of the 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Seoul, Republic of Korea, 14–17 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2833–2838. [Google Scholar]
  53. While, L.; Hingston, P.; Barone, L.; Huband, S. A faster algorithm for calculating hypervolume. IEEE Trans. Evol. Comput. 2006, 10, 29–38. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of trace-based adjacency.
Figure 1. Schematic representation of trace-based adjacency.
Electronics 13 01987 g001
Figure 2. Schematic representation of multi-task adjacency.
Figure 2. Schematic representation of multi-task adjacency.
Electronics 13 01987 g002
Figure 3. Schematic diagram of the genotype update.
Figure 3. Schematic diagram of the genotype update.
Electronics 13 01987 g003
Figure 4. Handcrafted dataset sample instance distribution diagram. (a) Square, (b) Long, (c) Smile, (d) Triangle, (e) Sizes.
Figure 4. Handcrafted dataset sample instance distribution diagram. (a) Square, (b) Long, (c) Smile, (d) Triangle, (e) Sizes.
Electronics 13 01987 g004
Figure 5. Plot of the variation of parameters p o p and g e n in (a) Sizes and (b) 50d-20c.
Figure 5. Plot of the variation of parameters p o p and g e n in (a) Sizes and (b) 50d-20c.
Electronics 13 01987 g005
Figure 6. Comparison of hypervolume value curves between MOCK and MFMOCK on three datasets (a) dataset1, (b) dataset2, (c) dataset3.
Figure 6. Comparison of hypervolume value curves between MOCK and MFMOCK on three datasets (a) dataset1, (b) dataset2, (c) dataset3.
Electronics 13 01987 g006
Table 1. Abbreviation declaration table.
Table 1. Abbreviation declaration table.
AbbreviationsMeaning
MOPmulti-objective optimization problem
MOEAmulti-objective evolutionary algorithm
LAGlocus-based adjacency graph
TAGtrajectory adjacency graph
PFPareto front
CFcontrol front
SNMFsymmetric non-negative matrix factorization
Devoverall deviation
Conconnectivity index
SBXsimulated binary crossover
ACCclustering accuracy
ARIadjusted Rand index
Avgaverage
CVcoefficient of variation
AGaverage gap
Table 2. Nomenclature declaration table.
Table 2. Nomenclature declaration table.
NomenclatureMeaning
mamount of tasks
Xdataset
namount of data in dataset
Msimilarity matrix between tasks
Cclusters vectors
Ddimension of task
hamount of cluster
Ycluster label matrix
rcluster label
lactual class label
Wregularization term
Gcluster correlation coefficient matrix
Ttask list
fobjective function
pindividual in evolution algorithm
Ppopulation in evolution algorithm
Ψ individual evaluation value
φ individual Fitness
τ skill factor
rclustering result
Ogene in evolution algorithm
Ndataset size
p o p population size
kamount of cluster
g e n population evolution generation
r m p random mating probability
Table 3. Manual dataset characteristics.
Table 3. Manual dataset characteristics.
DatasetSamplesFeature DimensionClasses
Square100024
Smile100024
Long100022
Sizes100024
Triangle100024
Table 4. Characteristics of a randomly generated dataset.
Table 4. Characteristics of a randomly generated dataset.
DatasetGeneratorSamplesFeature DimensionsClasses
10d-4c1966104
50d-4c2351504
100d-4c212861004
2d-10c13408210
10d-10c121611010
50d-10c229195010
100d-10c2210310010
2d-20c11231220
10d-20c112791020
50d-20c212545020
100d-20c2122010020
2d-40c12563240
10d-40c119371040
50d-40c221695040
100d-40c2196410040
Table 5. UCI dataset features.
Table 5. UCI dataset features.
DatasetSamplesFeature DimensionsClasses
Iris15043
Thyroid21553
Aggregate78827
Cancer68392
Jain37322
Wine178133
Append10672
Table 6. Clustering results for the manually generated dataset.
Table 6. Clustering results for the manually generated dataset.
ACC
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Square98.0025.3097.5099.0098.9099.1099.2099.00
Long90.6093.3068.4093.2093.4089.5091.7090.80
Smile98.6099.1099.1098.7098.50100.00100.00100.00
Sizes96.0076.8090.1063.5068.0096.3084.6098.40
Triangle86.8097.4097.3097.5083.3099.1099.2099.60
NMI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Square95.6030.3094.2495.6095.2396.1096.3095.87
Smile79.8674.6354.2178.8479.3572.1276.0573.52
Long98.4099.5099.5099.1099.40100.00100.00100.00
Sizes87.3969.3368.1937.9146.9688.0280.8188.56
Triangle95.9899.0698.4293.2892.7993.1995.3399.58
ARI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Square0.960.250.950.970.970.980.980.97
Smile0.840.820.520.830.840.740.790.77
Long0.970.980.980.980.991.001.001.00
Sizes0.910.310.850.240.390.910.820.95
Triangle0.880.970.970.960.810.990.991.00
Table 7. Clustering results for randomly generated datasets.
Table 7. Clustering results for randomly generated datasets.
ACC
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
10d-4c98.7654.1465.7399.1796.5899.7999.6999.90
50d-4c65.2458.4052.4278.3565.8194.8799.2999.72
100d-4c76.9167.9643.8677.9958.7983.8386.0091.52
2d-10c70.2828.3575.3577.5881.0455.1975.7692.08
10d-10c75.0123.6971.0374.3648.2282.9795.0592.32
50d-10c61.5317.3738.9279.1035.9091.7192.2699.55
100d-10c64.1517.7842.6180.8429.5885.9782.4589.28
2d-20c91.6382.9491.3177.1779.0485.6284.3297.32
10d-20c80.9275.5394.7669.2722.9996.7299.9899.92
50d-20c63.7216.6731.8260.5344.9858.5361.0882.38
100d-20c54.5117.3036.6454.4336.3170.1670.3378.03
2d-40c67.6978.9746.2082.4056.4682.2576.7186.73
10d-40c84.8288.0794.9049.2011.2096.0187.6187.35
50d-40c61.7820.0122.2233.9326.3755.1943.3477.59
100d-40c66.8012.8825.1535.7941.1451.9359.4176.78
NMI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
10d-4c95.0636.4163.2696.5087.8699.0299.4598.48
50d-4c57.8112.8742.4761.4857.5895.2791.5898.44
100d-4c65.6864.2038.6464.2456.2485.0791.2897.94
2d-10c77.8518.9875.5781.2482.0259.9474.4790.79
10d-10c82.2325.5771.4679.2839.0983.5594.2789.56
50d-10c61.2223.4235.8177.7232.7592.3591.2499.29
100d-10c66.0923.9242.6678.9022.4080.2278.4091.87
2d-20c94.2389.1894.2381.0389.0090.5790.2495.23
10d-20c92.8284.0295.6776.3927.4999.85100.0097.56
50d-20c66.5815.3234.4363.7053.7764.0067.8982.84
100d-20c69.2512.8731.9646.8145.1774.3573.5379.06
2d-40c82.8792.1266.0192.5172.9192.1290.8393.22
10d-40c94.7193.0498.2963.8217.8196.5492.6990.75
50d-40c72.0621.8233.6948.3335.3165.8444.2582.19
100d-40c70.6614.5533.7749.6556.8961.4462.1484.91
ARI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
10d-4c0.960.290.520.970.910.991.000.99
50d-4c0.420.090.230.490.420.880.850.89
100d-4c0.470.540.200.530.370.830.840.88
2d-10c0.640.280.710.700.740.510.670.88
10d-10c0.740.240.710.690.390.820.950.91
50d-10c0.340.270.090.620.260.940.941.00
100d-10c0.400.110.130.630.160.900.860.99
2d-20c0.730.340.560.700.670.810.440.81
10d-20c0.820.700.930.580.131.001.000.97
50d-20c0.280.020.060.430.320.330.360.66
100d-20c0.220.010.030.110.170.080.040.24
2d-40c0.660.800.380.820.480.820.760.85
10d-40c0.870.820.980.460.060.950.890.90
50d-40c0.270.020.040.240.090.300.080.54
100d-40c0.260.010.030.240.310.230.280.68
Table 8. Clustering results on UCI datasets.
Table 8. Clustering results on UCI datasets.
ACC
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Iris89.3377.3391.3388.6789.3386.6688.6089.30
Thyroid73.4075.8176.7457.1679.6379.5079.6082.30
Aggregate77.5586.0494.5476.4175.8598.6998.0699.71
Cancer44.9360.7662.3721.9230.6957.5758.4381.99
Jain77.8880.1693.8375.2377.4899.4080.4894.73
Wine63.9360.6754.4961.7468.5453.5456.4071.35
Append80.6682.0878.3075.4779.2565.9484.8187.74
NMI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Iris75.1565.6080.5473.6474.5072.6970.7474.60
Thyroid31.8221.8516.9019.1632.8132.1534.1035.60
Aggregate79.6189.9492.3498.4577.1880.0872.5899.07
Cancer32.6548.1249.6917.2826.4245.2644.3167.72
Jain32.9637.6360.2396.9232.4997.5678.0694.63
Wine41.2433.0735.6141.9541.6319.6533.5343.54
Append26.2628.2422.3126.2726.2122.6231.8832.07
ARI
DatasetKMSLALSCFCMMOCKMIE-MOCKMFMOCK
Iris0.730.580.770.720.730.810.720.83
Thyroid0.340.370.230.160.450.420.430.48
Aggregate0.710.910.930.990.690.810.600.97
Cancer0.200.290.320.140.180.300.310.46
Jain0.310.360.750.990.300.990.750.92
Wine0.360.270.320.360.350.220.300.41
Append0.320.280.300.290.290.220.330.32
Table 9. Comparison of algorithm test datasets.
Table 9. Comparison of algorithm test datasets.
DatasetTaskDataset NameSamplesFeatureClasses
Dataset1Task1Sizes100024
Task2Square100024
Dataset2Task1100d-40c196410040
Task2100d-10c122010010
Dataset3Task150d-20c-112455020
Task250d-20c-223345020
Table 10. Average performance of MFMOCK on dataset 1 problems.
Table 10. Average performance of MFMOCK on dataset 1 problems.
Dataset 1Task 1Task 2
Overall Deviation DevConnectivity Index ConOverall Deviation DevConnectivity Index Con
Avg3844.851.9483188.897.291
CV1.279%1.602%1.267%1.09%
AG1.866%1.306%2.879%1.967%
Table 11. Average performance of MFMOCK on dataset 2 problems.
Table 11. Average performance of MFMOCK on dataset 2 problems.
Dataset 2Task 1Task 2
Overall Deviation DevConnectivity Index ConOverall Deviation DevConnectivity Index Con
Avg1562.444.02904.838.95
CV3.682%7.019%2.208%0.91%
AG5.83%16.37%1.04%2.09%
Table 12. Average performance of MFMOCK on dataset 3 problems.
Table 12. Average performance of MFMOCK on dataset 3 problems.
Dataset 3Task 1Task 2
Overall Deviation DevConnectivity Index ConOverall Deviation DevConnectivity Index Con
Avg875.345.14825.756.07
CV5.981%3.267%2.014%4.459%
AG15.12%11.2%17.43%9.71%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Dang, K.; Yang, R.; Li, L.; Li, H.; Gong, M. Multi-Objective Automatic Clustering Algorithm Based on Evolutionary Multi-Tasking Optimization. Electronics 2024, 13, 1987. https://doi.org/10.3390/electronics13101987

AMA Style

Wang Y, Dang K, Yang R, Li L, Li H, Gong M. Multi-Objective Automatic Clustering Algorithm Based on Evolutionary Multi-Tasking Optimization. Electronics. 2024; 13(10):1987. https://doi.org/10.3390/electronics13101987

Chicago/Turabian Style

Wang, Ying, Kelin Dang, Rennong Yang, Leyan Li, Hao Li, and Maoguo Gong. 2024. "Multi-Objective Automatic Clustering Algorithm Based on Evolutionary Multi-Tasking Optimization" Electronics 13, no. 10: 1987. https://doi.org/10.3390/electronics13101987

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop