Next Article in Journal
Correction: Mergelyan, S.N.; Matevossian, H.A. Editorial: S. N. Mergelyan’s Dissertation “Best Approximations in the Complex Domain” (Short Version). Mathematics 2024, 12, 939
Previous Article in Journal
Complex t-Intuitionistic Fuzzy Graph with Applications of Rubber Industrial Water Wastes
Previous Article in Special Issue
A Novel Fuzzy Bi-Clustering Algorithm with Axiomatic Fuzzy Set for Identification of Co-Regulated Genes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Similarity-Based Three-Way Clustering by Using Dimensionality Reduction

by
Anlong Li
1,
Yiping Meng
2,* and
Pingxin Wang
2
1
School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China
2
School of Science, Jiangsu University of Science and Technology, Zhenjiang 212100, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(13), 1951; https://doi.org/10.3390/math12131951
Submission received: 15 May 2024 / Revised: 16 June 2024 / Accepted: 18 June 2024 / Published: 24 June 2024
(This article belongs to the Special Issue New Advances in Data Analytics and Mining)

Abstract

:
Three-way clustering uses core region and fringe region to describe a cluster, which divide the dataset into three parts. The division helps identify the central core and outer sparse regions of a cluster. One of the main challenges in three-way clustering is the meaningful construction of the two sets. Aimed at handling high-dimensional data and improving the stability of clustering, this paper proposes a novel three-way clustering method. The proposed method uses dimensionality reduction techniques to reduce data dimensions and eliminate noise. Based on the reduced dataset, random sampling and feature extraction are performed multiple times to introduce randomness and diversity, enhancing the algorithm’s robustness. Ensemble strategies are applied on these subsets, and the k-means algorithm is utilized to obtain multiple clustering results. Based on these results, we obtain co-association frequency between different samples and fused clustering result using the single-linkage method of hierarchical clustering. In order to describe the core region and fringe region of each cluster, the similar class of each sample is defined by co-association frequency. The lower and upper approximations of each cluster are obtained based on similar class. The samples in the lower approximation of each cluster belong to the core region of the cluster. The differences between lower and upper approximations of each cluster are defined as fringe region. Therefore, a three-way explanation of each cluster is naturally formed. By employing various UC Irvine Machine Learning Repository (UCI) datasets and comparing different clustering metrics such as Normalized Mutual Information (NMI), Adjusted Rand Index (ARI), and Accuracy (ACC), the experimental results show that the proposed strategy is effective in improving the structure of clustering results.

1. Introduction

As an unsupervised technique in data mining and machine learning, cluster analysis is widely used in various areas such as attribute reduction [1,2,3,4], feature selection [5,6,7], image processing [8,9], information granulation [10,11,12], and graph convolutional neural networks [13,14,15]. The primary objective of clustering is to organize heterogeneous data into meaningful groups based on their similarities, revealing the inherent structures and patterns within the dataset. To achieve this, various clustering algorithms [16] have been developed. However, it has been accepted that a single clustering algorithm cannot handle all types of data distribution effectively. Different algorithms or different parameters for an algorithm may lead to different clustering results. To enhance the robustness and stability of clustering algorithms, researchers have proposed ensemble clustering methods. In comparison to single clustering methods, ensemble clustering methods [17,18,19,20,21,22] integrate results from multiple foundational clustering algorithms, yielding more stable, robust, and accurate clustering solutions. Nevertheless, existing ensemble clustering methods typically adopt a hard clustering strategy, where an element can belong to only one cluster or none, and clear boundaries exist between different clusters. However, in situations with insufficient information on data samples, hard clustering algorithms often lead to higher decision risks.
To address this issue, three-way decision theory [23,24] was introduced to describe uncertainties in information. This method divides the sample universe into three mutually exclusive regions and adopts different decision strategies for each region [25,26]. The three-way decision framework can be integrated with various computational models for learning uncertainty, such as rough set theory [27,28,29], Bayesian networks [30,31], and fuzzy particle swarm optimization [32,33]. Inspired by the idea of three-way decision, Yu [34] presented the framework of three-way clustering by using core and the fringe regions to character a cluster. These two sets partition the sample space into three parts, which capture three kinds of relationships between objects and a cluster, namely, belonging to, partially belonging to, and not belonging to [35,36,37,38].
Recently, three-way clustering [39] has garnered widespread research interest, leading to the development of various three-way clustering algorithms within this theoretical framework. Wang and Yao [40] proposed a three-way clustering framework called CE3, derived from mathematical morphology’s erosion and dilation concepts. Li et al. [41] introduced sample’s stability to identify and establish relationships in ensemble clustering. Yu et al. [42] proposed an efficient three-way clustering algorithm based on the idea of universal gravitation. Jia et al. [43] developed an automatic three-way clustering approach by combining the proposed threshold selection and the cluster number selection method. Wang et al. [44] proposed a three-way adaptive density peak clustering (3W-ADPC) method by integrating natural nearest neighbors with DPC.
Most of the existing three-way clustering algorithms are based on the original dataset, which is not suitable for high-dimensional datasets. The processing of high-dimensional data poses a fundamental yet highly challenging problem in the current field of data science. The purpose of dimensionality reduction is to decrease the data’s dimensionality while retaining the most significant aspects of its characteristics. By reducing the data’s dimensionality, we can simplify the complexity of data analysis, enhance model training speed, reduce storage requirements, and facilitate a clearer understanding and interpretation of the model’s results. Various dimensionality reduction techniques are commonly employed to address this challenge, including Principal Component Analysis (PCA) [45,46,47], spectral clustering [48,49], factor analysis [50], and multidimensional scaling [51].
By integrating dimensionality reduction into three-way clustering, this paper presents an ensemble three-way clustering algorithm based on dimensionality reduction. The proposed method uses dimensionality reduction techniques to reduce data dimensions and eliminate noise. Based on the reduced dataset, random sampling and feature extraction are performed multiple times to introduce randomness and diversity, enhancing the algorithm’s robustness. Ensemble strategies are applied on these subsets, and the k-means algorithm is utilized to obtain multiple clustering results. Based on these results, the frequency of different data points being assigned to the same cluster is calculated to derive the co-occurrence frequency. If the co-occurrence frequency between data points exceeds a certain threshold, they are defined as similar classes. Finally, a three-way clustering approach was introduced by using the proposed similar relations. The main contributions of this research are as follows:
(1)
Ensemble three-way clustering framework based on dimensionality reduction.
We introduce a novel ensemble three-way clustering framework that combines dimensionality reduction techniques with clustering ensemble methods. This framework reduces data dimensions, eliminates noise, and enhances clustering stability. By leveraging multiple clustering results, the method enhances the algorithm’s robustness through randomness and diversity.
(2)
Integration of co-occurrence frequency, hierarchical clustering, and lifecycle analysis:
The proposed method calculates the co-occurrence frequency of data points being in the same cluster, aiding in accurately defining similar classes. It employs a single-linkage hierarchical clustering approach to fuse clustering results and constructs a dendrogram based on these probabilities. By analyzing the lifecycle of clusters, we determine the most stable clustering result, ensuring robustness and consistency.
These contributions collectively enhance the performance and applicability of three-way clustering algorithms, especially for high-dimensional datasets, providing a more accurate and stable clustering solution.
The remainder of this paper is organized as follows. In Section 2, we provide a comprehensive review of the concepts related to three-way clustering, the k-means algorithm, PCA, and data integration strategies. Section 3 outlines the methodology and algorithmic process employed in this study. The results and performance metrics obtained from the proposed algorithm on the UCI dataset are presented in Section 4. Section 5 encompasses the discussion of our findings and identifies areas for future improvement.

2. Related Work

2.1. Three-Way Clustering

Traditional hard clustering depicts a cluster by one set with a sharp boundary. Only two relationships between the sample and cluster are considered, i.e., belonging to and not belonging to. For the samples inside the cluster, they belong to this cluster, and for the samples outsider the cluster, they are not the elements of this cluster. Given a dataset X = { x 1 , x 2 , , x n } with n samples and k clusters in traditional hard clustering, the clustering results can be represented as C = { C 1 , C 2 , , C k } , where C i , i = 1 , 2 , , k satisfies
C i Ø , i = 1 , , k i = 1 k C i = U C i C j = Ø , i j .
In traditional clustering, each sample is unequivocally assigned to one cluster, and there are clear boundaries between different clusters. This two-way description of a cluster may not adequately show the uncertainty information in data. To address the limitation in traditional clustering, Yu [34,52] proposed three-way clustering by defining three types of membership relations between a sample and a cluster, namely, belonging to fully, belonging to partially and not belonging to. Three-way clustering utilizes the core region C o C i and the fringe region F r C i to depict a cluster, and the universe is split by these two sets into three sections, C o C i , F r C i , and T r ( C i ) = U C o ( C i ) F r ( C i ) , which obey the following conditions:
C o C i T r C i F r C i = U C o C i T r C i = Ø C o C i F r C i = Ø F r C i T r C i = Ø .
Three-way clustering results of dataset X are expressed as
C = { ( C o ( C 1 ) , F r ( C 1 ) ) , ( C o ( C 2 ) , F r ( C 2 ) ) , , ( C o ( C k ) , F r ( C k ) ) } .

2.2. PCA Dimensionality Reduction

As a powerful tool in the realm of data analysis, PCA [47] (Principal Component Analysis) offers a systematic approach to reduce the dimensionality of data while retaining the significant variance within the dataset. This not only makes data easier to visualize but also enhances the efficiency of subsequent analytical techniques. The fundamental idea of PCA involves a linear transformation that maps the original data onto a new coordinate system. The selection of this new coordinate system aims to maximize the variance of the data along specific axes. By choosing the first few principal components, the data can be projected onto these components, achieving dimensionality reduction.
In the computational process, the initial step involves calculating the covariance matrix of the original data. Subsequently, through eigenvalue decomposition, the eigenvalues and eigenvectors of the covariance matrix are obtained. Following this, a selection of the top eigenvalue-ordered eigenvectors forms the new coordinate system, representing the principal components. Finally, projecting the original data onto these principal components yields the reduced-dimensional data. Figure 1 illustrates the fundamental principle of PCA for dimensionality reduction. In Figure 1, the original distribution of the dataset is given on the plane, where the red and black dots represents different classes. Through PCA, these points are projected onto the principal component directions in the reduced-dimensional space, resulting in a new distribution of data. This process allows for the mapping of high-dimensional data into a lower-dimensional space while retaining the essential features of the original data, thus reducing dimensionality.
The application of PCA for dimensionality reduction offers the advantage of preserving the crucial features of the data while reducing their dimensionality. This enhances computational efficiency for subsequent analyses, providing robust support for research endeavors.

2.3. K-Means Algorithm

K-means algorithm [53] is a widely used clustering method with the goal of partitioning a dataset into k clusters, such that samples in the same cluster have high similarity, and samples in distinct clusters have low similarity. The main idea of k-means algorithm involves determining the positions of cluster centers by minimizing a loss function, which incorporates the Euclidean distance between sample and cluster centers. Specifically, the algorithm initiates by randomly selecting k sample points as initial cluster centers. It iteratively performs two key steps, i.e., assigning each sample point to the closest cluster center in Euclidean distance, and updating the position of each cluster’s center based on the samples assigned to it. This process repeats until the cluster centers no longer undergo significant changes, signifying convergence of the loss function. The mathematical formulation of the loss function is given by
J = i = 1 n j = 1 k w i j χ i μ j 2 ,
where w i j is the indicator function, indicating whether the sample χ i is assigned to cluster μ j . By minimizing this loss function, k-means algorithm efficiently identifies optimal cluster center positions, facilitating effective data clustering.

2.4. Hierarchical Clustering

Hierarchical clustering builds a tree-like structure (dendrogram) to represent the nested grouping of data points. It can be divided into agglomerative and divisive methods [54]. Agglomerative hierarchical clustering starts with each data point as an individual cluster and iteratively merges the closest pairs of clusters until a single cluster is formed. Conversely, divisive hierarchical clustering starts with the whole dataset as a single cluster and recursively splits it into smaller clusters. A well-known variation is the single-linkage method, which defines the distance between two clusters as the minimum distance between any pair of points from the two clusters. This method is effective in identifying clusters with irregular shapes.

2.5. Clustering Ensemble and Co-Association Frequency

Although there are many clustering methods, it has been accepted that there is not one clustering method that can identify all kinds of data structure distribution. In order to solve this problem, Strehl and Ghosh [16] proposed the cluster ensemble algorithm, which combines multiple clustering results of a set of objects into one clustering result without accessing the original features of the objects. The framework of clustering ensemble can be depicted by Figure 2.
The aim of clustering ensemble [55] is to consolidate multiple independent clustering results into a comprehensive outcome, aiming to overcome potential biases introduced by different clustering algorithms. Moreover, the rise of clustering ensemble has given birth to various clustering ensemble methods, such as the voting–merging approach proposed by Hornik [56]. This method leverages clustering ensemble algorithms to achieve more reliable and stable clustering results. It utilizes an unsupervised voting mechanism to amalgamate within the ensemble clustering, ultimately merging to derive the final clustering outcome. For family clustering results of a dataset, there are three types of relationships between two samples by qualitative observation. They may be always assigned to the same cluster, or they are assigned to the same cluster occasionally. The last circumstance is not assigned to the same group completely. In order to quantify a sample’s tendency of changing groups quantitatively, Li et al. [41] introduced a measurement named as co-association frequency by using the results of a family clustering.
Given a dataset X = { x 1 , x 2 , , x n } with n samples and C 1 , C 2 , , C L are family clustering results on U, we use C l ( x i ) to indicate the label of x i induced by clustering result C l . The co-association frequency p i j , which represents that two samples x i and x j appear in the same cluster, is calculated by
p i j = 1 L l = 1 L C l x i , C l x j ,
where
C l x i , C l x j = 1 C l ( x i ) = C l ( x j ) 0 C l ( x i ) C l ( x j ) .
We use an example to illustrate p i j . Figure 3 is a dataset X with 6 samples and four clustering results C 1 , C 2 , C 3 , C 4 of X . The samples x 1 and x 2 consistently remain in the same cluster across all results, indicating that co-association frequency p 12 = 1 . On the other hand, x 1 and x 3 are assigned to the same cluster only in C 1 and C 2 , showing that co-association frequency p 13 = 0.5 . The samples x 1 and x 5 are grouped into different clusters across all four clustering results, indicating that co-association frequency p 15 = 0 .
According to the above definition, we can obtain the co-association matrix of Figure 3 as Table 1.
Co-association frequency [57,58] is to measure the probability that two data samples are assigned to the same cluster in multiple clustering results. Specifically, if two samples are consistently assigned to the same cluster across multiple clustering results, their co-association frequency is 1. If two samples are not assigned to the same group completely, their co-association frequency is 0. By calculating the co-occurrence probability for all data points, a co-association frequency matrix is obtained. This matrix provides information about the similarity of data points. By setting a threshold for co-association frequency, samples with frequencies above the threshold are grouped into the same similarity class. This approach integrates information from multiple clustering runs, not relying solely on a single clustering result, thereby enhancing a comprehensive understanding of the data structure.

3. Similarity-Based Three-Way Clustering by Using Dimensionality Reduction

In this section, we propose a similarity theory [43,59] based on data dimensionality reduction and similarity-based three-way clustering. In contrast to traditional algorithms, our approach first employs the PCA algorithm for data preprocessing, transforming high-dimensional data into low-dimensional data. It incorporates an ensemble strategy by randomly extracting subsets of features from the samples in multiple iterations, generating diverse basic clustering results using the traditional k-means clustering algorithm. Subsequently, we calculate the co-association frequency between samples to derive similarity classes. By extracting only partial features of the samples, we significantly reduce the computational complexity compared to the existing traditional ensemble clustering methods. The algorithm proposed in this paper involves three main steps: the generation of basic clusters by using dimensionality-reduced data, the computation of co-association frequency and similarity classes, and the integration of these results into three-way clustering.

3.1. Dimensionality Reduction

In this study, we employed data dimensionality reduction techniques, specifically utilizing Principal Component Analysis (PCA) to reduce the dimensions of the data. PCA is a commonly used dimensionality reduction method, aiming to map the original data onto a lower-dimensional subspace while retaining the maximum variance in the data. Through PCA, we can transform high-dimensional data into lower-dimensional space, thereby enhancing our understanding of the intrinsic structure of the data.
To begin with, consider a dataset comprising n samples and D features, represented by matrix X , where each row corresponds to a sample, and each column represents a feature. Our objective is to project this D -dimensional dataset onto a K -dimensional subspace (where K < D ) and obtain a new feature matrix Z . The specific steps of dimensionality reduction by using PCA are as follows:
Step 1: Data normalization: The first step involves centralizing the original data by subtracting the mean of each feature, resulting in the centered matrix X .
Step 2: Covariance Matrix Computation: The covariance matrix represents the correlations between data features, with the specific formula
Ω = 1 N X T X .
Step 3: Eigenvalue and Eigenvector Computation: Eigenvalue decomposition is applied to the covariance matrix Ω , yielding eigenvalues λ 1 , λ 2 , , λ D and their corresponding eigenvectors v 1 , v 2 , , v D .
Ω v = λ i v i , i = 1 , 2 , , D .
Step 4: Selection of Top K Eigenvectors: The eigenvectors corresponding to the top K largest eigenvalues are chosen, forming the projection matrix V .
Step 5: Data Projection: The centered original data matrix X is projected onto the selected K -dimensional subspace, resulting in the reduced feature matrix Z , where each row represents a sample, and each column represents a reduced feature. The specific formula is
Z = X V .
Through the aforementioned steps, we obtain the reduced-dimensional data matrix. In this low-dimensional space, we conduct fundamental clustering operations. This data-driven foundational clustering method allows for clustering analysis in lower dimensions while preserving the primary features of the data. The key advantage of this approach lies in its ability to facilitate data visualization, reduce computational complexity, and enhance clustering effectiveness through dimensionality reduction.
Next, we randomly select parts of the sample’s features to obtain different clustering results. For a multidimensional dataset, different subsets of features try to describe the dataset from different views. Thus, a set of diverse clustering results will be obtained when distinguishing subsets of features are employed. Suppose that we randomly extract parts of the features and apply the k-means clustering method to divide the dataset into k clusters. This process is repeated L times, yielding multiple clustering results C 1 , C 2 , , C L . The process of foundational clustering based on data dimensionality reduction is outlined in Algorithm 1.
Algorithm 1: Foundational Clustering Based on Data Dimensionality Reduction
Mathematics 12 01951 i001

3.2. Clustering Ensemble

From multiple clustering iterations, we obtain basic clustering results C 1 , C 2 , , C L . Subsequently, we present a method for integrating the basic clustering results by using the co-occurrence frequency matrix. The aim is to employ the single-link method of hierarchical clustering to generate a more robust clustering result.
For a dataset X = { x 1 , x 2 , , x n } with n samples and C 1 , C 2 , , C L are family clustering results of X , we can construct an n × n co-association frequency matrix P , whose elements p i j represents the frequency that two samples x i and x j are simultaneously assigned to the same cluster.
We view p i j as the similarity between samples and utilize the single-linkage of hierarchical clustering to obtain an ensemble clustering result. In the process of clustering, each data sample is treated as an independent cluster, and then gradually the most similar cluster is merged based on their co-association frequencies. Clusters with the highest similarity are merged to form a new cluster node. This process iterates until the cluster result with the highest lifetime is chosen as the final merged result.
The schematic representation of the single-linkage clustering dendrogram is illustrated in Figure 4. Different colors in Figure 4 represent different clusters at present, and each color represents a set of samples with high similarity. This bottom-up merging strategy ensures that we fully consider the degree of association between samples, resulting in more accurate clustering results. By measuring the similarity between different clusters and visualizing them as a dendrogram, we could intuitively observe the structure and hierarchy of the clustering results. In the dendrogram, higher connecting points represented stronger associations between clusters with higher co-occurrence frequencies. These results were relatively stable and less susceptible to noise or changes in the data. Therefore, such clustering results were more reliable and better able to reflect the true structure and patterns of the data.
By constructing a single-linkage clustering dendrogram using co-association frequencies and selecting the clustering result with the highest lifetime as the final fusion result, we obtain more stable clustering results, thereby enhancing our understanding of the features and inherent structure of the dataset. The process of ensemble clustering is outlined in Algorithm 2.
Algorithm 2: Ensemble Clustering Results
  Input: Reduced data matrix C 1 , C 2 , , C L
  Output: C 1 , C 2 , , C L
1  Compute the co-occurrence frequency matrix P by (2).
2  Obtain the single-linkage dendrogram of P .
3  Achieve ensemble clustering results C with the highest lifetime.
4  Return ensemble clustering results C = ( C 1 , C 2 , , C k ) .

3.3. Similar Classes Based on Co-Association Frequency

This section introduces three-way clustering models based on the co-occurrence frequency derived from clustering ensemble, proposing a similarity relationship under the framework of co-association frequency. Firstly, we give the definition of similar relation between x i and x j .
Definition 1.
For a dataset X = { x 1 , x 2 , , x n } with n samples and C 1 , C 2 , , C L are family clustering results of X , p i j is the co-association frequency between samples x i and x j . The similarity relation S i m θ ( x i , x j ) based on a threshold θ is defined as:
S i m θ ( x i , x j ) = { ( x i , x j ) X × X | p i j θ } ,
where 0 θ 1 is a pre-defined parameter. For x i X , the similar class is computed by:
S i m θ ( x i ) = { x j X | p i j θ } .
We still use Figure 3 as an example. If we take θ = 0.7 , then S i m θ ( x 1 ) = { x 1 , x 2 } , S i m θ ( x 2 ) = { x 1 , x 2 } , S i m θ ( x 3 ) = { x 3 , x 4 } , S i m θ ( x 4 ) = { x 3 , x 4 } , S i m θ ( x 5 ) = { x 5 , x 6 } , and S i m θ ( x 6 ) = { x 5 , x 6 } .
From the above definition, we can find that the similar class S i m θ ( x i ) has the following properties:
1   x i S i m θ ( x i )   ;
2   if   x j S i m θ ( x i ) ,   then   x i S i m θ ( x j ) ;
3   i = 1 n S i m θ ( x i ) = X .
Clearly, the set of similar classes { S i m θ ( x i ) | { x i X } forms a covering of dataset X . For any subset C X , 0 θ 1 , the lower and upper approximations based on the co-association frequency are defined as follows:
A p r θ ( C ) ¯ = { x i X | S i m θ ( x i ) C } ;
A p r θ ( C ) ¯ = { x i X | S i m θ ( x i ) C Ø } .
Furthermore, we can use the positive region P o s θ ( C ) and the fringe region B n d θ ( C ) to describe the objective subset C . So, we define P o s θ ( C ) and B n d θ ( C ) , of C as
P o s θ ( C ) = A p r θ ( C ) ¯
B n d θ ( C ) = A p r θ ( C ) ¯ A p r θ ( C ) ¯
Usually, the positive region P o s θ ( C ) contains the samples that belong to C definitely, and the fringe region B n d θ ( C ) contains the samples that belong to C possibly. Based on the definitions and properties of P o s θ ( C ) and B n d θ ( C ) , for any cluster C i X , it is straightforward to obtain the core region C o ( C i ) and the fringe region F r ( C i ) by
C o ( C i ) = P o s θ ( C i ) ,
F r ( C i ) = B n d θ ( C i ) .
Algorithm 3 illustrates the calculation of core region C o ( C i ) and the fringe region F r ( C i ) based on co-association frequency.
Algorithm 3: Finding core region and fringe region
Mathematics 12 01951 i002

3.4. Similarity-Based Three-Way Clustering by Using Dimensionality Reduction

The stepwise execution of Algorithms 1–3 forms the framework of the proposed similarity-based three-way clustering by using dimensionality reduction, as illustrated in Algorithm 4.
Algorithm 4: Similarity-based three-way clustering algorithm
Input: Original data matrix X = { x 1 , x 2 , , x n } , Number of iteratiber of clusters L
Dimensionality reduction method (PCA), Threshold θ
Output: C = { ( C o ( C 1 ) , F r ( C 1 ) ) , ( C o ( C 2 ) , F r ( C 2 ) ) , , ( C o ( C k ) , F r ( C k ) ) }
1
  Initialize: ←Algorithm 1; Return C 1 , C 2 , , C L ;
2
  Ensemble: ←Algorithm 2; Return  C 1 , C 2 , , C k ;
3
  Identify Core and Fringe Regions: ←Algorithm 3;
4
  Return  C = { ( C o ( C 1 ) , F r ( C 1 ) ) , ( C o ( C 2 ) , F r ( C 2 ) ) , , ( C o ( C k ) , F r ( C k ) ) } .
In this framework, we first generate a set of base clustering results by employing dimensionality reduction techniques (Algorithm 1). Subsequently, by calculating co-association frequencies, we utilize the single-linkage of hierarchical clustering to obtain ensemble clustering results (Algorithm 2). Finally, by defining the similar classes of each sample, we derive the core and fringe regions, further adjusting the clustering structure to yield more accurate and representative three-way clustering outcomes.
The uniqueness of this framework lies in its integration of data dimensionality reduction, co-association frequency computation, and definition of similar classes, providing a comprehensive revelation of the intrinsic structure during the clustering ensemble process. Algorithm 4 outlines the overall process of the three-way clustering framework, demonstrating how optimized clustering results are generated through multiple iterations to better reflect the characteristics of the original data.
The proposed approach offers a powerful tool for clustering ensemble, aiding in the precise capture of complex relationships and distribution patterns in clustering analysis. The three-way clustering framework provides valuable insights seeking to uncover intricate structures within their datasets.

4. Experimental Analyses

4.1. Data Descriptions

In this section, we conduct some experiments to evaluate the effectiveness of the proposed algorithm. We employ datasets from 13 UCI machine learning repositories [60], spanning diverse domains such as biology, medicine, and finance. The detailed information about these datasets is presented in Table 2, including the number of clusters and other relevant details. The software used for implementation includes MATLAB2019a for statistical and matrix computations and Python 3.9 with libraries such as NumPy, SciPy, and scikit-learn for data processing and machine learning tasks, ensuring robust and efficient analysis.

4.2. Evaluation Indices

(1)
Adjusted Rand Index (ARI) [61,62] serves as a prominent external metric for assessing clustering performance in comparison to ground truth labels. The ARI, an extension of the Rand Index (RI), is designed to overcome the limitations of the RI by adjusting for chance agreements.
ARI adjusts the RI using the following formula:
A R I = R I E [ R I ] max ( R I ) E [ R I ] ,
where E [ R I ] represents the expected Rand Index under random conditions. The Rand Index (RI) is calculated by the formula:
R I = a + b a + b + c + d .
a : the number of sample pairs that belong to the same cluster in both the ground truth and clustering results.
b : the number of sample pairs that belong to different clusters in both the ground truth and clustering results.
c : the number of sample pairs that belong to the same cluster in the ground truth but to different clusters in the results.
d : the number of sample pairs that belong to different clusters in the ground truth but to the same cluster in the results.
ARI values provide insights into the agreement between clustering results and ground truth labels, with 1 indicating perfect agreement, 0 suggesting performance no better than random assignment, and negative values indicating worse than random allocation. The introduction of ARI offers a comprehensive and objective means for evaluating clustering algorithms, facilitating a more accurate understanding of their performance.
(2)
Adjusted Mutual Information (AMI) [63,64] is an internal metric commonly used to assess the performance of clustering results. It is designed to measure the similarity between clustering results and a ground truth (typically, actual labels) by quantifying the information gain between two distributions.
The computation of AMI involves the following formula:
A M I ( U , V ) = M I ( U , V ) E [ M I ( U , V ) ] m a x ( H ( U ) , H ( V ) ) E [ M I ( U , V ) ] ,
where M I ( U , V ) represents the mutual information between U and V . E [ M I ( U , V ) ] is the expected mutual information under random conditions. H ( U ) and H ( V ) are the entropies of U and V , respectively.
The numerator of AMI is an adjusted value of mutual information, while the denominator is an adjusted value of entropy. The values of AMI range from [ 0 , 1 ] , where 1 indicates a perfect match, 0 denotes random matching, and negative values signify matching below random levels.
(3)
Accuracy (ACC) [65] is a common metric used to assess the performance of a classification model. It measures the proportion of samples that the model correctly classifies and serves as a simple and intuitive performance indicator. The formula for calculating ACC is as follows:
A C C = T P + T N T P + T N + F P + F N ,
where T P (True Positives) represents the number of samples correctly classified as the positive class, T N (True Negatives) represents the number of samples correctly classified as the negative class, F P (False Positives) represents the number of samples actually belonging to the negative class but misclassified as the positive class, F N (False Negatives) represents the number of samples actually belonging to the positive class but misclassified as the negative class.
The range of ACC is [ 0 , 1 ] , where 1 indicates perfect classification and 0 indicates classification failure. While ACC is an intuitive and easy-to-understand metric, it may have limitations when dealing with class imbalance.

4.3. Experimental Performances

Firstly, the PCA dimensionality reduction method is applied to high-dimensional datasets to obtain processed low-dimensional data. Subsequently, a clustering ensemble strategy is employed for the low-dimensional data. This involves randomly sampling subsets of data and features and running the traditional k-means clustering strategy for 50 iterations on all datasets. Then, an automatic hierarchical clustering method is used to form the clustering structure, and the merged results can be visualized using a dendrogram. Finally, the upper and lower approximations of similar classes are derived, and the core and fringe regions of each cluster are determined. Additionally, similarity threshold θ is 0.7 in the experiments.
Because NMI, ARI, and ACC are only adopted to the hard clustering results, three-way clustering results cannot calculate these values directly. In order to present the performances of our proposed algorithm, this study uses the core regions to form a clustering result, then calculate the NMI, ARI, and ACC by using the core region to represent the corresponding cluster. The clustering ensemble strategy is executed 50 times on all datasets, with an ensemble size of 50, to calculate the average NMI, ARI, and ACC values. The performances of the proposed algorithm on these three indicators are displayed in Table 3 and Figure 5, Figure 6 and Figure 7. To compare clustering effects, the performances of k-means, FCM, and DBSCAN are also presented in Table 3 and Figure 5, Figure 6 and Figure 7. The best performances for each dataset are highlighted in bold.
Table 3. The performances of different algorithms.
Table 3. The performances of different algorithms.
DatasetsAlgorithmARIAMIACC
SeedsK-means0.75000.70540.9095
FCM0.71610.69150.8952
DBSCAN0.70210.43960.3667
Ours0.81980.76850.9356
CreditK-means0.00910.0320.3741
FCM0.02720.03170.3917
DBSCAN0.01100.00060.389
Ours0.11160.10730.3987
IonosphereK-means0.0110.00060.5783
FCM0.17130.12720.7094
DBSCAN0.21740.14260.3932
Ours0.25000.20170.7833
LibrasK-means0.18370.18420.3389
FCM0.05970.330.1778
DBSCAN0.00250.22150.1000
Ours0.51930.71440.6182
EcoilK-means0.45420.57090.5565
FCM0.36790.56190.497
DBSCAN0.00800.0050.4256
Ours0.39370.49990.6100
SegmentationK-means0.03310.07360.2455
FCM0.38750.50620.6100
DBSCAN0.10670.33010.2939
Ours0.55010.69960.6720
ThyroidK-means0.21450.39110.5721
FCM0.42940.1760.786
DBSCAN0.31230.03560.4465
Ours0.59640.56280.8950
WdbcK-means0.00190.00520.5202
FCM0.72990.61380.9279
DBSCAN0.02740.01450.6098
Ours0.64410.52950.9320
WineK-means0.44830.44850.6461
FCM0.34920.40750.6854
DBSCAN0.27000.31370.5169
Ours0.58310.66740.8118
Through a comparative analysis of the data presented in Table 3 and Figure 5, Figure 6 and Figure 7, the following conclusions can be drawn:
(1).
By comparing the performance of our proposed three-way clustering algorithm with traditional clustering methods, such as k-means, FCM (Fuzzy C-Means), and DBSCAN (Density-Based Spatial Clustering of Applications with Noise), on AMI, ARI, and ACC, it can be found that our proposed algorithm demonstrates significant advantages on most datasets. Taking the Libras dataset as an example, after running the proposed algorithm, the resulting AMI, ARI, and ACC values are 0.5193, 0.7144, and 0.6182, respectively. In contrast, the AMI, ARI, and ACC values for the traditional k-means algorithm are only 0.1837, 0.1842, and 0.3389, respectively. This improvement is attributed to the dimensionality reduction of original high-dimensional data, mapping it to a lower-dimensional space, thus reducing data complexity. The introduction of co-occurrence probability enables more precise delineation of similar classes, allocating data points to core and fringe regions, better capturing the inherent structure of the data.
(2).
By comparing the proposed three-way clustering algorithm with other algorithms in terms of AMI, ARI, and ACC, we observed significant improvements in the proposed algorithm relative to others. Specifically, across all datasets, the proposed algorithm exhibited an average improvement of approximately 20% to 30% in ARI and ACC, and an average increase of about 15% to 35% in AMI. There are several potential reasons behind these improvements. Firstly, the proposed three-way clustering algorithm adopts an ensemble strategy, integrating concepts of data dimensionality reduction, co-occurrence frequencies, and similarity classes, thereby offering a more comprehensive consideration of the inherent structure of the data. Secondly, leveraging the single-linkage method of hierarchical clustering, the proposed three-way clustering algorithm effectively captures the degree of correlation among data points, resulting in more precise classification of data points into clusters. Additionally, by selecting the clustering result with the highest lifetime as the final merged result, the proposed three-way algorithm ensures the stability and consistency of the clustering results, rendering it more suitable for various data types and complex structures. The suboptimal performance on the Wdbc dataset may be due to algorithm sensitivity to different parameter settings, and parameter selection may vary across different datasets. Although our proposed algorithm shows significant improvements, certain algorithms may perform better under specific conditions due to their inherent characteristics. For example, algorithms like DBSCAN are particularly effective for datasets with noise and density variations, while hierarchical clustering can capture nested cluster structures. By comparing the actual runtime with the computational time complexity, it is concluded that the proposed algorithm strikes a balance between accuracy and computational efficiency. Although it is not the fastest, its robustness and ability to handle high-dimensional and noisy data make it a valuable tool in practical applications.
In summary, the proposed three-way clustering algorithm amalgamates ideas from data dimensionality reduction, co-occurrence frequency calculation, and similar class partitioning. Compared to traditional clustering algorithms, it demonstrates advantages in more nuanced data analysis and accurate clustering results, making it more feasible and effective in practical applications.

5. Conclusions

The theoretical contribution of this paper lies in the proposal of a novel three-way clustering framework that integrates dimensionality reduction, co-occurrence frequencies, and similarity classes with three-way clustering. The objective is to efficiently cluster heterogeneous data from multiple sources by leveraging inherent structural information. Initially, we employ principal component analysis (PCA) to reduce the dimensionality of the data, mapping high-dimensional data into a lower-dimensional space. This not only decreases computational complexity but also enhances clustering efficiency.
Subsequently, we introduce the concept of co-occurrence frequencies, considering the co-occurrence relationships between samples. By applying a threshold to the co-occurrence probability, samples are classified into similar classes, combined with the division into core and fringe regions. This ensures that the proposed algorithm not only accurately describes the intrinsic structure of the data but also exhibits robustness. The experimental results show that the proposed algorithm can improve clustering accuracy, particularly when dealing with complex data structures and significant noise interference. To further enhance the clustering process, we integrate these co-occurrence probabilities with a single-linkage hierarchical clustering method. This fusion enables us to construct a dendrogram that captures the similarity between different clusters. Lifecycle analysis is then employed to select the most stable clustering result, ensuring consistency and robustness.
The practical contribution of this paper is the improvement in clustering accuracy. Experimental results demonstrate that the proposed algorithm significantly enhances clustering precision, especially when handling complex data structures and substantial noise interference. This proves its practical effectiveness in various real-world scenarios. The method shows significant advantages across multiple datasets, highlighting its versatility and robustness in dealing with diverse and high-dimensional data. This adaptability makes it suitable for a wide range of applications, from bioinformatics to market segmentation.
Although the algorithm demonstrates significant advantages across multiple datasets during experimental validation, it does not consistently exhibit the expected improvements on certain specific datasets. This discrepancy may arise due to a partial mismatch between data characteristics and algorithm design, necessitating further exploration and refinement.
In future research, we will focus on the following aspects:
(1).
Adaptability of parameter selection:
The subjective nature of parameter thresholds in the algorithm may impact the stability of experimental results. To enhance algorithm robustness, considering more objective and adaptive parameter selection methods to accommodate different dataset requirements and application scenarios is essential.
(2).
Improving the Quality of Base Clustering:
The generation of base clustering using different feature subsets may lead to poor-quality results, negatively affecting the final ensemble clustering outcome. To enhance the quality of base clustering, we can employ automatic evaluation mechanisms based on the data’s intrinsic structure or utilize advanced clustering performance metrics. Additionally, introducing other methods such as setting evaluation functions will help eliminate the impact of low-quality base clustering, effectively improving the overall performance of ensemble clustering.
(3).
Adaptation Improvements for Specific Datasets:
The observation that the algorithm did not consistently exhibit expected improvements on specific datasets suggests a potential mismatch between data characteristics and algorithm design. Further work can include adapting the algorithm specifically for certain datasets, enhancing its generality and adaptability.

Author Contributions

Conceptualization, Y.M.; Data curation, A.L.; Formal analysis, A.L.; Funding acquisition, P.W.; Methodology, P.W.; Supervision, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (nos. 62076111, 61906078) and the Key Laboratory of Oceanographic Big Data Mining and Application of Zhejiang Province (no. OBDMA202002).

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, X.; Yao, Y. Ensemble selector for attribute reduction. Appl. Soft Comput. 2018, 70, 1–11. [Google Scholar] [CrossRef]
  2. Jiang, Z.; Yang, X.; Yu, H.; Liu, D.; Wang, P.; Qian, Y. Accelerator for multi-granularity attribute reduction. Knowl.-Based Syst. 2019, 177, 145–158. [Google Scholar] [CrossRef]
  3. Li, J.; Yang, X.; Song, X.; Li, J.; Wang, P.; Yu, D.J. Neighborhood attribute reduction: A multi-criterion approach. Int. J. Mach. Learn. Cybern. 2019, 10, 731–742. [Google Scholar] [CrossRef]
  4. Liu, K.; Yang, X.; Yu, H.; Fujita, H.; Chen, X.; Liu, D. Supervised information granulation strategy for attribute reduction. Int. J. Mach. Learn. Cybern. 2020, 11, 2149–2163. [Google Scholar] [CrossRef]
  5. Xu, S.; Yang, X.; Yu, H.; Yu, D.-J.; Yang, J.; Tsang, E.C. Multi-label learning with label-specific feature reduction. Knowl.-Based Syst. 2016, 104, 52–61. [Google Scholar] [CrossRef]
  6. Liu, K.; Yang, X.; Fujita, H. An efficient selector for multi-granularity attribute reduction. Inf. Sci. 2019, 505, 457–472. [Google Scholar] [CrossRef]
  7. Liu, K.; Yang, X.; Yu, H.; Mi, J.; Wang, P.; Chen, X. Rough set based semi-supervised feature selection via ensemble selector. Knowl.-Based Syst. 2020, 165, 282–296. [Google Scholar] [CrossRef]
  8. Xu, M.; Li, C.; Zhang, S.; Le Callet, P. State-of-the-art in 360 video/image processing: Perception, assessment and compression. IEEE J. Sel. Top. Signal Process. 2020, 14, 5–26. [Google Scholar] [CrossRef]
  9. Tov, O.; Alaluf, Y.; Nitzan, Y.; Patashnik, O.; Cohen-Or, D. Designing an encoder for StyleGAN image manipulation. ACM Tran. Graph. 2021, 40, 1–14. [Google Scholar] [CrossRef]
  10. Yang, X.; Qi, Y.; Song, X.; Yang, J. Test cost sensitive multigranulation rough set: Model and minimal cost selection. Inf. Sci. 2013, 250, 184–199. [Google Scholar] [CrossRef]
  11. Xu, W.; Guo, Y. Generalized multigranulation double-quantitative decision-theoretic rough set. Knowl.-Based Syst. 2016, 105, 190–205. [Google Scholar] [CrossRef]
  12. Li, W.; Xu, W.; Zhang, X.; Zhang, J. Updating approximations with dynamic objects based on local multigranulation rough sets in ordered information systems. Artif. Intell. Rev. 2021, 55, 1821–1855. [Google Scholar] [CrossRef]
  13. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 11. [Google Scholar] [CrossRef]
  14. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  15. Coley, C.W.; Jin, W.; Rogers, L.; Jamison, T.F.; Jaakkola, T.S.; Green, W.H.; Barzilay, R.; Jensen, K.F. A graph-convolutional neural network model for the prediction of chemical reactivity. Chem. Sci. 2019, 10, 370–377. [Google Scholar] [CrossRef] [PubMed]
  16. Strehl, A.; Ghosh, J. Cluster ensembles-a knowledge reuse framework for combing multiple partitions. J. Mach. Learn. Res. 2003, 3, 583–617. [Google Scholar]
  17. Fred, A.L.N.; Jain, A.K. Combining multiple clusterings using evidence accumulation. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 835–850. [Google Scholar] [CrossRef] [PubMed]
  18. Zhou, Z.; Tang, W. Cluster ensemble. Knowl.-Based Syst. 2006, 19, 77–83. [Google Scholar] [CrossRef]
  19. Huang, D.; Lai, J.; Wang, C. Ensemble clustering using factor graph. Pattern Recognit. 2016, 50, 131–142. [Google Scholar] [CrossRef]
  20. Huang, D.; Wang, C.; Lai, J. Locally weighted ensemble clustering. IEEE Trans. Cybern. 2018, 48, 1460–1473. [Google Scholar] [CrossRef]
  21. Xu, L.; Ding, S. A novel clustering ensemble model based on granular computing. Appl. Intell. 2021, 51, 5474–5488. [Google Scholar] [CrossRef]
  22. Zhou, P.; Wang, X.; Du, L.; Li, X.J. Clustering ensemble via structured hypergraph learning. Inf. Fusion 2022, 78, 171–178. [Google Scholar] [CrossRef]
  23. Yao, Y. The superiority of three-way decisions in probabilistic rough set models. Inf. Sci. 2011, 81, 1080–1096. [Google Scholar] [CrossRef]
  24. Yao, Y. Three-way decisions and cognitive computing. Cogn. Comput. 2016, 8, 543–554. [Google Scholar] [CrossRef]
  25. Yao, Y. Tri-level thinking: Models of three-way decision. Int. J. Mach. Learn. Cybern. 2020, 11, 947–959. [Google Scholar] [CrossRef]
  26. Yao, Y. The geometry of three-way decision. Appl. Intell. 2021, 51, 6298–6325. [Google Scholar] [CrossRef]
  27. Pawlak, Z. Rough sets. Int. J. Comput. Inf. Sci. 1982, 11, 314–356. [Google Scholar] [CrossRef]
  28. Yang, X.; Liang, S.; Yu, H.; Gao, S.; Qian, Y. Pseudo-label neighborhood rough set: Measures and attribute reductions. Int. J. Approx. Reason. 2019, 105, 112–129. [Google Scholar] [CrossRef]
  29. Dou, H.; Yang, X.; Song, X.; Yu, H.; Wu, W.Z.; Yang, J. Decision-theoretic rough set: A multi-cost strategy. Knowl.-Based Syst. 2016, 91, 71–83. [Google Scholar] [CrossRef]
  30. Darwiche, A. Bayesian networks. Found. Artif. Intell. 2008, 3, 467–509. [Google Scholar]
  31. Daly, R.; Shen, Q.; Aitken, S. Learning Bayesian networks: Approaches and issues. Knowl. Eng. Rev. 2011, 26, 99–157. [Google Scholar] [CrossRef]
  32. Yang, X.; Liu, D.; Yang, X.; Liu, K.; Li, T. Incremental fuzzy probability decision-theoretic approaches to dynamic three-way approximations. Inf. Sci. 2021, 550, 71–90. [Google Scholar] [CrossRef]
  33. Li, C.; Zhou, J.; Kou, P.; Xiao, J. A novel chaotic particle swarm optimization based fuzzy clustering algorithm. Neurocomputing 2012, 83, 98–109. [Google Scholar] [CrossRef]
  34. Yu, H. A framework of three-way cluster analysis. In Proceedings of the International Joint Conference on Rough Sets, Olsztyn, Poland, 3–7 July 2017; pp. 300–312. [Google Scholar]
  35. Wu, T.; Fan, J.; Wang, P. An improved three-way clustering based on ensemble strategy. Mathematics 2022, 10, 1457. [Google Scholar] [CrossRef]
  36. Wang, P.; Yang, X. Three-way clustering method based on stability theory. IEEE Access 2021, 9, 33944–33953. [Google Scholar] [CrossRef]
  37. Wang, P.; Shi, H.; Yang, X.; Mi, J. Three-way k-means: Integrating k-means and three-way decision. Int. J. Mach. Learn. Cybern. 2019, 10, 2767–2777. [Google Scholar] [CrossRef]
  38. Fan, J.; Wang, P.; Jiang, C.; Yang, X.; Song, J. Ensemble learning using three-way density-sensitive spectral clustering. Int. J. Approx. Reason. 2022, 149, 70–84. [Google Scholar] [CrossRef]
  39. Wang, P.; Yang, X.; Ding, W.; Zhan, J.; Yao, Y. Three-way clustering: Foundations, survey and challenges. Appl. Soft Comput. 2024, 151, 111131. [Google Scholar] [CrossRef]
  40. Wang, P.; Yao, Y. CE3: A three-way clustering method based on mathematical morphology. Knowl.-Based Syst. 2018, 155, 54–65. [Google Scholar] [CrossRef]
  41. Li, F.; Qian, Y.; Wang, J.; Dang, C.; Jing, L. Clustering ensemble based on sample’s stability. Artif. Intell. 2019, 273, 37–55. [Google Scholar] [CrossRef]
  42. Yu, H.; Chang, Z.; Wang, G.; Chen, X. An efficient three-way clustering algorithm based on gravitational search. Int. J. Mach. Learn. Cyber. 2020, 11, 1003–1016. [Google Scholar] [CrossRef]
  43. Jia, X.; Rao, Y.; Li, W.; Yang, S.; Yu, H. An automatic three-way clustering method based on sample similarity. Int. J. Mach. Learn. Cybern. 2021, 12, 1545–1556. [Google Scholar] [CrossRef]
  44. Wang, P.; Wu, T.; Yao, Y. A three-way adaptive density peak clustering (3W-ADPC) method. Appl. Intell. 2023, 53, 23966–23982. [Google Scholar] [CrossRef]
  45. Vittoria, B.; Lucia, M.C.; Domenico, V. A short review on minimum description length: An application to dimension reduction in PCA. Entropy 2022, 24, 269. [Google Scholar] [CrossRef] [PubMed]
  46. Goparaju, B.; Rao, S.B. A DDoS attack detection using PCA dimensionality reduction and support vector machine. Int. J. Commun. Netw. Inf. Sec. 2022, 14, 1–8. [Google Scholar] [CrossRef]
  47. Boushaba, A.; Cauet, S.; Chamroo, A.; Etien, E.; Rambault, L. Comparative study between physics-informed CNN and PCA in induction motor broken bars MCSA Detection. Sensors 2022, 22, 9494. [Google Scholar] [CrossRef] [PubMed]
  48. Geophys, J. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar]
  49. Ng, A.; Jordan, M.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2001, 14, 849–856. [Google Scholar]
  50. Ford, J.K.; MacCallum, R.C.; Tait, M. The application of exploratory factor analysis in applied psychology: A critical review and analysis. Pers. Psychol. 1986, 39, 291–314. [Google Scholar] [CrossRef]
  51. Torgerson, W.S. Multidimensional scaling: I. Theory and method. Psychometrika 1952, 17, 401–419. [Google Scholar] [CrossRef]
  52. Yu, H.; Jiao, P.; Yao, Y.Y.; Wang, G.Y. Detecting and refining overlapping regions in complex networks with three-way decisions. Inf. Sci. 2016, 373, 21–41. [Google Scholar] [CrossRef]
  53. MacQueen, J. Some methods for classification and analysis of multivariate observations. Berkeley Symp. Math. Stat. Probab. 1967, 5, 281–297. [Google Scholar]
  54. Unrau, R.C.; Krieger, O.; Gamsa, B.; Stumm, M. Hierarchical clustering: A structure for scalable multiprocessor operating system design. J. Supercomput. 1995, 9, 105–134. [Google Scholar] [CrossRef]
  55. Shi, N.; Liu, X.; Guan, Y. Research on k-means clustering algorithm: An improved k-means clustering algorithm. In Proceedings of the 2010 Third International Symposium on Intelligent Information Technology and Security Informatics, Jian, China, 2–4 April 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 63–67. [Google Scholar]
  56. Dimitriadou, E.; Weingessel, A.; Hornik, K. Voting-merging: An ensemble method for clustering. In Proceedings of the 2010 International Conference on Artificial Neural Networks, Vienna, Austria, 21–25 August 2001; pp. 217–224. [Google Scholar]
  57. Fan, J.; Wang, X.; Wu, T.; Zhu, J.; Wang, P. Three-way ensemble clustering based on sample’s perturbation theory. Mathematics 2022, 10, 2598. [Google Scholar] [CrossRef]
  58. Wang, Y.; Liu, J.; Huang, Y.; Feng, X. Using hashtag graph-based topic model to connect semantically-related words without co-occurrence in microblogs. IEEE Tran. Knowl. Data Eng. 2016, 28, 1919–1933. [Google Scholar] [CrossRef]
  59. Abdalrada, A.S.; Abawajy, J.; Al-Quraishi, T.; Islam, S.M.S. Machine learning models for prediction of co-occurrence of diabetes and cardiovascular diseases: A retrospective cohort study. J. Diab. Meta. Disord. 2022, 21, 251–261. [Google Scholar] [CrossRef] [PubMed]
  60. Bache, K.; Lichman, M. UCI Machine Learning Repository; University of California: Irvine, CA, USA, 2013. [Google Scholar]
  61. Shi, H.; Wang, P.; Yang, X.; Yu, H. An improved mean imputation clustering algorithm for incomplete data. Neural Process. Lett. 2021, 54, 3537–3550. [Google Scholar] [CrossRef]
  62. Jiang, W.; Xu, X.; Wen, Z.; Wei, L. Applying the similarity theory to model dust dispersion during coal-mine tunneling. Process Saf. Environ. 2021, 148, 415–427. [Google Scholar] [CrossRef]
  63. Hoffman, M.; Steinley, D.; Brusco, M.J. A note on using the adjusted rand index for link prediction in networks. Soc. Netw. 2015, 42, 72–79. [Google Scholar] [CrossRef]
  64. Steinley, D.; Brusco, J.M. A note on the expected value of the Rand index. Brit. J. Math. Stat. Psychol. 2018, 71, 287–299. [Google Scholar] [CrossRef]
  65. Amodio, S.; D’Ambrosio, A.; Iorio, C.; Siciliano, R. Adjusted concordance index: An extension of the adjusted rand index to fuzzy partitions. J. Classif. 2021, 38, 112–128. [Google Scholar]
Figure 1. Illustration of Dimensionality Reduction by PCA.
Figure 1. Illustration of Dimensionality Reduction by PCA.
Mathematics 12 01951 g001
Figure 2. The framework of clustering ensemble.
Figure 2. The framework of clustering ensemble.
Mathematics 12 01951 g002
Figure 3. An example of a dataset with four clustering results.
Figure 3. An example of a dataset with four clustering results.
Mathematics 12 01951 g003
Figure 4. Single-linkage dendrogram of hierarchical clustering.
Figure 4. Single-linkage dendrogram of hierarchical clustering.
Mathematics 12 01951 g004
Figure 5. ARI Comparison of Algorithms Embedded in K-means.
Figure 5. ARI Comparison of Algorithms Embedded in K-means.
Mathematics 12 01951 g005
Figure 6. AMI Comparison of Algorithms Embedded in K-means.
Figure 6. AMI Comparison of Algorithms Embedded in K-means.
Mathematics 12 01951 g006
Figure 7. ACC Comparison of Algorithms Embedded in K-means.
Figure 7. ACC Comparison of Algorithms Embedded in K-means.
Mathematics 12 01951 g007
Table 1. The co-association frequency matrix of Figure 3.
Table 1. The co-association frequency matrix of Figure 3.
p i j x 1 x 2 x 3 x 4 x 5 x 6
x 1 110.50.2500
x 2 110.50.2500
x 3 0.50.510.750.250.25
x 4 0.250.250.7510.50.25
x 5 000.250.511
x 6 000.250.2511
Table 2. Datasets Used in Experiments.
Table 2. Datasets Used in Experiments.
IDDatasetsNumbersDimensionsCategories
1Seeds27073
2Credit149393
3Ionosphere351342
4Libras3609015
5Ecoil21073
6Segmentation2310197
7Thyroid21593
8Wdbc569302
9Wine178133
10Waveform5000403
11Iris15043
12Yeast1484810
13Dermatology366346
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, A.; Meng, Y.; Wang, P. Similarity-Based Three-Way Clustering by Using Dimensionality Reduction. Mathematics 2024, 12, 1951. https://doi.org/10.3390/math12131951

AMA Style

Li A, Meng Y, Wang P. Similarity-Based Three-Way Clustering by Using Dimensionality Reduction. Mathematics. 2024; 12(13):1951. https://doi.org/10.3390/math12131951

Chicago/Turabian Style

Li, Anlong, Yiping Meng, and Pingxin Wang. 2024. "Similarity-Based Three-Way Clustering by Using Dimensionality Reduction" Mathematics 12, no. 13: 1951. https://doi.org/10.3390/math12131951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop