Next Article in Journal
Evaluation of Urban Transportation Resilience under Extreme Weather Events
Previous Article in Journal
Using Machine Learning to Predict Pedestrian Compliance at Crosswalks in Jordan
Previous Article in Special Issue
Cross-Domain Person Re-Identification Based on Feature Fusion Invariance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Spectral Clustering Algorithm for Non-Linear Graph Embedding in Information Networks

1
College of Information and Electronic Engineering, Hunan City University, Yiyang 413000, China
2
College of Information and Engineer, Hunan University, Changsha 410008, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2024, 14(11), 4946; https://doi.org/10.3390/app14114946
Submission received: 11 April 2024 / Revised: 4 June 2024 / Accepted: 5 June 2024 / Published: 6 June 2024

Abstract

:
With the development of network technology, information networks have become one of the most important means for people to understand society. As the scale of information networks expands, the construction of network graphs and high-dimensional feature representation will become major factors affecting the performance of spectral clustering algorithms. To address this issue, in this paper, we propose a spectral clustering algorithm based on similarity graphs and non-linear deep embedding, named S E G _ S C . This algorithm introduces a new spectral clustering model that explores the underlying structure of graphs through sparse similarity graphs and deep graph representation learning, thereby enhancing graph clustering performance. Experimental analysis with multiple types of real datasets shows that the performance of this model surpasses several advanced benchmark algorithms and performs well in clustering on medium- to large-scale information networks.

1. Introduction

Information networks, an essential form for expressing complex relationships between objects, are ubiquitous in our daily lives and work. They include social information networks formed by social media platforms such as WeChat and semantic information networks constituted by webpages on the Internet.
Spectral clustering, detailed in numerous studies [1,2,3,4], comprises three main steps: constructing similarity graphs, transforming graphs into low-dimensional vectors, and utilizing the K-means clustering algorithm. For example, Cai et al. [1,2] proved that a similarity graph based on the representative point strategy could be successfully applied to spectral clustering for information networks. Luo et al. [3] used K-means to select representative points from original data and constructed a multi-layer kernel-free similarity graph based on representative points. Shanham et al. [4] used the deep learning network approach to overcome the scalability and generalization issues of low-dimensional embedding in spectral clustering. In summary, at present, the optimization of spectral clustering algorithms mainly consists of the construction of similarity graphs and low-dimensional embedding representations.
With the development of information networks, designing effective algorithms to construct information network graphs to represent their semantic information has become a focal research issue. Traditional methods [5,6] generally adopt high-dimensional sparse vectors, representing the structure and semantic information of information networks through adjacency matrices and involving the design of various application algorithms based on the above. However, high-dimensional sparse representation requires more computational space and time as the network scale increases, leading to poor scalability and accuracy in a graph structure.
The k-nearest neighbor (KNN) graph is a manifold graph construction learning method, which includes methods such as Isomap [7] and local linear embedding [8]. This particular model has good generalization capabilities and can map topological manifolds. If high-dimensional manifolds are embedded in low-dimensional manifold spaces, the properties of Euclidean space still exist locally. The KNN algorithm is known for its stability, simplicity, and efficiency. However, it presents limitations when handling large datasets or high-dimensional data, which can result in increased computational complexity and limited clustering results. In recent years, considering nearest neighbor search methods, D. Cai et al. [1,3,9] improved the KNN graph by introducing a KNN graph clustering algorithm that utilizes representative points. These representative points represent each cluster during computation, and noise has minimal influence on this approach. Moreover, for KNN graph construction, the representative point graph has been applied in spectral clustering. By computing a low-rank approximation of the affinity tensor graph, the initial unified affinity matrix is refined with reliable information from the confidence affinity matrix.
For low-dimensional vector representation, there are many effective methods available. Most models [10,11] that learn graph construction first use linear encoding and then apply singular value decomposition (SVD) [12] for linear dimension reduction. Linear encoding [6,13] produces sparse representation, where each data point is represented as a linear representation of a small number of base vectors. Mathematically, linear encoding is defined as finding two matrices U and Z, whose product can best approximate the data X U Z . Large-scale sparse coding clustering (LSC) [9] combines linear encoding, SVD, and K-means into a spectral clustering method. It has a solid theoretical basis and has contributed to significant advancements in clustering research on large-scale data. GraRep [14] is a novel model designed for learning weighted graph vertex representations by generating low-dimensional vectors to represent graphs. However, these methods all involve the use of SVD for linear dimension reduction, and more effective non-linear dimension reduction techniques have yet to be explored.
In recent years, deep learning [4,15] has been widely applied in graph representation learning. Implementing a greedy layer training strategy can improve computational performance. Stacked autoencoders [16] are considered an effective method for mapping from the original representation space to low-dimensional spaces, with the capability to produce highly non-linear projections.
More recently, sparse autoencoders have been used to replace the SVD step in spectral clustering. Shao et al. [17] proposed a deep architecture called DLC for fast graph clustering with linear encoders. Their method combines feature embedding and linear encoding and was implemented through a deep hierarchical architecture. Cao et al. [18] proposed the DNGR model for learning graph representations. The model utilizes pointwise mutual information (PMI) to represent the original graph. It then projects this low-dimensional representation using stacked denoising autoencoders. The authors theoretically and empirically demonstrate the effectiveness of the model in clustering and classification processes. Tian et al. [13] indicated in their study that the computational complexity of autoencoders, O ( n ) , is considerably lower than the characteristic decomposition time, O ( n 3 ) , of spectral clustering. They proposed a non-linear spectral clustering algorithm based on deep learning. In this method, the graph is embedded in deep learning through autoencoders, followed by clustering using the K-means algorithm. This study on non-linear mapping spectral clustering algorithms made significant contributions to the field; however, the study presented the following issues: (1) for large-scale data, directly representing graph structures with high-dimensional similarity matrices does not reflect the essential features of graph structures; and (2) the authors simply propose a framework for non-linear spectral clustering algorithms without providing theoretical proof.
In our study, we utilized sparse representation for the graph similarity matrix, which typically enhances clustering accuracy by eliminating noise information that can impact clustering results. In representation learning, deep structures are used to construct features. Deep structures can benefit from hierarchical training for fine-tuning, which helps in finding improved local solutions to meet the expectation maximization objectives. The method proposed in this paper is called sparse encoder graph spectral clustering ( S E G _ S C ) .
The S E G _ S C algorithm proposed in this paper makes the following contributions:
  • In this study, we study the manifold learning similarity matrix and linear low-dimensional mapping of semantic information networks. We define a model structure using the K-means clustering algorithm, similar to spectral algorithms.
  • Theoretically, the results presented in this paper prove that the model adopts a clustering method consistent with the spectral clustering objective function and captures the non-linear features of semantic information networks.
  • Empirically, our results demonstrate that the model can enhance the clustering performance of semantic information networks by capturing meaningful semantic, relational, and structural information. Our experimental results show that the model can effectively complete clustering tasks.
The remainder of this paper is arranged as follows: in Section 2, we review related works; in Section 3, we describe the proposed framework, its theoretical analysis, computation complexity analysis, and the optimization algorithm; in Section 4, we report the experimental results; and finally, we conclude the paper in Section 5.

2. Related Work

2.1. Spectral Clustering

Spectral clustering is an algorithm derived from graph theory and has been widely utilized in clustering. The primary aim is to represent all nodes using a Laplacian matrix. In this matrix, the edge weights between nodes that are far apart are lower; in contrast, those that are closer are higher. The Laplacian matrix is then subjected to eigenvalue decomposition to identify its k-smallest eigenvalues, which are utilized to derive a reduced-dimension representation of the network. Finally, the K-means algorithm is employed for clustering. The process is illustrated in Figure 1.
There are three methods for constructing the similarity graph in the Laplacian matrix: the adjacency method, the KNN method, and the fully connected method. The adjacency method sets a distance threshold and then measures the similarity between any two nodes using the Euclidean distance. The KNN algorithm traverses all sample points, selecting the nearest k points as neighbors for each sample in order to calculate their similarity. The fully connected method calculates the similarity values between all points.
The objective of spectral clustering is to partition the graph into k-disconnected subgraphs while minimizing the graph cut. The normalized cut criterion [19] is defined as follows:
Definition 1.
(Normalized Cut Criterion)
N o r m a l i z e d C u t ( C 1 , C 2 , , C k ) = i = 1 k c u t ( C i , C i ¯ ) v o l ( C i ) ,
where v o l ( C i ) = v C i d v , C i C j = Φ , i = 1 k C i = V , C i ¯ denotes the complement of C i and c u t ( C i , C i ¯ ) represents the number of connections between the two subgraphs C i and C i ¯ .
Solving the normalized cut set is an NP-hard problem; according to N o r m a l i z e d C u t ( C 1 , C 2 , , C k ) = i = 1 k h i T L h i = i = 1 k ( H T L H ) i i = t r ( H T L H ) , its approximation can be described as a trace function minimization Formula (2).
arg min H t r ( H T L H )   s . t .   H T D H = I ,
where L = D S , D = d i a g ( d 1 , d 2 , , d n ) , and properties of the Laplacian matrix L are as follows:
  • L is a symmetric and positive semi-definite matrix;
  • The number of graph connections equals the number of eigenvalues equal to zero, i.e., Rank(L) = nk;
  • For vector f R n , there is f T L h = 1 2 i , j = 1 n S i j ( f i f j ) 2 .
This objective function calculates the eigenvectors of the matrix L. According to the matrix computation Theorem 1, H is composed of the eigenvectors corresponding to the k-smallest non-negative singular values of the normalized Laplacian matrix. The final clustering result is obtained by running the K-means algorithm on the graph’s embedding matrix H.
Theorem 1.
(Eckart–Young–Mirsky Theorem) Given an r-rank matrix P, its SVD: P = V Σ U T , where V T V = I , U T U = I and then
arg min r a n k ( P ) = k | | P P | | F = V Σ U T ,
where Σ ; besides only containing the k largest singular values and replacing other singular values with 0, the rest remains the same as the matrix Σ. This theorem indicates that reconstructing the matrix using the k largest singular vectors from singular value decomposition provides the best rank-k approximation of the original matrix.
The effectiveness of spectral clustering depends on the similarity matrix, making it very effective for clustering sparse data, a task that is challenging for the K-means algorithm. Moreover, due to the utilization of dimensionality reduction techniques, spectral clustering handles the complexity of high-dimensional clustering more effectively than traditional clustering methods such as K-means.

2.2. Linear Coding

Linear coding (LC) [9] is a matrix decomposition technique that represents the original data as a combination of a set of base vectors and their corresponding “compressed” representations. Its data representation form is
X U Z ,
where X = ( x 1 , x 2 , , x n ) R m × n , U R m × p , and the column vectors of U are the basis vectors. Z R p × n and the column vectors of Z represent p-dimensional vectors with respect to the base vectors of U. Therefore, the matrix decomposition of LC is defined as the following optimization function problem:
arg min U , Z | | X U Z | | 2 .
LC adds sparse constraints to each column of Z in the optimization problem described in Formula (5). LC implements a data-sparse representation based on matrix decomposition, where each data point is a linear combination of a few base vectors. This approach allows for a more accurate interpretation of each data point.

2.3. Low-Dimensional Embedding Learning of Information Networks

Dimension reduction is the process of transforming the original high-dimensional attribute space into a lower-dimensional subspace data representation through mathematical transformations. At present, dimension reduction techniques are mainly divided into two categories: linear and non-linear methods.
Classic linear dimension reduction methods include principal component analysis (PCA), multi-dimensional scaling (MDS), linear discriminant analysis (LDA), etc. PCA [20] projects data into a low-dimensional space representation through linear transformation. Its optimal objective function is to minimize Formula (6):
arg min W t r a c e ( W T X X T W )   s . t .   W T W = I .
The goal of MDS is to construct a low-dimensional space where the similarity relationships between nodes can be represented. In this space, the distances are consistent with those in the original data, taking into account the global Euclidean distance of samples during sample dimension reduction. LDA minimizes differences between nodes in the same cluster and maximizes differences between nodes in different clusters through linear projection.
Recently, neural information network learning has achieved significant success in various applications, including image classification, speech recognition, and natural language processing. Collobert et al.’s research [21,22,23,24,25] made training in deep architectural structures possible and effective. One of the strategies employed in their research was the greedy layered unsupervised pre-training strategy, whose aim was to learn a useful representation layer by layer and then set the output features as inputs for the next layer. Each layer in this process involves some form of non-linearity, such as non-linear activation functions (e.g., sigmoid or tanh) or regularization of features (e.g., sparse constraints). By stacking these non-linear layers together, deep learning can be used to generate improved representations. The process from the input layer to the hidden layer is called encoding, and the process from the hidden layer to the output layer is called decoding, as illustrated in the model process in Figure 2.

3. Model Framework

The model proposed in this paper is driven by a three-step scheme of spectral clustering, which includes the construction of similarity graphs, low-dimensional graph embedding, and the K-means clustering algorithm. In our study, we first implemented the reconstruction of the similarity matrix for spectral clustering, followed by creating a low-dimensional embedded graph. Our spectral clustering model was utilized for large-scale clustering. The basis of this model is to design an effective method for constructing a similarity graph and performing Laplacian matrix eigen decomposition. Specifically, in the similarity graph S, we attempted to design it as symmetric and positive semi-definite, satisfying the following form:
S = Z T Z ,
where Z R p × n represents the p-dimensional representation of the original graph, p n .
The purpose of the S E G _ S C spectral clustering model is to first represent the information network as a graph G = ( V , S ) , where S is the node similarity graph. Next, it embeds S into a low-dimensional graph embedding matrix R and then identifies a disjoint partition C = { C 1 , C 2 , , C k } using the K-means algorithm, where k represents the number of clusters.

3.1. Similarity Matrix

KNN [26] is a commonly used manifold learning method. By utilizing a test sample, we integrated the concept of k-nearest neighbor information and distance measurement with the normalized Laplacian matrix of spectral clustering to derive a suitable weighted information network similarity matrix. This matrix can be represented by the following Formula (8):
S ( v i , v j ) = A i j v t N ( v i ) A i t ,
where v t is the k-nearest neighbor node of v i .
S defines the similarity relationship between each node and all its adjacent nodes; however, this representation is very time-consuming, especially for large-scale data. According to the SC method, for any data point v i , our approximation can be represented as shown in Formula (9):
v i ˙ = j = 1 p Z j i U j ,
where the value of U is randomly selected from the information network graph to create a set of p landmark nodes, generating the landmark node set U = { u 1 , u 2 , , u p } . Calculating the relationship values of U with the k-nearest neighbor nodes, we obtained the similarity graph. According to the Nadaraya–Watson method [27], the sparse representation Z is defined as shown in Formula (10):
Z ( u j , v i ) = K ( v i , u j ) u t U , t = { i 1 , i 2 , , i k } K ( v i , u t ) ,
where K(:) is the Gaussian kernel function. Therefore, the definition of the similarity matrix S is as shown in Formula (11):
S = Z T Z .

3.2. Non-Linear Low-Dimensional Graph Embedding

First, in the present paper, we analyze the non-linear spectral clustering properties of the single-layer network. Second, the model representation was studied, and we describe the deep embedding graph of the multi-layer network. Third, we analyze the computation process of the multi-layer network based on the objection function. Finally, we analyze its time performance and the algorithm.

3.2.1. Spectral Analysis on Single-Layer Graph

Using the method of graph dimension reduction based on non-linear projection, the projection of sample points in the low-dimensional coordinate system is as follows:
R i j = f ( W j T S i ) ,
where S i is the i-th row vector of the similarity matrix S, W j is the j-th row vector of parameter W, and R i = ( R i 1 , , R i d ) is the low-dimensional embedding landmark of S i . R i approximates the reconstruction of X i , which can be represented as
S ^ i = j = 1 d g ( R i j W j ) .
The objective formula for the difference in distance between the original sample S i and the projected reconstructed sample S ^ i is as follows:
i = 1 n | | j = 1 d g ( R i j W j ) S i | | 2 2 = i = 1 n g ( R i T W ) g ( W T R i ) 2 i = 1 n g ( R i T W ) S i + c o n s t t r ( i = 1 n g ( W R i T ) g ( R i W T ) ) ,
where W = ( w 1 , w 2 , , w d ) ; hence, the objective formula solving process can be approximated as
a r g m a x W , R i t r ( i = 1 n g ( W R i T ) g ( R i W T ) ) s . t . W T W = I .
Therefore, the model ensures that the distance to the sample points on the hyperplane is sufficiently close, and the projections of the sample points on the hyperplane are separated as far as possible.
Notably, when the function g() is a linear function, this objective model transforms into the objective function of regular spectral clustering, as shown in Formula (16):
m a x S t r ( W S W T ) ,   s . t .   W T W = I .
Therefore, Formula (15) is defined as the non-linear objective model of spectral clustering.

3.2.2. Deep Graph Embedding on a Multi-Layer Graph

As previously stated, we can consider S as a training set containing n nodes. We consider multi-layer information networks in each layer. The objective formula is as follows:
arg max W ( l ) , R i l = 1 m t r ( i = 1 n g ( W ( l ) ( R i ( l ) ) T ) g ( R i ( l ) ( W ( l ) ) T ) ) s . t . W T W = I .
Assuming the multi-layer information network deep architecture as shown in Figure 2, the input of the first layer is denoted as X ( 1 ) = S . More hidden layers need to be added to the multi-layer architecture.
It is widely known that multi-layer information networks are divided into an input layer, hidden layers, and an output layer; the structure of the i-th neuron in layer (l + 1) is shown in Figure 3.
The formula for h W ( l + 1 ) , b ( l ) ( l + 1 ) ( X ( l ) ) is:
h W ( l + 1 ) , b ( l ) ( l + 1 ) ( X ( l ) ) = f ( ( W ( l + 1 ) ) T X ) = f ( j = 1 q w i j ( l ) x j ( l ) + b i ( l ) ) ,
where f is the neuron activation function, which is generally a sigmoid or tanh function. X ( l ) represents the data of the layer l, assuming that the number of neural units in layer l is q + 1 .
Assuming there are r training samples X 1 , , X r through the deep neural information networks, if the input sample is X i , the output of the output layer is h W , b ( X i ) , and the actual output should be Y i . Therefore, a loss function with regularization terms is defined as
J ( W , b ) = 1 r i = 1 r J ( W , b , X i , Y i ) + λ 2 l = 1 n l 1 i = 1 s l + 1 j = 1 s l ( w i j ( l ) ) 2 ,
where J ( W , b , X i , Y i ) = 1 2 | | h W , b ( X i ) Y i | | 2 is the sum of squared errors of the output layer. Here, n l denotes the number of layers in the entire information network, and s l indicates the number of units in layer l, excluding bias units. λ represents the regularization term parameter.

3.2.3. Computation Analysis Based on Deep Graph Embedding

The regularization term stabilizes the algorithm and suppresses overfitting. J ( W , b ) is a non-convex function; thus, using the gradient descent method can easily lead to local optimization results. In addition, the initialization of each parameter w i j ( l ) and b i ( l ) must be set randomly and not all set to zero. If this is not performed, the output will be the same for any input.
The step formula for parameters w i j ( l ) and b i ( l ) is
w i j ( l ) = w i j ( l ) λ J ( W , b ) w i j ( l ) b i ( l ) = b i ( l ) λ J ( W , b ) b i ( l ) ,
where λ is the step learning parameter.
J ( W , b , X i , Y i ) w i j ( l ) = J ( W , b , X i , Y i ) I i ( l + 1 ) I i ( l + 1 ) w i j ( l ) ,
where I i ( l + 1 ) = w i j ( l ) x j ( l ) + b i ( l ) . Assuming e i ( l + 1 ) = J ( W , b , X i , Y i ) I i ( l + 1 ) , then
J ( W , b , X i , Y i ) w i j ( l ) = x j ( l ) e i ( l + 1 ) .
Similarly,
J ( W , b , X i , Y i ) b i ( l ) = J ( W , b , X i , Y i ) I i ( l + 1 ) I i ( l + 1 ) b i ( l ) = e i ( l + 1 ) ,
calculate e i ( l + 1 ) ,
e i ( l + 1 ) = J ( W , b , X i , Y i ) I i ( l + 1 ) = J ( W , b , X i , Y i ) O i ( l + 1 ) O i ( l + 1 ) I i ( l + 1 ) ,
where O i ( l + 1 ) = f ( I i ( l + 1 ) ) ; hence, the output layer e i ( l + 1 ) = J ( W , b , X i , Y i ) O i ( l + 1 ) f ( I i ( l + 1 ) ) ; thus, the entire output vector E ( l + 1 ) = O J ( W , b , X i , Y i ) f ( I ( l + 1 ) ) , where O J ( W , b , X i , Y i ) = { J ( W , b , X i , Y i ) O 1 ( l + 1 ) , , J ( W , b , X i , Y i ) O n ( l + 1 ) } and ⊙ represents the Hadamard product. Therefore, the middle layer is
e i ( l + 1 ) = J ( W , b , X i , Y i ) I i ( l + 1 ) = k = 1 r J ( W , b , X i , Y i ) I k ( l + 2 ) I k ( l + 2 ) O i ( l + 1 ) O i ( l + 1 ) I k ( l + 1 ) = k = 1 r e k ( l + 2 ) ( w k i ( l + 2 ) O i ( l + 1 ) + b k ( l + 2 ) ) O i ( l + 1 ) f ( I k ( l + 1 ) ) = k = 1 r e k ( l + 2 ) w k ( l + 2 ) f ( I k ( l + 1 ) ) ,
where E ( l + 1 ) = ( ( W ( l + 2 ) ) T E ( l + 2 ) ) f ( I ( l + 1 ) ) . Thus, the parameters were trained using the gradient descent method.
W ( l ) = W ( l ) λ r i = 1 r E ( l ) ( X i ) X ( l 1 ) ( X i )
b ( l ) = b ( l ) λ r i = 1 r E ( l ) ( X i ) ,
where r represents the number of samples.

3.2.4. Computation Complexity Analysis of an Algorithm on a Multi-Layer Graph

The results presented in this paper refer to the deep model as Sparse Encoding Graph Spectral Clustering ( S E G _ S C ), and its implementation process is shown in Algorithm 1.
Suppose we possess the original data X R n × d and use p landmark points; we need O ( p d n ) to construct the similarity matrix and O ( n ) to complete the multi-layer graph embedding. K-means requires O ( n k t ) , where t is the number of iterations. In Algorithm 1, the computational complexity is O ( p d n + n + n k t ) .
Algorithm 1 S E G _ S C algorithm
Require: 
Original information network X, number of landmark nodes p, and m-layer information network structure;
Ensure: 
Clustering partitions C 1 , C 2 , , C k
  •   Calculate the similarity matrix S according to Formula (11);
  •   for i = 1 to n do
    • 1. Build a multi-layer model as shown in Figure 2 based on the similarity matrix S. Let X ( 1 ) = S be the input layer;
    • 2. Train the m-layer model by optimizing the target Formula (17). The node parameter calculation process is illustrated in Formulas (22) and (23), obtaining the activation h i of the hidden layer;
  •   end for
  •   Execute K-means for R ( m 2 )

4. Experiment

In the following section, we report on the experiments conducted in our study. First, we introduce the datasets used in the experiments, the evaluation metrics, and the benchmark algorithms. Subsequently, we demonstrate the effectiveness of the proposed S E G _ S C algorithm through empirical testing.

4.1. Experimental Data

To evaluate model performance, in this study we tested the model on various real-world datasets obtained from the UCI Machine Learning Repository and social information network datasets. A brief description of the datasets used can be found below as follows:
  • Wine: The Wine dataset, sourced from the UCI Machine Learning Repository (Asuncion and Newman, 2007), comprises 178 instances and 13 features. These data are the results of chemical analyses of mature wine grapes. All samples are labeled with three categories of wine. In our study, we established a similarity graph.
  • Pendigit: This dataset is a handwritten digit dataset consisting of 3498 samples and 16 features created by 44 writers.
  • Minist: This dataset consists of handwritten digits and is available on Yann LeCun’s webpage. Each image is represented as a 784-dimensional vector and there are 4000 images in total.
  • Blogdata: This dataset is a social relationship information network provided by blog authors. This information network contains 10,321 nodes, 333,983 edges, and 39 different cluster labels. These labels represent the inferred and provided blog interests of the blog authors.
  • 20 newsgroups: This dataset is a news social network containing 18,846 newsgroup documents and 26,214 features divided into 20 different groups.
  • Seimic: This dataset is used to classify different types of mobile vehicles, selecting 98,000 samples, 50 features, and 3 classifications.
  • Covtype: This dataset is a forest cover-type dataset containing 581,012 data, 54 features, and 2 classes. This dataset contains tree observations from four areas of the Roosevelt National Forest in Colorado, USA.
Detailed information on all of these semantic networks is summarized in Table 1.

4.2. Benchmark Algorithms

To demonstrate the effectiveness of S E G _ S C , the following methods were used as benchmark algorithms. A brief description of each method is found below as follows:
  • PCA [20]: A common matrix decomposition method used for dimensionality reduction or feature extraction. PCA defines a loss function based on the distance between the original sample and the reconstructed sample after projection. PCA is a linear graph embedding method. The clustering result is then obtained using the K-means algorithm.
  • SC [2]: Spectral clustering consists of an anchor graph based on set-to-set distances derived from local covariance matrix representation.
  • GraphEncoder [13]: A method for learning graph representations. This method explores non-linear graph embedding using deep learning and obtains clustering results by running the K-means algorithm on the embedding.
  • DCN (deep clustering network) [28]: The DCN performs joint dimensionality reduction and clustering. Reduction is achieved by training a deep neural information network to approximate any non-linear function.

4.3. Evaluation Criteria

In this paper, we evaluate the performance of the aforementioned clustering algorithms using the accuracy (ACC) and normalized mutual information (NMI) of the clustering results. ACC evaluates the performance of the algorithm by comparing the clustering labels it produces with the actual clustering labels, examining the clustering status of all pairs of nodes to determine if they belong to the same cluster. ACC is defined as follows:
A C C = i = 1 n ( c l u s t e r _ l a b e l i = = t r u t h _ l a b e l i ) n ,
where n represents the number of nodes, c l u s t e r _ l a b e l i represents the clustering label result of the i-th node by the algorithm, and t r u t h _ l a b e l i represents the real clustering label of the i-th node.
NMI is the mutual information entropy between the algorithm’s clustering labels and the actual clustering labels. Mutual information entropy considers partitions as the probability distribution of nodes falling into a cluster. It is defined as follows:
N M I ( π a , π b ) = i = 1 K a i = 1 K b n i , j l o g ( n n ˙ i , j n i a n j b ) ( i = 1 K a n i a l o g ( n i a n ) ) ( j = 1 K b n j b l o g ( n j b n ) ) ,
where π a and π b represent two types of clustering partitioning, n i , j represents the number of nodes belonging to both cluster C i a and cluster C j b , n i a indicates the number of nodes belonging to the i-th cluster C i a , and n j b indicates the number of nodes belonging to the j-th cluster C j b . NMI’s value ranges from 0 to 1, with a value of 1 being given when the results of a and b are the same.

4.4. Parameter Settings

In our experiments, each sample from all datasets was normalized to possess the same unit length. We set the neighborhood number in the sparse KNN graph search to five. We detected the relationship between parameters p (the number of landmark points) and m (depth level lay) and performance indicators ACC and NMI on real small- and medium-sized data from various sources.
Experiments were conducted on the S E G _ S C model, setting the landmark number parameter p from 200 to 1200, increasing in intervals of 100, with the derived ACC results as shown in Figure 4. From Figure 4, it can be seen that the ACC value increases with the change in landmark value, which is attributed to the random selection method. Evidently, having more landmarks can enhance the final ACC performance; however, the change was not found to be significant in our study. Figure 5 illustrates the relationship between the number of landmarks and NMI. The NMI value generally increases as the number of landmarks increases but with limited magnitude.
Examining the settings of the depth structure layer (lay), the ACC value of the first layer of S G E _ S C on the dataset in Figure 6 was obtained by directly applying K-means to the input normalized sparse similarity graph. It can be observed from Figure 6 that as the layers become deeper, the ACC value increases. This suggests that deep structures play a crucial role in generating high-quality clustering results. In Figure 7, as the number of layers increases, the NMI values for large-scale datasets such as blogdata and 20 newgroup increase.
In summary, the results presented in this paper suggest using five layers for medium- to large-scale datasets; small-scale datasets, in contrast, do not require more than three layers. The information presented in Table 2 shows the number of nodes in each layer. Moreover, without the loss of generality, in this study we set the number of landmarks to 1000.
All neural networks use the sigmoid function as the activation function for each layer, and the sparse target (the target value for hidden layer activation) is 0.05. Specific parameter information is shown in Table 3.

4.5. Algorithm Comparison

Table 4 presents the average running time of the S E G _ S C algorithm in comparison with similar deep neural network models such as Graphencoder, indicating that their running times are approximately equal. This result meets the purpose of sparse embedding representation time optimization.
In terms of cluster performance, Table 5 summarizes the clustering results of seven datasets, indicating that S E G _ S C offers relatively higher ACC. Table 6 displays the comparison values of various algorithms on large-scale data, demonstrating stronger NMI performance advantages as the data scale increases.
From the results shown in Table 5 and Table 6, it can be seen that the model proposed in this paper exhibits superior performance compared to the other models. As shown in Table 5, S E G _ S C achieved higher ACC. This improvement was significant compared to the other algorithms, attributed to the optimization of graph similarity. In addition, as the graph uses a non-linear projection, it can be observed from Table 6 that in the various types of large-scale datasets, the NMI values tend to be optimal.
The advantage of this model lies in using sparse similarity graphs to replace the original graph’s similarity matrix and adopting non-linear encoding schemes instead of feature decomposition. During the final step in particular, the low-dimensional features inputted into K-means are complemented by spectral clustering. Furthermore, the non-linear encoding scheme reduces the dimensionality of the data.

5. Conclusions and Future Research

In the present paper, a spectral clustering model based on deep neural networks named S E G _ S C is proposed. First, the model utilizes a sparse similarity graph based on landmark points and the KNN method, introducing a deep graph representation that encodes each vertex into a low-dimensional vector representation. Secondly, the model theoretically demonstrates the effectiveness of S E G _ S C . Experiments involving the use of real data demonstrate that spectral clustering using deep learning can enhance clustering performance and reduce time consumption in various types of large-scale datasets of information networks.
In future research, we plan to extend this model from general information networks to multi-dimensional information networks, enhancing its scalability on large-scale datasets. In addition, we plan to strengthen its impact by including case studies or real-world applications where the proposed method has been implemented. Conducting such studies will demonstrate its practical viability.

Author Contributions

L.N.: Conceptualization, methodology, software, investigation, formal analysis, writing—original draft preparation, writing—review and editing, funding acquisition; P.M.: visualization, validation, data curation, supervision, funding acquisition; W.Q.: investigation, formal analysis, resources. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China (Grant 2017YFB0202901, 2017YFB0202905), the Hunan Natural Science Foundation Project (No. 2023JJ50353).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cai, Z.; Li, R.; Wu, H. Learning unified anchor graph based on affinity relationships with strong consensus for multi-view spectral clustering. Multimed. Syst. 2023, 29, 261–273. [Google Scholar] [CrossRef]
  2. Qin, Y.; Quan, S.; Wei, C.; Ni, W.; Li, K.; Dong, X.; Ye, Y. Spectral clustering with anchor graph based on set-to-set distances for large-scale hyperspectral images. Int. J. Remote Sens. 2022, 43, 2438–2460. [Google Scholar] [CrossRef]
  3. Luo, X.; He, X.; Yang, X. Fast Spectral Clustering Based on Anchor Point Extraction with Bisecting k-means. Comput. Eng. Appl. 2023, 59, 74–81. [Google Scholar]
  4. Shaham, U.; Stanton, K.; Li, H.; Nadler, B.; Basri, R.; Kluger, Y. SpectralNet: Spectral Clustering using Deep Neural Networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–20. [Google Scholar]
  5. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the International Conference on Neural Information Processing Systems: Natural and Synthetic, Cambridge, MA, USA, 3–8 December 2001; pp. 849–856. [Google Scholar]
  6. Donoho, D.L.; Elad, M. Optimally Sparse Representation in General (Nonorthogonal) Dictionaries via Minimization. Proc. Natl. Acad. Sci. USA 2003, 100, 2197–2202. [Google Scholar] [CrossRef] [PubMed]
  7. Tenenbaum, J.B.; De, S.V.; Langford, J.C. A global geometric framework for nonlinear dimensionality reduction. Science 2000, 29, 2319–2323. [Google Scholar] [CrossRef] [PubMed]
  8. Mcivor, R.T.; Humphreys, P.K. A case-based reasoning approach to the make or buy decision. Integr. Manuf. Syst. 2000, 11, 295–310. [Google Scholar] [CrossRef]
  9. Cai, D.; Chen, X. Large Scale Spectral Clustering Via Landmark-Based Sparse Representation. IEEE Trans. Cybern. 2015, 45, 1669–1680. [Google Scholar] [PubMed]
  10. Zhang, Y.; Zhou, Z.H. Multilabel dimensionality reduction via dependence maximization. ACM Trans. Knowl. Discov. Data 2010, 4, 1–21. [Google Scholar] [CrossRef]
  11. Hou, J.; Nayak, R. Robust clustering of multi-type relational data via a heterogeneous manifold ensemble. In Proceedings of the IEEE International Conference on Data Engineering, Seoul, Republic of Korea, 13–17 April 2015; pp. 615–626. [Google Scholar]
  12. Golub, G.H.; Reinsch, C. Singular value decomposition and least squares solutions. Numer. Math. 1970, 14, 403–420. [Google Scholar] [CrossRef]
  13. Tian, F.; Gao, B.; Cui, Q.; Chen, E.; Liu, T.Y. Learning deep representations for graph clustering. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; pp. 1293–1299. [Google Scholar]
  14. Cao, S.; Lu, W.; Xu, Q. GraRep: Learning Graph Representations with Global Structural Information. In Proceedings of the ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 18–23 July 2015; pp. 891–900. [Google Scholar]
  15. Gupta, A.; Katarya, R. Deep embedding for mental health content on socialmedia using vector spacemodel with feature clusters. Concurr. Comput. Pract. Exp. 2022, 34, 1–21. [Google Scholar] [CrossRef]
  16. Chen, M.; Xu, Z.; Weinberger, K.; Fei, S. Marginalized Denoising Autoencoders for Domain Adaptation. In Proceedings of the 29th International Coference on International Conference on Machine Learning, Edinburgh, UK, 26 June–1 July 2012; pp. 1627–1634. [Google Scholar]
  17. Shao, M.; Li, S.; Ding, Z.; Fu, Y. Deep Linear Coding for Fast Graph Clustering. In Proceedings of the Twenty-Fourth IJCAI Conference, Buenos Aires, Argentina, 25–31 July 2015; pp. 3798–3804. [Google Scholar]
  18. Cao, S.; Lu, W.; Xu, Q. Deep neural networks for learning graph representations. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1145–1152. [Google Scholar]
  19. Shi, J.; Malik, J.M. Normalized Cuts and Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar]
  20. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  21. Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural Language Processing (almost) from Scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
  22. Chen, M.; Weinberger, K.; Sha, F.; Bengio, Y. Marginalized denoising auto-encoders for nonlinear representations. In Proceedings of the International Conference on International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 1476–1484. [Google Scholar]
  23. Kang, Z.; Peng, C.; Cheng, Q. Twin Learning for Similarity and Clustering: A Unified Kernel Approach. In Proceedings of the International Conference on Association for the Advancement of Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2080–2086. [Google Scholar]
  24. Hinton, G.E.; Zemel, R.S. Autoencoders, minimum description length and Helmholtz free energy. In Proceedings of the International Conference on Neural Information Processing Systems, Denver, CO, USA, 29 November–2 December 1993; Morgan: San Francisco, CA, USA, 1993; pp. 3–10. [Google Scholar]
  25. Bourlard, H.; Kamp, Y. Auto-association by multilayer perceptrons and singular value decomposition. Biol. Cybern. 1988, 59, 291–294. [Google Scholar] [CrossRef] [PubMed]
  26. Altman, N.S. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. Am. Stat. 1992, 46, 175–185. [Google Scholar] [CrossRef]
  27. Jones, M.C. Applied Nonparametric Regression. J. R. Stat. Soc. Ser. C Appl. Stat. 1992, 41, 431–432. [Google Scholar] [CrossRef]
  28. Yang, B.; Fu, X.; Sidiropoulos, N.D.; Hong, M. Towards k-means-friendly spaces: Simultaneous deep learning and clustering. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3861–3870. [Google Scholar]
Figure 1. Structural diagram of the spectral clustering algorithm.
Figure 1. Structural diagram of the spectral clustering algorithm.
Applsci 14 04946 g001
Figure 2. Multi-layer deep architecture.
Figure 2. Multi-layer deep architecture.
Applsci 14 04946 g002
Figure 3. Neuron structure.
Figure 3. Neuron structure.
Applsci 14 04946 g003
Figure 4. Relationship graph between ACC and the number of landmarks (Parameter p) on the dataset.
Figure 4. Relationship graph between ACC and the number of landmarks (Parameter p) on the dataset.
Applsci 14 04946 g004
Figure 5. Relationship graph between NMI and the number of landmarks (parameter p) on the dataset.
Figure 5. Relationship graph between NMI and the number of landmarks (parameter p) on the dataset.
Applsci 14 04946 g005
Figure 6. Relationship graph between ACC and depth level lay (parameter m) on the dataset.
Figure 6. Relationship graph between ACC and depth level lay (parameter m) on the dataset.
Applsci 14 04946 g006
Figure 7. Relationship graph between NMI and depth level lay (parameter m) on the dataset.
Figure 7. Relationship graph between NMI and depth level lay (parameter m) on the dataset.
Applsci 14 04946 g007
Table 1. S E G _ S C experimental data information.
Table 1. S E G _ S C experimental data information.
DatasetNode CountFeature CountCluster Count
Wine178133
Pendigit34981610
Mnist400078410
Blogdata10,312139
20 newgroups18,84626,21420
Seimic98,000503
Covtype581,012542
Table 2. Neural information network layer structure.
Table 2. Neural information network layer structure.
DatasetLayer Structure
Wine178-128-64
Pendigit3498-2400-560
Mnist4000-1280-640
Blogdata10,312-5096-2048-1280-640
20 newgroup18,846-5096-2048-1280-640
Seimic98,000-59,048-10,080-5090-2948
Covtype581,012-180,000-98,000-59,000-18,000
Table 3. Experimental parameter table.
Table 3. Experimental parameter table.
ParametersRangeSetpoints
p200, 300, 400, 500, 600, 700, 800, 1000, 12001000
gsigmoid function, tanh functionsigmoid function
m1, 2, 3, 4, 53 or 5
Table 4. S E G _ S C experimental time.
Table 4. S E G _ S C experimental time.
WinePendigitMnistBlogdata20 NewgroupsSeimicCovtype
Graphencoder2250325524620279410,124
S E G _ S C 1022831454861226419942
Table 5. Cluster comparison of ACC results.
Table 5. Cluster comparison of ACC results.
AlgorithmWinePendigitMnistBlogdata20 NewgroupsSeimicCovtype
PCA0.730.540.500.590.590.430.2
SC0.810.620.620.580.60.50.27
Graphencoder0.840.60.380.590.510.470.23
DCN0.830.630.830.640.440.650.21
SEG_SC0.860.650.60.650.620.540.24
Table 6. Cluster comparison of NMI results in large-scale data.
Table 6. Cluster comparison of NMI results in large-scale data.
AlgorithmBlogdata20 NewgroupsSeimicCovtype
Graphencoder0.120.140.150.07
PCA0.130.140.160.07
SC0.130.160.180.08
SEG_SC0.140.160.170.09
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ni, L.; Manman, P.; Qiang, W. A Spectral Clustering Algorithm for Non-Linear Graph Embedding in Information Networks. Appl. Sci. 2024, 14, 4946. https://doi.org/10.3390/app14114946

AMA Style

Ni L, Manman P, Qiang W. A Spectral Clustering Algorithm for Non-Linear Graph Embedding in Information Networks. Applied Sciences. 2024; 14(11):4946. https://doi.org/10.3390/app14114946

Chicago/Turabian Style

Ni, Li, Peng Manman, and Wu Qiang. 2024. "A Spectral Clustering Algorithm for Non-Linear Graph Embedding in Information Networks" Applied Sciences 14, no. 11: 4946. https://doi.org/10.3390/app14114946

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop