Next Article in Journal
Recognition and Implementation of Contact Simple Map Germs from (ℂ2, 0) → (ℂ2, 0)
Previous Article in Journal
Robust (Q,S,R)-γ-Dissipative and H2 Performances for Switched Systems with Mixed Time Delays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Community Detection in Multilayer Networks Based on Matrix Factorization and Spectral Embedding Method

1
School of Mathematics Sciences, Huaibei Normal University, Huaibei 235000, China
2
School of Mathematics and Statistics, Lanzhou University, Lanzhou 730033, China
3
School of Mathematics and Statistics, Xuzhou University of Technology, Xuzhou 221018, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(7), 1573; https://doi.org/10.3390/math11071573
Submission received: 28 February 2023 / Revised: 10 March 2023 / Accepted: 22 March 2023 / Published: 23 March 2023

Abstract

:
Community detection remains a challenging research hotspot in network analysis. With the complexity of the network data structures increasing, multilayer networks, in which entities interact through multiple types of connections, prove to be effective in describing complex networks. The layers in a multilayer network may not share a common community structure. In this paper, we propose a joint method based on matrix factorization and spectral embedding to recover the groups not only for the layers but also for nodes. Specifically, the layers are grouped via the matrix factorization method with layer similarity-based regularization in the perspective of a mixture multilayer stochastic block model, and then the node communities within a layer group are revealed by clustering a combination of the spectral embedding derived from the adjacency matrices and the shared approximation matrix. Numerical studies show that the proposed method achieves competitive clustering results as the number of nodes and/or number of layers vary, together with different topologies of network layers. Additionally, we apply the proposed method on two real-world multilayer networks and obtain interesting findings which again highlight the effectiveness of our method.

1. Introduction

Network analysis has emerged in many areas of research and applications, from sociology and ecology to economics [1,2]. For networks, nodes represent the entities of interest, and connecting edges represent the specific relationship that exists among nodes. As the complexity of real datasets increases, various interactions among the elementary components of these systems have emerged. For example, people may contact each other as workmates in a company and as friends on Facebook; different countries may trade in a variety of products. Traditional monolayer networks, which record only one type of interaction between nodes, are less capable to describe the multiple modes of interactions encountered and to understand the topology of the real datasets precisely. Fortunately, several network models have been in existence as various extensions of monolayer networks. One such example are the multilayer networks in which multiple modes of interactions are considered and each specific interaction forms a layer [3]. In this paper, we focus on the multilayer networks with layers sharing a common set of nodes and without inter-layer edges [4].
Community detection is a fundamental issue when analyzing network data. For monolayer network, community detection aims to cluster nodes into communities, and thus there exist more edges connecting nodes in the same community and fewer edges joining nodes across communities [5]. The community detection problem has significant applications due to the fact that entities in a community are likely to share common properties; for example, people in a community may come from the same place and/or they may share other similar interests [6]. Thus, revealing the community structure in a network can provide a better understanding of the overall functioning of the network. There exists a substantial volume of research about community detection techniques on monolayers, such as probabilistic methods [7,8] and criteria-based approaches [9,10]. For a comprehensive review of these works we refer to [11].
The community detection methods originally proposed for monolayer networks can be used directly to analyze multilayer networks in two ways. One way is to aggregate the layers into a graph and then apply traditional community detection methods on the aggregated network [12,13]. The other way is to use the traditional community detection methods on each layer independently and the clustering results are then merged [14,15]. Several other researchers have also proposed community detection methods for multilayer networks, including probability models [16,17,18,19], spectral methods based on various versions of aggregation adjacency matrices or Laplacian matrices [20,21], matrix factorization methods [22] and criteria based methods such as minimizing a squared objective function or maximizing a modularity function [23,24,25]. Most of these methods aim to detect communities that are common across layers and ignore the inhomogeneous nature of network layers. However, for many real multilayer networks, different layers may have different community structures, and some of the layers are interdependent while others are not. For instance, in an international trade multilayer network in which products represent layers and countries are nodes, the trade patterns among countries related to processed foods may be different from those related to unprocessed foods; on the other hand, the trade activities among countries related to different processed foods may be similar. These findings suggest that we should jointly find groups for layers and assign community memberships to nodes for each layer group.
A few recent works have focused on addressing the community detection problem in inhomogeneous multilayer networks. For example, Ref. [26] introduce a multilayer SBM which clusters layers into groups, the so-called strata, and assume that layers in each strata share a common community structure; Ref. [27] introduce a common subspace independent-edge multiple random graph model allowing a shared latent structure across layers, being layer-specific in each layer, and propose a joint spectral embedding of adjacency matrices to simultaneously estimate underlying parameters for each layer; Ref. [28] investigate the so called “Mixture MultiLayer Stochastic Block Model” (MMLSBM) and propose a regularized tensor decomposition to reveal the communities for the layers and nodes. Ref. [29] propose an alternating minimization algorithm(ALMA) that simultaneously recovers the layer partitions and node partitions. Although the methods for finding layer groups in the mentioned works are a bit different, given the layer groups, these works apply classical cluster methods either on the aggregated adjacency matrices or an estimated matrix of that layer group to find the node assignments. More related works we refer to are [30,31,32].
In this work, we are interested in inhomogeneous multilayer networks under the mixture multilayer stochastic block model (MMLSBM) setting. We consider a two-step procedure to detect the layer groups (between-layer clustering) as well as the node communities within a layer group (within-layer clustering). Unlike commonly used methods that describe the multilayer networks using tensor-wise data in the literature, we instead use the vectorization of the model. This technique can reduce the systematic errors due to the improper setting of parameters in the estimation of the tensor approximations. A combination of the Matrix Factorization and Spectral Embedding method (MFSE) is proposed to handle the joint community detection problem. The present paper makes the following contributions. First, under a vectorized perspective, we propose a matrix factorization based method to recover the layer communities. Moreover, since layers that are very structurally similar are more likely to belong to the same group, we integrate layer similarity regularization into the matrix factorization based objective function. Under the MMLSBM setting, each layer can be described as a stochastic block model (SBM) model and two layers have the same connection probability matrix once they belong to the same layer community. From the least squares perspective, the underlying membership matrix C of the layers is optimized by minimizing the objective function A C Q F 2 + P e n , where A is a vectorized version of the adjacency tensor and the l-th row of matrix Q can be treated as the potential representation of the l-th layer community and P e n is the regularization term. Subsequently, we employ the spectral embedding method to recover the node communities in each layer group. For a given layer group, we simultaneously apply the spectral embedding on the adjacency matrices and on the shared structural matrix derived from Q in that group. The final node latent representations are a combination of two types of spectral embeddings. Then, classical clustering methods can be carried out on the joint spectral embedding to search the node assignments. In addition, an efficient singular value decomposition is used to give a recommended value of the number of layer groups as well as the number of node communities in each layer group.
The rest of the paper is organized as follows. The framework is illustrated in Section 2. Section 3 gives the theoretical results of the proposed method. Section 4 provides several experimental evaluations to demonstrate the effectiveness of the proposed method and Section 5 summarizes the paper.

2. Model Framework

2.1. Notations

For a matrix A, let A F be the Frobenius norm and denote its transpose by A T . Let A ( i ; · ) and A ( · ; j ) be the i-th row and j-th column of a matrix A, respectively. Denote the Kronecker product of the matrices of matrices A and B by A B . Some more notations used throughout this paper are collected in Table 1.

2.2. Discussion of the Framework

We consider a multilayer network G = { G ( 1 ) , , G ( L ) } with L layers on the same set of n nodes. For the l-th layer, denote its adjacency matrix by A ( l ) R n × n , where A i j ( l ) = 1 if the nodes i and j are connected and A ( l ) ( i , j ) = 0 otherwise, further, A i j ( l ) = A j i ( l ) if the l-th layer is undirected and A i j ( l ) A j i ( l ) otherwise.
In the MMLSBM setting, A i j ( l ) are independent Bernoulli random variables with P ( A i j ( l ) = 1 ) = P i j ( l ) and P i i ( l ) = 0 . The probability matrices P ( l ) give a tensor form P R n × n × L . In this paper, we assume that the probability connection tensor P can be partitioned into M ( M < L ) groups. In particular, there exists a label function c : [ L ] [ M ] and a corresponding clustering matrix C { 0 , 1 } L × M such that
E [ A ( l ) | Z ] = P ( l ) = Z ( m ) B ( m ) ( Z ( m ) ) T , m = c ( l ) , 1 l L , 1 m M ,
where B ( m ) [ 0 , 1 ] K m × K m is the connectivity matrix satisfying B ( m ) ( i , j ) = B ( m ) ( j , i ) . We aim to recover the network groups G 1 , , G M as well as the within-layer community assignments { V k ( m ) ; k = 1 , , K m , m = 1 , , M } .

2.2.1. Between-Layer Cluster

Methods

Adopting the vectorized versions of P ( l ) and B ( l ) , it follows from [33] that
p ( l ) = ( Z ( c ( l ) ) Z ( c ( l ) ) ) b ( c ( l ) ) , l = 1 , , L .
Denote Θ ( c ( l ) ) = Z ( c ( l ) ) Z ( c ( l ) ) and recall that C is the membership matrix of the layers, i.e., for l = 1 , , L ; m = 1 , , M , C ( l , m ) = 1 if c ( l ) = m and 0 otherwise. We thus rewrite Equation (1) as
P = p ( 1 ) p ( L ) = Θ ( c ( 1 ) ) b ( c ( 1 ) ) Θ ( c ( L ) ) b ( c ( L ) ) = C Q
where Q = ( Θ ( 1 ) b ( 1 ) , , Θ ( M ) b ( M ) ) T R M × n 2 can be considered as the shared structure matrix of M layer groups.
Similarly, we introduce the matrix A = ( a ( 1 ) , a ( 2 ) , , a ( L ) ) T R L × n 2 . Note that P can not be observed directly but A can be; combining Equations (1)–(3) and from the perspective of the least-squares method, we propose to estimate the potential layer cluster matrix via the following objective function
min C C T = I ; Q A C Q F 2 .
In the estimation step, we relax the restrictions on C, i.e., C ( l , m ) = 1 if layer l belongs to group m, and impose the non-negativity together with C C T = I on C. For the non-negative feature of the matrices Θ ( l ) and b ( l ) , we can solve the optimization problem of Equation (4) by a nonnegative matrix factorization method.
Intuitively, the layers with high pairwise structural similarity tend to be partitioned into a group. Some effort has been devoted to designing various kinds of indices to measure the similarity between layers [34,35]. In this paper, we use the Pearson linear correlation coefficient S ( l , m ) to quantify the degree of correlation between two layers l and m [36]. The Pearson correlation coefficient is defined as
S ( l , m ) = i j ( A i j ( m ) A ( m ) ¯ ) ( A i j ( l ) A ( l ) ¯ ) σ ( A ( m ) ¯ ) σ ( A ( l ) ¯ )
where A ( · ) ¯ represents the mean degree and σ ( A ( · ) ¯ ) is the standard deviation of the layer A ( · ) . There is no doubt that S ( l , m ) [ 1 , 1 ] . The closer S ( l , m ) is to 1 ( 1 ), the more positively (negatively) correlated the two layers are. Additionally, layers l and m are almost linearly independent if S ( l , m ) closes to 0.
Above all, a layer similarity based regularization is defined as
S ( l , m ) C ( l ; · ) C ( m ; · ) F 2
where S ( l , m ) is the Pearson linear correlation and C ( l ; · ) represents the l-th row of matrix C. A simple calculation leads to the following matrix format of Equation (5)
1 2 l m S ( l , m ) C ( l ; · ) C ( m ; · ) F 2 = t r ( C T L S C )
where t r ( · ) is the trace function, S = ( S ( l , m ) ) is the layer similarity matrix and L S = D S S stands for the Laplacian matrix of the layer similarity matrix. D S is a diagonal matrix representing the degree matrix of S.
By integrating the regularization Equation (6) into Equation (4), we now formalize the between-layer cluster problem as a non-negative matrix factorization problem with layer similarity regularization:
min C C T = I , Q L ( C , Q ) = min C C T = I , Q A C Q F 2 + α t r ( C T L S C )
where α is a trade-off parameter for the regularization term. In this paper, for l m , we set the entities S ( l , m ) which are less than 0 to 0, and choose the parameter α in Equation (7) by the cross validation method.

Updating Rules

Note that L ( C , Q ) is not convex with respect to C and Q together; it is thus difficult to find the global minimum of the loss function. To this end, an iterative algorithm is adopted to pursue the solution of Equation (7). In addition, we introduce Lagrange multipliers Γ R L × M and Ξ R M × n 2 for the non-negative constraints on C and Q, respectively. Then the objective function Equation (7) is rewritten as
min C C T = I , Q L ( C , Q ) = min C C T = I , Q A C Q F 2 + α t r ( C T L S C ) + t r ( Γ C T ) + t r ( Ξ Q T ) .
The partial derivatives of Equation (8) with respect to C and Q are
L ( C , Q ) C = 2 C Q Q T 2 A Q T + 2 α L S C + Γ .
and
L ( C , Q ) Q = 2 C T C Q 2 C T A + Ξ .
A local minimum of Equation (8) is achieved by respectively setting
L ( C , Q ) C = 0
and
L ( C , Q ) Q = 0 .
After that, for l { 1 , 2 , , L } , m { 1 , 2 , , M } and i { 1 , 2 , , n 2 } , we multiply by C ( l , m ) and Q ( m , i ) on both sides of Equations (11) and (12), respectively. Since the KKT conditions imply that C ( l , m ) Γ ( l , m ) = 0 and Q ( m , i ) Ξ ( m , i ) = 0 , we obtain the optimal solutions of C ( l , m ) and Q ( m , i ) as follows:
C ^ ( l , m ) C ( l , m ) ( A Q T ) ( l , m ) ( C Q Q T + α L S C ) ( l , m )
and
Q ^ ( m , i ) Q ( m , i ) ( C T A ) ( m , i ) ( C T C Q ) ( m , i ) .
Now, we can recover the communities of network layers by applying K-means algorithm on the rows of C ^ . For each m, the estimated Q ^ ( m ; · ) represents the shared structural information of layers in the m-th layer group; we transform Q ^ into M separate matrices. Specifically, for m = 1 , , M , we consider matrices Q ^ m R n × n given by
Q ^ m ( i ; ) = Q ^ ( m ; ( ( i 1 ) n + 1 ) : i n ) .
We stress that Q ^ m ( i ; · ) represents the shared behavior of node i in the m-th layer group.
A toy example is further used to describe the procedure of Equation (7). As shown in Figure 1, the simple multilayer network consists of four layers and five nodes. Each row of the matrix A R 4 × 25 represents the vectorized version of the adjacency matrix related to the layer. Applying the updating rules in Equations (13) and (14), we obtain the estimated layer membership matrix C R 4 × 2 and the estimated shared structure information matrix Q R 2 × 25 . After that, we use K-means on C ^ to obtain the layer groups. Then, by Equation (15), we construct two 5 × 5 matrices Q ^ 1 and Q ^ 2 based on two rows of Q ^ , respectively. Each constructed matrix is treated as the potential representation of the related layer group.

2.2.2. Within-Layer Community Detection

Given the layer membership matrix C ^ and the shared structure matrix Q ^ , the issue that we focus on is how to use the structural information within a layer to detect node assignments. Refs. [28,29] applied a classical clustering method either on the aggregated adjacency matrices or an estimated embedding matrix associated to that layer group. To better describe the consistent topology structural across layers within a layer group, we propose to use an aggregated spectral embedding derived from both the observed adjacency matrices in that layer group and of the shared Q ^ m on the task of clustering nodes within a layer group.
Once the group layers are fixed, we average the adjacency matrices in the m-th group as A ¯ m = l : c ( l ) = m A ( l ) L m . Then, a standard spectral embedding method is applied on A ¯ m . We list the detailed steps below.
(Sa):
Calculate the graph Laplacian L A ¯ m = D A ¯ m 1 / 2 A ¯ m D A ¯ m 1 / 2 .
(Sb):
Find orthonormal eigenvectors U ^ A ¯ m 1 , U ^ A ¯ m 2 , , U ^ A ¯ m K m corresponding to the K m eigenvalues that are largest in absolute value of L A ¯ m and put them in a matrix U ^ A ¯ m = [ U ^ A ¯ m 1 , U ^ A ¯ m 2 , , U ^ A ¯ m K m ] R n × K m .
(Sc):
Form the normalized version U ˜ A ¯ m by normalizing each row of U ^ A ¯ m to have unit length. Then, U ˜ A ¯ m is considered to be the spectral embedding derived from A ¯ m .
Following the steps (Sa)–(Sc) of the spectral embedding, we are coming to find the embedding representations of nodes from Q ^ m . Denote the graph Laplacian by L Q ^ m = D Q ^ m 1 / 2 Q ^ m D Q ^ m 1 / 2 and the corresponding normalized embedding matrix by U ˜ Q ^ m .
Based on the above, for the m-th layer group, we add the node representations obtained in U ˜ A ¯ m with the U ˜ Q ^ m to obtain extended node representations:
Y ^ m = U ˜ A ¯ m + U ˜ Q ^ m 2 .
The within-layer community detection is then carried out by applying a K-means algorithm on Y ^ m . Up to now, we formalize a two-step community detection method for multilayer networks based on Matrix Factorization and joint Spectral Embedding (MFSE) which is outlined as Algorithm 1.
Remark 1
(Initializations of C and Q). It is known that random initializations may lead the algorithms into non-informative local minimals. The Pearson correlation matrix S illustrates the structural similarity among layers and thus it can provide a glancing estimation of the membership of the layers. Here, we apply K-means on the layer similarity matrix S and take the centers returned by the K-means algorithm as the initialization of C. In our proposed objective function, each row of Q stands for the potential representations of nodes in a specific layer group, following the way of initializing C; after carrying out K-means on A , the initialization of Q is given by the centers returned by the K-means algorithm.
Remark 2
(Pre-determination of M and K m ( m = 1 , , M ) ). Much work has been carried out for determining the number of communities in the community detection problem ([37,38]), either by maximizing specific objective function or searching all candidates. Among these, the eigengap heuristic method proves to be effective in applications ([39,40]). In this paper, we select the number of layer groups M and number of node communities K m of the m-th layer group by the eigengap heuristic method. For M, we compute the eigendecomposition of the layer similarity matrix S and then rank the eigenvalues λ 1 ( S ) > λ 2 ( S ) > . In practice, there is usually a significant drop between λ M ( S ) and λ M + 1 ( S ) and we set the number of layer groups as M.
Once the layer groups are fixed, we first obtain the averaged adjacency matrix A ¯ m in the m-th layer group and then apply eigen-decomposition on A ¯ m . After that, K m is chosen as the number of communities if the eigenvalue λ K m ( A ¯ m ) is relatively larger than λ K m + 1 ( A ¯ m ) , i.e., the gap between λ K m ( A ¯ m ) and λ K m + 1 ( A ¯ m ) is much greater than those of other adjacent eigenvalues.
Algorithm 1 MFSE Algorithm
Input: The adjacency matrices A ( l ) , l = 1 , 2 , , L ; number of layer groups M; number of the node groups { K m } m = 1 M ; tuning parameter α .
Output: The network groups { G 1 , , G M } and the within-layer node communities { V k ( m ) ; m = 1 , , M , k = 1 , , K m } .
Algorithm:
1. 
Initialize C and Q;
2. 
Iterate until convergence.
(a2). 
Update C ^ according to Equation (13);
(b2). 
Update Q ^ according to Equation (14);
3. 
Applying K-means on the rows of C ^ to get the layer groups G 1 , , G M .
4. 
For each layer group G m .
(a4). 
Applying the spectral embedding method, i.e., steps (Sa)–(Sc) on A ¯ m to obtain the latent node representation matrix U ˜ A ¯ m ;
(b4). 
Similar to (a4), estimate the node representation matrix U ˜ Q ^ m based on Q ^ m ;
(c4). 
The layer groups { V k ( m ) } are obtained by applying K-means on the rows of the final node embedding Y ^ m in Equation (16).

3. Theoretical Guarantees

Some additional assumptions are required to analyze the convergence issues. We collect them as follows:
(A1)
d Q m > 3 log 8 n ϵ for ϵ > 0 ;
(A2)
24 d Q m 1 / 2 λ K m 3 log 8 n ϵ 1 / 2 with λ K m representing the K m -th largest eigenvalue of L Q m L Q m ;
(A3)
cluster of layers and local communities balance condition: there exist positive constants γ ̲ < γ ¯ such that
γ ̲ M < L L m a x L L m i n < γ ¯ M , γ ̲ K m < n n m a x ( m ) n n m i n ( m ) < γ ¯ K m .
Note that Assumption (A1) restricts the sparsity of the network, Assumption (A2) requires that the K m -th eigenvalue of L Q m L Q m is not too small and Assumption (A3) is the so-called balanced network condition and is very common in the problem of community detection ([41]).
We need to obtain the estimators C ^ and Q ^ of C * R L × M and Q * R M × n 2 based on the observed { A ( l ) } . First of all, we give a theoretical result on the consistency of the estimators from the objective function L ( C , Q ) . We denote the estimators in the iteration by { Δ ^ [ t ] = ( C ^ [ t ] , Q ^ [ t ] ) } t = 1 .
Theorem 1.
{ Δ ^ [ t ] } t = 1 converges to a critical point of L ( C , Q ) .
The proof of Theorem 1 is given in Appendix A.
We denote by π ( m ) the permutation function of [ m ] . Given the true partition of the network layers [ L ] = m = 1 M G m and the estimated partition [ L ] = m = 1 M G ^ m , the misclassfication error rate of between-layer clustering can be defined as
E r r B L = min τ π ( m ) m = 1 M | G m G ^ τ ( m ) | / L .
Theorem 2.
Suppose that Assumption (A3) holds, there exists a sufficiently small function ε ( n , L ) depending on n and L such that the between-layer clustering error satisfies E r r B L ζ ε ε ( n , L ) max m L m 4 L , where ζ ε is a positive constant depending on ε.
The proof of Theorem 2 is given in Appendix B.
For the m-th layer group, the cluster centroids sought by K-means algorithm are defined as O = { O m R n × d : O m h a s K m unique rows } , satisfying O ^ m = arg min X ^ m O Y ^ m O m F . Let
Γ m = i : O ^ m ( i ; · ) Y m * ( i ; · ) Π 2 > O ^ m ( j ; · ) Y m * ( j ; · ) Π 2
be the set of misclustered nodes, where Y m * ( i ; · ) represents the parameter centroid corresponding to the i-th node while O ^ m ( i ; · ) is the estimated centroid, and Π is an orthonormal rotation.
Theorem 3.
Assume that the smallest eigenvalue of E A ¯ m is larger than α n and Assumptions (A1)- (A2) hold. There exist constants C ( ϵ ) > 0 , C 0 > 0 , 0 < ξ 1 and for any ϵ > 0 , with probability at least 1 ϵ ; the proportion of misclustered nodes in the m-th layer group satisfies | Γ m | n ξ 2 144 C 0 K m log ( 8 n / ϵ ) n λ K m 2 + C ( ϵ ) K m log n n α n 2 .
The proof of Theorem 3 is given in Appendix C.

4. Experiments

In this section, we assess the performance of the proposed MFSE method through simulated and real-world multi-layer networks. For synthetic multilayer networks, we compare our method with the TWIST method ([28]) and the ALMA method ([29]). Specifically, TWIST focuses on the MMSBM model and applies Tucker decomposition on the observed adjacency tensor to obtain its low-rank approximations. TWIST finds both the global and the local memberships of nodes, together with the memberships of layers. ALMA is an alternating minimization algorithm which aims at revealing the layer partition as well as the node partitions. We remark that both MFSE and ALMA aim at recovering the community assignment of layers together with the nodes. To make the comparison fair, all three competitive methods are only dedicated to detecting the layer partition and the node partition which are also named as “between-layer clustering” and “within-layer clustering”, respectively. For both synthetic multilayer networks and real-word networks, the between-layer clustering error is assessed by E r r B L (Equation (17)) and the proportion of misclassified nodes within layers (the within-layer clustering error) is evaluated by m = 1 M Γ m M n (Equation (18)).

4.1. Simulation Study

In this part, we test our method on synthetic multilayer networks. We carry out three simulations to evaluate the performance of the proposed method by varying the number of layers L, the number of nodes n and network sparsity parameter γ , respectively. The clustering results of all methods in three simulations averaged over 500 replications. Below, we elaborate on each of the simulation scenarios.
The purpose of the first simulation is to examine the ability of the proposed method to discover both layer and node clusters as the number of nodes n varies. Imagine a MMLSBM with L = 60 layers, M = 3 layer groups and K 1 = K 2 = K = 2 node communities for each layer group. The underlying community assignments of layers and nodes are generated by the multinomial distributions with equal class probabilities 1 / M for the layers, and 1 / K for the nodes within a layer group. For completeness, we take into account both assortative and disassortative SBM models ([42]) in the tests. A SBM model is assortative if B ( i , j ) < B ( i , i ) and disassortative if B ( i , j ) > B ( i , i ) for i j , with B ( i , j ) representing the connection probability ([41]). We set the connection matrices for three layer groups as
B m = p m γ m p m γ m p m γ m p m p m γ m p m γ m p m γ m p m p m , m = 1 , 2 , 3
where p = ( p 1 , p 2 , p 3 ) = ( 0.4 , 0.4 , 0.5 ) and γ = ( γ 1 , γ 2 , γ 3 ) = ( 1.2 , 0.8 , 0.6 ) . Notice that values of γ m govern the strength of the assortative (disassortative) feature of the network. Roughly speaking, the network is assortative if γ m < 1 and disassortative if γ m > 1 , and the community structure becomes obscure as γ m  closes to 1. We increase the number of nodes n from 40 to 400 and record the between-layer clustering error results and within-layer clustering error results of three methods in Figure 2a,b. As expected, both types of errors rates for all three methods decrease substantially as n becomes large, implying the effect of number of nodes on clustering errors. As n 80 , the plots demonstrate that MFSE always achieves superior or comparative performances in both between-layer and within-layer clustering.
In the second simulation, we evaluate the performances of three methods as the number of layers varies. The generated MMLSBM is similar to that in the first simulation except n = 100 is fixed. We set the number of layers L to increase from 40 to 200. The clustering errors are reported in Figure 2c,d. For all methods, both types of errors decrease as L increases, which is consistent with the theoretical results of these methods. Specifically, for between-layer clustering, MFSE outperforms TWIST and ALMA once L 80 , and MFSE continues to obtain superior results for the within-layer clustering.
The last simulation studies the impact of parameter γ which controls the ratio of the connection probabilities of the inter-groups of nodes versus intra-groups of nodes. The settings are the same as previous simulation scenarios except n = 100 and L = 120 are fixed. We set γ = { γ 1 , γ 1 , γ 1 } , and the performances of all methods on two types of clustering with γ 1 increasing from 0.1 to 1.9 are investigated. The results of this comparison are presented in Figure 2e,f. Notice that the smaller the γ 1 , the more pronounced the within-group and between-group distinction becomes. Additionally, the larger γ 1 (>1) indicates the stronger disassortative feature of the network. Further, the within-layer clustering becomes harder when γ 1 closes to 1, which also makes the between-layer clustering difficult. The performances of the three methods in Figure 2e,f agree with those findings.
In summary, for between-layer clustering, the proposed method MFSE achieves better results in all parameter settings and ALMA almost gives the second best results. For within-layer clustering, MFSE still gives superior or comparable performances in different settings. We claim that TWIST works notably even in the cases where the between-layer errors are higher than those of ALAM. This could be due to the fact that TWIST recovers the node partitions by clustering the adjacency matrices in that layer group, while ALAM clusters the low-rank approximation of the adjacency tensor in that layer group which may leave out some direct structure information. We use both direct adjacency matrices and the low-rank approximations to reveal the node memberships within a layer group, which makes our method more effective in various model settings.

4.2. Real-World Network Data

In this section, we demonstrate the merits of MFSE in community detection on two real-world multilayer networks.

4.2.1. FAO data

The data are collected from the Worldwide Food Trading Networks data that are available at https://www.fao.org/faostat/en/#data/TM. The data was collected since 1986 and the last update was given in 23 December 2022. In our analysis, the data was accessed on 1 January 2010, before 31 December 2010. The dataset contains export/import trading among 245 countries/regions for more than 300 products. In the multilayer network, layers represent different types of commodities and nodes represent countries. Two countries have a link if there exist import/export relationships of a product between them. We convert the original directed and weighted networks to undirected and unweighted ones by ignoring the directions and weights. We first remove the layers with links less than 1135 (the third quartile) and retain the layers with average degree greater than 10 (the second quartile). Finally, we choose the countries that are active in trading at least 90% of the selected products, reducing a multilayer network consisting of 46 layers and 100 nodes.
Subsequently, we use Algorithm 1 to recover the partitions of layers (products) and nodes (countries) of the multilayer network. By applying the eigengap heuristic method, we choose M = 2 and record the between-layer clustering results in Table 2.
From Table 2, we see that the first layer group consists mostly of solid and liquid beverage products as well as processed food products with a certain shelf life, and the second group contains mostly cereals and raw materials for food products, together with processed food products with a long shelf life.
Applying the eigengap heuristic method again, we choose K 1 = K 2 = 5 , which is consistent with the number of continents (Africa, Americas, Asia, Europe, and Oceania) and can be considered as a natural partition of countries. Figure 3 shows the node communities recovered by MFSE. We have the following comments: (1) for both layer groups, there exist close trading activities among some developed and developing countries in Asia and Oceania, for example, China, Japan, Australia, New Zealand, etc, see Community 1 in Figure 3a,b, respectively; (2) for the products in group 1, since North America and Europe are not far apart by ocean transportation, countries in these two regions have more continual demands for trade than the countries in the Americas which have a longer shipping distance, please refer to Communities 2 and 3 in Figure 3a; (3) for the products in group 2, the considerable shelf life of some products allows for close trade among countries with long marine transportation distance, such as among countries in North America and South America (Community 3 in Figure 3b); (4) the countries in Southeast Europe (Community 4 in Figure 3a,b, respectively) have their independent trade patterns regardless of the types of products, probably due to the Free Trade Agreement which was carried out in 2006.

4.2.2. AU-CS Network

The network is constructed from a dataset that describes five online and offline relationships including Facebook, leisure, work, coauthor and lunch among 61 employees of a university research department ([43]). The employees are represented by nodes and one specific relationship is represented by one layer, together with links standing for an existing relation between nodes. Thus, a five-layer multilayer network with 620 edges is formed. As a pre-processing step, we computed the eigenvalues of the graph Laplacian based on the layer-similarity matrix and chose M = 2 by eigengap heuristic. After that, by using the eigengap heuristic method again, we chose K 1 = K 2 = 3 as the numbers of node communities in two layer groups, respectively. We apply Algorithm 1 on the data and list the between-layer clustering result in Table 3. The intuition of these layer partitions is that the employees have a natural co-occurrence of work and lunch interactions as colleagues; for example, they may work together and have lunch together in the canteen. Simultaneously, it is natural to group together the online/offline relationships based on personal/research interests.
Then, we investigate the within-layer clustering results. Unlike the FAO data, we may not explain intuitively the node partitions and have no idea about the groundtruth communities. For completeness, we use the density metric to evaluate the quality of the clustering method. The density at the l-th network layer is defined as follows:
d e n s i t y = k = 1 K l | { ( v i ( l ) , v j ( l ) ) | v i ( l ) , v j ( l ) C k ( l ) , ( v i ( l ) , v j ( l ) ) E ( l ) } | | E ( l ) |
where, for the l-th layer, v i ( l ) , C k ( l ) and E ( l ) represent the i-th node, the k-th node community and the edge set, respectively. Additionally, to make the results comparable, we introduce a standard Spectral Clustering based method (SPC-Mean) in the literature which uses the Mean adjacency matrix A ¯ = l = 1 L A ( l ) L to recover the node communities. We record the densities of each layer returned by the proposed MFSE method and SPC-Mean method in Table 4. We further plot the adjacency matrix of each layer returned by MFSE (Figure 4) where nodes are arranged according to the community labels. Table 4 shows that MFSE gives highly competitive densities for all layers on the AU-CS network and Figure 4 implies that the connection patterns of the first three layers are significantly different from those of the remaining two layers.

5. Conclusions

In this paper, we propose a matrix factorization and spectral embedding based community detection method for multilayer networks. The proposed method recovers the layer groups by clustering the low-rank approximate of a vectorized version matrix derived by the adjacency matrices. After that, the node communities within a layer group can be recovered by both the observed adjacency matrices and a shared structural matrix in that layer group. We give the theoretical guarantees of the method and the simulation results confirm the fact that the proposed method performs efficiently under different model settings. Notice that the layers of networks may have various degrees of sparsity in many real applications; it is thus of interest to extent our method to more sparse scenarios in future studies. Another direction worth exploring is where one can generalize the matrix factorization-based representation incorporating node attributes to extract communities in multilayer networks.

Author Contributions

Conceptualization, F.T.; methodology, F.T.; software, X.Z.; validation, X.Z.; formal analysis, X.Z.; investigation, F.T.; resources, C.L.; data curation, C.L.; writing—original draft preparation, C.L.; writing—review and editing, C.L.; visualization, C.L.; supervision, C.L.; project administration, F.T.; funding acquisition, F.T. and X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (12201235, 11971214), Natural Science Foundation of Anhui Province (2108085QA14), the Excellent Young Talents Fund Program of Higher Education Institutions of Anhui Province (gxyqZD2022044) and Qinglan Project of Jiangsu Province.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editor, associate editor and referees.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MFSEMatrix factorization and spectral embedding method
MMLSBMMixture multilayer stochastic block model
TWISTTucker decomposition with integrated SVD transformation
ALMAAlternating minimization algorithm

Appendix A

The proof of Theorem 1 is based on the following lemmas and the Kurdyka–Lojasiewicz property.
Lemma A1.
There exists a constant ι > 0 such that for t = 1 , 2 , , Δ ^ [ t ] F 2 ι .
Proof. 
Since L ( C , Q ) 0 and the sequence of of { Δ [ t ] } t = 1 is monotone decreasing, the result is obtained. □
Lemma A2.
For each t = 1 , 2 , , it holds that ι 1 Δ ^ [ t + 1 ] Δ ^ [ t ] F 2 ( L ( Δ ^ [ t ] ) L ( Δ ^ [ t + 1 ] ) ) , where ι 1 = 2 min { α 2 ( λ 1 ( L S ) ) 2 , 1 } and λ 1 ( L S ) is the largest eigenvalue of L S .
Proof. 
Denote C ¨ ^ = v e c ( C ^ [ t ] C ^ [ t + 1 ] ) . For each t = 1 , 2 , , we have
L ( C ^ [ t ] , Q ^ [ t ] ) L ( C ^ [ t + 1 ] , Q ^ [ t ] ) = < L ( C ^ [ t + 1 ] , Q ^ [ t ] ) C ^ , C ¨ ^ > + 2 C ¨ ^ T ( Q ^ [ t ] ) T Q ^ [ t ] I M 2 × M 2 + α L S T I n 2 × n 2 C ¨ ^ F 2 2 C ^ [ t ] C ^ [ t + 1 ] F 2 α 2 ( λ 1 ( L S ) ) 2
where ⊗ is the Kronecker product of two matrices.
Similarly, we have
L ( C ^ [ t ] , Q ^ [ t ] ) L ( C ^ [ t ] , Q ^ [ t + 1 ] ) = < L ( C ^ [ t ] , Q ^ [ t + 1 ] ) Q ^ , Q ^ [ t ] Q ^ [ t + 1 ] > + 2 ( Q ^ [ t ] Q ^ [ t + 1 ] ) T ( Q ^ [ t ] Q ^ [ t + 1 ] ) F 2 2 Q ^ [ t ] Q ^ [ t + 1 ] F 2
Combining Equations (A1) and (A2) yields that ι 1 Δ ^ [ t + 1 ] Δ ^ [ t ] F 2 ( L ( Δ ^ [ t ] ) L ( Δ ^ [ t + 1 ] ) ) .
Lemma A3.
For any t = 1 , 2 , and l [ t + 1 ] L ( Δ ^ [ t + 1 ] ) , it holds that l [ t + 1 ] F ι 2 Δ ^ [ t ] Δ ^ [ t + 1 ] F , where ι 2 = 2 ι 2 + A F .
Proof. 
It follows from Equation (11) that
L ( Δ ^ [ t + 1 ] ) C ^ = 2 C ^ [ t + 1 ] Q ^ [ t + 1 ] ( Q ^ [ t + 1 ] ) T 2 A ( Q ^ [ t + 1 ] ) T + 2 α L S C ^ [ t + 1 ] + Γ .
On the other hand,
L ( C ^ [ t + 1 ] , Q ^ [ t ] ) C ^ = 2 C ^ [ t + 1 ] Q ^ [ t ] ( Q ^ [ t ] ) T 2 A ( Q ^ [ t ] ) T + 2 α L S C ^ [ t + 1 ] + Γ = 0 .
Combining Equations (A3) and (A4), we can get
L ( Δ ^ [ t + 1 ] ) C ^ F 2 C ^ [ t + 1 ] Q ^ [ t + 1 ] ( Q ^ [ t + 1 ] ) T C ^ [ t + 1 ] Q ^ [ t ] ( Q ^ [ t ] ) T F + 2 A ( Q ^ [ t + 1 ] ) T A ( Q ^ [ t ] ) T 2 ι 2 Q ^ [ t + 1 ] Q ^ [ t ] F + 2 A F Q ^ [ t + 1 ] Q ^ [ t ] F 2 ι 2 + A F Q ^ [ t + 1 ] Q ^ [ t ] F
Similarly, the computation of Equation (10) implies that
L ( Δ ^ [ t + 1 ] ) Q ^ = 2 ( C ^ [ t + 1 ] ) T C ^ [ t + 1 ] Q ^ [ t + 1 ] 2 ( C ^ [ t + 1 ] ) T A + Ξ .
and
L ( C ^ [ t ] , Q ^ [ t + 1 ] ) Q ^ = 2 ( C ^ [ t ] ) T C ^ [ t ] Q ^ [ t + 1 ] 2 ( C ^ [ t + 1 ] ) T A + Ξ = 0 .
Plugging Equation (A6) into Equation (A5) yields that
L ( Δ ^ [ t + 1 ] ) Q ^ F 2 C ^ [ t + 1 ] C ^ [ t + 1 ] ( Q ^ [ t + 1 ] ) T C ^ [ t ] C ^ [ t ] ( Q ^ [ t + 1 ] ) T F + 2 A C ^ [ t + 1 ] A Q ^ [ t ] 2 γ 2 C ^ [ t + 1 ] C ^ [ t ] F + 2 A F C ^ [ t + 1 ] C ^ [ t ] F 2 γ 2 + A F C ^ [ t + 1 ] C ^ [ t ] F .
Together with Equation (A5) and Equation (A7), we complete the proof. □
Let x , X be a point and a point set, respectively. Define d i s t ( x , X ) = min y X d i s t ( x , y ) as the shortest distance between x and X . Denote Ψ υ the class of all concave and continuous functions ϱ : [ 0 , υ ] ( 0 , + ) satisfying: (1) ϱ ( 0 ) = 0 ; (2) ϱ is continuous differentiable on ( 0 , υ ) and (3) for all υ ˙ ( 0 , υ ) , ϱ ( υ ˙ ) > 0 . Now, we are presenting the Kurdyka–Lojasiewicz (KL) property.
Definition A1.
(Kurdyka–Lojasiewicz (KL) property [44]) Let κ : R d ( , + ] be a proper lower semicontinuous function. For x ˙ d o m κ = { x R d : κ ( x ) } if there exists an υ ( 0 , + ] , a neighborhood X of x ˙ and a function ϱ Ψ υ such that for all x X { κ ( x ˙ ) < κ ( x ) < κ ( x ˙ ) + υ } , it holds that
ϱ ( κ ( x ) κ ( x ˙ x ) ) d i s t ( 0 , κ ( x ) ) > 1 ,
then κ satisfies the KL property at x ˙ . Further, if κ satisfies the KL property at each point in d o m κ , we then call κ a KL function.
Proof of Theorem 1. 
The proof of Theorem 1 is based on the sufficient conditions provided by Lemmas A1–A3 and the KL property in Definition A1. An analogous analysis of convergence about the integrative matrix factorization problem was completed in [45]. The proof of Theorem 1 can be conducted according to Theorem 1 in [45]. We omit the details here. □

Appendix B

The proof of Theorem 2 requires the following lemma simplified from [32], which provides an error bound for any ( 1 + ε ) -approximate K-means solution.
Lemma A4.
Let Ψ R n × d with K distinct rows satisfying min i j Ψ ( i ; · ) Ψ ( j ; · ) F 2 γ with γ a positive constant. Let Ψ ^ R n × d and Y ^ be an ( 1 + ε ) solution to K-means problem with input Ψ ^ ; then, the number of errors in Y ^ is no larger than ζ ε Ψ Ψ ^ F 2 γ 2 for some constant ζ ε depending only on ε.
Proof of Theorem 2. 
It follows from Theorem 1 that there exist a sufficiently small function ε ( n , L ) depending on n and L such that C ^ C * F 2 ε ( n , L ) . Note that C * R L × M with min C * ( i ; · ) C * ( j ; · ) 2 L m a x . Combining this fact with Lemma A4, we conclude that the misclassification rate is not larger than ζ ε L m a x 4 L C ^ C * F 2 ζ ε ε ( n , L ) L m a x 4 L .

Appendix C

Proof of Theorem 3. 
Denote max { max i O m ( i ; · ) , max i Y m * ( i ; · ) } = ξ ( ξ 1 ) . Define Γ ˜ m = { i : O m ( i ; · ) Y m * ( i ; · ) Π 2 1 ξ } . It is clear that
| Γ m | | Γ ˜ m | = i Γ ˜ m 1 ξ i Γ ˜ m O ^ m ( i ; · ) Y m * ( i ; · ) Π 2 ξ O ^ m Y m * Π F 2 ξ O ^ m Y ^ m Π F 2 + Y ^ m Y m * Π F 2 2 ξ Y ^ m Y m * Π F 2 ξ 2 U ˜ A ¯ m Y m * Π F 2 + U ˜ Q ^ m Y m * Π F = ξ 2 ( I 1 + I 2 )
We first deal with I 1 . For some constant C ( ϵ ) > 0 , it immediately follows from Theorem 3.1 of [46] that
I 1 C ( ϵ ) K m log n α n 2 .
Under Assumptions (A1) and (A2), Theorem 3 of [47] confirms that
I 2 144 C 0 K m log ( 8 n / ϵ ) λ K m 2
A combination of Equations (A9) and (A10) lead to the desired result. □

References

  1. Girvan, M.; Newman, M.E.J. Communitystructureinsocial and biological networks. Proc. Natl. Acad. Sci. USA 2002, 99, 7821–7826. [Google Scholar] [CrossRef] [Green Version]
  2. Javed, M.A.; Younis, M.S.; Latif, S.; Qadir, J.; Baig, A. Community detection in networks: A multidisciplinary review. J. Netw. Comput. Appl. 2018, 108, 87–111. [Google Scholar] [CrossRef]
  3. Kim, J.; Lee, J.G. Community detection in multi-layer graphs: A survey. ACM Sigmod Rec. 2015, 44, 37–48. [Google Scholar] [CrossRef] [Green Version]
  4. Kivelä, M.; Arenas, A.; Barthelemy, M.; Gleeson, J.P.; Moreno, Y.; Porter, M.A. Multilayer networks. J. Complex Netw. 2014, 2, 203–271. [Google Scholar] [CrossRef] [Green Version]
  5. Hric, D.; Darst, R.K.; Fortunato, S. Community detection in networks: Structural communities versus ground truth. Phys. Rev. E 2014, 90, 062805. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Traud, A.L.; Kelsic, E.D.; Mucha, P.J.; Porter, M.A. Comparing community structure to characteristics in online collegiate social networks. SIAM Rev. 2011, 53, 526–543. [Google Scholar]
  7. Mariadassou, M.; Robin, S.; Vacher, C. Uncovering latent structure in valued graphs: A variarion approach. Ann. Appl. Stat. 2010, 4, 715–742. [Google Scholar] [CrossRef]
  8. Amini, A.A.; Chen, A.; Bickel, P.J.; Levina, E. Pseudo-likelihood methods for community detection in large sparse networks. Ann. Stat. 2013, 41, 2097–2122. [Google Scholar] [CrossRef]
  9. Newman, M.E.J. Modularity and community structure in networks. Proc. Natl. Acad. Sci. USA 2006, 103, 8577–8582. [Google Scholar] [CrossRef] [Green Version]
  10. Karrer, B.; Newman, M.E. Stochastic blockmodels and community structure in networks. Phys. Rev. E 2011, 83, 016107. [Google Scholar] [CrossRef] [Green Version]
  11. Fortunato, S.; Darko, H. Community detection in networks: A user guide. Phys. Rep. 2016, 659, 1–44. [Google Scholar] [CrossRef] [Green Version]
  12. Taylor, D.; Caceres, R.S.; Mucha, P.J. Super-resolution community detection for layer-aggregated multilayer networks. Phys. Rev. X 2017, 7, 031056. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Taylor, D.; Shai, S.; Stanley, N.; Mucha, P.J. Enhanced detectability of community structure in multilayer networks through layer aggregation. Phys. Rev. Lett. 2016, 116, 228301. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Dong, X.; Frossard, P.; Vandergheynst, P.; Nefedov, N. Clustering on multi-layer graphs via subspace analysis on grassmann manifolds. IEEE Trans. Signal. Proces. 2013, 62, 905–918. [Google Scholar] [CrossRef] [Green Version]
  15. Tang, L.; Wang, X.; Liu, H. Community detection via hetero- geneous interaction analysis. Data Min. Knowl. Discov. 2012, 25, 1–33. [Google Scholar] [CrossRef]
  16. Xu, K.S.; Hero, A.O. Dynamic stochastic blockmodels for time-evolving social networks. IEEE J. Sel. Top. Signal Process. 2014, 8, 552–562. [Google Scholar] [CrossRef] [Green Version]
  17. Han, Q.; Kevin, X.; Edoardo, A. Consistent estimation of dynamic and multi-layer block models. In Proceedings of the International Conference on Machine Learning, Miami, FL, USA, 9–11 December 2015; pp. 1511–1520. [Google Scholar]
  18. Tang, R.; Tang, M.; Vogelstein, J.T.; Priebe, C.E. Robust estimation from multiple graphs under gross error contamination. arXiv 2017, arXiv:1707.03487. [Google Scholar]
  19. De Bacco, C.; Power, E.A.; Larremore, D.B.; Moore, C. Community detection, link prediction, and layer interdependence in multilayer networks. Phys. Rev. E 2017, 95, 042317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Dong, X.; Frossard, P.V.; Ergheynst, P.; Nefedov, N. Clustering with multi-layer graphs: A spectral perspective. IEEE. Trans. Signal. Proces. 2012, 60, 5820–5831. [Google Scholar] [CrossRef] [Green Version]
  21. Bhattacharyya, S.; Chatterjee, S. Spectral clustering for multiple sparse networks: I. arXiv 2018, arXiv:1805.10594. [Google Scholar]
  22. Tang, W.; Lu, Z.; Dhillon, I.S. Clustering with multiple graphs. In Proceedings of the Ninth IEEE International Conference on Data Mining, Miami, FL, USA, 6–9 December 2009; pp. 1016–1021. [Google Scholar]
  23. Lei, J.; Chen, K.; Lynch, B. Consistent community detection in multi-layer network data. Biometrika 2020, 107, 61–73. [Google Scholar] [CrossRef]
  24. Mucha, P.J.; Richardson, T.; Macon, K.; Porter, M.A.; Onnela, J.P. Community structure in time-dependent, multiscale, and multiplex networks. Science 2010, 328, 876–878. [Google Scholar] [CrossRef]
  25. Bazzi, M.; Porter, M.A.; Williams, S.; McDonald, M.; Fenn, D.J.; Howison, S.D. Community detection in temporal multilayer networks, with an application to correlation networks. Multiscale Model. Simul. 2016, 14, 1–41. [Google Scholar] [CrossRef]
  26. Stanley, N.; Shai, S.; Taylor, D.; Mucha, P.J. Clusteringnetwork layers with the strata multilayer stochastic block model. IEEE. Trans. Netw. Sci. Eng. 2016, 3, 95–105. [Google Scholar] [CrossRef] [Green Version]
  27. Arroyo, J.; Athreya, A.; Cape, J.; Chen, G.; Priebe, C.E.; Vogelstein, J.T. Inference for multiple heterogeneous networks with a common invariant subspace. J. Mach. Learn. Res. 2021, 22, 6303–6351. [Google Scholar]
  28. Jing, B.Y.; Li, T.; Lyu, Z.; Xia, D. Community detection on mixture multilayer networks via regularized tensor decomposition. Ann. Stat. 2021, 49, 3181–3205. [Google Scholar] [CrossRef]
  29. Fan, X.; Pensky, M.; Yu, F.; Zhang, T. ALMA: Alternating Minimization Algorithm for Clustering Mixture Multilayer Network. J. Mach. Learn. Res. 2022, 23, 6303–6351. [Google Scholar]
  30. Le, C.M.; Levin, K.; Levina, E. Estimating a network from multiple noisy realizations. Electron. J. Stat 2018, 12, 4697–4740. [Google Scholar] [CrossRef]
  31. Chen, S.; Liu, S.; Ma, Z. Global and individualized community detection in inhomogeneous multilayer networks. Ann. Stat. 2022, 50, 2664–2693. [Google Scholar] [CrossRef]
  32. Lei, J.; Lin, K.Z. Bias-adjusted spectral clustering in multi-layer stochastic block models. J. Am. Stat. Assoc. 2022, 1–13. [Google Scholar] [CrossRef]
  33. Gupta, A.K.; Nagar, D.K. Matrix Variate Distributions; Chapman Hall/CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
  34. De Domenico, M.; Granell, C.; Porter, M.A.; Arenas, A. The physics of spreading processes in multilayer networks. Nat. Phys. 2016, 12, 901–906. [Google Scholar] [CrossRef] [Green Version]
  35. De Domenico, M.; Biamonte, J. Spectral entropies as information-theoretic tools for complex network comparison. Phys. Rev. X 2016, 6, 041062. [Google Scholar] [CrossRef] [Green Version]
  36. Nicosia, V.; Latora, V. Measuring and modeling correlations in multiplex networks. Phys. Rev. X 2015, 92, 032805. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Yang, L.; Cao, X.; Jin, D.; Wang, X.; Meng, D. A unified semi-supervised community detection framework using latent space graph regularization. IEEE Trans. Cybern. 2015, 45, 2585–2598. [Google Scholar] [CrossRef] [PubMed]
  38. Ma, X.; Gao, L.; Yong, X.; Fu, L. Semi-supervised clustering algorithm for community structure detection in complex networks. Phys. A 2010, 389, 187–197. [Google Scholar] [CrossRef]
  39. Luxburg, V. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  40. Lu, H.; Sang, X.; Zhao, Q.; Lu, J. Community detection algorithm based on nonnegative matrix factorization and pairwise constraints. Phys. A 2020, 545, 123491. [Google Scholar] [CrossRef]
  41. Amini, A.A.; Levina, E. On semidefinite relaxations for the block model. Ann. Stat. 2018, 46, 149–179. [Google Scholar] [CrossRef] [Green Version]
  42. Tang, F.; Wang, C.; Su, J.; Wang, Y. Spectral clustering-based community detection using graph distance and node attributes. Computation. Stat. 2020, 35, 69–94. [Google Scholar] [CrossRef]
  43. Rossi, L.; Magnani, M. Towards effective visual analytics on multiplex and multilayer networks. Chaos. Soliton. Farct. 2015, 72, 68–76. [Google Scholar] [CrossRef] [Green Version]
  44. Attouch, H.; Bolte, J.; Svaiter, B.F. Convergence of descent methods for semi-algebraic and tame problems: Proximal algorithms, forward-backward splitting, and regularized gauss-seidel methods. Math. Programm. 2013, 137, 91–129. [Google Scholar] [CrossRef] [Green Version]
  45. Liu, J.; Wang, J.; Liu, B. Community detection of multi-Layer attributed networks via penalized alternating factorization. Mathematics 2020, 8, 239. [Google Scholar] [CrossRef] [Green Version]
  46. Lei, J.; Rinaldo, A. Consistency of spectral clustering in stochastic block models. Ann. Stat. 2015, 43, 215–237. [Google Scholar] [CrossRef]
  47. Binkiewicz, N.; Vogelstein, J.T.; Rohe, K. Covariate-assisted spectral clustering. Biometrika 2017, 104, 361–377. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. A toy example describes how to estimate C and Q in the objective function Equation (7).
Figure 1. A toy example describes how to estimate C and Q in the objective function Equation (7).
Mathematics 11 01573 g001
Figure 2. (a,b) Between-layer clustering results and within-layer clustering results returned by three methods as the number of nodes n increases; (c,d) Between-layer clustering results and within-layer clustering results returned by three methods as the number of layers L increases; (e,f) Between-layer clustering results and within-layer clustering results returned by three methods as γ varies. Each data point shows the average result and the error bar determines the range of one standard deviation from the average.
Figure 2. (a,b) Between-layer clustering results and within-layer clustering results returned by three methods as the number of nodes n increases; (c,d) Between-layer clustering results and within-layer clustering results returned by three methods as the number of layers L increases; (e,f) Between-layer clustering results and within-layer clustering results returned by three methods as γ varies. Each data point shows the average result and the error bar determines the range of one standard deviation from the average.
Mathematics 11 01573 g002
Figure 3. (a) Node communities found in layer group 1 by MSFE; (b) Node communities found in layer group 2 by MSFE.
Figure 3. (a) Node communities found in layer group 1 by MSFE; (b) Node communities found in layer group 2 by MSFE.
Mathematics 11 01573 g003
Figure 4. (ae) Adjacency matrices of five layers on AU-CS network with nodes arranged according the community labels obtained by MFSE.
Figure 4. (ae) Adjacency matrices of five layers on AU-CS network with nodes arranged according the community labels obtained by MFSE.
Mathematics 11 01573 g004
Table 1. Notation definitions.
Table 1. Notation definitions.
NotationsExplanations
[ n ] [ n ] = { 1 , 2 , , n }
n ; L number of nodes, number of network layers
G ( l ) ; A ( l ) R n × n the l-th layer; the corresponding adjacency matrix
A R n × n × L , A ( · ; · ; l ) = A ( l ) the adjacency tensor
M ; L m ( m = 1 M L m = L ) number of layer groups; number of layers in the m-th layer group
K m ( m = 1 , 2 , , M ) number of node communities in the m-th layer group
G 1 , , G M the network groups such that [ L ] = m = 1 M G m
V k ( m ) , m = 1 , , M , k = 1 , , K m the k-th node community in the m-th layer group
L m a x ; L m i n L m a x = max m L m ; L m i n = min m L m
n m a x ( m ) ; n m i n ( m ) n m a x ( m ) = max k { 1 , , K m } | V k ( m ) | ; n m i n ( m ) = min k { 1 , , K m } | V k ( m ) |
Z ( m ) { 0 , 1 } n × K m the membership (clustering) matrix of the m-th layer group
B ( m ) [ 0 , 1 ] K m × K m link probability matrix of the m-th layer group
P ( m ) = E ( A ( m ) ) = ( P i j ( m ) ) R n × n the expected matrix of the m-th layer
C { 0 , 1 } L × M the membership matrix of layers
a ( m ) the vectoried version of matrix A ( m ) by sequentially stacking its columns
b ( m ) , p ( m ) the vectoried versions of matrices B ( m ) , P ( m ) , respectively
A R L × n 2 , P R L × n 2 matrices with rows a ( m ) and p ( m ) , respectively
A ( i ; · ) , A ( · ; j ) the i-th row and j-th column of a matrix A, respectively
A ( r ; i : j ) the i-to-jth entities in the r-th row of matrix A
L Q the Laplacian matrix of Q
D Q a diagonal-matrix with ( i , i ) -th element D Q ( i , i ) = j Q ( i , j )
d Q m d Q m = min i D Q m ( i , i )
U ^ Q = [ U ^ Q 1 , , U ^ Q K m ] ; U ˜ Q eigenvector matrix corresponding to L Q ; normalized version of U ^ Q
A ^ ; A * an estimator of matrix A; the population version of A
λ k ( W ) the k-th absolute large eigenvalue of matrix W
Π an orthonormal rotation whose rank varies from one line to another
Table 2. Between-layer clustering results obtained by MFSE.
Table 2. Between-layer clustering results obtained by MFSE.
Layer GroupsProducts
Group 1Cheese_whole_cow_milk, Bread, Chocolate_products_nes, Fatty_acids, Anise_badian_fennel_coriander, Beer_of_barley, Chillies_and_peppers_dry, Beverages_distilled_alcoholic, Coffee_extracts, Coffee_roasted, Fruit_dried_nes, Juice_fruit
_nes, Pepper_(piper_spp.), Sugar_nes, Tea_mate_extracts, Vegetables_frozen
Group 2Beverages_non_alcoholic, Macaroni, Food_prep_nes, Cigarettes, Beans_dry, Flour_wheat, Crude_materials, Food_preparations_flour_malt_extract, Food_wastes,
Fruit_prepared_nes, Cereals_breakfast, Coffee_green, Cocoa_powder_and_cake, Pastry, Pet_food, Tea, Nuts_prepared_(exc._groundnuts), Mixes_and_doughs, Oil_vegetable_origin_nes, Oil_essential_nes, Rice_milled, Sugar_refined, Sugar_confectionery, Wine, Spices_nes, Vegetables_dehydrated, Vegetables_preserved_nes, Waters ice_etc, Vegetables_in_vinegar, Tobacco_unmanufactured
Table 3. Between-layer clustering results obtained by MFSE on AU-CS network.
Table 3. Between-layer clustering results obtained by MFSE on AU-CS network.
Layer GroupsLayers
Group 1Coauthor, Leisure, Facebook
Group 2Lunch, Work
Table 4. Densities obtained by MFSE and SPC-Mean on each layer of AU-CS network.
Table 4. Densities obtained by MFSE and SPC-Mean on each layer of AU-CS network.
LayersMFSESPC-Mean
Coauthor0.950.76
Facebook0.760.70
Leisure0.830.76
Lunch0.720.62
Work0.770.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, F.; Zhao, X.; Li, C. Community Detection in Multilayer Networks Based on Matrix Factorization and Spectral Embedding Method. Mathematics 2023, 11, 1573. https://doi.org/10.3390/math11071573

AMA Style

Tang F, Zhao X, Li C. Community Detection in Multilayer Networks Based on Matrix Factorization and Spectral Embedding Method. Mathematics. 2023; 11(7):1573. https://doi.org/10.3390/math11071573

Chicago/Turabian Style

Tang, Fengqin, Xuejing Zhao, and Cuixia Li. 2023. "Community Detection in Multilayer Networks Based on Matrix Factorization and Spectral Embedding Method" Mathematics 11, no. 7: 1573. https://doi.org/10.3390/math11071573

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop