Next Article in Journal
Statistical Adaptation Loss Improved SMALL Sample Ship Detection Method Based on an Attention Mechanism and Data Enhancement
Previous Article in Journal
Secure Image Signal Transmission Scheme Using Poly-Polarization Filtering and Orthogonal Matrix
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Discovering Community Structure in Multiplex Networks via a Co-Regularized Robust Tensor-Based Spectral Approach

by
Esraa Al-sharoa
1,*,
Mohammad Al-wardat
1,2,
Mahmood Al-khassaweneh
2,3 and
Ali Al Bataineh
4
1
Electrical Engineering Department, Jordan University of Science and Technology, Irbid 22110, Jordan
2
Computer Engineering Department, Yarmouk University, Irbid 22110, Jordan
3
Engineering, Computing and Mathematical Sciences, Lewis University, Romeoville, IL 60446, USA
4
Department of Electrical and Computer Engineering, Norwich University, Northfield, VT 05663, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(4), 2514; https://doi.org/10.3390/app13042514
Submission received: 23 January 2023 / Revised: 8 February 2023 / Accepted: 9 February 2023 / Published: 15 February 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Complex networks arise in various fields, such as biology, sociology and communication, to model interactions among entities. Entities in many real-world systems exhibit different types of interactions, which requires modeling these type of systems properly. Multiplex networks are used to model these systems, as they can reflect the nodes’ pair-wise interactions as multiple distinct types of links across layers. Community detection is a widely studied application in network analysis as it provides insights into the structure and organization of the network. Even though multiple algorithms have been developed in the community detection field, many of them have a limited performance in the presence of noise. In this article, we develop a novel algorithm that combines tensor low-rank representation, spectral clustering and distance regularization to improve the accuracy in discovering communities in multiplex networks. The low-rank representation leads to reducing the noise and errors existing in the network and the optimization of an accurate consensus set of eigenvectors that reveals the communities in the network. Moreover, the proposed approach balances the agreement between the eigenvectors of each layer, i.e., individual subspaces, and the consensus set of eigenvectors, i.e., common subspaces, by minimizing the projection distance between them. The common and individual subspaces are computed efficiently through Tucker decomposition and modified spectral clustering, respectively. Finally, multiple experiments are conducted on real and simulated networks to evaluate the proposed approach and compare it to state-of-the-art algorithms. The proposed approach shows its robustness and efficiency in discovering the communities in multiplex networks.

1. Introduction

Real-world systems such as communications, transportation and biology, to name a few, can be presented as complex networks that reflect the intercommunication between their objects [1]. Real-world systems were first represented by simple graphs, and this representation did not take into account multiple or different types of interactions between the system’s objects. Consequently, simple graphs are considered inappropriate for modeling systems that exhibit different types of interactions between their objects. Multiplex networks have emerged to tackle this issue, which represent a multi-layer network consisting of the same nodes across layers, while the edges in the separate layers encode the different types of interactions [2,3].
Community detection is an important application that has been of crucial interest to researchers in analyzing networks [4]. The purpose of community detection is to reveal the set of communities in the network and reduce its dimensionality. A community is typically defined as a batch of strongly connected nodes that are weakly connected to the rest of the nodes in the network [5]. The majority of the early literature in community detection was on discovering communities in monoplex or single-layer networks [6]. Recently, various approaches have evolved to discover communities in multiplex networks [7]. Earlier multiplex network community detection approaches interpreted the multiplex network as a monoplex network by aggregating the different layers into a single-layer, then discovering the community structure by applying traditional simple graph community detection techniques [8]. However, these techniques are not appropriate to reveal the true community structure, as they do not consider the proprieties carried in each layer. To cope with the disadvantages of the aggregation-based methods, other multiplex community detection techniques have been recently proposed.
In [9], a unified non-negative matrix factorization algorithm that aimed to reveal the set of within- and across-layer clusters in fully-connected multi-layer networks was introduced, where the nodes were connected within and across the layers. The authors in [10] presented an approach to investigate a multi-view network structure based on a spectral clustering technique. However, they did not consider the effect of the individual layers on the common subspace. In [11], the authors presented a one parameter family of matrix power means (PM) to investigate the communities in multi-layer networks. The PM method merges the Laplacians from the separate layers and then analyzes them in the stochastic block model. In [12], two methods that aim to investigate the Laplacian rank constrained graph, which is considered as the centroid of the constructed graph for each layer with different confidences, were proposed. In the first method, the layers’ weights are learnt by introducing a hyper-parameter, whereas the second method is self-weighted. A top-bottom network partitioning strategy was proposed in [13], employing cross-layer edge clustering coefficients to assess the possibility of separating the communities. In [14], a robust multi-view spectral clustering (RMSC) technique was proposed. The proposed RMSC uncovered the clusters in the network based on decomposing the transition probability matrix into low rank and sparse components. In [15], the clusters in multi-layer networks were detected based on the Grassmann manifold theory through a modified spectral clustering approach (SCML). The proposed SCLM revealed the community structure of each layer by applying normalized spectral clustering for each layer individually. On the other hand, a centroid-based co-regularization (CentroidCoreg) for spectral clustering was proposed in [16], detecting the community structure based on a common subspace across the layers. In [17], the proposed method relied on the probability distribution of each layer to cluster the multiplex network. In [18], an approach that relied on graph factorization to explore the consensus, as well as the complementary information coming from the different views (2CMV), was presented. In [19], the authors proposed an algorithm (DiMMA) that captured the diversity and complexity of the data from multiple perspectives in order to provide a more accurate and reliable clustering solution. This was achieved by incorporating the distance information from the different views. Nevertheless, the abovementioned techniques are less robust in real-world conditions, as they do not perform well in noisy networks. In [20], the authors proposed to detect the clusters in multiplex networks through joint non-negative matrix factorization (NMF), where multiple formulations of the objective function were introduced: (i) collective symmetric NMF (CSNMF), (ii) collective projective NMF (CPNMF) and (iii) collective symmetric non-negative matrix trifactorization (CSNMTF). In these approaches, a set of non-negative low-dimensional feature representations were found first for the individual layers. The individual non-negative embeddings were then fused into a common feature representation through collective factorization.
In this article, we propose a new approach to discover the communities in multiplex networks called the Co-regularized Robust Tensor-based Spectral Approach (Mx-CRTSA). In particular, we develop an objective function that merges multiple terms, including tensor decomposition, spectral clustering and a centroid-based co-regularization. The goal of the proposed objective function is to optimize a common centroid set of eigenvectors to reveal the communities in the network after recovering a clean representation of the multiplex network. The introduced model is solved through an alternating minimization scheme, where the clean representation is recovered through alternating direction methods of multipliers and the consensus set of eigenvectors through Tucker decomposition. A summary of the proposed Mx-CRTSA is presented in Figure 1. The contributions of Mx-CRTSA are outlined as follows:
1.
The proposed Mx-CRTSA constructs an adjacency tensor form the multiplex network and minimizes its rank through nuclear norm minimization while pruning the noise by adopting the L 2 , 1 for error minimization.
2.
The proposed Mx-CRTSA balances the agreement between the eigenvectors of each layer, i.e., the individual subspace, and a consensus set of eigenvectors, i.e., common or centroid subspace. This is achieved by minimizing the projection distance between the individual and common subspaces.
3.
A co-regularization term is considered in the objective function that includes a weighting parameter, which is optimized efficiently along with the consensus subspace through high order orthogonal iteration (HOOI), which can provide a more accurate and robust solution than other matrix factorization methods. With the aid of HOOI, the contribution of the individual subspaces to the common subspace is optimized.
4.
To assess the performance, robustness and effectiveness of the proposed Mx-CRTSA, multiple experiments are conducted to highlight its potential in clustering multiplex networks. The results indicate its superiority compared to other existing techniques.
The structure of this paper is as follows: The background is given in Section 2. The Mx-CRTSA approach is presented and solved in Section 3. The results are discussed in Section 4. Lastly, the conclusions are outlined in Section 5. A summary of the notation used in this paper is presented in Table 1.

2. Background

2.1. Tensor Operations

In this section, the tensor operations that are used throughout this paper are presented. Suppose we have a 3-mode tensor, A R n 1 × n 2 × n 3 , the basic tensor operations can be then listed as follows:
  • Block circulant matrix, b c i r c ( A ) : Converts the tensor A to a matrix by merging all frontal slices vertically and horizontally, which results in a matrix of the dimension n 3 n 1 × n 3 n 2 , defined as:
    b c i r c ( A ) = A ( 1 ) A ( n 3 ) A ( 2 ) A ( 2 ) A ( 1 ) A ( 3 ) A ( n 3 ) A ( n 3 1 ) A ( 1 ) ,
    where A ( i ) represents the i t h frontal slice of A .
  • Block vectorizing, b v e c ( A ) : Converts the tensor A to a block matrix by aggregating all frontal slices vertically and this results in a matrix of the dimension n 3 n 1 × n 3 defined as follows:
    b v e c ( A ) = A ( 1 ) A ( 2 ) A ( n 3 ) .
    Furthermore, the opposite operation of the block vectorizing that transforms the aggregated matrix into a tensor can be defined as:
    b v f o l d ( b v e c ( A ) ) = A .
  • Block diagonal matrix, b d i a g ( A ) : Converts the tensor A to a block diagonal matrix. This is executed by stacking all frontal slices on the diagonal of the matrix, leading to a matrix of the dimension n 3 n 1 × n 3 n 2 that can be defined as follows:
    b d i a g ( A ) = A ( 1 ) A ( n 3 ) .
    On the other hand, the block diagonal opposite operation b d f o l d ( b d i a g ( A ) ) = A recovers the tensor from the block diagonal matrix.
  • Tensor Product (t-Product): Assuming that we have two 3-mode tensors, namely A 1 and A 2 . Then, the t-product between A 1 and A 2 can be calculated as follows:
    A 3 = A 1 * A 2 = b v f o l d { b c i r c ( A 1 ) b v e c ( A 2 ) } .
  • Tensor-Singular Value Decomposition (T-SVD):
    A = Q l e f t * S f * Q r i g h t ,
    where Q l e f t and Q r i g h t are the left and right orthonormal singular vectors of the tensor A , respectively. S f is the singular value core tensor.
  • T-SVD-Based Tensor Nuclear Norm (T-SVD-TNN): The SVD of A via TNN in the frequency domain and can be calculated as follows:
    A = i = 1 m i n ( n 1 , n 2 ) j = 1 n 3 | A f ( i , i , j ) | ,
    where A f is the SVD of the tensor A in the frequency domain.
  • Tucker Decomposition: Provides orthogonal subspace decomposition along each mode of the tensor A [21]:
    A B × 1 U ( 1 ) × 2 U ( 2 ) × 3 U ( 3 ) ,
    where B R r 1 × r 2 × r 3 is the core tensor, { U ( i ) R n i × r i } i = 1 3 are the low-dimensional orthogonal matrices and r i is the rank of { U ( i ) } with r i n i . Tucker decomposition is solved through an iterative optimization procedure known as HOOI, which provides the best projection matrices [22].

2.2. Graph Theory

A simple network or graph, G, is defined by the set of nodes, V, the set of edges, E, and the adjacency matrix, A. The adjacency matrix A R n × n is an undirected, symmetric and non-negative matrix that reflects the pair-wise affinity between nodes, where a i j [ 0 , 1 ] for weighted networks and a i j { 0 , 1 } for binary networks. A diagonal degree matrix, D, is defined from A, where d i i = j = 1 n a i j [23,24]. A N = D 0.5 A D 0.5 defines a normalized version of the adjacency matrix, which is a positive semi-definite matrix.
An M-layer and n-vertex multiplex network, G , is utilized as a group of simple graphs, { G ( m ) } , where G ( m ) = ( V ( m ) , E ( m ) , A ( m ) ) with m { 1 , , M } . A ( m ) R n × n refers to the adjacency matrix of the m t h layer.

2.3. Spectral Clustering

Spectral clustering [23] is a famous approach in graph clustering to optimize the solution of:
max U t r ( U A N U ) , s . t U U = I ,
where t r refers to the matrix trace operator. As the matrix A N is positive semi-definite, Equation (9) can be reformulated using the Frobenius norm as:
max U U A N U F 2 , s . t U U = I .
Both maximization objective functions in Equations (9) and (10) can be solved by eigenvalue decomposition of A N , where U R n × c is the matrix of the eigenvectors set that corresponds to the largest c eigenvalues of A N . After the indicator matrix U is computed, k-means is used to reveal the vertices community assignment [23].

2.4. Projection Distance

Let U 1 R n × c and U 2 R n × c be two orthonormal basis sets that correspond to two subspaces, span ( U 1 ) and span ( U 2 ) , respectively. The distance between the two subspaces can be quantified by the principal angles between span ( U 1 ) and span ( U 2 ) . Then, the squared projection distance between span ( U 1 ) and span ( U 2 ) can be formulated as [25]:
d p r o j 2 ( span ( U 1 ) , span ( U 2 ) ) = i = 1 c sin 2 ( θ i ) = c i = 1 c cos 2 ( θ i ) = c t r ( U 1 U 1 U 2 U 2 ) = 1 2 2 c 2 t r ( U 1 U 1 U 2 U 2 ) = 1 2 t r ( U 1 U 1 ) + t r ( U 2 U 2 ) 2 t r ( U 1 U 1 U 2 U 2 ) = 1 2 U 1 U 1 U 2 U 2 F 2 ,
where θ i is the i th principal angle between span ( U 1 ) and span ( U 2 ) .

3. Discovering Community Structure in Multiplex Networks: Co-regularized Robust Tensor-Based Spectral Approach (Mx-CRTSA)

In this paper, we approach the problem of discovering community structure in noisy multiplex networks and propose grouping the nodes by efficiently incorporating the information furnished by the different layers. More precisely, we propose reducing the effect of noise and outliers in the network by extracting a set of clean graphs from the original multiplex network. These clean graphs conserve the significant characteristics of the multiplex network and can be used to compute the lower dimensional subspaces by exploiting concepts from subspace analysis. The contribution of these individual subspaces is then optimized to find a representative common subspace of the network.

3.1. Proposed Objective Function

Given a noisy M-layer multiplex network, an adjacency tensor, A R n × n × M , is constructed from the network by stacking the normalized adjacencies, { A N ( m ) } m = 1 M R n × n , of the individual layers as the frontal slices of the tensor. The proposed Mx-CRTSA aims to detect the community structure of the multiplex network by first extracting a clean self-representation tensor from the noisy adjacency tensor. The extracted clean tensor is then used to compute the low-dimensional individual spectral embeddings or subspaces, { U m } m = 1 M R n × c , and a common centroid spectral embedding across all layers, U * R n × c , where U * is enforced to be close to all { U m } m = 1 M . The proposed objective function is formulated as follows:
min U * , U ( m ) , Z , E , w Z + γ E 2 , 1 λ m = 1 M t r U ( m ) Z N ( m ) U ( m ) + w m t r U * U ( m ) U ( m ) U * s . t . A ( m ) = A ( m ) Z ( m ) + E ( m ) , E = [ E ( 1 ) ; E ( 2 ) ; ; E ( M ) ] , Z ( m ) = Z ( m ) , U ( m ) U ( m ) = I , m = { 1 , , M } , Z = Φ ( Z ( 1 ) , Z ( 2 ) , , Z ( M ) ) , w 2 = 1 , w 0 , U * U * = I ,
where
  • Z is considered to recover the low-rank approximation tensor, Z , from the adjacency tensor, A . Φ ( · ) is adopted to construct the clean tensor from the recovered self-representation matrices, { Z ( m ) } m = 1 M , and rotate it so that the computational complexity is reduced while the tensor’s nuclear norm is minimized.
  • E 2 , 1 represents the L 2 , 1 of the error matrix E R n M × n , which is included to remove the noise from the adjacency tensor, A . The matrix E is formed by arranging the individual error matrices, { E ( m ) } m = 1 M , vertically.
  • The last term consists of two sub-terms. The first sub-term, m = 1 M t r ( U ( m ) Z N ( m ) U ( m ) ) , refers to the normalized cut of each layer which is included to find the individual spectral embedding of the separate layers, { U ( m ) } m = 1 M , where Z N ( m ) is the normalized clean adjacency matrix of the m t h layer. The second sub-term, w m t r ( U * U ( m ) U ( m ) U * ) , represents the co-regularization across layers and corresponds to the projection distance between the individual subspaces, { U ( m ) } m = 1 M , and the centroid or common subspace, U * . By minimizing the projection distance, the common subspace is constrained to be close to all { U ( m ) } m = 1 M , while the node connectivity in each layer is still preserved. The degree of the consistency between the individual subspaces and the common subspace is controlled by the weighting vector, w = [ w 1 , w 2 , , w m ] T , which also will be optimized.
  • γ and λ are regularization parameters to balance the trade-off between the different terms in the proposed objective function.

3.2. Optimization of the Proposed Objective Function

The proposed objective function is solved through an alternating minimization scheme. More precisely, we solve for Z and E while fixing { U ( m ) } m = 1 M , U * and w through the alternating direction method of multipliers (ADMM). Then, the individual subspaces, { U ( m ) } m = 1 M , and the common subspace, U * , across all layers, along with w, are estimated efficiently using eigendecomposition and HOOI, respectively.
To solve the proposed objective function described in Equation (12), an auxiliary variable, H , is introduced to allow variables separability. By introducing H , the objective function is reformulated as follows:
min U * , U ( m ) , Z , E , w , H H + γ E 2 , 1 λ m = 1 M t r U ( m ) Z N ( m ) U ( m ) + w m t r U * U ( m ) U ( m ) U * s . t . A ( m ) = A ( m ) Z ( m ) + E ( m ) , E = [ E ( 1 ) ; E ( 2 ) ; ; E ( M ) ] , Z ( m ) = Z ( m ) U ( m ) U ( m ) = I , m = { 1 , , M } , Z = H , H = Φ ( H ( 1 ) , H ( 2 ) , , H ( M ) ) , w 2 = 1 , w 0 , U * U * = I .
The augmented Lagrangian function can be then written as follows:
min U * , U ( m ) , Z , E , w , H H + γ E 2 , 1 λ m = 1 M t r U ( m ) Z N ( m ) U ( m ) + w m t r U * U ( m ) U ( m ) U * + m = 1 M X 1 ( m ) , A ( m ) A ( m ) Z ( m ) E ( m ) + β 1 2 A ( m ) A ( m ) Z ( m ) E ( m ) F 2 + X 2 , Z H + β 2 2 Z H F 2 s . t . E = [ E ( 1 ) ; E ( 2 ) ; ; E ( M ) ] , Z ( m ) = Z ( m ) , U ( m ) U ( m ) = I , m = { 1 , , M } , H = Φ ( H ( 1 ) , H ( 2 ) , , H ( M ) ) , w 2 = 1 , w 0 , U * U * = I ,
where the matrix X 1 ( m ) and tensor X 2 are the Lagrangian multipliers. β 1 and β 2 are penalty parameters. · is the trace inner product. To this end, Equation (14) can be solved by an alternating minimization scheme as follows:
  • Updating H : Fixing all the variables except H and considering all the terms that includes the variable H from Equation (14) gives:
    H k + 1 = min H H + X 2 k , H k Z k + β 2 2 H k Z k F 2 = min H H + β 2 2 H k P k F 2 ,
    where P k = Z k + X 2 k β 2 . The solution of Equation (15) can be found by applying tensor tubal-shrinkage operator [26]:
    H k + 1 = F τ P k = Q l e f t * F τ S f * Q r i g h t ,
    where P k = Q l e f t * S f * Q r i g h t , F τ S f = S f * D and τ = M β 2 . The diagonal tensor D R n × n × M contains the singular values in the frequency domain and can be calculated as D ( i , i , j ) = ( 1 τ S f ( j ) ( i , i ) ) + .
  • Updating E: The sub-problem of E can be written as:
    E k + 1 = min E γ E 2 , 1 + m = 1 M ( X 1 k ( m ) , A ( m ) A ( m ) Z k ( m ) E k ( m ) + β 1 2 A ( m ) A ( m ) Z k ( m ) E k ( m ) F 2 ) ,
    where Equation (17) can be written as:
    E k + 1 = min E γ E 2 , 1 + m = 1 M β 1 2 E ( m ) N k ( m ) F 2 ,
    where N k ( m ) = A ( m ) A ( m ) Z k ( m ) + X 1 k ( m ) β 1 . To simplify Equation (18), we construct a matrix N by stacking the layers of N ( m ) , m = { 1 , , M } vertically which leads to the following equation:
    E k + 1 = min E γ E 2 , 1 + β 1 2 E N F 2 ,
    where E can be computed as suggested in [27]:
    E k + 1 = N : , j 2 γ β 1 N : , j 2 , N : , j 2 > γ β 1 0 , elsewhere ,
    where N : , j 2 refers to the L 2 -norm of the j t h column in N.
  • Updating Z ( m ) : Considering all the terms with Z ( m ) and fixing the rest of the variables, the sub-problem of Z k + 1 ( m ) is written as:
    Z k + 1 ( m ) = min Z ( m ) X 1 k ( m ) , A ( m ) A ( m ) Z k ( m ) E k + 1 ( m ) + β 1 2 A ( m ) A ( m ) Z k ( m ) E k + 1 ( m ) F 2 + X 2 , Z k H k + 1 + β 2 2 Z k H k + 1 F 2 λ t r U k ( m ) Z N k ( m ) U k ( m ) .
    The updated Z ( m ) is calculated by finding the gradient of Equation (21) and setting it to zero [28] which results in a closed form solution of Z l + 1 ( m ) as:
    Z k + 1 ( m ) = β 2 I + β 1 A ( m ) A ( m ) 1 · ( λ D ( m ) ( 0.5 ) U k ( m ) U k ( m ) D ( m ) ( 0.5 ) + β 2 H k + 1 ( m ) X 2 k ( m ) + β 1 A ( m ) T k ( m ) ) ,
    where T k ( m ) = A ( m ) E k + 1 ( m ) + X 1 k ( m ) β 1 and the matrix Z ( m ) is found for each layer separately.
  • Updating U ( m ) : Fixing all the variables except U ( m ) in Equation (14), we obtain:
    U k + 1 ( m ) = min U ( m ) λ t r U ( m ) Z N ( m ) U ( m ) + w m t r U * U ( m ) U ( m ) U * s . t . U ( m ) U ( m ) = I .
    An equivalent form of Equation (23) is:
    U k + 1 ( m ) = max U ( m ) λ t r U ( m ) Z m o d ( m ) U ( m ) s . t . U ( m ) U ( m ) = I ,
    where Z m o d ( m ) = D ( m ) ( 0.5 ) Z ( m ) D ( m ) ( 0.5 ) + w m U * U * . The resultant problem in Equation (24) is equivalent to the maximization problem in Equation (9), but with an adjusted or modified normalized adjacency matrix. Consequently, by the Rayleigh–Ritz theorem, Equation (24) can be solved through eigenvalue decomposition (EVD) of Z m o d ( m ) , where U ( m ) contains the set of orthonormal eigenvectors that correspond to the largest c eigenvalues.
  • Updating U * : The U * sub-problem can be defined as:
    U k + 1 * = min U * , w λ t r U * m = 1 M w m U ( m ) U ( m ) U * s . t . w 2 = 1 , w 0 , U * U * = I .
    Equation (25) can be written as a trace maximization problem:
    U k + 1 * = max U * , w λ t r U * m = 1 M w m U ( m ) U ( m ) U * s . t . w 2 = 1 , w 0 , U * U * = I .
    Equation (26) is rewritten in terms of Tucker decomposition as [4,29]:
    U k + 1 * = max U * , w Γ × 1 U * × 2 U * × 3 w F 2 s . t . w 2 = 1 , w 0 , U * U * = I ,
    where Γ ( m ) = U ( m ) U ( m ) , m = { 1 , , M } . The common subspace between all layers of the multiplex network, U * , and the weighting vector across layers, w m , can be found efficiently using HOOI.
  • Finally, the Lagrange multipliers, X 1 k + 1 ( m ) , X 2 k + 1 , and the penalty parameters, β 1 k + 1 , β 2 k + 1 , are computed as:
    X 1 k + 1 ( m ) = X 1 k ( m ) + β 1 k A ( m ) A ( m ) Z k + 1 ( m ) E k + 1 ( m ) ,
    X 2 k + 1 = X 2 k + β 2 k + 1 Z k + 1 H k + 1 ,
    β 1 k + 1 = min ( ρ β 1 k , β m a x ) ,
    β 2 k + 1 = min ( ρ β 2 k , β m a x ) ,
    where ρ is set to be 2. The proposed algorithm keeps iterating until it meets the following stopping criterion:
    A ( m ) A ( m ) Z ( m ) E ( m ) ϵ , Z H ϵ .
A summary of the proposed algorithm is described in Algorithm 1.
Algorithm 1Mx-CRTSA
Input :
A , c, γ , λ , ρ = 2 , ϵ = 0.01 , β 1 = β 2 = 0.01 , γ m a x = 10 10 .
Output :
Clustering labels.
1:
Initialize H = X 2 = 0 , E ( m ) = Z ( m ) = X 1 ( m ) = 0 , { m = 1 , , M } .
2:
while not converge do
3:
   Update H k + 1 through Equation (16).
4:
   Update E k + 1 ( m ) through Equation (20).
5:
   Update Z k + 1 ( m ) through Equation (22).
6:
   Update U k + 1 ( m ) through Equation (24).
7:
   Update U k + 1 * through Equation (27).
8:
   Update X 1 k + 1 ( m ) , X 2 k + 1 , β 1 k + 1 , and β 2 k + 1 via Equations (28)–(31).
9:
end while
10:
Obtain community assignment by applying k-means to U * .
11:
return Clustering labels.

3.3. Complexity of Mx-CRTSA

In Mx-CRTSA, multiple variables are updated in each iteration including H , E, U ( m ) and U * . The computational cost of H and E is O ( M n 2 l o g ( n ) + M 2 n 2 ) and O ( M n 2 ) , respectively. The cost of computing U ( m ) is the same as computing U * and equals O ( M n 3 ) . Therefore, the total cost of the Mx-CRTSA algorithm is equal to O ( L ( M n 2 l o g ( n ) + M 2 n 2 + M n 2 + 2 M n 3 ) ) O ( M L n 3 ) , where L, M and n refer to the number of iterations, layers and vertices, respectively.

4. Results and Discussion

Multiple experiments on simulated and real multiplex networks have been conducted to investigate the effectiveness and robustness of the proposed Mx-CRTSA compared to other community detection methods, including PM [11], RMSC [14], SCML [15], CentroidCoreg [16], CSNMF [20], CPNMF [20], CSNMTF [20], DiMMA [19] and 2CMV [18]. A comparison between the objective functions of the different methods considered in the experiments is illustrated in Table 2. The performance of the different methods is quantified using accuracy (ACC), normalized mutual information (NMI), adjusted Rand index (AR) and F-score [30,31]. All the adopted quality metrics are normalized in the range of [ 0 , 1 ] , where higher values indicate more accurate clustering results. The experiments were carried out using MATLAB 2020a on a PC (Intel(R) Core(TM) i7 processor and 16 GB of RAM). For reproducibility purposes, the implemented MATLAB code of the proposed Mx-CRTSA approach is publicly available in our GitHub repository https://github.com/wardat99/Mx-CRTSA, accessed on 8 February 2023.

4.1. Simulated Multiplex Networks

In this experiment, the adopted simulated multiplex networks are developed in [32]. The structure of these networks is controlled by two parameters: μ 0 and p 0 . μ 0 is the community mixing parameter that controls the network’s modularity, while p 0 is the value of the probability of the inter-layer dependency tensor which is used to control the dependency across the layers of the network. The values of μ 0 and p 0 can vary in the range of [ 0 , 1 ] . As the value of μ 0 increases, the noise level in the network increases.
The simulated multiplex networks are generated by varying the parameter μ 0 to control the noise level in the multiplex network so that the robustness of the proposed Mx-CRTSA can be evaluated. In our experiment, we kept the value of p 0 = 1 and varied the value of μ 0 .
The effectiveness of the different algorithms, including Mx-CRTSA, in discovering the community structure in multiplex networks as μ 0 changes in the range [ 0.1 , 0.8 ] was validated by NMI and AR under multiple scenarios. In the first scenario, 2-layer multiplex networks with n = 128 nodes and c = 4 communities were generated. Examples of some of the networks tested in the first scenario for different values of μ 0 are shown Figure 2a–d. In the second scenario, the number of layers was increased to three, while the number of nodes and clusters were fixed, i.e., n = 128 and c = 4 , to explore the performance of the proposed Mx-CRTSA method with networks with a larger number of layers. Furthermore, some examples of the generated networks from scenario 2 are shown in Figure 3a–f. In the third scenario, 2-layer multiplex networks were generated with n = 128 nodes and c = 10 communities, to explore the performance of the proposed Mx-CRTSA with networks that consist of small-sized clusters. More examples of the generated networks from scenario 3 are shown in Figure 4a–d.
In all scenarios, the experiments were replicated 50 times and the variation in the average values of NMI and AR with respect to μ 0 for the different methods are reported in Figure 5a–f. As the figure shows, the proposed Mx-CRTSA method scores the best results in all networks in terms of NMI and AR compared to the other methods. This experiment reflects the robustness and effectiveness of the proposed Mx-CRTSA compared to other methods under comparison. These results indicate that the proposed algorithm mitigates the impact of noise by estimating the clean low-rank representation tensor which leads to more accurate clustering results.

Scalability Analysis

To test the proposed Mx-CRTSA’s scalability, a set of simulated networks were generated where the number of nodes in the networks was increased from 32 to 2048 on a logarithmic scale. The multiplex networks were generated following [32] with c = 10 , M = 2 and μ 0 = 0.1 . The comparison between the run time taken by Mx-CRTSA and the other methods is reported in Figure 6. As the figure demonstrates, Mx-CRTSA takes a longer time compared to most of the other methods and is comparable to CSNMTF and DiMMA. However, the Mx-CRTSA method obtains the best results over the compared methods in terms of community detection.

4.2. Real-World Multiplex Networks

Multiple well-known real-world multiplex networks were adopted to validate the performance of the proposed Mx-CRTSA, including:
  • Nokia Research Center (NRC). (https://bitbucket.org/uuinfolab/20csur/src/master/algorithms/spectral/data/) (accessed on 15 January 2023) Mobile phone data of 136 users living and working in Switzerland with eight communities and three layers, including physical location, Bluetooth scans and phone calls [33].
  • CKM physicians innovation. (https://manliodedomenico.com/data.php) (accessed on 15 January 2023) A network that reflects data on medical innovation, considering physicians in four towns in Illinois (Peoria, Bloomington, Quincy and Galesburg) with three layers and four communities [34].
  • C. Elegans. A three-layer multiplex network of the Caenorhabditis Elegans connectome that reflects the synaptic junctions between the neurons including electric, chemical monadic and polyadic junctions. The network consists of 279 nodes and 12 communities representing the neuron groups [35,36].
  • Cora. (https://people.cs.umass.edu/~mccallum/data.html) (accessed on 15 January 2023) A subset of the Cora bibliographic dataset consisting of 292 research papers with three communities, including natural language processing, data mining and robotics, and three layers, including, title similarity, abstract similarity and citation relationship.
  • BBCSPORT. (http://mlg.ucd.ie/datasets/bbc.html) (accessed on 15 January 2023) A collection of BBC sport articles ( n = 544 ) with two layers and grouped into five clusters: athletics, cricket, football, rugby and tennis. [37].
  • Wikipedia. (http://www.svcl.ucsd.edu/projects/crossmodal/) (accessed on 15 January 2023) A dataset containing 693 articles, where each article is considered a node with two layers and ten communities.
  • COIL-20. (https://www.cs.columbia.edu/CAVE/software/softlib/) (accessed on 15 January 2023) An image dataset that consists of 1440 images collected from the Columbia object image library with three layers that reflect intensity, LBP and Gabor, and twenty communities.
  • 100Leaves. (https://archive.ics.uci.edu/ml/datasets/One-hundred+plant+species+leaves+) (accessed on 15 January 2023) A plant species leaf database with 1600 samples, 100 communities and 3 layers, where the edges in the layers reflect shape descriptor, fine-scale margin and texture.
A description of the real-world multiplex networks is shown in Table 3.

4.2.1. Real-World Multiplex Networks Clustering

In this section, the experimental results of the proposed Mx-CRTSA method on real multiplex networks are presented and compared to other well-known methods in clustering multiplex networks through several clustering quality measures. These quality measures include ACC, NMI, AR and F-score. The comparison results are summarized in Table 4, where the best performance is in bold font. The evaluation metrics results reported in Table 4 illustrate that the proposed method achieves the best performance across the different networks. For instance, Mx-CRTSA is the only method that detects the correct community structure, with a score of 1 in the different quality metrics in both CKM and Cora networks. In particular, the proposed Mx-CRTSA enhances the performance by 3.7 % , 98.3 % , 17.9 % , 4.2 % , 17.9 % , 96.8 % , 5.2 % , 52.8 % and 22.2 % in the CKM network in terms of NMI over the CSNMTF, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV, respectively, whereas in Cora, the performance is boosted by 49.2 % , 55.8 % , 39.7 % , 2.1 % , 49.7 % , 2.1 % , 1.8 % , 54.5 % and 49.3 % in terms of ACC over the CSNMTF,, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV, respectively. Another example is the NRC network, where Mx-CRTSA improves the performance by 26.9 % , 39.6 % , 21.1 % , 15.7 % , 17.3 % , 17.6 % , 19.2 % , 24.7 % and 20.8 % in terms of the NMI over CSNMTF, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV, respectively. In the C. Elegans network, the Mx-CRTSA algorithm improves the quality of the detected structure by approximately 4.8 % , 12.7 % , 6.7 % , 6.1 % , 20.6 % , 8.2 % , 7.7 % , 16.8 % and 17.5 % in terms of the NMI over the CSNMTF, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV method, respectively. Furthermore, Mx-CRTSA performs better than CSNMTF, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV by 70.4 % , 75.3 % , 71.7 % , 25.4 % , 9.4 % , 46.9 % , 22.8 % , 53.7 % and 27.5 % in the BBCSport network in terms of ACC, respectively. For the 100leaves network, the proposed Mx-CRTSA improves the performance by 53.8 % , 71.7 % , 13.9 % , 3.3 % , 16.6 % , 28.8 % , 5.1 % , 35.2 % and 11.7 % in terms of the F-score over the CSNMTF, CPNMF, CSNMF, SCML, RMSC, PM, CentroidCoreg, DiMMA and 2CMV, respectively. The improvement in performance of the proposed approach with respect to the other algorithms and the different metrics is presented in Table 5. These results demonstrate the enhancement achieved by the proposed algorithm with respect to the other methods, where estimating a clean adjacency tensor and employing it to discover the communities by optimizing the contribution of the individual subspaces to the common subspace boosts the accuracy of the detected community structure.

4.2.2. Convergence Discussion

The solution of the proposed Mx-CRTSA is conducted through an alternating minimization scheme. We start with estimating the low-rank component through ADMM. The individual and centroid spectral embeddings are then estimated through spectral clustering. More precisely, the individual subspaces, { U ( m ) } m = 1 M , and the common subspace across all layers, U * , along with the weights vector, w, are estimated efficiently using eigenvalue decomposition and HOOI, respectively.

4.2.3. Parameters Selection

In the proposed objective function there are two regularization parameters that need to be tuned, ( γ , λ ) , where γ is used to control the error component, E, whereas λ controls the individual and the common normalized cut terms across the layers. To investigate the performance sensitivity of the proposed Mx-CRTSA to these parameters, both γ and λ were varied from 0.01 to 10 and the effect of tuning one of the parameters is explored while keeping the other one fixed. According to Figure 7, the optimal range for λ and γ to achieve the best results is found to be λ [ 0.1 , 0.7 ] and γ [ 0.04 , 0.09 ] .

4.2.4. Run Time Comparison

The average run time of Mx-CRTSA was compared to the run times of the other methods, including CentroidCoreg, PM, SCML, RMSC, CPNMF, CSNMF and CSNMTF. The reported results refer to the average run times over 10 realizations and are displayed in the bar graph shown in Figure 8. As the bar graph shows, Mx-CRTSA is slower than all the other methods except CSNMTF, and has a comparable speed to PM and CPNMF. Nevertheless, the proposed algorithm surpasses the rest of the algorithms regarding detecting the community structure, as demonstrated by the experiments.

5. Conclusions

In this paper, a co-regularized tensor-based community detection approach for multiplex networks is proposed. In particular, in the proposed algorithm, a tensor low-rank representation, a normalized cut and a centroid-based co-regularization are combined to reveal the clusters in multiplex networks. The proposed objective function decomposes the multiplex network’s adjacency tensor into a clean low-rank representation and error components. Moreover, the proposed Mx-CRTSA reveals a common centroid set of eigenvectors to uncover the communities in the network by optimizing the contribution of the individual subspaces to the common subspace. The performance of Mx-CRTSA is tested and compared in various synthetic and real multiplex networks with other community detection algorithms.
The proposed approach exhibits robustness to noise as a consequence of estimating the low-rank representation of the original tensor. Furthermore, the distance co-regularization term considered in the proposed objective optimizes the contribution of the individual subspaces that correspond to the separate layers to the common subspace by minimizing the projection distance between them. The contributions of the individual subspaces are optimized along with the consensus subspace through HOOI efficiently. Finally, the experimental results demonstrate that the Mx-CRTSA approach is capable of discovering the communities in multiplex networks with higher accuracy compared to the other algorithms.
For future work, we will consider reducing the high computational complexity of the proposed approach. In fact, the high complexity comes from the computation of the individual subspaces through EVD and the common subspace through high order SVD. Consequently, truncated SVD and partial EVD will be considered to solve this problem.

Author Contributions

Conceptualization, E.A.-s.; methodology, E.A.-s. and M.A.-w.; software, E.A.-s. and M.A.-w.; validation, E.A.-s., M.A.-w., M.A.-k. and A.A.B.; formal analysis, E.A.-s. and M.A.-w.; validation, E.A.-s., M.A.-w., M.A.-k. and A.A.B.; investigation, E.A.-s. and M.A.-w.; resources, E.A.-s., M.A.-w., M.A.-k. and A.A.B.; data curation, E.A.-s. and M.A.-w.; writing—original draft preparation, E.A.-s., and M.A.-w.; writing—review and editing, E.A.-s., M.A.-w., M.A.-k. and A.A.B.; visualization, E.A.-s., M.A.-w., M.A.-k. and A.A.B.; supervision, E.A.-s.; project administration, E.A.-s.; funding acquisition, E.A.-s., M.A.-k. and A.A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by: Jordan University of Science and Technology under Grant 20220436, Senator Patrick Leahy School of Cybersecurity and Advanced Computing, Norwich University and Lewis University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Barabási, A.L. Network science. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2013, 371, 20120375. [Google Scholar] [CrossRef] [PubMed]
  2. Kivelä, M.; Arenas, A.; Barthelemy, M.; Gleeson, J.P.; Moreno, Y.; Porter, M.A. Multilayer networks. J. Complex Netw. 2014, 2, 203–271. [Google Scholar] [CrossRef] [Green Version]
  3. Boccaletti, S.; Bianconi, G.; Criado, R.; Del Genio, C.I.; Gómez-Gardenes, J.; Romance, M.; Sendina-Nadal, I.; Wang, Z.; Zanin, M. The structure and dynamics of multilayer networks. Phys. Rep. 2014, 544, 1–122. [Google Scholar] [CrossRef] [Green Version]
  4. Al-Sharoa, E.; Al-Khassaweneh, M.; Aviyente, S. Tensor based temporal and multilayer community detection for studying brain dynamics during resting state fMRI. IEEE Trans. Biomed. Eng. 2018, 66, 695–709. [Google Scholar] [CrossRef] [PubMed]
  5. Lancichinetti, A.; Fortunato, S. Community detection algorithms: A comparative analysis. Phys. Rev. E 2009, 80, 056117. [Google Scholar] [CrossRef] [Green Version]
  6. Fortunato, S.; Hric, D. Community detection in networks: A user guide. Phys. Rep. 2016, 659, 1–44. [Google Scholar] [CrossRef] [Green Version]
  7. Huang, X.; Chen, D.; Ren, T.; Wang, D. A survey of community detection methods in multilayer networks. Data Min. Knowl. Discov. 2021, 35, 1–45. [Google Scholar] [CrossRef]
  8. Al-Sharoa, E.M.; Bara’M, A.; Alkhassaweneh, M.A. Robust Community Detection in Graphs. IEEE Access 2021, 9, 118757–118770. [Google Scholar] [CrossRef]
  9. Al-Sharoa, E.M.; Aviyente, S. Community Detection in Fully-Connected Multi-layer Networks Through Joint Nonnegative Matrix Factorization. IEEE Access 2022, 10, 43022–43043. [Google Scholar] [CrossRef]
  10. Al-Sharoa, E.M.; Al-Wardat, M.A. Multi-View Robust Tensor-Based Subspace Clustering. IEEE Access 2022, 10, 134292–134306. [Google Scholar] [CrossRef]
  11. Mercado, P.; Gautier, A.; Tudisco, F.; Hein, M. The power mean Laplacian for multilayer graph clustering. In Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR, Playa Blanca, Lanzarote, 9–11 April 2018; pp. 1828–1838. [Google Scholar]
  12. Nie, F.; Li, J.; Li, X. Self-weighted Multiview Clustering with Multiple Graphs. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 2564–2570. [Google Scholar]
  13. Bródka, P.; Filipowski, T.; Kazienko, P. An introduction to community detection in multi-layered social network. In World Summit on Knowledge Society; Springer: Berlin/Heidelberg, Germany, 2011; pp. 185–190. [Google Scholar]
  14. Xia, R.; Pan, Y.; Du, L.; Yin, J. Robust multi-view spectral clustering via low-rank and sparse decomposition. In Proceedings of the AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  15. Dong, X.; Frossard, P.; Vandergheynst, P.; Nefedov, N. Clustering on multi-layer graphs via subspace analysis on Grassmann manifolds. IEEE Trans. Signal Process. 2013, 62, 905–918. [Google Scholar] [CrossRef] [Green Version]
  16. Kumar, A.; Rai, P.; Daume, H. Co-regularized multi-view spectral clustering. Adv. Neural Inf. Process. Syst. 2011, 24, 1413–1421. [Google Scholar]
  17. Li, X.; Ng, M.K.; Ye, Y. Multicomm: Finding community structure in multi-dimensional networks. IEEE Trans. Knowl. Data Eng. 2013, 26, 929–941. [Google Scholar] [CrossRef] [Green Version]
  18. Luong, K.; Nayak, R. A novel approach to learning consensus and complementary information for multi-view data clustering. In Proceedings of the 2020 IEEE 36th International Conference on Data Engineering (ICDE), Dallas, TX, USA, 20–24 April 2020; pp. 865–876. [Google Scholar]
  19. Luong, K.; Nayak, R. Learning Inter- and Intra-Manifolds for Matrix Factorization-Based Multi-Aspect Data Clustering. IEEE Trans. Knowl. Data Eng. 2022, 34, 3349–3362. [Google Scholar] [CrossRef]
  20. Gligorijević, V.; Panagakis, Y.; Zafeiriou, S. Non-negative matrix factorizations for multiplex network analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 928–940. [Google Scholar] [CrossRef] [Green Version]
  21. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  22. Kolda, T.G. Multilinear Operators for Higher-Order Decompositions; Technical Report; Sandia National Laboratories (SNL): Albuquerque, NM, Mexico; Livermore, CA, USA, 2006. [Google Scholar]
  23. Von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  24. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 9–14 December 2002; pp. 849–856. [Google Scholar]
  25. Hamm, J.; Lee, D.D. Grassmann discriminant analysis: A unifying view on subspace-based learning. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 376–383. [Google Scholar]
  26. Xie, Y.; Tao, D.; Zhang, W.; Liu, Y.; Zhang, L.; Qu, Y. On unifying multi-view self-representations for clustering by tensor multi-rank minimization. Int. J. Comput. Vis. 2018, 126, 1157–1179. [Google Scholar] [CrossRef] [Green Version]
  27. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 171–184. [Google Scholar] [CrossRef] [Green Version]
  28. Birgin, E.G.; Martínez, J.M.; Raydan, M. Spectral projected gradient methods: Review and perspectives. J. Stat. Softw. 2014, 60, 1–21. [Google Scholar] [CrossRef]
  29. Al-Sharoa, E.; Al-Khassaweneh, M.; Aviyente, S. A tensor based framework for community detection in dynamic networks. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2312–2316. [Google Scholar]
  30. Christopher, D.M.; Prabhakar, R.; Hinrich, S. Introduction to Information Retrieval; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  31. Hubert, L.; Arabie, P. Comparing partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
  32. Jeub, L.; Bazzi, M. A generative model for mesoscale structure in multilayer networks implemented in MATLAB, preprint. arXiv 2016, arXiv:1608.06196. [Google Scholar]
  33. Kiukkonen, N.; Blom, J.; Dousse, O.; Gatica-Perez, D.; Laurila, J. Towards rich mobile phone datasets: Lausanne data collection campaign. In Proceedings of the ICPS 2010: The 7th ACM International Conference on Pervasive Services (ICPS 2010), Berlin, Germany, 13–16 July 2010. [Google Scholar]
  34. Coleman, J.; Katz, E.; Menzel, H. The diffusion of an innovation among physicians. Sociometry 1957, 20, 253–270. [Google Scholar] [CrossRef]
  35. Chen, B.L.; Hall, D.H.; Chklovskii, D.B. Wiring optimization can relate neuronal structure and function. Proc. Natl. Acad. Sci. USA 2006, 103, 4723–4728. [Google Scholar] [CrossRef] [Green Version]
  36. De Domenico, M.; Porter, M.A.; Arenas, A. MuxViz: A tool for multilayer analysis and visualization of networks. J. Complex Networks 2015, 3, 159–176. [Google Scholar] [CrossRef]
  37. Greene, D.; Cunningham, P. Practical solutions to the problem of diagonal dominance in kernel document clustering. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 377–384. [Google Scholar]
Figure 1. A flowchart of the proposed Mx-CRTSA approach.
Figure 1. A flowchart of the proposed Mx-CRTSA approach.
Applsci 13 02514 g001
Figure 2. Examples showing some of the simulated 2-layer multiplex networks generated with n = 128 , c = 4 and variable μ 0 . (a,b) Adjacency matrices with μ 0 = 0.1 ; (c,d) Adjacency matrices with μ 0 = 0.2 .
Figure 2. Examples showing some of the simulated 2-layer multiplex networks generated with n = 128 , c = 4 and variable μ 0 . (a,b) Adjacency matrices with μ 0 = 0.1 ; (c,d) Adjacency matrices with μ 0 = 0.2 .
Applsci 13 02514 g002
Figure 3. Examples showing some of the simulated 3-layer multiplex networks generated with n = 128 , c = 4 and variable μ 0 . (ac) Adjacency matrices with μ 0 = 0.1 ; (df) Adjacency matrices with μ 0 = 0.6 .
Figure 3. Examples showing some of the simulated 3-layer multiplex networks generated with n = 128 , c = 4 and variable μ 0 . (ac) Adjacency matrices with μ 0 = 0.1 ; (df) Adjacency matrices with μ 0 = 0.6 .
Applsci 13 02514 g003
Figure 4. Examples showing some of the simulated 2-layer multiplex networks generated with n = 128 , c = 10 and variable μ 0 . (a,b) Adjacency matrices with μ 0 = 0.1 ; (c,d) Adjacency matrices with μ 0 = 0.2 .
Figure 4. Examples showing some of the simulated 2-layer multiplex networks generated with n = 128 , c = 10 and variable μ 0 . (a,b) Adjacency matrices with μ 0 = 0.1 ; (c,d) Adjacency matrices with μ 0 = 0.2 .
Applsci 13 02514 g004aApplsci 13 02514 g004b
Figure 5. Average NMI and AR scored by the different methods for simulated multiplex networks with n = 128 generated with variable μ 0 . (a,b) Multiplex networks with c = 4 and M = 2 ; (c,d) Multiplex networks with c = 4 and M = 3 ; (e,f) Multiplex networks with c = 10 and M = 2 .
Figure 5. Average NMI and AR scored by the different methods for simulated multiplex networks with n = 128 generated with variable μ 0 . (a,b) Multiplex networks with c = 4 and M = 2 ; (c,d) Multiplex networks with c = 4 and M = 3 ; (e,f) Multiplex networks with c = 10 and M = 2 .
Applsci 13 02514 g005aApplsci 13 02514 g005b
Figure 6. The scalability of Mx-CRTSA compared with other existing methods as the number of nodes increases.
Figure 6. The scalability of Mx-CRTSA compared with other existing methods as the number of nodes increases.
Applsci 13 02514 g006
Figure 7. Sensitivity of the proposed Mx-CRTSA in terms of NMI to the regularization parameters in (a) Cora, (b) C. Elegans, (c) CKM and (d) Wikipedia networks.
Figure 7. Sensitivity of the proposed Mx-CRTSA in terms of NMI to the regularization parameters in (a) Cora, (b) C. Elegans, (c) CKM and (d) Wikipedia networks.
Applsci 13 02514 g007aApplsci 13 02514 g007b
Figure 8. Average run time (in seconds) of the different methods.
Figure 8. Average run time (in seconds) of the different methods.
Applsci 13 02514 g008
Table 1. List of notations.
Table 1. List of notations.
SymbolDescription
A , A, aTensor, matrix, vector
A F Frobenius norm of the matrix A, A F 2 = i , j a ij 2
A F 2 Frobenius norm of the tensor A , A F 2 i , j , k a ijk 2
A * Nuclear norm of A, A * = i σ i ( A )
σ i ( A ) i t h singular value of A
A ( j , : , : ) j t h horizontal slice of A
A ( : , j , : ) j t h lateral slice of A
A ( : , : , j ) j t h frontal slice of A
A ( j ) j t h frontal slice of A
A ( : , i , j ) Mode-1 fiber of A
A ( i , : , j ) Mode-2 fiber of A
A ( i , j , : ) Mode-3 fiber of A
A + The positive values of matrix A
Table 2. Comparison between the objective functions of the different methods considered in the experiments.
Table 2. Comparison between the objective functions of the different methods considered in the experiments.
MethodObjective FunctionConstraints
CentroidCoreg max U ( m ) , U * m = 1 M t r ( U ( m ) L ( m ) U ( m ) + λ m t r ( U ( m ) U ( m ) U * U * ) ) U * U * = I U ( m ) U ( m ) = I .
SCML min U m = 1 M t r ( U L ( m ) U ) λ m = 1 M t r ( U U * U ( m ) U ( m ) ) U * U * = I .
CSNMF min H A ( m ) U * U * F 2 + U U * U ( m ) U ( m ) F 2 U 0 U * U * = I .
CPNMF min H A ( m ) U * U * A ( m ) F 2 + U U * U ( m ) U ( m ) F 2 U 0 ,   U * U * = I .
CSNMTF min H A ( m ) U * S ( m ) U * F 2 + U U * U ( m ) U ( m ) F 2 U 0 ,   U * U * = I .
Mx-CRTSA min U * , U ( m ) , Z , E , w Z + γ E 2 , 1 λ m = 1 M t r U ( m ) Z N ( m ) U ( m ) + w m t r U * U ( m ) U ( m ) U * A ( m ) = A ( m ) Z ( m ) + E ( m ) , U * U * = I , U ( m ) U ( m ) = I , w 2 = 1 ,      w 0
Table 3. Description of the real-world multiplex networks.
Table 3. Description of the real-world multiplex networks.
NetworkNodesLayersClusters
NRC13638
CKM24634
C. Elegans279312
Cora29233
BBCSPORT54425
Wikipedia693210
COIL20 1440320
100Leaves16003100
Table 4. Performance comparison between the different algorithms including the proposed Co-regularized Robust Tensor-based Spectral Approach (Mx-CRTSA) for the real-world multiplex networks. The best performance is in boldface.
Table 4. Performance comparison between the different algorithms including the proposed Co-regularized Robust Tensor-based Spectral Approach (Mx-CRTSA) for the real-world multiplex networks. The best performance is in boldface.
NetworkMetricCentroidCoregPMRMSCSCMLCSNMFCPNMFCSNMTFDiMMA2CMVMx-CRTSA
NRCACC 0.426 0.419 0.455 0.441 0.416 0.439 0.384 0.375 0.412 0 . 541
γ = 0.03 NMI 0.352 0.368 0.371 0.387 0.333 0.148 0.275 0.297 0.336 0 . 544
λ = 0.008 AR 0.208 0.153 0.216 0.218 0.153 0.027 0.121 0.147 0.194 0 . 331
F-score 0.337 0.419 0.358 0.357 0.312 0.302 0.296 0.286 0.314 0 . 448
CKMACC 0.983 0.479 0.947 0.987 0.991 0.317 0.987 0.605 0.836 1 . 00
γ = 0.04 NMI 0.948 0.032 0.821 0.958 0.821 0.017 0.963 0.472 0.778 1 . 00
λ = 0.1 AR 0.975 0.010 0.825 0.980 0.851 0.009 0.979 0.215 0.794 1 . 00
F-score 0.983 0.483 0.866 0.986 0.990 0.280 0.985 0.491 0.812 1 . 00
C. ElegansACC 0.419 0.369 0.301 0.401 0.322 0.369 0.333 0.290 0.286 0 . 438
γ = 0.01 NMI 0.432 0.427 0.303 0.448 0.442 0.382 0.461 0.341 0.334 0 . 509
λ = 0.02 AR 0.183 0.217 0.124 0.184 0.219 0.157 0.157 0.095 0.077 0 . 270
F-score 0 . 433 0.332 0.266 0.402 0.275 0.313 0.293 0.207 0.201 0.362
CoraACC 0.982 0.979 0.503 0.979 0.603 0.442 0.508 0.455 0.507 1 . 00
γ = 0.04 NMI 0.981 0.906 0.528 0.906 0.476 0.123 0.447 0.354 0.441 1 . 00
λ = 0.1 AR 0.949 0.939 0.389 0.939 0.356 0.065 0.353 0.176 0.337 1 . 00
F-score 0.966 0.959 0.592 0.959 0.620 0.473 0.579 0.533 0.569 1 . 00
BBCSportACC 0.768 0.527 0.902 0.742 0.279 0.237 0.292 0.459 0.721 0 . 996
γ = 0.3 NMI 0.604 0.337 0.835 0.563 0.031 0.006 0.025 0.193 0.656 0 . 989
λ = 0.02 AR 0.552 0.179 0.853 0.494 0.004 0.001 0.011 0.088 0.599 0 . 996
F-score 0.659 0.474 0.889 0.616 0.235 0.216 0.234 0.406 0.699 0 . 997
WikipediaACC 0.565 0.551 0.546 0.568 0.574 0.216 0.572 0.341 0.398 0 . 583
γ = 0.04 NMI 0.477 0.488 0.346 0.493 0.527 0.030 0.525 0.309 0.425 0 . 573
λ = 0.5 AR 0.399 0.405 0.223 0.412 0.404 0.003 0 . 450 0.058 0.215 0.437
F-score 0.492 0.419 0.482 0.492 0.437 0.141 0.449 0.216 0.324 0 . 505
COIL20 ACC 0.698 0.551 0.732 0.681 0.648 0.096 0.662 0.657 0.633 0 . 744
γ = 0.1 NMI 0.796 0.732 0.812 0.806 0.786 0.041 0.750 0.752 0.769 0 . 831
λ = 0.1 AR 0.641 0.491 0.663 0.627 0.594 0.008 0.533 0.556 0.577 0 . 676
F-score 0.660 0.522 0.681 0.647 0.616 0.049 0.558 0.580 0.600 0 . 692
100leavesACC 0.765 0.581 0.685 0.789 0.648 0.096 0.662 0.513 0.708 0 . 816
γ = 0.06 NMI 0.902 0.813 0.856 0.905 0.786 0.041 0.750 0.673 0.882 0 . 915
λ = 0.1 AR 0.701 0.461 0.586 0.719 0.594 0.025 0.533 0.363 0.634 0 . 753
F-score 0.704 0.467 0.589 0.722 0.616 0.038 0.217 0.403 0.638 0 . 755
Table 5. Percent of improvement achieved by the proposed method over the other methods.
Table 5. Percent of improvement achieved by the proposed method over the other methods.
NetworkMetricCentroidCoregPMRMSCSCMLCSNMFCPNMFCSNMTFDiMMA2CMV
NRCACC + 11.5 % + 12.2 % + 8.6 % + 10.0 % + 12.5 % + 10.2 % + 15.7 % + 16.5 % + 12.9 %
NMI + 19.5 % + 17.6 % + 17.3 % + 15.7 % + 21.1 % + 39.6 % + 26.9 % + 24.7 % + 20.8 %
AR + 12.3 % + 17.8 % + 11.5 % + 11.3 % + 17.8 % + 30.4 % + 21.0 % + 18.4 % + 13.7 %
F-score + 11.1 % + 2.9 % + 9.0 % + 9.1 % + 13.6 % + 14.6 % + 15.2 % + 16.2 % + 13.4 %
CKMACC + 1.7 % + 52.1 % + 5.3 % + 1.3 % + 0.9 % + 68.3 % + 1.3 % + 39.5 % + 58.8 %
NMI + 5.2 % + 96.8 % + 17.9 % + 4.2 % + 17.9 % + 98.3 % + 3.7 % + 52.8 % + 22.2 %
AR + 2.5 % + 99.0 % + 17.5 % + 2.0 % + 14.9 % + 99.1 % + 2.1 % + 78.5 % + 20.6 %
F-score + 1.7 % + 51.7 % + 13.4 % + 1.4 % + 1.0 % + 72.0 % + 1.5 % + 50.9 % + 18.8 %
C. ElegansACC + 1.9 % + 6.9 % + 13.7 % + 3.7 % + 11.6 % + 6.9 % + 10.5 % + 14.8 % + 15.2 %
NMI + 7.7 % + 8.2 % + 20.6 % + 6.1 % + 6.7 % + 12.7 % + 4.8 % + 16.8 % + 17.5 %
AR + 8.7 % + 5.3 % + 14.6 % + 8.6 % + 5.1 % + 11.3 % + 11.3 % + 17.5 % + 19.6 %
F-score 7.1 % + 3.0 % + 9.6 % 4.0 % + 8.7 % + 4.9 % + 6.9 % + 15.5 % + 16.1 %
CoraACC + 1.8 % + 2.1 % + 49.7 % + 2.1 % + 39.7 % + 55.8 % + 49.2 % + 54.5 % + 49.3 %
NMI + 1.9 % + 9.4 % + 47.2 % + 9.4 % + 52.4 % + 87.7 % + 55.3 % + 64.6 % + 55.9 %
AR + 5.1 % + 6.1 % + 61.1 % + 6.1 % + 64.4 % + 96.5 % + 64.7 % + 82.4 % + 66.3 %
F-score + 3.4 % + 4.1 % + 40.8 % + 4.1 % + 38.0 % + 52.7 % + 42.1 % + 46.7 % + 43.1 %
BBCSportACC + 22.8 % + 46.9 % + 9.4 % + 25.4 % + 71.7 % + 75.9 % + 70.4 % + 53.7 % + 27.5 %
NMI + 38.5 % + 65.2 % + 15.4 % + 42.6 % + 95.9 % + 98.3 % + 96.4 % + 79.6 % + 33.3 %
AR + 44.4 % + 81.7 % + 14.3 % + 50.2 % + 99.2 % + 99.5 % + 98.5 % + 90.8 % + 39.7 %
F-score + 33.8 % + 52.3 % + 10.8 % + 38.1 % + 76.2 % + 78.1 % + 76.3 % + 59.1 % + 29.8 %
WikipediaACC + 1.8 % + 3.2 % + 3.7 % + 1.5 % + 0.9 % + 36.7 % + 1.1 % + 24.2 % + 21.5 %
NMI + 9.6 % + 8.5 % + 22.7 % + 8.0 % + 4.6 % + 54.3 % + 4.8 % + 26.4 % + 14.8 %
AR + 3.8 % + 3.2 % + 21.4 % + 2.5 % + 3.3 % + 43.4 % 1.3 % + 37.9 % + 22.2 %
F-score + 1.3 % + 8.6 % + 2.3 % + 1.3 % + 6.8 % + 36.4 % + 5.6 % + 28.9 % + 18.1 %
COIL20 ACC + 4.6 % + 19.3 % + 1.2 % + 6.3 % + 9.6 % + 64.8 % + 8.2 % + 8.7 % + 11.1 %
NMI + 3.5 % + 9.9 % + 1.9 % + 2.5 % + 4.5 % + 79.0 % + 8.1 % + 7.9 % + 6.2 %
AR + 3.5 % + 18.5 % + 1.3 % + 4.9 % + 8.2 % + 66.8 % + 14.3 % + 12.0 % + 9.9 %
F-score + 3.2 % + 17.0 % + 1.1 % + 4.5 % + 7.6 % + 64.3 % + 13.4 % + 11.2 % + 9.2 %
100leavesACC + 5.1 % + 23.5 % + 13.1 % + 2.7 % + 16.8 % + 72.0 % + 15.4 % + 30.3 % + 10.8 %
NMI + 1.3 % + 10.2 % + 5.9 % + 1.0 % + 12.9 % + 87.4 % + 16.5 % + 24.2 % + 3.3 %
AR + 5.2 % + 29.2 % + 16.7 % + 3.4 % + 15.9 % + 72.8 % + 22.0 % + 39.0 % + 11.9 %
F-score + 5.1 % + 28.8 % + 16.6 % + 3.3 % + 13.9 % + 71.7 % + 53.8 % + 35.2 % + 11.7 %
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-sharoa, E.; Al-wardat, M.; Al-khassaweneh, M.; Al Bataineh, A. Discovering Community Structure in Multiplex Networks via a Co-Regularized Robust Tensor-Based Spectral Approach. Appl. Sci. 2023, 13, 2514. https://doi.org/10.3390/app13042514

AMA Style

Al-sharoa E, Al-wardat M, Al-khassaweneh M, Al Bataineh A. Discovering Community Structure in Multiplex Networks via a Co-Regularized Robust Tensor-Based Spectral Approach. Applied Sciences. 2023; 13(4):2514. https://doi.org/10.3390/app13042514

Chicago/Turabian Style

Al-sharoa, Esraa, Mohammad Al-wardat, Mahmood Al-khassaweneh, and Ali Al Bataineh. 2023. "Discovering Community Structure in Multiplex Networks via a Co-Regularized Robust Tensor-Based Spectral Approach" Applied Sciences 13, no. 4: 2514. https://doi.org/10.3390/app13042514

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop