Next Article in Journal
Molecular-Memory-Induced Counter-Intuitive Noise Attenuator in Protein Polymerization
Previous Article in Journal
Indeterminate Stieltjes Moment Problem: Entropy Convergence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Object Part Learning Using Deep Lp Smooth Symmetric Non-Negative Matrix Factorization

1
School of Mathematical Sciences, Guizhou Normal University, Guiyang 550025, China
2
College of Mathematics and Information Science, Guiyang University, Guiyang 550005, China
3
School of Mathematical Sciences, Xiamen University, Xiamen 361005, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Symmetry 2024, 16(3), 312; https://doi.org/10.3390/sym16030312
Submission received: 4 February 2024 / Revised: 27 February 2024 / Accepted: 29 February 2024 / Published: 6 March 2024
(This article belongs to the Section Mathematics)

Abstract

:
Nowadays, deep representations have gained significant attention due to their outstanding performance in a wide range of tasks. However, the interpretability of deep representations in specific applications poses a significant challenge. For instances where the generated quantity matrices exhibit symmetry, this paper introduces a variant of deep matrix factorization (deep MF) called deep L p smooth symmetric non-negative matrix factorization (DSSNMF), which aims to improve the extraction of clustering structures inherent in complex hierarchical and graphical representations in high-dimensional datasets by improving the sparsity of the factor matrices. We successfully applied DSSNMF to synthetic datasets as well as datasets related to post-traumatic stress disorder (PTSD) to extract several hierarchical communities. Specifically, we identified non-disjoint communities within the partial correlation networks of PTSD psychiatric symptoms, resulting in highly meaningful clinical interpretations. Numerical experiments demonstrate the promising applications of DSSNMF in fields like network analysis and medicine.

1. Introduction

Non-negative matrix factorization (NMF) [1] is a valuable technique that has been widely used in various areas of feature learning, including document clustering [2,3], recognition of objects and faces [4,5], signal analysis [6,7], and community detection [8,9], among others.
In the era of artificial intelligence, extracting meaningful features from massive and complex data is a significant challenge. Consequently, matrix factorization has gained significant prominence as a technique for extracting features and reducing dimensionality in recent years. NMF is a technique used to approximately decompose a non-negative data matrix X R + n × m into the product of two non-negative factorization matrices U R + n × r and V R + r × m , with the objective of approximating X UV . Here, X represents a data matrix where each column corresponds to a data point in dimension m, and r denotes the factorization rank, indicating that each data point can be expressed as a linear combination of r basis vectors. In order to improve feature extraction and data compression, researchers have developed several variants of NMF. These variants introduce different constraints on the basis matrix U and the coefficient matrices V . Some examples of these constraints include non-negativity [8], sparsity [10], manifold constraints [4], non-smoothness [5], and orthogonality [11]. When the matrix X represents the similarity values between two data points and exhibits symmetry (i.e., X = X R n × n and V = U ), symmetric non-negative matrix factorization [12] is proposed as a suitable approach. This method is commonly applied to matrices such as word co-occurrence matrices in topic modeling [13,14] or adjacency matrices of undirected graphs. It is worth noting that while the input symmetric matrix X typically has non-negative values, it is not strictly required. Nonetheless, it is essential to highlight that the resulting factor matrices must be non-negative.
Although NMF delivers excellent results in feature extraction and data compression, researchers continue to explore the presence of hierarchical information within objects. More recently, advancements in matrix factorization techniques, such as multilayer MF [15,16,17,18] and deep MF [18], influenced by deep learning theory, have elevated the field to a new level. Deep non-negative matrix factorization (deep NMF) builds upon standard NMF and approximates it as a product of multiple factor matrices. More precisely, the matrix X undergoes L layers of factorizations resulting in an approximation represented by X such that X U L V L V L 1 V 1 . The hierarchical form of equivalence can be expressed as follows:
X U 1 V 1 , U 1 U 2 V 2 , U L 1 U L V L .
where U l R + m × r l and V l R + r l × r l 1 ( l = 1 , , L ) with r 0 = n . In scheme of (1) the dimension r l of each layer is the rank of the factorization at layer l, and  r l ( l = 1 , , L ) are assumed to be decreasing, that is, r 1 > r 2 > > r L , see [19,20] for more details.
Deep MF has the capability to conduct the hierarchical decomposition of intricate input datasets, enhancing the interpretability of the dataset. It is widely employed in various domains, including hierarchical feature extraction [21], recommender systems [17], and multi-view clustering [22,23].
In the existing literature, the combination of smoothness and symmetry in deep NMF has not been explored. We will investigate a deep NMF, namely deep L p smooth symmetric non-negative matrix factorization (DSSNMF). The main contributions and motivations for the algorithm are as follows:
  • We investigate the case where the input quantity matrix X is a symmetric matrix, such as the adjacency matrix of an undirected graph. If  X is constructed correctly, the approximate deep factorization of X generates a series of non-negative clustering assignment matrices that effectively capture the inherent clustering structure of each layer. Numerical experiments conducted on synthetic datasets with complex structures demonstrate that the algorithm provides good interpretability of the hierarchical clustering information.
  • In order to obtain smoother and more accurate solutions, we introduce the L p constraints of matrices into the framework of depth-symmetric non-negative matrix factorization and propose the DSSNMF method. The method fully preserves the smooth and more accurate optimized solution.
  • In this paper, we investigate the combination of smooth and symmetry in deep matrix factorization through the proposed DSSNMF algorithm. The motivation behind this research is to explore the hierarchical structure present in complex datasets by enhancing the sparsity of the factor matrices. Our results demonstrate the effectiveness of DSSNMF in extracting hierarchical communities both in synthetic datasets and in post-traumatic stress disorder (PTSD) datasets, specifically revealing non-disjoint communities in the PTSD psychiatric symptom partial correlation networks. Moreover, our numerical experiments highlight the promising applications of DSSNMF in fields such as network analysis and medicine.
This paper is organized into five sections. Section 2 shows the process of the construction of the DSSNMF model. The DSSNMF algorithm and its implementation conditions are presented in Section 3. To demonstrate the performance of the DSSNMF algorithm, experiments are conducted on both the PTSD dataset and the simulated dataset in Section 4. Finally, Section 5 provides a conclusion.

2. DSSNMF Modeling

In this study, matrices are represented by bold capital letters indicated earlier. For example, U or U l represents a matrix, U ( i , j ) indicates the element at position ( i , j ) in the matrix, U ( : , j ) and U ( i , : ) represent the j-th column and i-th row of the matrix, respectively. The symbols ( · ) and T r ( · ) denote the transpose and trace of the matrix, and · , · and · ¯ indicate the scalar product of two vectors and the mean of a vector, respectively.
To make the modeling process understandable, we take the example of a simple undirected graph. Let X R n × n be a symmetric adjacency matrix of a graph (see Figure 1). The element at position ( i , j ) in X represents the similarity value between data points x i and x j , which correspond to nodes. The objective of DSSNMF is to use L layers of factorizations to obtain, at each layer l, a non-negative symmetric approximation of rank r l for the original matrix X . Specifically, at the first layer, we have X U 1 U 1 , where U 1 R + n × r 1 , similar to symNMF [12]. At the second layer, we factorize the matrix U 1 as U 1 U 2 V 2 with U 2 R + n × r 2 and V 2 R + r 2 × r 1 , where r 2 < r 1 . This results in a new symmetric approximation of rank r 2 < r 1 for X , namely X U 2 V 2 V 2 U 2 . Similarly, the L-th layer is decomposed as: X U L V L V 2 V 2 V L U L , with  U l R + n × r l and V l R + r l × r l 1 , ( l = 1 , , L ), r L < r L 1 < < r 2 < r 1 . This is how the deep symmetric matrix factorization model is constructed. From the perspective of the entire decomposition process, in simple terms, the decomposition factor U 1 can be understood as comprising r 1 communities, each containing n points, and  U 1 ( i , j ) refers to the i-th point belonging to the j-th community. Furthermore, as the factorization continues and the rank decreases, the columns of U 2 can be interpreted as representing the membership degrees of the n data points in r 2 communities. The smaller communities gradually merge into larger ones as the factorization progresses.
To generate smoother and more accurate solutions at each layer of the approximate factorization, we employ a series of additional L p smoothing constraints with the goal of enhancing the accuracy of the subsequent factorizations. The interpretation of matrix factorization at each layer has two aspects: utilizing linear combinations of basis vectors enables an accurate reconstruction of the data, and the interpretation of the basis vectors is clear-cut. The rationale behind this approach is to find a method to decompose the data into mutually exclusive components, thereby enhancing the accuracy of the subsequent factorizations.
Nevertheless, deep NMF lacks consideration for the smoothing of the basis matrices in each layer of the factorization, which restricts its ability to generate solutions that are both smooth and accurate. To address this limitation, incorporating L p smoothing constraints into the deep symmetric NMF model offers a promising approach for achieving a smoother and more accurate solution to the optimization problem. The new deep NMF model is as follows:
The first layer of factorization is: X U 1 U 1 + 2 β U 1 p p , where U 1 R + n × r 1 ,
The second layer of factorization is: X U 2 V 2 V 2 U 2 + 2 β U 2 p p , where U 2 R + n × r 2 , V 2 R + r 2 × r 1 ,
Similarly, the remaining layers can be given in the same way, the L-th layer of factorization is: X U L V L V 2 V 2 V L U L + 2 β U L p p , where U L R + n × r L , V L R + r L × r L 1 .
Where β is the regularization parameter, U l p = ( Σ i , j | U L ( i , j ) | p ) 1 p , l = 1 , , L , p is smooth parameter which ranges: 1 < p < 2 . This is the model for deep L p smooth symmetric non-negative matrix factorization (DSSNMF) proposed in this paper.

3. Algorithm for DSSNMF

Given a non-negative symmetric matrix X R n × n , the standard symNMF involves solving the following optimization problem:
min U R + n × r X U U F 2 .
In order to design an effective alternating type algorithm for standard symNMF, Li, X. et al. [24] transfer this problem to a penalized asymmetric NMF in the following form:
min U R + n × r , V R + r × n X U V F 2 + γ U V F 2 .
where the penalty term γ U V F 2 is included to encourage similarity between the two factors U and V . For sufficiently large γ , it has been shown that U = V holds for the critical points of (3) [24].
Inspired by the work of [20,25], we combine the deep L p smooth symmetric matrix factorization model, as described in Section 2, with (3) to formulate a comprehensive loss function that incorporates weighted sums of the layer-wise contributions.
D S S N M F = 1 2 ( X U 1 V 1 F 2 + γ 1 U 1 V 1 F 2 + 2 β U 1 p p + λ 1 ( U 1 U 2 V 2 F 2 + γ 2 U 2 ( V 2 V 1 ) F 2 + β U 2 p p ) + + λ L 1 ( U L 1 U L V L F 2 + γ L U L ( V L V L 1 V 2 V 1 ) F 2 + 2 β U L p p ) ) .
This loss function represents the error in the execution of the symNMF at each layer, where the weighting parameter, denoted as λ l , determines the relative importance of each layer. Remarkably, this approach can be viewed as an extension of the multilayer symmetric NMF model proposed by Cichocki et al. [15]. In this extension, instead of considering the sum of errors from individual layer factorizations, we calculate the overall weighted error that accounts for the errors across all layer factorizations. The error for each layer is E r ( l ) = λ l 1 ( U l 1 U l V l F 2 + γ l U l ( V l V l 1 V 2 V 1 ) F 2 + 2 β U l p p ) for l = 1 , , L , with  U 0 = X and λ 0 = 1 . The error E r ( l ) at each layer consists of three components. The first component represents the reconstruction error between U l 1 and U l V l , capturing the quality of the rank r l approximation. The second component ensures that the factorization at each layer adheres to the symNMF framework. Lastly, the third component is the L p constraint term. Therefore, both (4) and (3) adopt a similar method to solve non-quadratic optimization subproblems. This allows us to derive iterative updating rules for the factor matrices U l and V l at each layer. These matrices are alternately optimized, resulting in a monotonically decreasing objective function.
To minimize (4), the block coordinate descent (BCD) [15,20,25] is employed, necessitating the blockwise updating of the factor matrices. In this study, we utilize the fast projected gradient method (FPGM) [26] to update the factor matrices U l or V l . FPGM is a widely recognized first-order optimization technique that can efficiently update any U l or V l , l = 1 , L , as outlined in Algorithm 1. Considering the convexity of each layer’s subproblems, we select 1 L as the step size, where L represents the Lipschitz constant. Here the symbol P U denotes the projection on the feasible set U .
Algorithm 1 Restarted fast projected gradient method (FPGM) for U
Require:  U ( 0 ) : Initial matrix, U : Feasible set, ( U ) : Loss function, α 1 ( 0 , 1 ) : Parameter
Ensure:  U that minimizes ( U ) such that ( U ) < ( U ( 0 ) )
  1: Compute Lipschitz constant L of , M = U ( 0 )
  2: for  k = 1 to N do                %Number of iterations
  3:      U ( k ) = P U ( M 1 L ( U ) )
  4:      M = U ( k ) + ε k ( U ( k ) U ( k 1 ) ) with ε k = α k ( 1 α k ) α k 2 + α k + 1 , α k + 1 = 1 2 ( α k 4 + 4 α k 2 α k 2 )
  5:     if  ( U ( k ) ) > ( U ( k 1 ) )  then
  6:           M = U ( k 1 ) , α k + 1 = α 1
Algorithm 1 requires the gradient of the loss function (4) with respect to U l and V l ( l = 1 , , L ). By introducing U 0 = X , λ 0 = γ 0 = 1 , it is straightforward to derive
D S S N M F U l = E r ( l ) U l , D S S N M F V l = E r ( l ) V l .
On the other hand, E r ( l ) can be deformed as
E r ( l ) = λ l 1 ( U l 1 U l V l F 2 + γ l U l ( V l V l 1 V 2 V 1 ) F 2 + 2 β U l p p ) = λ l 1 ( t r ( ( U l 1 U l V l ) ( U l 1 U l V l ) ) + γ l t r ( ( U l ( V l Q ) ) ( U l V l Q ) ) + 2 β U l p p ) = λ l 1 ( ( t r ( U l 1 U l 1 ) 2 t r ( U l 1 V l U l ) + t r ( U l V l V l U l ) ) + γ l ( t r ( U l U l ) 2 t r ( U l V l Q ) + t r ( Q V l V l Q ) ) + 2 β U l p p ) ,
where Q = V l 1 V 2 V 1 . If we take the derivatives with respect to U l and V l , we obtain
D S S N M F U l = λ l 1 ( ( 2 U l 1 V l + U l V l V l + U l V l V l ) + γ l ( 2 U l 2 Q V l ) + 2 p β U l p 1 ) = 2 λ l 1 ( ( U l V l V l U l 1 V l ) + γ l ( U l Q V l ) + p β U l p 1 ) D S S N M F V l = 2 λ l 1 ( ( U l U l V l U l U l 1 ) + γ l ( V l Q Q U l Q ) ) .
Next, Algorithm 2 gives the algorithm for minimizing the loss function (4) with general constraints on the factor matrices U l and V l .
Algorithm 2 DSSNMF
Input: Symmetric matrix X
Output: Hierarchical factor matrix U 1 , , U L and V 1 , , V L
  1: Choose the number of factorization layers L, factorization rank per layer r 1 , , r L , and initial matrices U l ( 0 ) and V l ( 0 ) for each layer, regularization parameter β , smooth parameter p
  2: for  k = 1 , , do
  3:     for  l = 1 , , L  do
  4:         Execute Algorithm 1 and output V l ( k )
  5:         Execute Algorithm 1 and output U l ( k )
Special attention should be given to the selection of hyperparameters, particularly the number of layers L and their factorization rank r l , which are crucial elements of the deep NMF model (refer to the first step of Algorithm 2). These hyperparameters can be manually specifled by the user if there is existing prior knowledge about the network. In line with [20], we utilize the multilayer matrix factorization (MulMF) method to initialize the hierarchical factor matrices U l ( 0 ) and V l ( 0 ) . Specifically, the L-layer deep L p smooth symmetric non-negative matrix factorization initializes the factors layer by layer, supplying the initial factor matrix before entering the DSSNMF procedure. If no a priori information about the network is available, we employ the widely recognized Louvain Method (LM) [27], which efficiently identifies modular partitions in large networks, yielding a comprehensive hierarchical community structure and providing various resolutions for community detection. After t iterations, the LM method generates a graph partition consisting of r t distinct communities, ensuring that each node belongs to only one community. This effectively extracts a bottom-up hierarchy of communities within the graph, allowing different interpretations in each iteration, similar to the MulMF algorithm. The main difference is observed in MulMF, where nodes can belong to multiple communities with varying proportions. Without prior network information to guide the selection of hyperparameters, we set the number of layers L in DSSNMF to be equal to the number of LM iterations t. The rank r l is determined by the number of layers r t successively extracted by the LM within the communities.

4. Experiments and Discussions

Deep MF has garnered significant attention in recent years due to its ability to learn hierarchical representations from complex data [23,28,29]. The incorporation of symmetry [23], non-negativity constraints [8], sparsity-constrained [30] and pairwise relationships [31] has significantly improved the ability to capture hierarchical structures and community patterns in complex networks. Additionally, the application of deep architectures has further enhanced the representation learning capabilities, leading to improved clustering and recommendation performances. These advancements open up exciting possibilities for exploring hierarchical information extraction in various domains and pave the way for future research in the field of deep MF.
In this section, we employ the DSSNMF algorithm to analyze both synthetic datasets and real psychiatric network datasets. We conduct numerical experiments to extract latent hierarchical features, aiming to uncover previously unknown hierarchical structures, despite the absence of any prior knowledge. This task presents considerable challenges.

4.1. Synthetic Dataset

From the given loss function 4, it is evident that setting the parameters λ l ( l = 1 , , L 1 ) , γ l ( l = 1 , , L ) and β is necessary. These parameters play a crucial role in balancing the significance of each layer and facilitating the progression to the subsequent step. Specifically, λ l is chosen such that E r ( 0 ) ( l ) is equal for all l > 0 , that is λ l = E r ( 0 ) ( l ) E r ( 0 ) ( l + 1 ) . Similarly, the γ l values are determined so that E r 1 ( 0 ) ( l ) = E r 2 ( 0 ) ( l ) for all l, which implies γ l = E r 1 ( 0 ) ( l ) E r 2 ( 0 ) ( l ) .
For our subsequent experiments, we will employ the following comparison algorithm:
  • DSNMF: Deep Symmetric Matrix Factorization, which is discussed in detail in [25].
  • Multilayer NMF (MulNMF): This method involves multilayer non-negative matrix factorization, as described in [16]. To ensure a fair comparison, we limit the input matrix to a symmetric matrix.
  • Fuzzy k-means (FKM) [32] applied at each layer: This approach is a variation of the popular k-means clustering algorithm. In traditional k-means, each data point is assigned to a single cluster. However, in this variant, data points can be assigned to multiple clusters with varying degrees of membership.
Next, we consider a simple undirected graph, such as the one depicted in Figure 1, to form our dataset. Specifically, we choose the parameters L = 2 , r 1 = 4 and r 2 = 2 . In this noise-free graph, there are two disjoint subgraphs of equal size. Each subgraph comprises two networks of the same size, with both networks sharing a common set of nodes. For instance, when considering n = 18 , n * = 1 , we obtain the scenario ilustrated in Figure 1. To add some randomness to the data, we incorporate symmetric white noise into the noiseless adjacency X ¯ . This is carried out by adding a small value ϵ as the standard deviation. By doing so, we create a noisy data matrix.
X = m a x ( 0 , X ¯ + ϵ X ¯ F N N F ) ,
where matrix N is symmetric, and its elements are randomly generated from a standard normal distribution.
We generated 10 random noise matrices, denoted for different levels of noise ϵ and combinations of n and n * . These matrices were used in running DSSNMF and three other comparison algorithms (DSNMF, MulNMF, and FKM) with different values of p. To assess the results, we employed the mean removed spectral angle (MRSA) as the performance metric. The MRSA measures the average angular difference between the columns of the ground truth and computed factor matrices U l , which are aligned to minimize differences. This calculation was performed over t runs. The MRSA between two vectors x and y can be calculated as follows:
M R S A ( x , y ) = 100 π arccos ( < x x ¯ , y y ¯ > x x ¯ 2 y y ¯ 2 ) .
We’re easy to obtain 0 M R S A ( x , y ) 100 .
In Table 1 and Table 2, we investigate the impact of varying the value of p on different combinations of n and n * while maintaining a fixed noise level ( ϵ ) of 0.01, 0.05, 0.1, and 0.4, respectively. The results demonstrate that the MRSA values remain consistently stable within the range of p between 1.1 and 1.9. This finding suggests that our DSSNMF algorithm is robust to noise when the parameter β is appropriately fixed. Additionally, Table 3 examines the effect of different β values on the MRSA values under the same noise level. The findings reveal a continuous increase in the MRSA value as the range of β expands from 0.001 to 10.
In Table 4 and Table 5, the experimental evaluation of the decomposition algorithm is presented, with the primary objective of assessing its performance. The quality of the decomposition is measured using the MRSA metric, which allows for an objective evaluation. Furthermore, a comparative analysis with other well-known algorithms is included, adding depth to the assessment. Overall, these comprehensive findings contribute to a better understanding of the effectiveness and reliability of decomposition algorithms in this context. The numerical results of each visualization are shown in Figure 2.
Transitioning to Table 4, we evaluate the MRSA of the first layer factorization, comparing the performance of DSSNMF, DSNMF, MulNMF, and FKM. Notably, DSSNMF, DSNMF, and MulNMF outperform FKM, illustrating the effectiveness of our proposed algorithm. The results obtained from these three methods are comparable, with MulNMF expected to perform better theoretically due to its independent optimization of the first layer, a result that is confirmed by our experimental findings. However, even though DSSNMF and DSNMF also consider the factorization of the second layer, both methods demonstrate promising experimental results, with DSSNMF exhibiting a slight advantage over DSNMF. Additionally, DSSNMF offers the benefit of further enhancing the MRSA by leveraging the parameter β .
Surprisingly, as shown in Table 5, which presents the results of the second layer factorization, DSSNMF and DSNMF excel at extracting internal features and capturing the hierarchical structure through deep factorization techniques, surpassing other algorithms. In contrast, MulNMF performs poorly, exposing the limitations of standalone layer-by-layer factorization. On the other hand, FKM achieves the best results in MRSA. However, it is important to acknowledge that FKM solely operates on the original data matrix, disregarding the hierarchical information embedded within the dataset. In contrast, as the number of layers of decomposition increases, DSSNMF outperforms DSNMF in extracting internal features and capturing the internal hierarchical structure of the dataset.
Overall, the results from these tables emphasize the effectiveness of our proposed DSSNMF algorithm in achieving robustness against noise, superior performance in comparison to existing methods, and its ability to uncover the hierarchical nature of the dataset.

4.2. Post-Traumatical Stress Disorder (PTSD)

Network analysis has recently gained traction in computational psychiatry and psychology, particularly in studying mood, attitudes, and psychological disorders. The existing literature in psychology offers explanations for certain networks [33,34]. Additionally, researchers have extensively discussed biased correlation and partial correlation networks as effective models for estimating psychological datasets. In our study, we work with an original data matrix that consists of ratings from m subjects on n symptoms denoted as Y R + m × n . Here, the number of rows represents the number of patients, while the number of columns represents the number of symptoms. Consequently, the matrix Y captures the assessment data for m patients with regard to n symptoms. The matrix of partial correlation coefficients for the n patients is X R + n × n and its elements are calculated as X ( i , j ) = K ( i , j ) K ( i , i ) K ( j , j ) , here, K denotes the inverse of the variance–covariance matrix Σ , i.e., K = Σ 1 . For a detailed explanation of this calculation and its underlying principles, please refer to [34].
Considering a graph represented by the adjacency matrix X , where each node is a symptom, it is important to identify the community to which each symptom belongs. This analysis would enable us to evaluate how symptoms respond to interventions like medication and examine the interconnectedness among the symptoms. However, existing research in this area has focused on extracting only one layer of disjoint communities, limiting our understanding of the intricate interactions within the network. To address this limitation, we aim to explore multiple layers of community structure to gain a more comprehensive understanding of symptom interactions.
Next, we examined a dataset of 359 women with post-traumatic stress disorder (PTSD) as assessed by the PTSD Symptom Self Rating Scale (PSS-SR) [35,36]. The scale consists of 17 items that diagnose PTSD according to the DSM-IV criteria [37] (the fifth edition is recorded as DSM-5 [38]) and measures the severity of PTSD symptoms. This scale is widely used in the field of mental health to classify and assess various mental disorders. The 17 symptoms are categorized into three clusters: the Reexperiencing cluster (including Intrusive thoughts, Nightmares, Flashbacks, and Emotionally upset), the Avoidance cluster (including Avoid thoughts and feelings, Avoid places and activities, Psychogenic amnesia, Loss of interest, Detached from others, Restricted affect, and Foreshortened sense of future), and the Arousal cluster (including Sleep disturbance, Iritbility, Diffculty concentrating, Hyperalertness, Increased startle, and Physical reactivity). Each item corresponds to the frequency of specific behaviors considered characteristic of the pathology.
The symptom network for the PTSD dataset was constructed in R, and regularized estimation of a partial correlation network was performed using extended bayesian information criterion (EBIC) [39] selection. The full R code for this analysis can be found in the supplementary material to the literature [36]. The visualized symptom network is shown in Figure 3.
The DSSNMF algorithm was initialized using the LM method and applied to the symptom network. The algorithm successfully extracted three layers of communities from the network. In the first layer, four communities were identified, followed by three communities in the second layer, and two communities in the third layer. The parameter values used for DSSNMF were as follows: L = 3 , r 1 = 4 , r 2 = 3 , and r 3 = 2 . To demonstrate the effectiveness of the algorithm, Table 6 presents the three-layer communities extracted by DSSNMF when the parameter β is set to 0.05. Similarly, Table 7 shows the three-layer communities obtained when the parameter β is set to 0.5. For the sake of simplicity, each node was assigned to a major cluster, thereby facilitating the understanding of how symptoms are grouped within the PTSD dataset. It is important to note that each node could belong to one or more communities based on its connections and interactions within the network.
The assignment of nodes to communities in the DSSNMF algorithm is based on the membership degrees of the nodes in the factor matrices. The largest membership degree determines the initial assignment of a node to a community. Subsequently, if the membership degree of a node in a new community is at least 60% of its membership degree in the latest assigned community, it is assigned to the new community as well.
By observing Table 6 and Table 7, it can be seen that changing the parameter β leads to changes in the communities to which some nodes belong. For example, node 4 in the first layer of the community changed when the parameter β changed from 0.05 to 0.5. This indicates that the parameter β affects the proportions of this node in the factor matrix.
These changes provide preliminary insights into the effect of this symptom (Take node 4 as an example) on other parts of the network. However, further analysis and interpretation are required to fully understand the implications of these changes and their significance in relation to the overall network structure.
The changes in the community assignments of nodes when varying the parameter β provide insights into the underlying structure and relationships within the network. By observing the shifting memberships of nodes across different communities, we can discern how certain symptoms are related to or interact with others.
For example, in Table 6, when β = 0.05, node 12 appears in communities C2, C3, and C4, while node 13 is present in communities C1, C2, C3, and C4 in the first layer. This suggests that these symptoms (represented by nodes 12 and 13) have connections and overlaps with multiple symptom clusters. The change in the community assignments indicates that they may play a role in different aspects of the disorder or have varied associations with other symptoms depeninding on the value of β .
These findings highlight the complex nature of the symptom network and provide preliminary insights into the potential interactions and relationships between symptoms. By examining the hierarchical information of the community structure, we gain a deeper understanding of how symptoms are grouped together and how they relate to each other within the PTSD dataset.
Indeed, the ability of DSSNMF to classify symptoms into distinct clusters and extract symptoms that are closely related to multiple clusters is valuable for understanding the interplay and relationships between different symptom clusters within the PTSD dataset. This information can provide insights into the underlying mechanisms and dynamics of the disorder, helping researchers and clinicians better understand the complex interactions between symptoms and potentially guide more targeted interventions and treatment approaches.
Furthermore, the extracted communities align with the transition from DSM-IV-TR to DSM-IV, demonstrating the consistency and interpretability of the hierarchical information obtained through the DSSNMF algorithm. This indicates the practical significance of the algorithm in providing meaningful clinical interpretations and potential applications in the psychological domain.
Overall, the application of the DSSNMF algorithm in this study contributes to the field of deep NMF and provides valuable insights into the hierarchical structure of datasets generated in fields such as network analysis and psychology. Further research can explore the potential of hierarchical community extraction in other domains, which presents a fascinating avenue for future investigations.

5. Conclusions

Deep matrix factorization (deep MF) methods have gained significant attention in recent years due to their capability to hierarchically extract human understandable features. When dealing with high-dimensional and noisy data, data analysis and processing can become a challenging task, making methods that enhance interpretability highly appreciated. Therefore, this paper aims to enhance the deep MF algorithm’s capability to extract information about the intrinsic hierarchical structure of datasets generated in fields such as network analysis and psychology, where the data matrices are symmetric non-negative matrices. To achieve this, we propose a deep L p smooth symmetric non-negative matrix factorization (DSSNMF) algorithm. Experimental results on the PTSD dataset and the synthetic dataset for network analysis demonstrate that the DSSNMF algorithm successfully achieves this goal. The L p smoothing condition imposed by the algorithm improves the noise immunity of the algorithm. Furthermore, we extracted non-disjoint communities in a network of psychiatric symptoms in PTSD, which resulted in meaningful clinical interpretations. Exploring the potential of hierarchical community extraction in the psychological domain could be a fascinating avenue for future research.

Author Contributions

Conceptualization, S.L. and C.S.; methodology, S.L. and L.L.; software, C.S. and Z.C.; writing—original draft preparation, S.L.; writing—review and editing, C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the National Natural Science Foundation of China under Grant 12161020, 12061025, and partially funded by the Natural Science Foundation of Educational Commission of Guizhou Province under Grant Qian-Jiao-He KY Zi [2021]298, Guizhou Provincial Basis Research Program (Natural Science) (QKHJC-ZK[2023]YB245, QKHJC [2020]1Z002).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors are grateful to the editor and reviewers for their valuable review comments on our work. In addition, we would like to express our special thanks to the Science and Technology Department of Guizhou Province for listing the research topic of this thesis, “Application Research of Deep Nonnegative Matrix Factorization in Hyperspectral Analysis and Image Recognition”, as a Guizhou Provincial Basis Research Program (Natural Science) starting from 2024 (Project Announcement No. 627 in the list, Applicant and Director: Shunli Li).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, D.; Seung, H.S. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems 13; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  2. Xu, W.; Gong, X.L.; Gong, Y. Document clustering based on non-negative matrix factorization. In Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Toronto, ON, Canada, 28 July–1 August 2003; ACM: New York, NY, USA, 2003; pp. 267–273. [Google Scholar]
  3. Shahnaz, F.; Berry, M.W.; Pauca, V.P.; Plemmons, R.J. Document clustering using nonnegative matrix factorization. Inf. Process. Manage. 2006, 42, 373–386. [Google Scholar] [CrossRef]
  4. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar]
  5. Pascual-Montano, A.; Carazo, J.M.; Kochi, K.; Lehmann, D.; Pascual-Marqui, R.D. Nonsmooth nonnegative matrix factorization (nsnmf). IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 403–415. [Google Scholar] [CrossRef]
  6. Fu, X.; Huang, K.; Sidiropoulos, N.D.; Ma, W.K. Nonnegative matrix factorization for signal and data analytics: Identifiability, algorithms, and applications. IEEE Signal. Proc. Mag. 2019, 36, 59–80. [Google Scholar] [CrossRef]
  7. Vaswani, N.; Bouwmans, T.; Javed, S.; Narayanamurthy, P. Robust subspace learning: Robust PCA, robust subspace tracking, and robust subspace recovery. IEEE Signal. Proc. Mag. 2018, 35, 32–55. [Google Scholar] [CrossRef]
  8. Liu, Z.; Yuan, G.; Luo, X. Symmetry and nonnegativity-constrained matrix factorization for community detection. IEEE/CAA J. Autom. Sin. 2022, 9, 1691–1693. [Google Scholar] [CrossRef]
  9. Shi, X.; Lu, H.; He, Y.; He, S. Community detection in social network with pairwise constrained symmetric non-negative matrix factorization. In Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Paris, France, 25–28 August 2015; pp. 541–546. [Google Scholar]
  10. Li, S.; Lu, L.Q.; Chen, Z. Graph-regularized, sparsity-constrained non-negative matrix factorization with earth mover’s distance metric. Mathematics 2023, 11, 1894. [Google Scholar] [CrossRef]
  11. Pan, J.; Ng, M.K.; Liu, Y.; Zhang, X.; Yan, H. Orthogonal nonnegative Tucker decomposition. SIAM J. Sci. Comput. 2021, 43, 55–81. [Google Scholar] [CrossRef]
  12. Kuang, D.; Ding, C.; Park, H. Symmetric nonnegative matrix factorization for graph clustering. In Proceedings of the 2012 SIAM International Conference on Data Mining, Anaheim, CA, USA, 26–28 April 2012; pp. 106–117. [Google Scholar]
  13. Huang, K.; Fu, X.; Sidiropoulos, N.D. Anchor-free correlated topic modeling: Identifiability and algorithm. In Proceedings of the Advances in Neural Information Processing Systems 29, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  14. Wang, J.; Zhang, X.L. Deep NMF topic modeling. Neurocomputing 2023, 515, 157–173. [Google Scholar] [CrossRef]
  15. Cichocki, A.; Zdunek, R. Multilayer nonnegative matrix factorisation. Electron. Lett. 2006, 42, 947. [Google Scholar] [CrossRef]
  16. Cichocki, A.; Zdunek, R. Multilayer nonnegative matrix factorization using projected gradient approaches. Int. J. Neural Syst. 2007, 17, 431–446. [Google Scholar] [CrossRef]
  17. Zhao, H.; Ding, Z.; Fu, Y. Multi-view clustering via deep matrix factorization. Mathematics 2017, 31. [Google Scholar] [CrossRef]
  18. Trigeorgis, G.; Bousmalis, K.; Zafeiriou, S.; Schuller, B.W. A deep matrix factorization method for learning attribute representations. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 417–429. [Google Scholar] [CrossRef]
  19. Chen, W.-S.; Zeng, Q.; Pan, B. A survey of deep nonnegative matrix factorization. Neurocomputing 2022, 491, 305–320. [Google Scholar] [CrossRef]
  20. De Handschutter, P.; Gillis, N. A consistent and flexible framework for deep matrix factorizations. Pattern Recognit. 2023, 134, 109102. [Google Scholar] [CrossRef]
  21. Smith, E.A.J. Hierarchical feature extraction through deep matrix factorization. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2356–2369. [Google Scholar]
  22. Xue, H.-J.; Dai, X.; Zhang, J.; Huang, S.; Chen, J. Deep matrix factorization models for recommender systems. IJCAI 2017, 17, 3203–3209. [Google Scholar]
  23. Hajiveiseh, A.; Seyedi, S.A.; Tab, F.A. Deep asymmetric nonnegative matrix factorization for graph clustering. Pattern Recognit. 2024, 148, 110179. [Google Scholar] [CrossRef]
  24. Li, X.; Zhu, Z.; Li, Q.; Liu, K. A provable splitting approach for symmetric nonnegative matrix factorization. IEEE Trans. Knowl. Data Eng. 2021, 35, 2206–2219. [Google Scholar] [CrossRef]
  25. De Handschutter, P.; Gillis, N.; Blekic, W. Deep symmetric matrix factorization. Proc. IEEE 2023, 635–639. [Google Scholar]
  26. Nesterov, Y.E. A method for solving the convex programming problem with convergence rate O(1/k2). Dokl. Akad. Nauk SSSR 1983, 269, 543–547. [Google Scholar]
  27. Blondel, V.D.; Guillaume, J.-L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 2008, 10008. [Google Scholar] [CrossRef]
  28. Luo, X.; Liu, Z.; Jin, L.; Zhou, Y.; Zhou, M. Symmetric nonnegative matrix factorization-based community detection models and their con-vergence analysis. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 1203–1215. [Google Scholar] [CrossRef]
  29. Luo, X.; Shang, M. Symmetric non-negative latent factor models for undirected large networks. IJCAI 2017, 2435–2442. [Google Scholar]
  30. Feng, X.-R.; Li, H.-C.; Li, J.; Du, Q.; Plaza, A.; Emery, W.J. Hyperspectral unmixing using sparsity-constrained deep nonnegative matrix factorization with total variation. IEEE Trans. Geosci. Remote. Sens. 2018, 56, 6245–6257. [Google Scholar] [CrossRef]
  31. Zhang, W.; Zhang, X.; Wang, H.; Chen, D. A deep variational matrix factorization method for recommendation on large scale sparse dataset. Neurocomputing 2019, 334, 206–218. [Google Scholar] [CrossRef]
  32. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Springer Science & Business Media: Berlin, Germany, 2013. [Google Scholar]
  33. Costantini, G.; Epskamp, S.; Borsboom, D.; Perugini, M.; Mottus, R.; Waldorp, L.J.; Cramer, A.O. State of the art personality research: A tutorial on network analysis of personality data in R. J. Res. Personal. 2015, 54, 13–29. [Google Scholar] [CrossRef]
  34. Epskamp, S.; Fried, E.I. A tutorial on regularized partial correlation networks. Psychol. Methods 2018, 23, 617. [Google Scholar] [CrossRef]
  35. Foa, E.B.; Riggs, D.S.; Dancu, C.V.; Rothbaum, B.O. Reliability and validity of a brief instrument for assessing post-traumatic stress disorder. J. Trauma. Stress 1993, 6, 459–473. [Google Scholar]
  36. Armour, C.; Fried, E.I.; Deserno, M.K.; Tsai, J.; Pietrzak, R.H. A network analysis of DSM-5 posttraumatic stress disorder symptoms and correlates in US military veterans. J. Anxiety Disord. 2017, 45, 49–59. [Google Scholar] [CrossRef]
  37. Segal, D.L. Diagnostic and Statistical Manual of Mental Disorders (DSM-IV-TR). Corsini Encycl. Psychol. 2010, 1–3. [Google Scholar]
  38. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders: DSM-5, 5th ed.; American Psychiatric Association: Washington, DC, USA, 2013. [Google Scholar]
  39. Chen, J.; Chen, Z. Extended Bayesian information criteria for model selection with large model spaces. Biometrika 2008, 95, 759–771. [Google Scholar] [CrossRef]
Figure 1. Simple undirected graphs with two layers of communities when r 1 = 4 and r 2 = 2 . The first layer of communities corresponds to the set of nodes surrounded by thin blue solid line ellipses. The second layer of communities corresponds to the set of nodes surrounded by the red thick dashed line ellipse.
Figure 1. Simple undirected graphs with two layers of communities when r 1 = 4 and r 2 = 2 . The first layer of communities corresponds to the set of nodes surrounded by thin blue solid line ellipses. The second layer of communities corresponds to the set of nodes surrounded by the red thick dashed line ellipse.
Symmetry 16 00312 g001
Figure 2. Take ( n , n * , ϵ ) = ( 50 , 5 , 0.05 ) as an example. MRSA visualization plots of DSSNMF and the three comparison algorithms run 20 times on synthetic data when r 1 = 4 , r 2 = 2 , β = 0.02 and p = 1.4 . The numerical results are shown in Figure 1. The x-axis of Figure 2 refers to the s-th experiments. (a) The first layer of factorization. (b) The second layer of factorization.
Figure 2. Take ( n , n * , ϵ ) = ( 50 , 5 , 0.05 ) as an example. MRSA visualization plots of DSSNMF and the three comparison algorithms run 20 times on synthetic data when r 1 = 4 , r 2 = 2 , β = 0.02 and p = 1.4 . The numerical results are shown in Figure 1. The x-axis of Figure 2 refers to the s-th experiments. (a) The first layer of factorization. (b) The second layer of factorization.
Symmetry 16 00312 g002
Figure 3. The symptom network for the PTSD dataset. This network containing the 17 DSM-IV symptoms of PTSD, and the thickness and brightness of an edge indicate the association strength.
Figure 3. The symptom network for the PTSD dataset. This network containing the 17 DSM-IV symptoms of PTSD, and the thickness and brightness of an edge indicate the association strength.
Symmetry 16 00312 g003
Table 1. The effect of varying the value of p on the MRSA values at different noise levels when n = 100 and n * = 10 , fixing β = 0.02.
Table 1. The effect of varying the value of p on the MRSA values at different noise levels when n = 100 and n * = 10 , fixing β = 0.02.
The first layer
( n , n * , ϵ ) p p = 1.1 p = 1.3 p = 1.5 p = 1.7 p = 1.9
(100, 10, 0.01)0.0521 ± 0.00280.0517 ± 0.00290.0519 ± 0.00280.0519 ± 0.00280.0519 ± 0.0028
(100, 10, 0.05)0.2639 ± 0.02030.2637 ± 0.01950.2609 ± 0.01520.2615 ± 0.01560.2643 ± 0.0166
(100, 10, 0.1)0.5440 ± 0.03350.5387 ± 0.04020.5340 ± 0.03380.5325 ± 0.03720.5355 ± 0.0428
(100, 10, 0.4)2.7410 ± 0.09882.6969 ± 0.10232.6784 ± 0.11412.7084 ± 0.12412.7384 ± 0.1200
The second layer
( n , n * , ϵ ) p p = 1.1 p = 1.3 p = 1.5 p = 1.7 p = 1.9
(100, 10, 0.01)0.0472 ± 0.00460.0473 ± 0.00450.0475 ± 0.00470.0479 ± 0.00430.0521 ± 0.0176
(100, 10, 0.05)0.3096 ± 0.08030.2978 ± 0.05790.2873 ± 0.04740.2881 ± 0.04800.3123 ±  0.0571
(100, 10, 0.1)0.5628 ± 0.11150.5660 ± 0.10940.5827 ± 0.14450.6472 ± 0.22830.6099 ± 0.1979
(100, 10, 0.4)2.8609 ± 0.75632.8609 ± 0.75632.8609 ± 0.75632.8609 ± 0.75632.8609 ± 0.7563
Table 2. The effect of varying the value of p on the MRSA values at different noise levels when n = 18 and n * = 1 , fixing β = 0.02.
Table 2. The effect of varying the value of p on the MRSA values at different noise levels when n = 18 and n * = 1 , fixing β = 0.02.
The first layer
( n , n * , ϵ ) p p = 1.2 p = 1.4 p = 1.6 p = 1.8
(18, 1, 0.01)0.1069 ± 0.01290.1043 ±  0.01320.1062 ± 0.01440.1049  ± 0.0131
(18, 1, 0.05)0.4734 ± 0.05780.4734  ± 0.06130.4758 ± 0.06270.4767 ± 0.0620
(18, 1, 0.1)1.0336 ± 0.12841.0330 ± 0.12801.0312 ± 0.12771.0338 ± 0.1261
(18, 1, 0.4)4.3235 ± 0.48124.3094 ± 0.49164.3201 ± 0.49114.3187 ± 0.4845
The second layer
( n , n * , ϵ ) p p = 1.2 p = 1.4 p = 1.6 p = 1.8
(18, 1, 0.01)1.5578 ± 6.35311.5455 ± 6.31131.5619 ±  6.34531.5616 ± 6.3286
(18, 1, 0.05)5.9477 ± 9.18885.1045 ± 8.40585.9908 ± 9.31665.9533 ±  8.9989
(18, 1, 0.1)4.8247 ± 8.81234.8180 ± 8.73124.8381 ± 8.80095.0625 ± 8.8486
(18, 1, 0.4)20.7649 ± 14.569520.931 ± 13.545821.2376 ± 13.999221.3150 ± 13.5245
Table 3. The effect of varying the value of β on the MRSA values at different noise levels when n = 50 and n * = 5 , fixing p = 1.4.
Table 3. The effect of varying the value of β on the MRSA values at different noise levels when n = 50 and n * = 5 , fixing p = 1.4.
The first layer
( n , n * , ϵ ) β β = 0.001 β = 0.01 β = 0.1 β = 1 β = 10 β = 100
(50, 5, 0.01)0.0709 ± 0.00370.0713 ± 0.00380.0722 ± 0.00380.0725 ± 0.00400.1119 ± 0.11940.2001 ± 0.4748
(50, 5, 0.05)0.3615 ± 0.02320.3618 ± 0.02360.3620 ± 0.02320.3734 ± 0.02590.5007 ± 0.16060.6832 ± 0.5689
(50, 5, 0.1)0.7388 ± 0.04590.7374 ± 0.04710.7369 ± 0.04780.7409 ± 0.04490.8479 ± 0.16691.5642 ± 2.0583
(50, 5, 0.4)3.1757 ± 0.17793.1794 ± 0.18113.1932 ± 0.18953.3092 ± 0.18373.4730 ± 0.18583.3764 ± 0.2636
The second layer
( n , n * , ϵ ) β β = 0.001 β = 0.01 β = 0.1 β = 1 β = 10 β = 100
(50, 5, 0.01)0.0642 ± 0.00620.0686 ± 0.02140.0768 ± 0.05990.0875 ± 0.08240.2264 ± 0.28700.4803 ± 0.6225
(50, 5, 0.05)0.3431 ± 0.04810.3228 ± 0.04810.3425 ± 0.04850.4626 ± 0.16910.5800 ± 0.48812.3267 ± 4.3214
(50, 5, 0.1)0.7351 ± 0.18280.6850 ± 0.09070.6934 ± 0.09010.7068 ± 0.11401.3429 ± 1.15593.2424 ± 4.6656
(50, 5, 0.4)3.6223 ± 1.04873.3539 ± 1.01013.4184 ± 0.77793.3132 ± 0.88883.7074 ± 0.83935.9160 ± 7.1708
Table 4. The MRSA of DSSNMF, DSNMF, MSNMF, and FKM on synthetic data was evaluated over 20 runs. The average and standard deviation of MRSA were compared with varying noise levels ϵ and network configurations, while keeping r 1 = 4 , β = 0.02 and p = 1.4 at the first layer.
Table 4. The MRSA of DSSNMF, DSNMF, MSNMF, and FKM on synthetic data was evaluated over 20 runs. The average and standard deviation of MRSA were compared with varying noise levels ϵ and network configurations, while keeping r 1 = 4 , β = 0.02 and p = 1.4 at the first layer.
( n , n * , ϵ ) ∖MethodsDSSNMFDSNMFMulNMFFKM
(18, 1, 0.01)0.1043 ± 0.01320.1038 ± 0.01310.1045 ± 0.01380.5239 ± 0.0160
(18, 1, 0.05)0.4734 ± 0.06130.4739 ± 0.05920.4727 ± 0.05940.7323 ± 0.0905
(18, 1, 0.1)1.0330 ± 0.12801.0329 ± 0.12761.0276 ± 0.12041.2282 ± 0.1601
(18, 1, 0.4)4.3049 ± 0.49164.3054 ± 0.48504.0792 ± 0.52274.5117 ± 0.5069
(50, 5, 0.01)0.0713 ± 0.00380.0709 ± 0.00370.0662 ± 0.00411.1242 ± 0.0068
(50, 5, 0.05)0.3612 ± 0.02330.3614 ± 0.02310.3359 ± 0.02501.1812 ± 0.0280
(50, 5, 0.1)0.7370 ± 0.04710.7375 ± 0.04640.6682 ± 0.04691.3218 ± 0.0625
(50, 5, 0.4)3.1759 ± 0.18103.1757 ± 0.17792.7132 ± 0.17863.1948 ± 0.1565
(100, 10, 0.01)0.0516 ± 0.00300.0522 ± 0.00280.0451 ± 0.00241.1249 ± 0.0033
(100, 10, 0.05)0.2602 ± 0.01540.2611 ± 0.01510.2301 ± 0.01121.1482 ± 0.0263
(100, 10, 0.1)0.5317 ± 0.03550.5441 ± 0.03350.4642 ± 0.02571.2321 ± 0.0435
(100, 10, 0.4)2.6838 ± 0.11252.7448 ± 0.10101.9678 ± 0.07632.4737 ± 0.1056
Table 5. The MRSA of DSSNMF, DSNMF, MSNMF, and FKM on synthetic data was evaluated over 20 runs. The average and standard deviation of MRSA were compared with varying noise levels ϵ and network configurations, while keeping r 2 = 2 , β = 0.02 and p = 1.4 at the second layer.
Table 5. The MRSA of DSSNMF, DSNMF, MSNMF, and FKM on synthetic data was evaluated over 20 runs. The average and standard deviation of MRSA were compared with varying noise levels ϵ and network configurations, while keeping r 2 = 2 , β = 0.02 and p = 1.4 at the second layer.
( n , n * , ϵ ) ∖MethodsDSSNMFDSNMFMulNMFFKM
(18, 1, 0.01)1.5455 ± 6.31131.5700 ± 6.34112.6463 ± 6.23550.1949 ± 0.0639
(18, 1, 0.05)5.1045 ± 8.40585.9290 ± 9.17479.9505 ± 10.03110.9365 ± 0.2825
(18, 1, 0.1)4.8180 ± 8.73124.8669 ± 8.828220.4606 ± 12.53112.0169 ± 0.6467
(18, 1, 0.4)20.932 ± 13.545821.1335 ± 14.572330.7843 ± 11.259411.0709 ± 5.7227
(50, 5, 0.01)0.0638 ± 0.00570.0645 ± 0.00638.3818 ± 3.95660.0712 ± 0.0074
(50, 5, 0.05)0.3208 ± 0.04900.3434 ± 0.04790.8620 ± 0.47400.3596 ± 0.0317
(50, 5, 0.1)0.6812 ± 0.09170.7218 ± 0.12177.1485 ± 4.06790.7038 ± 0.0856
(50, 5, 0.4)3.3458 ± 0.91423.6222 ± 1.048725.3546 ± 11.25023.1399 ± 0.4790
(100, 10, 0.01)0.0464 ± 0.00430.0472 ± 0.00477.6693 ± 4.83420.0504 ± 0.0048
(100, 10, 0.05)0.2880 ± 0.04720.2966 ± 0.05222.2166 ± 1.21000.2506 ± 0.0142
(100, 10, 0.1)0.5527 ± 0.17480.5630 ± 0.11131.5343 ± 1.07680.5023 ± 0.0317
(100, 10, 0.4)2.8309 ± 0.75632.8612 ± 0.755126.3690 ± 10.22892.1006 ± 0.1639
Table 6. Hierarchical communities extracted by DSSNMF when β = 0.05.
Table 6. Hierarchical communities extracted by DSSNMF when β = 0.05.
β = 0.05 First LayerSecond LayerThird Layer
C1{1, 3, 4, 13, 14}{1, 3, 4}{3, 4, 5, 6, 7, 8, 9, 10, 11}
C2{2, 10, 12, 13, 14, 15, 16, 17}{1, 2, 10, 12, 13, 14, 15, 16, 17}{1, 2, 3, 10, 12, 13, 14, 15, 16, 17}
C3{5, 6, 7, 8, 9, 10, 11, 12, 13}{5, 6, 7, 8, 9, 10, 11}
C4{12, 13, 14,17}
Table 7. Hierarchical communities extracted by DSSNMF when β = 0.5.
Table 7. Hierarchical communities extracted by DSSNMF when β = 0.5.
β = 0.5 First LayerSecond LayerThird Layer
C1{1, 3, 4, 13, 14}{1, 3, 4}{ 3, 4, 5, 6, 7, 8, 9, 10, 11}
C2{2, 12, 13, 14, 15, 16, 17}{1, 2, 3, 10, 12, 13, 14, 15, 16, 17}{1, 2, 3, 10, 12, 13, 14, 15, 16, 17}
C3{4, 5, 6, 7, 8, 9, 10, 11, 12, 13}{1, 3, 4, 5, 6, 7, 8, 9, 10, 11}
C4{12, 13, 14}
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Song, C.; Lu, L.; Chen, Z. Hierarchical Object Part Learning Using Deep Lp Smooth Symmetric Non-Negative Matrix Factorization. Symmetry 2024, 16, 312. https://doi.org/10.3390/sym16030312

AMA Style

Li S, Song C, Lu L, Chen Z. Hierarchical Object Part Learning Using Deep Lp Smooth Symmetric Non-Negative Matrix Factorization. Symmetry. 2024; 16(3):312. https://doi.org/10.3390/sym16030312

Chicago/Turabian Style

Li, Shunli, Chunli Song, Linzhang Lu, and Zhen Chen. 2024. "Hierarchical Object Part Learning Using Deep Lp Smooth Symmetric Non-Negative Matrix Factorization" Symmetry 16, no. 3: 312. https://doi.org/10.3390/sym16030312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop