Next Article in Journal
Diagnostic and Therapeutic Strategies for Fluoropyrimidine Treatment of Patients Carrying Multiple DPYD Variants
Next Article in Special Issue
Gene-Based Nonparametric Testing of Interactions Using Distance Correlation Coefficient in Case-Control Association Studies
Previous Article in Journal
Advances on Aptamers against Protozoan Parasites
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Co-differential Gene Selection and Clustering Based on Graph Regularized Multi-View NMF in Cancer Genomic Data

1
School of Information Science and Engineering, Qufu Normal University, Rizhao 276826, China
2
Library of Qufu Normal University, Qufu Normal University, Rizhao 276826, China
*
Author to whom correspondence should be addressed.
Genes 2018, 9(12), 586; https://doi.org/10.3390/genes9120586
Submission received: 25 October 2018 / Revised: 13 November 2018 / Accepted: 26 November 2018 / Published: 28 November 2018

Abstract

:
Cancer genomic data contain views from different sources that provide complementary information about genetic activity. This provides a new way for cancer research. Feature selection and multi-view clustering are hot topics in bioinformatics, and they can make full use of complementary information to improve the effect. In this paper, a novel integrated model called Multi-view Non-negative Matrix Factorization (MvNMF) is proposed for the selection of common differential genes (co-differential genes) and multi-view clustering. In order to encode the geometric information in the multi-view genomic data, graph regularized MvNMF (GMvNMF) is further proposed by applying the graph regularization constraint in the objective function. GMvNMF can not only obtain the potential shared feature structure and shared cluster group structure, but also capture the manifold structure of multi-view data. The validity of the proposed GMvNMF method was tested in four multi-view genomic data. Experimental results showed that the GMvNMF method has better performance than other representative methods.

1. Introduction

With the rapid development of gene sequencing technology, a large number of multi-view data have been generated. In essence, multi-view data are insightful and have multiple levels of genetic activity information. Exploring this information will provide us with an unprecedented opportunity to discover the molecular mechanisms of cancer [1]. The Cancer Genome Atlas (TCGA) is the largest genome-based platform. And it provides a large number of different types of omics data. In this paper, we use gene expression (GE), copy number variation (CNV), and methylation (ME) data of four cancers in the TCGA database. They are mutually dependent on each other [2].
In the field of bioinformatics, feature selection and clustering are two important ways to explore genomic data [3,4]. To some extent, feature selection can reduce computational complexity and also find differentially expressed genes associated with cancer. It promotes cancer research at the molecular level of genes. Multi-view clustering is the division of samples or genes in a multi-view dataset into several subsets based on their potential group structure; the samples or genes in the same subset have similarities.
With the advent of the big data era, data are no longer single-view but multi-view data composed of different sources. The information of multiple views in multi-view data is complementary, and it is of great significance to conduct in-depth research on this information [5]. Therefore, a multi-view model that specializes in multi-view data has emerged. By learning a multi-view model, we can mine multiple views to achieve better performance than relying on a single view.
Since non-negative matrix factorization (NMF) is an effective matrix decomposition method, more and more researchers are paying attention to the integration model of NMF [6,7]. The integrated model of NMF has straightforward interpretability. Naturally, some improvements and enhancements to the integration model of NMF have been proposed. For example, Zhang et al. [2] proposed joint NMF (jNMF) to discover the common pattern in multi-dimensional genomic data. In order to enhance the robustness of the model to heterogeneous noise, Yang et al. [8] obtained the integrative NMF (iNMF) model by improving jNMF. After applying orthogonal constraints on jNMF, Stražar et al. [9] proposed integrative orthogonality-regularized NMF (iONMF) to predict protein-RNA interactions. In order to detect differentially expressed genes in transcriptomics data, Wang et al. [10] proposed a new method called joint non-negative matrix factorization meta-analysis (jNMFMA) by combining jNMF and meta-analysis.
Although the above methods are effective, their flexibility is limited. They cannot simultaneously perform feature selection and multi-view clustering. However, feature selection and multi-view clustering facilitate the multi-level understanding of the overall system of cancer. Therefore, in this paper, we present a novel method called Multi-view Non-negative Matrix Factorization (MvNMF). It can effectively perform selection of co-differential genes and multi-view clustering simultaneously. Specifically, we improved jNMF by decomposing the coefficient matrix H I into the product of the subspace transformation matrix U I and the shared coefficient matrix V . Then, the new model is X I = W U I V ( I = 1 , 2 , , d ) . d indicates the number of different types of non-negative matrices. The shared basis matrix W contains potential group structures between different views. The shared coefficient matrix V not only has the low rank characteristic, but also has the shared feature pattern for views with different sources. In order to enhance the robustness of MvNMF to data containing manifold structures, the graph regularized MvNMF (GMvNMF) method is further proposed. This can be obtained by adding the intrinsic geometric information of the data to the MvNMF method.
The main contributions of this paper are as follows:
  • In order to effectively cluster and select features for multi-view data at the same time, a novel integrated model called MvNMF is proposed. In the MvNMF framework, the shared basis matrix can reconstruct the potential cluster group structure, which contributed to the improvement of clustering performance. The selection of the co-differential genes can be performed because the shared coefficient matrix can recover the common feature pattern from different views.
  • The graph regularization was applied to the objective function to form the GMvNMF method, which ensured that GMvNMF can capture the manifold structure of the multi-view data. This makes sense for the performance improvement of the integrated model.
  • Scientific and rational experiments were designed on the cancer genomic data to illustrate the validity of the GMvNMF method and achieve satisfactory results.
In what follows, jNMF and its representative variants, graph regularization are reviewed in Section 2; a detailed description of the proposed GMvNMF method is also included. The results of multi-view clustering and co-differential gene selection are presented in Section 3. Finally, the paper is concluded in Section 4.

2. Materials and Methods

2.1. Joint Non-Negative Matrix Factorization and Representative Variants

Joint NMF (jNMF) [2] is a popular matrix decomposition algorithm. In the field of bioinformatics, each type of genomic data can be represented as an original matrix. The row of the matrix represents the sample, and the column represents the expression level of genomic data.
Given d different types of non-negative matrices, the goal of jNMF is to find that the product of a shared basis matrix W n × k and the corresponding coefficient matrix ( H I ) k × m I is similar to the input data matrix ( X I ) n × m I , i.e., X I W H I ( I = 1 , 2 , , d ) . n represents the number of rows of the input data matrix. The value of k means the degree of dimensionality reduction of data. m I represents the number of columns of the i -th input data matrix. Further, the objective function of jNMF can be expressed as:
min I = 1 d X I W H I F 2        s . t . W 0 , H I 0 ,
where F denotes the Frobenius norm of the matrix. The shared basis matrix can reflect the sharing pattern of multi-view data matrices from different sources [11]. It is obvious that jNMF is equivalent to NMF when the value of d is 1. In other words, jNMF is a flexible and clever NMF extension model for the integration of multi-view data. Then, the updated rules are as follows:
W i l W i l I = 1 d ( X I H I ) i l I = 1 d ( W H I H I T ) i l ,
H l j H l j ( W T X I ) l j ( W T W H I ) l j ,      I = 1 , 2 , , d ,
where W i l and H l j refer to specific elements in the matrices W and H I .
jNMF is effective in finding homogenous effects of data from different sources, however, it does not consider the effects of heterogeneous noise between multi-view data. Therefore, Yang et al. [8] proposed an integrative NMF (iNMF) model. Specifically, the objective function of iNMF was defined as follows:
min I = 1 d X I ( W + V I ) H I F 2 + λ I = 1 d V I H I F 2 s . t .      W 0 , H I 0 , V I 0 , I = 1 , , d ,
where λ represents a balance parameter. W H I means taking into account the homogenous effect, and V I H I means considering the heterogeneous effect. In other words, V I H I can be used as an approximation of the heterogeneous effect.
In order to obtain non-overlapping and sparse solutions, Stražar et al. [9] proposed integrative orthogonality-regularized non-negative matrix factorization (iONMF) by applying orthogonal regularization in the jNMF framework. Thus, its cost function can be written as:
min I = 1 d ( X I W H I F 2 + α H I H I T I F 2 ) s . t .      W 0 , H I 0 , I = 1 , , d ,
where I represents a unit matrix, and α is a trade-off parameter. iONMF exhibits better performance in predicting protein-RNA interactions.

2.2. Graph Regularization

The maturity of manifold learning theory made people pay more attention to the internal geometric structure in the original data. The basic idea of graph regularization is to reconstruct the low-dimensional manifold structure embedded in high dimensional sample space. That is to say, adjacent sample points in the high dimensional sample space should be as close as possible in the low dimensional space [12]. If each sample point is used as the vertices of the graph to construct a K-nearest neighbor graph, then a symmetric matrix E is obtained [13]. E i j represents the weight of the edge connecting vertex i and vertex j . Thus, the degree of proximity between the vertices can be measured using E i j . The definition of E i j can be shown as follows:
E i j = { 1   if    x i N K ( x j )       or      x j N K ( x i ) , 0        otherwise ,
where N K ( x i ) is the set of K nearest neighbors of x i . x i is the i -th sample point.
Mathematically, the graph regularization can be formulated as follows:
  min V i , j s i s j 2 E i , j = min V tr ( V ( D E ) V T ) = min V tr ( V L V T ) ,
where D is a diagonal matrix and the elements on the diagonal are composed of the sum of the rows or columns of E . s i and s j are the low dimensional representation of x i and x j , respectively. tr ( ) is the trace of the matrix. The matrix V represents the coefficient matrix produced by the decomposition of NMF. Finally, L = D E is a graph Laplacian matrix [14].

2.3. Graph Regularized Multi-View Non-Negative Matrix Factorization

It is well known that cancer genomic datasets contain many types of data. In order to effectively utilize the information of multiple views, we proposed the MvNMF model and further improved it to get GMvNMF. The GMvNMF algorithm is introduced in detail; the proposed algorithm is given in Algorithm 1.

2.3.1. Objective Function

jNMF is a good integration model that can fully explore potential shared structures in multiple views [2]. However, its flexibility is not sufficient to explore multi-view clustering and selecting co-differentially expressed genes at the same time. For multi-view clustering, the same sample points in different views are likely to be grouped together. Therefore, we required the basis matrix to exhibit the potential cluster structure that is shared by different views. For the selection of co-differential genes, the expression of the same gene in different views should be considered comprehensively. Therefore, we required that the coefficient matrix reflected the shared feature structure from different views.
In view of the above requirements, a model called MvNMF was designed. It can simultaneously perform multi-view clustering and selection of co-differentially expressed genes. ( X I ) n × m I can be approximated by W U I V ( I = 1 , 2 , , d ) . Specifically, the objective function of MvNMF can be formulated as optimization problem:
min I = 1 d X I W U I V F 2            s . t .          W 0 , U I 0 , V 0 ,    I = 1 , 2 , , d ,
where W n × k is the shared basis matrix, ( U I ) k × r is the subspace transformation matrix, and V r × m is the shared coefficient matrix. n is the number of samples in the dataset. m is the number of features in the dataset. k denotes dimensionality reduction and r denotes the rank of the matrix. From the objective function we can see that the H I of jNMF can be approximated by U I V .
We further considered the low-dimensional manifold structure embedded in the high-dimensional multi-view data space. Thus, the GMvNMF model was obtained by combining graph regularization and MvNMF.
Its objective function can be written as the following minimization problem:
min I = 1 d X I W U I V F 2    + λ I tr ( W T L I W )           s . t .       W 0 , U I 0 , V 0 ,    I = 1 , 2 , , d ,
where λ I 0 is the balance parameter that controls the Laplacian regularization. It is worth mentioning that the different values of λ I represent the heterogeneity of multi-view data. If λ I = 0 , GMvNMF will be simplified to MvNMF. In other words, MvNMF is a special case of GMvNMF. Therefore, the following section only shows the optimization algorithm of GMvNMF.

2.3.2. Optimization of GMvNMF

The Equation (9) can be rewritten as:
min I = 1 d tr ( ( X I W U I V ) T ( X I W U I V ) ) + λ I tr ( W T L I W ) = tr ( X I T X I ) 2 tr ( X I T W U I V ) + tr ( V T U I T W T W U I V ) + λ I tr ( W T L I W ) .
The multiplicative iterative method was used to solve the optimization problem in Equation (10). Then the Lagrangian function f was constructed as follows:
f = tr ( X I T X I ) 2 tr ( X I T W U I V ) + tr ( V T U I T W T W U I V ) + λ I tr ( W T L I W )                    + tr ( ψ W T ) + tr ( φ I U I T ) + tr ( μ V T ) ,
where ψ = [ ψ i l ] , φ I = [ φ l a ] I and μ = [ μ a j ] are Lagrange multipliers that constrain W 0 , U I 0 and V 0 , respectively. i , l , a and j represent the subscripts of the elements in the matrix.
We separately derived the partial derivatives of W , U I and V of the Lagrangian function as follows:
f W = 2 X I V T U I T + 2 W U I V V T U I T + 2 λ I L I W + ψ ,
f U I = 2 W T X I V T + 2 W T W U I V V T + φ I ,       I = 1 , 2 , , d ,
f V = 2 U I T W T X I + 2 U I T W T W U I V + μ .
It is well known that Karush-Kuhn-Tucher (KKT) conditions [15] can be applied to solve an optimization problem with inequality constraints. By using the KKT conditions ψ W = 0 , φ I U I = 0 and μ V = 0 , we can get the following update rules:
W i l W i l I = 1 d ( X I V T U I T ) i l I = 1 d ( W U I V V T U I T + λ I L I W ) i l ,
( U I ) l a ( U I ) l a ( W T X I V T ) l a ( W T W U I V V T ) l a ,         I = 1 , 2 , , d ,
V a j V a j I = 1 d ( U I T W T X I ) a j I = 1 d ( U I T W T W U I V ) a j .
Finally, we summarize the iterative process of the proposed GMvNMF model in Algorithm 1.
Algorithm 1: GMvNMF
Data Input: ( X I ) n × m I
Parameters: λ I
Output: W , U I and V
Initialization: W 0 , U I 0 and V 0
Set r = 1
Repeat
 Update W by (15);
 Update U I by (16);
 Update V by (17);
r = r + 1 ;
Until convergence

3. Results

We performed experiments with multi-view clustering and selection of co-differentially expressed genes to verify the effectiveness of the proposed method. In addition, we used jNMF [2], iNMF [8] and iONMF [9] as comparison methods. Detailed information on the experimental settings and results are shown in the following section.

3.1. Datasets

The Cancer Genome Atlas (TCGA) program intends to analyze the genomic variation map of cancer by using high-throughput sequencing technology [16]. As the largest cancer genome database, TCGA contains a lot of valuable and incredible information. An in-depth study of this information can help us understand, prevent, and treat cancer. In this paper, we used four multi-view datasets to analyze the performance of the proposed method. These datasets included pancreatic adenocarcinoma (PAAD), esophageal carcinoma (ESCA), colon adenocarcinoma (COAD), and head and neck squamous cell carcinoma (HNSC). Each cancer dataset contained three different types of data, such as GE, CNV, and ME. All of the above data can be downloaded from the TCGA (https://tcgadata.nci.nih.gov/tcga/). In the experiment, we performed preprocessing on the data. First, principal component analysis (PCA) was used to reduce dimensionality and remove redundant information and noise on the data. Then, the data matrix was normalized such that each row of the matrix was distributed between 0 and 1. More descriptions of multi-view datasets are summarized in Table 1.

3.2. Parameter Setting

In the MvNMF and GMvNMF methods, we needed to choose parameters such as k , r and λ I . The values of k and r determined the size of the shared basis matrix, the subspace transformation matrix and the shared coefficient matrix. Thus, choosing a reasonable parameter value will promote the experimental results. Since the value of k means the degree of dimensionality reduction of data, it had a significant impact on the experiment. r is the rank of the matrix. If the value of r is more appropriate, then a better genetic selection result will be obtained. In other words, MvNMF and GMvNMF were sensitive to the choice of k and r . The graph regularization parameters λ I controlled the extent to which the internal geometric structure of the original data was preserved. In addition, λ I reflected the heterogeneity of data from different sources in multi-view data.
In the experiment, we empirically set λ I corresponding to different views in a multi-view data to the same value [17]. For convenience, we used the grid search algorithm to select the optimal value of the parameter. When k , r , and λ I were selected in the interval [1, 50], [1, k − 1], and [1, 100,000], respectively, MvNMF and GMvNMF achieved the best performance. The specific conditions of the selected parameters can be seen from the following figures. It should be noted that, as we can see from Figure 1a, when k = 2 , MvNMF had a higher accuracy in the HNSC dataset. That is to say, the value of r in MvNMF can only be 1 on the HNSC dataset. Therefore, it is not shown in Figure 1c. In summary, we can select the optimal parameters through Figure 1 and Figure 2.

3.3. Convergence and Computational Time Analysis

We iterated the updated rules of MvNMF and GMvNMF to approximate the local optimal solution of the objective function. In Figure 3, the convergence curves for the five methods are given (to save space, only the convergence curves on the ESCA dataset are shown). These five methods consisted of jNMF, iNMF, iONMF, MvNMF, and GMvNMF. It can be observed from Figure 3 that these five methods converged in 100 iterations. The error value is the loss function value. Additionally, the convergence criterion was that the error value tended to zero. Since MvNMF and GMvNMF had smaller error values than other methods, the convergence of MvNMF and GMvNMF is better.
In addition, we compared the execution time of these five algorithms. Experiments were executed on a PC with 3.50 GHz Intel(R) (Santa Clara, CA, USA) Xeon(R) CPU and 16G RAM. In the experiment, each method was repeated 10 times. The mean and variance were calculated. The statistics of the computational time are listed in Table 2. As can be seen from Table 2, all five methods had satisfactory running times. Because iNMF takes into account the heterogeneous effect, its computational time was affected. IONMF imposed orthogonal constraints, thus, it had the longest running time. MvNMF and GMvNMF had lower running times. This is because the experimental results showed that the decomposed matrix of our proposed method had better sparsity and lower rank than the matrix after jNMF decomposition.

3.4. Clustering Results

We performed clustering experiments on four multi-view datasets to verify the effectiveness of the proposed method. Multi-view clustering was performed on the shared basis matrix using the K-means algorithm.

3.4.1. Evaluation Metrics

In order to strictly analyze the performance of multi-view clustering, we adopted multiple measures, including accuracy (AC), recall, precision, and F-measure [18,19]. The AC is defined as follows:
A C = i = 1 n δ ( s i , m a p ( r i ) ) n ,
where n is the number of samples contained in the dataset, r i is the clustering label obtained using the clustering algorithm, and s i is the real data label. m a p ( r i ) is a permutation mapping function that maps clustering labels to real data labels. The real clusters refer to the known sample labels. In addition, δ ( x , y ) is a delta function.
Recall, precision, and F-measure are a set of metrics that are widely used in clustering applications. Recall can also be called sensitivity. True-positive (TP) indicates that two data points of the same cluster are divided into the same cluster. True-negative (TN) means that two data points of the same cluster are divided into different clusters. False-positive (FP) indicates that data points of two different clusters are divided into the same cluster. False-negative (FN) refers to two data points in different clusters divided into different clusters. Recall, precision and F-measure are defined as follows:
r e c a l l = T P T P + F N ,
p r e c i s i o n = T P T P + F P ,
F m e a s u r e = 2 1 / r e c a l l + 1 / p r e c i s i o n ,
where F-measure is a comprehensive evaluation indicator that takes into account recall and precision.

3.4.2. Multi-View Clustering Results

In the experiment, each algorithm is executed 50 times to reduce the impact of random initialization on multi-view clustering results. The mean and variance of performance on each multi-view data are recorded in Table 3.
According to Table 3, we can draw the following conclusions:
  • The clustering performance of jNMF on PAAD and COAD datasets was better than iNMF, iONMF, and MvNMF. This demonstrates that improvements to the traditional NMF integration model may result in the loss of useful information, which in turn affected the clustering results. However, in the ESCA and HNSC datasets, MvNMF outperformed jNMF, iNMF, and iONMF from the overall perspective of the evaluation metrics. This shows the validity of our proposed MvNMF model, which better preserved the complementary information between multiple views.
  • From Table 3, we can see that the precision of the GMvNMF method and the precision of the MvNMF method were similar in the four multi-view datasets. However, the GMvNMF method was at least about 18, 32 and 20% higher than the MvNMF method in terms of AC, recall, and F-measure. Therefore, the GMvNMF method had better clustering performance. This shows that it is necessary to consider the manifold structure that exists in multi-view data.
  • Taking the four multi-view datasets in Table 3 as a whole, the proposed GMvNMF method had the best clustering performance. GMvNMF outperformed other methods by about 23, 39, 0.67, and 25%, with respect to the average values of the metrics AC, recall, precision, and F-measure. Therefore, GMvNMF is an effective integration model that takes into account the latent group structure and intrinsic geometric information between multi-view data.

3.5. Gene Selection Results

3.5.1. Co-Differentially Expressed Gene Selection Results

It is well known that genomic alterations and genetic mutations can cause cancer [20,21]. Therefore, research on cancer genomic data is an urgent task. In the feature selection experiment, we used genomic data including GE, CNV, and ME to verify the effectiveness of the proposed method. The co-differential genes were selected on a shared coefficient matrix. Since the differential genes we selected are genes expressed in GE, CNV, and ME, the selected co-differential genes have more important biological significance.
In the experiment, we scored all the genes. These genes were then ranked in descending order of score. The higher the score of a gene, the greater its significance. We chose such a gene as a differentially expressed gene. In practice, we selected the top 500 genes of each method as co-differentially expressed genes for comparison. Then, the selected genes were placed in the GeneGards (http://www.genecards.org/) for analysis. GeneCards is a comprehensive database of human genes that provides a variety of valuable information for studying genes [22]. Table 4 lists the results of five methods for selecting co-differential genes.
In Table 4, N is obtained by matching the co-differential genes selected by each method to the virulence gene pool of PAAD, ESCA, COAD, and HNSC. A larger N indicates a higher accuracy in identifying co-differentially expressed genes. HRS represents the highest relevant score, and ARS represents the average relevant score. Relevant scores represent the degree to which a gene is associated with a disease. A higher relevant score for a gene means that the gene is likely to be a pathogenic gene. Although the ARS of MvNMF was slightly higher than GMvNMF in COAD and HNSC, the performance of co-differentially expressed genes selected by GMvNMF was better on the whole. This indicates that our method was reasonable. In order to retain the geometric structure in the data, it was necessary to add graph regularization to the method.

3.5.2. Discussion of Co-Differential Genes

Table 5 lists the relevant information of the top 10 co-differential genes selected by the GMvNMF method (to save space, we only listed the top 10 genes selected in the COAD dataset.). From Table 5, we can see that BRCA1 has the highest relevance score. BRCA1 is a protein-coding gene involved in DNA repair. When it is mutated, the tumor suppressor protein does not form normally. This leads to the emergence of cancer. BRCA1 has been confirmed to be related to COAD [23]. BRCA2 is a tumor suppressor gene, which is mainly involved in the repair of DNA damage and regulation of transcription. There is literature that BRCA2 is related to COAD [24]. As we all know, mutations in BRCA1 and BRCA2 increase the risk of breast or ovarian cancer [25]. Therefore, mutations in one gene may be related to the production of multiple cancers. This suggests that biologists can further study the link between COAD and breast or ovarian cancer. The protein encoded by the epidermal growth factor receptor (EGFR) is a transmembrane glycoprotein that is a member of the protein kinase superfamily. In addition, mutation or overexpression of EGFR generally triggers COAD [26].
Table 6 lists the co-differentially expressed genes with the highest relevance score selected by GMvNMF on the multi-view dataset of PAAD, ESCA and HNSC. These co-differential genes were highly likely to cause cancer. The relevance score of EGFR is 168.23 in the HNSC. EGFR is a protein-coding gene. Among its related pathways are extracellular regulated protein kinases (ERK) signaling and GE. Moreover, the importance of EGFR in HNSC has been widely recognized [27,28]. EGFR also appears in Table 5, which indicates that EGFR has to do with the occurrence of COAD and HNSC. This provides a new way for biologists to study COAD and HNSC.

4. Conclusions

In this paper, we proposed a new integrated NMF model called MvNMF for multi-view clustering and selection of co-differentially expressed genes. Considering the low-dimensional manifold structure existing in the high-dimensional multi-view sample space, the graph regularization constraint was added to the objective function of MvNMF. This new model is called GMvNMF. It effectively encodes the geometric information inherent in the data. Numerous experiments on cancer genomic data showed that our proposed GMvNMF method is more effective.
For future work, we continue to improve the model to enhance its robustness and sparsity.

Author Contributions

N.Y. and J.-X.L. jointly contributed to the design of the study. N.Y. designed and implemented the MvNMF and GMvNMF method, performed the experiments, and drafted the manuscript. Y.-L.G. and R.Z. participated in the design of the study and performed the statistical analysis. L.-Y.D. and J.S. contributed to the data analysis. All authors read and approved the final manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant Nos. 61872220, 61572284 and 61502272; and China Postdoctoral Science Foundation under Grant Nos. 2018M642635. This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, W.; Liu, Y.; Sun, N.; Wang, D.; Boydkirkup, J.; Dou, X.; Han, J.D.J. Integrating genomic, epigenomic, and transcriptomic features reveals modular signatures underlying poor prognosis in ovarian cancer. Cell Rep. 2013, 4, 542–553. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, S.; Liu, C.C.; Li, W.; Shen, H.; Laird, P.W.; Zhou, X.J. Discovery of multi-dimensional modules by integrative analysis of cancer genomic data. Nucleic Acids Res. 2012, 40, 9379–9391. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Yan, Q.; Ding, Y.; Xia, Y.; Chong, Y.; Zheng, C.; Yan, Q.; Ding, Y.; Xia, Y.; Chong, Y.; Zheng, C. Class-probability propagation of supervised information based on sparse subspace clustering for hyperspectral images. Remote Sens. 2017, 9, 1017. [Google Scholar] [CrossRef]
  4. Liu, J.X.; Wang, D.; Gao, Y.L.; Zheng, C.H.; Xu, Y.; Yu, J. Regularized non-negative matrix factorization for identifying differential genes and clustering samples: A survey. IEEE/ACM Trans. Comput. Biol. Bioinform. 2018, PP, 1. [Google Scholar]
  5. Sun, S. A survey of multi-view machine learning. Neural Comput. Appl. 2013, 23, 2031–2038. [Google Scholar] [CrossRef]
  6. Wang, J.; Tian, F.; Yu, H.; Liu, C.H.; Zhan, K.; Wang, X. Diverse non-negative matrix factorization for multiview data representation. IEEE Trans. Cybern. 2017, PP, 1–13. [Google Scholar] [CrossRef]
  7. Zhang, L.; Zhang, S. A unified joint matrix factorization framework for data integration. arXiv, 2017; arXiv:1707.08183. [Google Scholar]
  8. Yang, Z.; Michailidis, G. A non-negative matrix factorization method for detecting modules in heterogeneous omics multi-modal data. Bioinformatics 2016, 32, 1–8. [Google Scholar] [CrossRef] [PubMed]
  9. Stražar, M.; Žitnik, M.; Zupan, B.; Ule, J.; Curk, T. Orthogonal matrix factorization enables integrative analysis of multiple RNA binding proteins. Bioinformatics 2016, 32, 1527–1535. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, H.Q.; Zheng, C.H.; Zhao, X.M. jNMFMA: A joint non-negative matrix factorization meta-analysis of transcriptomics data. Bioinformatics 2015, 31, 572–580. [Google Scholar] [CrossRef]
  11. Zhang, C.; Zhang, S. Bayesian joint matrix decomposition for data integration with heterogeneous noise. arXiv, 2017; arXiv:1712.03337. [Google Scholar]
  12. Belkin, M.; Niyogi, P. Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv. Neural Inf. Process. Syst. 2001, 14, 585–591. [Google Scholar]
  13. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1548–1560. [Google Scholar] [PubMed]
  14. Liu, X.; Zhai, D.; Zhao, D.; Zhai, G.; Gao, W. Progressive image denoising through hybrid graph Laplacian regularization: A unified framework. IEEE Trans. Image Process. 2014, 23, 1491–1503. [Google Scholar] [PubMed]
  15. Facchinei, F.; Kanzow, C.; Sagratella, S. Solving quasi-variational inequalities via their KKT conditions. Math. Program. 2014, 144, 369–412. [Google Scholar] [CrossRef]
  16. Katarzyna, T.; Patrycja, C.; Maciej, W. The cancer genome atlas (TCGA): An immeasurable source of knowledge. Contemp. Oncol. 2015, 19, 68–77. [Google Scholar]
  17. Liu, J.; Wang, C.; Gao, J.; Han, J. Multi-View Clustering Via Joint Nonnegative Matrix Factorization. In Proceedings of the 2013 SIAM International Conference on Data Mining, Austin, TX, USA, 2–4 May 2013; pp. 252–260. [Google Scholar]
  18. Shahnaz, F.; Berry, M.W.; Pauca, V.P.; Plemmons, R.J. Document clustering using nonnegative matrix factorization. Inf. Process. Manag. 2004, 42, 373–386. [Google Scholar] [CrossRef]
  19. Jing, P.J.; Shen, H.B. MACOED: A multi-objective ant colony optimization algorithm for SNP epistasis detection in genome-wide association studies. Bioinformatics 2015, 31, 634–641. [Google Scholar] [CrossRef] [PubMed]
  20. Ponder, B.A.J. Cancer genetics. Nature 2001, 411, 336–341. [Google Scholar] [CrossRef]
  21. Liu, J.X.; Gao, Y.L.; Zheng, C.H.; Xu, Y.; Yu, J. Block-constraint robust principal component analysis and its application to integrated analysis of TCGA data. IEEE Trans. Nanobiosci. 2016, 15, 510–516. [Google Scholar] [CrossRef] [PubMed]
  22. Safran, M.; Dalah, I.; Alexander, J.; Rosen, N.; Iny, S.T.; Shmoish, M.; Nativ, N.; Bahir, I.; Doniger, T.; Krug, H. Genecards version 3: The human gene integrator. Database 2010, 2010, baq020. [Google Scholar] [CrossRef] [PubMed]
  23. Chakree, K.; Ovatlarnporn, C.; Dyson, P.J.; Ratanaphan, A. Altered DNA binding and amplification of human breast cancer suppressor gene BRCA1 induced by a novel antitumor compound, [Ru(η6-p-phenylethacrynate)Cl2(pta)]. Int. J. Mol. Sci. 2012, 13, 13183–13202. [Google Scholar] [CrossRef] [PubMed]
  24. Grabsch, H.; Dattani, M.; Barker, L.; Maughan, N.; Maude, K.; Hansen, O.; Gabbert, H.E.; Quirke, P.; Mueller, W. Expression of DNA double-strand break repair proteins ATM and BRCA1 predicts survival in colorectal cancer. Clin. Cancer Res. 2006, 12, 1494–1500. [Google Scholar] [CrossRef] [PubMed]
  25. Antoniou, A.; Pharoah, P.D.; Narod, S.; Risch, H.A.; Eyfjord, J.E.; Hopper, J.L.; Loman, N.; Olsson, H.; Johannsson, O.; Borg, A. Average risks of breast and ovarian cancer associated with BRCA1 or BRCA2 mutations detected in case series unselected for family history: A combined analysis of 22 studies. Am. J. Hum. Genet. 2003, 72, 1117–1130. [Google Scholar] [CrossRef] [PubMed]
  26. Rigopoulos, D.N.; Tsiambas, E.; Lazaris, A.C.; Kavantzas, N.; Papazachariou, I.; Kravvaritis, C.; Tsounis, D.; Koliopoulou, A.; Athanasiou, A.E.; Karameris, A. Deregulation of EGFR/VEGF/HIF-1a signaling pathway in colon adenocarcinoma based on tissue microarrays analysis. J. BUON 2010, 15, 107–115. [Google Scholar]
  27. Lee, J.W.; Soung, Y.H.; Kim, S.Y.; Nam, H.K.; Park, W.S.; Nam, S.W.; Kim, M.S.; Sun, D.I.; Lee, Y.S.; Jang, J.J. Somatic mutations of EGFR gene in squamous cell carcinoma of the head and neck. Clin. Cancer Res. 2005, 11, 2879–2882. [Google Scholar] [CrossRef] [PubMed]
  28. Bossi, P.; Resteghini, C.; Paielli, N.; Licitra, L.; Pilotti, S.; Perrone, F. Prognostic and predictive value of EGFR in head and neck squamous cell carcinoma. Oncotarget 2016, 7, 74362–74379. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Performance of the Multi-view Non-negative Matrix Factorization (MvNMF) set with different values of k and r . (a) is the clustering performance of MvNMF on PAAD and HNSC about k ; (b) is the clustering performance of MvNMF on ESCA and COAD about k ; (c) is the clustering performance of MvNMF on PAAD, ESCA and COAD about r .
Figure 1. Performance of the Multi-view Non-negative Matrix Factorization (MvNMF) set with different values of k and r . (a) is the clustering performance of MvNMF on PAAD and HNSC about k ; (b) is the clustering performance of MvNMF on ESCA and COAD about k ; (c) is the clustering performance of MvNMF on PAAD, ESCA and COAD about r .
Genes 09 00586 g001
Figure 2. Performance of the graph regularized MvNMF (GMvNMF) set with different values of k , r and λ . (a) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about k ; (b) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about r ; (c) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about λ .
Figure 2. Performance of the graph regularized MvNMF (GMvNMF) set with different values of k , r and λ . (a) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about k ; (b) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about r ; (c) is the clustering performance of GMvNMF on PAAD, HNSC, ESCA and COAD about λ .
Genes 09 00586 g002
Figure 3. Convergence curves of joint Non-negative Matrix Factorization (jNMF), integrated NMF (iNMF), integrative orthogonality-regularized NMF (iONMF), MvNMF, and GMvNMF.
Figure 3. Convergence curves of joint Non-negative Matrix Factorization (jNMF), integrated NMF (iNMF), integrative orthogonality-regularized NMF (iONMF), MvNMF, and GMvNMF.
Genes 09 00586 g003
Table 1. Description of four multi-view datasets.
Table 1. Description of four multi-view datasets.
DatasetsData TypesNormal SamplesTumor SamplesGenes
PAADGE, CNV, ME176419,877
ESCAGE, CNV, ME183919,877
HNSCGE, CNV, ME3982019,877
COADGE, CNV, ME2621916,977
Note: Datasets represent different multi-view data. PAAD: pancreatic adenocarcinoma; ESCA: esophageal carcinoma; HNSC: head and neck squamous cell carcinoma; COAD: colon adenocarcinoma; GE: gene expression; CNV: copy number variation; ME: methylation.
Table 2. Computational time on ESCA.
Table 2. Computational time on ESCA.
MethodsTimes (s)
jNMF2.8808 ± 1.7 × 10−4
iNMF3.4647 ± 1.3 × 10−3
iONMF5.7375 ± 2.8 × 10−3
MvNMF1.3495 ± 7.0 × 10−5
GMvNMF1.0767 ± 1.4 × 10−4
Table 3. The clustering performance on PAAD, ESCA, COAD and HNSC.
Table 3. The clustering performance on PAAD, ESCA, COAD and HNSC.
MethodsMetricsjNMFiNMFiONMFMvNMFGMvNMF
PAADAC (%)70.39 ± 3.7170.30 ± 3.7165.01 ± 2.7363.86 ± 0.7895.59 ± 0.05
Recall (%)61.78 ± 7.3456.49 ± 8.3053.17 ± 5.4856.30 ± 2.7791.90 ± 5.28
Precision (%)97.93 ± 0.0398.35 ± 0.0197.89 ± 0.0097.88 ± 0.0395.99 ± 1.92
F-measure (%)71.99 ± 5.2666.92 ± 7.0665.65 ± 4.6569.89 ± 1.9292.12 ± 5.03
ESCAAC (%)65.32 ± 3.7066.42 ± 3.4957.64 ± 0.2168.04 ± 0.7093.23 ± 0.21
Recall (%)51.48 ± 6.6754.39 ± 6.5551.90 ± 0.6751.10 ± 3.7597.21 ± 0.74
Precision (%)88.16 ± 5.8488.29 ± 6.2194.70 ± 0.2093.51 ± 0.5195.20 ± 0.00
F-measure (%)62.81 ± 6.6065.61 ± 6.2567.16 ± 0.5564.47 ± 3.3995.97 ± 0.37
COADAC (%)73.91 ± 1.8471.00 ± 1.3366.99 ± 0.6865.13 ± 0.0392.01 ± 0.01
Recall (%)57.15 ± 6.5451.28 ± 5.1650.24 ± 2.9547.15 ± 1.5894.70 ± 3.61
Precision (%)90.02 ± 3.2987.60 ± 4.5290.18 ± 1.8889.94 ± 0.6493.42 ± 0.00
F-measure (%)68.25 ± 5.3463.53 ± 5.1163.79 ± 2.861.45 ± 1.5892.22 ± 3.23
HNSCAC (%)66.75 ± 0.0066.16 ± 0.0166.39 ± 0.0067.70 ± 0.0386.18 ± 2.67
Recall (%)53.62 ± 2.1951.18 ± 2.2150.68 ± 2.2055.23 ± 2.4487.07 ± 5.57
Precision (%)95.22 ± 0.3894.30 ± 0.3994.03 ± 0.3995.53 ± 0.3394.93 ± 0.05
F-measure (%)67.85 ± 2.0165.61 ± 1.9665.09 ± 2.0369.10 ± 2.2088.63 ± 3.30
Note: The best experimental results are highlighted in bold.
Table 4. Co-differential genes selection results on four multi-view datasets.
Table 4. Co-differential genes selection results on four multi-view datasets.
MethodsPAADESCACOADHNSC
NHRSARSNHRSARSNHRSARSNHRSARS
jNMF37484.934.8916876.155.19142103.77.02175168.2317.75
iNMF37584.934.8417176.155.31144103.77.71175102.9816.65
iONMF375100.565.1917076.155.36141165.658.64175168.2317.52
MvNMF365100.565.2317076.155.52145165.658.66177168.2318.00
GMvNMF376100.565.5318276.155.69152173.128.37177168.2317.60
Note: N is obtained by matching the co-differential genes selected by each method to the virulence gene pool of PAAD, ESCA, COAD, and HNSC. HRS represents the highest relevant score, and ARS represents the average relevant score. The best experimental results are highlighted in bold.
Table 5. Summary of the co-differential genes selected by the GMvNMF method.
Table 5. Summary of the co-differential genes selected by the GMvNMF method.
Gene IDGene EDRelated Go AnnotationsRelated DiseasesRelevance Score
672BRCA1RNA binding and ligase activityBreast-Ovarian Cancer, Familial 1 and Pancreatic Cancer 4173.12
675BRCA2protease binding and histone acetyltransferase activityFanconi Anemia, Complementation Group D1 and Breast Cancer135.87
1956EGFRidentical protein binding and protein kinase activityInflammatory Skin and Bowel Disease, Neonatal, 2 and Lung Cancer104.16
3569IL6signaling receptor binding and growth factor activityKaposi Sarcoma and Rheumatoid Arthritis, Systemic Juvenile58.74
4318MMP9identical protein binding and metalloendopeptidase activityMetaphyseal Anadysplasia 2 and Metaphyseal Anadysplasia45.57
1495CTNNA1actin filament bindingMacular Dystrophy, Patterned, 2 and Butterfly-Shaped Pigment Dystrophy41.99
1950EGFcalcium ion binding and epidermal growth factor receptor bindingHypomagnesemia 4, Renal and Familial Primary Hypomagnesemia with Normocalciuria and Normocalcemia40.84
5594MAPK1transferase activity, transferring phosphorus-containing groups and protein tyrosine kinase activityChromosome 22Q11.2 Deletion Syndrome, Distal and Pertussis39.23
2475MTORtransferase activity, transferring phosphorus-containing groups and protein serine/threonine kinase activityFocal Cortical Dysplasia, Type II and Smith-Kingsmore Syndrome34.07
887CCKBRG-protein coupled receptor activity and 1-phosphatidylinositol-3-kinase regulator activityPanic Disorder and Anxiety23.43
Note: Gene ID represents the number of the gene. Gene ED represents the gene name.
Table 6. Summary of the co-differential genes selected on PAAD, ESCA and HNSC.
Table 6. Summary of the co-differential genes selected on PAAD, ESCA and HNSC.
Gene IDGene EDRelated Go AnnotationsRelated DiseasesParalog Gene
999CDH1calcium ion binding and protein phosphatase bindingGastric Cancer, Hereditary Diffuse and Blepharocheilodontic Syndrome 1CDH3
1499CTNNB1DNA binding transcription factor activity and bindingMental Retardation, Autosomal Dominant 19 and PilomatrixomaJUP
1956EGFRidentical protein binding and protein kinase activityInflammatory Skin and Bowel Disease, Neonatal, 2 and Lung CancerERBB4
Note: Paralog gene produced via gene duplication within a genome.

Share and Cite

MDPI and ACS Style

Yu, N.; Gao, Y.-L.; Liu, J.-X.; Shang, J.; Zhu, R.; Dai, L.-Y. Co-differential Gene Selection and Clustering Based on Graph Regularized Multi-View NMF in Cancer Genomic Data. Genes 2018, 9, 586. https://doi.org/10.3390/genes9120586

AMA Style

Yu N, Gao Y-L, Liu J-X, Shang J, Zhu R, Dai L-Y. Co-differential Gene Selection and Clustering Based on Graph Regularized Multi-View NMF in Cancer Genomic Data. Genes. 2018; 9(12):586. https://doi.org/10.3390/genes9120586

Chicago/Turabian Style

Yu, Na, Ying-Lian Gao, Jin-Xing Liu, Junliang Shang, Rong Zhu, and Ling-Yun Dai. 2018. "Co-differential Gene Selection and Clustering Based on Graph Regularized Multi-View NMF in Cancer Genomic Data" Genes 9, no. 12: 586. https://doi.org/10.3390/genes9120586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop