Next Article in Journal
Physiotherapy versus Consecutive Physiotherapy and Cognitive Treatment in People with Parkinson’s Disease: A Pilot Randomized Cross-Over Study
Previous Article in Journal
Risk Factors in Third and Fourth Degree Perineal Tears in Women in a Tertiary Centre: An Observational Ambispective Cohort Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

JDSNMF: Joint Deep Semi-Non-Negative Matrix Factorization for Learning Integrative Representation of Molecular Signals in Alzheimer’s Disease

School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology, Gwangju 61005, Korea
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2021, 11(8), 686; https://doi.org/10.3390/jpm11080686
Submission received: 2 June 2021 / Revised: 16 July 2021 / Accepted: 18 July 2021 / Published: 21 July 2021

Abstract

:
High dimensional multi-omics data integration can enhance our understanding of the complex biological interactions in human diseases. However, most studies involving unsupervised integration of multi-omics data focus on linear integration methods. In this study, we propose a joint deep semi-non-negative matrix factorization (JDSNMF) model, which uses a hierarchical non-linear feature extraction approach that can capture shared latent features from the complex multi-omics data. The extracted latent features obtained from JDSNMF enabled a variety of downstream tasks, including prediction of disease and module analysis. The proposed model is applicable not only to sample-matched multiple data (e.g., multi-omics data from one cohort) but also to feature-matched multiple data (e.g., omics data from multiple cohorts), and therefore it can be flexibly applied to various cases. We demonstrate the capabilities of JDSNMF using sample-matched simulated data and feature-matched multi-omics data from Alzheimer’s disease cohorts, evaluating the feature extraction performance in the context of classification. In a test application, we identify AD- and age-related modules from the latent matrices using an explainable artificial intelligence and regression model. These results show that the JDSNMF model is effective in identifying latent features having a complex interplay of potential biological signatures.

1. Introduction

With the development of high-throughput technology, large amounts of omics data have been produced. These resources enable us to understand disease- and patient-specific variations. Each omics datapoint contains different forms of information and is complementary to other data. By integrating multi-omics data, we can capture more complex interactions of the biological system.
Motivated by this, many studies using various machine learning methods have been performed to integrate multi-omics data, including MOFA [1] based on factor analysis, MSFA [2] based on factor analysis, RGCCA [3] based on canonical correlation analysis, joint NMF (jNMF) [4] based on non-negative matrix factorization (NMF), integrative NMF [5] based on NMF, intNMF [6] based on NMF, iCluster [7] based on Gaussian latent variable model, and JIVE [8] based on principal component analysis. A common aim of these approaches is to characterize heterogeneity between samples. However, these approaches have two limitations. First, non-linear relationships between features can be missed. Second, these approaches require matching samples of multiple omics data. Analysis can be performed only for samples having complete multi-omics data.
To enhance the understanding of complex relationships from multi-omics data, we developed a novel method called the joint deep semi-NMF model (JDSNMF) (Figure 1A). This model is an improvement for multi-omics data over the deep semi-NMF algorithm [9], which is a non-linear feature extraction approach for a single data type in computer vision tasks. JDSNMF is applicable to multiple data with matching features or samples. Cohorts that do not have multi-omics data for a specific disease can be used for integrated analysis. JDSNMF can also be applied to various scenarios such as cross-species or cross-disease comparisons. We demonstrate the capability of JDSNMF and compare its feature extraction performance in the context of classification using simulated data and publicly available Alzheimer’s disease (AD) datasets (Figure 1B). In simulated experiments, we evaluate the feature extraction performance of JDSNMF on multiple data (sample-matched). In real dataset experiments, we evaluate the feature extraction performance of JDSNMF on multi-omics data (feature-matched). In addition we demonstrate the interpretability of JDSNMF on feature-matched AD datasets (Figure 1C,E).
In summary, the main contributions of this work include:
1.
It introduces a hierarchical non-linear unsupervised learning model for capturing complex information from multi-omics data.
2.
It introduces a flexible integration model for multiple data which does not require matching samples.
3.
We compared the feature extraction performance of JDSNMF using sample-matched simulated data and feature-matched AD data.
4.
We identified the interpretability of JDSNMF through module analysis using an explainable artificial intelligence (xAI) (Figure 1D) and regression model (Figure 1E).

2. Related Work

In this section, a brief review of the most relevant works, i.e., NMF, joint NMF, semi-NMF, and deep semi-NMF, is provided.

2.1. Non-Negative Matrix Factorization

NMF [10,11] is a matrix factorization algorithm that contains non-negative constraints at the element of all matrices. It is a widely used linear feature extraction method in many fields. NMF approximately decomposes X Z + H + by minimizing the following objective function.
min Z , H X Z H F 2 , s . t . Z 0 , H 0 ,
where X R m × n is the original non-negative matrix, Z R m × k is the basis matrix, H R k × n is the coefficient matrix, the rank k < min { m , n } , and the operator · F is the Frobenius norm. The multiplicative update rule [11] has been frequently used to minimize the objective function.

2.2. Joint NMF

jNMF is an advanced NMF model that decomposes multiple data into a common basis matrix and multiple coefficient matrices [4]. Given non-negative I matrices X 1 R m × n 1 , …, X I R m × n I , jNMF approximately decomposes with X 1 Z + H 1 + , …, X I Z + H I + as the following objective function.
min Z , H 1 , , H I i = 1 I X i Z H i F 2 , s . t . Z 0 , H i 0 , i = 1 , , I ,
where Z R m × k is a common basis matrix, H i R k × n i is a ith coefficient matrix, and the rank k < min { m , n 1 , , n I } . The basis matrix Z is a feature extracted matrix in which multiple data are integrated, and a coefficient matrix H i is a feature extracted matrix from each i dataset. It is useful for integrating heterogeneous data such as multi-omics data.

2.3. Semi-NMF

Semi-NMF [12] is a relaxed NMF algorithm that restricts only a non-negative constraint of the H matrix. Matrix X and matrix Z may have mixed signs. Semi-NMF decomposes with X ± Z ± H + , minimizing the following objective function.
min Z , H X Z H F 2 , s . t . H 0 ,
where X R m × n , Z R m × k , H R k × n , and the rank k < min { m , n } . From the perspective of clustering, we can also see the Z matrix as the cluster centroid, and the H matrix can be viewed as the cluster indicator.

2.4. Deep Semi-NMF

Deep semi-NMF [9] is a multi-layer feature extraction algorithm used to learn complex hierarchical representations of a dataset. Deep semi-NMF is a semi-NMF with a multi-layer concept that can learn hidden representations. Deep semi-NMF decomposes a given original X matrix into M + 1 factors with X ± Z 1 ± Z 2 ± Z M ± H M + , minimizing the following objective function.
min Z m , H M X Z 1 ( Z 2 ( ( Z M H M ) ) ) F 2 , s . t . H m 0 , m = 1 , , M ,
where H m are restricted to be non-negative that are implicit representations of m layers as follows:
H M 1 + Z M ± H M + , H 2 + Z 3 ± Z M ± H M + , H 1 + Z 2 ± Z M ± H M +
Implicit representations of layers are suitable clusters according to the corresponding attributes. The multi-layer model helps learning multiple low-dimensional representations of the original matrix.

3. Proposed Method

3.1. Joint Deep Semi-NMF

Our proposed model, JDSNMF, has advantages of non-linear dimensionality reduction and heterogeneous data integration. It adapts several advanced NMF approaches for feature extraction to learn complex non-linear manifolds from complex heterogeneous data. JDSNMF adapts joint NMF [4] to integrate multi heterogeneous data and adapts the multi-layer NMF principle and a non-linear activation function to represent non-linear manifolds. In addition, it uses a regularization term to prevent overfitting. Given matrices X 1 R C × M 1 , …, X I R C × M I , I is the number of multiple datasets, C is the number of samples (or features), and M is the number of individual features (or samples). JDSNMF decomposes the given original matrices by minimizing the following objective function.
min i = 1 I X i U g ( Z i 1 g ( Z i 2 g ( Z i N H i N ) ) ) F 2 + λ S F , s . t . H i n 1 = g ( Z i n H i n ) , H i 0 H i N 1 0 , S { U , Z i 1 , , Z i N , H i N } , n = 1 , , N
where U R C × K 0 is a common sample (or feature) latent matrix, H i 0 R K 0 × M i is a feature (or sample) latent matrix of the first layer, H i n R K n × M i is a feature (or sample) latent matrix of a sublayer, and Z n R K n 1 × K n is a junction latent matrix. The reduced dimension of the first layer is K < min { C , M i } , and the reduced dimension K n is smaller than the reduced dimension K n 1 of the parent layer. JDSNMF uses a non-linear activation function g such as a sigmoid activation function and a ReLU activation function to make the decomposed H matrix non-negative with non-linearity.

3.2. Constructing Modules

The feature latent matrix and the sample latent matrix of the first layer are used in module construction. In the application of NMF, researchers have used two different approaches to construct modules. One is to assign each feature (or sample) to a module with a maximum value [13,14], and the other is to assign a set of significant features (or samples) from z-distribution to each module [4]. Because the distribution difference among modules may assign too many features and samples to specific modules in the former case, we used the latter case. In the sample latent matrix, samples larger than a predefined positive threshold (empirically set to μ + 1.6449 σ , 5%) were selected for each module. In the feature latent matrix, features larger than a given positive threshold and features smaller than a given negative threshold (empirically set to μ ± 2.5758 σ , 1%) were selected for each module.
Figure 2 demonstrates a process of constructing modules using an example of feature-matched gene expression (GE) and DNA methylation (DM) data. In some AD samples, Gene 1 and Gene 6 were overexpressed and Gene 4 was underexpressed; Gene 1 and Gene 6 were hypermethylated and Gene 4 was hypomethylated. Figure 2A shows that GE and DM data were decomposed using the JDSNMF model. U is a feature latent matrix, and H 1 0 and H 2 0 are sample latent matrices of the first layer. Cyan, green, and blue elements in U, H 1 0 , and H 2 0 are bigger than a positive threshold for each module, whereas red elements in U are smaller than a negative threshold for each module. We assumed that genes with values larger than a positive threshold (positive genes) increase GE or DM values of corresponding samples in the original data matrix and that genes with values less than a negative threshold (negative genes) decrease GE or DM values of corresponding samples in the original data matrix. In the K3 module, AD samples from the GE data and NL samples from the DM data were selected by thresholds; Gene 1 and Gene 4 are positive and negative genes, respectively. This result indicates that Gene 1 is more likely to be overexpressed and hypermethylated, and Gene 6 to be underexpressed and hypomethylated, in some of the AD samples. Figure 2B geometrically shows the process of K3 module construction. We coordinated genes and samples in three dimensions, each representing one module. The cyan circle, red circle, and purple squares indicate positive genes, negative genes, and thresholds for the K3 module, respectively.

4. Results

We conducted four experiments to validate the effectiveness of the proposed JDSNMF. First, we evaluated the feature extraction performance in the context of classification on sample-matched simulated data. Second, we evaluated the feature extraction performance on feature-matched multi-omics data.
The performance of JDSNMF using the GE data of the Addneuromed (ANM) and DM data of the MRC London Brainbank cohort as feature-matched data was compared with models using only GE data of the ANM cohort. We provide details of the comparison of the performances of AD/normal (NL) classification and mild cognitive impairment (MCI)/NL classification on the GE data of the ANM cohort. Third, we compared the performance of AD/NL classification according to the reduced dimension K on the Alzheimer’s Disease Neuroimaging Initiative (ADNI) cohort. Fourth, we highlight AD- and age-related modules and match the module to the known database, indicating that JDSNMF is effective in module analysis.

4.1. Dataset

We used a blood GE profile from the ANM consortium, a large cross-European AD biomarker study, and a follow-on Dementia Case Register (DCR) cohort in London. We downloaded the normalized RNA profile from the GEO: ANM Cohort 1 (GSE63060) and ANM Cohort 2 (GSE63061). The ANM Cohort 1 contains blood GE data of 145 patients with AD, 80 MCIs, and 104 NL, and the ANM Cohort 2 has those of 140 patients with AD, 109 MCIs, and 135 NLs after removing three borderline MCI samples and one other sample. We averaged the expression levels of probesets with the same RefSeq IDs, and then filtered out the uncommon RefSeq IDs between two cohorts (n = 20,017). We normalized ANM Cohorts 1 and 2 to reduce the batch effect using the Combat method from the ‘SVA’ R package [15].
We used a peripheral whole blood DM profile from the MRC London Brainbank for Neurodegenerative Disease [16]. We downloaded the DM profile from the GEO (GSE59685) that were converted to beta-values. The DM dataset had 80 subjects, some of which were annotated. We annotated 23 subjects that were not annotated in that study by the Braak stages [17,18] (i.e., Braak stages I and II—preclinical AD, Braak stages III and IV—prodromal AD, Braak stages V and VI—clinical AD). The DM data consisted of 68 patients with AD and 12 NLs. We mapped the CpG sites located within the promoter region, which is within a 200-base pair upstream region of the transcription start site (TSS200), to RefSeq IDs and averaged duplicated RefSeq IDs (n = 26,310). We filtered out the uncommon RefSeq IDs between the GE profiles and the DM profile. We also mapped RefSeq IDs to gene symbols and averaged the values of the same gene symbols in the GE and DM profiles (n = 12,011).
For the ADNI database, we downloaded robust multichip average normalized GE and raw DM profiles (IDAT format) from http://adni.loni.usc.edu (accessed on 1 September 2019). Both GE and DM profiles were obtained from whole blood. Because the GE profile did not contain diagnostic information, we obtained diagnosis of GE samples using the subject ID and visit code of the diagnostic information file (DXSUM_PDXCONV_ADNIALL.csv). The GE profile from ADNI consisted of 362 participants (NL: 246, AD: 116). We averaged the expression values of the same gene symbols (n = 18,727). Because ADNI has longitudinal data, there are samples with different diagnosis for each participant. Thus, we selected the DM samples examined on similar dates with GE samples for each participant. We obtained 294 participants (NL: 199, AD: 95), and we converted the IDAT file into a beta-value format using a “minifi” R package [19]. We mapped the CpG sites located within promoter regions (TSS200) to gene symbols. Then, we averaged β -values of the same gene symbols (n = 21,099). We filtered out the uncommon genes between the GE profiles and the DM profile (n = 15,922).

4.2. Hyperparameters and Settings for Experiments

In this section, the hyperparameters and settings of JDSNMF, NMF, deep semi-NMF, and classification models are described.
Our JDSNMF model has four key hyperparameters: the number of layers, the reduced dimensions of each layer, a layer to use for classification, and an L2 norm parameter. In the NMF method, the hyperparameters were the updated algorithm, the number of reduced dimensions, and the initialization method. The hyperparameters in the deep semi-NMF method were the same as those in the JDSNMF model.
For JDSNMF and deep semi-NMF, the Adam Optimizer [20] and early stopping were used. We used a sigmoid function as an activity function to make the decomposed sample latent matrix non-negative with non-linearity. The initialization strategy for a basis matrix and a coefficient matrix is an important step that affects the results. We used singular value decomposition (SVD), a popular initialization strategy in NMF tasks [21]. In a study by [22] using collective deep matrix factorization, SVD performed best when initializing each matrix to be decomposed.
We used an NMF method in the scikit-learn library [23] and implemented the deep semi-NMF method ourselves using Tensorflow [24]. JDSNMF is implemented using Tensorflow and code is available at https://github.com/dmcb-gist/Joint-deep-semi-NMF, accessed on 30 January 2020. The scikit-learn python library [23] for the linear support vector machine (SVM) and random forest (RF) based on 10,000 trees was used. For the deep neural network (DNN), the Tensorflow library [24] was used. For the DNN, we set three fully connected layers with L2 regularization and early stopping.

4.3. Sample-Matched Simulated Dataset Experiments

We evaluate the robustness performance and feature extraction performance of our proposed JDSNMF model on sample-matched simulated datasets.
Three datasets with small sample sizes and high dimensionality are simulated, while most of the features are irrelevant to the objective. We generated the predictor matrix X R S × F , drawn from a standard normal distribution, where S is the number of samples and F is the number of features. We set x s j as the sth sample and jth feature. The class label probability y s is generated by the logistic regression model as follows [25]:
y s = e x p β 0   +   f = 1 F β f x s f   +   ϵ 1   +   e x p β 0   +   f = 1 F β f x s f   +   ϵ ,
where β f represents the coefficient of jth feature, and ϵ represents random error generated by N ( 0 , 1.6 ) . Three simulated data are generated by the above procedure. We set 20 nonzero coefficient β and 980 zero coefficient β . The value of nonzero coefficient is from { ± 10 } . The nonzero coefficient features represent core features of three simulated data. We also choose the sample such that the class labels are consistent across three simulated data. We set training sample size as 750, validation sample size as 50, and test sample size as 100.
We evaluate the feature extraction performance with NMF, deep semi-NMF, and JDSNMF on simulated data (A,B,C). We measure the classification performance of the output from each feature extraction model using SVM. The optimal hyperparameters of each feature extraction model are searched by validation data. The hyperparameters of SVM were fixed. We repeated the analysis 20 times with different simulated datasets. Figure 3 shows the box plot analysis of the area under the curve (AUC) from test sets. We can easily show that integrating multi-omics data using JDSNMF improves the performance. We also observed that deep semi-NMF outperformed NMF. These results suggest that non-linearity and integration of multiple data are required to improve classification performance and that more important information for classification is extracted.

4.4. Feature-Matched Real Dataset Experiments

To evaluate our model’s ability to identify latent information predictive of AD from feature-matched omics data, we compared it with a feature selection approach based on differentially expressed genes, NMF, and a deep semi-NMF. The GE data of the ANM cohort were used for all models. Additionally, the DM data of the MRC London Brainbank cohort were used for JDSNMF.
Five-fold cross validation (CV) was employed to evaluate performance. In AD/NL classification, we divided samples into training, validation, and test sets (3:1:1 ratio) in each fold. The optimal hyperparameters were searched by validation data. In MCI/NL classification, we used optimal hyperparameter sets that had the best performance in the AD/NL classifications (see Supplementary Table S1 and the “Hyperparameters in MCI/NL classification on Addneuromed cohort ” section in the Additional Files for details). For the feature selection method, differentially expressed genes were selected using a training set with the “lmFit” function of the limma R package [26]. Statistically significant genes were obtained based on the adjusted p-value. For the baseline, we randomly selected the same number of genes selected in the feature selection model (an adjusted p-values < 5.00 × 10 4 ) in each CV fold. We measured the classification performance of the output from each model using DNN, SVM, and RF. The hyperparameters of DNN, SVM, and RF were fixed.
Figure 4 shows the averages of AUC values between true and false positives of three classifiers for each CV fold. The classification performances of each of the three classifiers are shown in Supplementary Figure S1. JDSNMF outperformed the other models in the classification of AD and NL (the average AUC of the three classifiers was 0.801) and in the classification of MCI and NL (the average AUC of the three classifiers was 0.771). When we applied a Wilcoxon signed-rank test for 15 AUC values (5 CVs × 3 classifiers), JDSNMF statistically significantly outperformed NMF and deep semi-NMF for the AD/NL classification and statistically significantly outperformed feature selection methods, NMF, and deep semi-NMF for MCI/NL classification. In the comparison of validation sets using various hyperparameters, we also confirmed that JDSNMF performed better in most situations (Supplementary Figures S2–S4). These results suggest that the DM data of other cohorts were helpful in extracting latent features and that JDSNMF is applicable to feature-matched dataset experiments as well.

4.5. Classification Performance Comparison According to the Reduced Dimension K on ADNI Cohort

The reduced dimension K is a critical hyperparameter in all matrix factorization models. If K is set too small, the data do not fit well to the model. On the other hand, if K is too large, overfitting occurs. In our experiments, K is determined through validation data. To further demonstrate the effects of K on the performance of JDSNMF, we compared the performance according to K. The classification performance is the average of the performances of the DNN, SVM, and RF models. The hyperparameters of DNN, SVM, and RF were fixed.
Figure 5 shows the classification performance where K varies from 10 to 150 for NMF, deep semi-NMF, and JDSNMF on the ADNI cohort. We observed that K did influence the classification performance of JDSNMF. However, JDSNMF was consistently outperformed by existing methods under a wide range of K values. In all K except for dimensions 10 and 130, the JDSNMF model outperformed other models. When we applied a Wilcoxon signed-rank test for 24 AUC values (8 reduced dimensions × 3 classifiers), JDSNMF statistically significantly outperformed deep semi-NMF (p-value < 0.05). The classification performances of each of the three classifiers are shown in Supplementary Figure S5.

4.6. Module Analysis

In this section, we attempt to interpret the clinical and biological meanings of the latent matrices from the GE data of the ANM cohort and the DM data of the MRC London Brainbank cohort. We identified AD- and age-related modules and matched the module to the known database. We used the GE sample latent matrix, DM sample latent matrix, and feature latent matrix from the JDSNMF model with the best performance in the “Feature-matched real dataset experiments” section (number of layers, three; reduced dimensions of each layer, 60, 59, 44; layer used for classification, third layer; and L2 norm parameter, 0.01). We constructed the 60 modules from the GE sample latent matrix, DM sample latent matrix, and feature latent matrix.

4.6.1. AD-Related Module

We used a locally interpretable model-agnostic explanation (LIME) [27] as the xAI method to identify the most important AD-related module in AD/NL classification. LIME is a model-agnostic technique that interprets complex models by approximating them locally. LIME learns to understand a black-box classifier by perturbing the input data and interpreting changes in prediction. We used LIME to locally approximate SVM, RF, and DNN models to the simple linear models to understand the models and identify the module that is important for AD classification. In addition, we calculated the relative importance of each module using RF.
Module 7 had the highest sum of LIME values for the three classifiers (Supplementary Figure S6). At the same time, module 7 was also the most important module for classification according to RF importance feature measures. Samples and significant genes included in module 7 are shown in Supplementary Figures S7 and S8. We performed a functional enrichment test of module 7 with KEGG pathways [28] using enrichR [29,30]. Table 1 shows the significantly enriched pathways (q-value < 0.0001) in module 7. Among them, the “Alzheimer disease” pathway was statistically significant (q-value = 9.00 × 10 5 ). In detail, 15 genes in module 7 overlapped with genes in the AD pathway. Among them, 3 (ATP5F1A, TNFRSF1A, and SNCA) were positive genes and 12 were negative genes (ATP5PF, ATP5PO, COX6A1, COX6C, COX7B, COX7C, NDUFB3, NDUFB2, NDUFA1, NDUFS5, NDUFS4, and UQCRQ). We surveyed these genes from previous studies for relevance to AD. SNCA ( α -synuclein) is widely known to play a role in the pathogenesis of AD. In a recent paper, elevated mRNA expression and low methylation of SNCA were observed in AD [31]. The samples included in module 7 showed similar results, with 12 AD samples and 1 NL sample in GE and 2 AD samples and 1 NL sample in DM. SNCA mRNA expression values of the module 7 samples were significantly higher than those of the rest of the samples (p-value of the two tailed t-test = 1.49 ×   10 2 ). SNCA methylation β -values of module 7 samples were significantly lower than those of the rest of the samples (p-value of the two tailed t-test = 2.40 ×   10 6 ). TNFRSF1A (TNF Receptor Superfamily Member 1A), known as TNFR1, may lead to neurotoxicity and mediates the A β -induced caspase-dependent death pathway in neuronal cells [32]. In a previous study [33], although the protein level of TNFR1 was increased in the neurons of patients with AD, no significant differences in TNFR1 mRNA expression were identified using in situ hybridization. However, we observed that TNFR1 mRNA expression levels of the module 7 samples (12 AD and 1 NL samples) were significantly higher than those of the rest of the samples (p-value of the two tailed t-test = 4.30 ×   10 4 ). As the previous study analyzed frontal cortex samples of only 12 AD and 12 NL cases, it is probable that TNFR1 mRNA expression levels would have differentially increased in a subgroup of Alzheimer’s patients. The ATP5F1A gene, among the positive genes, and the 12 negative genes are translated into subunits of ATP synthase, cytochrome c oxidase (COX), NADH-ubiquinone oxidoreductase, and ubiquinol-cytochrome c oxidoreductase, which are components of the mitochondrial respiratory chain. Previous studies reported that mitochondrial dysfunction is related to AD [34]. Reduced COX expression in the AD brain was observed in another study [35].
Other pathways are known to be directly or indirectly related to AD. For example, differential expression of oxidative phosphorylation genes is observed in AD [29]. In wild-type mice, non-alcoholic fatty liver disease has been shown to cause AD [36]. A relationship between natural killer cells and AD was also reported [37,38].
For each module, we performed linear regression analysis and calculated the Pearson correlation coefficient between the LIME importance score in the AD classification and the number of genes overlapped with the AD pathway in KEGG pathway analysis. Figure 6 shows the results for three example models: the best performing two-layer, three-layer, and four-layer models in the “Classification performance comparison on Addneuromed cohort” section (see Supplementary Figures S6, S9, and S10 for detail). In all three cases, there was a statistically significant relationship between importance in the AD prediction and the number of genes overlapped with the AD pathway.

4.6.2. Age-Related Module

To study other modules further, we used age information and found age-related modules by applying the following linear regression model to the GE and DM sample latent matrix.
Y i j = β j + γ j A g e i + δ j S e x i + μ j D i a g n o s i s i + k = 1 N α j k P C i k ,
where Y i j is the sample latent matrix value corresponding to sample i in module j, A g e i is the age of the sample i with coefficient γ j , S e x i is the sex of the sample i with coefficient δ j , D i a g n o s i s i is the AD state of the sample i with coefficient δ j , and P C i k denotes the value of the k-th principal component of the sample latent matrix for the i-th sample with coefficient α i j . N is the number of PCs that are not related to age. β j is an intercept for regression. We identified a relation with age for each module from γ j . As a result, modules 1, 5, and 22 were statistically significantly related to age (Supplementary Figure S11). We focused on module 1, the most statistically significant (p-value = 4.58 ×   10 3 and 3.53 ×   10 2 for GE and DM, respectively). In particular, γ j was positive for GE, indicating that GE values of genes in module 1 were positively correlated with age. In contrast, γ j was negative for DM, indicating that DM values of genes in module 1 were negatively correlated with age.
We overlapped the genes included in module 1 with gene ontology terms [39] and KEGG pathways using enrichR. A total of 34 GO terms and 44 pathways were significantly enriched (q-value < 0.05) in module 1 (Supplementary Table S2). The top terms in the list, such as “neutrophil-mediated immunity” (q-value: 2.51 ×   10 14 ), “neutrophil activation involved in immune response” (q-value = 2.9 ×   10 14 ), and “neutrophil activation involved in immune response” (q-value = 4.49 ×   10 14 ) are related to innate immunity functions of neutrophils. It is well known that the aging process deteriorates neutrophil functions such as phagocytic capacity, degranulation, and superoxide generation [40]. “Longevity regulating pathway” (q-value = 1.31 ×   10 2 ) is also significantly enriched, which is related to the aging process.
In addition, when we examined the overlaps between 240 genes in module 1 and 305 aging-related genes from the GenAge database [41] with Fisher’s exact test, the p-value was 9.99 ×   10 3 , showing the significance of the overlaps.

4.7. Experiments on Bioimaging Data

As brain imaging data are also important for the diagnosis of AD, we evaluated the feature extraction performance of our proposed JDSNMF model on sample-matched bioimaging datasets. The experiments were performed on magnetic resonance imaging (MRI) regions of interest (ROIs) and AV-45 amyloid positron emission tomography (PET) ROI datasets from ADNI-TADPOLE ( http://adni.loni.usc.edu, accessed on 13 July 2021). We selected 750 ADNI2 samples with both data types. The numbers of NL, significant memory concern (SMC), early MCI (EMCI), late MCI (LMCI), and AD are 180, 102, 170, 156, and 142, respectively.
We measured the classification performance of extracted features from each feature extraction model using SVM. The reduced dimension K of each model was selected by a validation set from each fold. The AUC values for one-vs-rest were calculated for multiclass classification. Table 2 shows the average AUC values of five-fold CV for NL/MCI, MCI/AD, NL/AD, and NL/SMC/EMCI/LMCI/AD classification tasks. JDSNMF had higher AUC values than others in the MCI/AD classification task. In the NL/AD and NL/SMC/EMCI/LMCI/AD classification tasks, JDSNMF had similar performance to or slightly higher than PET-based joint deep semi-NMF. We also observed that deep semi-NMF outperformed NMF on these three classification tasks. However, the performance of the non-linearity based models were worse than that of NMF in NL/MCI classification. Taken together, these results suggest that our model is generally applicable for various data types.

5. Discussion and Conclusions

We introduced a novel JDSNMF model—a non-linear feature extraction model—that can capture shared latent representations from complex multi-omics data. The proposed model is applicable to the existing sample-matched multi-omics integration and the new feature-matched multi-omics integration. JDSNMF demonstrated significant improvements in multi-omics feature extraction tasks on sample-matched simulated data and feature-matched multi-omics data. In AD experiments, we identified AD-related and age-related modules through the recent xAI model and traditional regression approaches. We observed that many AD-related genes were included in modules important for AD prediction and that genes in the age-related modules were significantly enriched with known aging-related genes. These results suggest that it is applicable not only to analysis of large cohorts with multi-omics data but also to integration analysis between single omics cohorts. Furthermore, feature-matched integration analysis can be applied to cross-species comparison, cross-disease comparison, and cross-cohort comparison. We also confirmed that our model is applicable to bioimaging data by further experiments.
However, this method has some limitations. Our model involves simple multi-task learning with uniform loss weights. This can lead to an optimization conflict while learning to reconstruct each omic datum. In the future, we will apply advanced multi-task learning methods [42] to the JDSNMF method.

Supplementary Materials

The following figures and tables are available online at https://www.mdpi.com/article/10.3390/jpm11080686/s1. Table S1: Top AUC scores of validation sets with various hyper-parameters in Addneuromed cohort. Figure S1: Comparison of the feature selection method, joint deep semi-NMF (JDSNMF) model, and other dimension reduction models using three classifiers on the Addneuromed cohort. Figure S2: Average AUC of validation sets using the two-layer model with hyperparameters (a layer used for classification and the number of reduced dimensions) on the Addneuromed cohort. Figure S3: Average AUC of validation sets using the three-layer model with hyperparameters (a layer used for classification and the number of reduced dimensions) on the Addneuromed cohort. Figure S4: Average AUCs of validation sets using the four-layer model with hyperparameters (a layer to use for classification and the number of reduced dimensions) on the Addneuromed cohort. Figure S5: Classification performance of the feature selection method, joint deep semi-NMF model (JDSNMF), and other dimension reduction models using three classifiers on the ADNI cohort. Figure S6: Identification of AD-related modules using the best performing three-layer model on the ANM validation set. Figure S7: Top 10 positive genes and negative genes in the AD-related module. Figure S8: Samples larger than a predefined positive threshold (empirically set to μ + 1.6449 σ ) in the AD-related module (module 7 of the best performing model in the ANM validation set). Figure S9: Identification of the AD-related module using best performing two-layer model in the ANM validation set. Figure S10: Identification of AD-related modules using the best performing four-layer model in the ANM validation set. Figure S11: Identification of modules related to age. Table S2: Significantly enriched gene ontology terms and KEGG pathways of genes in module 1.

Author Contributions

H.L. initiated and supervised the study. S.M. and H.L. developed the method. S.M. performed the experiments. All authors participated in writing the manuscript. Both authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Bio & Medical Technology Development Program of NRF funded by the Korean government (MSIT) (NRF-2018M3C7A1054935) and a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health & Welfare, Republic of Korea (grant number: HI18C0460).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Genetic data and bioimaging data used in this article were obtained from the Alzheimer’s disease Neuroimaging Initiative (ADNI) database https://ida.loni.usc.edu (accessed on 1 September 2019 and 13 July 2021, respectively). JDSNMF is publicly available at https://github.com/dmcb-gist/Joint-deep-semi-NMF (accessed on 30 January 2020).

Acknowledgments

Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf (accessed on 28 February 2021). Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). ADNI is funded by the National Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer’s Association; Drug Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Cogstate; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd. and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the National Institutes of Health (www.fnih.org). The grantee organization is the Northern California Institute for Research and Education, and the study is coordinated by the Alzheimer’s Therapeutic Research Institute at the University of Southern California. ADNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California.

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Argelaguet, R.; Velten, B.; Arnol, D.; Dietrich, S.; Zenz, T.; Marioni, J.C.; Buettner, F.; Huber, W.; Stegle, O. Multi-Omics Factor Analysis—A framework for unsupervised integration of multi-omics data sets. Mol. Syst. Biol. 2018, 14, e8124. [Google Scholar] [CrossRef]
  2. De Vito, R.; Bellio, R.; Trippa, L.; Parmigiani, G. Multi-study factor analysis. Biometrics 2019, 75, 337–346. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Žitnik, M.; Zupan, B. Data fusion by matrix factorization. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 41–53. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zhang, S.; Liu, C.C.; Li, W.; Shen, H.; Laird, P.W.; Zhou, X.J. Discovery of multi-dimensional modules by integrative analysis of cancer genomic data. Nucleic Acids Res. 2012, 40, 9379–9391. [Google Scholar] [CrossRef]
  5. Yang, Z.; Michailidis, G. A non-negative matrix factorization method for detecting modules in heterogeneous omics multi-modal data. Bioinformatics 2015, 32, 1–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Chalise, P.; Fridley, B.L. Integrative clustering of multi-level ‘omic data based on non-negative matrix factorization algorithm. PLoS ONE 2017, 12, e0176278. [Google Scholar]
  7. Shen, R.; Olshen, A.B.; Ladanyi, M. Integrative clustering of multiple genomic data types using a joint latent variable model with application to breast and lung cancer subtype analysis. Bioinformatics 2009, 25, 2906–2912. [Google Scholar] [CrossRef] [PubMed]
  8. Lock, E.F.; Hoadley, K.A.; Marron, J.S.; Nobel, A.B. Joint and individual variation explained (JIVE) for integrated analysis of multiple data types. Ann. Appl. Stat. 2013, 7, 523. [Google Scholar] [CrossRef]
  9. Trigeorgis, G.; Bousmalis, K.; Zafeiriou, S.; Schuller, B.W. A Deep semi-NMF Model for Learning Hidden Representations. In Proceedings of the 31st International Conference on International Conference on Machine Learning—Volume 32 (ICML’14), Bejing, China, 22–24 June 2014; pp. II-1692–II-1700. [Google Scholar]
  10. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788. [Google Scholar] [CrossRef] [PubMed]
  11. Seung, D.; Lee, L. Algorithms for non-negative matrix factorization. Adv. Neural Inf. Process. Syst. 2001, 13, 556–562. [Google Scholar]
  12. Ding, C.H.Q.; Li, T.; Jordan, M.I. Convex and Semi-Nonnegative Matrix Factorizations. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 45–55. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Brunet, J.P.; Tamayo, P.; Golub, T.R.; Mesirov, J.P. Metagenes and molecular pattern discovery using matrix factorization. Proc. Natl. Acad. Sci. USA 2004, 101, 4164–4169. [Google Scholar] [CrossRef] [Green Version]
  14. Kim, P.M.; Tidor, B. Subsystem identification through dimensionality reduction of large-scale gene expression data. Genome Res. 2003, 13, 1706–1718. [Google Scholar] [CrossRef] [Green Version]
  15. Leek, J.T.; Johnson, W.E.; Parker, H.S.; Jaffe, A.E.; Storey, J.D. The sva package for removing batch effects and other unwanted variation in high-throughput experiments. Bioinformatics 2012, 28, 882–883. [Google Scholar] [CrossRef]
  16. Lunnon, K.; Smith, R.; Hannon, E.; Jager, P.L.D.; Srivastava, G.; Volta, M.; Troakes, C.; Al-Sarraj, S.; Burrage, J.; Macdonald, R.; et al. Methylomic profiling implicates cortical deregulation of ANK1 in Alzheimer’s disease. Nat. Neurosci. 2014, 17, 1164–1170. [Google Scholar] [CrossRef] [Green Version]
  17. Goedert, M.; Clavaguera, F.; Tolnay, M. The propagation of prion-like protein inclusions in neurodegenerative diseases. Trends Neurosci. 2010, 33, 317–325. [Google Scholar] [CrossRef]
  18. Braak, H.; Braak, E. Neuropathological stageing of Alzheimer-related changes. Acta Neuropathol. 1991, 82, 239–259. [Google Scholar] [CrossRef]
  19. Aryee, M.J.; Jaffe, A.E.; Corrada-Bravo, H.; Ladd-Acosta, C.; Feinberg, A.P.; Hansen, K.D.; Irizarry, R.A. Minfi: A flexible and comprehensive Bioconductor package for the analysis of Infinium DNA methylation microarrays. Bioinformatics 2014, 30, 1363–1369. [Google Scholar] [CrossRef] [Green Version]
  20. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  21. Boutsidis, C.; Gallopoulos, E. SVD based initialization: A head start for nonnegative matrix factorization. Pattern Recognit. 2008, 41, 1350–1362. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, Q.; Sun, M.; Zhan, L.; Thompson, P.; Ji, S.; Zhou, J. Multi-Modality Disease Modeling via Collective Deep Matrix Factorization. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, USA, 13–17 August 2017; ACM: New York, NY, USA, 2017; pp. 1155–1164. [Google Scholar]
  23. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  24. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th Symposium on Operating Systems Design and Implementation (16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  25. Xia, L.Y.; Wang, Y.W.; Meng, D.Y.; Yao, X.J.; Chai, H.; Liang, Y. Descriptor selection via log-sum regularization for the biological activities of chemical structure. Int. J. Mol. Sci. 2018, 19, 30. [Google Scholar] [CrossRef] [Green Version]
  26. Smyth, G.K. Limma: Linear models for microarray data. In Bioinformatics and Computational Biology Solutions Using R and Bioconductor; Springer: Berlin/Heidelberg, Germany, 2005; pp. 397–420. [Google Scholar]
  27. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  28. Kanehisa, M.; Goto, S. KEGG: Kyoto encyclopedia of genes and genomes. Nucleic Acids Res. 2000, 28, 27–30. [Google Scholar] [CrossRef]
  29. Chen, E.Y.; Tan, C.M.; Kou, Y.; Duan, Q.; Wang, Z.; Meirelles, G.V.; Clark, N.R.; Ma’ayan, A. Enrichr: Interactive and collaborative HTML5 gene list enrichment analysis tool. BMC Bioinform. 2013, 14, 128. [Google Scholar] [CrossRef] [Green Version]
  30. Kuleshov, M.V.; Jones, M.R.; Rouillard, A.D.; Fernandez, N.F.; Duan, Q.; Wang, Z.; Koplev, S.; Jenkins, S.L.; Jagodnik, K.M.; Lachmann, A.; et al. Enrichr: A comprehensive gene set enrichment analysis web server 2016 update. Nucleic Acids Res. 2016, 44, W90–W97. [Google Scholar] [CrossRef] [Green Version]
  31. Yoshino, Y.; Mori, T.; Yoshida, T.; Yamazaki, K.; Ozaki, Y.; Sao, T.; Funahashi, Y.; Iga, J.i.; Ueno, S.i. Elevated mRNA expression and low methylation of SNCA in Japanese Alzheimer’s disease subjects. J. Alzheimer’s Dis. 2016, 54, 1349–1357. [Google Scholar] [CrossRef]
  32. Li, R.; Yang, L.; Lindholm, K.; Konishi, Y.; Yue, X.; Hampel, H.; Zhang, D.; Shen, Y. Tumor necrosis factor death receptor signaling cascade is required for amyloid-β protein-induced neuron death. J. Neurosci. 2004, 24, 1760–1771. [Google Scholar] [CrossRef] [Green Version]
  33. Cheng, X.; Yang, L.; He, P.; Li, R.; Shen, Y. Differential activation of tumor necrosis factor receptors distinguishes between brains from Alzheimer’s disease and non-demented patients. J. Alzheimer’s Dis. 2010, 19, 621–630. [Google Scholar] [CrossRef] [Green Version]
  34. Lin, M.T.; Beal, M.F. Mitochondrial dysfunction and oxidative stress in neurodegenerative diseases. Nature 2006, 443, 787. [Google Scholar] [CrossRef]
  35. Chandrasekaran, K.; Giordano, T.; Brady, D.R.; Stoll, J.; Martin, L.J.; Rapoport, S.I. Impairment in mitochondrial cytochrome oxidase gene expression in Alzheimer disease. Mol. Brain Res. 1994, 24, 336–340. [Google Scholar] [CrossRef]
  36. Kim, D.G.; Krenz, A.; Toussaint, L.E.; Maurer, K.J.; Robinson, S.A.; Yan, A.; Torres, L.; Bynoe, M.S. Non-alcoholic fatty liver disease induces signs of Alzheimer’s disease (AD) in wild-type mice and accelerates pathological signs of AD in an AD model. J. Neuroinflamm. 2016, 13, 1–18. [Google Scholar] [CrossRef] [Green Version]
  37. Solerte, S.B.; Fioravanti, M.; Severgnini, S.; Locatelli, M.; Renzullo, M.; Pezza, N.; Cerutti, N.; Ferrari, E. Enhanced cytotoxic response of natural killer cells to lnterleukin-2 in alzheimer’s disease. Dement. Geriatr. Cogn. Disord. 1996, 7, 343–348. [Google Scholar] [CrossRef]
  38. Jadidi-Niaragh, F.; Shegarfi, H.; Naddafi, F.; Mirshafiey, A. The role of natural killer cells in Alzheimer’s disease. Scand. J. Immunol. 2012, 76, 451–456. [Google Scholar] [CrossRef] [PubMed]
  39. Ashburner, M.; Ball, C.A.; Blake, J.A.; Botstein, D.; Butler, H.; Cherry, J.M.; Davis, A.P.; Dolinski, K.; Dwight, S.S.; Eppig, J.T.; et al. Gene ontology: Tool for the unification of biology. Nat. Genet. 2000, 25, 25–29. [Google Scholar] [CrossRef] [Green Version]
  40. Lord, J.M.; Butcher, S.; Killampali, V.; Lascelles, D.; Salmon, M. Neutrophil ageing and immunesenescence. Mech. Ageing Dev. 2001, 122, 1521–1535. [Google Scholar] [CrossRef]
  41. de Magalhaes, J.P.; Toussaint, O. GenAge: A genomic and proteomic network map of human ageing. FEBS Lett. 2004, 571, 243–247. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar]
Figure 1. Joint deep-semi NMF (JDSNMF) overview and downstreamanalyses. (A)Model overview: JDSNMF decomposes feature-matched (or sample-matched) multiple data. Taking Alzheimer’s disease experiments as an example, JDSNMF decomposes gene expression (GE) (X1) and DNA methylation (DM) (X2) data into a feature latent matrix (U), GE sample latent matrices (H10, …, H1N), DM sample latent matrices (H20, …, H2N), and junction latent matrices (Z11, …, Z1N, Z21, …, Z2N). (B) JDSNMF is used to integrate multi-omics datasets to latent matrices. These latent matrices, which capture the potential biological signatures that should be used in classification, are fed into the machine learning algorithmto generate a classification model. (C) Modules of features and samples were constructed using a feature latent matrix (U), a sample latent matrix of GE (H10), and a sample latent matrix of DM (H20). In U, H10, and H20, cyan, green, and blue elements were larger than positive thresholds, and in U, red elements were smaller than a negative threshold. (D,E) Biologically meaningful modules identified using an explainable artificial intelligence (xAI) and regression model.
Figure 1. Joint deep-semi NMF (JDSNMF) overview and downstreamanalyses. (A)Model overview: JDSNMF decomposes feature-matched (or sample-matched) multiple data. Taking Alzheimer’s disease experiments as an example, JDSNMF decomposes gene expression (GE) (X1) and DNA methylation (DM) (X2) data into a feature latent matrix (U), GE sample latent matrices (H10, …, H1N), DM sample latent matrices (H20, …, H2N), and junction latent matrices (Z11, …, Z1N, Z21, …, Z2N). (B) JDSNMF is used to integrate multi-omics datasets to latent matrices. These latent matrices, which capture the potential biological signatures that should be used in classification, are fed into the machine learning algorithmto generate a classification model. (C) Modules of features and samples were constructed using a feature latent matrix (U), a sample latent matrix of GE (H10), and a sample latent matrix of DM (H20). In U, H10, and H20, cyan, green, and blue elements were larger than positive thresholds, and in U, red elements were smaller than a negative threshold. (D,E) Biologically meaningful modules identified using an explainable artificial intelligence (xAI) and regression model.
Jpm 11 00686 g001
Figure 2. An example of the module construction. (A) Result of JDSNMF decomposition. In U, H 1 0 , and H 2 0 , cyan, green, and blue elements were larger than positive thresholds, and in U, H 1 0 , and H 2 0 , red elements were smaller than a negative threshold in U. (B) Genes and samples are coordinated for each axis, each of which represents a module. In U, H 1 0 , and H 2 0 , cyan, green, and blue elements are larger than positive thresholds, and in U, red elements are smaller than a negative threshold. Purple squares represent thresholds of the K3 module.
Figure 2. An example of the module construction. (A) Result of JDSNMF decomposition. In U, H 1 0 , and H 2 0 , cyan, green, and blue elements were larger than positive thresholds, and in U, H 1 0 , and H 2 0 , red elements were smaller than a negative threshold in U. (B) Genes and samples are coordinated for each axis, each of which represents a module. In U, H 1 0 , and H 2 0 , cyan, green, and blue elements are larger than positive thresholds, and in U, red elements are smaller than a negative threshold. Purple squares represent thresholds of the K3 module.
Jpm 11 00686 g002
Figure 3. Test AUC comparison of NMF, deep semi-NMF, and JDSNMF on simulated data.
Figure 3. Test AUC comparison of NMF, deep semi-NMF, and JDSNMF on simulated data.
Jpm 11 00686 g003
Figure 4. Comparison of the feature selection methods, NMF, deep semi-NMF, and JDSNMF, using the Addneuromed cohort. (A) AD/NL classification. (B) MCI/NL classification. Three classifiers (deep neural network, support vector machine, and random forest) were used as classifiers, and area under the curve (AUC) values between true positives and false positive rates were measured for classification performance. Then, the average AUC values of the three classifiers are shown for each CV fold. Stars indicate statistical significance with a Wilcoxon signed-rank test (*, p-value < 0.1; **, p-value < 0.05; ***, p-value < 0.01).
Figure 4. Comparison of the feature selection methods, NMF, deep semi-NMF, and JDSNMF, using the Addneuromed cohort. (A) AD/NL classification. (B) MCI/NL classification. Three classifiers (deep neural network, support vector machine, and random forest) were used as classifiers, and area under the curve (AUC) values between true positives and false positive rates were measured for classification performance. Then, the average AUC values of the three classifiers are shown for each CV fold. Stars indicate statistical significance with a Wilcoxon signed-rank test (*, p-value < 0.1; **, p-value < 0.05; ***, p-value < 0.01).
Jpm 11 00686 g004
Figure 5. Comparison of classification performance on the Alzheimer’s Disease Neuroimaging Initiative cohort. Average area under the curve (AUC) values of SVM, RF, and DNN for AD/NL classification by each model with respect to reduced dimensions.
Figure 5. Comparison of classification performance on the Alzheimer’s Disease Neuroimaging Initiative cohort. Average area under the curve (AUC) values of SVM, RF, and DNN for AD/NL classification by each model with respect to reduced dimensions.
Jpm 11 00686 g005
Figure 6. Relationship between the genes overlapping with Alzheimer’s disease (AD) KEGG pathway and importance in AD prediction. For each module, linear regression was performed, and the Pearson correlation was calculated for the relationship between the number of genes overlapping with the AD KEGG pathway and LIME importance for AD prediction. Three examples with different hyperparameters are shown: the best performing two-layer model (left), the best performing three-layer model (middle), and the best performing four-layer model (right) in the AD/NL classification task on the Addneuromed cohort.
Figure 6. Relationship between the genes overlapping with Alzheimer’s disease (AD) KEGG pathway and importance in AD prediction. For each module, linear regression was performed, and the Pearson correlation was calculated for the relationship between the number of genes overlapping with the AD KEGG pathway and LIME importance for AD prediction. Three examples with different hyperparameters are shown: the best performing two-layer model (left), the best performing three-layer model (middle), and the best performing four-layer model (right) in the AD/NL classification task on the Addneuromed cohort.
Jpm 11 00686 g006
Table 1. Enriched KEGG pathways of the module 7.
Table 1. Enriched KEGG pathways of the module 7.
KEGG PathwaysOverlapq-Value
Ribosome23/1537.35 ×  10 11
Oxidative phosphorylation19/1331.13 ×  10 8
Parkinson’s disease18/1421.75 ×  10 7
Phagosome18/1523.96 ×  10 7
Influenza A19/1713.46 ×  10 7
Thermogenesis21/2311.65 ×  10 6
Viral myocarditis11/591.99 ×  10 6
Graft-versus-host disease9/416.75 ×  10 6
Non-alcoholic fatty liver disease (NAFLD)15/1492.83 ×  10 5
Leukocyte transendothelial migration13/1122.73 ×  10 5
Natural killer cell-mediated cytotoxicity14/1312.61 ×  10 5
Toxoplasmosis13/1132.52 ×  10 5
Tuberculosis16/1794.18 ×  10 5
Pathogenic Escherichia coli infection9/555.36 ×  10 5
Leishmaniasis10/748.38 ×  10 5
Alzheimer disease15/1719.00 ×  10 5
Table 2. Performance comparison of NMF, deep semi-NMF, and JDSNMF on bioimaging data. Bold text indicates the best performance.
Table 2. Performance comparison of NMF, deep semi-NMF, and JDSNMF on bioimaging data. Bold text indicates the best performance.
TasksNMFDeep semi-NMFJDSNMF
MRIPETMRIPETMRI, PET
NL/MCI0.6300.7030.5690.6960.665
MCI/AD0.8310.8110.8380.8260.843
NL/AD0.9350.9630.9490.9720.973
NL/SMC/EMCI /LMCI/AD0.6240.6770.6780.7170.718
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Moon, S.; Lee, H. JDSNMF: Joint Deep Semi-Non-Negative Matrix Factorization for Learning Integrative Representation of Molecular Signals in Alzheimer’s Disease. J. Pers. Med. 2021, 11, 686. https://doi.org/10.3390/jpm11080686

AMA Style

Moon S, Lee H. JDSNMF: Joint Deep Semi-Non-Negative Matrix Factorization for Learning Integrative Representation of Molecular Signals in Alzheimer’s Disease. Journal of Personalized Medicine. 2021; 11(8):686. https://doi.org/10.3390/jpm11080686

Chicago/Turabian Style

Moon, Sehwan, and Hyunju Lee. 2021. "JDSNMF: Joint Deep Semi-Non-Negative Matrix Factorization for Learning Integrative Representation of Molecular Signals in Alzheimer’s Disease" Journal of Personalized Medicine 11, no. 8: 686. https://doi.org/10.3390/jpm11080686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop