Next Article in Journal
Split-GFP Reassembly Assay: Strengths and Caveats from a Multiparametric Analysis
Next Article in Special Issue
MetaSEM: Gene Regulatory Network Inference from Single-Cell RNA Data by Meta-Learning
Previous Article in Journal
Systematic Review of the Diagnostic and Clinical Utility of Salivary microRNAs in Traumatic Brain Injury (TBI)
Previous Article in Special Issue
Melatonin Inhibits hIAPP Oligomerization by Preventing β-Sheet and Hydrogen Bond Formation of the Amyloidogenic Region Revealed by Replica-Exchange Molecular Dynamics Simulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HLGNN-MDA: Heuristic Learning Based on Graph Neural Networks for miRNA–Disease Association Prediction

School of Computer Science and Technology, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(21), 13155; https://doi.org/10.3390/ijms232113155
Submission received: 16 August 2022 / Revised: 23 October 2022 / Accepted: 26 October 2022 / Published: 29 October 2022
(This article belongs to the Special Issue State-of-the-Art Molecular Biophysics in China)

Abstract

:
Identifying disease-related miRNAs can improve the understanding of complex diseases. However, experimentally finding the association between miRNAs and diseases is expensive in terms of time and resources. The computational screening of reliable miRNA–disease associations has thus become a necessary tool to guide biological experiments. “Similar miRNAs will be associated with the same disease” is the assumption on which most current miRNA–disease association prediction methods rely; however, biased prior knowledge, and incomplete and inaccurate miRNA similarity data and disease similarity data limit the performance of the model. Here, we propose heuristic learning based on graph neural networks to predict microRNA–disease associations (HLGNN-MDA). We learn the local graph topology features of the predicted miRNA–disease node pairs using graph neural networks. In particular, our improvements to the graph convolution layer of the graph neural network enable it to learn information among homogeneous nodes and among heterogeneous nodes. We illustrate the performance of HLGNN-MDA by performing tenfold cross-validation against excellent baseline models. The results show that we have promising performance in multiple metrics. We also focus on the role of the improvements to the graph convolution layer in the model. The case studies are supported by evidence on breast cancer, hepatocellular carcinoma and renal cell carcinoma. Given the above, the experiments demonstrate that HLGNN-MDA can serve as a reliable method to identify novel miRNA–disease associations.

1. Introduction

Since the discovery of microRNAs, an increasing number of researchers have investigated these molecules [1,2,3,4,5,6,7,8]. In particular, the discovery of a regulatory role for microRNAs in cellular activity suggests that these molecules are inextricably linked to many diseases [9,10,11,12,13]. Uncovering microRNA–disease associations has important implications for understanding disease mechanisms and assisting disease treatment [14,15,16,17,18,19,20]. However, due to the long time period required for biological experiments and the high resource costs, the use of computational methods to predict miRNA–disease associations has now become an important means for guiding traditional biological experiments, and it has greatly improved the efficiency of discovering disease-related miRNAs.
Some miRNA–disease association prediction models derive from combinatorial optimization theory and metric learning ideas, such as matrix-related operations and score estimation. MCMDA (matrix completion for miRNA–disease association) [21] performs matrix completion by applying a singular value thresholding algorithm on known miRNA–disease associations, and ILRMR (improved low-rank matrix recovery) [22] improves low-rank matrix recovery by referencing a weight matrix to enhance the prediction accuracy. MDMF (miRNA–disease Based on Matrix Factorization) [23] uses matrix factorization with disease similarity constraints to identify potential miRNA–disease associations. MDHGI (Decomposition and Heterogeneous Graph Inference) [24] discovers new miRNA–diseases associations by integrating the predicted association probability obtained from matrix decomposition through the sparse learning method. IMIPMF (inferring miRNA–disease interactions using probabilistic matrix factorization) [25] is a novel method for predicting miRNA–disease associations using probabilistic matrix factorization. WBSMDA (Within and Between Score for MiRNA–Disease Association prediction) [26] calculates a “Within and Between Score” for each miRNA–disease pair to predict the association between them. MLMD (Metric Learning for predicting miRNA–Disease) [27] is a novel computational model of metric learning for predicting miRNA–disease associations. It aims at learning miRNA–disease metrics to unravel not only novel disease-related miRNAs but also miRNA–miRNA and disease–disease similarities. DBNMDA (deep-belief network for predicting miRNA–disease associations) [28] constructs feature vectors to pre-train restricted Boltzmann machines for all miRNA–disease pairs and applies positive samples and the same number of selected negative samples to fine-tune a deep-belief network to obtain the final predicted scores.
Machine learning is also a class of methods [29,30] that are commonly applied to predict miRNA–disease associations [20,31,32,33,34,35,36,37]. RBMMMDA (restricted Boltzmann machine for multiple types of miRNA–disease associations) [38] proposes the restricted Boltzmann machine model to predict various types of miRNA–disease associations.
With the development of graph neural networks and the accumulation of large-scale graph data, in addition to traditional machine learning algorithms, DGCNN (multi-view multi-layer convolutional neural network) [39] and other deep learning [40,41,42,43,44,45] models have also been developed to deal with similar tasks. DGCNN focuses on large-scale and irregular network structures and adapts to the dynamic structure of local regions in the graph by flexibly designing convolutional filters. DeepMDA (predict miRNA–disease associations using deep learning) [46] uses a stacked self-encoder to obtain low-dimensional features from two high-dimensional feature vectors of miRNAs and diseases. A three-layer deep neural network [47] has then been developed to train classifiers of miRNA–disease feature pairs. MDA-CNN [48] constructs a three-layer miRNA–gene–disease association (MDA) network, and the network-based features of miRNAs and diseases are extracted using genes as the intermediate medium. The features are then downscaled using a self-encoder. Convolutional neural networks (CNNs) [49] are then used to further learn features from the miRNA–disease feature pairs. MDA-GCNFTG [50] predicts associations based on graph convolutional networks via graph sampling through the feature and topology graph to improve the training efficiency and accuracy. Instead of using heterogeneous graphs, MDA-GCNFTG constructs a homogeneous graph with MDPs (miRNA–disease pairs) as the nodes, which is the biggest difference with respect to our method. Although both use the GCN algorithm, the background graphs and the model focused on are completely different. For this task, the models focus on solving different problems.
However, there are still shortcomings in these recently proposed excellent computational models. Most of the current methods for predicting miRNA–disease associations are based on a strong assumption of similarity data. However, different models have different definitions of similarity, which makes the prediction results inaccurate. In addition, the miRNA functional similarity is incomplete and derived from known associations. Additionally, inconsistencies and incompleteness further lead to inaccurate prediction results. In this article, a heuristic learning method based on graph neural networks for miRNA–disease association prediction (HLGNN-MDA) is proposed. Inputting the whole miRNA–disease association network into the graph neural network for the model training causes high computational costs. To overcome this, we choose to train the graph neural network on enclosing subgraphs.
Figure 1 shows the overall framework of HLGNN-MDA. Our HLGNN-MDA model improves the graph neural network so that it can learn the information between miRNA and disease nodes and the topological relationships among homogeneous nodes at the same time. Compared with the previous computational models, our proposed method does not require too many similarity data and improves the applicability of graph neural networks in bipartite graph networks. More importantly, under such conditions, HLGNN-MDA can also achieve more accurate predictions.
Figure 1. Flowchart of HLGNN-MDA. (a) The enclosing subgraph of all pairs of nodes are extracted, and all the nodes in each enclosing subgraph are labeled. (b) The enclosing subgraphs are input into the graph neural network. The graph convolution layers adopt the three-layer structure model as shown in Figure 2 where each sub-figure of three in the second part corresponds to the processing results of three convolution modules. As shown in Figure 3, each convolution module has the same structure. The prediction results are obtained through the graph convolution layer, the graph pooling layer, 1D convolution and fully connected layers. Finally, the predicted results are verified against databases.
Figure 1. Flowchart of HLGNN-MDA. (a) The enclosing subgraph of all pairs of nodes are extracted, and all the nodes in each enclosing subgraph are labeled. (b) The enclosing subgraphs are input into the graph neural network. The graph convolution layers adopt the three-layer structure model as shown in Figure 2 where each sub-figure of three in the second part corresponds to the processing results of three convolution modules. As shown in Figure 3, each convolution module has the same structure. The prediction results are obtained through the graph convolution layer, the graph pooling layer, 1D convolution and fully connected layers. Finally, the predicted results are verified against databases.
Ijms 23 13155 g001

2. Results and Discussion

2.1. Performance Analysis of HLGNN-MDA Mode

In this section, we compare HLGNN-MDA with other related methods using tenfold cross-validation [20]. The algorithms selected for comparison were BLHARMDA (bipartite local models and hubness-aware regression for miRNA–disease association prediction) [51], BNPMDA (bipartite network projection for miRNA–disease association prediction) [52], IMCMDA (inductive matrix completion for miRNA–disease association prediction) [52], LFEMDA (predict miRNA–disease associations by latent features extraction) [53] and MKRMDA (multiple kernel learning-based Kronecker regularized least squares for miRNA–disease association prediction) [54]. All models for performance analysis experiments adopted the association data from HMDD v2.0. We also performed a performance comparison with the DGCNN model in the subsequent experiments of graph convolutional layer analysis, which is the baseline of our HLGNN-MDA convolutional module.
The miRNA functional similarity and disease semantic similarity were downloaded directly from IMCMDA. HLGNN-MDA used 5430 known miRNA–disease associations as positive samples and a random sample of the same number of candidate associations as negative samples. The dataset with the positive and negative samples together was then randomly divided into ten parts. One copy of each round was selected as the test set, and the remaining nine copies were used as the training set. The tenfold cross-validation was completed in turn. In each round, HLGNN-MDA removed the positive samples in the test set from the adjacency matrix. All five comparison algorithms completed tenfold cross-validation on the miRNA–disease association matrix. There were six evaluation metrics to analyze the model: ACC, precision, recall, AUROC, AUPR and MCC. The results are shown in Table 1, and their corresponding ROC curves are shown in Figure 4. The corresponding precision–recall (PR) curves are shown in Figure 5.
In Table 1, HLGNN-MDA-hopx represents the HLGNN-MDA model in which the enclosing subgraph’s hop is x (x = 1, 2, 3 or 4). Hops were used to extract subgraphs. The extraction process is described in Section 3.2.1. Compared with other methods, HLGNN-MDA-hopx had high performance in five of the six indicators. The minimum AUROC of HLGNN-MDA-hopx was greater than those of BNPMDA, IMCMDA, LFEMDA and MKRMDA. The minimum value of the AUPR of HLGNN-MDA-hopx was greater than the results of all other algorithms.
From the ROC curve in Figure 4, HLGNN-MDA-hop4 could cover the curves of BNPMDA, IMCMDA and MKRMDA. The maximum AUROC was 0.25% larger than the AUROC of BLHARMDA. Similarly, in the PR curve of Figure 5, HLGNN-MDA-hop4 could cover the PR curves of BNPMDA, IMCMDA and MKRMDA, which were roughly the same as that of BLHARMDA. The maximum AUPR value of HLGNN-MDA-hop4 was 0.55% larger than the maximum value of BLHARMDA.
Compared with the other algorithms, HLGNN-MDA had a relatively large advantage. Moreover, HLGNN-MDA used less similarity information to obtain better prediction results.

2.2. Influence of Different Hops in the Enclosing Subgraph

In this section, we discuss how HLGNN-MDA was affected by different enclosing subgraph hops. The required experimental data were miRNA–disease associations. The positive samples represented the known associations, and a random sample of the same number of candidates represented the negative samples. The test set was one-tenth the size of the entire dataset.
First, we evaluated the results of the HLGNN-MDA model using different hops in the enclosed subgraph. Each model was trained with the same training set and evaluated on the same test set. The range of the number of hops in the enclosing subgraph was 1 to 4. Their corresponding results are shown in Table 2. The ROC, PR and accuracy curves under different thresholds are shown in Figure 6, Figure 7 and Figure 8.
As shown in Table 2, HLGNN-MDA obtained the best value for all six metrics when the hops were 4. In Figure 6, Figure 7 and Figure 8, it can be seen that the curves with less hops were always covered by curves with larger hops. In particular, it should be noted that the results of HLGNN-MDA-hop2 had a larger increase than the results of HLGNN-MDA-hop1. The results of HLGNN-MDA-hop4 had a larger increase than the results of HLGNN-MDA-hop3. There was only a slight increase between HLGNN-MDA-hop3 and HLGNN-MDA-hop2, as shown in the three figures. These findings show that using an even larger hop in enclosing subgraphs for miRNA–disease association prediction could obtain better results.

2.3. Analysis of the Improved Graph Convolutional Layer

To show that our improvement in the graph neural network was effective, in this section, we present the analysis of the improved graph convolutional layer. After considering the information transfer between homogeneous nodes and heterogeneous nodes, HLGNN-MDA aggregated four propagation functions in the graph convolutional layer: A , A 2 , D 1 A and D 1 2 A D 1 2 .
Next, we explored the role of each propagation function in graph neural networks. If we deleted a certain propagation function from the current HLGNN-MDA but obtained a better result, it showed that the propagation function had a negative effect on the graph neural network. If the result of deleting a certain propagation function was worse, it had a positive effect on the graph neural network. Therefore, we discussed the role of the four propagation functions in turn in this way. The HLGNN-MDA model that removed propagation function A was marked as HLGNN-MDA-a. In this order, the HLGNN-MDA model deleting A 2 was marked as HLGNN-MDA-b; the model without D 1 A was marked as HLGNN-MDA-c; and the model without D 1 2 A D 1 2 was marked as HLGNN-MDA-d.
Then, we compared the HLGNN-MDA model with its four variations: HLGNN-MDA-a, HLGNN-MDA-b, HLGNN-MDA-c and HLGNN-MDA-d. As shown in Figure 9, the horizontal axis is the hop in the enclosing subgraph, and the vertical axis is the AUROC. The red line indicates HLGNN-MDA, and the yellow line indicates HLGNN-MDA-b, wherein A 2 was deleted from HLGNN-MDA; these variations were the most disparate. These results showed that A 2 had a greater impact on the model. The difference between the green fold of HLGNN-MDA-a and the red fold of HLGNN-MDA was minimal. However, when the hop number was 1, 2 or 4, HLGNN-MDA gave better results than HLGNN-MDA-a. Meanwhile, HLGNN-MDA-a and HLGNN-MDA-b all reached a maximum with hop = 3 and decreased slightly with hop = 4. Both propagation functions A and A 2 played an effective role in improving the predictive performance of the HLGNN-MDA model. From the figures, we could find that D 1 A (purple line) and D 1 2 A D 1 2 (blue line) had similar roles in graph neural networks and complemented each other. Meanwhile, if the evaluation metrics in Figure 9 were replaced with the AUPR and the ACC, a similar trend could be obtained.
Furthermore, since the graph neural network [55,56] of HLGNN-MDA is improved with respect to DGCNN, in order to show the effectiveness of HLGNN-MDA model, we also compared the performance of four variants of HLGNN-MDA with DGCNN with different hops. From Table 3, it could be concluded that DGCNN and HLGNN-MDA-b performed similarly. HLGNN-MDA-b was slightly higher than DGCNN when the enclosing subgraph hops were 1, 2 and 3. Compared with other HLGNN-MDA variant models, HLGNN-MDA-b, i.e., the A2-deleted model, had the worst performance among all models. As the value of hops increased, the six measures of HLGNN-MDA-b also increased slowly, and when hop = 3, HLGNN-MDA-b reached a maximum. This finding was consistent with the results presented in the above figures; that is, A 2 played a large role in the HLGNN-MDA model. Overall, HLGNN-MDA performed better than DGCNN in predicting miRNA–disease associations.
In summary, the four propagation functions in HLGNN-MDA all played a positive role, thus leading HLGNN-MDA to achieve good predictive performance. Of these, A and A 2 had a stronger role in improving the predictive performance of the model, while D 1 A and D 1 2 A D 1 2 enabled the model to be stable in its results with different hops of the enclosing subgraph. The combined use of these four propagation functions allowed HLGNN-MDA to perform well in the task of predicting miRNA–disease associations.

2.4. Validation of Prediction Results

In this section, all the known correlations in HMDD v2.0 were the training set for the model, and as many potential associations as possible were sampled as negative samples. According to the above analysis, HLGNN-MDA-hop4 gave the best predictions, so training was performed on it. All potential association relationships in HMDD v2.0 were then extracted and predicted on the trained HLGNN-MDA-hop4.
The prediction results were validated using the following databases: HMDD v3.0 [57], dbDEMC [58] and miR2Disease [59]. One association was considered to be validated if it was found in at least one database.
The final validation results were as follows: 10 out of the top 10 predictions were validated; a total of 49 out of the top 50 predictions were verified; a total of 97 out of the top 100 predictions were verified; and 169 out of the top 180 predictions were verified. The results demonstrated the effectiveness of HLGNN-MDA and its ability to predict potential novel associations.

2.5. Case Study

2.5.1. Breast Cancer

Breast cancer is one of the most dangerous malignancies to human health, especially for women. Globally, breast cancer accounts for 2.088 million new cases and 627,000 deaths per year, making it the number one malignancy in women [60]. The top 50 miRNAs predicted to be associated with breast cancer are listed in Table 4. In total, 49 of them were found in the relevant validation database. Only hsa-mir-362 (validation = no) does not present a record related to breast cancer in the three databases at present. Based on the literature validation [61], the hERG potassium channel, which enhances tumor aggressiveness and breast cancer proliferation, is transcriptionally regulated by hsa-miR-362-3p and thus associated with breast cancer growth. Another study [62] compared MDA-MB-231 and MCF7 breast cancer cell lines to the control CCD-1095Sk cell line, where hsa-miR-362-5p showed significant upregulation. The inhibition of hsa-miR-362-5p was found to significantly inhibit the diffusion, migration and invasion of MCF7 human breast cancer cells.

2.5.2. Hepatocellular Carcinoma

Primary liver cancer is the fifth most common cancer worldwide, mainly including hepatocellular carcinoma (HCC) [63,64]. A total of 49 of the top 50 miRNAs predicted to be related to hepatocellular carcinoma could be validated in three validation databases. The results are shown in Table 5. At present, no clear association between hsa-mir-495 and hepatocellular carcinoma could be found in these databases. However, a previous study [65] reported that hsa-mir-495 expression was frequently downregulated in hepatocellular carcinoma tissues and cell lines. Its expression levels were significantly correlated with tumor size, tumor lymph node metastasis (TNM) stage and lymph node metastasis in patients with hepatocellular carcinoma [65].

2.5.3. Renal Cell Carcinoma

Approximately 270,000 kidney cancer cases and 116,000 deaths are diagnosed annually worldwide [66]. Ninety percent of kidney cancers are tumors originating from the kidney epithelium and renal cell carcinoma [67]. The miRNAs predicted to be associated with the top 50 renal cell carcinomas are listed in Table 6. A total of 47 of the top 50 miRNAs could be validated clearly.

3. Materials and Methods

3.1. Data Resources

We collected human miRNA–disease associations from the HMDD v2.0 database [68] and obtained 5430 miRNA–disease associations between 495 miRNAs and 383 diseases. Therefore, many miRNA–disease associations were organized into an adjacency matrix YNnm × nd, where n m and n d represent the number of miRNAs and the number of diseases, respectively. If an association between miRNA m i and disease d j was recorded in HMDD v2.0, then Y ( i , j ) equaled 1; otherwise, it equaled 0.

3.2. Methods

3.2.1. Extraction of the Enclosing Subgraph of Node Pair

The design of our HLGNN-MDA model is inspired to the SEAL (learning from subgraphs, embedding and attributes for link prediction) [69] framework, which uses graph neural networks for link prediction. SEAL proves that most high-order heuristics can be approximated by learning from local enclosing subgraphs. Therefore, the first step of HLGNN-MDA is to extract the enclosing subgraph of all pairs of nodes. The enclosing subgraph is composed of two nodes, which are marked as central nodes, and their surrounding nodes. Then, the h -hop enclosing subgraph of central nodes m and d consists of all nodes whose distance from node m or node d is less than h steps.
HLGNN-MDA inputs the enclosing subgraphs into an “end-to-end” graph neural network for training. The association information of the central node pair of the enclosing subgraph is used as the supervisory label of the graph neural network. In association matrix Y , miRNA–disease associations equal to 1 are considered known, while those equal to 0 are treated as potential. The training set extracts known associations as positive samples and randomly samples an equal number of potential associations as negative samples. Subsequently, the enclosing subgraphs of all samples in the training set are extracted. To prevent the leakage of supervised labels, the edges between the central node pairs and the positive samples is removed during the extraction of the enclosing subgraph. The enclosing subgraph is denoted with A .

3.2.2. Label Nodes

In this section, we label all the nodes in each enclosing subgraph. Node labeling is the process of assigning an integer to a node in an enclosing subgraph and can be defined as f l : V N . Node labeling uses the DRNL (double-radius node labeling) method proposed in SEAL. First, we label the central two nodes with label 1. Other nodes with the same distance from the central two nodes are labeled with the same value. The farther the distance is, the greater the value is. The label of a node i can be derived from the following hash function:
f l ( i ) = 1 + min ( d x , d y ) + ( d / 2 ) [ ( d / 2 ) + ( d % 2 ) 1 ]
where d x and d y represent the distances from node i to central nodes x and y , respectively, with d = d x + d y ; and d / 2 and d % 2 indicate division and remainder, respectively. When a node is disconnected from the central nodes, it is labeled with 0.
In this way, we can label all the nodes in the enclosing subgraphs. Then, before being input into the graph neural network, the label of each node is expanded into one-hot encoding as its feature. This one-hot feature represents the position information of the node in the enclosing subgraph.

3.2.3. Construct Graph Neural Network

After extracting the enclosing subgraphs and labeling nodes in each enclosing subgraph, we can input the labeled enclosing subgraphs into the graph neural network for predicting miRNA–disease associations. At present, most of the proposed graph neural networks are applicable to homogeneous networks, whereas the miRNA–disease associations we use are in a bipartite graph network. Therefore, we improve the graph neural network DGCNN (deep graph convolutional neural network) to obtain better performance in heterogeneous networks.
Graph convolution layers. The role of the graph convolution layer is to learn the node representations [70,71]. A graph is input into the graph convolution layer through multilayer convolution, and the vector representation of each node can be extracted. The vector contains local substructural features of the graph. The process of graph convolution is the aggregation of feature information from the neighbors around each node.
Given the adjacency matrix of a graph A R n × n , its information matrix is X R n × d , where n indicates the number of nodes and d represents the dimension of the features. Graph convolution can be represented as follows:
Z = σ ( f ( A ) X W )
where W R d × c indicates the parameters to be trained, which converts the d-dimensional signal into c-dimensional signals; and f ( A ) represents the propagation function of adjacency matrix A . Usually, f ( A ) = D ~ 1 A ~ or f ( A ) = D ~ 1 2 A ~ D ~ 1 2 , where A ~ = A + I and D ~ are the degree matrices of A ~ . Moreover, f ( A ) X W indicates that the feature vector of each node is aggregated in the manner of a propagation function and then undergoes information conversion. In addition, σ ( ) is the activation function.
In the bipartite undirected network, miRNAs are only connected with diseases. That is, after a two-step jump, one disease node can only reach another disease, which is also true for miRNA nodes. Therefore, we use a propagation function for second-order topological information that allows information between homogeneous nodes to be aggregated together directly, i.e., defining a propagation function f ( A ) = A 2 .
With different propagation functions selected, the topological characteristics of the graph learned by the graph convolutional network are slightly different [72]. By splicing the neighbor information of nodes aggregated by different propagation functions, the graph convolutional layer can learn better graph topological features. The graph convolutional layer of our HLGNN-MDA model is shown in Figure 3 and can be represented by Formula (3):
Z t + 1 = σ ( [ A Z t , A 2 Z t , D 1 A Z t , D 1 2 A D 1 2 Z t ] W t )
where Z 0 = X and Z t R n × d t represent the output of the t -th graph convolutional layer; d t is the output dimension of the t -th layer graph convolution; and [ ] represents the splicing of row vectors, which splices the node vectors obtained through different propagation functions. The propagation functions used in our model are A , A 2 , D 1 A and D 1 2 A D 1 2 . After splicing, the topological features captured by these several propagation functions can be considered at the same time. W 4 d t × d t + 1 is the parameter to be trained, which maps the spliced node features from the 4 d t dimension to the d t + 1 dimension. Dimension 4 d t is used here because there are four propagation functions involved in the graph convolution process.
Enclosing subgraph A and its node features X go through t graph convolution layers to produce output Z t , t = 1 , , T . The overall graph convolution layer uses a global structure where the results of each layer are stitched together at the end, resulting in a result denoted as Z = [ Z 1 , , Z T ] . Each row of Z R n × 1 T d t is the 1 T d t -dimensional embedding vector representation of a node that contains rich topological information about that node in that graph. Figure 2 shows the overall architecture of the graph convolutional layer.
Graph pooling layers. A graph convolutional layer is used to learn a latent vector representation for each node. Here, the graph pooling layer of HLGNN-MDA selects the k most important nodes from the nodes of the graph to represent the graph. The importance of the node is evaluated using the result of the graph convolutional layer. The last layer of graph convolution maps the result of the previous layer to 1 dimension. That is, the parameter, W T , of the last layer of graph convolution maps dimension 4 d T 1 to 1 dimension. In this way, a value is obtained for each node, and this value indicates how important the node is in the graph.
We sort Z in descending order according to its last dimension. If two nodes appear to be equal in the last dimension of Z , we compare their penultimate dimension and so on, until the two nodes can be separated. The graph pooling layer takes the top k nodes in the ranking as its output, which helps subsequent conventional neural network layers to obtain a tensor with a fixed specification. When the number of nodes, n , is less than k , ( n k ) zero vectors are added after the sorted nodes.
Convolution layers and fully connected layer. After the graph pooling layer, a tensor Z = k × 1 T d t is obtained. In the remaining layers, we first use the traditional one-dimensional convolutional neural network combined with the max pooling layer to further refine the graph representation features and then make the final prediction using a fully connected layer.
To train a one-dimensional convolutional neural network on tensor Z , it first needs to be reshaped into a one-dimensional vector. To apply the filter for each node feature, the filter size and step length of the first one-dimensional convolutional neural network are set to 1 T d t ; that is, each node feature is convolved first. Then, after a max pooling layer, a one-dimensional convolutional neural layer is used to further learn the graph to represent the local features in the sequence features. Finally, it is connected to the fully connected layer. We use NLLLoss as the loss function, which adds up the predicted values of all predicted samples under the true labels. The predicted value here is a negative number in the logarithmic form of a normalized exponential function (softmax), mapping the prediction range from ( 0 , 1 ) to ( 0 , + ) . If all prediction samples are predicted correctly, NLLLoss is closer to 0. Finally, the fully connected layer outputs the probability of miRNA and disease node pair connections.

3.2.4. Evaluation Metrics

In order to verify the model performance, we choose the following metrics: AUROC (Receiver Operating Characteristic curve), ACC (accuracy), precision, recall, AUPR (area under the precision–recall curve) and MCC (Matthews correlation coefficient). The relevant definitions are as follows:
Precision = TP TP + FP
Recall = TP TP + TN
FPR = FP FP + TN
ACC = TP + TN TP + TN + FP + FN
MCC = TP TN × FP × FN ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN )
TP (true positive), FP (false positive), TN (true negative) and FN (false negative) were all derived from the confusion matrix. The ROC curve has the FPR (false positive rate) and TPR (recall or true positive rate) as the horizontal and vertical coordinates, respectively, and the area under the ROC curve is the AUC value. The area under the curve with recall and precision as the horizontal and vertical coordinates is the AUPR value.

4. Conclusions

Understanding the relationship between miRNAs and disease has important implications for disease prevention, detection and treatment. This paper proposes the HLGNN-MDA method, which is a heuristic for learning miRNA–disease association prediction from known miRNA–disease relationships based on graph neural networks. HLGNN-MDA first extracts the enclosing subgraphs around each miRNA–disease pair to be predicted to obtain the local network structure. Each node in the enclosing subgraphs is then labeled. The labeled subgraphs are then input into the graph neural network for classification. In particular, second-order topological information is added to the convolutional layer of the graph neural network to enable it to learn information between similar nodes. Second, different combinations of propagation functions are designed to improve the accuracy and stability of the graph neural network. We compared the model with the same type of miRNA–disease association prediction model using tenfold cross-validation. The results showed that HLGNN-MDA was able to obtain better performance than most miRNA–disease association prediction models. After discussing the effect of hop count on extracting closed subgraphs, we successively evaluated each propagation function combined in the model. Finally, we used the trained HLGNN-MDA model to make predictions and performed case studies on breast cancer, hepatocellular carcinoma and renal cell carcinoma. A total of 49 of the top 50 predicted miRNAs for breast cancer could be found in the validation database. The remaining hsa-mir-362 was also found to be associated with breast cancer, as supported by the literature. Similarly, 49 of the top 50 miRNAs predicted for hepatocellular carcinoma could be validated against the database. The remaining hsa-mir-495 was also found to be related to hepatocellular carcinoma, as supported by the literature. Finally, 47 of the top 50 miRNAs associated with renal cell carcinoma could be validated.
Generally, HLGNN-MDA has the following advantages: First, HLGNN-MDA can select arbitrary links for prediction without having to predict all potential miRNA–disease associations in the adjacency matrix. In particular, when predicting individual diseases or miRNAs, HLGNN-MDA can directly obtain the corresponding results. Second, HLGNN-MDA does not strictly require the corresponding similarity data because it can learn information through the topology of the network.
However, there are still some aspects for improvement. Valid miRNA and disease signatures are also important for prediction. Therefore, adding valid miRNA and disease signatures to HLGNN-MDA should be further investigated. Second, HLGNN-MDA is an end-to-end supervised learning algorithm framework. One of the problems with miRNA–disease association prediction is that there is no definite negative sample. For the potential associations, only a small proportion of them are truly associated, and most are unassociated. In this paper, the prediction of miRNA–disease association is approximated as a supervised learning model [73] with insufficient samples. Therefore, a new linkage prediction heuristic represents a future researchable direction.

Author Contributions

Conceptualization, L.Y. and B.J.; methodology, B.J.; software, B.J.; validation, B.J. and S.R.; formal analysis, B.J.; investigation, L.Y.; resources, L.Y.; data curation, B.J.; writing—original draft preparation, B.J.; writing—review and editing, S.R.; visualization, B.J. and S.R.; supervision, L.Y.; project administration, L.Y.; funding acquisition, L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China (grants Nos. 62072353 and 62132015).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Code of the model, datasets and results can be downloaded from GitHub (https://github.com/LiangYu-Xidian/HLGNN-MDA).

Acknowledgments

Thanks to all those who maintain excellent databases and to all experimentalists who enabled this work by making their data publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, R.C.; Feinbaum, R.L.; Ambros, V. The C. elegans heterochronic gene lin-4 encodes small RNAs with antisense com-plementarity to lin-14. Cell 1993, 75, 843–854. [Google Scholar] [CrossRef]
  2. Perez-Rodriguez, D.; Lopez-Fernandez, H.; Agis-Balboa, R.C. Application of miRNA-seq in neuropsychiatry: A method-ological perspective. Comput. Biol. Med. 2021, 135, 104603. [Google Scholar] [CrossRef]
  3. Cui, F.; Zhou, M.; Zou, Q. Computational biology and chemistry Special section editorial: Computational analyses for miRNA. Comput. Biol. Chem. 2021, 91, 107448. [Google Scholar] [CrossRef]
  4. Shaker, F.; Nikravesh, A.; Arezumand, R.; Aghaee-Bakhtiari, S.H. Web-based tools for miRNA studies analysis. Comput. Biol. Med. 2020, 127, 104060. [Google Scholar] [CrossRef]
  5. Wang, G.; Wang, Y.; Teng, M.; Zhang, D.; Li, L.; Liu, Y. Signal transducers and activators of transcription-1 (STAT1) regulates microRNA transcription in interferon gamma-stimulated HeLa cells. PLoS ONE 2010, 5, e11794. [Google Scholar]
  6. Zhao, Y.; Wang, F.; Juan, L. MicroRNA Promoter Identification in Arabidopsis using Multiple Histone Markers. BioMed Res. Int. 2015, 2015, 861402. [Google Scholar] [CrossRef] [Green Version]
  7. Zhao, Y.; Wang, F.; Chen, S.; Wan, J.; Wang, G. Methods of MicroRNA Promoter Prediction and Transcription Factor Mediated Regulatory Network. BioMed Res. Int. 2017, 2017, 7049406. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, X.; Yang, Y.; Liu, J.; Wang, G. The stacking strategy-based hybrid framework for identifying non-coding RNAs. Brief. Bioinform. 2021, 22, bbab023. [Google Scholar] [CrossRef]
  9. Tétreault, N.; De Guire, V. miRNAs: Their discovery, biogenesis and mechanism of action. Clin. Biochem. 2013, 46, 842–845. [Google Scholar] [CrossRef]
  10. Tian, L.; Wang, S.L. Exploring miRNA Sponge Networks of Breast Cancer by Combining miRNA-disease-lncRNA and miRNA-target Networks. Curr. Bioinform. 2021, 16, 385–394. [Google Scholar] [CrossRef]
  11. Han, W.; Lu, D.; Wang, C.; Cui, M.; Lu, K. Identification of Key mRNAs, miRNAs, and mRNA-miRNA Network Involved in Papillary Thyroid Carcinoma. Curr. Bioinform. 2021, 16, 146–153. [Google Scholar] [CrossRef]
  12. Sarkar, J.P.; Saha, I.; Sarkar, A.; Maulik, U. Machine learning integrated ensemble of feature selection methods followed by survival analysis for predicting breast cancer subtype specific miRNA biomarkers. Comput. Biol. Med. 2021, 131, 104244. [Google Scholar] [CrossRef] [PubMed]
  13. Liao, Z.; Li, D.; Wang, X.; Li, L.; Zou, Q. Cancer Diagnosis Through IsomiR Expression with Machine Learning Method. Curr. Bioinform. 2018, 13, 57–63. [Google Scholar] [CrossRef]
  14. Calin, G.A.; Dumitru, C.D.; Shimizu, M.; Bichi, R.; Zupo, S.; Noch, E.; Aldler, H.; Rattan, S.; Keating, M.; Rai, K.; et al. Frequent deletions and down-regulation of micro-RNA genes miR15 and miR16 at 13q14 in chronic lymphocytic leukemia. Proc. Natl. Acad. Sci. USA 2002, 99, 15524–15529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Lawrie, C.H.; Gal, S.; Dunlop, H.M.; Pushkaran, B.; Liggins, A.P.; Pulford, K.; Banham, A.H.; Pezzella, F.; Boultwood, J.; Wainscoat, J.S.; et al. Detection of elevated levels of tumour-associated microRNAs in serum of patients with diffuse large B-cell lymphoma. Br. J. Haematol. 2008, 141, 672–675. [Google Scholar] [CrossRef] [PubMed]
  16. Reddy, K.B. MicroRNA (miRNA) in cancer. Cancer Cell Int. 2015, 15, 38. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Khan, A.; Zahra, A.; Mumtaz, S.; Fatmi, M.Q.; Khan, M.J. Integrated In-silico Analysis to Study the Role of microRNAs in the Detection of Chronic Kidney Diseases. Curr. Bioinform. 2020, 15, 144–154. [Google Scholar] [CrossRef]
  18. Porta, C.; Figlin, R.A. MiR-193a-3p and miR-224 mediate renal cell carcinoma progression by targeting al-pha-2,3-sialyltransferase IV and the phosphatidylinositol 3 kinase/Akt pathway. Mol. Carcinog. 2019, 58, 1926–1927. [Google Scholar]
  19. Zhao, Z.; Zhang, C.; Li, M.; Yu, X.; Liu, H.; Chen, Q.; Wang, J.; Shen, S.; Jiang, J. Integrative Analysis of miRNA-mediated Competing Endogenous RNA Network Reveals the lncRNAs-mRNAs Interaction in Glioblastoma Stem Cell Differentiation. Curr. Bioinform. 2020, 15, 1187–1196. [Google Scholar] [CrossRef]
  20. Zhu, Q.; Fan, Y.; Pan, X. Fusing Multiple Biological Networks to Effectively Predict miRNA-disease Associations. Curr. Bioinform. 2021, 16, 371–384. [Google Scholar] [CrossRef]
  21. Li, J.Q.; Rong, Z.H.; Chen, X.; Yan, G.Y.; You, Z.H. MCMDA: Matrix completion for MiRNA-disease association prediction. Oncotarget 2017, 8, 21187–21199. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Peng, L.; Peng, M.; Liao, B.; Huang, G.; Liang, W.; Li, K. Improved low-rank matrix recovery method for predicting miRNA-disease association. Sci. Rep. 2017, 7, 6007. [Google Scholar] [CrossRef] [PubMed]
  23. Ha, J. MDMF: Predicting miRNA-Disease Association Based on Matrix Factorization with Disease Similarity Constraint. J. Pers. Med. 2022, 12, 885. [Google Scholar] [CrossRef]
  24. Chen, X.; Yin, J.; Qu, J.; Huang, L. MDHGI: Matrix Decomposition and Heterogeneous Graph Inference for miRNA-disease association prediction. PLoS Comput. Biol. 2018, 14, e1006418. [Google Scholar] [CrossRef] [PubMed]
  25. Ha, J.; Park, C.; Park, S. IMIPMF: Inferring miRNA-disease interactions using probabilistic matrix factorization. J. Biomed. Inform. 2020, 102, 103358. [Google Scholar] [CrossRef] [PubMed]
  26. Chen, X.; Yan, C.C.; Zhang, X.; You, Z.H.; Deng, L.; Liu, Y.; Zhang, Y.; Dai, Q. WBSMDA: Within and Between Score for MiRNA-Disease Association prediction. Sci. Rep. 2016, 6, 21106. [Google Scholar] [CrossRef] [PubMed]
  27. Ha, J.; Park, C. MLMD, MLMD: Metric Learning for predicting miRNA-Disease associations. IEEE Access 2021, 9, 78847–78858. [Google Scholar] [CrossRef]
  28. Chen, X.; Li, T.H.; Zhao, Y.; Wang, C.C.; Zhu, C.C. Deep-belief network for predicting potential miRNA-disease associations. Brief. Bioinform. 2021, 22, bbaa186. [Google Scholar]
  29. Zhang, Z.-Y.; Sun, Z.-J.; Yang, Y.-H.; Lin, H. Towards a better prediction of subcellular location of long non-coding RNA. Front. Comput. Sci. 2022, 16, 165903. [Google Scholar] [CrossRef]
  30. Yang, H.; Luo, Y.; Ren, X.; Wu, M.; He, X.; Peng, B.; Deng, K.; Yan, D.; Tang, H.; Lin, H. Risk Prediction of Diabetes: Big data mining with fusion of multi-farious physical examination indicators. Inf. Fusion 2021, 75, 140–149. [Google Scholar] [CrossRef]
  31. Lu, X.; Gao, Y.; Zhu, Z.; Ding, L.; Wang, X.; Liu, F.; Li, J. A Constrained Probabilistic Matrix Decomposition Method for Predicting miRNA-disease Associations. Curr. Bioinform. 2021, 16, 524–533. [Google Scholar] [CrossRef]
  32. Liu, Z.P. Predicting lncRNA-protein Interactions by Machine Learning Methods: A Review. Curr. Bioinform. 2020, 15, 831–840. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Duan, G.; Yan, C.; Yi, H.; Wu, F.X.; Wang, J. MDAPlatform: A Component-based Platform for Constructing and Assessing miRNA-disease Association Prediction Methods. Curr. Bioinform. 2021, 16, 710–721. [Google Scholar] [CrossRef]
  34. Zhang, X.; Zou, Q.; Rodriguez-Paton, A.; Zeng, X. Meta-Path Methods for Prioritizing Candidate Disease miRNAs. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 16, 283–291. [Google Scholar] [CrossRef] [PubMed]
  35. Zeng, X.; Liu, L.; Lü, L.; Zou, Q. Prediction of potential disease-associated microRNAs using structural perturbation method. Bioinformatics 2018, 34, 2425–2432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Dai, Q.; Chu, Y.; Li, Z.; Zhao, Y.; Mao, X.; Wang, Y.; Xiong, Y.; Wei, D.Q. MDA-CF: Predicting MiRNA-Disease associations based on a cascade forest model by fusing multi-source information. Comput. Biol. Med. 2021, 136, 104706. [Google Scholar] [CrossRef]
  37. Jiang, Q.; Wang, G.; Jin, S.; Li, Y.; Wang, Y. Predicting human microRNA-disease associations based on support vector machine. Int. J. Data Min. Bioinform. 2013, 8, 282–293. [Google Scholar] [CrossRef]
  38. Chen, X.; Clarence Yan, C.; Zhang, X.; Li, Z.; Deng, L.; Zhang, Y.; Dai, Q. RBMMMDA: Predicting multiple types of disease-microRNA associations. Sci. Rep. 2015, 5, 13877. [Google Scholar] [CrossRef] [Green Version]
  39. Phan, A.V.; Le Nguyen, M.; Nguyen, Y.L.; Bui, L.T. Dgcnn: A convolutional neural network over large-scale labeled graphs. Neural. Netw. 2018, 108, 533–543. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Yan, J.; Chen, S.; Gong, M.; Gao, D.; Zhu, M.; Gan, W. Review of the Applications of Deep Learning in Bioinformatics. Curr. Bioinform. 2021, 15, 898–911. [Google Scholar] [CrossRef]
  41. Ayachit, G.; Shaikh, I.; Pandya, H.; Das, J. Salient Features, Data and Algorithms for MicroRNA Screening from Plants: A Review on the Gains and Pitfalls of Machine Learning Techniques. Curr. Bioinform. 2021, 15, 1091–1103. [Google Scholar] [CrossRef]
  42. Chen, L.; Zhou, J.-P. Identification of Carcinogenic Chemicals with Network Embedding and Deep Learning Methods. Curr. Bioinform. 2021, 15, 1017–1026. [Google Scholar] [CrossRef]
  43. Wang, D.; Zhang, Z.; Jiang, Y.; Mao, Z.; Wang, D.; Lin, H.; Xu, D. DM3Loc: Multi-label mRNA subcellular localization prediction and analysis based on multi-head self-attention mechanism. Nucleic Acids Res. 2021, 49, e46. [Google Scholar] [CrossRef] [PubMed]
  44. Lv, H.; Dao, F.Y.; Zulfiqar, H.; Lin, H. DeepIPs: Comprehensive assessment and computational identification of phosphorylation sites of SARS-CoV-2 infection using a deep learning-based approach. Brief Bioinform. 2021, 22, bbab244. [Google Scholar] [CrossRef] [PubMed]
  45. Lv, H.; Dao, F.-Y.; Zulfiqar, H.; Su, W.; Ding, H.; Liu, L.; Lin, H. A sequence-based deep learning approach to predict CTCF-mediated chromatin loop. Brief. Bioinform. 2021, 22, bbab031. [Google Scholar] [CrossRef]
  46. Fu, L.; Peng, Q. A deep ensemble model to predict miRNA-disease association. Sci. Rep. 2017, 7, 14482. [Google Scholar] [CrossRef] [Green Version]
  47. Geete, K.; Pandey, M. Robust Transcription Factor Binding Site Prediction Using Deep Neural Networks. Curr. Bioinform. 2020, 15, 1137–1152. [Google Scholar] [CrossRef]
  48. Peng, J.; Hui, W.; Li, Q.; Chen, B.; Hao, J.; Jiang, Q.; Shang, X.; Wei, Z. A learning-based framework for miRNA-disease association identification using neural networks. Bioinformatics 2019, 35, 4364–4371. [Google Scholar] [CrossRef] [Green Version]
  49. Wu, B.; Zhang, H.; Lin, L.; Wang, H.; Gao, Y.; Zhao, L.; Chen, Y.-P.P.; Chen, R.; Gu, L. A Similarity Searching System for Biological Phenotype Images Using Deep Convolutional Encoder-decoder Architecture. Curr. Bioinform. 2019, 14, 628–639. [Google Scholar] [CrossRef]
  50. Chu, Y.; Wang, X.; Dai, Q.; Wang, Y.; Wang, Q.; Peng, S.; Wei, X.; Qiu, J.; Salahub, D.R.; Xiong, Y.; et al. MDA-GCNFTG: Identifying miRNA-disease associations based on graph convolutional networks via graph sampling through the feature and topology graph. Brief. Bioinform. 2021, 22, bbab165. [Google Scholar] [CrossRef]
  51. Chen, X.; Cheng, J.-Y.; Yin, J. Predicting microRNA-disease associations using bipartite local models and hubness-aware regression. RNA Biol. 2018, 15, 1192–1205. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Chen, X.; Wang, L.; Qu, J.; Guan, N.N.; Li, J.Q. Predicting miRNA-disease association based on inductive matrix completion. Bioinformatics 2018, 34, 4256–4265. [Google Scholar] [CrossRef] [PubMed]
  53. Che, K.; Guo, M.; Wang, C.; Liu, X.; Chen, X. Predicting MiRNA-Disease Association by Latent Feature Extraction with Positive Samples. Genes 2019, 10, 80. [Google Scholar] [CrossRef]
  54. Chen, X.; Niu, Y.W.; Wang, G.H.; Yan, G.Y. MKRMDA: Multiple kernel learning-based Kronecker regularized least squares for MiRNA-disease association prediction. J. Transl. Med. 2017, 15, 251. [Google Scholar] [CrossRef] [Green Version]
  55. Zhang, M.; Cui, Z.; Neumann, M.; Chen, Y. An End-to-End Deep Learning Architecture for Graph Classification. Proc. Conf. AAAI Artif. Intell. 2018, 32, 4438–4445. [Google Scholar] [CrossRef]
  56. Du, X.; Yao, Y. ConvsPPIS: Identifying Protein-protein Interaction Sites by an Ensemble Convolutional Neural Network with Feature Graph. Curr. Bioinform. 2020, 15, 368–378. [Google Scholar] [CrossRef]
  57. Huang, Z.; Shi, J.; Gao, Y.; Cui, C.; Zhang, S.; Li, J.; Zhou, Y.; Cui, Q. HMDD v3.0: A database for experimentally supported human microRNA-disease associations. Nucleic Acids Res. 2019, 47, D1013–D1017. [Google Scholar] [CrossRef] [Green Version]
  58. Yang, Z.; Wu, L.; Wang, A.; Tang, W.; Zhao, Y.; Zhao, H.; Teschendorff, A.E. dbDEMC 2.0: Updated database of differentially expressed miRNAs in human cancers. Nucleic Acids Res. 2016, 45, D812–D818. [Google Scholar] [CrossRef]
  59. Jiang, Q.; Wang, Y.; Hao, Y.; Juan, L.; Teng, M.; Zhang, X.; Li, M.; Wang, G.; Liu, Y. miR2Disease: A manually curated database for microRNA deregulation in human disease. Nucleic Acids Res. 2009, 37, D98–D104. [Google Scholar] [CrossRef] [Green Version]
  60. Tao, Z.; Shi, A.; Lu, C.; Song, T.; Zhang, Z.; Zhao, J. Breast Cancer: Epidemiology and Etiology. Cell Biophys. 2014, 72, 333–338. [Google Scholar] [CrossRef]
  61. Assiri, A.A.; Mourad, N.; Shao, M.; Kiel, P.; Liu, W.; Skaar, T.C.; Overholser, B.R. MicroRNA 362-3p Reduces hERG-related Current and Inhibits Breast Cancer Cells Proliferation. Cancer Genom. Proteom. 2019, 16, 433–442. [Google Scholar] [CrossRef] [PubMed]
  62. NI, F.; Gui, Z.; Guo, Q.; Hu, Z.; Wang, X.; Chen, D.; Wang, S. Downregulation of miR-362-5p inhibits proliferation, migration and invasion of human breast cancer MCF7 cells. Oncol. Lett. 2015, 11, 1155–1160. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. El-Serag, B.H.; Rudolph, L. Hepatocellular carcinoma: Epidemiology and molecular carcinogenesis. Gastroenterology 2007, 132, 2557–2576. [Google Scholar] [CrossRef] [PubMed]
  64. Jeyaram, C.; Philip, M.; Perumal, R.C.; Benny, J.; Jayakumari, J.M.; Ramasamy, M.S. A Computational Approach to Identify Novel Potential Precursor miRNAs and their Targets from Hepatocellular Carcinoma Cells. Curr. Bioinform. 2019, 14, 24–32. [Google Scholar] [CrossRef]
  65. Ye, Y.; Zhuang, J.; Wang, G.; He, S.; Zhang, S.; Wang, G.; Ni, J.; Wang, J.; Xia, W. MicroRNA-495 suppresses cell proliferation and invasion of hepatocellular carcinoma by directly targeting insulin-like growth factor receptor-1. Exp. Ther. Med. 2018, 15, 1150–1158. [Google Scholar]
  66. Ljungberg, B.; Campbell, S.C.; Cho, H.Y.; Jacqmin, D.; Lee, J.E.; Weikert, S.; Kiemeney, L.A. The Epidemiology of Renal Cell Carcinoma. Eur. Urol. 2011, 60, 1317. [Google Scholar] [CrossRef]
  67. Hsieh, J.J.; Purdue, M.P.; Signoretti, S.; Swanton, C.; Albiges, L.; Schmidinger, M.; Heng, D.Y.; Larkin, J.; Ficarra, V. Renal cell carcinoma. Nat. Rev. Dis. Prim. 2017, 3, 17009. [Google Scholar] [CrossRef]
  68. Li, Y.; Qiu, C.; Tu, J.; Geng, B.; Yang, J.; Jiang, T.; Cui, Q. HMDD v2.0: A database for experimentally supported human mi-croRNA and disease associations. Nucleic Acids Res. 2014, 42, D1070–D1074. [Google Scholar] [CrossRef] [Green Version]
  69. Zhang, M.; Chen, Y. Link prediction based on graph neural networks. In Advances in Neural Information Processing Systems 31, Proceedings of the 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2018. [Google Scholar]
  70. Thomas, N.; Kipf, M.W. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  71. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Advances in Neural Information Processing Systems 30, Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2017; pp. 1025–1035. [Google Scholar]
  72. Dehmamy, N.; Barabási, A.L.; Yu, R. Understanding the Representation Power of Graph Neural Networks in Learning Graph Topology. In Advances in Neural Information Processing Systems 32, Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2019. [Google Scholar]
  73. Lan, Y.; Li, Q. Supervised Learning in Spiking Neural Networks with Synaptic Delay Plasticity: An Overview. Curr. Bioinform. 2020, 15, 854–865. [Google Scholar] [CrossRef]
Figure 2. Three-layer graph convolutional layer structure of HLGNN-MDA.
Figure 2. Three-layer graph convolutional layer structure of HLGNN-MDA.
Ijms 23 13155 g002
Figure 3. Design of the graph convolution block for HLGNN-MDA.
Figure 3. Design of the graph convolution block for HLGNN-MDA.
Ijms 23 13155 g003
Figure 4. ROC curve of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms.
Figure 4. ROC curve of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms.
Ijms 23 13155 g004
Figure 5. PR curve of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms.
Figure 5. PR curve of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms.
Ijms 23 13155 g005
Figure 6. ROC curve of HLGNN-MDA in different enclosing subgraphs.
Figure 6. ROC curve of HLGNN-MDA in different enclosing subgraphs.
Ijms 23 13155 g006
Figure 7. PR curve of HLGNN-MDA in different enclosing subgraphs.
Figure 7. PR curve of HLGNN-MDA in different enclosing subgraphs.
Ijms 23 13155 g007
Figure 8. Accuracy curve of HLGNN-MDA using different enclosing subgraph hops.
Figure 8. Accuracy curve of HLGNN-MDA using different enclosing subgraph hops.
Ijms 23 13155 g008
Figure 9. AUC comparison between HLGNN-MDA and its four variations under different hops.
Figure 9. AUC comparison between HLGNN-MDA and its four variations under different hops.
Ijms 23 13155 g009
Table 1. Performance analysis of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms. The bold part indicates the maximum value in the corresponding column.
Table 1. Performance analysis of HLGNN-MDA and state-of-the-art miRNA–disease association prediction algorithms. The bold part indicates the maximum value in the corresponding column.
ModelACCPrecisionRecallAUROCAUPRMCC
BNPMDA0.790880.870690.683240.856480.882750.59574
IMCMDA0.772740.801020.725780.840040.849890.54791
LFEMDA0.847510.855900.835730.900390.912890.69522
BLHARMDA0.854420.856190.851930.928380.926990.70885
MKRMDA0.845490.876100.804790.896580.919710.69328
HLGNN-MDA-hop10.854420.862630.843090.929740.927790.70902
HLGNN-MDA-hop20.859760.859170.860590.928330.929270.71952
HLGNN-MDA-hop30.856350.867450.841250.928630.930070.71303
HLGNN-MDA-hop40.859120.867090.848250.930860.932470.71840
Table 2. Influence of different hops in enclosing subgraphs on HLGNN-MDA. The bold part indicates the maximum value in the corresponding column.
Table 2. Influence of different hops in enclosing subgraphs on HLGNN-MDA. The bold part indicates the maximum value in the corresponding column.
ModelACCPrecisionRecallAUROCAUPRMCC
HLGNN-MDA-hop10.881220.904300.852670.935350.932810.76368
HLGNN-MDA-hop20.927260.939390.913440.972120.975640.85484
HLGNN-MDA-hop30.938310.959460.915290.972660.977440.87754
HLGNN-MDA-hop40.968690.966900.970530.991780.993320.93739
Table 3. Comparison between the four variations of HLGNN-MDA and DGCNN with different hops.
Table 3. Comparison between the four variations of HLGNN-MDA and DGCNN with different hops.
ModelACCPrecisionRecallAUROCAUPRMCC
HLGNN-MDA-a-hop10.858200.911210.793740.927950.923030.72242
HLGNN-MDA-a-hop20.900550.905030.895030.946810.946290.80115
HLGNN-MDA-a-hop30.948430.945160.952120.985380.986260.89689
HLGNN-MDA-a-hop40.931860.951830.909760.975350.979450.86456
HLGNN-MDA-b-hop10.860960.901640.810310.933690.934120.72565
HLGNN-MDA-b-hop20.878450.923710.825050.937210.941060.76126
HLGNN-MDA-b-hop30.890420.915690.860040.942650.945750.78229
HLGNN-MDA-b-hop40.888580.907340.865560.942570.948770.77799
HLGNN-MDA-c-hop10.856350.841270.878450.927290.922410.71340
HLGNN-MDA-c-hop20.922650.916520.930020.966730.967290.84540
HLGNN-MDA-c-hop30.883980.924640.836100.936910.943000.77150
HLGNN-MDA-c-hop40.939230.963110.913440.976100.979010.87962
HLGNN-MDA-d-hop10.862800.898790.817680.929990.925860.72857
HLGNN-MDA-d-hop20.938310.942380.933700.980120.980810.87665
HLGNN-MDA-d-hop30.875690.923240.819520.932660.940600.75617
HLGNN-MDA-d-hop40.940150.954370.924490.985850.987020.88073
DGCNN-hop10.858200.888220.819520.928890.928890.71854
DGCNN-hop20.872010.904000.832410.935090.938310.74636
DGCNN-hop30.885820.905220.861880.942410.940830.88302
DGCNN-hop40.894110.919610.863720.952500.957070.89079
Table 4. Top 50 miRNAs predicted by HLGNN-MDA to be associated with breast cancer.
Table 4. Top 50 miRNAs predicted by HLGNN-MDA to be associated with breast cancer.
RankMicroRNAValidationRankMicroRNAValidation
1hsa-mir-211yes <H, D>26hsa-mir-30eyes <H, D>
2hsa-mir-186yes <D>27hsa-mir-494yes <H, D>
3hsa-mir-744yes <H, D>28hsa-mir-421yes <H, D>
4hsa-mir-138yes <H, D>29hsa-mir-501yes <H, D>
5hsa-mir-154yes <D>30hsa-mir-99byes <H, D>
6hsa-mir-216byes <H, D>31hsa-mir-196byes <H, D>
7hsa-mir-106ayes <H, D>32hsa-mir-185yes <H, D>
8hsa-mir-432yes <H, D>33hsa-mir-484yes <H, D>
9hsa-mir-32yes <H, D>34hsa-mir-144yes <H, D>
10hsa-mir-381yes <H, D>35hsa-mir-592yes <H, D>
11hsa-mir-142yes <H, D>36hsa-mir-130ayes <H, D>
12hsa-mir-150yes <H, D>37hsa-mir-542yes <H, D>
13hsa-mir-491yes <H, D>38hsa-mir-1224yes <H, D>
14hsa-mir-449ayes <H, D>39hsa-mir-376ayes <H, D>
15hsa-mir-362no40hsa-mir-451yes <H, D, M>
16hsa-mir-28yes <H, D>41hsa-mir-433yes <H, D>
17hsa-mir-378ayes <H, D>42hsa-mir-483yes <H, D>
18hsa-mir-212yes <H, D>43hsa-mir-1207yes <H, D>
19hsa-mir-98yes <H, D, M>44hsa-mir-33byes <H, D>
20hsa-mir-92byes <H, D>45hsa-mir-15byes <H, D>
21hsa-mir-455yes <H, D>46hsa-mir-630yes <H, D>
22hsa-mir-590yes <H, D>47hsa-mir-622yes <H, D>
23hsa-mir-330yes <H, D>48hsa-mir-1271yes <H, D>
24hsa-mir-675yes <H, D>49hsa-mir-424yes <H, D>
25hsa-mir-217yes <H, D>50hsa-mir-95yes <H, D>
Note: H <HMDD v3.0>, D <dbDEMC> and M <miR2Disease > represent the databases in which the relations could be validated.
Table 5. Top 50 miRNAs predicted by HLGNN-MDA to be associated with hepatocellular carcinoma.
Table 5. Top 50 miRNAs predicted by HLGNN-MDA to be associated with hepatocellular carcinoma.
RankMicroRNAValidationRankMicroRNAValidation
1hsa-mir-143yes <H, D, M>26hsa-mir-23byes <H, D, M>
2hsa-mir-196byes <H, D>27hsa-mir-574yes <H, D>
3hsa-mir-137yes <H, D, M>28hsa-mir-26byes <H, D, M>
4hsa-mir-520cyes <H, D>29hsa-mir-495no
5hsa-mir-376cyes <H, D>30hsa-mir-328yes <H, D, M>
6hsa-mir-184yes <H, D>31hsa-mir-452yes <H, D>
7hsa-mir-215yes <H, D, M>32hsa-mir-204yes <H, D>
8hsa-mir-302ayes <H, D>33hsa-mir-135byes <H, D>
9hsa-mir-34byes <H, D>34hsa-mir-95yes <H, D>
10hsa-mir-339yes <H, D>35hsa-mir-185yes <H, D, M>
11hsa-mir-708yes <H, D>36hsa-mir-206yes <H, D>
12hsa-mir-193yes <H, D>37hsa-mir-449ayes <H, D>
13hsa-mir-30eyes <H, D, M>38hsa-mir-520ayes <H, D>
14hsa-mir-488yes <H, D>39hsa-mir-194yes <H, D, M>
15hsa-mir-200yes <H, M>40hsa-mir-451yes <H, D>
16hsa-mir-342yes <H, D>41hsa-mir-149yes <H, D>
17hsa-mir-367yes <H, D>42hsa-mir-153yes <H, D>
18hsa-mir-302dyes <H, D>43hsa-mir-299yes <H, D>
19hsa-mir-494yes <H, D>44hsa-mir133ayes <H, D, M>
20hsa-mir-128yes <H, D, M>45hsa-mir-633yes <D>
21hsa-mir-340yes <H, D>46hsa-mir-132yes <H, D, M>
22hsa-mir-33byes <H, D>47hsa-mir-27byes <H, D>
23hsa-mir-625yes <H, D>48hsa-mir-935yes <H, D>
24hsa-mir-424yes <H, D>49hsa-mir-32yes <H, D>
25hsa-mir-151byes <H, D>50hsa-mir-186yes <H, D, M>
Note: H <HMDD v3.0>, D <dbDEMC> and M <miR2Disease > represent the databases in which the relations could be validated.
Table 6. Top 50 miRNAs predicted by HLGNN-MDA to be associated with renal cell carcinoma.
Table 6. Top 50 miRNAs predicted by HLGNN-MDA to be associated with renal cell carcinoma.
RankMicroRNAValidationRankMicroRNAValidation
1hsa-mir-20ayes <H, D, M>26hsa-mir-181ayes <H, D>
2hsa-mir-17yes <H, D, M>27hsa-mir-192yes <H, D>
3hsa-mir-27byes <H, D>28hsa-mir-22yes <H, D>
4hsa-mir-221yes <H, D, M>29hsa-mir-182yes <H, D, M>
5hsa-mir-223yes <H, D, M>30hsa-mir-29byes <H, D, M>
6hsa-mir-31yes <H, D>31hsa-mir-15ayes <H, D, M>
7hsa-mir-29ayes <H, D, M>32hsa-mir-375yes <H, D>
8hsa-mir-125byes <H, D>33hsa-mir-486yes <D>
9hsa-mir-133ayes <H, D, M>34hsa-mir-15byes <H, D>
10hsa-mir-125ayes <H, D>35hsa-mir-107yes <H, D>
11hsa-mir-18ayes <H, D>36hsa-mir-328yes <D>
12hsa-mir-1yes <H, D>37hsa-mir-23ayes <D>
13hsa-mir-30ayes <H, D, M>38hsa-mir-194yes <H, D>
14hsa-mir-181byes <H, D>39hsa-mir-193byes <H, D>
15hsa-mir-19byes <H, D, M>40hsa-mir-196byes <D>
16hsa-mir-214yes <H, D, M>41hsa-mir-137yes <H, D>
17hsa-mir-130ayes <H, D>42hsa-mir-191yes <H, D, M>
18hsa-mir-222yes <H, D>43hsa-mir-302ano
19hsa-mir-148ayes <H, D>44hsa-mir-135byes <D>
20hsa-mir-25yes <D>45hsa-mir-451bno
21hsa-mir-133byes <H, D>46hsa-mir-342yes <D, M>
22hsa-mir-183yes <H, D>47hsa-mir-30byes <H, D>
23hsa-mir-106ayes <H, D, M>48hsa-mir-373no
24hsa-mir-24yes <D>49hsa-mir-212yes <D>
25hsa-mir-132yes <D>50hsa-mir-193ayes <H, D>
Note: H <HMDD v3.0>, D <dbDEMC> and M <miR2Disease > represent the databases in which the relations could be validated.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yu, L.; Ju, B.; Ren, S. HLGNN-MDA: Heuristic Learning Based on Graph Neural Networks for miRNA–Disease Association Prediction. Int. J. Mol. Sci. 2022, 23, 13155. https://doi.org/10.3390/ijms232113155

AMA Style

Yu L, Ju B, Ren S. HLGNN-MDA: Heuristic Learning Based on Graph Neural Networks for miRNA–Disease Association Prediction. International Journal of Molecular Sciences. 2022; 23(21):13155. https://doi.org/10.3390/ijms232113155

Chicago/Turabian Style

Yu, Liang, Bingyi Ju, and Shujie Ren. 2022. "HLGNN-MDA: Heuristic Learning Based on Graph Neural Networks for miRNA–Disease Association Prediction" International Journal of Molecular Sciences 23, no. 21: 13155. https://doi.org/10.3390/ijms232113155

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop