Skip to Content
EntropyEntropy
  • Article
  • Open Access

27 April 2022

Deep Link-Prediction Based on the Local Structure of Bipartite Networks

,
,
and
1
School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
2
School of Computer and Information Engineering, Shanghai Polytechnic University, Shanghai 201209, China
*
Author to whom correspondence should be addressed.

Abstract

Link prediction based on bipartite networks can not only mine hidden relationships between different types of nodes, but also reveal the inherent law of network evolution. Existing bipartite network link prediction is mainly based on the global structure that cannot analyze the role of the local structure in link prediction. To tackle this problem, this paper proposes a deep link-prediction (DLP) method by leveraging the local structure of bipartite networks. The method first extracts the local structure between target nodes and observes structural information between nodes from a local perspective. Then, representation learning of the local structure is performed on the basis of the graph neural network to extract latent features between target nodes. Lastly, a deep-link prediction model is trained on the basis of latent features between target nodes to achieve link prediction. Experimental results on five datasets showed that DLP achieved significant improvement over existing state-of-the-art link prediction methods. In addition, this paper analyzes the relationship between local structure and link prediction, confirming the effectiveness of a local structure in link prediction.

1. Introduction

A bipartite network contains two types of nodes, and there are edges between different types of nodes [1]. Link prediction based on bipartite networks aims to predict the possibility of a link between two different types of nodes in the network [2,3]. Bipartite networks are ubiquitous, such as user–item purchase networks, drug–disease treatment networks, and author–subject research networks [4,5,6]. Link prediction based on bipartite networks can not only mine the hidden edge relationship between different types of nodes to provide the basis for network application, but also reveal the inherent law of network evolution to realize the dynamic analysis of the network.
Traditional link prediction is mainly based on the structure of the homogeneous network for target nodes. For example, common neighbor (CN) and preferential attachment (PA) mainly perform link prediction on the basis of common neighbors of target nodes by only involving the single-hop neighbors of target nodes [7]. Adamic–Adar (AA) and resource allocation (RA) are based on common neighbors of the neighbors of the target nodes as the link prediction basis, mainly based on two-hop neighbors of target nodes [8]. Katz is based on all paths connected by target nodes, involving multihop neighbors between target nodes [9]. Different from homogeneous networks, the neighbor of node u in a bipartite network is always node v, and the neighbor of node v is always node u, where u and v denote two types of nodes. For the structural characteristics of a bipartite network, similar relationships between neighbors still exist. For example, in a bipartite network, edges only exist between nodes of different types, and there are no edges between nodes of the same type. As shown in Figure 1, for different types of nodes u 0 and v 0 , node u 0 reaches node v 0 through three hops, which is the shortest path between different types of nodes. Therefore, three-hop paths between node u 0 and v 0 can be used as indicators for link prediction. For example, by modeling drug–protein bipartite networks in biological networks, and on the basis of connectivity paths between target nodes, interactions between drugs and target proteins can be effectively achieved [10]. The connection of user–item–user–item in the recommender system is a recommendation indicator. The essence of recommendation is to predict the relationship between users and items, which is actually a variant of link prediction [11,12]. These link predictions are mainly based on the connectivity paths between target nodes in which network structure information is extracted. On the basis of the length of the connected paths between target nodes, extracted network structure information is different.
Figure 1. Link relationship between u 0 and v 0 in a bipartite network, where dashed lines represent possible edges.
With the development of deep learning, deep-link prediction based on network structure is receiving attention, such as link prediction combined with deep learning and random walks [13,14,15,16,17]. First, these methods extract structural network features on the basis of random walks to capture latent information in bipartite networks. Then, on the basis of the deep-learning framework, the results of random walks are trained to obtain a deep-link prediction model for between target nodes. The strategy based on deep matrix factorization transforms the link prediction problem into a matrix completion problem. This strategy stores the structural information of the bipartite network on the basis of the adjacency matrix and realizes the completion of the matrix by learning explicit scoring and implicit feedback in the matrix through the deep neural network [18,19]. In fact, the network structure contains the feature information of links between nodes, and the idea is to automatically learn these features from the network structure. Link prediction based on heterogeneous graph neural networks can automatically learn the structural features in the network and realize link prediction between different types of nodes [20,21,22,23].
Through the above analysis, existing link prediction methods are mainly based on the global structure of a bipartite network, relying on structural and feature information between nodes for analysis, and they lack more refined local structural analysis. To tackle this problem, this paper proposes deep-link prediction (DLP) based on the local structure of bipartite networks to leverage the local structural information of the network to enhance prediction performance. First, the local structure between target nodes is extracted, the representation learning of the local structure is then performed on the basis of the graph neural network, and deep-link prediction is lastly realized on the basis of local structure features. The main contributions in our paper are as follows:
  • The DLP method is proposed on the basis of the local structure of the bipartite network, demonstrating the rationality of the local structure in link prediction.
  • A local structure extraction algorithm was designed to effectively extract local structures between target nodes of the bipartitie network and provide local structural information for link prediction.
  • Experimental results on five datasets demonstrate the superiority of DLP over state-of-the-art link prediction methods.

3. DLP Method

The specific framework of DLP is shown in Figure 2. The method can extract the local structure around the target nodes and obtain the possibility of links through local structural features. First, DLP extracts the local structure between target nodes. Then, the local structure and node features are input into the GCN model to obtain the feature representation of the local structure. Lastly, a deep-link prediction model is trained on the basis of feature representation and the link relationship between target nodes to achieve link prediction.
Figure 2. Framework of DLP method.

3.1. Local Structure Extraction

DLP extracts the local structure between target nodes, and on the basis of the local structure, it is used as the basis to examine whether there is a link between target nodes u and v. We designed a local structure extraction algorithm of bipartite networks that can generate all k-hop paths between target nodes u and v in a bipartite network and realize local structure extraction. The algorithm generates paths between nodes on the basis of depth-first-search, and the specific process is shown in Algorithm 1. First, target nodes u and v, bipartite network G, and path hops k are inputs. Then, all k-hop paths of node u are obtained on the basis of input path length k. If node u reaches node v through k hops, paths are preserved, and nodes in the paths are used to induce a local structure from bipartite network G, which contains the path where nodes u and v are directly connected by k hops. Lastly, an edge between u and v in the local structure is used as the label of the local structure.
Algorithm 1 Local Structure Extraction
Input: Target nodes u and v, bipartite network G, path hops k
Output: Local k-hop structure G k (u, v) between node pair (u, v)
1:  N o d e = {}
2:  P a t h = {u: u u k }; path is the set of k-hop paths of u, and u k is the end point of k-hop path
3: for i in P a t h do
4:  if u k == v then
5:     n o d e = {set of nodes in path i }
6:     N o d e = N o d e n o d e
7: Let G k (u, v) be the induced subgraph from G using vertices N o d e
8: Remove the edge between u and v in local structure G k (u, v), and use the edge as the label of the local structure
9: return G k (u, v), label

3.2. Node Feature Labeling

The local structure between target nodes is obtained by Algorithm 1. The features of nodes are usually not available, so the feature labeling of different nodes is required. The purposes of node labeling are as follows:
(1)
Using different labels to mark nodes’ different roles in the local structure. There are two types of nodes in a bipartite network. Distinguishing types of nodes can assist in fully understanding roles (user or item) in the local structure and effectively identify the target node.
(2)
Distinguishing positions of nodes in the local structure. Nodes with different relative positions to the target node have different structural importance to the link. Distinguishing the position information of nodes can effectively extract the semantic information of target nodes in local structures.
On the basis of the above criteria, different nodes were labelled in the local structure. First, different types of nodes are labeled. There are two types of nodes, U and V, in bipartite networks. In this paper, the U- and V-type nodes were labeled as 0 and 1, respectively, to effectively distinguish between different types of nodes. Then, labels of other nodes were determined according to which hop of the target node in the local structure. If the target node went through h hops to the other node, we labeled it h, which could distinguish the node’s position in the local structure. Lastly, one-hot encoding was performed on the labels of the node, and the final feature representation of the node was obtained by splicing.

3.3. Local Structure Representation

The local structure and node features were used as the input of GCN for the feature representation learning of the local structure. First, message passing is performed on the basis of the adjacency matrix of the local structure and node features to realize the feature update of each node. We used the graph convolutional neural network framework to update nodes, and the formula is as follows:
h i l + 1 = σ ( h i l + j N r ( i ) W j l h j l )
where h i l represents the feature vector of node i in layer l, N r ( i ) represents the neighbors of node i, h j l represents the feature vector of node j in layer l, W j l is a learnable parameter, and σ is the ReLU activation function. Lastly, feature h i l + 1 of the l+1 layer of node i is obtained.
Then, a representation method of the local structure was designed to realize feature representation. The features of the nodes in the local structure were averaged here as the feature representation of the final local structure, which not only contained the feature information of all nodes, but also ensured that the dimensions of local-structure and node features were the same.
H l s = M e a n ( h 1 , h 2 , . . . , h n + m )
where M e a n represents the mean function, h 1 , h 2 , …, h n + m represent the node feature vector in the local structure, n and m represent the number of two types of nodes, and H l s represents the final feature representation of the local structure.

3.4. Link Prediction

After local structure representation is obtained, the link score between target nodes is output on the basis of multilayer perceptron (MLP):
r p r e = σ ( W H l s )
where W is a parameter of the MLP that maps local structural representation H l s to a scalar rating r p r e , and σ is the activation function. The loss function is mean squared error ( M S E ), which is used to minimize the difference between predictions and ground-truth ratings:
L o s s = 1 m n i = 0 n j = 0 m ( R i j R i j * ) 2
where we used R i j and R i j * to denote the true rating and predicted rating of (i, j), respectively.

4. Experiments

Our experiments were carried out on a workstation equipped with Intel(R) Xeon(R) Gold 6132 CPU @ 2.60 GHz, NVIDIA Geforce GTX 1080Ti GPU and 192 GB RAM. The modules of DLP in the experiments were implemented with Python 3.8.3 with Pytorch 1.11.0. The DLP model adopted the SGD optimizer with parameters lr = 1 e 2 , momentum = 0.9, weight _ decay = 1 e 5 . Model training parameters: epoch = 80, batch _ size = 5. The main source codes are provided here: https://github.com/lvhehe/ISAT-Laboratory, accessed on 29 March 2022.

4.1. Datasets

The five datasets used in this paper were collected from the KONECT Project [42]: Corporate Club Memberships Dataset (CCMD), Corporate Leaderships Dataset (CLD), American Revolution Dataset (ARD), Crime Dataset (CD), and Unicode Languages Dataset (ULD). Dataset statistics are summarized in Table 1.
Table 1. Statistics of five commonly used link prediction datasets.

4.2. Method Comparison

DMF [43]: This method is based on a multilayer neural network that characterizes users and items into low-dimensional features, and synthesizes a recommendation matrix via potential low-dimensional features to achieve item recommendations.
VAE [44]: This method is based on autoencoder (AE) in which the encoder represents potential features on the basis of probability distribution and reconstructs the input on the basis of the decoder.
DAE [45]: This method is based on the backpropagation algorithm, which compresses the input into a low-dimensional feature space, then reconstructs the output, and represents nodes through the low-dimensional feature space.
IGMC [46]: This method trains a graph neural network (GNN) on the basis of 1-hop subgraphs around pairs (user, item) generated from the rating matrix, and maps these subgraphs to their corresponding ratings, enabling user rating predictions for items.

4.3. Evaluation Indicators

To evaluate the performance of the method, we randomly removed 20% of the existing links from each dataset as positive testing data. On the basis of standard learned link prediction, the same number of nonexistent links (unconnected node pairs) were sampled as negative testing data. The remaining 80% of existing links and the same number of additionally sampled nonexistent links were used to construct training data. Mean absolute error (MAE) and root mean square error (RMSE) were used for performance evaluation.
M A E = 1 l ( i , j ) l | r i j r i j * |
R M S E = ( i , j ) l ( r i j r i j * ) 2 l
where r i j is the true value between nodes i and j, r i j * is the predicted value of the DLP method, and l is the sample size of the experimental data.

4.4. Result

4.4.1. Local Structural Analysis

In this paper, the extraction of the local structure is mainly based on the k-hop path between target nodes; therefore, the influence of local structural depth on method performance is studied. Results are shown in Table 2; when local structural depth k was 1, DLP achieved the best performance of all five datasets in terms of R M S E and M A E . In addition, the performance of DLP on CCMD, CLD, and ULD was first worsened and improved when k increased from 1 to 3. This shows that, when k increases, the method captures deeper structural information among target nodes. However, on ARD and CD, the performance of the method gradually worsened as depth k increased. This shows that, when k increases, information between target nodes captured by the local structure tends to be similar, and the method cannot accurately distinguish structural differences between target nodes.
Table 2. Relationship between DLP performance and local structural depth.

4.4.2. Performance Comparison

The performance of DLP was compared with that of the aforementioned state-of-the-art methods, and results are shown in Table 3. The following conclusions were drawn:
Table 3. Performance comparison of DLP and various state-of-the-art methods.
(1)
In terms of R M S E and M A E , DLP achieved the best performance on the five datasets, indicating that DLP could effectively predict links existing between nodes.
(2)
On the five datasets, the overall reduction in M A E showed that DLP could reduce the error between predicted and actual values, and achieved effective prediction of edge weights between nodes.
(3)
The overall reduction in R M S E of the DLP method on the five datasets was 0.498, 0.619, 0.285, 0.605 and 0.601. DLP achieved the smallest R M S E on all datasets, indicating that the prediction results of the method were stable.

4.4.3. Influence of Link Sparsity

In order to verify the applicability of the method, sparsity was experimentally analyzed on five different datasets. The influence of link sparsity on DLP performance was investigated by randomly retaining linked 0.2, 0.4, 0.6, 0.8, and 1.0 in the original dataset. Results are shown in Figure 3. On the five datasets, experiments showed that the performance of DLP improved with increasing sparsity, indicating that the method relies on dense bipartite networks. At the same time, on CCMD and CLD, R M S E and M A E decrease rapidly on the basis of the increase in links. When sparsity reached 0.8, as it increased, although the performance of the method improved, R M S E and M A E gently decreased. On CD and ULD, the R M S E and M A E of DLP showed a steadily decreasing trend, which indicated that the sparsity of these two datasets more smoothly impacted DLP performance. In addition, on ARD, DLP performance showed fluctuating improvement based on the increase in links. When sparsity increased from 0.8 to 1.0, the performance of DLP suddenly improved, which indicated that DLP is more affected by sparsity on ARD, that is, the denser the edges in the bipartite network are, the better the performance of DLP is.
Figure 3. Results of sparsity analysis.

4.4.4. Model Analysis

DLP can extract the local structure of a bipartite network for analysis, analyze the local relationship between nodes, and study the importance of the local structure in link prediction. Experiments showed that DLP outperformed methods based on a global structure, proving the effectiveness of the local structure in link prediction. However, analysis of link sparseness showed that DLP can achieve excellent link prediction performance on dense bipartite networks, but relatively poor performance on sparse bipartite networks. This is understandable because the extraction of a local structure depends on the link relationship in the network. When the link relationship between nodes is sparse, the local structure is simple, and thefeatures of the local structure cannot be effectively extracted. This method is only based on artificially defined features and cannot effectively express features of nodes. Therefore, future research will collect node feature information as supplementary network information to achieve more effective link prediction.

5. Conclusions

Novel link-prediction method DLP was proposed that could mine feature information on the basis of the local structure between target nodes of bipartite networks. The following conclusions were drawn:
(1)
The effectiveness of the local structure in link prediction was confirmed, and experiments on five datasets showed that, when the local structure is smaller, the performance of link prediction is better.
(2)
Compared with existing state-of-the-art methods, DLP showed excellent performance on R M S E and M A E , indicating that this method could achieve effective prediction of link presence and link weight between nodes.
(3)
In addition, the performance of the method was improved on the basis of the increase in the sparsity of the dataset, indicating that the method is more suitable for dealing with dense bipartite networks.

Author Contributions

H.L., experimental design and writing—original draft; B.Z., supervision, funding acquisition, and writing—review and editing; S.H. and Z.X., writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R&D Program of China grant number 2017YFC0907 505.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guillaume, J.L.; Latapy, M. Bipartite graphs as models of complex networks. Phys. A 2006, 371, 795–813. [Google Scholar] [CrossRef]
  2. Aslan, S.; Kaya, B. Time-aware link prediction based on strengthened projection in bipartite networks. Inf. Sci. 2020, 506, 217–233. [Google Scholar] [CrossRef]
  3. Gao, M.; Chen, L.; Li, B.; Li, Y.; Liu, W.; Xu, Y.C. Projection-based link prediction in a bipartite network. Inf. Sci. 2017, 376, 158–171. [Google Scholar] [CrossRef]
  4. Zhang, J.W.; Gao, M.; Yu, J.L.; Yang, L.D.; Wang, Z.W.; Xiong, Q.Y. Path-based reasoning over heterogeneous networks for recommendation via bidirectional modeling. Neurocomputing 2021, 461, 438–449. [Google Scholar] [CrossRef]
  5. Yu, Z.X.; Huang, F.; Zhao, X.H.; Xiao, W.J.; Zhang, W. Predicting drug–disease associations through layer attention graph convolutional network. Brief. Bioinform. 2021, 22, bbaa243. [Google Scholar] [CrossRef]
  6. Bai, X.M.; Zhang, F.L.; Li, J.Z.; Xu, Z.; Patoli, Z.; Lee, I. Quantifying scientific collaboration impact by exploiting collaboration-citation network. Scientometrics 2021, 126, 7993–8008. [Google Scholar] [CrossRef]
  7. Barabasi, A.; Albert, R. Emergence of Scaling in Random Networks. Science 1999, 286, 509–512. [Google Scholar] [CrossRef] [Green Version]
  8. Zhou, T.; Lu, L.Y.; Zhang, Y.C. Predicting missing links via local information. EPJ B 2009, 71, 623–630. [Google Scholar] [CrossRef] [Green Version]
  9. Katz, L. A new status index derived from sociometric analysis. Psychometrika 1953, 18, 39–43. [Google Scholar] [CrossRef]
  10. Wang, W.; Lv, H.H.; Zhao, Y.; Liu, D.; Wang, Y.Q.; Zhang, Y. DLS: A Link Prediction Method Based on Network Local Structure for Predicting Drug-Protein Interactions. Front. Bioeng. Biotechnol. 2020, 8, 330. [Google Scholar] [CrossRef]
  11. Fan, S.H.; Zhu, J.X.; Han, X.T.; Shi, C.; Hu, L.M.; Ma, B.; Li, Y.L. Metapath-guided Heterogeneous Graph Neural Network for Intent Recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  12. Jin, J.R.; Qin, J.R.; Fang, Y.C.; Du, K.; Zhang, W.N.; Yu, Y.; Zhang, Z.; Smola, A.J. An Efficient Neighborhood-based Interaction Model for Recommendation on Heterogeneous Graph. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, CA, USA, 23–27 August 2020. [Google Scholar]
  13. Perozzi, B.; Al-Rfou, R.; Skiena, S. DeepWalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 24–27 August 2014. [Google Scholar]
  14. Gao, M.; Chen, L.H.; He, X.G.; Zhou, A. BiNE: Bipartite Network Embedding. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018. [Google Scholar]
  15. Grover, A.; Leskovec, J. node2vec: Scalable Feature Learning for Networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  16. Dong, Y.X.; Chawla, N.V.; Swami, A. metapath2vec: Scalable Representation Learning for Heterogeneous Networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Halifax, NS, Canada, 13–17 August 2017. [Google Scholar]
  17. Fu, T.Y.; Lee, W.C.; Lei, Z. HIN2Vec: Explore Meta-paths in Heterogeneous Information Networks for Representation Learning. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017. [Google Scholar]
  18. Yang, M.M.; Xu, S.H. A novel deep quantile matrix completion model for top-N recommendation. Knowl. Based Syst. 2021, 228, 107302. [Google Scholar] [CrossRef]
  19. Wang, M.N.; You, Z.H.; Wang, L.; Li, L.P.; Zheng, K. LDGRNMF: LncRNA-disease associations prediction based on graph regularized non-negative matrix factorization. Neurocomputing 2021, 424, 236–245. [Google Scholar] [CrossRef]
  20. Yang, Y.M.; Guan, Z.Y.; Li, J.X.; Zhao, W.; Cui, J.T.; Wang, Q. Interpretable and Efficient Heterogeneous Graph Convolutional Network. IEEE Trans. Knowl. Data Eng. 2021, 99, 1–14. [Google Scholar] [CrossRef]
  21. Jiang, X.Q.; Lu, Y.F.; Fang, Y.; Shi, C. Contrastive Pre-Training of GNNs on Heterogeneous Graphs. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management, Gold Coast, QLD, Australia, 1–5 November 2021. [Google Scholar]
  22. Fu, X.Y.; Zhang, J.N.; Meng, Z.Q.; King, I. MAGNN: Metapath Aggregated Graph Neural Network for Heterogeneous Graph Embedding. In Proceedings of the Web Conference 2020, Taipei, China, 20–24 April 2020. [Google Scholar]
  23. Wang, X.; Zhang, Y.D.; Shi, C. Hyperbolic Heterogeneous Information Network Embedding. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  24. Ahmad, I.; Akhtar, M.U.; Noor, S.; Shahnaz, A. Missing Link Prediction using Common Neighbor and Centrality based Parameterized Algorithm. Sci. Rep. 2020, 10, 364. [Google Scholar] [CrossRef] [PubMed]
  25. Verma, V.; Aggarwal, R.K. A comparative analysis of similarity measures akin to the Jaccard index in collaborative recommendations: Empirical and theoretical perspective. Soc. Netw. Anal. Min. 2020, 10, 43. [Google Scholar] [CrossRef]
  26. Wang, W.; Lv, H.H.; Zhao, Y. Predicting DNA binding protein-drug interactions based on network similarity. BMC Bioinform. 2020, 21, 322. [Google Scholar] [CrossRef]
  27. Lu, L.Y.; Jin, C.H.; Zhou, T. Similarity index based on local paths for link prediction of complex networks. Phys. Rev. E 2009, 80, 046122. [Google Scholar] [CrossRef] [Green Version]
  28. Brin, S.; Page, L. The anatomy of a large-scale hypertextual Web sea rch engine. Comput. Netw. Isdn Syst. 1998, 30, 107–117. [Google Scholar] [CrossRef]
  29. Fouss, F.; Pirotte, A.; Renders, J.M.; Saerens, M. Random-Walk Computation of Similarities between Nodes of a Graph with Application to Collaborative Recommendation. IEEE Trans. Knowl. Data Eng. 2007, 19, 355–369. [Google Scholar] [CrossRef]
  30. Chen, B.; Guo, W.; Tang, R.M.; Xin, X.; Ding, Y.; He, X.Q.; Wang, D. TGCN: Tag Graph Convolutional Network for Tag-Aware Recommendation. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, 19–23 October 2020. [Google Scholar]
  31. Qi, T.; Wu, F.Z.; Wu, C.H.; Huang, Y.F. Personalized News Recommendation with Knowledge-aware Interactive Matching. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021. [Google Scholar]
  32. Xu, Z.H.; Chen, C.; Thomas, L.; Miao, Y.S.; Meng, X.W. Tag-Aware Personalized Recommendation Using a Deep-Semantic Similarity Model with Negative Sampling. In Proceedings of the 25th ACM International Conference on Information and Knowledge Management, Indianapolis, IN, USA, 24–28 October 2016. [Google Scholar]
  33. Xu, Z.H.; Lukasiewicz, T.; Chen, C.; Miao, Y.S.; Meng, X.W. Tag-Aware Personalized Recommendation Using a Hybrid Deep Model. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  34. Sachin, M.; Rik, K.; Mohammad, R.; Hannaneh, H. DeFINE: DEep Factorized INput Word Embeddings for Neural Sequence Modeling. In Proceedings of the 8th International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  35. Ma, C.S.; Li, J.J.; Pan, P.; Li, G.H.; Du, J.B. BDMF: A Biased Deep Matrix Factorization Model for Recommendation. In Proceedings of the 2019 IEEE SmartWorld, Leicester, UK, 19–23 August 2019. [Google Scholar]
  36. Fan, J.C.; Cheng, J.Y. Matrix completion by deep matrix factorization. Neural Netw. 2018, 98, 34–41. [Google Scholar] [CrossRef]
  37. Zhang, C.X.; Song, D.L.; Huang, C.; Swami, A.; Chawla, N.V. Heterogeneous Graph Neural Network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  38. He, C.Y.; Xie, T.; Rong, Y.; Huang, W.B.; Li, Y.F.; Huang, J.Z.; Ren, X.; Shahabi, C. Bipartite Graph Neural Networks for Efficient Node Representation Learning. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  39. Wu, Y.X.; Liu, H.X.; Yang, Y.M. Graph Convolutional Matrix Completion for Bipartite Edge Prediction. In Proceedings of the 10th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management, Seville, Spain, 18–20 September 2018. [Google Scholar]
  40. Xu, S.Y.; Yang, C.; Shi, C.; Fang, Y.; Guo, Y.X.; Yang, T.C.; Zhang, L.H.; Hu, M.D. Topic-aware Heterogeneous Graph Neural Network for Link Prediction. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management, Gold Coast, QLD, Australia, 1–5 November 2021. [Google Scholar]
  41. Li, X.X.; Shang, Y.M.; Cao, Y.N.; Li, Y.G.; Tan, J.L.; Liu, Y.B. Type-Aware Anchor Link Prediction across Heterogeneous Networks Based on Graph Attention Network. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  42. Kunegis, J. Handbook of Network Analysis [KONECT—The Koblenz Network Collection]. Comput. Sci. 2014, 2, 1343–1350. [Google Scholar]
  43. Xue, H.J.; Dai, X.Y.; Zhang, J.B.; Huang, S.J.; Chen, J.J. Deep Matrix Factorization Models for Recommender Systems. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  44. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  45. Hinton, G.E.; Osindero, S.; Teh, Y.W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  46. Zhang, M.H.; Chen, Y.X. Inductive Matrix Completion Based on Graph Neural Networks. In Proceedings of the 8th International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.