Next Article in Journal
Symmetric-Key-Based Authentication among the Nodes in a Wireless Sensor and Actuator Network
Previous Article in Journal
Damage Detection in Largely Unobserved Structures under Varying Environmental Conditions: An AutoRegressive Spectrum and Multi-Level Machine Learning Methodology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Heterogeneous User Generated Contents-Driven Relation Assessment via Graph Representation Learning

1
School of Information Science & Engineering, East China University of Science and Technology, Shanghai 200237, China
2
School of Computer Science and Electronic Engineering, University of Essex, Colchester CO4 3SQ, UK
3
Department of Electronic and Electrical Engineering, University of Sheffield, Sheffield S10 3JD, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1402; https://doi.org/10.3390/s22041402
Submission received: 6 January 2022 / Revised: 8 February 2022 / Accepted: 9 February 2022 / Published: 11 February 2022
(This article belongs to the Topic Advances in Perceptual Quality Assessment of User Generated Contents)
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Cross-domain decision-making systems are suffering a huge challenge with the rapidly emerging uneven quality of user-generated data, which poses a heavy responsibility to online platforms. Current content analysis methods primarily concentrate on non-textual contents, such as images and videos themselves, while ignoring the interrelationship between each user post’s contents. In this paper, we propose a novel framework named community-aware dynamic heterogeneous graph embedding (CDHNE) for relationship assessment, capable of mining heterogeneous information, latent community structure and dynamic characteristics from user-generated contents (UGC), which aims to solve complex non-euclidean structured problems. Specifically, we introduce the Markov-chain-based metapath to extract heterogeneous contents and semantics in UGC. A edge-centric attention mechanism is elaborated for localized feature aggregation. Thereafter, we obtain the node representations from micro perspective and apply it to the discovery of global structure by a clustering technique. In order to uncover the temporal evolutionary patterns, we devise an encoder–decoder structure, containing multiple recurrent memory units, which helps to capture the dynamics for relation assessment efficiently and effectively. Extensive experiments on four real-world datasets are conducted in this work, which demonstrate that CDHNE outperforms other baselines due to the comprehensive node representation, while also exhibiting the superiority of CDHNE in relation assessment. The proposed model is presented as a method of breaking down the barriers between traditional UGC analysis and their abstract network analysis.

1. Introduction

Nowadays, user-generated contents (UGCs) are riddled in various large-scale online platforms such as e-commerce platforms, discussion forums, live streaming platforms and social networks [1,2,3,4]. The research on UGC can be roughly divided into intrinsic quality improvement and their interrelation analysis, which are indispensable parts of the online decision-making platform. Normally, researchers focus on the possible contents distortion or quality degradation, while neglecting the importance of relation assessment among various UGCs. Tapping into relationships of those high-quality UGCs can attract general attentions and produce great social benefits. In reality, many tangled contents can be described and analyzed by the relevant characteristics of a complex network.
The research on entities’ relationships via abstract network structures has always been a hotspot in many fields [5,6,7,8]. Accurate relation assessment and prediction are helpful to analyze the UGC network evolution patterns and assist network maintenance, which is of great significance to enhance the survivability and to improve the reliability in both static and dynamic networks. More precisely, relation prediction in network refers to forecasting the underlying existence of a link between two nodes based on the network structural information and the intrinsic information of nodes [9].
Among the existing relation prediction methods, the heuristic methods measure the connectivity of two nodes and the graph statistical features from the perspective of similarity, such as the degree of node, node centrality and the clustering coefficient [10,11], while the graph representation learning methods concentrate on encoding the intricate network structures into low-dimensional vector space so as to capture the multi-scale and high-level node features from the underlying topology [12]. These approaches [13,14,15], on the other hand, mainly represent the network under the assumption that the network is homogeneous, while focusing on the static attributes of the network. Actually, the abstract UGC network in the real world is normally composed of multiple types of nodes or edges, while the relationship between nodes is complex and evolves over time, which presents heterogeneous and dynamic characteristics. This renders the straightforward application of most existing relation prediction methods infeasible [16].
Dynamic heterogeneous networks are composed of different types of entities and relations, which usually evolve over time. Taking Figure 1 as an example, above the timeline is a commodity supply–demand network containing three types of nodes (i.e., customer, item and merchant) and edges (i.e., customer–customer, customer–item and merchant–item). The demands for T-shirts and shorts are greater in summer, while, as Christmas approaches, customers show demands for other products, such as Christmas trees and stockings, which demonstrates how the structure of networks varies over time with dynamic characteristics.
Up to now, limited attempts have been made to investigate the embedding of dynamic heterogeneous networks. Kong et al. [17] used the graph convolutional network (GCN) to extract the spatial structural features from heterogeneous information networks (HINs) and employed the long short-term memory (LSTM) network to forecast the existence of links. However, this shallow GCN merely captured neighbors with low proximity while ignoring the heterogeneous characteristics, and cannot be stacked for multiple layers due to the over-smoothing problem. DHNE [18] devises a historical–current networks structure, which takes all neighbors in the time step into consideration in order to learn the latent node embedding from the dynamic condition; moreover, a metapath-based random walk is conducted to capture the heterogeneous semantic information. DyHATR [19] integrates a hierarchical attention module and recurrent neural network based model to learn heterogeneity as well as temporal evolution. More recently, Ji et al. [20] introduce a Hawkes-process-based method to model the formation process of heterogeneous events adequately and use the importance sampling strategy to capture representative events for influence propagation. Xie et al. [21] propose the DyHINE method, comprising a temporal dynamic embedding module and an online updating module, which can deploy real-time updated embedding when the network evolves.
Although the above approaches work well in many types of applications, they still have some drawbacks. In general, traditional matrix factorization methods usually perform relatively poorly due to the high computational cost of decomposing the massive matrix, while most graph embedding techniques mainly focus on the homogeneous network. To address this challenge, our dynamic heterogeneous graph embedding method tends to learn a map function that converts complicated input networks into low-dimensional space for better representation while capturing the evolutionary properties of networks. The Markov-chain-optimized metapath is able to preserve the heterogeneous structure and semantics while improving computational efficiency. Despite multi-scale features on networks have been thoroughly explored, the current network embedding methods solely consider the low- or high-order proximity characteristics of nodes in a limited perspective, while ignoring the global features represented by community structures. To solve the structural information loss in feature extraction, we not only integrate other side information, but also capture global structure semantics via a clustering technique. Moreover, there is a lack of research about capturing temporal evolution characteristics, which is also of great importance in dynamic relation prediction. To address this issue, we propose an encoder–decoder structure to learn the temporal dependencies after obtaining the nodes representation, which contains comprehensive heterogeneous topology information and can be delivered for distilling the implicit correlations between each time step. Our main contributions can be summarized as follows:
  • User generated contents-driven method: We proposed a graph representation learning-based method, named the community-aware dynamic heterogeneous graph embedding method (CDHNE), for predicting and assessing the relationships between different user-generated contents. We consider the generalized user-generated contents as abstract nodes, which form a dynamic heterogeneous network, so as to introduce the graph embedding methods. The objective of this work is to explore the semantics of human activities and perform a comprehensive relation assessment of these activities.
  • Multi-level representation learning: We facilitated the metapath-based random walk, utilizing Markov chain approximation, for localized heterogeneous contents learning. An edge-centric attention mechanism is introduced for subgraph-level feature aggregation. The clustering technique, depending on node embedding, offers effective global structure semantic extraction without prior information.
  • Temporal dynamics extraction: We devised an encoder–decoder structure with two variants, i.e., CDHNE-GRU and CDHNE-LSTM, to learn the temporal evolutionary patterns in dynamic heterogeneous networks. Concretely, we split the dynamic heterogeneous network into several snapshots and leveraged the recurrent memory unit to capture long-term dependencies over time steps. In each hidden unit, the generated parameters are delivered to trigger the next gate. Finally, we can obtain the output through a fully connected decoder.
  • Experimental results: We constructed datasets containing a series of human activities, which included academic collaboration, commercial promotions and social interactions, and conducted extensive experiments to demonstrate the effectiveness of CDHNE under the user-generated contents scenarios. Specifically, we evaluated our proposed model on relation prediction problems and conduct community detection tasks to validate the effectiveness of CDHNE in global structure semantic extraction. The experimental results on four real-world datasets show that our proposed model outperforms the other competitive baselines in terms of AUROC and AUPRC.
The rest of this paper is organized as follows. Section 2 provides an overview of existing relation assessment methods. In Section 3, we introduce some necessary definitions that will be used in this paper and formulate the relation prediction task for dynamic heterogeneous networks. Section 4 describes our proposed model in detail. Extensive experimental results and analyses are presented in Section 5. Finally, we conclude this paper in Section 6. The main notations used in this paper are listed in Table 1.

2. Literature Review

In this section, we briefly review the research development trends concerning relation prediction and divide the approaches into two categories, including traditional heuristic methods and graph representation learning methods.
The traditional heuristic methods for relation prediction, such as common neighbors [22], Jaccard coefficient [23], Adamic Adar (AA) [24], resource allocation (RA) [25] and preferential attachment (PA) [26], mainly exist in static homogeneous networks where nodes and edges are of the same type. These methods attempt to quantify neighborhood overlap between nodes, while minimizing deviations caused by degrees. The basic idea of heuristic-based methods is typically based on the measures of nodes similarity. In general terms, the more similar the two nodes are, the more likely to have a link between them. For instance, both RA index and AA index attach more importance to common neighbor nodes in low degrees, which intuitively provide more information than common neighbor nodes in high degrees. Albeit simple, these strategies can still achieve competitive performance compared with other methods in a specific scenario.
The idea of graph representation learning is to extract the latent network features from the complicated topological structure and to encode features, such as node embedding vectors in a low-dimensional space. For a relation prediction task, the learned hidden features should preserve network properties in local-wise and global-wise, so that the unknown links can be accurately predicted. However, most graph representation learning methods focus on the topological properties of nodes (e.g., in-degree or out-degree, random walk distance, first-order proximity and so on). Early solutions [13,14] learned the hidden representations of vertices by deploying random walk in the static graph, which samples the first- and second-order similarity of nodes. Subsequently, node2vec [15] made some innovations in generating the random walk strategy, which combines depth-first sampling (DFS) and the breadth-first sampling (BFS), while efficiently exploring spatial contextual information in graph data. Ribeiro et al. [27] proposed struc2vec, which measures the structural similarity from a global perspective without requiring two nodes to be nearby. GraphSAGE [28] sampled a certain number of neighborhoods for further aggregation. Graph attention network (GAT) [29] utilized the self-attention mechanism to learn the weighting function for neighborhood matching.
Considering that networks based on UGC in real life normally vary over time, which exhibits dynamic evolution characteristics, several approaches were proposed to capture the temporal features. A straightforward way of modeling dynamic network is to split it into snapshot sequences along the timeline, which discretizes the continuous changes. Goyal et al. [30] proposed the dyngraph2vec, which learns the underlying network dynamics of evolution by constructing a deep recurrent architecture. Similarly, [31,32] adopted the long short-term memory network (LSTM) to learn the temporal dependencies of all dynamic network snapshots. Such recurrent methods capture temporal features through selective memory and forgetting mechanism which enables the model to handle long sequences. However, the shortage of these methods is obvious, which depend heavily on the time span of aggregated snapshot. Different from the above methods, researchers in DySAT [33] extract the structural information and dynamics node embedding simultaneously, via a self-attention mechanism. EvolveGCN [34] adapted the graph convolutional network (GCN) throughout the time dimension and utilized an RNN architecture to update the GCN parameters.
Furthermore, with the increasing diversity of the real-world, user-generated network, the nodes and edges in networks have gradually developed from a single type to a mixture of multiple types, which shows multi-source heterogeneous characteristics. The heterogeneous graph neural network (HetGNN) [35] conducted heterogeneous graph embedding by gathering the nodes with same type via a correlated sampling. Dong et al. [36] developed two metapath-based representation learning methods, namely metapath2vec, and its variant metapath2vec++, which learn the topological and semantic correlations in heterogeneous networks. Cen et al. [37] divided a node embedding procedure into two portions, namely the base embedding and the edge embedding, which share parameter information across different edge types, allowing the extraction of heterogeneous information.
Currently, limited attempts have been made to investigate the embedding of dynamic heterogeneous networks. HA-LSTM [17] firstly employ GCN to extract structural features from heterogeneous information network, then leverage a broad learning- and attention-based structure to capture the dynamic changes over timeline. DHNE [18] constructed a historical–current network structure with consecutive snapshots to capture temporal dependencies in the dynamic heterogeneous network, then a metapath-based random walk was conducted to extract intricate semantic information. The DyHATR [19] model built a hierarchical attention model to better learn the heterogeneity of static snapshots and captured the temporal evolution patterns via an attentive RNN structure. Nevertheless, none of the preceding techniques have established a multi-view embedding strategy, instead of focusing on the localized features while neglecting the global characteristics.

3. Preliminaries

In this section, we formulate the relation prediction problem for dynamic heterogeneous network. Firstly, we introduce some definitions and necessary notations that will be used in this paper, as follows:
Definition 1
(Dynamic heterogeneous network). A dynamic heterogeneous network (DyHN) can be represented as a set of observed graph, G = G 1 , G 2 , , G T , which contains T snapshots. G t = V t , E t , F , φ , with a adjacency matrix A t , denotes the snapshot at time t, where V t is the set of nodes and E t is the set of edges. F : V t T denotes the mapping function for node types, while φ : E t R is the edge type mapping function. For a dynamic heterogeneous network, the following conditions must be satisfied that T + R > 2 .
Definition 2
(Dynamic heterogeneous network representation learning). Given a dynamic heterogeneous network, G = G 1 , G 2 , , G T , the objective of graph representation learning is to encode the node as a low-dimensional vector that contains graph structure and local neighborhood information, that is, mapping snapshot G t into a hidden space, Z t R V t × d , where d is the final embedding dimension.
Definition 3
(Relation prediction). Given a series of observed network snapshots, G = { G 1 , G 2 , , G T } , relation prediction tasks in dynamic heterogeneous network can be viewed as a prediction on an evolving network with multiple types of nodes and edges. Besides, each snapshot is considered as a static heterogeneous network in this work, the link connection at time stamp, t + 1 , is determined by the spatial features and temporal evolutionary trajectory extracted from historical snapshots.

4. Community-Aware Dynamic Heterogeneous Network Embedding Method

In this section, we propose a community-aware dynamic heterogeneous network embedding method. In Section 4.1, we derive heterogeneous properties guided by a Markov-chain-optimized metapath. For the sake of computation efficiency, we perform negative sampling for learning node embedding in the skip-gram model. Then, we aggregate subgraph-level features using the edge-centric attention mechanism in Section 4.2, which preserves the low-order proximity of nodes. In Section 4.3, we extract the structure semantic of network from a macro perspective using clustering technique. To capture the temporal evolutionary patterns, we develop an encoder–decoder structure based on recurrent neural network in Section 4.4.

4.1. Heterogeneous Contents Encoding

Different from the homogeneous network, a heterogeneous network is made up of multiple types of nodes and edges. To extract the heterogeneity, inspired by metapath2vec [36], we perform random walk under the guidance of metapath on each heterogeneous network snapshot in order to capture the spatial topology and semantic information. Generally, a metapath strategy, P , can be described as a path that is predefined as: P : T 1 R 1 T 2 R 2 T t R t R n T n + 1 ; therein, R = R 1 R 2 R n represents the multi-level relationships between nodes T 1 and T n + 1 . For illustration, consider the commodity supply–demand network shown in Figure 2, the metapath “CIC” denotes the common interest in items between two customers, and “CIMIC” represents that two customers purchase similar items from the same merchant.
The key point of conducting random walk on heterogeneous networks is to determine the transition probabilities for each step. For the sake of efficiency and effectiveness, we associate the metapath-based random walk with higher-order Markov chains [38] to facilitate the learning of nodes distribution.
Theorem 1.
Given an arbitrary metapath, P : T 1 R 1 T 2 R 2 T t R t T t + 1 , there exists a k-order Markov chain if—and only if— P can be decomposed into a collection of distinct k-length metapaths, { T l R l T l + 1 R l + 1 R l + k 1 T l + k } , while satisfying the condition that current state, T l + k , is only determined by { T l , T l + 1 , , T l + k } . We can then leverage the transition probabilities obtained by k-length Markov chain to guide P -concerned random walks.
Note that the metapath decomposition mentioned above can be interpreted as a process of factor extraction. For instance, we can decompose the metapath “CIMIC” into a set of metapath factors: “CIM”, “IMI”, “MIC” and “ICI”. It is evident that the present state can only be determined by the last two states. Therefore, motivated by this, we utilize a second-order Markov chain to represent the metapaths of this “CIMIC” type and introduce the node transition probability matrix, as follows:
M k | j , i P T l + 1 T l + 2 , F : v i , j , k T l , l + 1 , l + 2 0 , o t h e r w i s e
where M k | j , i indicates the transition probability to node k, given the last hop node, j, and the second last hop node, i. F denotes the type mapping function for nodes i, j and k. P T l + 1 T l + 2 represents the transition probabilities proposed in path ranking algorithm (PRA) [39], which can be calculated as follows:
P T l + 1 T l + 2 = D T l + 1 T l + 2 1 A T l + 1 T l + 2
where A T l + 1 T l + 2 is the adjacency matrix between nodes in type T l + 1 and nodes in type T l + 2 , and  D T l + 1 T l + 2 is the degree matrix. Thus, given a dynamic heterogeneous network snapshot, G t = V t , E t , F , φ , and a meta path scheme, P : T 1 R 1 T 2 R 2 T t R t T t + 1 satisfies the second-order Markov chain. The transition probability at step t is defined as follows:
P r o b X t = k | X t 1 = j , X t 2 = i , P M k | j , i
To ease exposition, we abbreviate term P r o b · as p · . Subsequently, we utilize the skip-gram to learn node representations on the dynamic heterogeneous networks snapshot by maximizing the probability of the existence of heterogeneous neighbor nodes. After given a dynamic heterogeneous network snapshot G t with | T | types of nodes and the neighborhood sampling corpus, V P , guided by metapath P , we define the objective function as follows:
a r g max θ i V P c T m N c ( i ) log p m | i ; θ
where N c ( i ) denotes the neighborhoods of node i with type c, and θ is the set of parameters. Generally, the transition probability p m | i ; θ is normalized by the softmax function [18,36,40].
p m | i ; θ = exp X c m t · X i t m V P exp X m t · X i t
where X c m t is the context vector of m and X i t indicates the embedding of node i.
To relieve the burden of calculation, we deploy a negative sampling, the same as [40], which yields great performance in practice. We firstly define a negative sample size, N, then the final objective function is shown as follows:
log σ X c m t · X i t + n = 1 N E u n P u log σ X u n t · X i t
where σ · is the sigmoid function that limits the value to a range of 0 , 1 , P u is the predefined sampling distribution and u n means node u has been negative sampled for n times. The stochastic mini-batch gradient descent (SMGD) algorithm is utilized to optimize the objective function, which reduces the computational overhead and randomness while maintaining a fast convergence rate.

4.2. Subgraph-Level Feature Aggregation

After conducting metapath-based random walk on dynamic heterogeneous snapshots, we obtain the node-centric embedding, while the heterogeneous network also contains different types of edges. In order to incorporate edge-wise heterogeneous information to each node representation, we introduce edge-centric embedding method, including embedding for the same types of edges and different types of edges, with attention mechanism. In general, we perform a subgraph-level aggregation operation to obtain the representation of each snapshot in a micro perspective.
For the edge in same type, the edge-centric attention model aims to learn the importance coefficient of each node’s neighborhoods in same type and aggregate these features to generate hidden representations. Taking Figure 3 as an example, there are five neighbors around the node v 1 , while node v 2 , v 3 and v 6 are the same type as node v 1 . Therefore, this step involves handling with the set of nodes v 1 , v 2 , v 3 , v 6 .
Suppose the input features consist of node embedding of each snapshot, G t G ; we deploy edge-centric attention model for each node pair with the same edge type. The importance coefficient of node j N i r to node i with edge type r and t-th snapshot can be formulated as follows:
w ( i : j ) r t = exp σ α r T W r X i t W r X j t u N i r t exp σ α r T W r X i t W r X u t
where N i r t denotes the neighbors of node i with type r at t-th snapshot, and W r is the learnable parametric matrix. σ · is the LeakyReLU activation function. X i t , X j t and X u t are the embedding for node i, j, and u at t-th snapshot, respectively, and ⊕ represents the concatenation operation. Then, we aggregate the features of neighbors with same type edge by employing nonlinear transformation, as follows:
X ˜ i r t = σ j N i r t w ( i : j ) r t · W r X j t
where X ˜ i r t is the aggregated embedding of node i for the same type of edges at t-th snapshot. Here, σ · is the tanh function.
Subsequently, we further explore the impact of edge-type-based neighbors for certain nodes. Firstly, we carry out the feature transformation that map aggregated node embedding into high-level space σ W e · X ˜ i r t + b e through a nonlinear function, σ · , such as ReLU, tanh and sigmoid. W e and b e are the learnable parametric matrix and bias vector, respectively. Then, we measure the influence of different types of edges to a specific node by implementing an edge-centric attention model. The weight coefficient of node i with edge type r at t-th snapshot β i r t is normalized by the softmax function:
β i r t = exp q T · σ W e · X ˜ i r t + b e r R exp q T · σ W e · X ˜ i r t + b e
where q is the attentive parameterized vector and tanh function is utilized as activation function. With the normalized attention weights, we can finally obtain the embedding for node i at t-th snapshot from different edge-levels, which can be expressed as follows:
X ˜ i t = r = 1 R β i r t · X ˜ i r t
After obtaining the representation of each node in the snapshot, the overall node embedding at t-th snapshot can be described as follows: Z m i c r o t = X ˜ 1 t , X ˜ 2 t , , X ˜ i t , , X ˜ V t t , where X ˜ i t R F , F V t is the number of feature embedding dimensions.

4.3. Community-Level Semantic Learning

The local structure of nodes is crucial for dynamic heterogeneous network representation, while the global structure also plays an important role in portraying the network. Community structure exists in many real-world networks, whether they are homogeneous or heterogeneous, which reflect the global structure of networks in a macroscopic perspective. Motivated by this intuition, we present a community-aware graph embedding method with network clustering technique for extracting the structure semantic in community-level, which encode node information into low-dimensional representations.
However, in dynamic heterogeneous networks, community structures are generally regarded as a priori information, which can not be known in advance. Following the work DEC [41], we first initialize a set of K cluster centroids { c j } j = 1 K by a random selection procedure. The clustering objective function is defined as a Kullback–Leibler divergence (KL) loss between the soft probability distribution, Q, and the auxiliary probability distribution, P, which can be expressed as follows:
L c = K L ( P Q ) = i j p i j log p i j q i j
where q i j can be interpreted as the probability that measures the similarity between node i and cluster center j by Student’s t-distribution, as follows:
q i j = ( 1 + X ˜ i t c j 2 / n ) n + 1 2 j K 1 + X ˜ i t c j 2 / n n + 1 2
where n denotes the degree of freedom of the Student’s t-distribution, and X ˜ i t is the micro node embedding of node i generated in Section 4.2. Empirically, we set n = 1 in our experiments. p i j is the auxiliary probability distribution calculated by the following:
p i j = q i j 2 / i q i j j K q i j 2 / i q i j
During backpropagation, the stochastic gradient descent algorithm is utilized to iteratively optimize the cluster loss function so as to bring the node closer to its cluster centroid. The partial derivative of L c with respect to variables X ˜ i t and c j are shown as follows:
L c X ˜ i t = n + 1 n j = 1 K ( 1 + X ˜ i t c j 2 n ) 1 · p i j q i j X ˜ i t c j
L c c j = n + 1 n i = 1 V t ( 1 + X ˜ i t c j 2 n ) 1 · p i j q i j c j X ˜ i t
After sending the gradients down to update the parameters, we can obtain the node embedding, Z m a c r o , and cluster centroids, c j , from a macro perspective, where Z m a c r o R V t × d is constructed by updated X ˜ i t , and d is the dimension of the embedded features.
Finally, we construct the ultimate node embedding using a combination of the micro node embedding Z m i c r o and macro node embedding Z m a c r o , which is to model the overall structure of a dynamic heterogeneous network. Specifically, the overall node embedding Z t at t-th snapshot is defined as follows:
Z t = λ Z m i c r o t + ( 1 λ ) Z m a c r o t
where λ denote the trade-off parameter that balances the weight of the micro and macro node embedding at ultimate node representations. The overall framework of CDHNE is presented in Figure 4.

4.4. Temporal Evolutionary RNN Model

One of the major characteristics of dynamic heterogeneous networks is the time-varying characteristic. There are many scenarios of changing network structures, such as establishing citation relationships between authors, product recommendations between users, those newly added or removed UGCs, and so forth. For this reason, we propose two variations of our method, as presented in Figure 4: (i) CDHNE-GRU, (ii) CDHNE-LSTM. Two high-profile, RNN-based models—gated recurrent unit (GRU) and long short-term memory (LSTM)—are leveraged in our proposed method with encoder–decoder structure to enable the capability of capturing a network’s evolutionary patterns and to further extract comprehensive information along continuous snapshots.
Gated recurrent unit is a modification to the RNN hidden layer that makes it much better for capturing long-range connections and helps a lot with the vanishing gradient problems. Through the above structural semantic learning and feature aggregation operation, we can finally obtain the node embedding for all snapshots, Z 1 , Z 2 , , Z T , where Z t R V t × d and V t denote the number of nodes at t-th snapshot, and d is the size of the embedded dimension. Then, we take these embeddings as the input of GRU. These following are the equations that govern the computation of a GRU unit:
Γ u t = σ W u c t 1 Z t + b u Γ r t = σ W r c t 1 Z t + b r c ˜ t = tanh ( W c Γ r c t 1 Z t + b c ) c t = Γ u c ˜ t + 1 Γ u c t 1 a t = c t
where Γ u , Γ r R F denote the update and reset gates, respectively. c t , c ˜ t are the memory unit and the candidate value, respectively. W u , W r , W c R F × 2 d and b u , b r , b c R F are trainable parameters. F is the dimension of output embedding. σ ( · ) is the activation function. ⊕ indicates the concatenation operation, and * is the the operation of Hadamard product.
Compared with the GRU, the LSTM model achieves better representation and introduces the forget gate, Γ f , for controlling the information of previous moments more independently. The formulations of the single LSTM network at t-th snapshot are shown as follows:
Γ u t = σ W u a t 1 Z t + b u Γ f t = σ W f a t 1 Z t + b f Γ o t = σ W o a t 1 Z t + b o c ˜ t = t a n h ( W c a t 1 Z t + b c ) c t = Γ u c ˜ t + Γ f c t 1 a t = Γ o t tanh c t
where Γ u t , Γ f t , Γ o t R F denote the update, forget and output gate, respectively. State vector a t is the element-wise product of the output gate, Γ u t , and the memory unit, c t . W u , W f , W o , W c R F × 2 d and b u , b f , b o , b c R F are trainable parameters. The other notations represent the same meaning as the GRU model.
As depicted in Figure 4, in order to achieve a more effective temporal evolution learning, we stack several LSTM networks to construct a multi-layer structure. In each layer, there can be T recurrent memory units arranged like a chain, which is to deliver parameters to next timestamp. Firstly, we encode the input node embedding into hidden representation via an RNN-based model, then apply several fully connected layers as the decoder for the final relation prediction between two nodes.

4.5. Complexity Analysis of CDHNE

In this work, dynamic heterogeneous networks are represented by T static snapshots. Therefore, for each snapshot, we mainly consider the time complexity on two stages. Firstly, for the subgraph-level information extraction, which consists of heterogeneous contents learning and feature aggregation. During the metapath-based random walk, given a metapath set P with V t nodes, walk length, l, and walks per node, w, the time complexity is O ( d w V t l 2 ) , in which d is the embedding dimension. In the process of calculating node embeddings, the theoretical time complexity of skip-gram is extremely high; we employ negative sampling for reducing complexity as much as possible. During the feature aggregation stage, the time complexity is O ( d l R 2 E t ) , where R is the number of edge types, and E t is the number of edges at t-th snapshot.
For the community-level semantic learning, in each snapshot, the time complexity of calculating node embedding in macro perspective can be divided into two parts. To calculate the probability distribution q i j and p i j , our method takes O d K V t and O K V t , respectively, where K is the initial number of cluster centroids. The other is O d K V t , in calculating the gradient of parameters X ˜ i t and c j simultaneously. Since d V t , the time complexity of this part is almost linear with the number of nodes.
Besides, for the extraction of temporal evolution patterns, we utilize the RNN-based model, of which the time complexity is normally related to hardware execution. Thus, we introduce the model complexity in this subsection. The number of parameter for each cell in the LSTM model is 4 d n V t + n + n 2 , where n denotes the size of output term. Finally, the pseudocode for CDHNE is shown in Algorithm 1.
Algorithm 1 The CDHNE algorithm
Input: A dynamic heterogeneous network G with T snapshots, the predefined embedding
dimension d.
Output: The probability of the linkage between two nodes
1:
for each snapshot G t G 1 , G 2 , , G T  do
2:
    X i t A Markov-chain-optimized metapath random walk
3:
   for each edge in specific type r R  do
4:
     Calculate the importance weight w ( i : j ) r t of node pair i , j by Equation (7)
5:
     Aggregate the features of neighbors in same type and obtain the type-specific node embedding X ˜ i r t
6:
   end for
7:
   Calculate the weight coefficient β i r t of node i with edge type r at t-th snapshot by Equation (9)
8:
   Obtain the node representation at t-th snapshot Z m i c r o t in a micro perspective through feature aggregation with different type of edges
9:
   Obtain the community-level semantic information Z m a c r o by a clustering process
10:
end for
11:
Generate comprehensive node embedding Z t for each snapshot by Equation (16)
12:
Get the probability of the link existence through GRU/LSTM encoder–decoder structure

5. Experiments and Result Analysis

In this section, we evaluate the performance of the proposed model on four real-world datasets. We firstly introduce the datasets and the configuration of the experimental environment. Then, we introduce the baselines in detail. We also conduct elaborate experiments for relation prediction tasks and demonstrate the effectiveness of main components. Moreover, we analyze the influence of sampling granularity and investigate the sensitivity of hyper-parameters. The detailed statistics of datasets are summarized in Table 2. (https://github.com/ZijianChen1998/CDHNE.git) The implementation of our CDHNE model are publicly accessed on 8 February 2022.

5.1. Experiment Setup and Dataset Description

We abstract various user-generated contents as heterogeneous nodes and select four dynamic heterogeneous networks covering academic, commerce and social interaction fields as our experimental datasets. The detailed description are presented as follows:
  • DBLP [42] (https://dblp.uni-trier.de) (15 December 2021): The DBLP dataset comprises academic literature information in the field of computer science. In this experiment, we adopt a subset of the DBLP dataset collected by [18], and compress the information into 19 snapshots, which contains 3 types of nodes, i.e., authors, papers and venues.
  • AMiner [43] (https://www.aminer.cn/data) (15 December 2021): AMiner is a big data mining and service system platform which helps researchers to mine rich academic information. In this paper, we use the evolved dynamic heterogeneous network released by [18], which establishes the relationships among authors, articles and conferences.
  • EComm (https://tianchi.aliyun.com/competition/entrance/231721) (15 December 2021): The EComm dataset was launched in CIKM-2019 E-Commerce AI Challenge, which records the consumers’ shopping behavior over an 11-day period from 10 June 2019, to 20 June 2019. It consists of three files (i.e., user behavior files, user information sheets and product information tables).
  • Math-Overflow (https://snap.stanford.edu/data/sx-mathoverflow.html) (15 December 2021): This dataset [44] contains interactions of users over time, which are sampled from the stack exchange website Math-Overflow. There are three different types of directed edges (i.e., answer–question, comment–question, comment–answer) over a time span of up to 2350 days.
Evaluation Metrics: We choose two commonly used evaluation indicators, AUROC and AUPRC [45], to compare the relation prediction performance of different methods in dynamic heterogeneous networks. Among them, AUROC is the abbreviation of the area under the receiver operating characteristic curve (ROC), while AUPRC is the abbreviation of the area under the precision recall curve (PRC).
Noting that, AUROC can be interpreted as the probability of a randomly chosen missing link being ranked higher than a randomly chosen nonexistent link. Then, the AUROC can be formulated as follows:
A U R O C = n + 0.5 n N
where N denotes the number of independent processes, and there are n times that the score of missing link is greater than the score of nonexistent link and n times when the opposite scenario occurs. Similarly, the PRC is plotted by precision–recall pairs. Precision measures the capacity of the classifier to label the missing links correctly, while the recall reflects the completeness of the model to discover all missing links. Intuitively, the closer the value of AUROC and AUPRC are to one, the more discriminative the model is.
In the task of community detection, we take modularity to measure the capacity of discovering community structure in dynamic heterogeneous networks, which does not need prior information about the ground truth. Let C u denotes the affiliation of node u. The calculation formula of modularity is defined as follows:
Q = 1 2 m u v [ A u v k u k v 2 m ] δ C u , C v
where Q denotes the modularity, the closer it is to 1, the better the effect of community division. m is the number of overall edges in network. A u v indicates the linkage between node u and v. δ C u , C v = 1 only if node u and v belong to the same community, otherwise δ C u , C v = 0 .
During the phase of heterogeneous information processing, we employ metapath guided random walk for node representations, which based on the transition probability of k-order Markov chain. In dataset DBLP, we mainly consider metapaths involving APA (i.e., the coauthors semantic), and APCPA (i.e., the sharing publication on conferences from different authors). In dataset Aminer, we also pay attention to metapaths including APA and APCPA, with similar meanings. For the EComm dataset, we consider the relations between customers and items including browse, buy, add-to-cart and add as favorite. For the Math Overflow dataset, we are interested in the interactions between users, which manifests as answering, questioning and commenting.
To make a fair comparison, all experiments are conducted using the Windows(64-bit) PC with Intel Core i5-9300HF CPU 2.4 GHz, 16 GB RAM and NVIDIA GeForce GTX 1660Ti 6G GPU. The programming environment of CDHNE-GRU/LSTM are Python 3.7 and Tensorflow 1.15. The detailed experiment configurations are represented in Table 3.

5.2. Baseline Description

We compare our proposed model against eleven methods in four categories, including static homogeneous network embedding methods, static heterogeneous network embedding methods, dynamic homogeneous network embedding methods, and dynamic heterogeneous network embedding methods. The detailed descriptions are as follows.

5.2.1. Static Homogeneous Network Embedding

  • DeepWalk [13]: DeepWalk is a homogeneous network embedding method, which conducts random walk to learn the node representation in static network.
  • node2vec [15]: The idea of node2vec is similar to DeepWalk, while considering the DFS and BFS neighborhoods simultaneously, thus improving the effect of network embedding.
  • GAT [29]: Graph attention network leverages the attention mechanism to assign different weights to each neighbor, which adaptively realized the matching weights of different neighbors.
  • GraphSAGE [28]: GraphSAGE samples the neighbor nodes of each vertex on the static homogeneous graph, then aggregates the feature information from neighbors.

5.2.2. Static Heterogeneous Network Embedding

  • metapath2vec [36]: metapath2vec performs a path determined random walk in heterogeneous network and leverages the skip-gram model to generate the node embedding.
  • metapath2vec++ [36]: metapath2vec++ improves the metapath2vec by further extracting the structural and semantic correlations in heterogeneous networks.
  • HetGNN [35]: HetGNN is a graph neural network-based model for static heterogeneous network, which sample-correlated the neighbors for heterogeneous nodes with a restart random walk and obtained deep feature interactions through a encoding module.

5.2.3. Dynamic Homogeneous Network Embedding

  • dyngraph2vec-RNN [30]: A deep architecture with sparsely connected long short-term memory networks, which is able to learn the evolution patterns in homogeneous graph structures.
  • dyngraph2vec-AERNN [30]: An improved version of dyngraph2vec-RNN, which leverage multiple fully connected layer to learn the initially hidden representations.
  • DySAT [33]: DySAT obtains the node representations by jointly conducting self-attention operation to extract the structural information and temporal dynamics.

5.2.4. Dynamic Heterogeneous Network Embedding

  • DHNE [18]: A network representation learning method for dynamic heterogeneous networks, which construct historical–current networks snapshots in timelines and capture heterogeneous semantic information under the guidance of metapaths.
  • DyHATR [19]: DyHATR utilize a hierarchical attention mechanism to learn heterogeneous information and capture network dynamic evolutional patterns via a temporal-attention-based recurrent neural network.

5.3. Relation Prediction

The objective of a relation prediction task for dynamic heterogeneous networks is to learn various node representations from previous t-th snapshots, then forecast the relation existence at t + 1 -th snapshots. Concretely, we take previous t snapshots, G 1 , G 2 , , G t , as inputs and feed them into the model, then we can obtain the node embedding at t + 1 -th snapshots, which contain rich information about the network, and thus are used to predict relations.
Abundant experiments are conducted on four datasets. For datasets DBLP and AMiner, we hide a certain percentage of edges to generate the training set, respectively. Meanwhile, due to the fact demonstrated in the previous work that significant metapaths help solving downstream tasks [46,47], we set different weights for APA and APCPA in DBLP as 0.8 , 0.2 . Similarly, we allocate weights in AMiner as 0.6 , 0.4 . As for dataset EComm, there is only one type of metapath CI with full weight. Moreover, metapath strategy in Math Overflow is based on homogeneous nodes due to the single type of nodes. We trained our proposed models for a maximum of 1000 epochs with Adam optimizer [48], which is built into tensorflow. Early stopping mechanism is utilized for better efficiency. Weights are initialized through Xavier uniform initialization [49]. We conducted all the experimental tests five times, independently.
Experimental results are summarized in Table 4. Overall, CDHNE achieves the best performance among the four datasets on two criteria, namely, AUROC and AUPRC. The notable improvement validates the effectiveness of our model in extracting comprehensive features from dynamic heterogeneous networks, while avoiding high computational overheads. Profiting from the appropriate encoding and aggregating for subgraph-level features, CDHNE significantly outperforms metapath2vec. We also concentrate on the representation of macro semantic information, which also lead to high practical connotation in heterogeneous network relationships mining. Specifically, our proposed model achieves highest score (0.903 for AUROC and 0.885 for AUPRC) in DBLP, which surpasses the static homogeneous baselines by an average of 14.7% and 10.9%, respectively. Besides, our CDHNE conduces to performance gains over latest dynamic heterogeneous methods DHNE and DyHATR with 14.6% and 4%, respectively, for AUROC in DBLP. However, on the Math Overflow dataset, our model performs slightly higher than static homogeneous methods, which attributes to the fact that there is only one type of node in Math Overflow, so that our heterogeneous encoding component scarcely take effects.
As presented in Figure 5, we further vary the ratio of training set from 20% to 80% with the step of 10%. Five typical methods are selected for comparing the impact of different training ratio. Obviously, our model outperforms other methods, whether the training set is large or small, which validates the efficiency of CDHNE in extracting comprehensive node features. It should be noted that practically all methods perform poorly at low training ratios, while the value of AUROC grows rapidly when training ratio reaches 40% and becomes smooth at high training ratios. Thus, we can conclude that sufficient learning of node features brings advantages in handling relation prediction tasks.

5.4. Community Detection

To evaluate the performance of our proposed model in community structure detection, we use modularity as an assessment criteria. Since the concept of modularity was introduced, various related approaches have been proposed [50,51,52,53,54,55]. Among them, Louvain [50] attempts to discover communities by maximizing modularity with the greedy mechanism. Quick community adaptation (QCA) [52] is elaborated for tracking the evolution of community over time and updating the community structure simultaneously. Batch [51] is a batch-based incremental technique that relies on predefined strategies. GreMod [53] also performs incremental updating for capturing dynamic changes of communities. M-NMF [54] aims to preserve community structure in network embedding through matrix factorization. LBTR-SVM [55] utilized vertex classifier to affect community assignments.
Figure 6 shows the modularity comparison between six typical algorithm and our proposed model on four real-world datasets, respectively. Apparently, we can observe that CDHNE substantially outperforms the other algorithms. Specifically, CDHNE achieves on average 19%, 10.1%, 12.2% and 6.7% percent higher than GreMod on all the snapshots of DBLP, AMiner, EComm and Math Overflow, respectively, which demonstrate the superiority of our proposed model in assigning communities. Even compared with Louvain, our method still reaches only 0.2%, 0.7%, 1.6% and 2.3% percent lower on all the snapshots of DBLP, AMiner, EComm and Math Overflow, respectively. The effectiveness of CDHNE in detecting community structure comes from the pre-learning strategy, with which we first intensify the heterogeneous feature extraction at subgraph-level, then conduct clustering in the embedded space by minimizing the KL divergence. Different from capturing the tiny changes for each snapshot in a dynamic heterogeneous network, we highlight the importance of node representation learning in community detection. We visualize the first snapshot of four datasets in a low dimensional space from the learned node embedding during community-level semantic extraction. The result displayed in Figure 7 demonstrates the capacity of CDHNE in assigning communities, while embodying the drawbacks of randomly initialized centroids in handling large scale networks, according to Figure 7d.

5.5. Granularity of Snapshots Sampling

In this work, we heuristically analyze the impact of sampling granularity on relation prediction tasks, which mainly reflected in the number of snapshots. Normally, we consider the dynamic heterogeneous network as an observed series, G = G 1 , G 2 , , G T , which, to some extent, discretize the network at a certain sampling frequency. It is evident that different sampling granularity can bring varied effect on node representation learning, which is a problem worth pondering over. Taking Math Overflow as an example, the time span is 2350 days, which we divided into 11 snapshots in previous tasks, in which the duration of each snapshot is nearly 214 days. Figure 8 displays the experiment results of CDHNE on Math Overflow. We can observe that there exists certain regularity of AUROC variations in terms of embedding dimension and number of snapshots. Specifically, the value of AUROC grows as the number of sampled snapshots increases when the embedding dimension is fixed. The performance of CDHNE-LSTM continuously improves from 0.731 up to 0.784 in the range of 7 , 9 , 11 , 13 , 15 , 17 , 19 snapshots under the condition that the embedded dimension is configured as 128. Thus, as the number of snapshots grows, our proposed model is capable of capturing comprehensive information over the dynamic network and perceiving more explicit information and implicit associations for predicting link existence.
In addition, we also explore the effect of the embedding dimension in node representation learning on the Math Overflow dataset. The value of AUROC keeps growing in the range of 4 , 8 , 16 , 32 , 64 , 128 embedding dimension. However, the AUROC drops with embedding dimension of 256, indicating that oversize embedding dimension can cause overfitting problems. We can conclude that the finer embedding dimension and granularity are necessary ingredients in learning node representations and capturing dynamic patterns.

5.6. Sensitivity of Hyper-Parameters

In this section, we investigate the effect of different hyper-parameter setting on relation prediction tasks. We evaluate two variants of our proposed model, i.e., CDHNE-GRU and CDHNE-LSTM, which differ in capturing temporal dynamics, on four real-world datasets. As Figure 9 shows, the value of trade-off λ between micro and macro node embedding influences the performance of our model, which measures the importance of micro and macro node embedding. All models are consistently reinforced with larger value of λ and reach highest AUROC at λ = 0.8 , which indicates the significance of micro node embedding involving localized heterogeneous contents. Interestingly, we find that the value of AUROC drops significantly when λ decreases from 0.4 to 0.2 on Math Overflow compared with other datasets. Due to the fact that the node types in the Math Overflow are single, while the heterogeneity that is mainly reflected in the types of edges and its community structure is not obvious according to previous analysis, thereby performing relatively poorly under this circumstance.
Moreover, we further explore the effect of different batch size during the model training phase. Figure 10 shows that with more data added in one batch, the performance of two variants fluctuate slightly in a uptrend, accelerating the training speed and increasing the parallelism. The best AUROC achieved by CDHNE-LSTM in DBLP, AMiner, EComm and Math Overflow are 0.903, 0.854, 0.725 and 0.775, respectively. The investigation of hyper-parameters helps us to find the best setting of our models.

6. Conclusions

In this paper, we abstract the source of user-generated content into a dynamic heterogeneous network and present a novel graph representation learning method, named community-aware dynamic heterogeneous network embedding, for assessing complicated relations in whole graph, abbreviated as CDHNE. Our proposed model mainly consists of three components for micro, macro node representation learning and for capturing evolutionary patterns, respectively. Based on the Markov-chain-optimized metapath, our model is able to learn the heterogeneous contents using skip-gram model. The procedure of edge-centric attention makes features aggregation at the subgraph level. We further explore the latent community structure through the clustering technique and obtain the node embedding from a macro perspective. Using an intuitive aggregation mechanism, these two parts jointly incorporate both graph structure and heterogeneous side information (e.g., node and edge features). Ultimately, we present two variants of our model with different recurrent memory unit, i.e., CDHNE-GRU and CDHNE-LSTM, for dynamics learning. Our experimental analysis shows that the well-learned useful and discriminative network information, resulting in an omnipotent representation space, leads to the effectiveness of our proposed model in various downstream tasks, such as relation assessment and community detection, compared with other state-of-the-art methods. Moreover, the visualization of CDHNE on four datasets highlights its validity in network global information extraction. The stable and competitive performance also shows the reliability of our model while under different sampling granularity.
However, there are also some problems and salient drawbacks in our work. Our proposed model is de facto a combination of graph representation learning with sequence models. The graph embedding techniques captures the heterogeneous node information from multiple perspectives, while the sequence model captures the long-term dependencies within network evolution. This solution converts the dynamic network into a static network, which enables the use of various techniques for static networks, while enlarging the potential loss of information. Further research is needed in three main directions. Firstly, we will concentrate more on digging continuous time information with less information loss. Secondly, theoretical work is needed to analysis the stability of the CDHNE to network perturbations. Finally, the implementation of our method will be reconstructed in an end-to-end way.

Author Contributions

Conceptualization, R.H. and Z.C.; methodology, R.H. and Z.C.; software, Z.C.; validation, R.H., Z.C., J.H. and X.C.; formal analysis, R.H. and Z.C.; investigation, R.H. and Z.C.; resources, R.H.; data curation, Z.C.; writing—original draft preparation, Z.C.; writing—review and editing, R.H., Z.C., J.H. and X.C.; visualization, R.H. and Z.C.; supervision, R.H.; project administration, R.H.; funding acquisition, R.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Natural Science Foundation of China under Grant 61673178 and 61922063; in part by Natural Science Foundation of Shanghai under Grant 20ZR1413800; in part by European Union’s Horizon 2020 research and innovation programme under the Marie Skodowska-Curie grant agreement No 824019 and 101022280.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Narang, K.; Krishnan, A.; Wang, J.; Yang, C.; Sundaram, H.; Sutter, C. Ranking User-Generated Content via Multi-Relational Graph Convolution. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 470–480. [Google Scholar]
  2. Yi, F.; Chen, M.; Sun, W.; Min, X.; Tian, Y.; Zhai, G. Attention Based Network For No-Reference UGC Video Quality Assessment. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1414–1418. [Google Scholar]
  3. Sun, W.; Wang, T.; Min, X.; Yi, F.; Zhai, G. Deep learning based full-reference and no-reference quality assessment models for compressed ugc videos. In Proceedings of the 2021 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  4. Gao, Z.; Zhai, G.; Deng, H.; Yang, X. Extended geometric models for stereoscopic 3D with vertical screen disparity. Displays 2020, 65, 101972. [Google Scholar] [CrossRef]
  5. Li, L.; Lin, X.; Zhai, Y.; Yuan, C.; Zhou, Y.; Qi, J. User communities and contents co-ranking for user-generated content quality evaluation in social networks. Int. J. Commun. Syst. 2016, 29, 2147–2168. [Google Scholar] [CrossRef]
  6. Wang, D.; Jiang, M.; Syed, M.; Conway, O.; Juneja, V.; Subramanian, S.; Chawla, N.V. Calendar graph neural networks for modeling time structures in spatiotemporal user behaviors. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Washington, DC, USA, 14–18 August 2020; pp. 2581–2589. [Google Scholar]
  7. Wang, H.; Le, Z. Seven-layer model in complex networks link prediction: A survey. Sensors 2020, 20, 6560. [Google Scholar] [CrossRef] [PubMed]
  8. Xu, M.; Liu, W.; Xu, J.; Xia, Y.; Mao, J.; Xu, C.; Hu, S.; Huang, D. Recurrent Neural Network Based Link Quality Prediction for Fluctuating Low Power Wireless Links. Sensors 2022, 22, 1212. [Google Scholar] [CrossRef]
  9. Zhang, Z.; Cui, L.; Wu, J. Exploring an edge convolution and normalization based approach for link prediction in complex networks. J. Netw. Comput. Appl. 2021, 189, 103113. [Google Scholar] [CrossRef]
  10. Leicht, E.A.; Holme, P.; Newman, M.E. Vertex similarity in networks. Phys. Rev. E 2006, 73, 026120. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Lü, L.; Zhou, T. Link prediction in complex networks: A survey. Phys. A Stat. Mech. Its Appl. 2011, 390, 1150–1170. [Google Scholar] [CrossRef] [Green Version]
  12. Hamilton, W.L. Graph representation learning. Synth. Lect. Artifical Intell. Mach. Learn. 2020, 14, 1–159. [Google Scholar] [CrossRef]
  13. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  14. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  15. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  16. Koptelov, M.; Zimmermann, A.; Crémilleux, B.; Soualmia, L. Link prediction via community detection in bipartite multi-layer graphs. In Proceedings of the 35th Annual ACM Symposium on Applied Computing, Online, 30 March 2020; pp. 430–439. [Google Scholar]
  17. Kong, C.; Li, H.; Zhang, L.; Zhu, H.; Liu, T. Link prediction on dynamic heterogeneous information networks. In International Conference on Computational Data and Social Networks; Springer: Berlin/Heidelberg, Germany, 2019; pp. 339–350. [Google Scholar]
  18. Yin, Y.; Ji, L.X.; Zhang, J.P.; Pei, Y.L. Dhne: Network representation learning method for dynamic heterogeneous networks. IEEE Access 2019, 7, 134782–134792. [Google Scholar] [CrossRef]
  19. Xue, H.; Yang, L.; Jiang, W.; Wei, Y.; Hu, Y.; Lin, Y. Modeling dynamic heterogeneous network for link prediction using hierarchical attention with temporal rnn. arXiv 2020, arXiv:2004.01024. [Google Scholar]
  20. Ji, Y.; Jia, T.; Fang, Y.; Shi, C. Dynamic Heterogeneous Graph Embedding via Heterogeneous Hawkes Process. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2021; pp. 388–403. [Google Scholar]
  21. Xie, Y.; Ou, Z.; Chen, L.; Liu, Y.; Xu, K.; Yang, C.; Zheng, Z. Learning and Updating Node Embedding on Dynamic Heterogeneous Information Network. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, Virtual, 8–12 March 2021; pp. 184–192. [Google Scholar]
  22. Liben-Nowell, D.; Kleinberg, J. The link-prediction problem for social networks. J. Am. Soc. Inf. Sci. Technol. 2007, 58, 1019–1031. [Google Scholar] [CrossRef] [Green Version]
  23. Niwattanakul, S.; Singthongchai, J.; Naenudorn, E.; Wanapu, S. Using of Jaccard coefficient for keywords similarity. In Proceedings of the International Multiconference of Engineers and Computer Scientists, Hong Kong, China, 13–15 March 2013; Volume 1, pp. 380–384. [Google Scholar]
  24. Adamic, L.A.; Adar, E. Friends and neighbors on the web. Soc. Netw. 2003, 25, 211–230. [Google Scholar] [CrossRef] [Green Version]
  25. Zhou, T.; Lü, L.; Zhang, Y.C. Predicting missing links via local information. Eur. Phys. J. B 2009, 71, 623–630. [Google Scholar] [CrossRef] [Green Version]
  26. Barabási, A.; Jeong, H.; Néda, Z.; Ravasz, E.; Schubert, A.; Vicsek, T. Evolution of the social network of scientific collaborations. Phys. A Stat. Mech. Its Appl. 2002, 311, 590–614. [Google Scholar] [CrossRef] [Green Version]
  27. Ribeiro, L.F.; Saverese, P.H.; Figueiredo, D.R. struc2vec: Learning node representations from structural identity. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 385–394. [Google Scholar]
  28. Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1025–1035. [Google Scholar]
  29. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  30. Goyal, P.; Chhetri, S.R.; Canedo, A. dyngraph2vec: Capturing network dynamics using dynamic graph representation learning. Knowl.-Based Syst. 2020, 187, 104816. [Google Scholar] [CrossRef]
  31. Chen, J.; Wang, X.; Xu, X. GC-LSTM: Graph convolution embedded LSTM for dynamic network link prediction. Appl. Intell. 2021, 1–16. [Google Scholar] [CrossRef]
  32. Selvarajah, K.; Ragunathan, K.; Kobti, Z.; Kargar, M. Dynamic Network Link Prediction by Learning Effective Subgraphs using CNN-LSTM. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  33. Sankar, A.; Wu, Y.; Gou, L.; Zhang, W.; Yang, H. Dysat: Deep neural representation learning on dynamic graphs via self-attention networks. In Proceedings of the 13th International Conference on Web Search and Data Mining, Houston, TX, USA, 3–7 February 2020; pp. 519–527. [Google Scholar]
  34. Pareja, A.; Domeniconi, G.; Chen, J.; Ma, T.; Suzumura, T.; Kanezashi, H.; Kaler, T.; Schardl, T.; Leiserson, C. Evolvegcn: Evolving graph convolutional networks for dynamic graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, Houston, TX, USA, 3–7 February 2020; Volume 34, pp. 5363–5370. [Google Scholar]
  35. Zhang, C.; Song, D.; Huang, C.; Swami, A.; Chawla, N.V. Heterogeneous graph neural network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 793–803. [Google Scholar]
  36. Dong, Y.; Chawla, N.V.; Swami, A. metapath2vec: Scalable representation learning for heterogeneous networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 135–144. [Google Scholar]
  37. Cen, Y.; Zou, X.; Zhang, J.; Yang, H.; Zhou, J.; Tang, J. Representation learning for attributed multiplex heterogeneous network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1358–1368. [Google Scholar]
  38. He, Y.; Song, Y.; Li, J.; Ji, C.; Peng, J.; Peng, H. Hetespaceywalk: A heterogeneous spacey random walk for heterogeneous information network embedding. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 639–648. [Google Scholar]
  39. Lao, N.; Cohen, W.W. Relational retrieval using a combination of path-constrained random walks. Mach. Learn. 2010, 81, 53–67. [Google Scholar] [CrossRef] [Green Version]
  40. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems; Elsevier: Amsterdam, The Netherlands, 2013; pp. 3111–3119. [Google Scholar]
  41. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
  42. Moreira, C.; Calado, P.; Martins, B. Learning to rank academic experts in the DBLP dataset. Expert Syst. 2015, 32, 477–493. [Google Scholar] [CrossRef] [Green Version]
  43. Tang, J.; Zhang, J.; Yao, L.; Li, J.; Zhang, L.; Su, Z. Arnetminer: Extraction and mining of academic social networks. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 990–998. [Google Scholar]
  44. Paranjape, A.; Benson, A.R.; Leskovec, J. Motifs in temporal networks. In Proceedings of the tenth ACM International Conference on Web Search and Data Mining, Cambridge, UK, 6–10 February 2017; pp. 601–610. [Google Scholar]
  45. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [Green Version]
  46. Shang, J.; Qu, M.; Liu, J.; Kaplan, L.M.; Han, J.; Peng, J. Meta-path guided embedding for similarity search in large-scale heterogeneous information networks. arXiv 2016, arXiv:1610.09769. [Google Scholar]
  47. Li, X.; Wu, Y.; Ester, M.; Kao, B.; Wang, X.; Zheng, Y. Semi-supervised clustering in attributed heterogeneous information networks. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 1621–1629. [Google Scholar]
  48. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  49. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  50. Blondel, V.D.; Guillaume, J.L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 2008, P10008. [Google Scholar] [CrossRef] [Green Version]
  51. Chong, W.H.; Teow, L.N. An incremental batch technique for community detection. In Proceedings of the 16th International Conference on Information Fusion, Istanbul, Turkey, 9–12 July 2013 2013; pp. 750–757. [Google Scholar]
  52. Nguyen, N.P.; Dinh, T.N.; Xuan, Y.; Thai, M.T. Adaptive algorithms for detecting community structure in dynamic social networks. In Proceedings of the 2011 Proceedings IEEE INFOCOM, Shanghai, China, 10–15 April 2011; pp. 2282–2290. [Google Scholar]
  53. Shang, J.; Liu, L.; Xie, F.; Chen, Z.; Miao, J.; Fang, X.; Wu, C. A real-time detecting algorithm for tracking community structure of dynamic networks. arXiv 2014, arXiv:1407.2683. [Google Scholar]
  54. Wang, X.; Cui, P.; Wang, J.; Pei, J.; Zhu, W.; Yang, S. Community preserving network embedding. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  55. Shang, J.; Liu, L.; Li, X.; Xie, F.; Wu, C. Targeted revision: A learning-based approach for incremental community detection in dynamic networks. Phys. A Stat. Mech. Its Appl. 2016, 443, 70–85. [Google Scholar] [CrossRef]
Figure 1. A user generated dynamic heterogeneous network example: a commodity supply–demand network and its generalization of abstract graph structure. The solid line denotes the relations between customers and items, the dotted lines indicate the relations between customers, and the dashed lines indicate the relations between merchants and the produced items.
Figure 1. A user generated dynamic heterogeneous network example: a commodity supply–demand network and its generalization of abstract graph structure. The solid line denotes the relations between customers and items, the dotted lines indicate the relations between customers, and the dashed lines indicate the relations between merchants and the produced items.
Sensors 22 01402 g001
Figure 2. An illustrative example of a heterogeneous commodity supply–demand network guided by different metapaths.
Figure 2. An illustrative example of a heterogeneous commodity supply–demand network guided by different metapaths.
Sensors 22 01402 g002
Figure 3. The schematic diagram of calculating the weight coefficient of edge based neighbors for the middle node.
Figure 3. The schematic diagram of calculating the weight coefficient of edge based neighbors for the middle node.
Sensors 22 01402 g003
Figure 4. The overall architecture of the proposed CDHNE. At the t-th snapshot, CDHNE extracts micro node representations and macro structure semantic sequentially, and fuse them with parameterized weights as the input of temporal dynamic extraction model. The changes of entities and semantics in DyHN are captured with RNN-based model due to its superiority in handling long sequences. Finally, the output are transformed to the probability distribution through the fully-connected decoder.
Figure 4. The overall architecture of the proposed CDHNE. At the t-th snapshot, CDHNE extracts micro node representations and macro structure semantic sequentially, and fuse them with parameterized weights as the input of temporal dynamic extraction model. The changes of entities and semantics in DyHN are captured with RNN-based model due to its superiority in handling long sequences. Finally, the output are transformed to the probability distribution through the fully-connected decoder.
Sensors 22 01402 g004
Figure 5. The impact of training ratio on DBLP, AMiner, EComm and Math Overflow datasets in term of area under the receiver operating characteristic curve (AUROC). Five typical baselines, i.e., node2vec, GAT, matapath2vec, DHNE and DyHATR, are selected to compare with our proposed models, which validates that CDHNE-GRU/LSTM can effectively handle relation prediction tasks regardless of the training set size.
Figure 5. The impact of training ratio on DBLP, AMiner, EComm and Math Overflow datasets in term of area under the receiver operating characteristic curve (AUROC). Five typical baselines, i.e., node2vec, GAT, matapath2vec, DHNE and DyHATR, are selected to compare with our proposed models, which validates that CDHNE-GRU/LSTM can effectively handle relation prediction tasks regardless of the training set size.
Sensors 22 01402 g005
Figure 6. The modularity results on four real-world networks. (a) DBLP. (b) AMiner. (c) EComm. (d) Math Overflow. Six typical community detection algorithms, i.e., Louvain, QCA, Batch, GreMod, M-NMF and LBTR-SVM, are selected as competing algorithms. Observe that using the proposed CDHNE can effectively discover the community structure, while reaching a relatively high modularity compared with other methods.
Figure 6. The modularity results on four real-world networks. (a) DBLP. (b) AMiner. (c) EComm. (d) Math Overflow. Six typical community detection algorithms, i.e., Louvain, QCA, Batch, GreMod, M-NMF and LBTR-SVM, are selected as competing algorithms. Observe that using the proposed CDHNE can effectively discover the community structure, while reaching a relatively high modularity compared with other methods.
Sensors 22 01402 g006
Figure 7. A 2D visualization of CDHNE on the first snapshot of DBLP, AMiner, EComm and Math Overflow, respectively. Here, heterogeneous nodes are depicted in the same type. Top 10 communities in scale are colored for better distinction.
Figure 7. A 2D visualization of CDHNE on the first snapshot of DBLP, AMiner, EComm and Math Overflow, respectively. Here, heterogeneous nodes are depicted in the same type. Top 10 communities in scale are colored for better distinction.
Sensors 22 01402 g007
Figure 8. The performance of proposed model CDHNE under different sampling granularity on Math Overflow. The x-axis represents the sampling granularity, that is the number of snapshots, while the y-axis varies with the embedding dimension. The AUROC value for relation prediction task is given in each cell accompanied with a color bar on the right.
Figure 8. The performance of proposed model CDHNE under different sampling granularity on Math Overflow. The x-axis represents the sampling granularity, that is the number of snapshots, while the y-axis varies with the embedding dimension. The AUROC value for relation prediction task is given in each cell accompanied with a color bar on the right.
Sensors 22 01402 g008
Figure 9. The AUROC on four datasets with respect to the trade-off λ between micro and macro node embedding. CDHNE-GRU and CDHNE-LSTM mean the two variants of our model use gated recurrent unit and long short-term memory as temporal dynamic encoder, respectively.
Figure 9. The AUROC on four datasets with respect to the trade-off λ between micro and macro node embedding. CDHNE-GRU and CDHNE-LSTM mean the two variants of our model use gated recurrent unit and long short-term memory as temporal dynamic encoder, respectively.
Sensors 22 01402 g009
Figure 10. The AUROC on four real-world datasets with respect to the training batch size.
Figure 10. The AUROC on four real-world datasets with respect to the training batch size.
Sensors 22 01402 g010
Table 1. Summary of Main Notations.
Table 1. Summary of Main Notations.
NotationDescription
G the set of observed network snapshots
G t the network snapshot at t-th snapshot
A t adjacency matrix of G t
V t the set of nodes at t-th snapshot
E t the set of edges at t-th snapshot
F the mapping function for node types
φ the mapping function for edge types
T the set of node types
R the set of edge types
Tnumber of snapshots in set G
Z t the overall node embedding at t-th snapshot
dnumber of final embedding dimension
Table 2. Statistics of relation prediction datasets.
Table 2. Statistics of relation prediction datasets.
DatasetsTypes#Nodes#Edges#Node-Type#Edge-Type#SnapshotsTime Span
DBLPAcademic132,582275,206331919 years
AminerAcademic41,90168,068331616 years
ECommCommercial37,72491,033241111 days
Math OverflowSocial24,818506,55013112350 days
Table 3. Parameter configurations of the proposed models for relation prediction.
Table 3. Parameter configurations of the proposed models for relation prediction.
ParametersSetting
The trade-off of final node embeddings λ = 0.8
The embedding dimension d = 128
The random walk length l = 100
The number of walks per node50
The number of sampled neighborhood25
The number of negative samples5
The number of filters64
The training rate0.001
The dropout rate0.2
Table 4. Performance comparison on four datasets for the task of relation prediction on dynamic heterogeneous networks. The best results are highlighted in bold.
Table 4. Performance comparison on four datasets for the task of relation prediction on dynamic heterogeneous networks. The best results are highlighted in bold.
MethodsDBLPAMinerECommMath Overflow
AUROC AUPRC AUROC AUPRC AUROC AUPRC AUROC AUPRC
DeepWalk [13]0.7430.7620.7190.7410.5640.5610.7070.756
node2vec [15]0.7470.7660.7240.7460.5970.5940.7130.714
GAT [29]0.7620.7740.7570.7510.6470.6420.7350.763
GraphSAGE [28]0.7730.8010.7610.7540.5950.5900.6370.660
metapath2vec [36]0.8520.8550.7950.8040.6030.6530.6960.749
metapath2vec++ [36]0.8530.8610.7980.8110.6210.6860.7060.751
HetGNN [35]0.8710.8630.7860.7930.6460.6920.7210.734
dyngraph2vec-RNN [30]0.6510.6790.6830.6770.4990.5060.5230.562
dyngraph2vec-AERNN [30]0.6720.6910.6850.6840.5120.5090.5820.597
DySAT [33]0.6590.7010.6930.6860.5080.5110.5010.538
DHNE [18]0.7570.7660.7760.7790.5530.6190.6780.721
DyHATR [19]0.8630.8690.8320.8170.6930.7310.7430.778
CDHNE-GRU (proposed)0.8860.8790.8510.8330.7170.7450.7620.791
CDHNE-LSTM (proposed)0.9030.8850.8540.8410.7250.7510.7750.797
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, R.; Chen, Z.; He, J.; Chu, X. Dynamic Heterogeneous User Generated Contents-Driven Relation Assessment via Graph Representation Learning. Sensors 2022, 22, 1402. https://doi.org/10.3390/s22041402

AMA Style

Huang R, Chen Z, He J, Chu X. Dynamic Heterogeneous User Generated Contents-Driven Relation Assessment via Graph Representation Learning. Sensors. 2022; 22(4):1402. https://doi.org/10.3390/s22041402

Chicago/Turabian Style

Huang, Ru, Zijian Chen, Jianhua He, and Xiaoli Chu. 2022. "Dynamic Heterogeneous User Generated Contents-Driven Relation Assessment via Graph Representation Learning" Sensors 22, no. 4: 1402. https://doi.org/10.3390/s22041402

APA Style

Huang, R., Chen, Z., He, J., & Chu, X. (2022). Dynamic Heterogeneous User Generated Contents-Driven Relation Assessment via Graph Representation Learning. Sensors, 22(4), 1402. https://doi.org/10.3390/s22041402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop