Next Article in Journal
A Backup Protection Based on Compensated Voltage for Transmission Lines Connected to Wind Power Plants
Next Article in Special Issue
GvdsSQL: Heterogeneous Database Unified Access Technology for Wide-Area Environments
Previous Article in Journal
Electromagnetic Environment Assessment and Safety Research of Electrified High-Speed Railway Carriages
Previous Article in Special Issue
A Job Recommendation Model Based on a Two-Layer Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-View Temporal Knowledge Graph Reasoning Framework with Interpretable Logic Rules and Feature Fusion

1
School of Computer Science and Technology, Xi’an Jiaotong University, Xi’an 710049, China
2
School of Mathematics and Statistics, Chongqing Jiaotong University, Chongqing 400074, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(4), 742; https://doi.org/10.3390/electronics13040742
Submission received: 12 January 2024 / Revised: 8 February 2024 / Accepted: 11 February 2024 / Published: 12 February 2024

Abstract

:
A temporal knowledge graph represents temporal information between entities in a multi-relational graph. Its reasoning aims to infer and predict potential links among entities. Predicting time-aware entities is a challenging task due to significant differences in entity appearances over time, such as different timestamps and frequencies. Current embedding-based similarity-matching methods have been introduced for predicting temporal facts. However, they lack deterministic logical explainability and cannot model the dynamic evolution of entities over time. To address these challenges, we propose a novel framework for temporal knowledge graph reasoning based on multi-view feature fusion (MVFF). First, MVFF extracts logical rules and uses the Gumbel-Softmax trick to sample high-quality rules. Second, it integrates logical rules, temporal quadruples, and factual triples to capture semantic features, temporal information, and structural information to solve link prediction tasks. Through experiments on four benchmark datasets, we show that MVFF outperforms state-of-the-art methods, providing not only better performance but also interpretable results.

1. Introduction

Knowledge graphs (KGs) contain facts/triples that indicate relationships between entities, such as (South Korea, Accuse, and North Korea). However, in the real world, many entities and relations are time-dependent. To capture the evolution of information over time, temporal knowledge graphs (TKGs) have been developed. In TKGs, triples (s, p, o) are expanded into quadruples (s, p, o, t), where the time t denotes the validity of the fact. For example, (South Korea, Accuse, North Korea, 17 October 2014) indicates that South Korea accused North Korea on 17 October 2014.
While the TKGs provide a comprehensive description of the dynamics of real-world relationships, specific complexities are introduced when dealing with such structures. Temporal knowledge graph reasoning is realized through a link prediction task that focuses on identifying candidates that are most likely to complete a timestamped query of the form (s, p, ?, t), where s represents the head entity, p represents the relationship, and t represents the timestamp. Traditional learning techniques for static KGs are unable to capture the dynamic interactions between entities. To address this issue, various temporal knowledge graph embedding (KGE) models are proposed to encode entities and relations in a low-dimensional space using translation-based functions [1,2], a deep neural network-based method [3,4], and a tensor decomposition method [5,6]. As these methodologies have progressed, numerous strategies for TKG completion have been proposed, including geometric methods (ChronoR [7], BoxTE [8], TLT-KGE [9], RotateQVS [10], HTTR [11], PTKE [12]); tensor decomposition methods (QDN [13], TBDRI [14]); deep learning and embedding-based methods (TASTER [15], TeAST [16], RoAN [17], BiQCap [18]); and graph neural network-based reasoning methods (TARGCN [19], T-GAP [20], TAL-TKGC [21]). Interpolation is a statistical method that utilizes the relevant known values to estimate an unknown value or set [19]. The goal of extrapolation is to infer “future” unknown values from the current information [22]. These models employ time-dependent embedding of entities and relations to enhance the KG reasoning process. While effective, these methods lack interpretability. As shown in Figure 1, the color shades in the heatmap indicate how often the entity appears at different timestamps. Temporal information can be regarded as perturbations to the representations of entities or relations. As a result, it can be challenging for current models to accurately represent their temporal dynamics over time.
To overcome this issue, logic rule-based symbolic techniques have been proposed as a substitute for embedding-based models. One advantage of these methods is that they produce transparent and comprehensible inference results. However, the subtleties of event complexities can make the manual rule formulation process challenging, which might impede the acquisition of knowledge. Recent works, such as TLogic [23] and LCGE [24], propose methods for temporal KG reasoning and automatically extract knowledge rules. However, there are still several challenges that need to be addressed. First, these techniques used to determine confidence in individual rules sometimes ignore the interaction between several rules, resulting in an oversight of certain rules. Second, temporal rule selection with high-confidence metrics may include inconsistencies and irrelevant facts.
To overcome the limitations of both embedding-based and rule-based methods, we propose the MVFF framework that integrates temporal logical rules, tensor decomposition, time-encoding module, and GNN-based method. MVFF employs closed-loop restrictions and the temporal random walk algorithm [25] to produce a set of candidate rules based on sample entity sequences. The rules obtained after Gumbel sampling are combined with tensor decomposition embedding, time-encoding, and relationship-aware graph neural networks [26] for both scalability and interpretability.
The main contributions of our work include:
  • The proposed rule sampling method is differentiable, which aids in network backpropagation and weight updates. Additionally, it allows for the exploration of rule diversity or the utilization of rules with high confidence through temperature parameters.
  • To capture semantic, structural, and temporal information about temporal knowledge graphs, MVFF combines temporal coding techniques, relation-aware graph neural networks, and tensor decomposition.
  • Experimental results on four benchmark temporal knowledge graph datasets show that our model significantly outperforms many state-of-the-art baseline methods. Our framework improves 2.98%/4.96%/23.97%/6.78% on the MRR evaluation metrics, and 5.03%/3.29%/26.0%/7.82% on the Hit@1 evaluation metrics for datasets ICEWS14/ICEWS05-15/Wikidata12k/GDELT, respectively.
The rest of this paper is organized as follows. Section 2 briefly reviews the related work of temporal knowledge graph reasoning. In Section 3, we introduce the proposed method (MVFF), including the temporal logical rule learning module, temporal logical rule learning module, and multi-view feature fusion module. Afterward, we discuss experiments of temporal knowledge graph reasoning in Section 4. Section 5 delineates the limitations of our method. Section 6 concludes this paper and discusses the future research directions.

2. Related Work

2.1. Embedding-Based Methods

Embedding-based methods mainly include translation-based models, tensor factorization-based models, and neural network-based models. The translation-based models originate from the TransE [27] series of methods, which are extended in TKG by embedding timestamps as a vector. Leblay et al. [1] proposed the TTransE model. Jiang et al. [28] capture the temporal relationship between relationships using the temporal evolution matrix. García-Durán et al. [29] learn the time-aware representation of relationship types. The above methods consider entities and relationships in the same semantic space. However, in different timestamps, entities and relationships may correspond to different tail entities. To solve this problem, Dasgupta et al. [2] borrow the idea of TransH [30] and represent time as a hyperplane. Xu et al. [31] propose the TeRo model to solve the limitations. Yuan et al. [32] proposed the TRHyTE model, which represents temporal relations through a series of hyperplane combinations. The tensor decomposition-based model views TKG as a fourth-order tensor and represents entities, relations, and timestamps as low-dimensional matrices through tensor decomposition. Lacroix et al. [6] decomposed the tensor into a complex space. Goel et al. [4] proposed a time-entity embedding method based on temporal embeddings of entities. Shao et al. [33] used the Tucker decomposition to represent entities, relations, and timestamps by viewing TKG as a fourth-order tensor. A model for the polar temporal knowledge graph reasoning method (PTKE) is proposed by [12]. It is a member of the translation-based model family and integrates time-aware facts into a polar coordinate system. TASTER is a new and efficient TKG reasoning technique that offers a framework for leveraging both global and local information [15]. A relation-oriented attention network (RoAN) is constructed to model temporal embeddings of relations in accordance with the attention mechanism [17]. In order to fully capture the semantics, Wang et al. [13] propose a novel TKG reasoning method called the quadruplet distributor network (QDN), which builds the QDN to facilitate the information aggregation and distribution among items. In the first section on multi-view feature fusion, we employ a fourth-order tensor decomposition method that can effectively encode the basic information of quadruples in a temporal knowledge graph.

2.2. Graph Neural Network-Based Reasoning Methods

Graph reasoning techniques, which represent the correlations in graph structures to accomplish the task of reasoning about temporal knowledge graphs, have been consistently proposed as temporal knowledge graphs and include rich information about graph structures. Wu et al. [3] divided the entity representation into structural and temporal representations. Jin et al. [34] proposed a recursive event network (RE-NET), which uses a neighborhood aggregator to model facts. He et al. [35] proposed a historical information propagation model (HIP). Li et al. [36] proposed a time-directed recurrent graph network (TiRGN) with local–global historical patterns. Jung et al. [20] proposed a new graph neural network encoder that can effectively capture query-related information from TKGs. Nie et al. [21] presented TAL-TKGC, a temporal attention learning-based TKG reasoning approach that consists of an importance-weighted GCN and a temporal attention module. Zhu et al. [37] proposed employing the WinGNN framework to model dynamic graphs. It is based on a straightforward GNN model that incorporates a novel random gradient aggregation mechanism and a meta-learning technique. In the second part of multi-view feature fusion, we utilize a relation-aware graph neural network to encode the structural information of static knowledge graphs, effectively mining the structural associations between entities.

2.3. Temporal Rule-Based Methods

Currently, the use of temporal logic rules in TKGs is relatively underexplored due to three reasons. Firstly, there is a lack of fast and accurate rule extraction methods that do not require manual correction in TKGs. Secondly, logic rule-based methods mostly have a large search space. Additionally, when rules cannot cover all relationships, rule-based systems may not provide answers. A temporal graph encoder and a logical decoder are the two fundamental components of the explainable extrapolation reasoning framework called Temporal LogiCal Graph Networks (TECHS), which was proposed by Lin et al. [38]. Bai et al. [39] proposed a temporal logic-based reasoning model to generate a set of candidate rules and select the optimal rule through pruning strategies. Liu et al. [23] used a random walk method to extract rules. However, this method only considers rules where atomic propositions increase over time and does not consider rules with other time patterns. In the methods section, we propose a method for the extraction and learning of temporal logic rules.
Unlike embedding-based and graph neural network-based methods, our MVFF framework integrates multiple feature representations, offering a more nuanced understanding of temporal dynamics in TKGs. While embedding-based methods focus on static representations and graph neural networks emphasize structural relationships, MVFF captures both static and dynamic features, providing a holistic view of temporal entities and relations. Additionally, while embedding methods and graph neural network approaches lack the ability to explain their reasoning results, our differentiable rule selection method offers interpretability of the reasoning process.

3. Approach

3.1. Overview of MVFF

MVFF can be divided into four modules: (1) the temporal rule module, which conducts explainable reasoning; (2) the four-order tensor decomposition module, which captures both temporal evolution and semantic dependence simultaneously; (3) the RGNN module, which effectively captures structural information; and (4) the time-encoding module, which represents the temporarily valid entities. Figure 2 illustrates an example of interpretable reasoning for temporal rules at the top, and the multi-view feature encoding process is depicted in the subsequent three rectangles.

3.2. Temporal Logical Rule Learning

3.2.1. Temporal Logical Rules

Given the symbolic nature of knowledge graphs (KGs), logic rules are ideally suited for tasks involving temporal knowledge graph link prediction. In rule-based TKG reasoning, temporal logic rules are generally expressed using Horn clauses. They consist of a head and a body, where the head is a single atomic proposition and the body is a conjunction of several atomic propositions. Horn clauses can be represented as Equation (1).
u p q v ,
where u denotes the head of the rule, and p q v refers to the body of the rule. The validity of facts is time-sensitive. Thus, Horn clauses can be used for temporal relationships between facts, as shown in Equation (2).
s , p h , o , t s , p 1 , z 1 , t 1 z n , r n , o , t n ,
where p i ( i = 1 , , h ) denotes the relationships between head and tail entities; s, o, and z j ( j = 1 , , n ) refer to different entities; and t k ( k = 1 , , m ) means timestamp. To enable horn clauses to represent temporal relationships between quadruple propositions, the temporal information needs to satisfy the constraint t 1 t 2 t n t . Based on the length of the temporal logic rule’s body and the temporal relationships, the rules can be divided into different forms. Table 1 shows all temporal logic rules with lengths of 1 and 2.
Figure 3a,b show the general pattern of rules with a length of 1 and 2, respectively, which can be inferred through chain propagation. Figure 3c–f show other patterns of links. To mine temporal logic rules using such links, we add inverse relations to mine potential temporal logic patterns.
We utilize the temporal random walk to extract temporal sequences from TKGs. By adding constraints from cycles during the random walk, circular sequences like those in Figure 3c,d can be obtained by the following Equations (3) and (4). In our environment conditions, the number of tuples impacted by the temporal order is typically rather significant. Stated differently, it is challenging for a single formulation of the temporal logical rule to address them to the greatest extent feasible. We define three compositions of different patterns of linking between entities. When o and s are linked by relation p 1 at timestamp t 1 , and o and z are linked by relation p 2 at t 2 , we can infer that z and s will be linked by relation p h at timestamp t 1 , as shown in Figure 3d. The differences between the three subgraphs, namely Figure 3d–f, lie in the orientation of the arrows. When ‘s’, ‘o’, and ‘z’ are in different directions, they serve as distinct starting and ending entities for the links, resulting in different temporal logical rules being generated.
s , p 1 , o , t 1 , o , p h , s , t t 1 t ,
s , p 1 , o , t 1 , o , p 2 , z , t 2 , z , p h , s , t t 1 t 2 t ,
To extract temporal logic rules from circular sequences of entities, we introduce inverse relationships between connected entities, which are expressed in Equations (5) and (6).
s , p h 1 , o , t s , p 1 , o , t 1 t 1 t ,
s , p h 1 , z , t s , p 1 , o , t 1 o , p 2 , z , t 2 t 1 t 2 t ,
For a rule head relationship, multiple rules may be generated during temporal random walks, which may overlap or conflict with each other. Therefore, it is necessary to determine the confidence of a rule to evaluate the likelihood of its correctness. Using Equation (6) as an example, s, o, and z in the equation represent entity variables, which are replaced with actual entities. The number of specific sequences that satisfy the rule body in the TKGs is referred to as body support, and the number of sequences that satisfy the rule grounding is referred to as rule support. Rule confidence is defined as the ratio of rule support to body support.

3.2.2. Representation of Temporal Logic Rules

To embed temporal logical rules, we represent the relationships as atomic propositions and adopt a linear layer as a time transition operator (W), which maps the relationships in the rule body to the same time as the rule head relationship for correlation calculation via the Hadamard product, as shown in Table 1. We obtain different patterns of rules by setting different time constraints of temporal random walks and then translate the relationships of different times involved in the rules into the same vector space for representation learning, using an approach similar to TransR [40].
By performing temporal random walks, a large number of temporal logical rules can be obtained. Existing methods for applying rules to specific queries directly select the rule with the maximum confidence [23] or use all the rules [24]. The method of selecting the rule with the maximum confidence does not take into account the diversity of the rules. Using all the rules may lead to conflicting rules and cause a significant increase in computational complexity. In this paper, we employ the Gumbel-softmax [41] to sample temporal rules, which address the issues of resource scarcity and slow reasoning that arise from using all rules, as well as the problem of always selecting the rule with the maximum confidence using Argmax. Furthermore, this method retains gradient information during the sampling process and can be optimized using backpropagation.
Suppose there are k rules under the relation p h in a certain pattern, we sample k random numbers ε 1 , , ε k from a uniform distribution and use the Gumbel distribution to transform these random numbers. The resulting values are added to the original probability of each rule. The Gumbel distribution is expressed as follows:
g i Gumbel ( 0 , 1 ) = log ( log u i ) , u i Uniform ( 0 , 1 ) ,
By using the Gumbel-Softmax method, we transfer the randomness of rule sampling to the sampling of the uniform distribution, which has no unknown parameters. This completes the reparameterization process of discrete rules. To maintain differentiability, we apply the softmax function to smooth the approximate result of Argmax. The final rule is obtained as follows:
r u l e = softmax g i + log a i τ i 1 k = exp g i + log a i τ j = 1 k exp g i + log a i τ i 1 k ,
where a i is the probability corresponding to the rule; g i is the Gumbel noise added to the rule; τ is the temperature coefficient, and the smaller its value, the closer the result is to Argmax; k represents the number of temporal logical rules under the patterns with the head relation p h .

3.3. Multi-View Feature Fusion

3.3.1. Tensor Decomposition

In this paper, we utilize a four-order tensor decomposition method (e.g., [6]) that incorporates temporal information to represent entities and relations. Specifically, some relations in the TKG are influenced by timestamps, while others are not. Therefore, we represent the relationship in the four-tuple as a time-constrained representation and a time-independent representation. The tensor decomposition is defined as follows:
X ^ = Re ( S , P t , O ¯ , T ) + ( S , P , O ¯ , 1 ) X ^ i , j , k , l = Re s i , p j t t l + p j , o k ¯ ,
where Re denotes the real part of the complex numbers; S denotes the tensor matrix of the head entity; P t denotes the time-dependent relational tensor matrix; O ¯ denotes the conjugate of the tensor matrix of the tail entity; T denotes the tensor matrix of time; P denotes the time-independent relational tensor matrix; and i , j , k , l denote the tensor indexes in the corresponding matrices, respectively. The time-independent relation p i is embedded using the embedding obtained by learning the temporal logic rules.
In TKGs, some facts will never occur because there is no possible connection between certain types of entities, as shown in Figure 1. At the same time, the method based on tensor decomposition can obtain representations of entities, relationships, and time but does not sufficiently model the structural information in TKGs. Based on these considerations, this paper converts TKG into a static graph by removing the time and then uses RGNN to encode the static knowledge graph.

3.3.2. RGNN Module and Time-Encoding Module

The RGNN model (e.g., [26]) is trained on a static knowledge graph to obtain the static constraints between entities and relations. In a relational multi-graph propagation model, the entity s embedding is updated as follows:
h s ( l + 1 ) = σ p P j N s p 1 c s , p W p ( l ) h j ( l ) + W 0 ( l ) h s ( l ) ,
where a set of neighbor entities with a p-relationship to entity s is represented by the symbol N s p . c s , p is a regularization constant, the value of c s , p is N i r , W p ( l ) is a linear transformation function.
To enable “knowledge” and entities to keep pace with time and update automatically, the time coding module (TE) encodes the temporal information through a neural network and concatenates it with the entity features encoded by the relation-aware graph neural network to complete the coding of the temporal information, as shown in Figure 2.

3.3.3. Learning

For a given quadruple, the rationality of the quadruple is evaluated using the scoring function in Equation (11). Similarly, the triple is scored using the function in Equation (12).
f 1 ( s , p , o , t ) = Re i = 1 d [ s ] i · p t + p r i · [ o ¯ ] i ,
where R e ( · ) represents the real part of a complex number; [ s ] i represents the ith dimension of the complex vector of the head entity; p t represents the relationship representation related to timestamp t; p r represents the relationship representation; o ¯ represents the conjugate of the vector of the tail entity.
f 2 ( s , p , o ) = i = 1 k [ s ] i · [ p ] i · [ o ] i ,
During model optimization, the cross-entropy function is used to calculate the loss of positive and negative samples of quadruples and triples. N3 regularization is used for the embedding representation of entities and relationships based on tensor decomposition. The following displays the objective functions.
L 1 = log exp f 1 ( s , p , o , t ) s ε exp f 1 s , p , o , t log exp f 1 ( s , p , o , t ) o ε exp f 1 s , p , o , t ,
L 2 = log exp f 2 ( s , p , o ) s ε exp f 2 s , p , o log exp f 2 ( s , p , o ) o ε exp f 2 s , p , o ,
L = ( s , p , o , t ) τ L 1 + α · s 3 3 + p t 3 3 + p r 3 3 + o 3 3 + λ · L 2 ,
where α represents the weight of the regularization term, and λ controls the proportion of static embeddings.

4. Experiments

4.1. Datasets

In this paper, we utilized four datasets to validate the effectiveness of the proposed model, including ICEWS14 [10], ICEWS0515 [10], Wikidata12k [2], and GDELT [10]. Dataset statistics are shown in Table 2. The following explains the selection of the above datasets in this paper: First, training distinct time sets can satisfy the variable time intervals in these datasets. Second, the temporal representation of Wikidata12k is similar to that of ICEWS. Third, to give a more precise comparison of the experimental results, we chose the same dataset as the benchmark model.

4.2. Experimental Setup

RGNN was first trained on static knowledge graphs without temporal information. The embedding dimensions of entities and relations ranged from 100, 150, 200, 250, and the learning rate was optimized within 0.01, 0.05, 0.1, 0.2. The batch size of each training data set was set to 30,000. The rule confidence threshold range used was 0.1, 0.2, 0.3, 0.4, 0.5. The range of embedding dimensions of entities and relations in the tensor decomposition is 500, 1000, 1500, 2000, 2500. The learning rate was set within 0.01, 0.05, 0.1, 0.5, and the weight range of N3 regularization was set to 0.01. The ratio of static knowledge features in the loss calculation was set to 2.5, and the batch size was set to 1000. The maximum number of iterations for the model was set to 500. During model training, early stopping was used.

4.3. Baseline Models

We compare MVFF with static knowledge graph methods and temporal knowledge graph methods. Static knowledge graph methods include TransE [27], DisMult [42], ComplEx [5], SimplE [43]. TKG methods are divided into geometric methods (TTransE [1], HyTE [2], TeRo [31], ChronoR [7], BoxTE [8], TLT-KGE [9], RotateQVS [10], HTTR [11], PTKE [12], TGeomE [44]); tensor decomposition methods (TNTComplEx [6], TuckERTNT [33], QDN [13], TBDRI [14]); deep learning and embedding-based methods (TeMP-SA [3], DE-Simple [4], TeLM [45], TASTER [15], TeAST [16], RoAN [17], BiQCap [18]); and graph neural network based-reasoning methods (TARGCN [19], T-GAP [20], TAL-TKGC [21]).

4.4. Measures

We evaluate MVFF using mean reciprocal rank (MRR) and Hits@{1, 3, 10}(H@1, 3, 10).
M R R = 1 N · i N 1 rank i ,
where N is the total amount of test data.
H i t s @ n = 1 N · i N f rank i n ,
If r a n k i < n is true, the value of function f is 1. Typically, n is usually set as 1, 3, or 10. Keep in mind that better performance is indicated by higher MRR and Hits@n.

4.5. Overall Results

In this section, we compare the MVFF model with other TKG reasoning models, including traditional static knowledge graph-based reasoning models, embedding-based TKG reasoning models, and GNN-based TKG reasoning models. The comparison results on the ICEWS2014 and ICEWS05-15 datasets are shown in Table 3, and the comparison results on the Wikidata12k and GDELT datasets are shown in Table 4.
As shown in the tables, the MVFF model achieved the best experimental results in all evaluation metrics. We notice that, across all datasets, the MVFF model consistently and significantly outperforms all baselines. Our technique achieves significant performance increases, particularly when compared to the baselines that perform the best overall. Compared to the state-of-the-art approach, our framework improves 2.98%/4.96%/ 23.97%/6.78% on the MRR evaluation metrics and 5.03%/3.29%/26.0%/7.82% on the Hit@1 evaluation metrics for datasets ICEWS14/ICEWS05-15/Wikidata12k/GDELT, respectively.

4.6. The Impact of Different Hyperparameters

This section investigates the impact of different hyperparameters on model performance.

4.6.1. Comparison of Different Rule Confidence

Based on the knowledge graph reasoning algorithm that uses temporal logical rules, the quality of the rules plays a crucial role in the algorithm’s performance. The rule confidence threshold is a metric utilized to measure the minimum probability value at which a rule is considered to be correct. In this study, we conducted experiments to determine the optimal range of rule confidence thresholds, which include values of 0.1, 0.2, 0.3, 0.4, 0.5. We specified a minimum body support of 20 for the experiment to prevent small body support from leading to high overall rule confidence. Moreover, we held all other parameters constant to disentangle the influence of the rule confidence threshold from other factors. The results, which are displayed in Table 5, illustrate how various rule confidence criteria affect the algorithm’s performance for the ICEWS14 and ICEWS05-15 datasets.

4.6.2. Comparison of Embedding Dimensions

In this study, we investigated the impact of embedding dimensions on the TKG reasoning method based on embedding representation, where entities, relations, and timestamps are embedded into a low-dimensional vector space for link prediction tasks. While higher embedding dimensions can improve model performance, excessively high dimensions can also increase the computational cost. To examine the influence of embedding dimensions, with a batch size of 1000, an RGNN embedding dimension of 250, a learning rate of 0.1, and a rule confidence threshold of 0.2, we conducted comparison analyses on the ICEWS14 dataset. Except for the embedding dimension, all other parameters remained unchanged. According to the experimental results, as illustrated in Figure 4, the model’s performance increases with increasing embedding dimensions—that is, until a certain point, at which point it begins to decrease. As Figure 4 shows, the embedding dimension is fixed at 2000.

4.6.3. The Impact of Rule Sampling Methods

The experiment used a Gumbel-Softmax method to sample rules that meet the conditions for specific relationships. Other methods utilized Argmax to select the rule with the highest confidence score and used all the rules that meet the conditions without sampling. Comparative experiments were conducted on the ICEWS14 dataset while keeping other model parameters constant. The experimental results, as shown in Figure 5, demonstrated that the Gumbel-Softmax method yielded the best performance.
In the comparative analysis of temporal rule sampling methods, the Gumbel-softmax method can flexibly sample the appropriate rules, the Argmax method selects the rules with maximum confidence, and the third method uses all the rules; the experiment was performed on the ICEWS14 dataset. As can be seen in Figure 5, the training evaluation curve of the Gumbel-softmax method has been kept the highest, and as a result, it can be confirmed that this study utilizes differentiable sampling effectively.

4.6.4. Comparison of Different Temperature Coefficients

To explore the impact of the Gumbel-Softmax method on rule sampling, we conducted experiments using different temperature coefficients ( τ ). A higher τ value means more new rules are explored during the sampling process, while a lower τ value means more rules with higher probabilities are exploited. It can be seen from Figure 6 that τ is set to 10, and the result is the best. This implies that the model prefers to explore diverse temporal logic rules.

4.7. Statistics of Extracted Temporal Logic Rules

Table 6 displays the statistics of the rules collected from the ICEWS14 dataset. The number of excursions across each relation was set to 200, and the sequence lengths of the excursions were 2 and 3, corresponding to rules of lengths 1 and 2. Table 7 includes the number of rule instances and the number of rules for various rule confidence levels. Our study presents six distinct temporal reasoning patterns, each with its own confidence level, reflecting the likelihood of one event leading to another within temporal knowledge graphs. These patterns range from direct actions such as accusations leading to rejections (Pattern 1, confidence 0.186) to more complex interactions involving multiple parties and actions, like the influence of making statements and hosting visits on subsequent consultations (Pattern 3, confidence 0.143). Notably, Pattern 2 (confidence 0.506) underscores a significant likelihood of property confiscation following investigations, highlighting a strong predictive relationship in legal or regulatory contexts. By synthesizing these examples, we gain insights into the reasoning power and practical implications of these patterns. This not only validates the effectiveness of our method in capturing complex temporal relationships but also opens avenues for further exploration into the impact of various rule extraction methods on reasoning accuracy.

4.8. Ablation Study

To further analyze the contribution of model components (logic rules, RGNN, and TE), we perform a set of ablation studies. In the MVFF model, the temporal logic rule models the causal relationship using extracted rules, the RGNN encodes structural information for the static knowledge graph, and TE learns the temporal information. To ensure the comparability of the results, the same parameters were used in the experiments. The embedding dimension of RGNN was set to 250, the embedding dimension of tensor decomposition was set to 2000, and the rule confidence threshold was set to 0.2. The experimental results are shown in Table 8. As shown in the table, the performance of MVFF decreases no matter which module is removed, especially after the removal of the RGNN module, the performance of MVFF decreases the most, thus verifying the importance of a static knowledge graph in temporal knowledge graph reasoning.

4.9. Case Study

Our framework provides interpretable reasoning processes, as can be seen in the explainable reasoning module in the top right corner of Figure 2; for the query (South Korea, Reject, ?, 30 April 2024), based on the temporal rules R e j e c t ( s , o , t ) A c c u s e ( s , o , t 1 ) and the existing quadruple (South Korea, Accuse, North Korea, 17 October 2024), we can interpretably answer the query as to where the answer (North Korea) is located.

5. Limitation

Our proposed method integrates logic rules and multi-view feature encoding modules for the task of temporal knowledge graph reasoning. However, the rules extracted under different settings may vary, potentially impacting the final reasoning outcomes. In subsequent research, we will further explore and study the effects of different rule extraction methods on temporal knowledge graph reasoning. Although our multi-view feature fusion approach has already achieved optimal results in reasoning tasks, it lacks mathematical theoretical proof. In future studies, we will continue to investigate how to theoretically explain the advanced nature of our proposed method.

6. Conclusions and Future Work

In this paper, we propose a novel framework that combines logic rule-based and embedding-based methods for temporal knowledge graphs (TKG). The proposed framework provides both the interpretability of logic rule-based methods and the scalability of embedding-based methods. We introduced the pattern of temporal logic rules and demonstrated how they can be automatically extracted in TKG. Additionally, we introduced the Gumbel-Softmax sampling method used in rule sampling and the multi-view representation of TKG, where entities and relations are embedded separately in the static knowledge graph and TKG. Our experimental results demonstrate that the MVFF performs well on the link prediction task. This provides strong evidence that the combination of rule-based and embedding-based methods improves the performance of TKG reasoning.
In the future, we will learn how entities and relations in quaternions evolve as time goes on, transitioning from interpolative to extrapolative reasoning. Additionally, we will theoretically investigate the contribution of static knowledge graph representations to temporal knowledge graph reasoning. Due to the powerful understanding capabilities of large models, we further explore the task of temporal knowledge graph reasoning by integrating it with symbolic reasoning in large models [46].

Author Contributions

Conceptualization, H.X. and J.B.; methodology, H.X. and H.L.; software, H.X., H.L. and J.B.; validation, H.X., J.B. and H.L.; formal analysis, H.X., J.B., C.H. and F.C.; resources, J.B.; writing—original draft preparation, H.X. and J.B.; writing—review and editing, H.X., J.B., C.H. and F.C.; project administration, J.B.; funding acquisition, J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the CAAI-Huawei MindSpore Open Fund and the Science and Technology Research Program of Chongqing Municipal Education Commission (Grant No. KJQN202200718).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Leblay, J.; Chekol, M.W. Deriving validity time in knowledge graph. In Proceedings of the Companion Proceedings of the Web Conference 2018, Lyon, France, 23–27 April 2018; pp. 1771–1776. [Google Scholar]
  2. Dasgupta, S.S.; Ray, S.N.; Talukdar, P. Hyte: Hyperplane-based temporally aware knowledge graph embedding. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2001–2011. [Google Scholar]
  3. Wu, J.; Cao, M.; Cheung, J.C.K.; Hamilton, W.L. Temp: Temporal message passing for temporal knowledge graph completion. arXiv 2020, arXiv:2010.03526. [Google Scholar]
  4. Goel, R.; Kazemi, S.M.; Brubaker, M.; Poupart, P. Diachronic embedding for temporal knowledge graph completion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3988–3995. [Google Scholar]
  5. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex embeddings for simple link prediction. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 20–22 June 2016; pp. 2071–2080. [Google Scholar]
  6. Lacroix, T.; Obozinski, G.; Usunier, N. Tensor Decompositions for Temporal Knowledge Base Completion. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  7. Sadeghian, A.; Armandpour, M.; Colas, A.; Wang, D.Z. Chronor: Rotation based temporal knowledge graph embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 6471–6479. [Google Scholar]
  8. Messner, J.; Abboud, R.; Ceylan, I.I. Temporal knowledge graph completion using box embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 7779–7787. [Google Scholar]
  9. Zhang, F.; Zhang, Z.; Ao, X.; Zhuang, F.; Xu, Y.; He, Q. Along the Time: Timeline-traced Embedding for Temporal Knowledge Graph Completion. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 2529–2538. [Google Scholar]
  10. Chen, K.; Wang, Y.; Li, Y.; Li, A. RotateQVS: Representing Temporal Information as Rotations in Quaternion Vector Space for Temporal Knowledge Graph Completion. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 5843–5857. [Google Scholar] [CrossRef]
  11. Zhao, X.; Li, A.; Jiang, R.; Chen, K.; Peng, Z. Householder Transformation-Based Temporal Knowledge Graph Reasoning. Electronics 2023, 12, 2001. [Google Scholar] [CrossRef]
  12. Liu, R.; Yin, G.; Liu, Z.; Zhang, L. PTKE: Translation-based temporal knowledge graph embedding in polar coordinate system. Neurocomputing 2023, 529, 80–91. [Google Scholar] [CrossRef]
  13. Wang, J.; Wang, B.; Gao, J.; Li, X.; Hu, Y.; Yin, B. QDN: A Quadruplet Distributor Network for Temporal Knowledge Graph Completion. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–13. [Google Scholar] [CrossRef] [PubMed]
  14. Yu, M.; Guo, J.; Yu, J.; Xu, T.; Zhao, M.; Liu, H.; Li, X.; Yu, R. TBDRI: Block decomposition based on relational interaction for temporal knowledge graph completion. Appl. Intell. 2023, 53, 5072–5084. [Google Scholar] [CrossRef]
  15. Wang, X.; Lyu, S.; Wang, X.; Wu, X.; Chen, H. Temporal knowledge graph embedding via sparse transfer matrix. Inf. Sci. 2023, 623, 56–69. [Google Scholar] [CrossRef]
  16. Li, J.; Su, X.; Gao, G. TeAST: Temporal Knowledge Graph Embedding via Archimedean Spiral Timeline. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 15460–15474. [Google Scholar] [CrossRef]
  17. Bai, L.; Ma, X.; Meng, X.; Ren, X.; Ke, Y. RoAN: A relation-oriented attention network for temporal knowledge graph completion. Eng. Appl. Artif. Intell. 2023, 123, 106308. [Google Scholar] [CrossRef]
  18. Zhang, S.; Liang, X.; Li, Z.; Feng, J.; Zheng, X.; Wu, B. BiQCap: A Biquaternion and Capsule Network-Based Embedding Model for Temporal Knowledge Graph Completion. In Proceedings of the International Conference on Database Systems for Advanced Applications, Tianjin, China, 17–20 April 2023; pp. 673–688. [Google Scholar]
  19. Ding, Z.; Ma, Y.; He, B.; Tresp, V. A simple but powerful graph encoder for temporal knowledge graph completion. arXiv 2021, arXiv:2112.07791. [Google Scholar]
  20. Jung, J.; Jung, J.; Kang, U. Learning to walk across time for interpretable temporal knowledge graph completion. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 786–795. [Google Scholar]
  21. Nie, H.; Zhao, X.; Yao, X.; Jiang, Q.; Bi, X.; Ma, Y.; Sun, Y. Temporal-structural importance weighted graph convolutional network for temporal knowledge graph completion. Future Gener. Comput. Syst. 2023, 143, 30–39. [Google Scholar] [CrossRef]
  22. Zhang, M.; Xia, Y.; Liu, Q.; Wu, S.; Wang, L. Learning Latent Relations for Temporal Knowledge Graph Reasoning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 12617–12631. [Google Scholar] [CrossRef]
  23. Liu, Y.; Ma, Y.; Hildebrandt, M.; Joblin, M.; Tresp, V. Tlogic: Temporal logical rules for explainable link forecasting on temporal knowledge graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 4120–4127. [Google Scholar]
  24. Niu, G.; Li, B. Logic and Commonsense-Guided Temporal Knowledge Graph Completion. In Proceedings of the AAAI, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  25. Nguyen, G.H.; Lee, J.B.; Rossi, R.A.; Ahmed, N.K.; Koh, E.; Kim, S. Dynamic network embeddings: From random walks to temporal random walks. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 1085–1092. [Google Scholar]
  26. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the Semantic Web: 15th International Conference, ESWC 2018, Heraklion, Crete, Greece, 3–7 June 2018; pp. 593–607. [Google Scholar]
  27. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 2013, 26, 2787–2795. [Google Scholar]
  28. Jiang, T.; Liu, T.; Ge, T.; Sha, L.; Li, S.; Chang, B.; Sui, Z. Encoding temporal information for time-aware link prediction. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 2350–2354. [Google Scholar]
  29. García-Durán, A.; Dumančić, S.; Niepert, M. Learning Sequence Encoders for Temporal Knowledge Graph Completion. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 4816–4821. [Google Scholar]
  30. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  31. Xu, C.; Nayyeri, M.; Alkhoury, F.; Yazdi, H.S.; Lehmann, J. TeRo: A time-aware knowledge graph embedding via temporal rotation. arXiv 2020, arXiv:2010.01029. [Google Scholar]
  32. Yuan, L.; Li, Z.; Qu, J.; Zhang, T.; Liu, A.; Zhao, L.; Chen, Z. TRHyTE: Temporal Knowledge Graph Embedding Based on Temporal-Relational Hyperplanes. In Proceedings of the Database Systems for Advanced Applications: 27th International Conference, DASFAA 2022, Virtual Event, 11–14 April 2022; pp. 137–152. [Google Scholar]
  33. Shao, P.; Zhang, D.; Yang, G.; Tao, J.; Che, F.; Liu, T. Tucker decomposition-based temporal knowledge graph completion. Knowl.-Based Syst. 2022, 238, 107841. [Google Scholar] [CrossRef]
  34. Jin, W.; Qu, M.; Jin, X.; Ren, X. Recurrent event network: Autoregressive structure inference over temporal knowledge graphs. arXiv 2019, arXiv:1904.05530. [Google Scholar]
  35. He, Y.; Zhang, P.; Liu, L.; Liang, Q.; Zhang, W.; Zhang, C. HIP Network: Historical Information Passing Network for Extrapolation Reasoning on Temporal Knowledge Graph. In Proceedings of the IJCAI, Virtual Event, 19–27 August 2021; pp. 1915–1921. [Google Scholar]
  36. Li, Y.; Sun, S.; Zhao, J. Tirgn: Time-guided recurrent graph network with local-global historical patterns for temporal knowledge graph reasoning. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23–29 July 2022; pp. 2152–2158. [Google Scholar]
  37. Zhu, Y.; Cong, F.; Zhang, D.; Gong, W.; Lin, Q.; Feng, W.; Dong, Y.; Tang, J. WinGNN: Dynamic Graph Neural Networks with Random Gradient Aggregation Window. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA, 6–10 August 2023; pp. 3650–3662. [Google Scholar]
  38. Lin, Q.; Liu, J.; Mao, R.; Xu, F.; Cambria, E. TECHS: Temporal logical graph networks for explainable extrapolation reasoning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 1281–1293. [Google Scholar]
  39. Bai, L.; Yu, W.; Chai, D.; Zhao, W.; Chen, M. Temporal knowledge graphs reasoning with iterative guidance by temporal logical rules. Inf. Sci. 2023, 621, 22–35. [Google Scholar] [CrossRef]
  40. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Volume 29. [Google Scholar]
  41. Jang, E.; Gu, S.; Poole, B. Categorical Reparameterization with Gumbel-Softmax. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  42. Yang, B.; Yih, W.t.; He, X.; Gao, J.; Deng, L. Embedding entities and relations for learning and inference in knowledge bases. arXiv 2014, arXiv:1412.6575. [Google Scholar]
  43. Kazemi, S.M.; Poole, D. Simple embedding for link prediction in knowledge graphs. Adv. Neural Inf. Process. Syst. 2018, 31, 4289–4300. [Google Scholar]
  44. Xu, C.; Nayyeri, M.; Chen, Y.Y.; Lehmann, J. Geometric Algebra Based Embeddings for Static and Temporal Knowledge Graph Completion. IEEE Trans. Knowl. Data Eng. 2022, 35, 4838–4851. [Google Scholar] [CrossRef]
  45. Xu, C.; Chen, Y.Y.; Nayyeri, M.; Lehmann, J. Temporal knowledge graph completion using a linear temporal regularizer and multivector embeddings. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 2569–2578. [Google Scholar]
  46. Xu, F.; Wu, Z.; Sun, Q.; Ren, S.; Yuan, F.; Yuan, S.; Lin, Q.; Qiao, Y.; Liu, J. Symbol-llm: Towards foundational symbol-centric interface for large language models. arXiv 2023, arXiv:2311.09278. [Google Scholar]
Figure 1. Differences in periods and frequencies of entity appearances over time on ICEWS14.
Figure 1. Differences in periods and frequencies of entity appearances over time on ICEWS14.
Electronics 13 00742 g001
Figure 2. Framework of MVFF. Rule-based reasoning is illustrated by explainable reasoning.
Figure 2. Framework of MVFF. Rule-based reasoning is illustrated by explainable reasoning.
Electronics 13 00742 g002
Figure 3. Different patterns of linking between entities. (a,b) show the general pattern of rules with a length of 1 and 2. (cf) are four compositions of temporal logic rules.
Figure 3. Different patterns of linking between entities. (a,b) show the general pattern of rules with a length of 1 and 2. (cf) are four compositions of temporal logic rules.
Electronics 13 00742 g003
Figure 4. Training curves on ICEWS14 for various embedding dimensions.
Figure 4. Training curves on ICEWS14 for various embedding dimensions.
Electronics 13 00742 g004
Figure 5. Training curves on ICEWS14 for different sample methods.
Figure 5. Training curves on ICEWS14 for different sample methods.
Electronics 13 00742 g005
Figure 6. Results of link prediction using different τ (10, 1, 0.1) temperature coefficients.
Figure 6. Results of link prediction using different τ (10, 1, 0.1) temperature coefficients.
Electronics 13 00742 g006
Table 1. Representation of different temporal rule patterns.
Table 1. Representation of different temporal rule patterns.
Rule PatternsTemporal RulesTemporal ConstraintTemporal Rule Representation
Pattern 1 ( s , p h , o , t ) ( s , p 1 , o , t 1 ) t 1 = t p r 1 p r h
Pattern 2 ( s , p h , o , t ) ( s , p 1 , o , t 1 ) t 1 < t ( W t p r 1 ) p r h
Pattern 3 ( s , p h , o , t ) ( s , p 1 , o , t 1 ) ( z 1 , p 2 , o , t 2 ) t 1 < t 2 < t ( W t W t p r 1 ) + ( W t p r 2 ) p r h
Pattern 4 ( s , p h , o , t ) ( s , p 1 , z 1 , t 1 ) ( z 1 , p 2 , o , t 2 ) t 1 = t 2 < t ( W t p r 1 ) + ( W t p r 2 ) p r h
Pattern 5 ( s , p h , o , t ) ( s , p 1 , z 1 , t 1 ) ( z 1 , p 2 , o , t 2 ) t 1 < t 2 = t W t p r 1 + p r 2 p r h
Pattern 6 ( s , p h , o , t ) ( s , p 1 , z 1 , t 1 ) ( z 1 , p 2 , o , t 2 ) t 1 = t 2 = t p r 1 + p r 2 p r h
Table 2. The datasets used in experiments.
Table 2. The datasets used in experiments.
DatasetEntitiesRelationsTime SpanGranularityTrainValidTest
ICEWS1471282301 yearDaily72,82689418963
ICEWS05-1510,48825111 yearDaily368,96246,27546,092
Wikidata12k12,55424539 yearYear32,49740624062
GDELT500201 yearDaily2,735,685341,961341,961
Table 3. Temporal knowledge graph link prediction results on datasets ICEWS14 and ICEWS05-15. The highest score is in the bold, and the second highest score is underlined.
Table 3. Temporal knowledge graph link prediction results on datasets ICEWS14 and ICEWS05-15. The highest score is in the bold, and the second highest score is underlined.
ModelsICEWS14ICEWS05-15
MRR H@1 H@3 H@10 MRR H@1 H@3 H@10
The static KG methods
TransE 10.2800.094-0.6370.2940.090-0.663
DisMult 10.4390.323-0.6720.4560.337-0.691
ComplEx 20.4700.3500.5400.7100.4900.3700.5500.730
SimplE 20.4580.3410.5160.6870.4780.3590.5390.708
The temporal KG methods
Geometric methods
TTransE 10.2550.074-0.6010.2710.084-0.616
HyTE 10.2970.1080.4160.6550.3160.1160.4450.681
TeRo0.5620.4680.6210.7320.5860.4690.6680.795
ChronoR0.6250.5470.6690.7730.6750.5960.7230.820
BoxTE0.6130.5280.6640.7630.6670.5820.7190.820
TLT-KGE0.6340.5510.6840.7860.6900.6090.7410.835
RotateQVS0.5910.5070.6420.7540.6330.5290.7090.813
PTKE0.5540.4440.6240.7600.5720.4540.6490.789
TGeomE0.6290.5460.6800.7800.6860.6050.7360.833
HTTR0.6160.5440.6660.7700.6630.5810.7170.815
Tensor decomposition methods
TNTComplEx0.5600.5200.6600.7600.6700.5900.7100.810
TuckERTNT0.6250.5440.6730.7330.6750.5930.7250.819
QDN0.6430.5670.6880.7840.6920.6110.7430.838
TBDRI0.6520.5520.6970.7850.7090.6460.7570.821
Deep learning and embedding-based methods
TA-DisMult 10.4770.363-0.6860.4740.346-0.728
DE-SimplE 10.5260.4180.5920.7250.5130.3920.5780.748
TeLM0.6250.5450.6730.7740.6780.5990.7280.823
TASTER0.6110.527-0.7670.6540.562-0.818
TeAST0.6370.5600.6820.7820.6830.6040.7320.829
RoAN0.5880.4760.6610.7880.5990.4790.6790.823
BiQCap0.6430.5630.6870.7980.6910.6210.7380.837
Graph neural network-based reasoning methods
TARGCN0.6360.5760.6720.7460.7020.6350.7430.823
T-GAP0.6100.5090.6770.7900.6700.5680.7430.845
TAL-TKGC0.3120.2150.4900.6900.3540.2260.4930.692
MVFF (ours)0.6720.5970.7310.8400.7460.6680.7890.873
1 The Results from RotateQVS [10]. 2 The Results from ChronoR [7]. The other results are from the published paper.
Table 4. Temporal knowledge graph link prediction results on datasets Wikidata12k and GDELT.
Table 4. Temporal knowledge graph link prediction results on datasets Wikidata12k and GDELT.
ModelsWikidata12kGDELT
MRR H@1 H@3 H@10 MRR H@1 H@3 H@10
TeRo0.2990.1980.3290.5070.2450.154 0.420
TeLM0.3330.2310.3600.542----
RotateQVS----0.2700.1750.2930.458
TLT-KGE----0.2480.1650.2680.407
PTKE0.2800.1790.3000.482----
TeAST----0.3710.2830.4010.544
RoAN----0.2900.1870.3150.496
TGeomE0.3330.2320.3620.546----
BiQCap0.3120.2160.3430.5170.2730.1830.3080.469
MVFF (ours)0.4380.3170.5060.6850.3980.3070.4560.573
The highest score is in the bold, and the second highest score is underlined.
Table 5. Results of link prediction on ICEWS14 and ICEWS05-15 with different levels of confidence.
Table 5. Results of link prediction on ICEWS14 and ICEWS05-15 with different levels of confidence.
ModelsICEWS14ICEWS05-15
MRR H@1 H@3 H@10 MRR H@1 H@3 H@10
0.10.6640.5820.7230.8350.7330.6540.7760.863
0.20.6720.5970.7310.8400.7460.6680.7890.873
0.30.6660.5880.7270.8360.7350.6570.7780.865
0.40.6680.5910.7270.8350.7350.6570.7820.865
0.50.6700.5930.7340.8370.7430.6590.7800.870
Table 6. Statistical information on rules in the ICEWS14 dataset.
Table 6. Statistical information on rules in the ICEWS14 dataset.
Pattern 1Pattern 2Pattern 3Pattern 4Pattern 5Pattern 6
Rule length112222
Number of walks per relationship200200200200200200
Time relationship t 1 = t t 1 < t t 1 < t 2 < t t 1 = t 2 < t t 1 < t 2 = t t 1 = t 2 = t
Relationship number376412378321335250
Number of rules407779567105206121481203
Table 7. Examples of rules in ICEWS14.
Table 7. Examples of rules in ICEWS14.
PatternCaseConfidence
Pattern 1 R e j e c t ( s , o , t ) A c c u s e ( s , o , t ) 0.186
Pattern 2 C o n f i s c a t e _ p r o p e r t y ( s , o , t ) I n v e s t i g a t e ( s , o , t 1 ) 0.506
Pattern 3 C o n s u l t ( s , o , t ) M a k e _ s t a t e m e n t ( s , z , t 1 ) H o s t _ a _ v i s i t ( z , o , t 2 ) 0.143
Pattern 4 C o n s u l t ( s , o , t ) M a k e _ s t a t e m e n t ( s , z , t 1 ) A c c u s e ( z , o , t 2 ) 0.141
Pattern 5 I n v e s t i g a t e ( s , o , t ) _ M a k e _ s t a t e m e n t ( s , z , t 1 ) D e m a n d ( z , o , t 2 ) 0.124
Pattern 6 A c c u s e ( s , o , t ) M a k e _ s t a t e m e n t ( s , z , t 1 ) T h r e a t e n ( z , o , t 2 ) 0.150
Table 8. Ablation study of removing the rule module, RGNN, and TE from the whole model MVFF.
Table 8. Ablation study of removing the rule module, RGNN, and TE from the whole model MVFF.
ModelsICEWS14ICEWS05-15
MRRH@1H@3H@10MRRH@1H@3H@10
w/o Rule0.6350.5880.7200.8270.6990.6150.7440.841
w/o RGNN0.6310.5770.7060.8210.6870.6260.7390.840
w/o TE0.6620.5840.7230.8330.7400.6550.7760.862
MVFF0.6720.5970.7310.8400.7460.6680.7890.873
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, H.; Bao, J.; Li, H.; He, C.; Chen, F. A Multi-View Temporal Knowledge Graph Reasoning Framework with Interpretable Logic Rules and Feature Fusion. Electronics 2024, 13, 742. https://doi.org/10.3390/electronics13040742

AMA Style

Xu H, Bao J, Li H, He C, Chen F. A Multi-View Temporal Knowledge Graph Reasoning Framework with Interpretable Logic Rules and Feature Fusion. Electronics. 2024; 13(4):742. https://doi.org/10.3390/electronics13040742

Chicago/Turabian Style

Xu, Hongcai, Junpeng Bao, Hui Li, Chao He, and Feng Chen. 2024. "A Multi-View Temporal Knowledge Graph Reasoning Framework with Interpretable Logic Rules and Feature Fusion" Electronics 13, no. 4: 742. https://doi.org/10.3390/electronics13040742

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop