Next Article in Journal
The BiCG Algorithm for Solving the Minimal Frobenius Norm Solution of Generalized Sylvester Tensor Equation over the Quaternions
Previous Article in Journal
Blood Damage Analysis within the FDA Benchmark Nozzle Geometry at Laminar Conditions: Prediction Sensitivities to Software and Non-Newtonian Viscosity Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unleashing the Power of Decoders: Temporal Knowledge Graph Extrapolation with Householder Transformation

1
School of Data and Computer Science, Shandong Women’s University, Jinan 250300, China
2
School of Information Engineering, Shandong Youth University of Political Science, Jinan 250103, China
3
School of Journalism and Communication, Tsinghua University, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(9), 1166; https://doi.org/10.3390/sym16091166
Submission received: 2 July 2024 / Revised: 27 August 2024 / Accepted: 31 August 2024 / Published: 6 September 2024

Abstract

:
In the realm of artificial intelligence, knowledge graphs (KGs) serve as an essential structured framework, capturing intricate relationships between diverse entities and supporting a broad spectrum of AI applications. Despite their utility, the static characteristic of KGs poses challenges in dynamically evolving information landscapes. This has catalyzed the development of temporal knowledge graphs (TKGs), which introduce a temporal layer to KGs, facilitating the representation of knowledge progression through time. This study zeroes in on the critical task of TKG extrapolation, which is vital for forecasting future occurrences and offering foresight into emerging situations across a variety of fields. Most contemporary approaches to TKG extrapolation are predicated on the symmetrical encoder–decoder paradigm, wherein the processes of representation learning and reasoning are harmoniously intertwined. While the encoder often garners the most attention due to its role in capturing and encoding information, the pivotal role of the decoder, which is often overlooked, is essential for direct inference and the accurate projection of temporal dynamics. To this end, we present the Householder-transformation-based temporal knowledge graph extrapolation (HTKGE) method: a groundbreaking encoder–decoder framework that reimagines the decoder’s contribution to TKG extrapolation. Our approach spotlights an adaptive decoder propelled by Householder transformations, which engage dynamically with the temporal encoding from the encoder. This interaction fosters a nuanced comprehension of the TKG’s temporal trajectory. Our empirical evaluations across four benchmark TKG datasets substantiate HTKGE’s consistent efficacy in TKG extrapolation tasks.

1. Introduction

Knowledge graphs (KGs) have become an indispensable asset in the realm of artificial intelligence, providing a structured and comprehensive framework for representing complex interconnections between disparate entities [1,2,3]. They are the bedrock of modern AI systems, underpinning applications ranging from personalized recommendation engines to advanced question-answering services [4,5]. The ability of KGs to encode semantic relationships has made them a critical component in the development of large-scale AI applications, from virtual assistants to autonomous vehicles. Despite the transformative impact of KGs, their static representation of knowledge poses a significant limitation in a world where information is inherently dynamic. This has catalyzed the evolution of temporal knowledge graphs (TKGs), which introduce a temporal layer to traditional KGs. TKGs not only capture the intricate web of relationships between entities but can also track the evolution of these relationships over time. By embedding timestamps within the graph, TKGs offer a temporal perspective that is crucial for understanding the progression of events, the trajectory of entity interactions, and the shift in relationships—a feature that is particularly valuable for historical analysis, trend prediction and real-time decision-making in various sectors, including finance, healthcare, and social sciences [6,7]. The integration of time as a first-class citizen in knowledge representation marks a significant advancement in the field of AI, enabling more nuanced and context-aware applications that can adapt to the ever-changing landscape of information.
Temporal knowledge graphs (TKGs) are instrumental in modeling the evolution of knowledge over time and encompass facts from an initial time point t 0 up to a terminal time point t T . The reasoning within TKGs is essential for addressing their inherent incompleteness, which can impede their utility in various applications. TKG reasoning is typically categorized into two fundamental tasks: interpolation [8,9,10] and extrapolation [11,12,13]. Interpolation is the process of deducing missing entities or relationships that are expected to exist within the known time frame, i.e., between t 0 and t T . This task is grounded in the existing data structure and is, to a certain extent, informed by the historical context provided by the TKG. On the other hand, extrapolation is the more complex endeavor of forecasting facts that are anticipated to occur after t T . Unlike interpolation, extrapolation must project beyond the given data points to make predictions that are not directly supported by the existing TKG. This task requires an understanding of underlying trends and patterns and the ability to anticipate future developments, which is significantly more challenging and is the primary focus of this paper [14]. Given these considerations, our research endeavors to concentrate on the extrapolation aspect of TKGs. The ability to perform effective extrapolation is crucial for applications that aim to predict future events or to provide insights into forthcoming scenarios, such as financial forecasting, disaster preparedness, and strategic planning.
The prevalent approach to TKG reasoning leverages efficient encoder–decoder [15,16,17] frameworks, which have been widely adopted for their ability to distill and utilize the complex structural and semantic information present in TKGs [11,18]. The encoder component is tasked with learning meaningful representations that capture the essence of the TKG’s topology and the temporal dynamics of its facts. These representations are then passed to the decoder, which is tailored to perform reasoning for specific downstream tasks, such as link prediction or fact retrieval. Among the various methods, Recurrent Event Graph Convolutional Network (REGCN) [14] employs a recurrent structure to sequentially process the temporal information, treating the TKG as an evolving graph wherein nodes and edges are updated over time. This allows the REGCN to capture the temporal dependencies and structural changes within the TKG. On the other hand, the Time-Guided Recurrent Graph Network (TiRGN) [19] introduces a mechanism to integrate both local and global historical patterns and uses a time-guided approach to enhance the reasoning process. The TiRGN model excels at discerning recurrent and periodic behaviors: a capability indispensable for precise forecasting, particularly in sectors where the ubiquity of such patterns is a defining characteristic. While symmetrical encoder–decoder architectures have illuminated their promise in propelling the frontiers of TKG reasoning, a critical examination reveals a potential shortcoming from a technical perspective. The prevailing inclination has been to concentrate enhancements on the encoder component, seeking to distill efficient and information-rich representations from the input data [11,14,19,20,21]. This focus, while beneficial, has inadvertently resulted in relative neglect of the decoder’s equally crucial function within the architecture. The decoder’s role is not merely to reconstruct or generate output from the encoded representation; it is the linchpin that translates compressed information back into a useful form for reasoning tasks. The oversight could lead to suboptimal reasoning capabilities, as the decoder may not be sufficiently adept at interpreting the encoder’s output, especially when dealing with the complexities of temporal dynamics and evolving relationships in TKGs.
Analogous to the findings in the Convolutional Neural Networks for Multi-Relational Learning (ConvR) [22], where the interaction between the convolutional filters and the input features is paramount for capturing the nuanced relationships in multi-relational data, the decoder in TKG reasoning holds a similarly critical position. The decoder’s interaction with the encoded representations is what ultimately translates the learned features into actionable insights and predictions. It is the decoder’s responsibility to interpret the entity and relation embeddings in a manner that is aligned with the task at hand, whether it be predicting future interactions or inferring missing links within the TKG. This underscores the necessity to develop decoders that are not only capable of understanding the complex patterns encoded within the representations but are also adept at leveraging this understanding to make informed predictions. The decoder’s design must facilitate a nuanced exploration of the multifaceted interactions between entities and relations across different temporal contexts. It is this very interaction that enables the model to transcend a simplistic mapping of inputs to outputs to allow for a deeper and more contextualized reasoning process.
In this work, we introduce the Householder-transformation-based temporal knowledge graph extrapolation (HTKGE) method: a novel encoder–decoder architecture specifically tailored to address the intricacies of predicting future facts within temporal knowledge graphs (TKGs). This method stands out due to its innovative approach to the decoder’s role in the reasoning process, which we deem indispensable to the overall performance of TKG extrapolation. While the encoder in our method is acknowledged for its essential function of capturing the temporal dynamics and relational structures of TKGs through a sophisticated recurrent mechanism, it is the decoder that represents our unique contribution. The proposed Householder-transformation-based adaptive decoder is meticulously designed to capitalize on the rich temporal encoding provided by the encoder, enhancing the model’s predictive capabilities through a series of transformative operations. The decoder distinguishes itself by leveraging Householder transformations: a mathematical technique that allows for the creation of dynamic filters. These filters interact intimately with the encoded representations, adeptly capturing the nuanced and evolving interactions between entities and relationships over time. This adaptive feature is inspired by the adaptive filters [22], which enable a deep and contextual understanding of the TKG’s temporal evolution, which is indispensable for making accurate predictions. HTKGE offers a robust and versatile framework for not only predicting future states within TKGs but also for uncovering the latent temporal dependencies that are often obscured in complex temporal data structures.
We summarize our main contributions as follows:
  • We introduce a novel Householder-transformation-based temporal knowledge graph extrapolation (HTKGE) method: an innovative encoder–decoder architecture that addresses the challenge of predicting future facts in temporal knowledge graphs (TKGs).
  • We propose a new Householder-transformation-based adaptive decoder, which is meticulously designed to capitalize on the rich temporal encoding provided by the encoder, enhancing the model’s predictive capabilities through a series of transformative operations.
  • Experiments on four public TKG datasets demonstrate that HTKGE is consistently effective for the TKG extrapolation task.

2. Related Work

2.1. Static KG Reasoning

Within the domain of static KG reasoning, a variety of approaches have been introduced and broadly fall into three categories. The first category includes translation-based models that represent relationships between entities through vector arithmetic, with TransE [23] and its extensions [24,25,26] being particularly noteworthy. The second category, semantic matching models, assesses the plausibility of entity relationships using triangular norms, with DistMult [27] and ComPlEx [28] recognized for employing bilinear and complex number asymmetry, respectively. The third category is query-centric methods, where models as cited in [29,30] aim to capture the distributed nature of KG entities and relations. Enhancements to graph convolutional networks, such as R-GCN [31] and CompGCN [32], have also been proposed to improve relation-awareness in KGs. However, these models mainly focus on static reasoning and often do not account for the temporal dynamics inherent in KGs.

2.2. Interpolation TKG Reasoning

TKG interpolation methods build upon static KG models by incorporating temporal dimensions. TTransE [33], extending TransE, integrates temporal constraints to refine the modeling of time-sensitive relationships. HyTE [34] incorporates timestamps with hyperplanes, blending time directly into the entity-relation context. TA-DistMult [35] extends this approach by intertwining relations and timestamps through an LSTM, addressing the diversity in temporal expressions. TComplEx [36], advancing from ComPlEx, employs tensor decomposition to create timestamp embeddings, thus enhancing temporal adaptability. ATiSE [37] introduces additive time series decomposition to embed temporal information, aligning TKG representations with multidimensional Gaussian distributions to account for uncertainty. TeRo [8] proposes dual embeddings for relations in order to handle the temporal interval’s inception and conclusion. TIMEPLEX-base [38] harmonizes entity, relation, and temporal embeddings, leveraging recurrent facts and temporal interactions for improved predictive tasks. RotateQVS [10] and T-GAP [39] prioritize model interpretability, with RotateQVS utilizing quaternion space for temporal evolution and T-GAP focusing on temporal displacement for inference. TeLM [9] investigates the role of time granularity in TKGs through tensor decomposition, deepening the understanding of temporal dynamics within knowledge graphs.

2.3. Extrapolation TKG Reasoning

Extrapolation reasoning in TKGs, which aim to predict future relationships, has become increasingly popular. The RE-GCN model [14] approaches TKG modeling by factoring in structural and temporal features along with static attributes in a unified manner without separate query encoding. Other models like RE-NET [11] and CyGNeT [40] employ subgraph aggregators and sequential copy networks to represent sequences of facts. TANGO-TuckER [41] and TANGO-DistMult [41] utilize neural ordinary differential equations (ODEs) within GCNs to encode temporal dynamics for better predictive performance. TLogic [42] provides an interpretable framework based on logical rules derived from random walks, while xERTE [43] incorporates a temporal association attention mechanism to preserve causal relationships. TITer [12] stands out due to its use of reinforcement learning and has a time coding function and a reward system based on the Dirichlet distribution. CEN [13] tackles evolutionary patterns with a CNN that is sensitive to sequence lengths and a curriculum learning approach. TiRGN [19] delves into historical dependencies using recurrent graph encoders at both the local and global levels, enhancing the model’s ability to leverage historical contexts for extrapolation. At present, the creators of the majority of these cutting-edge encoder–decoder-based methods for TKG extrapolation have predominantly focused their research efforts on the encoder component, with the aim of learning efficient knowledge representations. This focus has often overshadowed the equally crucial role of the decoder in the reasoning process. This imbalance has emerged as a significant gap in the literature and serves as the primary motivation for our research.

3. Preliminaries

3.1. TKG Extrapolation

TKG extrapolation involves predicting future temporal facts based on a collection of known temporal facts. A TKG is defined as G = E , R , T , F and encompasses entities E , relations R , timestamps T , and facts F . Each fact is articulated as a quadruple ( e s , r , e o , t ) , where the relation r R is predicated between a subject e s E and an object e o E at the specified timestamp t T . To enhance the TKG with temporal dynamics, inverse relations are introduced, augmenting the set of facts with the inverse quadruple ( e o , r 1 , e s , t ) for each existing fact. The observable facts, denoted by 𝒪 , are confined to a time interval [ t 0 , T ] and are represented as 𝒪 = { ( e s , r , e o , t ) | t [ t 0 , T ] } . Given a query Q = ( e q , r q , , t q ) , where ∗ denotes the entity to be determined, the goal of TKG extrapolation is to infer the unknown object entity, presupposing that t q exceeds the latest timestamp T in the observable facts, thus signifying extrapolation into the future beyond the known time frame.

3.2. Householder Transformation

The Householder transformation, a pivotal technique in numerical linear algebra, is adept at performing elementary reflections within a vector space. This method is particularly useful for transforming a given vector into a canonical form wherein all but one of its components are zero.
Given a non-zero vector v R n , the Householder transformation projects v onto a hyperplane defined by a unit vector ω that is normal to the desired reflection direction. The reflection operation results in a new vector v , which is a modified version of v with specific components annihilated. This is graphically illustrated in Figure 1, where the vector v is decomposed into x and y , with v representing x y .
Mathematically, the Householder matrix H ( ω ) is constructed from the unit vector ω and is defined as:
H ( ω ) = I 2 ω ω T
where I is the identity matrix, and ω T denotes the transpose of ω . This matrix H ( ω ) is symmetric and orthogonal and has the property that H ( ω ) = [ H ( ω ) ] T = [ H ( ω ) ] 1 . When applied to a vector v , the transformation H v yields v , maintaining the vector’s two-norm.
The Householder transformation is not only instrumental in QR decomposition but also advantageous in terms of numerical stability and efficiency, as it requires less computational memory and is more precise compared to other methods like the classical Gram–Schmidt process. The QR decomposition using Householder transformations is especially effective for constructing an orthogonal matrix Q from a given matrix A , which can simplify subsequent matrix operations.

4. Proposed Method

In this section, we introduce a novel Householder-transformation-based temporal knowledge graph extrapolation (HTKGE) method, as illustrated in Figure 2. Our HTKGE is an encoder–decoder architecture designed to capture the multi-relational and temporal dynamics of knowledge graphs by implementing a novel application of Householder transformations in the decoder.

4.1. Encoder Architecture

The encoder module in our HTKGE framework is designed to encapsulate the temporal evolution and multi-relational aspects of knowledge graphs, with a focus on capturing the sequential dependencies and historical patterns inherent in temporal knowledge graphs.

4.1.1. RecurrentEvolutionary Embedding

At the core of our encoder is the recurrent evolutionary embedding, which is responsible for updating the representations of entities and relations over successive timestamps. This process is modeled as a recurrent graph convolutional operation that leverages the structural information within the knowledge graph at each timestamp t.
For an entity e and a relation r at timestamp t, the entity’s embedding h e q t is updated by aggregating information from its neighboring entities through a graph convolutional layer, which can be expressed as:
h e q t = GCN θ { h e t 1 e N ( e ) } , W r ,
where N ( e ) denotes the neighbors of entity e, W r represents the weight matrix corresponding to relation r, and GCN θ is the graph convolutional operation parameterized by θ .
Our recurrent evolutionary embedding module is designed to progressively refine entity and relation representations as new information becomes available over time. This component is pivotal as it allows our model to adapt to the ever-changing landscape of TKGs, ensuring that the encoded representations are as current and relevant as possible. The recurrent nature of this embedding enables it to capture the temporal flow and evolving relationships within the graph.

4.1.2. Temporal Dependency Modeling

To capture the temporal dependencies, our encoder employs a recurrent mechanism that attends to the sequence of graph states across timestamps. This is achieved by a gated unit that updates the entity embeddings while retaining the informative historical structure. The updated entity embedding h ¯ e t at time t is computed as:
h ¯ e t = GRU h e q t , h ¯ e t 1 ; U , W , b ,
where GRU is the gated recurrent unit with the update gate parameter U , the hidden state parameter W , and the bias term b . And h ¯ e t 1 is the entity embedding from the previous timestamp t 1 .
Temporal dependency modeling is a cornerstone of our encoder and focuses on the temporal sequences that characterize TKGs. By employing a recurrent mechanism with gating units, we ensure that our model can maintain historical memory while continuously updating its understanding of the graph’s state. This component is essential for capturing the causal relationships and temporal progressions that are vital for accurate TKG extrapolation.

4.1.3. Global Historical Pattern Integration

In addition to the local temporal dependencies, our encoder also incorporates global historical patterns to enhance the reasoning capabilities. This is accomplished by a global history encoder network that accumulates repeated historical facts and identifies periodic behaviors over time. The global embedding g e for an entity e is obtained by:
g e = AGG { h e q t t = 1 , , T } ,
where AGG is an aggregation function, such as mean or sum, over the entity’s embeddings across all timestamps T.
Recognizing the importance of historical context in TKGs, our global historical pattern integration component provides a macroscopic view of the graph’s evolution. By identifying and aggregating repeated historical facts and periodic behaviors, this component enriches the encoder’s understanding with a global perspective. This integration of global patterns is crucial for uncovering long-term trends and cyclical behaviors that may influence future states of the TKG.

4.1.4. Local–Global Interaction

The final step in our encoder architecture is to integrate the local sequential patterns with the global historical patterns. This is achieved by a fusion mechanism that balances the contributions from both the local and global contexts. The fused embedding h e q for each entity e is given by:
h e q = α · h ¯ e T + ( 1 α ) · g e ,
where α is a weighting factor that determines the influence of the local pattern h ¯ e T (at the latest timestamp T) versus the global pattern g e . And we use d to denote the dimension for all the embeddings above.
The resulting fused embeddings h e q serve as the input to the decoder module, which will generate the final predictions. This encoder framework provides a comprehensive representation of the temporal knowledge graph by jointly modeling the local sequential dependencies and the global historical patterns.
The fusion of local sequential patterns with global historical patterns is a key innovation in our encoder architecture. This interaction allows our model to balance the fine-grained, immediate dependencies with broader, long-term trends. The resulting fused embeddings are thus a nuanced representation that encapsulates both the micro and macro dynamics of the TKG, providing a solid foundation for the decoder’s predictive tasks.

4.2. Decoder Architecture with Adaptive Convolution

The decoder in our HTKGE framework is designed to translate the comprehensive entity representations from the encoder into accurate predictive outcomes. This decoder employs adaptive convolution techniques integrated with Householder transformations for embedding combination that align with the numerical linear algebra practices established in the Householder transformation subsection. The framework of our decoder is depicted in Figure 3.

4.2.1. Householder Transformation for Embedding Integration

For a given query q = ( e q , r q , , t q ) , we integrate relation embeddings and time embeddings within the query using the Householder transformation, which is a method known for its utility in numerical linear algebra, particularly for performing elementary reflections within a vector space. Given a relation embedding r q and a time embedding t q , we construct a Householder matrix H ( ω ) parameterized by a unit vector ω , which is derived from the embeddings r q and t q through a predefined function ϕ that captures the desired reflection direction:
ω = ϕ ( r q , t q ) .
The Householder matrix H ( ω ) is then defined as:
H ( ω ) = I 2 ω ω T ,
where I is the identity matrix. This matrix is applied to the relation embedding r to produce a transformed embedding r q that is reflective of the temporal context provided by t q and the relation-specific characteristics of r q :
r q = H ( ω ) r q .

4.2.2. Dynamic Filter Generation

The transformed relation embeddings are then reshaped to create dynamic convolutional filters (the channel number is denoted as n) following the adaptive convolutional network strategy outlined in [22]. Each filter F r q , which is tailored for the relation r q , is generated by reshaping the Householder-transformed relation embedding r q into a 2D filter format conducive for convolution. And considering the channels, we actually have c various filters F r q ( i ) , i = 1 , 2 , · · · n , for each relation:
F r q ( 1 ) , F r q ( 2 ) , · · · F r q ( n ) = split ( RESHAPE r ) ,
where the operation RESHAPE is used to convert a d-dimensional embedding into a three-dimensional matrix suitable for convolutional processing, i.e., d = h × w × n ; here, h and w represent the dimensions of the filter. After a split operation, we have n different filters, where each filter F r q ( i ) R h × w .

4.2.3. Adaptive Convolution Operation

The decoder performs an adaptive convolution operation with the dynamic filters F r q across the entity embeddings h e q from the encoder. This operation is detailed as follows:
Filter application: Each filter F r q ( i ) is convolved with the query entity embedding h e q (obtained from Equation (4)) to produce a feature map c ( i ) . The convolution operation at a specific position is defined as:
c ( i ) ( m , n ) = j = 0 h 1 k = 0 w 1 F r q ( i ) ( j , k ) · h e ( m + j , n + k ) ,
where c ( i ) ( m , n ) is the value at position ( m , n ) in the feature map for the i-th filter, and h e ( m + j , n + k ) is the element of the entity embedding h e q at the corresponding position after padding, if necessary.
Feature map generation: The feature map c is constructed by aggregating the results of the convolution operation across multiple convolution filters:
c = AGG ( c ( i ) | for i = 1 , 2 , · · · n ) .
This adaptive convolution operation is central to the decoder’s capability to capture the intricate interplay between entities and relations within the temporal knowledge graph, enabling HTKGE to forecast future facts with enhanced accuracy. The dynamic nature of the filters allows for relation-specific feature extraction that is sensitive to the unique characteristics of each relation, thereby improving the predictive power of the model.
Scoring for prediction: The final prediction is articulated by a scoring function:
Score q , e a = W c · h e o 0 ,
where W is a learnable weight matrix, and h e a 0 is the initial embedding for a candidate answer entity. In our approach, we refine the prediction process by meticulously selecting the candidate entities that achieve the highest scores as the definitive answers.

4.3. Training Objective

Our HTKGE is refined through the optimization of a cross-entropy loss function. This function quantifies the discrepancy between the anticipated and actual results for a specified query. The loss function L is designed for entity prediction and is formalized below:
L = q , e a T train Y e a log score e t e h , r h , t , t ,
where T train is the set of training queries, and score q , e a is the probabilistic score for the facts calculated by Equation (11). Y e a are the label vectors, where the element is 1 if the fact occurs and 0 otherwise.

5. Experiments

5.1. Experimental Setup

5.1.1. Benchmark Datasets

For assessing our HTKGE method, we utilize four prominent datasets from the TKG extrapolation benchmarks: ICEWS14* [43], ICEWS05-15 [35], WIKI [33], and GDELT [11]. A detailed depiction of these datasets is provided in Table 1. The ICEWS14* and ICEWS05-15 datasets are derived from subsets of the Integrated Crisis Early Warning System (ICEWS) [44], which is an expansive event-based database. The former encompasses events from the year 2014, while the latter includes a decade-long span from 2005 to 2015. For the WIKI dataset, we streamlined the data by excluding month and day information to achieve year-level granularity consistent with the methodology in [11]. The GDELT dataset, extracted from a comprehensive global event repository [45], boasts a detailed 15 min time granularity.
For extrapolation tasks, we organize the factual data from these four datasets chronologically and segment each into training, validation, and test subsets in accordance with the temporal stratification approach of Jin et al. [11], Han et al. [43] in order to guarantee (timestamps for training) < (timestamps for validation) < (timestamps for testing). This chronological division ensures a logical sequence from past events (training set) to future events (test set), enabling the utilization of historical data for predictive analysis.

5.1.2. Evaluation Protocol

The assessment of our proposed method’s capacity for extrapolation reasoning within TKGs is facilitated through a link prediction task. This task is specifically designed to deduce missing entities within temporal facts, which are symbolically represented as either ( e s , r , , t ) or ( , r , e o , t ) , where e s and e o are entities, r symbolizes the relationship, and t indicates the timestamp. For a given test sample ( e s , r , e o , t ) , the process commences with the generation of a candidate quadruple set C, defined as C = { ( e s , r , e o ¯ , t ) : e o ¯ E } { ( e s ¯ , r , e o , t ) : e s ¯ E } , by substituting e s or e o with every entity from the entity set E . Following this, the quadruples are ranked according to their scores, as determined by Equation (11). For reporting our experimental results, we adhere to the time-wise filtered setting as referenced in [37,46]. This approach is selected due to its alignment with TKG reasoning and ensures that facts that are not observable at a specific timestamp t are still accounted for in the assessment of the test quadruple, as elucidated in [8]. This method stands in contrast to time-unwise filtering settings, which, as indicated in [14,23], may lead to potentially incorrect higher-ranking scores.
Performance metrics are reported using the standard benchmarks: the proportion of correct triples ranked within the top 1, 3, and 10 positions (H@1, H@3, and H@10, respectively) and the mean reciprocal rank (MRR). These metrics are designed such that higher values are indicative of better performance. For the sake of consistency, we present the mean results from five experimental runs, electing to omit the variance due to its general minimality.

5.1.3. Baselines

In our comprehensive evaluation, the HTKGE method is rigorously compared against a suite of eleven state-of-the-art models that are recognized for their proficiency in TKG extrapolation. Our baselines include the neural ordinary-equation-based TANGO [41], the probabilistic model CyGNet [40], the reinforcement-learning-based TITer [12], the subgraph-focused xERTE [43], and the rule-based TLogic [42]. Particularly, we place a strong emphasis on comparing with methods that are founded on the encoder–decoder framework, such as RE-NET [11], RE-GCN [14], TiRGN [19], HiSMatch [18], DaeMon [20], and RPC [21].

5.2. Performance Comparison

We present a comparative analysis of our proposed HTKGE method against eleven state-of-the-art models on four experimented datasets: GDELT, ICEWS14*, WIKI, and ICEWS05-15. The performance is evaluated using standard link prediction metrics, including the mean reciprocal rank (MRR) and the proportion of correct triples ranked in the top 1 (H@1), top 3 (H@3), and top 10 (H@10) positions. The results, as presented in Table 2, evidence that HTKGE achieves state-of-the-art (SOTA) performance across the board, with particularly notable superiority on the GDELT, ICEWS14*, and WIKI datasets, where it outperforms existing SOTA methods across all evaluated metrics. These results are indicative of HTKGE’s robustness at capturing the nuances of temporal dependencies within TKGs and its effectiveness at predicting future facts. Specifically, our HTKGE demonstrates a marked improvement over the state-of-the-art (SOTA) method RPC, which also employs an encoder–decoder architecture, across four benchmark datasets. When considering the comprehensive metric MRR, HTKGE achieves a notable enhancement in performance: it outperforms RPC by 6.7% on the GDELT dataset, by 3.4% on the ICEWS14* dataset, by 1.8% on the WIKI dataset, and by 1.9% on the ICEWS05-15 dataset. HTKGE’s consistent outperforming across various datasets underscores its generalizability and reliability in the context of TKG reasoning. The adaptive Householder transformations within our decoder are particularly instrumental to these results and allow for nuanced understanding of and dynamic interaction with the temporal evolution of TKGs. This innovation sets HTKGE apart from other methods and positions it as a formidable candidate for future research and application in the field of TKG extrapolation.
Regarding the fact that our HTKGE does not surpass HiSMatch [18] on ICEWS05-15, we believe this could be due to several factors. Firstly, the ICEWS05-15 dataset may possess unique attributes that are better-suited to the approach taken by HiSMatch. Secondly, despite HTKGE’s consistent performance across most datasets, HiSMatch’s dominance on ICEWS05-15 could be a result of its model architecture, training process, or other undisclosed proprietary enhancements that contribute to its performance.

5.3. Ablation Study

Ablation studies are conducted on two datasets—namely, ICEWS14* and ICEWS05-15—to investigate the impact of individual components on the performance of our HTKGE. The H@1 performances are shown in Figure 4, where five sub-models are compared, including (1) the original HTKGE, (2) a variant of HTKGE that lacks temporal dependency modeling, denoted as “−TP”, (3) another variant without the incorporation of global historical patterns, denoted as “−GP”, (4) a modified version wherein the Householder transformation is replaced by direct addition of the relational and temporal representations, denoted as “+Add”, and (5) a version wherein the Householder transformation is substituted with an element-wise product of the relational and temporal representations, denoted as “+Prod”. Based on the experimental outcomes, the following conclusions can be drawn: (a) Both the “−TP” and “−GP” models exhibit a decline in performance compared to the original HTKGE. This suggests that within the encoder’s design, both temporal dependency modeling and the inclusion of global historical patterns significantly contribute to the overall performance. (b) The “+Add” and “+Prod” models also demonstrate a decrease in performance relative to the original HTKGE. This indicates that the Householder transformation utilized within the decoder’s architecture plays a crucial role in enhancing the model’s predictive capabilities. The observed performance drop underscores the importance of this specific transformation in the decoding process.

5.4. Parameter Analysis on the Decoder

Our exploration into the adaptability and efficiency of the proposed Householder-transformation-based adaptive decoder within the HTKGE framework extends to a thorough parameter analysis. This analysis is pivotal for optimizing the decoder’s configuration to achieve a balance between computational resources and performance metrics. We specifically focus on two critical hyperparameters: the channel number n and the filter size h × w , which are known to substantially influence the decoder’s functional capacity. In our analysis, we meticulously adjust the channel number n across a spectrum of values { 20 , 40 , 60 , 80 , 100 } and concurrently vary the filter size h × w through a set of dimensions { 2 × 2 , 3 × 3 , 4 × 4 , 5 × 5 } . This methodical tuning process is designed to elucidate the relationship between the parameter count and the decoder’s performance, as measured by its ability to accurately predict future facts within TKGs.
The findings of our parameter analysis are encapsulated in Figure 5. As the table illustrates, the parameter count of the decoder increases monotonically with the growth in the channel number n and the dimensions of the filter size h × w . However, the performance metrics, particularly the H@10 score, exhibit a far less pronounced variation. This observation is particularly noteworthy as it suggests that our decoder is capable of achieving a commendable level of performance with a relatively modest parameter count. The relatively flat performance curve across different parameter configurations underscores the robustness of our decoder’s design. It suggests that the adaptive Householder transformations are adept at capturing the nuances of the temporal encoding, even when the complexity of the model is scaled back. This finding is a testament to the effectiveness of our approach and paves the way for further optimizations that could lead to even more efficient models.

6. Conclusions

In this paper, we introduce a novel Householder-based temporal knowledge graph extrapolation (HTKGE) method that makes significant strides in the domain of TKG reasoning by introducing an innovative encoder–decoder architecture. The unique adaptive decoder utilizes Householder transformations and is proven to be instrumental in enhancing predictive capabilities, allowing for a nuanced understanding of temporal dynamics within TKGs. Empirical results on four benchmark datasets confirm HTKGE’s effectiveness, highlighting its potential as a robust framework for future state prediction and the revelation of latent temporal dependencies in TKGs. This work paves the way for further advancements in the field and emphasizes the pivotal role of the decoder in TKG extrapolation tasks.
Limitations and future work: Our model’s performance is contingent upon the quality of input data: high-quality data are essential for precise extrapolations within sparse TKGs. Given the deep learning framework, the quality of parameter training is paramount to our model’s performance. We acknowledge the possibility that our model may have been inadvertently tuned to the specific data distribution and temporal properties of the datasets. In our future work, we will address scalability by refining our approach to manage large TKGs efficiently. We will explore data augmentation to counter data sparsity and seek algorithms that maintain the precision of Householder transformations while enhancing computational efficiency. Additionally, we will focus on improving the decoder’s interpretability by examining its decision-making process with the aim of optimizing its parameters for superior performance. We also intend to explore the causal relationships between events across different datasets, recognizing the complexity and potential impact of such interconnections on our model’s predictive capabilities.

Author Contributions

Conceptualization, Y.Z.; Methodology, F.Y.; Writing—original draft, X.Z.; Writing—review and editing, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bollacker, K.D.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the SIGMOD 2008, Vancouver, BC, Canada, 9–12 June 2008. [Google Scholar]
  2. Miller, G.A. WordNet: A lexical database for English. Commun. ACM 1995, 38, 39–41. [Google Scholar] [CrossRef]
  3. Suchanek, F.M.; Kasneci, G.; Weikum, G. Yago: A Core of Semantic Knowledge. In Proceedings of the 16th International Conference on World Wide Web, WWW ’07, New York, NY, USA, 8–12 May 2007; pp. 697–706. [Google Scholar]
  4. Jiang, L.; Yan, G.; Luo, H.; Chang, W. Improved Collaborative Recommendation Model: Integrating Knowledge Embedding and Graph Contrastive Learning. Electronics 2023, 12, 4238. [Google Scholar] [CrossRef]
  5. Liu, K.; Wu, J.; Sun, Q.; Yang, H.; Wan, R. Harnessing Test-Oriented Knowledge Graphs for Enhanced Test Function Recommendation. Electronics 2024, 13, 1547. [Google Scholar] [CrossRef]
  6. Gao, H.; Wang, Y.; Shao, J.; Shen, H.; Cheng, X. User Identity Linkage across Social Networks with the Enhancement of Knowledge Graph and Time Decay Function. Entropy 2022, 24, 1603. [Google Scholar] [CrossRef] [PubMed]
  7. Du, Z.; Qu, L.; Liang, Z.; Huang, K.; Cui, L.; Gao, Z. IMF: Interpretable multi-hop forecasting on temporal knowledge graphs. Entropy 2023, 25, 666. [Google Scholar] [CrossRef] [PubMed]
  8. Xu, C.; Nayyeri, M.; Alkhoury, F.; Shariat Yazdi, H.; Lehmann, J. TeRo: A Time-aware Knowledge Graph Embedding via Temporal Rotation. In Proceedings of the COLING 2020, Barcelona, Spain, 8–13 December 2020. [Google Scholar]
  9. Xu, C.; Chen, Y.; Nayyeri, M.; Lehmann, J. Temporal Knowledge Graph Completion using a Linear Temporal Regularizer and Multivector Embeddings. In Proceedings of the NAACL-HLT 2021, Online, 6–11 June 2021. [Google Scholar] [CrossRef]
  10. Chen, K.; Wang, Y.; Li, Y.; Li, A. RotateQVS: Representing Temporal Information as Rotations in Quaternion Vector Space for Temporal Knowledge Graph Completion. In Proceedings of the ACL 2022, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
  11. Jin, W.; Qu, M.; Jin, X.; Ren, X. Recurrent Event Network: Autoregressive Structure Inference over Temporal Knowledge Graphs. In Proceedings of the EMNLP 2020, Online, 16–20 November 2020; pp. 6669–6683. [Google Scholar]
  12. Sun, H.; Zhong, J.; Ma, Y.; Han, Z.; He, K. TimeTraveler: Reinforcement Learning for Temporal Knowledge Graph Forecasting. In Proceedings of the EMNLP 2021, Punta Cana, Dominican Republic, 16–20 November 2021; pp. 8306–8319. [Google Scholar] [CrossRef]
  13. Li, Z.; Guan, S.; Jin, X.; Peng, W.; Lyu, Y.; Zhu, Y.; Bai, L.; Li, W.; Guo, J.; Cheng, X. Complex Evolutional Pattern Learning for Temporal Knowledge Graph Reasoning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), ACL 2022, Dublin, Ireland, 22–27 May 2022; Muresan, S., Nakov, P., Villavicencio, A., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 290–296. [Google Scholar] [CrossRef]
  14. Li, Z.; Jin, X.; Li, W.; Guan, S.; Guo, J.; Shen, H.; Wang, Y.; Cheng, X. Temporal Knowledge Graph Reasoning Based on Evolutional Representation Learning. In Proceedings of the SIGIR 2021, Virtual Event, 11–15 July 2021. [Google Scholar] [CrossRef]
  15. Mak, H.W.L.; Han, R.; Yin, H.H.F. Application of Variational AutoEncoder (VAE) Model and Image Processing Approaches in Game Design. Sensors 2023, 23, 3457. [Google Scholar] [CrossRef] [PubMed]
  16. Zhong, G.; Gao, W.; Liu, Y.; Yang, Y.; Wang, D.H.; Huang, K. Generative adversarial networks with decoder–encoder output noises. Neural Netw. 2020, 127, 19–28. [Google Scholar] [CrossRef] [PubMed]
  17. Saha, U.; Saha, S.; Fattah, S.A.; Saquib, M. Npix2Cpix: A GAN-Based Image-to-Image Translation Network With Retrieval- Classification Integration for Watermark Retrieval From Historical Document Images. IEEE Access 2024, 12, 95857–95870. [Google Scholar] [CrossRef]
  18. Li, Z.; Hou, Z.; Guan, S.; Jin, X.; Peng, W.; Bai, L.; Lyu, Y.; Li, W.; Guo, J.; Cheng, X. HiSMatch: Historical Structure Matching based Temporal Knowledge Graph Reasoning. In Proceedings of the EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 7328–7338. [Google Scholar] [CrossRef]
  19. Li, Y.; Sun, S.; Zhao, J. TiRGN: Time-Guided Recurrent Graph Network with Local-Global Historical Patterns for Temporal Knowledge Graph Reasoning. In Proceedings of the IJCAI 2022, Vienna, Austria, 23–29 July 2022. [Google Scholar] [CrossRef]
  20. Dong, H.; Ning, Z.; Wang, P.; Qiao, Z.; Wang, P.; Zhou, Y.; Fu, Y. Adaptive Path-Memory Network for Temporal Knowledge Graph Reasoning. In Proceedings of the IJCAI 2023, Macao, China, 19–25 August 2023. [Google Scholar] [CrossRef]
  21. Liang, K.; Meng, L.; Liu, M.; Liu, Y.; Tu, W.; Wang, S.; Zhou, S.; Liu, X. Learn from Relational Correlations and Periodic Events for Temporal Knowledge Graph Reasoning. In Proceedings of the SIGIR 2023, Taipei, Taiwan, 23–27 July 2023. [Google Scholar] [CrossRef]
  22. Jiang, X.; Wang, Q.; Wang, B. Adaptive convolution for multi-relational learning. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 978–987. [Google Scholar]
  23. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–8 December 2013; pp. 2787–2795. [Google Scholar]
  24. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge Graph Embedding by Translating on Hyperplanes. In Proceedings of the AAAI 2014, Québec, QC, Canada, 27–31 July 2014. [Google Scholar]
  25. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning Entity and Relation Embeddings for Knowledge Graph Completion. In Proceedings of the AAAI 2015, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  26. Sun, Z.; Deng, Z.; Nie, J.; Tang, J. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  27. Yang, B.; Yih, W.; He, X.; Gao, J.; Deng, L. Embedding Entities and Relations for Learning and Inference in Knowledge Bases. In Proceedings of the ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  28. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex Embeddings for Simple Link Prediction. In Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York, NY, USA, 19–24 June 2016; Volume 48, pp. 2071–2080. [Google Scholar]
  29. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D Knowledge Graph Embeddings. In Proceedings of the AAAI 2018, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  30. Shang, C.; Tang, Y.; Huang, J.; Bi, J.; He, X.; Zhou, B. End-to-End Structure-Aware Convolutional Networks for Knowledge Base Completion. In Proceedings of the AAAI 2019, Honolulu, HI, USA, 27 January–1 February 2019; pp. 3060–3067. [Google Scholar]
  31. Schlichtkrull, M.S.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In Proceedings of the ESWC 2018, Heraklion, Crete, Greece, 3–7 June 2018. [Google Scholar]
  32. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P.P. Composition-based Multi-Relational Graph Convolutional Networks. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  33. Leblay, J.; Chekol, M.W. Deriving validity time in knowledge graph. In Proceedings of the Web Conference 2018, Lyon, France, 23–27 April 2018; pp. 1771–1776. [Google Scholar]
  34. Dasgupta, S.S.; Ray, S.N.; Talukdar, P. Hyte: Hyperplane-based temporally aware knowledge graph embedding. In Proceedings of the EMNLP 2018, Brussels, Belgium, 31 October–2 November 2018; pp. 2001–2011. [Google Scholar]
  35. García-Durán, A.; Dumancic, S.; Niepert, M. Learning Sequence Encoders for Temporal Knowledge Graph Completion. In Proceedings of the EMNLP 2018, Brussels, Belgium, 31 October–2 November 2018. [Google Scholar]
  36. Lacroix, T.; Obozinski, G.; Usunier, N. Tensor Decompositions for Temporal Knowledge Base Completion. In Proceedings of the ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  37. Xu, C.; Nayyeri, M.; Alkhoury, F.; Lehmann, J.; Yazdi, H.S. Temporal Knowledge Graph Embedding Model based on Additive Time Series Decomposition. arXiv 2019, arXiv:1911.07893. [Google Scholar]
  38. Jain, P.; Rathi, S.; Mausam; Chakrabarti, S. Temporal Knowledge Base Completion: New Algorithms and Evaluation Protocols. In Proceedings of the EMNLP 2020, Online, 16–20 November 2020; 20 November 2020. [Google Scholar] [CrossRef]
  39. Jung, J.; Jung, J.; Kang, U. Learning to Walk across Time for Interpretable Temporal Knowledge Graph Completion. In Proceedings of the KDD ’21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, Singapore, 14–18 August 2021; pp. 786–795. [Google Scholar] [CrossRef]
  40. Zhu, C.; Chen, M.; Fan, C.; Cheng, G.; Zhang, Y. Learning from History: Modeling Temporal Knowledge Graphs with Sequential Copy-Generation Networks. In Proceedings of the AAAI 2021, Virtual, 2–9 February 2021. [Google Scholar]
  41. Han, Z.; Ding, Z.; Ma, Y.; Gu, Y.; Tresp, V. Learning Neural Ordinary Equations for Forecasting Future Links on Temporal Knowledge Graphs. In Proceedings of the EMNLP 2021, Punta Cana, Dominican Republic, 16–20 November 2021. [Google Scholar] [CrossRef]
  42. Liu, Y.; Ma, Y.; Hildebrandt, M.; Joblin, M.; Tresp, V. TLogic: Temporal Logical Rules for Explainable Link Forecasting on Temporal Knowledge Graphs. In Proceedings of the AAAI 2022, Virtual, 22 February–1 March 2022. [Google Scholar]
  43. Han, Z.; Chen, P.; Ma, Y.; Tresp, V. Explainable Subgraph Reasoning for Forecasting on Temporal Knowledge Graphs. In Proceedings of the ICLR 2021, Vienna, Austria, 4 May 2021. [Google Scholar]
  44. Lautenschlager, J.; Shellman, S.; Ward, M. ICEWS Events and Aggregations. Harv. Dataverse 2015, 3. [Google Scholar]
  45. Leetaru, K.; Schrodt, P.A. Gdelt: Global data on events, location, and tone, 1979–2012. In Proceedings of the ISA Annual Convention, Citeseer, San Francisco, CA, USA, 2–4 April 2013; pp. 1–49. [Google Scholar]
  46. Goel, R.; Kazemi, S.M.; Brubaker, M.A.; Poupart, P. Diachronic Embedding for Temporal Knowledge Graph Completion. In Proceedings of the AAAI 2020, New York, NY, USA, 2–7 February 2020. [Google Scholar] [CrossRef]
Figure 1. An illustration of Householder transformation.
Figure 1. An illustration of Householder transformation.
Symmetry 16 01166 g001
Figure 2. An overview of our HTKGE, for which the decoder details can be found in Figure 3.
Figure 2. An overview of our HTKGE, for which the decoder details can be found in Figure 3.
Symmetry 16 01166 g002
Figure 3. An illustration of our Householder-transformation-based adaptive decoder.
Figure 3. An illustration of our Householder-transformation-based adaptive decoder.
Symmetry 16 01166 g003
Figure 4. Ablation studies on ICEWS14* and ICEWS05-15.
Figure 4. Ablation studies on ICEWS14* and ICEWS05-15.
Symmetry 16 01166 g004
Figure 5. H@10 performance comparison of different decoder parameters on ICEWS14*.
Figure 5. H@10 performance comparison of different decoder parameters on ICEWS14*.
Symmetry 16 01166 g005
Table 1. Statistics of the datasets.
Table 1. Statistics of the datasets.
DatasetGDELTICEWS14*WIKIICEWS05-15
Entities7691712812,55410,094
Relations24023024251
Train1,734,39963,685539,286368,868
Validation238,76513,82367,53846,302
Test305,24113,22263,11046,159
Time granularity15 min24 h1 year24 h
Timestamps29753652324017
Table 2. Results on link prediction task for four experimented datasets, where the results of RE-NET are from HiSMatch. The best score is in bold.
Table 2. Results on link prediction task for four experimented datasets, where the results of RE-NET are from HiSMatch. The best score is in bold.
DatasetGDELTICEWS14*
Metric MRR H@1 H@3 H@10 MRR H@1 H@3 H@10
RE-NET [11]19.612.421.034.038.328.741.354.5
TANGO [41]19.212.220.432.826.317.329.144.2
CyGNet [40]18.511.519.632.032.723.736.350.7
TITer [12]20.214.122.231.241.732.746.558.4
RE-GCN [14]19.812.521.034.041.831.646.761.5
xERTE [43]18.912.320.130.340.832.745.757.3
TiRGN [19]21.713.623.337.645.134.451.365.0
HiSMatch [18]22.014.523.836.645.835.850.865.1
TLogic [42]19.812.221.735.643.033.648.361.2
DaeMon [20]20.713.722.534.2----
RPC [21]22.414.424.438.344.634.949.865.1
HTKGE (ours)23.916.226.138.746.136.251.566.4
± 0.022 ± 0.016 ± 0.029 ± 0.025 ± 0.033 ± 0.028 ± 0.035 ± 0.042
DatasetWIKIICEWS05-15
MetricMRRH@1H@3H@10MRRH@1H@3H@10
RE-NET [11]49.746.951.243.543.333.447.863.1
TANGO [41]50.448.551.553.642.932.748.162.3
CyGNet [40]58.847.966.478.736.826.641.656.2
TITer [12]73.971.775.477.047.738.052.965.8
RE-GCN [14]78.574.581.684.748.037.353.968.3
xERTE [43]71.168.176.179.046.637.852.363.9
TiRGN [19]81.777.885.187.150.039.356.170.7
HiSMatch [18]78.173.981.3284.752.942.059.173.3
TLogic [42]79.775.481.985.247.036.253.167.4
DaeMon [20]82.478.386.088.0----
RPC [21]81.276.385.488.751.439.957.071.8
HTKGE (ours)82.778.586.889.152.441.358.472.9
± 0.049 ± 0.036 ± 0.043 ± 0.046 ± 0.035 ± 0.027 ± 0.030 ± 0.034
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, F.; Zhang, Y.; Zhao, X.; Pang, S. Unleashing the Power of Decoders: Temporal Knowledge Graph Extrapolation with Householder Transformation. Symmetry 2024, 16, 1166. https://doi.org/10.3390/sym16091166

AMA Style

Yang F, Zhang Y, Zhao X, Pang S. Unleashing the Power of Decoders: Temporal Knowledge Graph Extrapolation with Householder Transformation. Symmetry. 2024; 16(9):1166. https://doi.org/10.3390/sym16091166

Chicago/Turabian Style

Yang, Fuqiang, Yue Zhang, Xuechen Zhao, and Shengnan Pang. 2024. "Unleashing the Power of Decoders: Temporal Knowledge Graph Extrapolation with Householder Transformation" Symmetry 16, no. 9: 1166. https://doi.org/10.3390/sym16091166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop