temporal knowledge graph representation learning with local and global evolutions

  • Journal Profile
  • Editor-in-Chief
  • Editorial Board
  • Ethics Guidelines
  • Open Access
  • Subscription

Computer Science ›› 2022 , Vol. 49 ›› Issue (9) : 162-171. doi: 10.11896/jsjkx.220500204

• Artificial Intelligence • Previous Articles     Next Articles

Temporal Knowledge Graph Representation Learning

XU Yong-xin 1,2 , ZHAO Jun-feng 1,2,3 , WANG Ya-sha 1,2,3 , XIE Bing 1,2,3 , YANG Kai 1,2,3   

  • 1 School of Computer Science,Peking University,Beijing 100871,China 2 Key Laboratory of High Confidence Software Technologies,Ministry of Education,Beijing 100871,China 3 Peking University Information Technology Institute(Tianjin Binhai),Tianjin 300450,China
  • Received: 2021-10-22 Revised: 2022-05-16 Online: 2022-09-15 Published: 2022-09-09
  • About author: XU Yong-xin,born in 1998,postgraduate.His main research interests include knowledge graph and so on. ZHAO Jun-feng,born in 1974,Ph.D,research professor,is a member of China Computer Federation.Her main research interests include big data analysis,knowledge graph,urban computing and so on.
  • Supported by: National Natural Science Foundation of China(62172011).
  • 1. 探讨2016版国际胰瘘研究小组定义和分级系统对胰腺术后患者胰瘘分级的影响.PDF (500KB)

Abstract: As a structured form of human knowledge,knowledge graphs have played a great supportive role in supporting the semantic intercommunication of massive,multi-source,heterogeneous data,and effectively support tasks such as data analysis,attracting the attention of academia and industry.At present,most knowledge graphs are constructed based on non-real-time static data,without considering the temporal characteristics of entities and relationships.However,data in application scenarios such as social network communication,financial trade,and epidemic spreading network are highly dynamic and exhibit complex temporal properties.How to use time series data to build knowledge graphs and effectively model them is a challenging problem.Recently,numerous studies use temporal information in time series data to enrich the characteristics of knowledge graphs,endowing know-ledge graphs with dynamic features,expanding fact triples into quadruple representation(head entity,relationship,tail entity,time).The knowledge graph which utilizes time-related quadruples to represent knowledge are collectively referred to as temporal knowledge graph.This paper summarizes the research work of temporal knowledge graph representation learning by sorting out and analyzing the corresponding literature.Specifically,it first briefly introduce the background and definition of temporal know-ledge graph.Next,it summarizes the advantages of the temporal knowledge graph representation learning method compared with the traditional knowledge graph representation learning method.Then it elaborates on the recent method of temporal knowledge graph representation learning from the perspective of the method modeling facts,introduces the dataset used by the above method and summarizes the main challenges of this technology.Finally,the future research direction is prospected.

Key words: Knowledge graph, Deep learning, Representation learning, Temporal information, Dynamic process

CLC Number: 

Cite this article

XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning[J].Computer Science, 2022, 49(9): 162-171.

share this article

Add to citation manager EndNote | Reference Manager | ProCite | BibTeX | RefWorks

URL: https://www.jsjkx.com/EN/10.11896/jsjkx.220500204

          https://www.jsjkx.com/EN/Y2022/V49/I9/162

Related Articles 15

Recommended 0.

  • Related Articles
  • Recommended

A Survey on Temporal Knowledge Graph: Representation Learning and Applications

Knowledge graphs have garnered significant research attention and are widely used to enhance downstream applications. However, most current studies mainly focus on static knowledge graphs, whose facts do not change with time, and disregard their dynamic evolution over time. As a result, temporal knowledge graphs have attracted more attention because a large amount of structured knowledge exists only within a specific period. Knowledge graph representation learning aims to learn low-dimensional vector embeddings for entities and relations in a knowledge graph. The representation learning of temporal knowledge graphs incorporates time information into the standard knowledge graph framework and can model the dynamics of entities and relations over time. In this paper, we conduct a comprehensive survey of temporal knowledge graph representation learning and its applications. We begin with an introduction to the definitions, datasets, and evaluation metrics for temporal knowledge graph representation learning. Next, we propose a taxonomy based on the core technologies of temporal knowledge graph representation learning methods, and provide an in-depth analysis of different methods in each category. Finally, we present various downstream applications related to the temporal knowledge graphs. In the end, we conclude the paper and have an outlook on the future research directions in this area.

[a]organization=School of Computer Science and Technology, East China Nomal University,addressline=3663 North Zhongshan Road, city=Shanghai, postcode=200062, country=China

[b]organization= College of Computer Science and Technology, Guizhou University,addressline=2708 South Huaxi Avenue, city=Guiyang, postcode=550025, state=Guizhou, country=China

[c]organization= Department of Industrial Engineering, Tsinghua University,addressline=30 Shuangqing Road, city=Beijing, postcode=100084, country=China

1 Introduction

Knowledge graphs (KGs) describe the real world with structured facts. A fact consists of two entities and a relation connecting them, which can be formally represented as a triple (head, relation, tail) , and an instance of a fact is (Barack Obama, make statement, Iran) . Knowledge graph representation learning (KGRL)   [ 33 ] seeks to learn the low-dimentional vector embeddings of entities and relations and use these embeddings for downstream tasks such as information retrieval  [ 16 ] , question answering  [ 29 ] , and recommender systems  [ 2 ] .

Existing KGs ignore the timestamp indicating when a fact occurred and cannot reflect their dynamic evolution over time. In order to represent KGs more accurately, Wikidata  [ 73 ] and YOGO2  [ 32 ] add temporal information to the facts, and some event knowledge graphs  [ 55 , 42 ] also contain the timestamps indicating when the events occurred. The knowledge graphs with temporal information are called temporal knowledge graphs (TKGs). Figure  1 is a subgraph of the temporal knowledge graph. The fact in TKGs are expanded into quadruple (head, relation, tail, timestamp) , a specific instance is (Barack Obama, make statement, Iran, 2014-6-19) .

Refer to caption

The emergence of TKGs has led to increased researcher interest in temporal knowledge graph representation learning (TKGRL)  [ 53 ] . The acquired low-dimensional vector representations are capable of modelling the dynamics of entities and relations over time, thereby improving downstream applications such as time-aware knowledge reasoning  [ 34 ] , entity alignment  [ 81 ] , and question answering  [ 59 ] .

Temporal knowledge graph representation learning and applications are at the forefront of current research. Nevertheless, as of now, a comprehensive survey on the topic is not yet available. Ji et al.  [ 33 ] provides a survey on KGs, which includes a section on TKGs. However, this section only covers a limited number of early methods related to TKGRL. The paper  [ 53 ] is a survey on TKGs, with one section dedicated to introducing representation learning methods of static knowledge graphs, and only five models related to TKGRL are elaborated in detail. The survey  [ 6 ] is about temporal knowledge graph completion (TKGC), and it focuses solely on the interpolation-based temporal knowledge graph reasoning application.

This paper comprehensively summarizes the current research of TKGRL and its related applications. Our main contributions are summarized as follows: (1) We conduct an extensive investigation on various TKGRL methods up to the present, analyze their core technologies, and propose a new classification taxonomy. (2) We divide the TKGRL methods into ten distinct categories. Within each category, we provide detailed information on the key components of different methods and analyze the strengths and weaknesses of these methods. (3) We introduce the latest development of different applications related to TKGs, including temporal knowledge graph reasoning, entity alignment between temporal knowledge graphs, and question answering over temporal knowledge graphs. (4) We summarize the existing research of TKGRL and point out the future directions which can guide further work.

The remainder of this paper is organized as follows: Chapter 2 introduces the background of temporal knowledge graphs, including definitions, datasets, and evaluation metrics. Chapter 3 summarizes various temporal knowledge graph representation learning methods, including transformation-based methods, decomposition-based methods, graph neural networks-based methods, capsule network-based methods, autoregression-based methods, temporal point process-based methods, interpretability-based methods, language model methods, few-shot learning methods and others. Chapter 4 introduces the related applications of the temporal knowledge graph, such as temporal knowledge graph reasoning, entity alignment between temporal knowledge graphs, and question answering over temporal knowledge graphs. Chapter 5 highlights the future directions of Temporal Knowledge Graph Representation Learning (TKGRL), encompassing Scalability, Interpretability, Information Fusion, and the Integration of Large Language Models. Chapter 6 gives a conclusion of this paper.

2 Background

2.1 problem formulation.

Take Figure  1 for example, where the entity set E contains (Barack Obama, Iran, Iraq) , the relation set contains (make statement, express intent to provide military aid, provide military aid, provide military protection or peacekeeping, receive deployment peacekeeper) , the time set contains (2014-6-18, 2014-6-19, 2014-6-20, 2014-6-22) , and the fact set contains ((Iran, Provide military protection or peacekeeping, Iraq, 2014-6-18)), (Iraq, receive deployment peacekeeper, Iran, 2014-6-18), (Barack Obama, Make statement, Iran, 2014-6-19), (Iran, Make statement, Barack Obama, 2014-6-20),(Barack Obama, express intent to provide military aid, Iraq, 2014-6-19), (Barack Obama, provide military aid, Iraq, 2014-6-22)) .

Refer to caption

2.2 Datasets

There are four commonly used datasets for temporal knowledge graph representation learning.

ICEWS The integrated crisis early warning system (ICEWS)  [ 55 ] captures and processes millions of data points from digitized news, social media, and other sources to predict, track and respond to events around the world, primarily for early warning. Three subsets are typically used: ICEWS14, ICEWS05-15, and ICEWS18, which contain events in 2014, 2005-2015, and 2018, respectively.

GDELT The global database of events, language, and tone (GDELT)  [ 42 ] is a global database of society. It includes the world’s broadcast, print, and web news from across every country in over 100 languages and continually updates every 15 minutes.

Wikidata The Wikidata  [ 73 ] is a collaborative, multilingual auxiliary knowledge base hosted by the Wikimedia Foundation to support resource sharing and other Wikimedia projects. It is a free and open knowledge base that can be read and edited by both humans and machines. Many items in Wikidata have temporal information.

YAGO The YAGO  [ 32 ] is a linked database developed by the Max Planck Institute in Germany. YAGO integrates data from Wikipedia, WordNet, and GeoNames. YAGO integrates WordNet’s word definitions with Wikipedia’s classification system, adding temporal and spatial information to many knowledge items.

The datasets of TKGs often require unique data processing methods for different downstream applications. Table   1 presents the statistics of datasets for various tasks of TKGs. In knowledge reasoning tasks, datasets are typically divided into different training sets, validation sets, and test sets based on task type (interpolation and extrapolation). In entity alignment tasks, as the same entites in the real world need to be aligned between different KGs, a dataset always includes two temporal knowledge graphs that must be learned simultaneously. For question answering tasks, the datasets not only include temporal knowledge graphs used to search for answers but also include temporal-related questions (which have not been showed here).

2.3 Evaluation Metrics

The evaluation metrics for verifying the performance of TKGRL are M ​ R ​ R 𝑀 𝑅 𝑅 MRR (mean reciprocal rank) and H ​ i ​ t ​ @ ​ k 𝐻 𝑖 𝑡 @ 𝑘 Hit@k .

𝑴 ​ 𝑹 ​ 𝑹 𝑴 𝑹 𝑹 \boldsymbol{MRR} The M ​ R ​ R 𝑀 𝑅 𝑅 MRR represents the average of the reciprocal ranks of the correct answers. It can be calculated as follows:

where S 𝑆 S is the set of all correct answers, | S | 𝑆 |S| is the number of the sets. The predicted result is a set sorted by the probability of the answer from high to low, and r ​ a ​ n ​ k i 𝑟 𝑎 𝑛 subscript 𝑘 𝑖 rank_{i} is the rank of the i-th correct answer in the prediction result. The higher the MRR, the better the performance.

𝑯 ​ 𝒊 ​ 𝒕 ​ @ ​ 𝒌 𝑯 𝒊 𝒕 bold-@ 𝒌 \boldsymbol{Hit@k} The H ​ i ​ t ​ s ​ @ ​ k 𝐻 𝑖 𝑡 𝑠 @ 𝑘 Hits@k reports the proportion of correct answers in the top k 𝑘 k predict results. It can be calculated by the following equation:

Refer to caption

3 Temporal Knowledge Graph Representation Learning Methods

Compared to KGs, TKGs contain additional timestamps, which are taken into account in the construction of TKGRL methods. These methods can be broadly categorized into transformation-based, decomposition-based, graph neural networks-based, and capsule network-based approaches. Additionally, temporal knowledge graphs can be viewed as sequences of snapshots captured at different timestamps or events that occur over continuous time, and can be learned using autoregressive and temporal point process techniques. Moreover, some methods prioritize interpretability, language model, and few-shot learning. Thus, based on the core technologies employed by TKGRL methods, we group them into the following categories: transformation-based methods, decomposition-based methods, graph neural networks-based methods, capsule network-based methods, autoregression-based methods, temporal point process-based methods, interpretability-based methods, language model methods, few-shot learning methods, and others. A visualization of the categorization of TKGRL methods is presented in Figure  3 .

The notations in these methods are varied, and we define our notations to describe them uniformly. We use lower-case letters to denote scalars, bold lower-case letters to denote vectors, bold upper-case letters to denote matrices, bold calligraphy upper-case letters to denote order 3 tensors, and bold script upper-case letters to denote order 4 tensors. The main notations and their descriptions are listed in table  2 .

3.1 Transformation-based Methods

In the transformation-based method, timestamps or relations are regarded as the transformation between entities. The representation learning of TKGs is carried out by integrating temporal information or mapping entities and relations to temporal hyperplanes based on the existing KGRL methods. There are translation-based transformations and rotation-based transformations.

𝒉 𝒓 𝒕 1 2 \|\boldsymbol{h}+\boldsymbol{r}-\boldsymbol{t}\|_{1/2} .

𝒉 𝒓 𝝉 𝒕 1 2 \|\boldsymbol{h}+\boldsymbol{r}+\boldsymbol{\tau}-\boldsymbol{t}\|_{1/2} .

𝒉 subscript 𝒓 𝑠 𝑒 𝑞 𝒕 2 \|\boldsymbol{h}+\boldsymbol{r}_{seq}-\boldsymbol{t}\|_{2} .

subscript 𝒉 𝜏 subscript 𝒓 𝜏 subscript 𝒕 𝜏 1 2 \|\boldsymbol{h}_{\tau}+\boldsymbol{r}_{\tau}-\boldsymbol{t}_{\tau}\|_{1/2} .

subscript 𝒉 𝜏 𝒓 subscript ¯ 𝒕 𝜏 1 \|\boldsymbol{h}_{\tau}+\boldsymbol{r}-\overline{\boldsymbol{t}}_{\tau}\|_{1} .

subscript 𝒉 𝜏 𝒓 subscript ¯ 𝒕 𝜏 2 \|\boldsymbol{h}_{\tau}+\boldsymbol{r}-\overline{\boldsymbol{t}}_{\tau}\|_{2} .

3.2 Decomposition-based Methods

The main task of representation learning is to learn the low-dimensional vector representation of the knowledge graph. Tensor decomposition has three applications: dimension reduction, missing data completion, and implicit relation mining, which meet the needs of knowledge graph representation learning. The knowledge graph consists of triples and can be represented by an order 3 tensor. For the temporal knowledge graph, the additional temporal information can be represented by an order 4 tensor, and each tensor dimension is the head entity, relation, tail entity, and timestamp, respectively. Tensor decomposition includes Canonical Polyadic (CP) decomposition  [ 30 ] and Tucker decomposition  [ 69 ] .

3.3 Graph Neural Networks-based Methods

Graph Neural Networks (GNN)  [ 60 ] have powerful structure modeling ability. The entity can enrich its representation with the attribute feature and the global structure feature by GNN. Typical graph neural networks include Graph Convolutional Networks (GCN)  [ 38 ] and Graph Attention Networks (GAT)  [ 72 ] . GCN gets the representation of nodes by aggregating neighbor embeddings, and GAT uses a multi-head attention mechanism to get the representation of nodes by aggregating weighted neighbor embedding. The knowledge graph is a kind of graph that has different relations. The relation-aware graph neural networks are developed to learn the representations of entities in the knowledge graph. Relational Graph Convolutional Networks (R-GCN)  [ 61 ] is a graph neural network model for relational data. It learns the representation for each relation and obtains entity representation by aggregating neighborhood information under different relation representations. Temporal knowledge graphs have additional temporal information, and some methods enhance the representation of entities by a time-aware mechanism.

TEA-GNN  [ 81 ] learns entity representations through a time-aware graph attention network, which incorporates relational and temporal information into the GNN structure. Specifically, it assigns different weights to different entities with orthogonal transformation matrices computed from the neighborhood’s relational embeddings and temporal embeddings and obtains the entity representations by aggregating the neighborhood.

TREA  [ 82 ] learns more expressive entity representation through a temporal relational graph attention mechanism. It first maps entities, relations, and timestamps into an embedding space, then integrates entities’ relational and temporal features through a temporal relational graph attention mechanism from their neighborhood, and finally, uses a margin-based log-loss to train the model and obtains the optimized representations.

DEGAT  [ 74 ] proposes a dynamic embedding graph attention network. It first uses the GAT to learn the static representations of entities by aggregating the features of neighbor nodes and relations, then adopts a diachronic embedding function to learn the dynamic representations of entities, and finally concatenates the two representations and uses the ConvKB as the decoder to obtain the score.

T 2 superscript 𝑇 2 T^{2} TKG [ 84 ] , or Latent relations Learning method for Temporal Knowledge Graph reasoning, is a novel approach that addresses the limitations of existing methods in explicitly capturing intra-time and inter-time latent relations for accurate prediction of future facts in Temporal Knowledge Graphs. It first employs a Structural Encoder (SE) to capture representations of entities at each timestamp, encoding their structural information. Then, it introduces a Latent Relations Learning (LRL) module to mine and exploit latent relations both within the same timestamp (intra-time) and across different timestamps (inter-time). Finally, the method fuses the temporal representations obtained from SE and LRL to enhance entity prediction tasks.

3.4 Capsule Network-based Methods

CapsNet is first proposed for computer vision tasks to solve the problem that CNN needs lots of training data and cannot recognize the spatial transformation of targets. The capsule network is composed of multiple capsules, and one capsule is a group of neurons. The capsules in the lowest layer are called primary capsules, usually implemented by convolution layers to detect the presence and pose of a particular pattern (such as eyes, nose, or mouth). The capsules in the higher level are called routing capsules, which are used to detect more complex patterns (such as faces). The output of a capsule is a vector whose length represents the probability that the pattern is present and whose orientation represents the pose of the pattern.

CapsE  [ 54 ] explores the application of capsule network in the knowledge graph. It represents a triplet as a three-column matrix in which each column represents the embedding of the head entity, relation, and tail entity, respectively. The matrix was fed to the capsule network, which first maps the different features of the triplet by a CNN layer, then captures the various patterns by the primary capsule layer, and finally routes the patterns to the next capsule layer to obtain the continuous output vector whose length indicates whether the triplet is valid.

TempCaps  [ 18 ] incorporates temporal information and proposes a capsule network for temporal knowledge graph completion. The model first selects the neighbors of the head entity in a time window and obtains the embeddings of these neighbors with the capsules, then adopts a dynamic routing process to connect the lower capsules and higher capsules and gets the head entity embedding, and finally uses a multi-layer perceptron (MLP)  [ 20 ] as the decoder to produce the scores of all candidate entities.

subscript 𝒂 1 subscript 𝒃 1 𝒊 subscript 𝒄 1 𝒋 subscript 𝒅 1 𝒌 subscript 𝒂 0 subscript 𝒃 0 subscript 𝒄 0 subscript 𝒅 0 subscript 𝒂 1 subscript 𝒃 1 subscript 𝒄 1 subscript 𝒅 1 superscript ℝ 𝑑 \mathbb{H}^{d},\ \boldsymbol{q=q_{0}+q_{1}\xi},\ \boldsymbol{q_{0}=a_{0}+}\\ \boldsymbol{b_{0}i+c_{0}j+d_{0}k},\ \boldsymbol{q_{1}=a_{1}+b_{1}i+c_{1}j+d_{1}k},\ \boldsymbol{a_{0},\,b_{0},\,c_{0},\,d_{0},\,a_{1},\,b_{1},\,c_{1},}\\ \boldsymbol{d_{1}}\in\mathbb{R}^{d} ). Dual quaternions enable modeling of both rotation and translation operations simultaneously. The model first transforms the head entity through relations and timestamps in the dual quaternion space, where the representation is close to that of the tail entity. The scoring function is ‖ 𝒉 ⊙ 𝒓 ⊙ 𝝉 − 𝒕 ‖ norm direct-product 𝒉 𝒓 𝝉 𝒕 \|\boldsymbol{h}\odot\boldsymbol{r}\odot\boldsymbol{\tau}-\boldsymbol{t}\| . The learned representations are then inputted into the capsule network to obtain the final representation.

3.5 Autoregression-based Methods

𝑇 1 G_{T+1} in the future.

RE-NET  [ 34 ] proposes a recurrent event network to model the TKG for predicting future facts. It believes that the facts in G T subscript 𝐺 𝑇 G_{T} at timestamp T 𝑇 T depend on the facts in the past m subgraphs G T − 1 : T − m subscript 𝐺 : 𝑇 1 𝑇 𝑚 G_{T-1:T-m} before T 𝑇 T . It first uses the R-GCN to learn the global structural representations and local neighborhood representations of the head entity at each timestamp. Then it utilizes the gated recurrent units (GRU)  [ 11 ] to update the above representations and pass these representations to an MLP as a decoder to infer the facts at timestamp T 𝑇 T . This method only models the representations of the specific entity and relation in the query triple, ignoring the structural dependency between all triples in each subgraph, which may lose some important information from the entities not in the query triple.

Glean  [ 14 ] thinks that most of the existing representation learning methods use the structural information of TKG, ignoring the unstructured information such as semantic information of words. It proposes a temporal graph neural network with heterogeneous data fusion. Specifically, it first constructs temporal event graphs based on historical facts at each timestamp and temporal word graphs from event summaries at each timestamp. Then it uses the CompGCN  [ 70 ] to learn the structural representations of entities and relations in the temporal event graphs and the GCN to learn the textual semantic representations of entities and relations in the temporal word graphs. Finally, it fuses the two representations and utilizes a recurrent encoder to model temporal features for final prediction.

RE-GCN  [ 46 ] splits the TKG into a sequence of KG according to the timestamps and encodes the facts in the past m steps recurrently to predict the entities and relations in the future. This model proposes a recurrent evolution network based on a graph convolution network to model the evolutional representations by incorporating the structural dependencies among concurrent facts, the sequential patterns across temporally adjacent facts, and the static properties. It uses the relation-aware GCN to capture the structural dependency and utilizes GRU to obtain the sequential pattern. Then it combines the static properties learned by R-GCN as a constraint to learn the evolutional representations of entities and relations and adopts ConvTransE  [ 62 ] as the decoder to predict the probability of entities and relations at next timestamp.

TiRGN  [ 43 ] argues that the above methods can only capture the local historical dependence of the adjacent timestamps and cannot fully learn the historical characteristics of the facts. It proposes a time-guided recurrent graph neural network with local-global historical patterns which can model the historical dependency of events at adjacent snapshots with a local recurrent encoder, the same as RE-GCN, and collect repeated historical facts by a global history encoder. The final representations are fed into a time-guided decoder named Time-ConvTransE/Time-ConvTransR to predict the entities and relations in the future.

Cen  [ 44 ] believes that modeling historical facts with fixed time steps could not discover the complex evolutional patterns that vary in length. It proposes a complex evolutional network that use the evolution unit in RE-GCN as a sequence encoder to learn the representations of entities in each subgraph and utilizes the CNN as the decoder to obtain the feature maps of historical snapshots with different length. The curriculum learning strategy is used to learn the complex evolution pattern with different lengths of historical facts from short to long and automatically select the optimal maximum length to promote the prediction.

3.6 Temporal Point Process-based Methods

The autoregression-based methods sample the TKG into discrete snapshots according to a fixed time interval, which cannot effectively model the facts with irregular time intervals. Temporal point process (TPP)  [ 12 ] is a stochastic process composed of a series of events in a continuous time domain. The representation learning methods based on TPP regard the TKG as a list of events changing continuously with time and formalize it as ( G , O ) 𝐺 𝑂 (G,O) , where G 𝐺 G is the initialized TKG at time τ 0 subscript 𝜏 0 \tau_{0} , O 𝑂 O is a series of observed events ( h , r , t , τ ) ℎ 𝑟 𝑡 𝜏 (h,r,t,\tau) . At any time τ > τ 0 𝜏 subscript 𝜏 0 \tau>\tau_{0} , the TKG can be updated by the events before time τ 𝜏 \tau . The TPP can be characterized by conditional intensity function λ ​ ( τ ) 𝜆 𝜏 \lambda(\tau) . Given the historical events before a timestamp, if we can find a conditional intensity function to characterize them, then we can Predict whether the events will occur in the future with a conditional density and when the events will occur with an expectation.

Know-Evolve  [ 67 ] combines the TPP and the deep neural network framework to model the occurrence of facts as a multi-dimensional TPP. It characterizes the TPP with the Rayleigh process and uses neural networks to simulate the intensity function. RNN is used to learn the dynamic representation of the entities, and bilinear relationship score is used to capture multiple relational interactions between entities to modulate the intensity function. Thus, it can predict whether and when the event will occur.

GHNN  [ 27 ] believes that the Hawkes process  [ 28 ] based on the neural network can effectively capture the influence of past facts on future facts, and proposes a graph Hawkes neural network (GHNN). Firstly, it solves the problem that Know-Evolve could not deal with co-occurrence facts and uses a neighborhood aggregation module to process multiple facts of entities co-occurring. Then, it utilizes the continuous-time LSTM (cLSTM) model  [ 51 ] to simulate the Hawkes process to capture the evolving dynamics of the facts to implement link prediction and time prediction.

EvoKG  [ 57 ] argues that the above methods based on TPP lack to model the evolving network structure, and the methods based on autoregression lack to model the event time. It proposes a model jointly modeling the evolving network structure and event time. First, it uses an extended R-GCN and RNN to learn the time-evolving structural representations of entities and relations and utilizes an MLP with softmax to model the conditional probability of event triple. Then, it uses the same framework to learn the time-evolving temporal representations and adopts the TPP based on conditional density estimation with a mixture of log-normal distributions to model the event time. Finally, it jointly trains the two tasks and predicts the event and time in the future.

3.7 Interpretability-based Methods

The aforementioned methods has resulted in a lack of interpretability and transparency in the generated results. As a result, we categorize interpretability-based methods as a separate category to underscore the crucial role of interpretability in developing reliable and transparent models. These methods aim to provide explanations for the predictions made by the models. Two popular types of such methods are subgraph reasoning-based and reinforcement learning-based approaches.

Subgraph Reasoning xERTE  [ 23 ] is a subgraph reasoning-based method that proposes an explainable reasoning framework for predicting facts in the future. It starts from the head entity in the query and utilizes a temporal relational graph attention mechanism to learn the entity representation and relation representation. Then it samples the edges and temporal neighbors iteratively to construct the subgraph after several rounds of expansion and pruning. Finally, it predicts the tail entity in the subgraph.

Reinforcement Learning Reinforcement learning (RL)  [ 35 ] is usually modeled as a Markov Decision Process (MDP)  [ 4 ] , which includes a specific environment and an agent. The agent has an initial state. After performing an action, it receives a reward from the environment and transitions to a new state. The goal of reinforcement learning is to find the policy network to obtain the maximum reward from all action strategies.

CluSTeR  [ 45 ] proposes a two-stage reasoning strategy to predict the facts in the future. First, the clue related to a given query is searched and deduced from the history based on reinforcement learning. Then, the clue at different timestamps is regarded as a subgraph related to the query, and the R-GCN and GRU are used to learn the evolving representations of entities in the subgraph. Finally, the two stages are jointly trained, and the prediction is inferred.

TITer  [ 65 ] directly uses the temporal path-based reinforcement learning model to learn the representations of the TKG and reasons for future facts. It adds temporal edges to connect each historical snapshot of the TKG. The agent starts from the head entity of the query, transitions to the new node according to the policy network, and searches for the answer node. The method designs a time-shaped reward based on Dirichlet distribution to guide the model learning.

3.8 Language Model

In the domain of TKG, the rapid development of language models has prompted researchers to explore their application for predictive tasks. The current methodologies employing language models in the TKG domain predominantly encompass two distinct approaches: In-Context Learning and Supervised Fine-Tuning.

In-Context Learning ICLTKG [ 41 ] introduces a novel TKG forecasting approach that leverages large language models (LLMs) through in-context learning (ICL) [ 5 ] to efficiently capture and utilize irregular patterns of historical facts for accurate predictions. The implementation algorithm of this paper involves a three-stage pipeline designed to harness the capabilities of large language models (LLMs) for temporal knowledge graph (TKG) forecasting. The first stage focuses on selecting relevant historical facts from the TKG based on the prediction query. These facts are then used as context for the LLM, enabling it to capture temporal patterns and relationships between entities. Then the contextual facts are transformed into a lexical prompt that represents the prediction task. Finally, the output of the LLM is decoded into a probability distribution over the entities within the TKG. Throughout this pipeline, the algorithm controls the selection of background knowledge, the prompting strategy, and the decoding process to ensure accurate and efficient TKG forecasting. By leveraging the capabilities of LLMs and harnessing the irregular patterns embedded within historical data, this approach achieves competitive performance across a diverse range of TKG benchmarks without the need for extensive supervised training or specialized architectures.

zrLLM [ 17 ] introduces a novel approach that leverages large language models (LLMs) to enhance zero-shot relational learning on temporal knowledge graphs (TKGs). It proposes a method to first use an LLM to generate enriched relation descriptions based on textual descriptions of KG relations, and then a second LLM is employed to generate relation representations, which capture semantic information. Additionally, a relation history learner is developed to capture temporal patterns in relations, further enabling better reasoning over TKGs. The zrLLM approach is shown to significantly improve the performance of TKGF models in recognizing and forecasting facts with previously unseen zero-shot relations. Importantly, zrLLM achieves this without further fine-tuning of the LLMs, demonstrating the potential of alignment between the natural language space of LLMs and the embedding space of TKGF models. Experimental results exhibit substantial gains in zero-shot relational learning on TKGs, confirming the effectiveness and adaptability of the proposed zrLLM approach.

Supervised Fine-Tuning ECOLA  [ 25 ] proposes a joint learning model by leveraging text knowledge to enhance temporal knowledge graph representations. As existing TKGs often lack fact description information, the authors construct three new datasets that contain such information. During model training, they jointly optimize the knowledge-text prediction (KTP) objective and the temporal knowledge embedding (tKE) objective to improve the representation of TKGs. KTP employs pre-trained language models such as transformers  [ 71 ] , while tKE can utilize an existing TKGRL model such as DyERNIE [ 26 ] . By augmenting the temporal knowledge graph representation with text descriptions, the model achieves significant performance gains.

Frameworks such as GenTKG  [ 47 ] and Chain of History  [ 50 ] adopt retrieval augmented generation method for prediction. They utilize specific strategies to retrieve historical facts with high temporal relevance and logical coherence. Subsequently, these frameworks apply supervised fine-tuning language models to predict the future based on the retrieved historical facts. The input of the language model comprises historical facts and prediction queries, with the model outputting forecasted results. The authors have constructed a bespoke dataset of instructional data, which is utilized to train the language model, resulting in exemplary performance.

3.9 Few-shot Learning Methods

Few-shot learning (FSL)  [ 76 ] is a type of machine learning problems that deals with the problem of learning new concepts or tasks from only a few examples. FSL has applications in a variety of domains, including computer vision  [ 36 ] , natural language processing  [ 77 ] , and robotics  [ 49 ] , where data may be scarce or expensive to acquire. In TKGs, some entities and relations are only exist in a limited number of facts, and new entities and relations emerge over time. The latest TKGRL models now have the ability to perform FSL  [ 8 ] , which is essential for better representing these limited data. Due to the differences in handling data and learning methods for few entities and few relations, we will introduce them separately.

Few Entities MetaTKG  [ 78 ] reveals that new entities emerge over time in TKGs, and appear in only a few facts. Consequently, learning their representation from limited historical information leads to poor performance. Therefore, the authors propose a temporal meta-learning framework to address this issue. Specifically, they first divide the TKG into multiple temporal meta-tasks, then employ a temporal meta-learner to learn evolving meta-knowledge across these tasks, finally, the learned meta-knowledge guides the backbone (which can be an existing TKGRL model, such as RE-GCN  [ 46 ] ) to adapt to new data.

MetaTKGR  [ 75 ] confirms that emerging entities, which exist in only a small number of facts, are insufficient to learn their representations using existing models. To address this issue, the authors propose a meta temporal knowledge graph reasonging framework. The model leverages the temporal supervision signal of future facts as feedback to dynamically adjust the sampling and aggregation neighborhood strategy, and encoder the new entity representations. The optimized parameters can be learned via a bi-level optimization, where inner optimization initializes entity-specific parameters using the global parameters and fine-tunes them on the support set, while outer optimization operates on the query set using a temporal adaptation regularizer to stabilize meta temporal reasoning over time. The learned parameters can be easily adapted to new entities.

As existing datasets themselves often contain new entities that are associated with only a few facts, it is possible to directly divide the existing dataset into tasks to construct the support set and query set without requiring the creation of a new dataset.

Few Relations TR-Match  [ 22 ] identifies that most relations in TKGs have only a few quadruples, and new relations are added over time. Existing models are inadequate in addressing the few-shot scenario and may not fully represent the evolving temporal and relational features of entities in TKGs. To address these issues, the authors propose a temporal-relational matching network. Specifically, the proposed approach incorporates a multi-scale time-relation attention encoder to adaptively capture local and global information based on time and relation to tackle the dynamic properties problem. A new matching processor is designed to address the few-shot problem by mapping the query to a few support quadruples in a relation-agnostic manner. To address the challenge posed by few relations in temporal knowledge graphs, three new datasets, namely ICEWS14-few, ICEWS05-15-few, and ICEWS18-few, are constructed based on existing TKG datasets. The proposed TR-Match framework is evaluated on these datasets, and the experimental results demonstrate its capability to achieve excellent performance in few-shot relation scenarios.

MTKGE  [ 10 ] recognizes that TKGs are subject to the emergence of unseen entities and relations over time. To address this challenge, the authors propose a meta-learning-based temporal knowledge graph extrapolation model. The proposed approach includes a Relative Position Pattern Graph (RPPG) to construct several position patterns, a Temporal Sequence Pattern Graph (TSPG) to learn different temporal sequence patterns, and a Graph Convolutional Network (GCN) module for extrapolation. This model leverages meta-learning techniques to adapt to new data and extract useful information from the existing TKG. The proposed MTKGE framework represents an important advancement in TKGRL by introducing a novel approach to knowledge graph extrapolation.

3.10 Other Methods

Other methods leverage the unique characteristics of TKGs to learn entity, relation, and timestamp representations. For instance, One approach explores the repetitive patterns of TKG and learns a more expressive representation using the copy-generation patterns. Alternatively, other methods employ various geometric and algorithmic techniques to capture the structural properties of TKGs and learn effective representations.

Copy Generation CygNet  [ 88 ] finds that many facts show repeated patterns along the timeline, which indicates that potential knowledge can be learned from historical facts. Therefore, it proposes a time-aware copy-generation model which can predict future facts concerning the known facts. It constructs a historical vocabulary with multi-hot vectors of head entity and relation in each snapshot. In copy mode, it generates an index vector with an MLP to obtain the probability of the tail entities in the historical entity vocabulary. In generation mode, it uses an MLP to get the probability of tail entities in the entire entity vocabulary. After combining the two probabilities, it predicts the tail entity.

Neural Ordinary Equation TANGO  [ 24 ] contends that existing approaches for modeling TKGs are inadequate in capturing the continuous evolution of TKGs over time due to their reliance on discrete state spaces. To address this limitation, TANGO proposes a novel model based on Neural Ordinary Differential Equations (NODEs). More specifically, the proposed model first employs a multi-relational graph convolutional network module to capture the graph structure information at each point in time. A graph transformation module is also utilized to model changes in edge connectivity between entities and their neighbors. The output of these modules is integrated to obtain dynamic representations of entities and relations. Subsequently, an ODE solver is adopted to solve these dynamic representations, thereby enabling TANGO to learn continuous-time representations of TKGs. TANGO’s novel approach based on ODEs offers a more effective and accurate method for modeling the dynamic evolution of TKGs compared to existing techniques that rely on discrete representation space.

Geometric model The transformation models represent TKGs in a Euclidean space, while Dyernie  [ 26 ] maps the TKGs to a non-Euclidean space and employs Riemannian manifold product to learn evolving entity representations. This approach can model multiple simultaneous non-Euclidean structures, such as hierarchical and cyclic structures, to more accurately capture the complex structural properties of TKGs. By leveraging these additional structural features, the Dyernie method can more effectively capture the relationships between entities in TKGs, resulting in improved performance on TKGRL tasks.

subscript 𝒃 𝒉 superscript 𝝉 𝒓 \boldsymbol{b_{h}+\tau^{r}} . Furthermore, BoxTE represents boxes using a time-induced relation head box 𝒓 𝒉 | 𝝉 = 𝒓 𝒉 − 𝝉 𝒓 superscript 𝒓 conditional 𝒉 𝝉 superscript 𝒓 𝒉 superscript 𝝉 𝒓 \boldsymbol{r^{h|\tau}}=\boldsymbol{r^{h}-\tau^{r}} and a time-induced relation tail box 𝒓 𝒕 | 𝝉 = 𝒓 𝒕 superscript 𝒓 conditional 𝒕 𝝉 superscript 𝒓 𝒕 \boldsymbol{r^{t|\tau}=\boldsymbol{r^{t}}} − 𝝉 𝒓 superscript 𝝉 𝒓 \boldsymbol{-\tau^{r}} . To determine whether a fact r ​ ( h , t | τ ) 𝑟 ℎ conditional 𝑡 𝜏 r(h,t|\tau) holds, BoxTE checks if the representations for h ℎ h and t 𝑡 t lie within their corresponding boxes. This is expressed mathematically as follows: 𝒉 𝒓 ​ ( 𝒉 , 𝒕 | 𝝉 ) ∈ 𝒓 𝒉 superscript 𝒉 𝒓 𝒉 conditional 𝒕 𝝉 superscript 𝒓 𝒉 \boldsymbol{h^{r(h,t|\tau)}}\in\boldsymbol{r^{h}} , and 𝒕 𝒓 ​ ( 𝒉 , 𝒕 | 𝝉 ) ∈ 𝒓 𝒕 superscript 𝒕 𝒓 𝒉 conditional 𝒕 𝝉 superscript 𝒓 𝒕 \boldsymbol{t^{r(h,t|\tau)}}\in\boldsymbol{r^{t}} . By substracting 𝝉 𝒓 superscript 𝝉 𝒓 \boldsymbol{\tau^{r}} from both sides, it obtains the equivalent expressions: 𝒉 𝒓 ​ ( 𝒉 , 𝒕 ) ∈ 𝒓 𝒉 | 𝝉 superscript 𝒉 𝒓 𝒉 𝒕 superscript 𝒓 conditional 𝒉 𝝉 \boldsymbol{h^{r(h,t)}}\in\boldsymbol{r^{h|\tau}} , 𝒕 𝒓 ​ ( 𝒉 , 𝒕 ) ∈ 𝒓 𝒕 | 𝝉 superscript 𝒕 𝒓 𝒉 𝒕 superscript 𝒓 conditional 𝒕 𝝉 \boldsymbol{t^{r(h,t)}}\in\boldsymbol{r^{t|\tau}} . BoxTE enables the model to capture rigid inference patterns and cross-time inference patterns, thereby making it a powerful tool for TKGRL.

TGeomE  [ 80 ] moves beyond the use of complex or hypercomplex spaces for TKGRL and instead proposes a novel geometric algebra-based embedding approach. This method utilizes multivector representations and performs fourth-order tensor factorization of TKGs while also introducing a new linear temporal regularization for time representation learning. The proposed TGeomE approach can naturally model temporal relations and enhance the performance of TKGRL models.

3.11 Summary

In this section, we divide the TKGRL methods into ten categories and introduce the core technologies of these methods in detail. Table  3 shows the summary of the methods, including the representation space, the encoder for mapping the entities and relations to the vector space, and the decoder for predicting the answer.

4 Applications of Temporal Knowledge Graph

By introducing temporal information, TKG can express the facts in the real world more accurately, improve the quality of knowledge graph representation learning, and answer temporal questions more reliably. It is helpful for the applications such as reasoning, entity alignment, and question answering.

4.1 Temporal Knowledge Graph Reasoning

TKGRL methods are widely used in temporal knowledge graph reasoning (TKGR) tasks which automatically infers new facts by learning the existing facts in the KG. TKGR usually has three subtasks: entity prediction, relation prediction, and time prediction. Entity prediction is the basic task of link prediction, which can be expressed as two queries ( ? , r , t , τ ) ? 𝑟 𝑡 𝜏 (?,r,t,\tau) and ( h , r , ? , τ ) ℎ 𝑟 ? 𝜏 (h,r,?,\tau) . Relation prediction and time prediction can be expressed as ( h , ? , t , τ ) ℎ ? 𝑡 𝜏 (h,?,t,\tau) and ( h , r , t , ? ) ℎ 𝑟 𝑡 ? (h,r,t,?) , respectively.

TKGR can be divided into two categories based on when the predictions of facts occur, namely interpolation and extrapolation. Suppose that a TKG is available from time τ 0 subscript 𝜏 0 \tau_{0} to τ T subscript 𝜏 𝑇 \tau_{T} . The primary objective of interpolation is to retrieve the missing facts at a specific point in time τ ​ ( τ 0 ≤ τ ≤ τ T ) 𝜏 subscript 𝜏 0 𝜏 subscript 𝜏 𝑇 \tau\ (\tau_{0}\leq\tau\leq\tau_{T}) . This process is also known as temporal knowledge graph completion (TKGC). On the other hand, extrapolation aims to predict the facts that will occur in the future ( τ ≥ τ T ) 𝜏 subscript 𝜏 𝑇 (\tau\geq\tau_{T}) and is referred to as temporal knowledge graph forecasting.

Several methods have been proposed for Temporal Knowledge Graph Completion (TKGC) including transformation-based, decomposition-based, graph neural networks-based, capsule Network-based, and other geometric methods. These techniques aim to address the problem of missing facts in TKGs by leveraging various mathematical models and neural networks. In contrast, predicting future facts in TKGs requires a different approach that can model the temporal evolution of the graph. Autoregression-based, temporal point process-based, and few-shot learning methods are commonly used for this task. Interpretability-based methods are used to increase the reliability of prediction results. These techniques provide evidence to support predictions, helping to establish trust and improving the overall quality of predictions made by the model. To further enhance the performance of TKGRL, semantic augmentation technology can be employed to improve the quality and quantity of semantic information of TKGs. Utilizing entity and relation names, as well as textual descriptions of fact associations, can enrich their representation and promote the development of downstream tasks of TKGs. In addition, large language models (LLMs) for natural language processing (NLP) can facilitate the acquisition of rich semantic information about entities and relations, further augmenting the performance of TKGRL models.

4.2 Entity Alignment Between Temporal Knowledge Graphs

Entity alignment (EA) aims to find equivalent entities between different KGs, which is important to promote the knowledge fusion between multi-source and multi-lingual KGs. Defining G 1 = ( E 1 , R 1 , T 1 , F 1 ) subscript 𝐺 1 subscript 𝐸 1 subscript 𝑅 1 subscript 𝑇 1 subscript 𝐹 1 G_{1}=(E_{1},R_{1},T_{1},F_{1}) and G 2 = ( E 2 , R 2 , T 2 , F 2 ) subscript 𝐺 2 subscript 𝐸 2 subscript 𝑅 2 subscript 𝑇 2 subscript 𝐹 2 G_{2}=(E_{2},R_{2},T_{2},F_{2}) to be two TKGs, S = { ( e 1 i , e 2 j ) | e 1 i ∈ E 1 , e 2 j ∈ E 2 } 𝑆 conditional-set subscript 𝑒 subscript 1 𝑖 subscript 𝑒 subscript 2 𝑗 formulae-sequence subscript 𝑒 subscript 1 𝑖 subscript 𝐸 1 subscript 𝑒 subscript 2 𝑗 subscript 𝐸 2 S=\{(e_{1_{i}},e_{2_{j}})|e_{1_{i}}\in E_{1},e_{2_{j}}\in E_{2}\} is the set of alignment seeds between G 1 subscript 𝐺 1 G_{1} and G 2 subscript 𝐺 2 G_{2} . EA seeks to find new alignment entities according to the alignment seeds S 𝑆 S . The methods of EA between TKGS mainly adopt the GNN-based model.

Currently, exploring the entity alignment (EA) between Temporal Knowledge Graphs (TKGs) is an active area of research. TEA-GNN  [ 81 ] was the first method to incorporate temporal information via a time-aware attention Graph Neural Network (GNN) to enhance EA. TREA  [ 82 ] utilizes a temporal relational attention GNN to integrate relational and temporal features of entities for improved EA performance. STEA  [ 7 ] identifies that the timestamps in many TKGs are uniform and proposes a simple GNN-based model with a temporal information matching mechanism to enhance EA. Initially, the structure and relation features of an entity are fused together to generate the entity embedding. Then, the entity embedding is updated using GNN aggregation from neighborhood. Finally, the entity embedding is obtained by concatenating the embedding of each layer of the GNN. STEA not only updates the representation of entities but also calculates time similarity by considering associated timestamps. The method combines both the similarities of entity embeddings and the similarities of entity timestamps to obtain aligned entities. Overall, STEA offers an effective way of improving entity representation in TKGs and provides a reliable solution for aligning entities over time.

4.3 Question Answering Over Temporal Knowledge Graphs

Question answering over KG (KGQA) aims to answer natural language questions based on KG. The answer to the question is usually an entity in the KG. In order to answer the question, one-hop or multi-hop reasoning is required on the KG. Question answering over TKG (TKGQA) aims to answer temporal natural language questions based on TKG, the answer to the question is entity or timestamp in the TKG, and the reasoning on TKG is more complex than it on KG.

Research on TKGQA is in progress. CRONKGQA  [ 59 ] release a new dataset named CRONQUESTIONS and propose a model combining representation of TKG and question for TKGQA. It first uses TComplEx to obtain the representation of entities and timestamps in the TKG, and utilizes BERT  [ 15 ] to obtain their representations in the question, then calculates the scores of all entities and times, and finally concatenated the score vectors to obtain the answer.

TSQA  [ 63 ] argues existing TKGQA methods haven’t explore the implicit temporal feature in TKGs and temporal questions. It proposes a time sensitive question answering model which consists of a time-aware TKG encoder and a time-sensitive question answering module. The time-aware TKG encoder uses TComplEx with time-order constraints to obtain the representations of entities and timestamps. The time-sensitive question answering module first decomposes the question into entities and a temporal expression. It uses the entities to extract the neighbor graph to reduce the search space of timestamps and answer entities. The temporal expression is fed into the BERT to learn the temporal question representations. Finally, entity and temporal question representations are combined to estimate the time and predict the entity with contrastive learning.

5 Future Directions

Despite the significant progress made in TKGRL research, there remain several important future directions. These include addressing scalability challenges, improving interpretability, incorporating information from multiple modalities, and leveraging large language models to enhance the ability of representing dynamic and evolving TKGs.

5.1 Scalability

The current datasets available for TKG are insufficient in size compared to real-world knowledge graphs. Moreover, TKGRL methods tend to prioritize improving task-specific performance and often overlook the issue of scalability. Therefore, there is a pressing need for research on effective methods of learning TKG representations that can accommodate the growing demand for data. A possible avenue for future research in this field is to investigate various strategies for enhancing the scalability of TKGRL models.

One approach for improving the scalability of TKGRL models is to employ distributed computing techniques, such as parallel processing or distributed training, to enable more efficient processing of large-scale knowledge graphs. Parallel processing involves partitioning the dataset into smaller subsets and processing each subset simultaneously. In contrast, distributed training trains the model on various machines concurrently, with the outcomes combined to enhance the overall accuracy of the model. This approach could prove especially beneficial for real-time processing of extensive knowledge graphs in applications that require quick response times.

Another approach is to use sampling techniques to reduce the size of the knowledge graph without sacrificing accuracy. For example, researchers could use clustering algorithms to identify groups of entities and events that are highly interconnected and then sample a representative subset of these groups for training the model. This approach could help to reduce the computational complexity of the model without sacrificing accuracy. Sampling techniques can also be used for negative sampling in TKGRL. Negative sampling involves selecting negative samples that are not present in the knowledge graph to balance out the positive samples during training. By employing efficient negative sampling techniques, researchers can significantly reduce the computational complexity of the TKGRL model while maintaining high accuracy levels.

Overall, addressing issues related to scalability will be critical for advancing the state-of-the-art in temporal knowledge graph research and enabling practical applications in real-world scenarios.

5.2 Interpretability

The enhancement of interpretability is a crucial research direction, as it allows for better understanding of how model outputs are derived and ensures the reliability and applicability of the model’s results. Despite the availability of interpretable methods, developing more interpretable models and techniques for temporal knowledge graphs remains a vital research direction.

One promising approach involves incorporating attention mechanisms to identify the most relevant entities and events in the knowledge graph at different points in time. This approach would allow users to understand which parts of the graph are most important for a given prediction, which could improve the interpretability of the model.

In addition, researchers could explore the use of visualization techniques to help users understand the structure and evolution of the knowledge graph over time. For example, interactive visualizations could enable users to explore the graph and understand how different entities and events are connected.

By making TKGRL more interpretable, we can gain greater insights into complex real-world phenomena, support decision-making processes, and ensure that these models are useful for practical applications.

5.3 Information Fusion

Most TKGRL methods only utilize the structural information of TKGs, with few models incorporating textual information of entities and relations. However, text data contains rich features that can be leveraged to enhance TKGs’ representation. Therefore, effectively fusing various features of TKGs, including entity feature, relation feature, time feature, structure feature and textual feature, represents a promising future research direction.

One approach to information fusion in TKGRL is to use multi-modal data sources. For example, researchers can combine textual data, such as news articles or social media posts, with structured data from knowledge graphs to improve the accuracy of the model. This approach can help the TKGRL model to capture more relationships between entities and events that may not be apparent from structured data alone.

Another approach is to use attention mechanisms to dynamically weight the importance of different sources of information at different points in time. This approach would allow the model to focus on the most relevant information for a given prediction, which could improve the accuracy of the model while reducing computational complexity.

In general, information fusion is a powerful tool in TKGRL that can help researchers improve the accuracy and reliability of the model by combining information from multiple sources. However, it is essential to carefully weigh the benefits and costs of using different fusion techniques, depending on the specific dataset and research goals.

5.4 Incorporating Large Language Models

Recent advances in natural language processing, such as the development of large language models (LLMs)   [ 87 ] has been largely advanced by both academia and industry. A notable achievement in the field of LLMs is the introduction of ChatGPT  1 1 1 https://openai.com/blog/chatgpt/ , a highly advanced AI chatbot. Developed using LLMs, ChatGPT has generated significant interest and attention from both the research community and society at large. ChatGPT uses the generative pre-trained transformer (GPT) such as GPT-4  [ 56 ] , have led to significant improvements in various natural language tasks. LLMs have been shown to be highly effective at capturing complex semantic relationships between words and phrases, and they may be able to provide valuable insights into the meaning and context of entities and relations in a knowledge graph. Efficiently combining LLMs with TKGRL is an novel research direction for the future.

One approach to incorporating LLMs into TKGRL is to use LLMs to generate embeddings for entities and relations. These embeddings could be used as input to a TKGRL model, enabling it to capture more rich feature of entities and relations over time.

Another potential approach is to use LLMs to generate textual descriptions of entities and facts in the TKGs. These descriptions could be used to enrich the TKGs with additional semantic information, which could then be used to improve the accuracy of predictions.

Aboveall, incorporating LLMs into TKGRL has the potential to significantly improve the accuracy and effectiveness of these models, and it is an exciting area for future research. However, it is essential to carefully consider the challenges and limitations of using LLMs, such as computational complexity and potential bias in the pre-trained data.

6 Conclusion

Temporal knowledge graphs (TKGs) provide a powerful framework for representing and analyzing complex real-world phenomena that evolve over time. Temporal knowledge graph representation learning (TKGRL) is an active area of research that investigates methods for automatically extracting meaningful representations from TKGs.

In this paper, we provide a detailed overview of the development, methods, and applications of TKGRL. We begin by the definition of TKGRL and discussing the datasets and evaluation metrics commonly used in this field. Next, we categorize TKGRL methods based on their core technologies and analyze the fundamental ideas and techniques employed by each category. Furthermore, we offer a comprehensive review of various applications of TKGRL, followed by a discussion of future directions for research in this area. By focusing on these areas, we can continue to drive advancements in TKGRL and enable practical applications in real-world scenarios.

Acknowledgements

We appreciate the support from National Natural Science Foundation of China with the Main Research Project on Machine Behavior and Human-Machine Collaborated Decision Making Methodology (72192820 & 72192824), Pudong New Area Science & Technology Development Fund (PKX2021-R05), Science and Technology Commission of Shanghai Municipality (22DZ-2229004), and Shanghai Trusted Industry Internet Software Collaborative Innovation Center.

  • Abboud et al. [2020] Abboud, R., Ceylan, İ.İ., Lukasiewicz, T., Salvatori, T., 2020. Boxe: A box embedding model for knowledge base completion, in: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (Eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual. URL: https://proceedings.neurips.cc/paper/2020/hash/6dbbe6abe5f14af882ff977fc3f35501-Abstract.html .
  • Anelli et al. [2021] Anelli, V.W., Di Noia, T., Di Sciascio, E., Ferrara, A., Mancino, A.C.M., 2021. Sparse feature factorization for recommender systems with knowledge graphs, in: Proceedings of the 15th ACM Conference on Recommender Systems, Association for Computing Machinery, New York, NY, USA. p. 154–165. URL: https://doi.org/10.1145/3460231.3474243 , doi: 10.1145/3460231.3474243 .
  • Balazevic et al. [2019] Balazevic, I., Allen, C., Hospedales, T., 2019. TuckER: Tensor factorization for knowledge graph completion, in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 5185–5194. URL: https://aclanthology.org/D19-1522 , doi: 10.18653/v1/D19-1522 .
  • Bellman [1957] Bellman, R., 1957. A markovian decision process. Indiana Univ. Math. J. 6, 679–684. URL: http://www.iumj.indiana.edu/IUMJ/FULLTEXT/1957/6/56038 .
  • Brown et al. [2020] Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al., 2020. Language models are few-shot learners. Advances in neural information processing systems 33, 1877–1901.
  • Cai et al. [2022a] Cai, B., Xiang, Y., Gao, L., Zhang, H., Li, Y., Li, J., 2022a. Temporal knowledge graph completion: A survey. CoRR abs/2201.08236. URL: https://arxiv.org/abs/2201.08236 , arXiv:2201.08236 .
  • Cai et al. [2022b] Cai, L., Mao, X., Ma, M., Yuan, H., Zhu, J., Lan, M., 2022b. A simple temporal information matching mechanism for entity alignment between temporal knowledge graphs, in: Proceedings of the 29th International Conference on Computational Linguistics, International Committee on Computational Linguistics, Gyeongju, Republic of Korea. pp. 2075–2086. URL: https://aclanthology.org/2022.coling-1.181 .
  • Chen et al. [2023a] Chen, J., Geng, Y., Chen, Z., Pan, J.Z., He, Y., Zhang, W., Horrocks, I., Chen, H., 2023a. Zero-shot and few-shot learning with knowledge graphs: A comprehensive survey. Proceedings of the IEEE , 1–33doi: 10.1109/JPROC.2023.3279374 .
  • Chen et al. [2022] Chen, K., Wang, Y., Li, Y., Li, A., 2022. RotateQVS: Representing temporal information as rotations in quaternion vector space for temporal knowledge graph completion, in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 5843–5857. URL: https://aclanthology.org/2022.acl-long.402 , doi: 10.18653/v1/2022.acl-long.402 .
  • Chen et al. [2023b] Chen, Z., Xu, C., Su, F., Huang, Z., Dou, Y., 2023b. Meta-learning based knowledge extrapolation for temporal knowledge graph, in: Proceedings of the ACM Web Conference 2023, Association for Computing Machinery, New York, NY, USA. p. 2433–2443. URL: https://doi.org/10.1145/3543507.3583279 , doi: 10.1145/3543507.3583279 .
  • Cho et al. [2014] Cho, K., van Merrienboer, B., Gülçehre, Ç., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y., 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation, in: Moschitti, A., Pang, B., Daelemans, W. (Eds.), Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 25-29, 2014, Doha, Qatar, A meeting of SIGDAT, a Special Interest Group of the ACL, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 1724–1734. URL: https://doi.org/10.3115/v1/d14-1179 , doi: 10.3115/v1/d14-1179 .
  • Daley and Vere-Jones [2008] Daley, D.J., Vere-Jones, D., 2008. An Introduction to the Theory of Point Processes. Volume II: General Theory and Structure. Springer, Berlin, GER. URL: https://link.springer.com/book/10.1007/978-0-387-49835-5 .
  • Dasgupta et al. [2018] Dasgupta, S.S., Ray, S.N., Talukdar, P., 2018. HyTE: Hyperplane-based temporally aware knowledge graph embedding, in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 2001–2011. URL: https://aclanthology.org/D18-1225 , doi: 10.18653/v1/D18-1225 .
  • Deng et al. [2020] Deng, S., Rangwala, H., Ning, Y., 2020. Dynamic knowledge graph based multi-event forecasting, in: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery; Data Mining, Association for Computing Machinery, New York, NY, USA. p. 1585–1595. URL: https://doi.org/10.1145/3394486.3403209 , doi: 10.1145/3394486.3403209 .
  • Devlin et al. [2019] Devlin, J., Chang, M., Lee, K., Toutanova, K., 2019. BERT: pre-training of deep bidirectional transformers for language understanding, in: Burstein, J., Doran, C., Solorio, T. (Eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 4171–4186. URL: https://doi.org/10.18653/v1/n19-1423 , doi: 10.18653/v1/n19-1423 .
  • Dietz et al. [2018] Dietz, L., Kotov, A., Meij, E., 2018. Utilizing knowledge graphs for text-centric information retrieval, in: The 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, Association for Computing Machinery, New York, NY, USA. p. 1387–1390. URL: https://doi.org/10.1145/3209978.3210187 , doi: 10.1145/3209978.3210187 .
  • Ding et al. [2023] Ding, Z., Cai, H., Wu, J., Ma, Y., Liao, R., Xiong, B., Tresp, V., 2023. Zero-shot relational learning on temporal knowledge graphs with large language models. arXiv preprint arXiv:2311.10112 .
  • Fu et al. [2022] Fu, G., Meng, Z., Han, Z., Ding, Z., Ma, Y., Schubert, M., Tresp, V., Wattenhofer, R., 2022. TempCaps: A capsule network-based embedding model for temporal knowledge graph completion, in: Proceedings of the Sixth Workshop on Structured Prediction for NLP, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 22–31. URL: https://aclanthology.org/2022.spnlp-1.3 , doi: 10.18653/v1/2022.spnlp-1.3 .
  • García-Durán et al. [2018] García-Durán, A., Dumancic, S., Niepert, M., 2018. Learning sequence encoders for temporal knowledge graph completion, in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 4816–4821. URL: https://aclanthology.org/D18-1516 , doi: 10.18653/v1/D18-1516 .
  • Gardner and Dorling [1998] Gardner, M.W., Dorling, S., 1998. Artificial neural networks (the multilayer perceptron)—a review of applications in the atmospheric sciences. Atmospheric environment 32, 2627–2636. URL: https://doi.org/10.1016/S1352-2310(97)00447-0 .
  • Goel et al. [2020] Goel, R., Kazemi, S.M., Brubaker, M.A., Poupart, P., 2020. Diachronic embedding for temporal knowledge graph completion, in: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, New York, NY, USA, February 7-12, 2020, AAAI Press, Palo Alto, CA, USA. pp. 3988–3995. URL: https://ojs.aaai.org/index.php/AAAI/article/view/5815 .
  • Gong et al. [2023] Gong, X., Qin, J., Chai, H., Ding, Y., Jia, Y., Liao, Q., 2023. Temporal-relational matching network for few-shot temporal knowledge graph completion, in: Database Systems for Advanced Applications: 28th International Conference, DASFAA 2023, Tianjin, China, April 17–20, 2023, Proceedings, Part II, Springer-Verlag, Berlin, Heidelberg. p. 768–783. URL: https://doi.org/10.1007/978-3-031-30672-3_52 , doi: 10.1007/978-3-031-30672-3_52 .
  • Han et al. [2021a] Han, Z., Chen, P., Ma, Y., Tresp, V., 2021a. Explainable subgraph reasoning for forecasting on temporal knowledge graphs, in: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021, OpenReview.net, Austria. URL: https://openreview.net/forum?id=pGIHq1m7PU .
  • Han et al. [2021b] Han, Z., Ding, Z., Ma, Y., Gu, Y., Tresp, V., 2021b. Learning neural ordinary equations for forecasting future links on temporal knowledge graphs, in: Moens, M., Huang, X., Specia, L., Yih, S.W. (Eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, Association for Computational Linguistics. pp. 8352–8364. URL: https://doi.org/10.18653/v1/2021.emnlp-main.658 , doi: 10.18653/v1/2021.emnlp-main.658 .
  • Han et al. [2023] Han, Z., Liao, R., Gu, J., Zhang, Y., Ding, Z., Gu, Y., Köppl, H., Schütze, H., Tresp, V., 2023. Ecola: Enhanced temporal knowledge embeddings with contextualized language representations. arXiv:2203.09590 .
  • Han et al. [2020a] Han, Z., Ma, Y., Chen, P., Tresp, V., 2020a. Dyernie: Dynamic evolution of riemannian manifold embeddings for temporal knowledge graph completion. arXiv:2011.03984 .
  • Han et al. [2020b] Han, Z., Ma, Y., Wang, Y., Günnemann, S., Tresp, V., 2020b. Graph hawkes neural network for forecasting on temporal knowledge graphs, in: Das, D., Hajishirzi, H., McCallum, A., Singh, S. (Eds.), Conference on Automated Knowledge Base Construction, AKBC 2020, Virtual, June 22-24, 2020, OpenReview.net, Virtual. URL: https://doi.org/10.24432/C50018 , doi: 10.24432/C50018 .
  • Hawkes [1971] Hawkes, A.G., 1971. Spectra of some self-exciting and mutually exciting point processes. Biometrika 58, 83–90. URL: https://doi.org/10.1093/biomet/58.1.83 .
  • He et al. [2021] He, G., Lan, Y., Jiang, J., Zhao, W.X., Wen, J.R., 2021. Improving multi-hop knowledge base question answering by learning intermediate supervision signals, in: Proceedings of the 14th ACM International Conference on Web Search and Data Mining, Association for Computing Machinery, New York, NY, USA. p. 553–561. URL: https://doi.org/10.1145/3437963.3441753 , doi: 10.1145/3437963.3441753 .
  • Hitchcock [1927] Hitchcock, F.L., 1927. The expression of a tensor or a polyadic as a sum of products. Journal of Mathematics and Physics 6, 164–189. URL: https://doi.org/10.1002/sapm192761164 , doi: 10.1002/sapm192761164 .
  • Hochreiter and Schmidhuber [1997] Hochreiter, S., Schmidhuber, J., 1997. Long short-term memory. Neural Comput. 9, 1735–1780. URL: https://doi.org/10.1162/neco.1997.9.8.1735 , doi: 10.1162/neco.1997.9.8.1735 .
  • Hoffart et al. [2011] Hoffart, J., Suchanek, F.M., Berberich, K., Lewis-Kelham, E., de Melo, G., Weikum, G., 2011. Yago2: Exploring and querying world knowledge in time, space, context, and many languages, in: Proceedings of the 20th International Conference Companion on World Wide Web, Association for Computing Machinery, New York, NY, USA. p. 229–232. URL: https://doi.org/10.1145/1963192.1963296 , doi: 10.1145/1963192.1963296 .
  • Ji et al. [2021] Ji, S., Pan, S., Cambria, E., Marttinen, P., Philip, S.Y., 2021. A survey on knowledge graphs: Representation, acquisition, and applications. IEEE transactions on neural networks and learning systems 33, 494–514. URL: https://doi.org/10.1109/TNNLS.2021.3070843 , doi: 10.1109/TNNLS.2021.3070843 .
  • Jin et al. [2020] Jin, W., Qu, M., Jin, X., Ren, X., 2020. Recurrent event network: Autoregressive structure inferenceover temporal knowledge graphs, in: Webber, B., Cohn, T., He, Y., Liu, Y. (Eds.), Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 6669–6683. URL: https://doi.org/10.18653/v1/2020.emnlp-main.541 , doi: 10.18653/v1/2020.emnlp-main.541 .
  • Kaelbling et al. [1996] Kaelbling, L.P., Littman, M.L., Moore, A.W., 1996. Reinforcement learning: A survey. J. Artif. Intell. Res. 4, 237–285. URL: https://doi.org/10.1613/jair.301 , doi: 10.1613/jair.301 .
  • Kang and Cho [2022] Kang, D., Cho, M., 2022. Integrative few-shot learning for classification and segmentation, in: IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, IEEE. pp. 9969–9980. URL: https://doi.org/10.1109/CVPR52688.2022.00974 , doi: 10.1109/CVPR52688.2022.00974 .
  • Kazemi and Poole [2018] Kazemi, S.M., Poole, D., 2018. Simple embedding for link prediction in knowledge graphs, in: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (Eds.), Advances in Neural Information Processing Systems, Curran Associates, Inc., New York, NY, USA. URL: https://proceedings.neurips.cc/paper/2018/file/b2ab001909a8a6f04b51920306046ce5-Paper.pdf .
  • Kipf and Welling [2017] Kipf, T.N., Welling, M., 2017. Semi-supervised classification with graph convolutional networks, in: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, OpenReview.net, Toulon, France. URL: https://openreview.net/forum?id=SJU4ayYgl .
  • Lacroix et al. [2020] Lacroix, T., Obozinski, G., Usunier, N., 2020. Tensor decompositions for temporal knowledge base completion, in: 8th International Conference on Learning Representations, ICLR 2020, OpenReview.net, Addis Ababa, Ethiopia. URL: https://openreview.net/forum?id=rke2P1BFwS .
  • Leblay and Chekol [2018] Leblay, J., Chekol, M.W., 2018. Deriving validity time in knowledge graph, in: Companion Proceedings of the The Web Conference 2018, International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE. p. 1771–1776. URL: https://doi.org/10.1145/3184558.3191639 , doi: 10.1145/3184558.3191639 .
  • Lee et al. [2023] Lee, D.H., Ahrabian, K., Jin, W., Morstatter, F., Pujara, J., 2023. Temporal knowledge graph forecasting without knowledge using in-context learning. arXiv preprint arXiv:2305.10613 .
  • Leetaru and Schrodt [2013] Leetaru, K., Schrodt, P.A., 2013. Gdelt: Global data on events, location, and tone, 1979–2012, in: ISA annual convention, Citeseer. Citeseer, State College, PA, USA. pp. 1–49. URL: http://data.gdeltproject.org/documentation/ISA.2013.GDELT.pdf .
  • Li et al. [2022a] Li, Y., Sun, S., Zhao, J., 2022a. Tirgn: Time-guided recurrent graph network with local-global historical patterns for temporal knowledge graph reasoning, in: Raedt, L.D. (Ed.), Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022, ijcai.org, San Francisco, CA, USA. pp. 2152–2158. URL: https://doi.org/10.24963/ijcai.2022/299 , doi: 10.24963/ijcai.2022/299 .
  • Li et al. [2022b] Li, Z., Guan, S., Jin, X., Peng, W., Lyu, Y., Zhu, Y., Bai, L., Li, W., Guo, J., Cheng, X., 2022b. Complex evolutional pattern learning for temporal knowledge graph reasoning, in: Muresan, S., Nakov, P., Villavicencio, A. (Eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 290–296. URL: https://doi.org/10.18653/v1/2022.acl-short.32 , doi: 10.18653/v1/2022.acl-short.32 .
  • Li et al. [2021a] Li, Z., Jin, X., Guan, S., Li, W., Guo, J., Wang, Y., Cheng, X., 2021a. Search from history and reason for future: Two-stage reasoning on temporal knowledge graphs, in: Zong, C., Xia, F., Li, W., Navigli, R. (Eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 4732–4743. URL: https://doi.org/10.18653/v1/2021.acl-long.365 , doi: 10.18653/v1/2021.acl-long.365 .
  • Li et al. [2021b] Li, Z., Jin, X., Li, W., Guan, S., Guo, J., Shen, H., Wang, Y., Cheng, X., 2021b. Temporal knowledge graph reasoning based on evolutional representation learning, in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Association for Computing Machinery, New York, NY, USA. p. 408–417. URL: https://doi.org/10.1145/3404835.3462963 , doi: 10.1145/3404835.3462963 .
  • Liao et al. [2023] Liao, R., Jia, X., Ma, Y., Tresp, V., 2023. Gentkg: Generative forecasting on temporal knowledge graph. arXiv preprint arXiv:2310.07793 .
  • Lin and She [2020] Lin, L., She, K., 2020. Tensor decomposition-based temporal knowledge graph embedding, in: 32nd IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2020, Baltimore, MD, USA, November 9-11, 2020, IEEE, Baltimore, MD, USA. pp. 969–975. URL: https://doi.org/10.1109/ICTAI50040.2020.00151 , doi: 10.1109/ICTAI50040.2020.00151 .
  • Liu et al. [2023] Liu, X., Zhang, Y., Shan, D., 2023. Unseen object few-shot semantic segmentation for robotic grasping. IEEE Robotics Autom. Lett. 8, 320–327. URL: https://doi.org/10.1109/LRA.2022.3226073 , doi: 10.1109/LRA.2022.3226073 .
  • Luo et al. [2024] Luo, R., Gu, T., Li, H., Li, J., Lin, Z., Li, J., Yang, Y., 2024. Chain of history: Learning and forecasting with llms for temporal knowledge graph completion. arXiv preprint arXiv:2401.06072 .
  • Mei and Eisner [2017] Mei, H., Eisner, J., 2017. The neural hawkes process: A neurally self-modulating multivariate point process, in: Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R. (Eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, Curran Associates, Inc., New York, NY, USA. pp. 6754–6764. URL: https://proceedings.neurips.cc/paper/2017/hash/6463c88460bd63bbe256e495c63aa40b-Abstract.html .
  • Messner et al. [2022] Messner, J., Abboud, R., Ceylan, İ.İ., 2022. Temporal knowledge graph completion using box embeddings, in: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, AAAI Press. pp. 7779–7787. URL: https://ojs.aaai.org/index.php/AAAI/article/view/20746 .
  • Mo et al. [2021] Mo, C., Wang, Y., Jia, Y., Liao, Q., 2021. Survey on temporal knowledge graph, in: Sixth IEEE International Conference on Data Science in Cyberspace, DSC 2021, Shenzhen, China, October 9-11, 2021, IEEE, Berkeley,CA, USA. pp. 294–300. URL: https://doi.org/10.1109/DSC53577.2021.00047 , doi: 10.1109/DSC53577.2021.00047 .
  • Nguyen et al. [2019] Nguyen, D.Q., Vu, T., Nguyen, T.D., Nguyen, D.Q., Phung, D., 2019. A capsule network-based embedding model for knowledge graph completion and search personalization, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 2180–2189. URL: https://aclanthology.org/N19-1226 , doi: 10.18653/v1/N19-1226 .
  • O’brien [2010] O’brien, S.P., 2010. Crisis early warning and decision support: Contemporary approaches and thoughts on future research. International studies review 12, 87–104. URL: https://doi.org/10.1111/j.1468-2486.2009.00914.x .
  • OpenAI [2023] OpenAI, 2023. Gpt-4 technical report. arXiv:2303.08774 .
  • Park et al. [2022] Park, N., Liu, F., Mehta, P., Cristofor, D., Faloutsos, C., Dong, Y., 2022. Evokg: Jointly modeling event time and network structure for reasoning over temporal knowledge graphs, in: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, Association for Computing Machinery, New York, NY, USA. p. 794–803. URL: https://doi.org/10.1145/3488560.3498451 , doi: 10.1145/3488560.3498451 .
  • Sadeghian et al. [2021] Sadeghian, A., Armandpour, M., Colas, A., Wang, D.Z., 2021. Chronor: Rotation based temporal knowledge graph embedding, in: Proceedings of the AAAI Conference on Artificial Intelligence, AAAI2021, AAAI Press, Palo Alto, CA, USA. pp. 6471–6479. URL: https://ojs.aaai.org/index.php/AAAI/article/view/16802 , doi: 10.1609/aaai.v35i7.16802 .
  • Saxena et al. [2021] Saxena, A., Chakrabarti, S., Talukdar, P.P., 2021. Question answering over temporal knowledge graphs, in: Zong, C., Xia, F., Li, W., Navigli, R. (Eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 6663–6676. URL: https://doi.org/10.18653/v1/2021.acl-long.520 , doi: 10.18653/v1/2021.acl-long.520 .
  • Scarselli et al. [2009] Scarselli, F., Gori, M., Tsoi, A.C., Hagenbuchner, M., Monfardini, G., 2009. The graph neural network model. IEEE Transactions on Neural Networks 20, 61–80. URL: https://ieeexplore.ieee.org/document/4700287 , doi: 10.1109/TNN.2008.2005605 .
  • Schlichtkrull et al. [2018] Schlichtkrull, M.S., Kipf, T.N., Bloem, P., van den Berg, R., Titov, I., Welling, M., 2018. Modeling relational data with graph convolutional networks, in: Gangemi, A., Navigli, R., Vidal, M., Hitzler, P., Troncy, R., Hollink, L., Tordai, A., Alam, M. (Eds.), The Semantic Web - 15th International Conference, ESWC 2018, Heraklion, Crete, Greece, June 3-7, 2018, Proceedings, Springer, Berlin, GER. pp. 593–607. URL: https://doi.org/10.1007/978-3-319-93417-4_38 , doi: 10.1007/978-3-319-93417-4\_38 .
  • Shang et al. [2019] Shang, C., Tang, Y., Huang, J., Bi, J., He, X., Zhou, B., 2019. End-to-end structure-aware convolutional networks for knowledge base completion, in: The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, AAAI Press, Palo Alto, CA, USA. pp. 3060–3067. URL: https://doi.org/10.1609/aaai.v33i01.33013060 , doi: 10.1609/aaai.v33i01.33013060 .
  • Shang et al. [2022] Shang, C., Wang, G., Qi, P., Huang, J., 2022. Improving time sensitivity for question answering over temporal knowledge graphs, in: Muresan, S., Nakov, P., Villavicencio, A. (Eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 8017–8026. URL: https://doi.org/10.18653/v1/2022.acl-long.552 , doi: 10.18653/v1/2022.acl-long.552 .
  • Shao et al. [2022] Shao, P., Zhang, D., Yang, G., Tao, J., Che, F., Liu, T., 2022. Tucker decomposition-based temporal knowledge graph completion. Knowl. Based Syst. 238, 107841. URL: https://doi.org/10.1016/j.knosys.2021.107841 , doi: 10.1016/j.knosys.2021.107841 .
  • Sun et al. [2021] Sun, H., Zhong, J., Ma, Y., Han, Z., He, K., 2021. Timetraveler: Reinforcement learning for temporal knowledge graph forecasting, in: Moens, M., Huang, X., Specia, L., Yih, S.W. (Eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 8306–8319. URL: https://doi.org/10.18653/v1/2021.emnlp-main.655 , doi: 10.18653/v1/2021.emnlp-main.655 .
  • Sun et al. [2019] Sun, Z., Deng, Z., Nie, J., Tang, J., 2019. Rotate: Knowledge graph embedding by relational rotation in complex space, in: 7th International Conference on Learning Representations, ICLR 2019, OpenReview.net, New Orleans, LA, USA. URL: https://openreview.net/forum?id=HkgEQnRqYQ .
  • Trivedi et al. [2017] Trivedi, R., Dai, H., Wang, Y., Song, L., 2017. Know-evolve: Deep temporal reasoning for dynamic knowledge graphs, in: Precup, D., Teh, Y.W. (Eds.), Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, PMLR, Cambridge, MA, USA. pp. 3462–3471. URL: http://proceedings.mlr.press/v70/trivedi17a.html .
  • Trouillon et al. [2016] Trouillon, T., Welbl, J., Riedel, S., Gaussier, É., Bouchard, G., 2016. Complex embeddings for simple link prediction, in: Proceedings of The 33rd International conference on machine learning, PMLR, New York, NY, USA. pp. 2071–2080. URL: https://proceedings.mlr.press/v48/trouillon16.html .
  • Tucker [1966] Tucker, L.R., 1966. Some mathematical notes on three-mode factor analysis. Psychometrika 31, 279–311. URL: http://hdl.handle.net/10.1007/BF02289464 , doi: 10.1007/BF02289464 .
  • Vashishth et al. [2020] Vashishth, S., Sanyal, S., Nitin, V., Talukdar, P.P., 2020. Composition-based multi-relational graph convolutional networks, in: 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, OpenReview.net, Addis Ababa, Ethiopia. URL: https://openreview.net/forum?id=BylA_C4tPr .
  • Vaswani et al. [2017] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L.u., Polosukhin, I., 2017. Attention is all you need, in: Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R. (Eds.), Advances in Neural Information Processing Systems, Curran Associates, Inc. URL: https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf .
  • Velickovic et al. [2018] Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y., 2018. Graph attention networks, in: 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, OpenReview.net, Vancouver, BC, Canada. URL: https://openreview.net/forum?id=rJXMpikCZ .
  • Vrandečić and Krötzsch [2014] Vrandečić, D., Krötzsch, M., 2014. Wikidata: A free collaborative knowledgebase. Commun. ACM 57, 78–85. URL: https://doi.org/10.1145/2629489 , doi: 10.1145/2629489 .
  • Wang et al. [2022a] Wang, J., Zhu, C., Zhu, W., 2022a. Dynamic embedding graph attention networks for temporal knowledge graph completion, in: Memmi, G., Yang, B., Kong, L., Zhang, T., Qiu, M. (Eds.), Knowledge Science, Engineering and Management, Springer International Publishing, Cham. pp. 722–734. URL: https://doi.org/10.1007/978-3-031-10983-6_55 .
  • Wang et al. [2022b] Wang, R., zheng li, Sun, D., Liu, S., Li, J., Yin, B., Abdelzaher, T., 2022b. Learning to sample and aggregate: Few-shot reasoning over temporal knowledge graphs, in: Oh, A.H., Agarwal, A., Belgrave, D., Cho, K. (Eds.), Advances in Neural Information Processing Systems. URL: https://openreview.net/forum?id=1LmgISIDZJ .
  • Wang et al. [2020] Wang, Y., Yao, Q., Kwok, J.T., Ni, L.M., 2020. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. 53. URL: https://doi.org/10.1145/3386252 , doi: 10.1145/3386252 .
  • Xia et al. [2020] Xia, C., Zhang, C., Zhang, J., Liang, T., Peng, H., Yu, P.S., 2020. Low-shot learning in natural language processing, in: 2020 IEEE Second International Conference on Cognitive Machine Intelligence (CogMI), pp. 185–189. doi: 10.1109/CogMI50398.2020.00031 .
  • Xia et al. [2022] Xia, Y., Zhang, M., Liu, Q., Wu, S., Zhang, X.Y., 2022. MetaTKG: Learning evolutionary meta-knowledge for temporal knowledge graph reasoning, in: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Abu Dhabi, United Arab Emirates. pp. 7230–7240. URL: https://aclanthology.org/2022.emnlp-main.487 .
  • Xu et al. [2020] Xu, C., Nayyeri, M., Alkhoury, F., Shariat Yazdi, H., Lehmann, J., 2020. TeRo: A time-aware knowledge graph embedding via temporal rotation, in: Proceedings of the 28th International Conference on Computational Linguistics, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 1583–1593. URL: https://aclanthology.org/2020.coling-main.139 , doi: 10.18653/v1/2020.coling-main.139 .
  • Xu et al. [2023] Xu, C., Nayyeri, M., Chen, Y.Y., Lehmann, J., 2023. Geometric algebra based embeddings for static and temporal knowledge graph completion. IEEE Transactions on Knowledge and Data Engineering 35, 4838–4851. doi: 10.1109/TKDE.2022.3151435 .
  • Xu et al. [2021] Xu, C., Su, F., Lehmann, J., 2021. Time-aware graph neural network for entity alignment between temporal knowledge graphs, in: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Stroudsburg, PA, USA. pp. 8999–9010. URL: https://aclanthology.org/2021.emnlp-main.709 , doi: 10.18653/v1/2021.emnlp-main.709 .
  • Xu et al. [2022] Xu, C., Su, F., Xiong, B., Lehmann, J., 2022. Time-aware entity alignment using temporal relational attention, in: Proceedings of the ACM Web Conference 2022, Association for Computing Machinery, New York, NY, USA. p. 788–797. URL: https://doi.org/10.1145/3485447.3511922 , doi: 10.1145/3485447.3511922 .
  • Zhang et al. [2022] Zhang, F., Zhang, Z., Ao, X., Zhuang, F., Xu, Y., He, Q., 2022. Along the time: Timeline-traced embedding for temporal knowledge graph completion, in: Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Association for Computing Machinery, New York, NY, USA. p. 2529–2538. URL: https://doi.org/10.1145/3511808.3557233 , doi: 10.1145/3511808.3557233 .
  • Zhang et al. [2023a] Zhang, M., Xia, Y., Liu, Q., Wu, S., Wang, L., 2023a. Learning latent relations for temporal knowledge graph reasoning, in: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 12617–12631.
  • Zhang et al. [2023b] Zhang, S., Liang, X., Li, Z., Feng, J., Zheng, X., Wu, B., 2023b. Biqcap: A biquaternion and capsule network-based embedding model for temporal knowledge graph completion, in: Wang, X., Sapino, M.L., Han, W., Abbadi, A.E., Dobbie, G., Feng, Z., Shao, Y., Yin, H. (Eds.), Database Systems for Advanced Applications - 28th International Conference, DASFAA 2023, Tianjin, China, April 17-20, 2023, Proceedings, Part II, Springer. pp. 673–688. URL: https://doi.org/10.1007/978-3-031-30672-3_45 , doi: 10.1007/978-3-031-30672-3\_45 .
  • Zhang et al. [2023c] Zhang, S., Liang, X., Tang, H., Zheng, X., X. Zhang, A., Ma, Y., 2023c. Ducape: Dual quaternion and capsule network–based temporal knowledge graph embedding. ACM Trans. Knowl. Discov. Data 17. URL: https://doi.org/10.1145/3589644 , doi: 10.1145/3589644 .
  • Zhao et al. [2023] Zhao, W.X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z., Du, Y., Yang, C., Chen, Y., Chen, Z., Jiang, J., Ren, R., Li, Y., Tang, X., Liu, Z., Liu, P., Nie, J.Y., Wen, J.R., 2023. A survey of large language models. arXiv:2303.18223 .
  • Zhu et al. [2021] Zhu, C., Chen, M., Fan, C., Cheng, G., Zhang, Y., 2021. Learning from history: Modeling temporal knowledge graphs with sequential copy-generation networks, in: Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, AAAI Press, Palo Alto, CA, USA. pp. 4732–4740. URL: https://ojs.aaai.org/index.php/AAAI/article/view/16604 .

ar5iv homepage

Subscribe to the PwC Newsletter

Join the community, edit social preview.

temporal knowledge graph representation learning with local and global evolutions

Add a new code entry for this paper

Remove a code repository from this paper.

temporal knowledge graph representation learning with local and global evolutions

Mark the official implementation from paper authors

Add a new evaluation result row.

  • GRAPH REPRESENTATION LEARNING
  • REPRESENTATION LEARNING

Remove a task

temporal knowledge graph representation learning with local and global evolutions

Add a method

Remove a method, edit datasets, temporal knowledge graph representation learning with local and global evolutions.

Knowledge-Based Systems 2022  ·  Jiasheng Zhang , Shuang Liang , Yongpan Sheng , Jie Shao · Edit social preview

Temporal knowledge graph (TKG) representation learning aims to project entities and relations in TKG to low-dimensional vector space while preserving the evolutionary nature of TKG. Most existing methods separately treat knowledge that happened at different times, which fails to explore how temporal knowledge graph evolves over time. Actually, TKG should evolve both on local and global structures. The local structure evolution describes the formation process of graph structure in a detailed manner, while the global structure evolution refers to the dynamic topology (e.g., community partition) of graph, which is derived from the continuous formation process. Both of them are key factors for understanding the evolutionary nature of TKG. Unfortunately, seldom attention has been paid to this aspect. In this paper, we propose a new TKG representation learning framework with local and global structure evolutions, named EvoExplore. Specifically, we define the local structure evolution as the establishment process of relations between entities, and propose a hierarchical-attention-based temporal point process to capture the formation process of graph structure in a fine-grained manner. For global structure evolution, we propose a novel soft modularity parameterized by entity representations to capture the dynamic community partition of TKG. Finally, we employ a multi-task loss function to jointly optimize the above two parts, which allows EvoExplore to learn the mutual influences of local and global structure evolutions. Experimental results on three realworld datasets demonstrate the superiority of EvoExplore compared with baseline methods.

Code Edit Add Remove Mark official

Tasks edit add remove, datasets edit, results from the paper edit add remove, methods edit add remove.

Complex Evolutional Pattern Learning for Temporal Knowledge Graph Reasoning

Zixuan Li , Saiping Guan , Xiaolong Jin , Weihua Peng , Yajuan Lyu , Yong Zhu , Long Bai , Wei Li , Jiafeng Guo , Xueqi Cheng

Export citation

  • Preformatted

Markdown (Informal)

[Complex Evolutional Pattern Learning for Temporal Knowledge Graph Reasoning](https://aclanthology.org/2022.acl-short.32) (Li et al., ACL 2022)

  • Complex Evolutional Pattern Learning for Temporal Knowledge Graph Reasoning (Li et al., ACL 2022)
  • Zixuan Li, Saiping Guan, Xiaolong Jin, Weihua Peng, Yajuan Lyu, Yong Zhu, Long Bai, Wei Li, Jiafeng Guo, and Xueqi Cheng. 2022. Complex Evolutional Pattern Learning for Temporal Knowledge Graph Reasoning . In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) , pages 290–296, Dublin, Ireland. Association for Computational Linguistics.

Papers on Temporal Knowledge Graph Embedding and Reasoning

Papers on tkge-r.

Github repo stars

Some papers on Temporal Knowledge Graph Embedding and Reasoning

Useful research resources

Graph-based Deep Learning Literature, Github

links to conference publications in graph-based deep learning

Reinforcement learning on graphs: A survey, Github

This collection of papers can be used to summarize research about graph reinforcement learning for the convenience of researchers.

Awesome Machine Learning for Combinatorial Optimization Resources, Github

Awesome machine learning for combinatorial optimization papers.

Awesome-TKGC, Github

A collection of papers and resources about temporal knowledge graph completion (TKGC).

AKGR: Awesome Knowledge Graph Reasoning, Github

AKGR: Awesome Knowledge Graph Reasoning is a collection of knowledge graph reasoning works, including papers, codes and datasets.

Awesome Knowledge Graph, Github

A curated list of Knowledge Graph related learning materials, databases, tools and other resources.

Awesome-DynamicGraphLearning, Github

Awesome papers about machine learning (deep learning) on dynamic (temporal) graphs (networks / knowledge graphs).

KGE, Github

Some papers on Knowledge Graph Embedding(KGE)

KGLQ, Github

Some papers about Logical Query on Knowledge Graphs (KGLQ)

ADGC: Awesome Deep Graph Clustering, Github

Awesome Deep Graph Clustering is a collection of SOTA, novel deep graph clustering methods (papers, codes, and datasets).

Graph Adversarial Learning Literature, Github

A curated list of adversarial attacks and defenses papers on graph-structured data.

[1] New Frontiers of Knowledge Graph Reasoning: Recent Advances and Future Trends - WWW 2024, Webpage

[1] Knowledge Graph Reasoning and Its Applications - KDD 2023, Webpage

[1] Reasoning on Knowledge Graphs: Symbolic or Neural? - AAAI 2022, Webpage

[1] All You Need to Know to Build a Product Knowledge Graph - KDD 2021 Tutorial, Webpage

[1] Fact Checking: Theory and Practice - KDD 2018 Tutorial, Webpage

Survey Papers

Knowledge Graph Embedding: An Overview , APSIPA Transactions on Signal and Information Processing, 2024. paper Ge, X., Wang, Y. C., Wang, B., & Kuo, C. C. J

Survey of Temporal Knowledge Graph Completion Methods , Journal of Computer Engineering & Applications, 2024. paper Lei, X. I. A. O., & Qi, L. I.

Overview of Knowledge Reasoning for Knowledge Graph , Neurocomputing, 2024. paper Liu, X., Mao, T., Shi, Y., & Ren, Y.

Knowledge graph embedding: A survey from the perspective of representation spaces , ACM Computing Surveys, 2024. paper Cao, J., Fang, J., Meng, Z., & Liang, S.

A survey on graph representation learning methods , ACM Transactions on Intelligent Systems and Technology, 2024. paper Khoshraftar, S., & An, A.

Knowledge Graphs Meet Multi-Modal Learning: A Comprehensive Survey , ArXiv, 2024. paper Chen, Z., Zhang, Y., Fang, Y., Geng, Y., Guo, L., Chen, X., … & Chen, H.

A survey for managing temporal data in RDF , Information Systems, 2024. paper Wu, Di, Hsien-Tseng Wang, and Abdullah Uz Tansel

A Survey on Temporal Knowledge Graph: Representation Learning and Applications , ArXiv, 2024. paper Cai, Li, Xin Mao, Yuhao Zhou, Zhaoguang Long, Changxu Wu, and Man Lan.

A Survey on Temporal Knowledge Graph Completion: Taxonomy, Progress, and Prospects , ArXiv, 2023. paper

Jiapu Wang, Boyue Wang, Meikang Qiu, Shirui Pan, Bo Xiong, Heng Liu, Linhao Luo, Tengfei Liu, Yongli Hu, Baocai Yin, Wen Gao

Knowledge Graphs: Opportunities and Challenges , Artificial Intelligence Review, 2023, paper

Ciyuan Peng, Feng Xia, Mehdi Naseriparsa & Francesco Osborne

Generalizing to Unseen Elements: A Survey on Knowledge Extrapolation for Knowledge Graphs , ArXiv, 2023. paper

Mingyang Chen, Wen Zhang, Yuxia Geng, Zezhong Xu, Jeff Z. Pan, Huajun Chen

A Comprehensive Survey on Automatic Knowledge Graph Construction , ArXiv, 2023. paper

Lingfeng Zhong, Jia Wu, Qian Li, Hao Peng, Xindong Wu

Temporal Knowledge Graph Completion: A Survey ArXiv, 2022. paper

Borui Cai, Yong Xiang, Longxiang Gao, He Zhang, Yunfeng Li, Jianxin Li.

Update : Borui Cai, Yong Xiang, Longxiang Gao, He Zhang, Yunfeng Li, Jianxin Li, Temporal Knowledge Graph Completion: A Survey , 2023 Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, Survey Track. Pages 6545-6553. paper

Reasoning over different types of knowledge graphs: Static, temporal and multi-modal , ArXiv, 2022. paper

A Survey on Temporal Knowledge Graphs-Extrapolation and Interpolation Tasks , Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery, paper

Sulin Chen & Jingbin Wang

Generalizing to Unseen Elements: A Survey on Knowledge Extrapolation for Knowledge Graphs . 2023 Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, Survey Track. paper

Survey on Temporal Knowledge Graph , 2021 IEEE Sixth International Conference on Data Science in Cyberspace (DSC). paper

Chong Mo; Ye Wang; Yan Jia; Qing Liao

Knowledge-Based Systems

[1] Yue, L., Ren, Y., Zeng, Y., Zhang, J., Zeng, K., Wan, J., & Zhou, M. (2024). Complex expressional characterizations learning based on block decomposition for temporal knowledge graph completion . Knowledge-Based Systems, 111591.

[2] Zhu, L., Zhang, H., & Bai, L. (2024). Hierarchical pattern-based complex query of temporal knowledge graph . Knowledge-Based Systems, 284, 111301.

[3] Huang, H., Xie, L., Liu, M., Lin, J., & Shen, H. (2024). An embedding model for temporal knowledge graphs with long and irregular intervals . Knowledge-Based Systems, 111893.

Applied Intelligence

[1] Wang, J., Wu, R., Wu, Y., Zhang, F., Zhang, S., & Guo, K. (2024). MPNet: temporal knowledge graph completion based on a multi-policy network . Applied Intelligence, 1-17. Github

[1] Li, X., Zhou, H., Yao, W., Li, W., Liu, B., & Lin, Y. (2024). Intricate Spatiotemporal Dependency Learning for Temporal Knowledge Graph Reasoning . ACM Transactions on Knowledge Discovery from Data.

Information Science

[1] (THOR) Lee, Y. C., Lee, J., Lee, D., & Kim, S. W. (2024). Learning to compensate for lack of information: Extracting latent knowledge for effective temporal knowledge graph completion . Information Sciences, 654, 119857.

Extended version from: Y. -C. Lee, J. Lee, D. Lee and S. -W. Kim, “THOR: Self-Supervised Temporal Knowledge Graph Embedding via Three-Tower Graph Convolutional Networks,” 2022 IEEE International Conference on Data Mining (ICDM), Orlando, FL, USA, 2022, pp. 1035-1040, doi: 10.1109/ICDM54844.2022.00127. Github

[2] (Joint-MTComplEx) Zhang, F., Chen, H., Shi, Y., Cheng, J., & Lin, J. (2024). Joint framework for tensor decomposition-based temporal knowledge graph completion . Information Sciences, 654, 119853.

[3] (DGTL) Liu, Z., Li, Z., Li, W., & Duan, L. (2024). Deep Graph Tensor Learning for Temporal Link Prediction . Information Sciences, 120085. Github

[4] (CRmod) Zhu, L., Chai, D., & Bai, L. (2024). CRmod: Context-Aware Rule-Guided reasoning over temporal knowledge graph . Information Sciences, 120343. Github

[5] Dai, Y., Guo, W., & Eickhoff, C. (2024). Wasserstein adversarial learning based temporal knowledge graph embedding . Information Sciences, 659, 120061.

[6] Xu, X., Jia, W., Yan, L., Lu, X., Wang, C., & Ma, Z. (2024). Spatiotemporal knowledge graph completion via diachronic and transregional word embedding . Information Sciences, 120477.

Information Fusion

[1] (MvTuckER) Wang, H., Yang, J., Yang, L. T., Gao, Y., Ding, J., Zhou, X., & Liu, H. (2024). MvTuckER: Multi-view knowledge graphs represention learning based on tensor tucker model . Information Fusion, 102249.

Information Processing & Management

[1] (STKGR-PR) Meng, X., Bai, L., Hu, J., & Zhu, L. (2024). Multi-hop path reasoning over sparse temporal knowledge graphs based on path completion and reward shaping . Information Processing & Management, 61(2), 103605. Github

Expert Systems with Applications

[1] (CDRGN-SDE) Zhang, D., Feng, W., Wu, Z., Li, G., & Ning, B. (2024). CDRGN-SDE: Cross-Dimensional Recurrent Graph Network with neural Stochastic Differential Equation for temporal knowledge graph embedding . Expert Systems with Applications, 123295. Github

[2] (TPComplEx) Yang, J., Ying, X., Shi, Y., & Xing, B. (2024). Tensor decompositions for temporal knowledge graph completion with time perspective . Expert Systems with Applications, 237, 121267. Github

Frontiers of Computer Science

[1] (EvolveKG) Liu, J., Yu, Z., Guo, B., Deng, C., Fu, L., Wang, X., & Zhou, C. (2024). EvolveKG: a general framework to learn evolving knowledge graphs . Frontiers of Computer Science, 18(3), 183309.

Neural networks

[1] Shao, P., Tao, J., & Zhang, D. (2024). Bayesian hypernetwork collaborates with time-difference evolutional network for temporal knowledge prediction . Neural Networks, 106146.

[2] Bai, L., Li, N., Li, G., Zhang, Z., & Zhu, L. (2024). Embedding-based Entity Alignment of Cross-Lingual Temporal Knowledge Graphs . Neural Networks, 106143.

[3] 🔥 Mei, X., Yang, L., Jiang, Z., Cai, X., Gao, D., Han, J., & Pan, S. (2024). An Inductive Reasoning Model based on Interpretable Logical Rules over temporal knowledge graph . Neural Networks, 106219. Github

Engineering Applications of Artificial Intelligence

[1] Zhu, L., Zhao, W., & Bai, L. (2024). Quadruple mention text-enhanced temporal knowledge graph reasoning . Engineering Applications of Artificial Intelligence, 133, 108058. Github

Journal of Intelligent Information Systems

[1] Du, C., Li, X., & Li, Z. (2024). Semantic-enhanced reasoning question answering over temporal knowledge graphs . Journal of Intelligent Information Systems, 1-23.

Artificial Intelligence

[1] Dong, H., Wang, P., Xiao, M., Ning, Z., Wang, P., & Zhou, Y. (2024). Temporal Inductive Path Neural Network for Temporal Knowledge Graph Reasoning . Artificial Intelligence, 104085. Github

IEEE Transactions on Fuzzy Systems

[1] Ji, H., Yan, L., & Ma, Z. (2023). FSTRE: Fuzzy Spatiotemporal RDF Knowledge Graph Embedding Using Uncertain Dynamic Vector Projection and Rotation . IEEE Transactions on Fuzzy Systems.

Electronics

[1] 🔥 Xu, H., Bao, J., Li, H., He, C., & Chen, F. (2024). A Multi-View Temporal Knowledge Graph Reasoning Framework with Interpretable Logic Rules and Feature Fusion . Electronics, 13(4), 742.

Neurocomputing

[1] He, M., Zhu, L., & Bai, L. (2024). ConvTKG: A query-aware convolutional neural network-based embedding model for temporal knowledge graph completion . Neurocomputing, 127680.

[1] Zhang, F., Zhang, Z., Zhuang, F., Zhao, Y., Wang, D., & Zheng, H. (2024). Temporal Knowledge Graph Reasoning With Dynamic Memory Enhancement . IEEE Transactions on Knowledge and Data Engineering.

Tsinghua Science and Technology

[1] Han, Y., Lu, G., Zhang, S., Zhang, L., Zou, C., & Wen, G. (2024). A Temporal Knowledge Graph Embedding Model Based on Variable Translation . Tsinghua Science and Technology, 29(5), 1554-1565.

Applied Soft Computing

[1] Bai, L., Chen, M., & Xiao, Q. (2024). Multi-Hop Temporal Knowledge Graph Reasoning with Multi-Agent Reinforcement Learning. Applied Soft Computing, 111727. Github

Semantic Web Journal

[1] (TRKGE) Song, B., Amouzouvi, K., Xu, C., Wang, M., Lehmann, J., & Vahdati, S. Temporal Relevance for Representing Learning over Temporal Knowledge Graphs .

[1] (TPRG) Bai, L., Chen, M., Zhu, L., & Meng, X. (2023). Multi-hop temporal knowledge graph reasoning with temporal path rules guidance . Expert Systems with Applications, 223, 119804. Github

The Journal of Supercomputing

[1] (TKGA) Wang, Z., You, X., & Lv, X. (2023). A relation enhanced model for temporal knowledge graph alignment . The Journal of Supercomputing, 1-23.

Information Systems

[1] (RITI) Liu, R., Yin, G., Liu, Z., & Tian, Y. (2023). Reinforcement learning with time intervals for temporal knowledge graph reasoning . Information Systems, 102292.

Information Sciences

[1] (T-GAE) Hou, X., Ma, R., Yan, L., & Ma, Z. (2023). T-GAE: A Timespan-Aware Graph Attention-based Embedding Model for Temporal Knowledge Graph Completion . Information Sciences, 119225.

[2] (TASTER) Wang, X., Lyu, S., Wang, X., Wu, X., & Chen, H. (2023). Temporal knowledge graph embedding via sparse transfer matrix . Information Sciences, 623, 56-69.

[3] (TLmod) Bai, L., Yu, W., Chai, D., Zhao, W., & Chen, M. (2023). Temporal knowledge graphs reasoning with iterative guidance by temporal logical rules . Information Sciences, 621, 22-35.

IEEE/ACM Transactions on Audio, Speech, and Language Processing

[1] (TARGAT) Xie, Z., Zhu, R., Liu, J., Zhou, G., & Huang, J. X. (2023). TARGAT: A Time-Aware Relational Graph Attention Model for Temporal Knowledge Graph Embedding . IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[1] (TBDRI) Yu, M., Guo, J., Yu, J., Xu, T., Zhao, M., Liu, H., … & Yu, R. (2023). TBDRI: block decomposition based on relational interaction for temporal knowledge graph completion . Applied Intelligence, 53(5), 5072-5084.

[2] (GLANet) Wang, J., Lin, X., Huang, H., Ke, X., Wu, R., You, C., & Guo, K. (2023). GLANet: temporal knowledge graph completion based on global and local information-aware network . Applied Intelligence, 1-17.

[3] (ChronoR-CP) Li, M., Sun, Z., Zhang, W., & Liu, W. (2023). Leveraging semantic property for temporal knowledge graph completion . Applied Intelligence, 53(8), 9247-9260.

[4] (TIAR) Mu, C., Zhang, L., Ma, Y., & Tian, L. (2023). Temporal knowledge subgraph inference based on time-aware relation representation . Applied Intelligence, 53(20), 24237-24252.

[5] (TNTSimplE) He, P., Zhou, G., Zhang, M., Wei, J., & Chen, J. (2023). Improving temporal knowledge graph embedding using tensor factorization . Applied Intelligence, 53(8), 8746-8760.

Neural Networks

[1] (TFSC) Zhang, H., & Bai, L. (2023). Few-shot link prediction for temporal knowledge graphs based on time-aware translation and attention mechanism . Neural Networks, 161, 371-381. Github

[2] Shao, P., Liu, T., Che, F., Zhang, D., & Tao, J. (2023). Adaptive pseudo-Siamese policy network for temporal knowledge prediction . Neural Networks.

[1] Shao, P., He, J., Li, G., Zhang, D., & Tao, J. (2023). Hierarchical Graph Attention Network for Temporal Knowledge Graph Reasoning . Neurocomputing, 126390.

[2] (TANGO) Wang, Z., Ding, D., Ren, M., & Conti, M. (2023). TANGO: A Temporal Spatial Dynamic Graph Model for Event Prediction . Neurocomputing, 126249.

IEEE Transactions on Neural Networks and Learning Systems

[1] (QDN) Wang, J., Wang, B., Gao, J., Li, X., Hu, Y., & Yin, B. (2023). QDN: A Quadruplet Distributor Network for Temporal Knowledge Graph Completion . IEEE Transactions on Neural Networks and Learning Systems. Github

Journal of Systems Science and Systems Engineering

[1] Yan, Z., & Tang, X. (2023). Narrative Graph: Telling Evolving Stories Based on Event-centric Temporal Knowledge Graph . Journal of Systems Science and Systems Engineering, 32(2), 206-221.

[1] (RoAN) Bai, L., Ma, X., Meng, X., Ren, X., & Ke, Y. (2023). RoAN: A relation-oriented attention network for temporal knowledge graph completion . Engineering Applications of Artificial Intelligence, 123, 106308. Github

Future Generation Computer Systems

[1] (TAL-TKGC) Nie, H., Zhao, X., Yao, X., Jiang, Q., Bi, X., Ma, Y., & Sun, Y. (2023). Temporal-structural importance weighted graph convolutional network for temporal knowledge graph completion . Future Generation Computer Systems.

Cognitive Computation

[2] (MsCNN) Liu, W., Wang, P., Zhang, Z., & Liu, Q. (2023). Multi-Scale Convolutional Neural Network for Temporal Knowledge Graph Completion . Cognitive Computation, 1-7.

ACM Transactions on Knowledge Discovery from Data

[1] (DuCape) Zhang, S., Liang, X., Tang, H., Zheng, X., Zhang, A. X., & Ma, Y. DuCape: Dual Quaternion and Capsule Network Based Temporal Knowledge Graph Embedding . ACM Transactions on Knowledge Discovery from Data.

IEEE Transactions on Knowledge and Data Engineering

[1] Li, Y., Chen, H., Li, Y., Li, L., Philip, S. Y., & Xu, G. (2023). Reinforcement Learning based Path Exploration for Sequential Explainable Recommendation . IEEE Transactions on Knowledge and Data Engineering. Github

[1] (RLAT) Bai, L., Chai, D., & Zhu, L. (2023). RLAT: Multi-hop temporal knowledge graph reasoning based on Reinforcement Learning and Attention Mechanism . Knowledge-Based Systems, 269, 110514.

Journal of Computational Design and Engineering

[1] (MetaRT) Zhu, L., Xing, Y., Bai, L., & Chen, X. (2023). Few-shot link prediction with meta-learning for temporal knowledge graphs . Journal of Computational Design and Engineering, 10(2), 711-721.

[1] 🔥 (IMF) Du, Z., Qu, L., Liang, Z., Huang, K., Cui, L., & Gao, Z. (2023). IMF: Interpretable Multi-Hop Forecasting on Temporal Knowledge Graphs . Entropy, 25(4), 666. Github

Complex & Intelligent Systems

[1] (FTMO) Zhu, L., Bai, L., Han, S., & Zhang, M. (2023). Few-shot temporal knowledge graph completion based on meta-optimization . Complex & Intelligent Systems, 9(6), 7461-7474. Github

World Wide Web

[1] (FTMF) Bai, L., Zhang, M., Zhang, H., & Zhang, H. (2023). FTMF: Few-shot temporal knowledge graph completion based on meta-optimization and fault-tolerant mechanism . World Wide Web, 26(3), 1243-1270. Github

[1] (OSLT) Ma, R., Mei, B., Ma, Y., Zhang, H., Liu, M., & Zhao, L. (2023). One-shot relational learning for extrapolation reasoning on temporal knowledge graphs . Data Mining and Knowledge Discovery, 1-18.

[1] (EvoExplore) Jiasheng Zhang, Shuang Liang, Yongpan Sheng, Jie Shao. “Temporal knowledge graph representation learning with local and global evolutions” . Knowledge-Based Systems 2022. Github

[2] (TuckERT) Pengpeng Shao, Dawei Zhang, Guohua Yang, Jianhua Tao, Feihu Che, Tong Liu. “Tucker decomposition-based temporal knowledge graph completion” . Knowledge Based Systems 2022. Github

[1] (BTDG) Yujing Lai, Chuan Chen, Zibin Zheng, Yangqing Zhang. “Block term decomposition with distinct time granularities for temporal knowledge graph completion” . Expert Systems with Applications 2022. Github

[1] (TPath) Luyi Bai, Wenting Yu, Mingzhuo Chen, Xiangnan Ma. “Multi-hop reasoning over paths in temporal knowledge graphs using reinforcement learning” . Applied Soft Computing 2021.

[1] (TPmod) Bai, L., Ma, X., Zhang, M., & Yu, W. (2021). Tpmod: A tendency-guided prediction model for temporal knowledge graph completion . ACM Transactions on Knowledge Discovery from Data, 15(3), 1-17. Github

[2] (Dacha) Chen, L., Tang, X., Chen, W., Qian, Y., Li, Y., & Zhang, Y. (2021). Dacha: A dual graph convolution based temporal knowledge graph representation learning method using historical relation . ACM Transactions on Knowledge Discovery from Data (TKDD), 16(3), 1-18.

IEEE Access

[1] (TDG2E) Tang, X., Yuan, R., Li, Q., Wang, T., Yang, H., Cai, Y., & Song, H. (2020). Timespan-aware dynamic knowledge graph embedding by incorporating temporal evolution . IEEE Access, 8, 6849-6860.

[2] (3DRTE) Wang, J., Zhang, W., Chen, X., Lei, J., & Lai, X. (2020). 3drte: 3d rotation embedding in temporal knowledge graph . IEEE Access, 8, 207515-207523.

Journal of Web Semantics

[1] (ConT) Ma, Y., Tresp, V., & Daxberger, E. A. (2019). Embedding models for episodic knowledge graphs . Journal of Web Semantics, 59, 100490.

Multi-task recommendation based on dynamic knowledge graph

  • Published: 03 June 2024

Cite this article

temporal knowledge graph representation learning with local and global evolutions

  • Minwei Wen 1 ,
  • Hongyan Mei   ORCID: orcid.org/0009-0005-4067-0150 1 ,
  • Wei Wang 2 ,
  • Xiaorong Xue 1 &
  • Xing Zhang 1  

15 Accesses

Explore all metrics

Introducing knowledge graphs into recommender systems effectively solves sparsity and cold start problems. However, existing KG recommendation methods such as MKR mostly rely on static knowledge graphs, ignoring that nodes and edges in the graph dynamically change over time, leading to problems such as insufficient timeliness, inability to describe context dependencies, data redundancy, and noise. We propose a multi-task learning method for recommendation enhancement based on dynamic knowledge graphs, MTRDKG, which models the dynamic knowledge graph as a series of continuous time events. Specifically, after node events (node-level or interactions between nodes) occur, the memory state of the nodes is updated through a temporal graph network (TGN), and node temporal embeddings are generated to capture the nodes’ attributes, contextual relationships, and dynamic change information. We use the node embeddings generated by TGN in conjunction with the recommended items as a shared part, aiming to integrate dynamic knowledge graph information into the recommendation task, thereby improving the recommendation effect. Extensive experiments were conducted with four real-world datasets and state-of-the-art baseline methods. The results show that MTRDKG outperforms existing methods in terms of recommendation accuracy and knowledge graph embedding quality, especially in dealing with datasets of different sparsity levels.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price includes VAT (Russian Federation)

Instant access to the full article PDF.

Rent this article via DeepDyve

Institutional subscriptions

temporal knowledge graph representation learning with local and global evolutions

Similar content being viewed by others

temporal knowledge graph representation learning with local and global evolutions

Artificial intelligence in recommender systems

temporal knowledge graph representation learning with local and global evolutions

Knowledge Graphs: Opportunities and Challenges

temporal knowledge graph representation learning with local and global evolutions

Advances in Collaborative Filtering

Data availability.

All links to data generated or analysed in this study are included in this published article.

http://snap.stanford.edu/jodie/reddit.csv

http://snap.stanford.edu/jodie/wikipedia.csv

https://grouplens.org/datasets/movielens/1m/

https://grouplens.org/datasets/hetrec-2011/

Wang Q, Mao Z, Wang B, Guo L (2017) Knowledge graph embedding: A survey of approaches and applications. IEEE Trans Knowl Data Eng 29(12):2724–2743

Article   Google Scholar  

Sang L, Xu M, Qian S, Wu X (2021) Knowledge graph enhanced neural collaborative recommendation. Expert Syst Appl 164:113992

Wang H, Zhao M, Xie X, Li W, Guo M (2019) Knowledge graph convolutional networks for recommender systems. In: The world wide web conference, pp 3307–3313

Guo Q, Zhuang F, Qin C, Zhu H, Xie X, Xiong H, He Q (2020) A survey on knowledge graph-based recommender systems. IEEE Trans Knowl Data Eng 34(8):3549–3568

Zhang F, Yuan NJ, Lian D, Xie X, Ma W-Y (2016) Collaborative knowledge base embedding for recommender systems. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp 353–362

Lin Y, Liu Z, Sun M, Liu Y, Zhu, X (2015) Learning entity and relation embeddings for knowledge graph completion. In: Proceedings of the AAAI conference on artificial intelligence, vol 29

Zhou Z, Wang C, Feng Y, Chen D (2022) Jointe: Jointly utilizing 1d and 2d convolution for knowledge graph embedding. Knowl-Based Syst 240:108100

Baghershahi P, Hosseini R, Moradi H (2023) Self-attention presents low-dimensional knowledge graph embeddings for link prediction. Knowl-Based Syst 260:110124

Zhang T, Tian X, Sun X, Yu M, Sun Y, Yu G (2021) Overview on knowledge graph embedding technology research. J Softw 34(1):277–311

Google Scholar  

Lin Q, Mao R, Liu J, Xu F, Cambria E (2023) Fusing topology contexts and logical rules in language models for knowledge graph completion. Inf Fus 90:253–264

Zhang Y, Yang Q (2021) A survey on multi-task learning. IEEE Trans Knowl Data Eng 34(12):5586–5609

Wang H, Zhang F, Zhao M, Li W, Xie X, Guo M (2019) Multi-task feature learning for knowledge graph enhanced recommendation. In: The world wide web conference, pp 2000–2010

Ye Q, Hsieh C-Y, Yang Z, Kang Y, Chen J, Cao D, He S, Hou T (2021) A unified drug-target interaction prediction framework based on knowledge graph and recommendation system. Nat Commun 12(1):6775

Du Y, Zhu X, Chen L, Fang Z, Gao Y (2022) Metakg: Meta-learning on knowledge graph for cold-start recommendation. IEEE Trans Knowl Data Eng

Huang Z, Liu Y, Zhan C, Lin C, Cai W, Chen Y (2021) A novel group recommendation model with two-stage deep learning. IEEE Trans Syst Man Cybern: Syst 52(9):5853–5864

Rossi E, Chamberlain B, Frasca F, Eynard D, Monti F, Bronstein M (2020) Temporal graph networks for deep learning on dynamic graphs. Preprint arXiv:2006.10637

Qin M, Zhang C, Bai B, Zhang G, Yeung D-Y (2023) High-quality temporal link prediction for weighted dynamic graphs via inductive embedding aggregation. IEEE Trans Knowl Data Eng

Wu T, Khan A, Yong M, Qi G, Wang M (2022) Efficiently embedding dynamic knowledge graphs. Knowl-Based Syst 250:109124

Zhou L, Yang Y, Ren X, Wu F, Zhuang Y (2018) Dynamic network embedding by modeling triadic closure process. In: Proceedings of the AAAI conference on artificial intelligence, vol 32

Goyal P, Chhetri SR, Canedo A (2020) dyngraph2vec: Capturing network dynamics using dynamic graph representation learning. Knowl-Based Syst 187:104816

Paudel R, Huang HH (2022) Pikachu: Temporal walk based dynamic graph embedding for network anomaly detection. In: NOMS 2022-2022 IEEE/IFIP network operations and management symposium, pp 1–7. IEEE

Sankar A, Wu Y, Gou L, Zhang W, Yang H (2020) Dysat: Deep neural representation learning on dynamic graphs via self-attention networks. In: Proceedings of the 13th international conference on web search and data mining, pp 519–527

Nguyen GH, Lee JB, Rossi RA, Ahmed NK, Koh E, Kim S (2018) Continuous-time dynamic network embeddings. In: Companion proceedings of the the web conference 2018, pp 969–976

Khoshraftar S, Mahdavi S, An A, Hu Y, Liu J (2019) Dynamic graph embedding via lstm history tracking. In: 2019 IEEE international conference on data science and advanced analytics (DSAA), pp 119–127. IEEE

Zuo Y, Liu G, Lin H, Guo J, Hu X, Wu J (2018) Embedding temporal network via neighborhood formation. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, pp 2857–2866

Pareja A, Domeniconi G, Chen J, Ma T, Suzumura T, Kanezashi H, Kaler T, Schardl T, Leiserson C (2020) Evolvegcn: Evolving graph convolutional networks for dynamic graphs. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 5363–5370

Xu D, Ruan C, Korpeoglu E, Kumar S, Achan K (2020) Inductive representation learning on temporal graphs. In: International conference on learning representations. https://openreview.net/forum?id=rJeW1yHYwH

Vandenhende S, Georgoulis S, Van Gansbeke W, Proesmans M, Dai D, Van Gool L (2021) Multi-task learning for dense prediction tasks: A survey. IEEE Trans Pattern Anal Mach Intell 44(7):3614–3633

Wang Y, Ding W, Zhang R, Li H (2020) Boundary-aware multitask learning for remote sensing imagery. IEEE J Sel Top Appl Earth Obs Remote Sens 14:951–963

Yang E, Pan J, Wang X, Yu H, Shen L, Chen X, Xiao L, Jiang J, Guo G (2023) Adatask: A task-aware adaptive learning rate approach to multi-task learning. In: Proceedings of the AAAI conference on artificial intelligence, vol 37, pp 10745–10753

Wang Y, Zhang J, Zhou X, Zhang Y (2022) Hierarchical aggregation based knowledge graph embedding for multi-task recommendation. In: Asia-Pacific Web (APWeb) and web-age information management (WAIM) joint international conference on web and big data, pp 174–181. Springer

Hu B, Ye Y, Zhong Y, Pan J, Hu M (2022) Transmkr: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation. Neurocomputing 474:107–114

Zhu J, Zhang Y, Wang Y, Liao W, Chen R, Yuan M (2023) Knowledge-enhanced multi-task recommendation in hyperbolic space. Appl Intell 53(23):28694–28710

Zhai H, Zheng W, Ouyang Y, Pan X, Zhang W (2024) Multi-focus image fusion via interactive transformer and asymmetric soft sharing. Eng Appl Artif Intell 133:107967

Huang W, Wu J, Song W, Wang Z (2022) Cross attention fusion for knowledge graph optimized recommendation. Appl Intell 1–10

Gao M, Li J-Y, Chen C-H, Li Y, Zhang J, Zhan Z-H (2023) Enhanced multi-task learning and knowledge graph-based recommender system. IEEE Trans Knowl Data Eng

Zhou Y, Guo J, Song B, Chen C, Chang J, Yu FR (2022) Trust-aware multi-task knowledge graph for recommendation. IEEE Trans Knowl Data Eng

Shu H, Huang J (2023) Multi-task feature and structure learning for user-preference based knowledge-aware recommendation. Neurocomputing 532:43–55

Zeb A, Saif S, Chen J, Haq AU, Gong Z, Zhang D (2022) Complex graph convolutional network for link prediction in knowledge graphs. Expert Syst Appl 200:116796

Kazemi SM, Goel R, Eghbali S, Ramanan J, Sahota J, Thakur S, Wu S, Smyth C, Poupart P, Brubaker M (2020) Time2vec: Learning a vector representation of time

Pennebaker JW, Francis ME, Booth RJ (2001) Linguistic inquiry and word count: Liwc 2001. Mahway: Lawrence Erlbaum Associates 71(2001), p 2001

Cheng H-T, Koc L, Harmsen J, Shaked T, Chandra T, Aradhye H, Anderson G, Corrado G, Chai W, Ispir M et al (2016) Wide & deep learning for recommender systems. In: Proceedings of the 1st workshop on deep learning for recommender systems, pp 7–10

Wang H, Zhang F, Wang J, Zhao M, Li W, Xie X, Guo M (2018) Ripplenet: Propagating user preferences on the knowledge graph for recommender systems. In: Proceedings of the 27th ACM international conference on information and knowledge management, pp 417–426

Zhang Y, Yuan M, Zhao C, Chen M, Liu X (2022) Aggregating knowledge-aware graph neural network and adaptive relational attention for recommendation. Appl Intell 52(15):17941–17953

Li H, Li C, Feng K, Yuan Y, Wang G, Zha H Robust knowledge adaptation for dynamic graph neural networks. Preprint arXiv:2207.10839

Zhang M, Wu S, Yu X, Liu Q, Wang L (2022) Dynamic graph neural networks for sequential recommendation. IEEE Trans Knowl Data Eng 35(5):4741–4753

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (No. 62273170), the Surface Project of Liaoning Provincial Department of Education (No. JYTMS20230869), the Scientific Research Project of Liaoning Provincial Department of Education (Nos. JZL202015404, LJKZ0625), and the Liaoning Provincial Higher Education Innovation Talent Support Project (No. LR2019034).

Author information

Authors and affiliations.

School of Electronics and Information Engineering, Liaoning University of Technology, Jinzhou, 121001, Liaoning, China

Minwei Wen, Hongyan Mei, Xiaorong Xue & Xing Zhang

School of Electrical Engineering, Liaoning University of Technology, Jinzhou, 121001, Liaoning, China

You can also search for this author in PubMed   Google Scholar

Contributions

Minwei Wen: Conceptualization, Methodology, Resources, Validation, Data Curation, Writing- Original Draft, Writing - Review and Editing. Hongyan Mei: Methodology, Software, Formal analysis, Data Curation, Writing - Review and Editing. Wei Wang, Xiaorong Xue and Xing Zhang: Formal analysis, Validation, Writing- Review and Editing.

Corresponding author

Correspondence to Hongyan Mei .

Ethics declarations

Conflict of interest.

We declare that the authors do not have any conflict of interest related to the content of this article.

Ethics approval

This article contains no studies with human participants or animals performed by authors.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Wen, M., Mei, H., Wang, W. et al. Multi-task recommendation based on dynamic knowledge graph. Appl Intell (2024). https://doi.org/10.1007/s10489-024-05548-1

Download citation

Accepted : 19 May 2024

Published : 03 June 2024

DOI : https://doi.org/10.1007/s10489-024-05548-1

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Recommender systems
  • Knowledge graphs
  • Multi-task learning
  • Temporal graph network
  • Find a journal
  • Publish with us
  • Track your research

IMAGES

  1. Diachronic embeddings for temporal knowledge graph completion

    temporal knowledge graph representation learning with local and global evolutions

  2. Temporal Knowledge Graph Representation Learning With Local And Global Evolutions

    temporal knowledge graph representation learning with local and global evolutions

  3. 【论文阅读】Temporal knowledge graph reasoning based on evolutional representation learning

    temporal knowledge graph representation learning with local and global evolutions

  4. Temporal Knowledge Graph Reasoning with Historical Contrastive Learning

    temporal knowledge graph representation learning with local and global evolutions

  5. Illustration of a temporal knowledge graph between persons and countries.

    temporal knowledge graph representation learning with local and global evolutions

  6. Deep learning on dynamic graphs

    temporal knowledge graph representation learning with local and global evolutions

VIDEO

  1. Visualizing Graph Neural Networks with CorGIE: Corresponding a Graph to Its Embedding

  2. 【蜻蜓点水】A Multimodal Translation Based Approach for Knowledge Graph Representation

  3. [phd0708] Temporal Knowledge Graph Extraction and Modeling across Multiple Documents for Health Risk

  4. Time-Series Representation Learning via Temporal and Contextual Contrasting

  5. [rfp0666] TATKC: A Temporal Graph Neural Network for Fast Approximate Temporal Katz Centrality Ranki

  6. A Model of Text Enhanced Knowledge Graph Representation Learning With Mutual Attention

COMMENTS

  1. Temporal knowledge graph representation learning with local and global evolutions

    3.1. Evolution of the temporal knowledge graph. Definition 1 Temporal Knowledge Graph. A temporal knowledge graph (TKG) is defined as a directed graph with timestamps G = (E, R, T), where E and R are the entity set and relation set, respectively. T is the set of valid timestamps. Each knowledge in TKG is represented as a quadruple (s, r, o, τ), in which s ∈ E and o ∈ E are the subject and ...

  2. Temporal knowledge graph representation learning with local and global

    Temporal knowledge graph (TKG) representation learning aims to project entities and relations in TKGs to a low-dimensional vector space while preserving the evolutionary nature of TKGs. Most ...

  3. Temporal knowledge graph representation learning with local and global

    2023. TLDR. A novel Knowledge Evolution with Time Duration (KETD) model is proposed for knowledge representation of temporal knowledge graphs with finer-time granularities and is the first attempt to learn embeddings of time durations and combines them into embeddings of entities and relations to predict potential periodic facts. Expand.

  4. PDF a arXiv:2403.04782v1 [cs.CL] 2 Mar 2024

    2.1. Problem Formulation. A temporal knowledge graph is a directed multi-relational graph containing structured facts. It is usually expressed as G = (E, R, T, F), where E, R, and. T are the sets of entities, relations, and timestamps, respectively, and F ⊂. E × R × E × T is the set of all possible facts.

  5. Local-Global History-aware Contrastive Learning for Temporal Knowledge

    Temporal knowledge graphs (TKGs) have been identified as a promising approach to represent the dynamics of facts along the timeline. The extrapolation of TKG is to predict unknowable facts happening in the future, holding significant practical value across diverse fields. Most extrapolation studies in TKGs focus on modeling global historical fact repeating and cyclic patterns, as well as local ...

  6. Temporal Knowledge Graph Representation Learning

    PDF (PC) 2094. Abstract: As a structured form of human knowledge,knowledge graphs have played a great supportive role in supporting the semantic intercommunication of massive,multi-source,heterogeneous data,and effectively support tasks such as data analysis,attracting the attention of academia and industry.At present,most knowledge graphs are ...

  7. Temporal Knowledge Graph Reasoning Based on Evolutional Representation

    Temporal Knowledge Graph Reasoning Based on Evolutional Representation Learning. Pages 408-417. Previous Chapter ... we propose a novel Recurrent Evolution network based on Graph Convolution Network (GCN), called RE-GCN, which learns the evolutional representations of entities and relations at each timestamp by modeling the KG sequence ...

  8. A Survey on Temporal Knowledge Graph: Representation Learning and

    Knowledge graphs (KGs) describe the real world with structured facts. A fact consists of two entities and a relation connecting them, which can be formally represented as a triple (head, relation, tail), and an instance of a fact is (Barack Obama, make statement, Iran).Knowledge graph representation learning (KGRL) [] seeks to learn the low-dimentional vector embeddings of entities and ...

  9. Temporal knowledge graph representation learning with local and global

    EvoExplore is a framework that projects entities and relations in temporal knowledge graph (TKG) to low-dimensional vector space while preserving the evolutionary nature of TKG. It captures the local structure evolution as the formation process of graph structure and the global structure evolution as the dynamic community partition of TKG.

  10. LogE-Net: Logic Evolution Network for Temporal Knowledge Graph

    Temporal knowledge graph representation learning with local and global evolutions. Abstract . Temporal knowledge graph (TKG) representation learning aims to project entities and relations in TKGs to a low-dimensional vector space while preserving the evolutionary nature of TKGs. Most existing methods treat knowledge that ...

  11. Temporal knowledge graph representation learning with local and global

    AbstractTemporal knowledge graph (TKG) representation learning aims to project entities and relations in TKGs to a low-dimensional vector space while preserving the evolutionary nature of TKGs. Mos...

  12. PDF LogE-Net: Logic Evolution Network for Temporal Knowledge Graph Forecasting

    At the same time, the logic memory module is introduced to assist reasoning based on long-term memory. In general, the contributions of our work are as follows: Aiming at scenarios in temporal knowledge graph forecasting, a logic rule-based reasoning model, logE-Net, is proposed to assist evolution embedding representation in making decisions ...

  13. PDF Learning Temporal and Spatial Embedding for Temporal Knowledge Graph

    Abstract. Temporal knowledge graphs store a large number of tempo-ral facts that simulate the dynamic interactions of entities along the timeline. Since existing temporal knowledge graphs often suffer from incompleteness, it is crucial to build time-aware representation learn-ing models that help to infer the missing temporal facts. However, most

  14. Temporal Knowledge Graph Entity Alignment via Representation Learning

    2.1 Temporal Knowledge Graph Embedding. In recent years, the representation learning of static KGs has been extensively studied. Some examples of typical embedding models include TransE [], DistMult [], RESCAL [], ComplEx [] and QuatE [].These methods have been proved to be effective for the completion task of static KGs, but cannot model the temporal evolution of KGs due to their neglect of ...

  15. A Survey on Temporal Knowledge Graph: Representation Learning and

    This paper reviews the methods and applications of temporal knowledge graph representation learning, which incorporates time information into the standard knowledge graph framework. It proposes a taxonomy based on the core technologies and provides an in-depth analysis of different methods in each category.

  16. Complex Evolutional Pattern Learning for Temporal Knowledge Graph

    A Temporal Knowledge Graph (TKG) is a sequence of KGs corresponding to different timestamps. TKG reasoning aims to predict potential facts in the future given the historical KG sequences. One key of this task is to mine and understand evolutional patterns of facts from these sequences. The evolutional patterns are complex in two aspects, length ...

  17. StreamE: Learning to Update Representations for Temporal Knowledge

    Temporal knowledge graph representation learning with local and global evolutions. Knowl. Based Syst., Vol. 251 (2022), 109234. Google Scholar Digital Library; Ying Zhou, Xuanang Chen, Ben He, Zheng Ye, and Le Sun. 2022. Re-thinking Knowledge Graph Completion Evaluation from an Information Retrieval Perspective. In SIGIR '22: The 45th ...

  18. Papers on Temporal Knowledge Graph Embedding and Reasoning

    A Survey on Temporal Knowledge Graph: Representation Learning and Applications, ArXiv, 2024. paper Cai, Li, Xin Mao, ... "Temporal knowledge graph representation learning with local and global evolutions". Knowledge-Based Systems 2022. Github [2] (TuckERT) Pengpeng Shao, Dawei Zhang, Guohua Yang, Jianhua Tao, Feihu Che, Tong Liu. ...

  19. Temporal Knowledge Graph Reasoning Based on Evolutional Representation

    the temporal reasoning tasks on six benchmark datasets. Especially, it achieves up to 11.46% improvement in MRR for entity predic-tion with up to 82 times speedup comparing to the state-of-the-art baseline. CCS CONCEPTS • Computing methodologies →Temporal reasoning. KEYWORDS Temporal knowledge graph, evolutional representation learning,

  20. Learning Temporal and Spatial Embedding for Temporal Knowledge Graph

    Abstract. Temporal knowledge graphs store a large number of temporal facts that simulate the dynamic interactions of entities along the timeline. Since existing temporal knowledge graphs often suffer from incompleteness, it is crucial to build time-aware representation learning models that help to infer the missing temporal facts.

  21. Temporal Knowledge Graph Reasoning Based on Evolutional Representation

    Knowledge Graph (KG) reasoning that predicts missing facts for incomplete KGs has been widely explored. However, reasoning over Temporal KG (TKG) that predicts facts in the future is still far from resolved. The key to predict future facts is to thoroughly understand the historical facts. A TKG is actually a sequence of KGs corresponding to different timestamps, where all concurrent facts in ...

  22. Large-scale knowledge graph representation learning

    The knowledge graph emerges as powerful data structures that provide a deep representation and understanding of the knowledge presented in networks. In the pursuit of representation learning of the knowledge graph, entities and relationships undergo an embedding process, where they are mapped onto a vector space with reduced dimensions. These embeddings are progressively used to extract their ...

  23. Unveiling LLMs: The Evolution of Latent Representations in a Temporal

    Following the framework proposed by Ghandeharioun et al. ( 2024), we leverage the activation patching technique to extract factual information embedded within the latent representations of LLMs and represent its evolution across the layers using a temporal knowledge graph. The procedure is outlined in Figure 1.

  24. Temporal Graph Rewiring with Expander Graphs

    Temporal graph learning (TGL) (Kazemi et al.,2020) emerged from graph representation learning to study evolv-ing relationships in temporal graph data. Given the dynamic nature of temporal graphs, temporal graph neural networks (TGNNs) (Longa et al.,2023) are developed to capture the evolution of graph topology introducing novel components

  25. Multi-task recommendation based on dynamic knowledge graph

    Introducing knowledge graphs into recommender systems effectively solves sparsity and cold start problems. However, existing KG recommendation methods such as MKR mostly rely on static knowledge graphs, ignoring that nodes and edges in the graph dynamically change over time, leading to problems such as insufficient timeliness, inability to describe context dependencies, data redundancy, and ...