Inductive Representation Learning in Temporal Networks via Causal Anonymous Walks
Abstract
Temporal networks serve as abstractions of many real-world dynamic systems. These networks typically evolve according to certain laws, such as the law of triadic closure, which is universal in social networks. Inductive representation learning of temporal networks should be able to capture such laws and further be applied to systems that follow the same laws but have not been unseen during the training stage. Previous works in this area depend on either network node identities or rich edge attributes and typically fail to extract these laws. Here, we propose Causal Anonymous Walks (CAWs) to inductively represent a temporal network. CAWs are extracted by temporal random walks and work as automatic retrieval of temporal network motifs to represent network dynamics while avoiding the time-consuming selection and counting of those motifs. CAWs adopt a novel anonymization strategy that replaces node identities with the hitting counts of the nodes based on a set of sampled walks to keep the method inductive, and simultaneously establish the correlation between motifs. We further propose a neural-network model CAW-N to encode CAWs, and pair it with a CAW sampling strategy with constant memory and time cost to support online training and inference. CAW-N is evaluated to predict links over 6 real temporal networks and outperforms previous SOTA methods in the inductive setting. CAW-N also outperforms previous methods in 4 out of the 6 networks in the transductive setting.
1 Introduction
Temporal networks consider dynamically interacting elements as nodes, interactions as temporal links, with labels of when those interactions happen. Such temporal networks provide abstractions to study many real-world dynamic systems (Holme & Saramäki, 2012). Researchers have investigated temporal networks in recent several decades and concluded many insightful laws that essentially reflect how these real-world systems evolve over time (Kovanen et al., 2011; Benson et al., 2016; Paranjape et al., 2017; Zitnik et al., 2019). For example, the law of triadic closure in social networks, describing that two nodes with common neighbors tend to have a mutual interaction later, reflects how people establish social connections (Simmel, 1950). Later, a more elaborate law on the correlation between the interaction frequency between two individuals and the degree that they share social connections, further got demonstrated (Granovetter, 1973; Toivonen et al., 2007). Feedforward control loops that consist of a direct interaction (from node to node ) and an indirect interaction (from through another node to ), also work as a law in the modulation of gene regulatory systems (Mangan & Alon, 2003) and also as the control principles of many engineering systems (Gorochowski et al., 2018). Although research on temporal networks has achieved the above success, it can hardly be generalized to study more complicated laws: Researchers have to investigate an exponentially increasing number of patterns when incorporating more interacting elements let alone their time-evolving aspects.
Recently, representation learning, via learning vector representations of data based on neural networks, has offered unprecedented possibilities to extract, albeit implicitly, more complex structural patterns (Hamilton et al., 2017b; Battaglia et al., 2018). However, as opposed to the study on static networks, representation learning of temporal networks is far from mature. Two challenges on temporal networks have been frequently discussed. First, the entanglement of structural and temporal patterns requires an elegant model to digest the two-side information. Second, the model scalability becomes more crucial over temporal networks as new arriving links need to be processed timely while a huge link set due to the repetitive links between two nodes needs to be digested simultaneously.
In contrast to the above two challenges, another challenge, the inductive capability of the temporal-network representation, is often ignored. However, it is equally important if not more, as the inductive capability indicates whether the models indeed capture the dynamic laws of the systems and can be further generalized to the system that share the same laws but have not been used to train these models. These laws may only depend on structures such as the triadic closure or feed-forward control loops as aforementioned. These laws may also correlate with node attributes, such as interactions between people affected by their gender and age (Kovanen et al., 2013). But in both cases, the laws should be independent from network node identities. Although previous works tend to learn inductive models by removing node identities (Trivedi et al., 2019; Xu et al., 2020), they run into other issues to inductively represent the dynamic laws, for which we leave more detailed discussion in Sec. 2.


Here we propose Causal Anonymous Walks (CAW) for modeling temporal networks. Our idea for inductive learning is inspired by the recent investigation on temporal network motifs that correspond to connected subgraphs with links that appear within a restricted time range (Kovanen et al., 2011; Paranjape et al., 2017). Temporal network motifs essentially reflect network dynamics: Both triadic closure and feed-forward control can be viewed as temporal network motifs evolving (Fig. 1); An inductive model should predict the 3rd link in both cases when it captures the correlation of these two links as they share a common node, while the model is agnostic to the node identities of these motifs.
Our CAW model has two important properties (Fig. 2): (1) Causality extraction — a CAW starts from a link of interest and backtracks several adjacent links over time to encode the underlying causality of network dynamics. Each walk essentially gives a temporal network motif; (2) Set-based anonymization — CAWs remove the node identities over the walks to guarantee inductive learning while encoding relative node identities based on the counts that they appear at a certain position according to a set of sampled walks. Relative node identities guarantee that the structures of motifs and their correlations are still kept after removing node identities. To predict temporal links between two nodes of interest, we propose a model CAW-Network (CAW-N) that samples a few CAWs related to the two nodes of interest, encodes and aggregates these CAWs via RNNs (Rumelhart et al., 1986) and set-pooling respectively to make the prediction.
Experiments show that CAW-N is extremely effective. CAW-N does not need to enumerate the types of motifs and count their numbers that have been used as features to predict network dynamics (Lahiri & Berger-Wolf, 2007; Rahman & Al Hasan, 2016; Rossi et al., 2019; AbuOda et al., 2019; Li & Milenkovic, 2017), which significantly saves feature-engineering effort. CAW-N also keeps all fine-grained temporal information along the walks that may be removed by directly counting motifs (Ahmed et al., 2015; Paranjape et al., 2017). CAWs share a similar idea as anonymous walks (AW) (Micali & Zhu, 2016) to remove node identities. However, AWs have only been used for entire static graph embedding (Ivanov & Burnaev, 2018) and are not directedly applied to represent temporal networks: AWs cannot capture causality; AWs get anonymized based on each single walk and hence lose the correlation between network motifs. In contrast, CAWs capture all the information, temporal, structural, motif-correlation that are needed, to represent temporal networks.
We conclude our contributions in three-folds: (1) A novel approach to represent temporal network CAW-N is proposed, which leverages CAWs to encode temporal network motifs to capture network dynamics while keeping fully inductive. CAW-N is evaluated to predict links over 6 real-world temporal networks. CAW-N outperforms all SOTA methods by about 15% averaged over 6 networks in the inductive setting and also significantly beat all SOTA methods over 5 networks in the transductive setting; (2) CAW-N significantly decreases the feature-engineering effort in traditional motif selection and counting approaches and keeps fine-grained temporal information; (3) CAW-N is paired with a CAW sampling method with constant memory and time cost, which conduces to online learning.
2 Related Work
Prior work on representation learning of temporal networks preprocesses the networks by simply aggregating the sequence of links within consecutive time windows into network snapshots, and use graph neural networks (GNN) (Scarselli et al., 2008; Kipf & Welling, 2017) and RNNs or transformer networks (Vaswani et al., 2017) to encode structural patterns and temporal patterns respectively (Pareja et al., 2020; Manessi et al., 2020; Goyal et al., 2020; Hajiramezanali et al., 2019; Sankar et al., 2020). The main drawback of these approaches is that they need to predetermine a time granularity for link aggregation, which is hard to learn structural dynamics in different time scales. Therefore, approaches that work on link streams directly have been recently proposed (Trivedi et al., 2017; 2019; Kumar et al., 2019; Xu et al., 2020). Know-E (Trivedi et al., 2017), DyRep (Trivedi et al., 2019) and JODIE (Kumar et al., 2019) use RNNs to propagate messages across interactions to update node representations. Know-E, JODIE consider message exchanges between two directly interacted nodes while DyRep considers an additional hop of interactions. Therefore, DyRep gives a more expressive model at a cost of high complexity. TGAT (Xu et al., 2020) in contrast mimics GraphSAGE (Hamilton et al., 2017a) and GAT (Veličković et al., 2018) to propagate messages in a GNN-like way from sampled historical neighbors of a node of interest. TGAT’s sampling strategy requires to store all historical neighbors, which is unscalable for online learning. Our CAW-N directly works on link streams and only requires to memorize constant many most recent links for each node.

Most of the above models are not inductive because they associate each node with an one-hot identity (or the corresponding row of the adjacency matrix, or a free-trained vector) (Li et al., 2018; Chen et al., 2019; Kumar et al., 2019; Hajiramezanali et al., 2019; Sankar et al., 2020; Manessi et al., 2020; Goyal et al., 2020). TGAT (Xu et al., 2020) claimed to be inductive by removing node identities and just encoding link timestamps and attributes. However, TGAT was only evaluated over networks with rich link attributes, where the structural dynamics is not captured essentially: If we focus on structural dynamics only, it is easy to show a case when TGAT confuses node representations and will fail: Suppose in the history, two node pairs and only interact within each pair but share the timestamps (Fig. 3). Intuitively, a proper model should predict that future links still appear within each pair. However, TGAT cannot distinguish v.s. , and v.s. , which leads to incorrect prediction. Note that GraphSAGE (Hamilton et al., 2017a) and GAT (Veličković et al., 2018) also share the similar issue when representing static networks for link prediction (Zhang et al., 2020; Srinivasan & Ribeiro, 2019). DyRep (Trivedi et al., 2019) is able to relieve such ambiguity by merging node representations with their neighbors’ via RNNs. However, when DyRep runs over a new network, it frequently encounters node representations unseen during its training and will fail to make correct prediction. Our CAW-N removes node identities and leverages relative node identities to avoid the issue in Fig. 3. Detailed explanations are given in Sec.4.2.
Network-embedding approaches may also be applied to temporal networks (Zhou et al., 2018; Du et al., 2018; Mahdavi et al., 2018; Singer et al., 2019; Nguyen et al., 2018). However, they directly assign each node with a learnable vector. Therefore, they are not inductive and cannot digest attributes.
3 Problem Formulation and Notations
Problem Formulation. A temporal network can be represented as a sequence of links that come in over time, i.e. where is a link and is the timestamp showing when arrives. Each link corresponds to a dyadic event between two nodes . For simplicity, we first assume those links to be undirected and without attributes while later we discuss how to generalized our method to directed attributed networks. The sequence of links encodes network dynamics. Therefore, the capability of a model for representation learning of temporal networks is typically evaluated by how accurately it may predict future links based on the historical links (Sarkar et al., 2012). In this work, we also use link prediction as the metric. Note that we care not only the link prediction between the nodes that have been seen during the training. We also expect the models to predict links between the nodes that has never been seen as the inductive evaluation.
Notations. We define to include the links attached to a node before certain time . A walk (reverse over time) on temporal networks can be represented as
(1) |
We use to denote the th node-time pair, and and to denote the corresponding node and time in correspondingly. Later, we also use as vector concatenation.
4 Proposed Method: Causal Anonymous Walk-Network
4.1 Preliminaries: Anonymous Walk and Temporal Network Motif
Anonymous walks were first considered by Micali & Zhu (2016) to study the reconstruction of a Markov process from the records without sharing a common “name space”. AWs can be directly rephrased in the network context. Specifically, an AW starts from a node, performs random walks over the graph to collect a walk of nodes, e.g. . AW has an important anonymization step to replace the node identities by the orders of their appearance in each walk, which we term relative node identities in AW and define it as
(2) |
Note that the set operation above removes duplicated elements. Although AW removes node identities, those nodes are still distinguishable within this walk. Therefore, an AW can be viewed as a network motif while the information on which specific nodes form this motif is removed. Examples of AWs are shown as follows. While these are two different walks, they may be mapped to the same AW when node identities get removed.
4.2 Causal Anonymous Walk
We propose CAW that shares the high-level concept with AW to remove the original node identities. However, CAW has a different causal sampling strategy and a novel set-based approach for node anonymization, which are specifically designed to encode temporal network dynamics (Fig. 2).
Causality Extraction. Alg. 1 shows our causal sampling: We sample connected links by backtracking over time to extract the underlying causality of network dynamics. More recent links may be more informative and thus we introduce a non-negative hyper-parameter to sample a link with a probability proportional to where are the timestamps of this link and the link previously sampled respectively. A large can emphasize more on recent links while zero leads to uniform sampling. In Sec.4.4, we will discuss an efficient sampling strategy for the step 5 in Alg.1, which avoids computing those probabilities by visiting the entire historical links.
Then, given a link and a time , we use Alg. 1 to collect many -step walks starting from both and , and record them in and respectively. For convenience, a walk from a starting node can be represented as Eq.1.
Set-based Anonymization. Based on and , we may anonymize each node identity that appears on at least one walk in and design relative node identity for . Our design has the following consideration. (Eq.2) only depends on a single path, which results from the original assumption that any two AWs do not even share the name space (i.e., node identities) (Micali & Zhu, 2016). However, in our case, node identities are actually accessible, though an inductive model is not allowed to use them directly. Instead, correlation across different walks could be a key to reflect laws of network dynamics: Consider the case when the link happens only if there is another node appearing in multiple links connected to (Fig. LABEL:fig:AW-issue). Therefore, we propose to use node identities to first establish such correlation and then remove the original identities.
for
Specifically, we define as follows: For , let count the times in node appears at certain positions, i.e., for . Further, define
(3) |
Essentially, and encode the correlation between walks within and respectively and the set operation in Eq.3 establishes the correlation across and . For the case in Fig.3, suppose include all historical one-step walks (before ) starting from respectively. Then, it is easy to show that that allows differentiating and , while TGAT (Xu et al., 2020) as discussed in Sec.2 fails. From the network-motif point of view, not only encodes each network motif that corresponds to one single walk as does but also establishes the correlation among these network motifs. cannot establish the correlation between motifs and will also fail to distinguish and in Fig.3. We see it as a significant breakthrough as such correlation often gets neglected in previous works that directly count motifs or adopt AW-type anonymization .
Later, we use for simplicity when the reference set can be inferred from the context. Then, each walk (Eq.1) can be anonymized as
(4) |
The following theorem indicates that does not depend on node identities to guarantee the inductive property of the models, which can be easily justified.
Theorem 4.1.
For two pairs of walk sets and , if there exists a bijective mapping between node identities such that each walk in can be bijectively mapped to one walk in according to for all . Then for all nodes that appear in at least one walk in .
4.3 Neural Encoding for Causal Anonymous Walks
After we collect CAWs, neural networks can be conveniently leveraged to extract their structural (in ) and fine-grained temporal information by encoding CAWs: We will propose the model CAW-N to first encode each walk (Eq.4) and then aggregate all encoded walks in .
Encode . Note that each walk is a sequence of node-time pairs. If we encode each node-time pair and plug those pairs in a sequence encoder, e.g., RNNs, we obtain the encoding of :
(5) |
where are two encoding function on and respectively. One may use transformer networks instead of RNNs to encode the sequences but as the sequences in our case are not long , RNNs have achieved good enough performance. Now, we specify the two encoding functions and as follows. Recall the definition of (Eq.3).
(6) |
Here the encoding of adopts the sum-pooling as the order of is not relevant. For , we adopt random Fourier features to encode time (Xu et al., 2019; Kazemi et al., 2019) which may approach any positive definite kernels according to the Bochner’s theorem (Bochner, 1992).
(7) |
Encode . After encoding each walk in , we aggregate all these walks to obtain the final representation for prediction. We suggest to use either mean-pooling for algorithmic efficiency or self-attention (Vaswani et al., 2017) followed by mean-pooling to further capture subtle interactions between different walks. Specifically, suppose are the CAWs in and each . We set as
-
•
Mean-AGG(): .
-
•
Self-Att-AGG(): where are two learnable parameter matrices.
We add a 2-layer perceptron over to make the final link prediction.
4.4 Extension and Discussion
Attributed nodes/links and directed links. In some real networks, nodes or links may have attributes available, e.g., the message content in the case of SMS networks. In this case, the walk in Eq.1 can be associated with node/link attributes where refers to the attributes on link or on the node or a concatenation of these two parts. Note that the direction of a link can also be viewed as a binary link attribute, where a 2-dimensional one-hot encoding can be used. To incorporate such information, we only need to change (Eq.5) as
(8) |
The derived encoding is then concatenated with to obtain the enhanced final encoding of .
Efficient link sampling. A naive implementation of the link sampling in step 5 in Alg.1 is to compute and normalize the sampling probabilities of all links in , which requires to memorize all historical links and costs much time and memory. To solve this problem, we propose a sampling strategy (Appendix A) with expected time and memory complexity if links in come in by following a Poisson process with intensity (Last & Penrose, 2017). This means that our sampling strategy with a positive allows the model only recording recent links for each node instead of the entire history. Our experiments in Sec.5.3 show that that achieves the best prediction performance makes 5 in different datasets. Since the time and memory complexity do not increase with respect to the number of links, our model can be used for online training and inference. Note that the case reduces to uniform sampling, mostly adopted by previous methods (Xu et al., 2020), which requires to record the entire history and thus is not scalable.
Measurement | Wikipedia | MOOC | Enron | Social Evo. | UCI | |
---|---|---|---|---|---|---|
nodes & temporal links | 10,985 & 672,447 | 9,227 & 157,474 | 7145 & 411,749 | 184 & 125,235 | 74 & 2,099,520 | 1,899 & 59,835 |
attributes for nodes & links | 172 & 172 | 172 & 172 | 0 & 4 | 0 & 0 | 0 & 0 | 0 & 0 |
avg. link stream intensity |
5 experiments
5.1 Experimental Setup
CAW-N Variants. We test CAW-N-mean and CAW-N-attn which uses mean and attention pooling respectively to encode (Sec. 4.3). Their code is provided in the supplement.
Baselines. Our method is compared with six previous state-of-the-art baselines on representation learning of temporal networks. They can be grouped into two categories based on their input data structure: (1) Snapshot-based methods, including DynAERNN (Goyal et al., 2020), VGRNN (Hajiramezanali et al., 2019) and EvolveGCN (Pareja et al., 2020); (2) Stream-based methods, including TGAT (Xu et al., 2020), JODIE (Kumar et al., 2019) and DyRep (Trivedi et al., 2019). We give their detailed introduction in Appendix C.2.2. For the snapshot-based methods, we view the link aggregation as a way to preprocess historical links. We adopt the aggregation ways suggested in their papers. These models are trained and evaluated over the same link sets as the stream-based methods.
Dataset. We use six real-world public datasets: Wikipedia is a network between wiki pages and human editors. Reddit is a network between posts and users on subreddits. MOOC is a network of students and online course content units. Social Evolution is a network recording the physical proximity between students. Enron is a email communication network. UCI is a network between online posts made by students. We summarize their statistics in Tab.1 and give their detailed description and access in Appendix C.1.
Evaluation Tasks. Two types of tasks are for evaluation: transductive and inductive link prediction.
Transductive link prediction task allows temporal links between all nodes to be observed up to a time point during the training phase, and uses all the remaining links after that time point for testing. In our implementation, we split the total time range [0, ] into three intervals: [0, ), [, ), [, ]. links occurring within each interval are dedicated to training, validation, and testing set, respectively. For all datasets, we fix /=0.7, and /=0.85 .
Inductive link prediction task predicts links associated with nodes that are not observed in the training set. There are two types of such links: 1) "old vs. new" links, which are links between an observed node and an unobserved node; 2) "new vs. new" links, which are links between two unobserved nodes. Since these two types of links suggest different types of inductiveness, we distinguish them by reporting their performance metrics separately. In practice, we follow two steps to split the data: 1) we use the same setting of the transductive task to first split the links chronologically into training / validation / testing sets; 2) we randomly select 10% nodes, remove any links associated with them from the training set, and remove any links not associated with them in the validation and testing sets.
Following most baselines, we randomly sample an equal amount of negative links and consider link prediction as a binary classification problem. For fair comparison, we use the same evaluation procedures for all baselines, including the snapshot-based methods.
Training configuration. We use binary cross entropy loss and Adam optimizer to train all the models, and early stopping strategy to select the best epoch to stop training. For hyperparameters, we primarily tune those that control the CAW sampling scheme including the number , the length of CAWs and the time decay . We will investigate their sensitivity in Sec.5.3. For all baselines, we adapt their implemented models into our evaluation pipeline and extensively tune them. Detailed description of all models’ tuning can be found in Appendix C. Finally, we adopt two metrics to evaluate the models’ performance: Area Under the ROC curve (AUC) and Average Precision (AP).
5.2 Results and Discussion
Task | Methods | Wikipedia | MOOC | Social Evo. | Enron | UCI | ||
---|---|---|---|---|---|---|---|---|
Inductive | new v.s. new | DynAERNN | 57.51 2.54 | 55.16 1.15 | 60.85 1.61 | 52.00 0.16 | 51.57 2.63 | 50.20 2.78 |
JODIE | 72.49 0.38 | 70.78 0.75 | 80.04 0.28† | 87.66 0.12† | 73.99 2.54† | 64.77 0.75 | ||
DyRep | 62.37 1.49 | 67.07 1.26 | 74.07 1.88 | 83.92 0.02 | 69.74 0.44 | 63.76 4.67 | ||
VGRNN | 61.93 0.72 | 60.64 0.68 | 63.01 0.29 | 54.49 1.21 | 72.01 0.08 | 61.35 1.10 | ||
EvolveGCN | 63.31 0.53 | 58.01 0.16 | 52.31 4.14 | 46.95 0.85 | 42.53 2.12 | 76.65 0.63† | ||
TGAT | 94.96 0.88† | 93.53 0.84† | 70.10 0.35 | 53.27 1.16 | 63.34 2.95 | 76.36 1.48 | ||
CAW-N-mean | 97.67 0.08∗ | 99.27 0.08∗ | 90.52 0.54∗ | 95.55 0.40∗ | 96.43 0.39∗ | 93.14 0.32∗ | ||
CAW-N-attn | 97.71 0.08∗ | 99.29 0.10∗ | 90.86 0.56∗ | 95.20 0.68∗ | 96.98 0.41∗ | 93.16 0.25∗ | ||
new v.s. old | DynAERNN | 58.79 3.01 | 57.97 2.38 | 80.99 1.35 | 52.31 0.59 | 54.36 1.48 | 52.26 1.36 | |
JODIE | 76.33 0.03 | 74.65 0.06 | 87.40 1.71 | 91.80 0.01† | 85.24 0.08 | 69.95 0.11 | ||
DyRep | 66.13 1.07 | 76.72 0.19 | 88.23 1.20† | 87.98 0.45 | 94.39 0.32† | 93.28 0.96† | ||
VGRNN | 54.11 0.74 | 62.93 0.69 | 60.10 0.88 | 64.66 0.41 | 76.33 0.05 | 62.39 1.08 | ||
EvolveGCN | 65.61 0.37 | 56.29 2.17 | 50.20 1.92 | 50.73 1.36 | 42.53 2.13 | 70.78 0.22 | ||
TGAT | 97.25 0.18† | 95.47 0.17† | 69.30 0.08 | 54.22 1.28 | 58.76 1.18 | 74.19 0.88 | ||
CAW-N-mean | 97.70 0.17∗ | 98.77 0.13∗ | 91.08 0.20∗ | 91.52 0.21 | 93.98 0.44 | 91.15 0.24 | ||
CAW-N-attn | 97.56 0.14∗ | 98.21 0.17∗ | 90.40 0.13∗ | 91.55 0.30 | 93.99 0.62 | 91.17 0.26 | ||
Transductive | DynAERNN | 83.37 1.48 | 71.00 1.10 | 89.34 0.24 | 67.78 0.80 | 63.11 1.13 | 83.72 1.79 | |
JODIE | 87.71 0.02 | 88.43 0.02 | 90.50 0.01† | 89.78 0.04 | 89.36 0.06 | 74.63 0.11 | ||
DyRep | 67.36 1.23 | 77.40 0.13 | 90.49 0.03 | 90.85 0.01† | 96.71 0.04† | 95.23 0.25† | ||
VGRNN | 51.89 0.92 | 71.20 0.65 | 90.03 0.32 | 72.84 0.73 | 79.08 0.02 | 89.43 0.27 | ||
EvolveGCN | 58.42 0.52 | 60.48 0.47 | 50.36 0.85 | 60.36 0.65 | 74.02 0.31 | 78.30 0.22 | ||
TGAT | 96.65 0.06† | 96.36 0.05† | 72.09 0.29 | 56.63 0.55 | 60.88 0.37 | 77.67 0.27 | ||
CAW-N-mean | 98.07 0.06∗ | 98.83 0.11∗ | 94.28 0.13∗ | 93.88 0.72∗ | 91.64 0.15 | 95.25 0.30∗ | ||
CAW-N-attn | 98.05 0.09∗ | 98.96 0.10∗ | 94.33 0.08∗ | 92.54 0.58∗ | 92.06 0.16 | 95.14 0.25 |
We report AUC scores in Tab.2, and report AP scores in Tab.6 of Appendix D.1. In the inductive setting and especially with "new vs new" links, our models significantly outperform all baselines on all datasets. Even in transductive setting where the baselines claim their primary contribution, our approaches still significantly outperform them on four out of six datasets. Note that our models achieve almost perfect scores (i.e. ) on Reddit and Wikpedia when the baselines are far from perfect. Meanwhile, the strongest baseline on these two attributed datasets, TGAT (Xu et al., 2020), suffers a lot on all the other datasets where informative node / link attributes become unavailable.
Comparing the two training settings, we observe that the performance of four baselines (JODIE, VGRNN, DynAERNN,DyRep) drop significantly when transiting from the transductive setting to the inductive one, as they mostly record node identities either explicitly (JODIE, VGRNN, DynAERNN) or implicitly (DyRep). TGAT and EvolveGCN do not use node identities and thus their performance gaps between the two settings are small, while they sometimes do not perform well in the transductive setting, as they encounter the ambiguity issue in Fig. 3. In contrast, our methods perform well in both the transductive and inductive settings. We attribute this superiority to the anonymization procedure: the set-based relative node identities well capture the correlation between walks to make good prediction while removing the original node identities to keep entirely inductive. Even when the network structures greatly change and new nodes come in as long as the network evolves according to the same law as the network used for training, CAW-N will always work. Comparing CAW-N-mean and CAW-N-attn, we see that our attention-based variant outperforms the mean-pooling variant, albeit at the cost of high computation complexity. Also note that the strongest baselines on all datasets are stream-based methods, which indicates that the aggregation of links into network snapshots may remove some useful time information (see more discussion in Appendix D.2).
No. | Ablation | Wikipedia | UCI | Social Evo. |
---|---|---|---|---|
1. | original method (CAW-N-mean) | 98.94 0.11 | 91.88 0.26 | 91.97 0.23 |
2. | remove | 96.28 0.66 | 79.45 0.71 | 53.69 0.21 |
3. | remove | 97.90 0.17 | 83.91 0.39 | 68.25 1.99 |
4. | remove , | 88.94 0.85 | 50.01 0.02 | 50.00 0.00 |
5 | replace by one-hot | 96.55 0.21 | 85.59 0.34 | 68.47 0.86 |
6. | fix | 74.38 2.05 | 80.44 0.56 | 77.72 0.44 |
We further conduct ablation studies on Wikipedia (attributed), UCI (non-attributed), and Social Evolution (non-attributed), to validate effectiveness of critical components of our model. Tab. 3 shows the results. By comparing Ab.1 with Ab.2, 3 and 4 respectively, we observe that our proposed node anonymization and encoding, , contributes most to the performance, though the time encoding also helps. Comparing performance across different datasets, we see that the impact of ablation is more prominent when informative node/link attributes are unavailable such as with UCI and Social Evolution. Therefore, in such scenarios our CAWs are highly crucial. In Ab.5, we replace our proposed with (Eq.2), which is used in standard AWs, and we use one-hot encoding of node new identities . We see by comparing Ab.1, 2, and 5 that such anonymization process is significantly less effective than our , though it helps to some extent. Finally, Ab.6 suggests that entirely uniform sampling of the history may hurt performance.
5.3 Hyperparameter Investigation of CAW Sampling

We systematically analyze the effect of hyperparameters used in CAW sampling schemes, including sampling number , temporal decay coefficient and walk length . The experiments are conducted on UCI and Wikipedia datasets using CAW-N-mean. When investigating each hyperparameter, we set the rest two to an optimal value found by grid search, and report the mean AUC performance on all inductive test links (i.e. old vs. new, new vs. new) and their 95% confidence intervals.
The results are summarized in Fig.5. From (a), we observe that only a small number of sampled CAWs are needed to achieve a competitive performance. Meanwhile, the performance gain is saturated as the sampling number increases. We analyze the temporal decay in (b): usually has an optimal interval, whose values and length also vary with different datasets to capture the different levels of temporal dynamics; a small suggests an almost uniform sampling of interaction history, which hurts the performance; an overly large also damages the model, since it makes the model only sample the most recent few interactions for computation and blind to the rest. Based on our efficient sampling strategy (Sec.4.4), we may combine the optimal with the average link intensity (Tab.1), and concludes that CAW-N only needs to online record and sample from about a constant times about 5 () most recent links for each node. Plot (c) suggests that the performance may peak at a certain CAW length, while the exact value may vary with datasets. Longer CAWs indicate that the corresponding networks evolve according to more complicated laws encoded in higher-order motifs.
5.4 Complexity Evaluation
We examine how the runtime of CAW-N depends on the number of edges used for training. We record the runtimes of CAW-N for training one epoch on the Wikipedia datasets using with batch-size=32. Specifics of the computing infrastructure are given in Appendix C.5. Fig. 6 (a) shows the accumulated runtime of executing the random walk extraction i.e. Alg.1 only. It well aligns with our theoretical analysis (Thm. A.2) that each step of the random walk extraction has constant complexity (i.e. accumulated runtime linear with ). Plot (b) shows the entire runtime for one-epoch training, which is also linear with . Note that is the time complexity that one at least needs to pay. The study demonstrates our method is scalable to long edge streams.

6 Conclusion
We proposed CAW-N to inductively represent the dynamics of temporal networks. CAW-N uses CAWs to implicitly extract network motifs via temporal random walks and adopts novel set-based anonymization to establish the correlation between network motifs. The success of CAW-N points out many promising future research directions on temporal networks: Pairing CAW-N with neural network interpretation techniques (Montavon et al., 2018) may give a chance to automatically discover larger and meaningful motifs/patterns of temporal networks; CAW-N may also be generalized to predict high-order structures (e.g., triangles) that correspond to some function units of temporal networks from different domains (Benson et al., 2016; 2018; Zitnik et al., 2019).
Acknowledgments
We thank Jiaxuan You and Rex Ying for their helpful discussion on the idea of Causal Anonymous Walks. We also thank Rok Sosič, Camilo Andres Ruiz and Maria Brbić for providing insightful feedback on the abstract. We also gratefully acknowledge the support of DARPA under Nos. FA865018C7880 (ASED), N660011924033 (MCS); ARO under Nos. W911NF-16-1-0342 (MURI), W911NF-16-1-0171 (DURIP); NSF under Nos. OAC-1835598 (CINES), OAC-1934578 (HDR), CCF-1918940 (Expeditions), IIS-2030477 (RAPID); Stanford Data Science Initiative, Wu Tsai Neurosciences Institute, Chan Zuckerberg Biohub, Amazon, Boeing, JPMorgan Chase, Docomo, Hitachi, JD.com, KDDI, NVIDIA, Dell. J. L. is a Chan Zuckerberg Biohub investigator.
References
- AbuOda et al. (2019) Ghadeer AbuOda, Gianmarco De Francisci Morales, and Ashraf Aboulnaga. Link prediction via higher-order motif features. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 2019.
- Ahmed et al. (2015) Nesreen K Ahmed, Jennifer Neville, Ryan A Rossi, and Nick Duffield. Efficient graphlet counting for large networks. In International Conference on Data Mining. IEEE, 2015.
- Battaglia et al. (2018) Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, et al. Relational inductive biases, deep learning, and graph networks. arXiv preprint arXiv:1806.01261, 2018.
- Benson et al. (2016) Austin R Benson, David F Gleich, and Jure Leskovec. Higher-order organization of complex networks. Science, 353(6295), 2016.
- Benson et al. (2018) Austin R Benson, Rediet Abebe, Michael T Schaub, Ali Jadbabaie, and Jon Kleinberg. Simplicial closure and higher-order link prediction. Proceedings of the National Academy of Sciences, 115(48):E11221–E11230, 2018.
- Bochner (1992) Salomon Bochner. A theorem on Fourier-Stieltjes integrals. Collected Papers of Salomon Bochner, 2, 1992.
- Chen et al. (2019) Jinyin Chen, Jian Zhang, Xuanheng Xu, Chenbo Fu, Dan Zhang, Qingpeng Zhang, and Qi Xuan. E-lstm-d: A deep learning framework for dynamic network link prediction. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2019.
- Du et al. (2018) Lun Du, Yun Wang, Guojie Song, Zhicong Lu, and Junshan Wang. Dynamic network embedding: An extended approach for skip-gram based network embedding. In International Joint Conference on Artificial Intelligence, 2018.
- Gorochowski et al. (2018) Thomas E Gorochowski, Claire S Grierson, and Mario di Bernardo. Organization of feed-forward loop motifs reveals architectural principles in natural and engineered networks. Science advances, 4(3), 2018.
- Goyal et al. (2020) Palash Goyal, Sujit Rokka Chhetri, and Arquimedes Canedo. dyngraph2vec: Capturing network dynamics using dynamic graph representation learning. Knowledge-Based Systems, 187, 2020.
- Granovetter (1973) Mark S Granovetter. The strength of weak ties. American journal of sociology, 78(6), 1973.
- Hajiramezanali et al. (2019) Ehsan Hajiramezanali, Arman Hasanzadeh, Krishna Narayanan, Nick Duffield, Mingyuan Zhou, and Xiaoning Qian. Variational graph recurrent neural networks. In Advances in Neural Information Processing Systems, 2019.
- Hamilton et al. (2017a) Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems, 2017a.
- Hamilton et al. (2017b) William L Hamilton, Rex Ying, and Jure Leskovec. Representation learning on graphs: Methods and applications. IEEE Data Engineering Bulletin, 40(3), 2017b.
- Holme & Saramäki (2012) Petter Holme and Jari Saramäki. Temporal networks. Physics reports, 519(3), 2012.
- Ivanov & Burnaev (2018) Sergey Ivanov and Evgeny Burnaev. Anonymous walk embeddings. In International Conference on Machine Learning, 2018.
- Kazemi et al. (2019) Seyed Mehran Kazemi, Rishab Goel, Sepehr Eghbali, Janahan Ramanan, Jaspreet Sahota, Sanjay Thakur, Stella Wu, Cathal Smyth, Pascal Poupart, and Marcus Brubaker. Time2vec: Learning a vector representation of time. arXiv preprint arXiv:1907.05321, 2019.
- Kipf & Welling (2016) Thomas N Kipf and Max Welling. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308, 2016.
- Kipf & Welling (2017) Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations, 2017.
- Kovanen et al. (2011) Lauri Kovanen, Márton Karsai, Kimmo Kaski, János Kertész, and Jari Saramäki. Temporal motifs in time-dependent networks. Journal of Statistical Mechanics: Theory and Experiment, 2011(11), 2011.
- Kovanen et al. (2013) Lauri Kovanen, Kimmo Kaski, János Kertész, and Jari Saramäki. Temporal motifs reveal homophily, gender-specific patterns, and group talk in call sequences. Proceedings of the National Academy of Sciences, 110(45), 2013.
- Kumar et al. (2019) Srijan Kumar, Xikun Zhang, and Jure Leskovec. Predicting dynamic embedding trajectory in temporal interaction networks. In International Conference on Knowledge Discovery Data Mining, 2019.
- Lahiri & Berger-Wolf (2007) Mayank Lahiri and Tanya Y Berger-Wolf. Structure prediction in temporal networks using frequent subgraphs. In IEEE Symposium on Computational Intelligence and Data Mining, 2007.
- Last & Penrose (2017) Günter Last and Mathew Penrose. Lectures on the Poisson process, volume 7. Cambridge University Press, 2017.
- Lavenberg (1983) Stephen Lavenberg. Computer performance modeling handbook. Elsevier, 1983.
- Li & Milenkovic (2017) Pan Li and Olgica Milenkovic. Inhomogeneous hypergraph clustering with applications. In Advances in Neural Information Processing Systems, 2017.
- Li et al. (2018) Taisong Li, Jiawei Zhang, S Yu Philip, Yan Zhang, and Yonghong Yan. Deep dynamic network embedding for link prediction. IEEE Access, 6, 2018.
- Mahdavi et al. (2018) Sedigheh Mahdavi, Shima Khoshraftar, and Aijun An. dynnode2vec: Scalable dynamic network embedding. In International Conference on Big Data (Big Data). IEEE, 2018.
- Manessi et al. (2020) Franco Manessi, Alessandro Rozza, and Mario Manzo. Dynamic graph convolutional networks. Pattern Recognition, 97, 2020.
- Mangan & Alon (2003) Shmoolik Mangan and Uri Alon. Structure and function of the feed-forward loop network motif. Proceedings of the National Academy of Sciences, 100(21), 2003.
- Micali & Zhu (2016) Silvio Micali and Zeyuan Allen Zhu. Reconstructing markov processes from independent and anonymous experiments. Discrete Applied Mathematics, 200, 2016.
- Montavon et al. (2018) Grégoire Montavon, Wojciech Samek, and Klaus-Robert Müller. Methods for interpreting and understanding deep neural networks. Digital Signal Processing, 73:1–15, 2018.
- Nguyen et al. (2018) Giang Hoang Nguyen, John Boaz Lee, Ryan A Rossi, Nesreen K Ahmed, Eunyee Koh, and Sungchul Kim. Continuous-time dynamic network embeddings. In Companion Proceedings of the The Web Conference, 2018.
- Paranjape et al. (2017) Ashwin Paranjape, Austin R Benson, and Jure Leskovec. Motifs in temporal networks. In International Conference on Web Search and Data Mining, 2017.
- Pareja et al. (2020) Aldo Pareja, Giacomo Domeniconi, Jie Chen, Tengfei Ma, Toyotaro Suzumura, Hiroki Kanezashi, Tim Kaler, Tao B Schardl, and Charles E Leiserson. EvolveGCN: Evolving graph convolutional networks for dynamic graphs. In AAAI Conference on Artificial Intelligence, 2020.
- Rahman & Al Hasan (2016) Mahmudur Rahman and Mohammad Al Hasan. Link prediction in dynamic networks using graphlet. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 2016.
- Rossi et al. (2020) Emanuele Rossi, Ben Chamberlain, Fabrizio Frasca, Davide Eynard, Federico Monti, and Michael Bronstein. Temporal graph networks for deep learning on dynamic graphs. arXiv preprint arXiv:2006.10637, 2020.
- Rossi et al. (2019) Ryan A Rossi, Anup Rao, Sungchul Kim, Eunyee Koh, Nesreen K Ahmed, and Gang Wu. Higher-order ranking and link prediction: From closing triangles to closing higher-order motifs. arXiv preprint arXiv:1906.05059, 2019.
- Rumelhart et al. (1986) David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 323(6088), 1986.
- Sankar et al. (2020) Aravind Sankar, Yanhong Wu, Liang Gou, Wei Zhang, and Hao Yang. DySAT: Deep neural representation learning on dynamic graphs via self-attention networks. In International Conference on Web Search and Data Mining, 2020.
- Sarkar et al. (2012) Purnamrita Sarkar, Deepayan Chakrabarti, and Michael I Jordan. Nonparametric link prediction in dynamic networks. In International Conference on Machine Learning, 2012.
- Scarselli et al. (2008) Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. The graph neural network model. IEEE Transactions on Neural Networks, 20(1), 2008.
- Simmel (1950) Georg Simmel. The sociology of georg simmel, volume 92892. Simon and Schuster, 1950.
- Singer et al. (2019) Uriel Singer, Ido Guy, and Kira Radinsky. Node embedding over temporal graphs. In International Joint Conference on Artificial Intelligence, 2019.
- Srinivasan & Ribeiro (2019) Balasubramaniam Srinivasan and Bruno Ribeiro. On the equivalence between positional node embeddings and structural graph representations. In International Conference on Learning Representations, 2019.
- Toivonen et al. (2007) Riitta Toivonen, Jussi M Kumpula, Jari Saramäki, Jukka-Pekka Onnela, János Kertész, and Kimmo Kaski. The role of edge weights in social networks: modelling structure and dynamics. In Noise and Stochastics in Complex Systems and Finance, volume 6601. International Society for Optics and Photonics, 2007.
- Trivedi et al. (2017) Rakshit Trivedi, Hanjun Dai, Yichen Wang, and Le Song. Know-evolve: deep temporal reasoning for dynamic knowledge graphs. In International Conference on Machine Learning, 2017.
- Trivedi et al. (2019) Rakshit Trivedi, Mehrdad Farajtabar, Prasenjeet Biswal, and Hongyuan Zha. Dyrep: Learning representations over dynamic graphs. In International Conference on Learning Representation, 2019.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems, 2017.
- Veličković et al. (2018) Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In International Conference on Learning Representations, 2018.
- Wang et al. (2020) Yanbang Wang, Pan Li, Chongyang Bai, VS Subrahmanian, and Jure Leskovec. Generic representation learning for dynamic social interaction. International Conference on Knowledge Discovery Data Mining, MLG workshop, 2020.
- Xu et al. (2019) Da Xu, Chuanwei Ruan, Evren Korpeoglu, Sushant Kumar, and Kannan Achan. Self-attention with functional time representation learning. In Advances in Neural Information Processing Systems, 2019.
- Xu et al. (2020) Da Xu, Chuanwei Ruan, Evren Korpeoglu, Sushant Kumar, and Kannan Achan. Inductive representation learning on temporal graphs. In International Conference on Learning Representation, 2020.
- Zhang et al. (2020) Muhan Zhang, Pan Li, Yinglong Xia, Kai Wang, and Long Jin. Revisiting graph neural networks for link prediction. arXiv preprint arXiv:2010.16103, 2020.
- Zhou et al. (2018) Le-kui Zhou, Yang Yang, Xiang Ren, Fei Wu, and Yueting Zhuang. Dynamic network embedding by modeling triadic closure process. In AAAI Conference on Artificial Intelligence, 2018.
- Zitnik et al. (2019) Marinka Zitnik, Marcus W Feldman, Jure Leskovec, et al. Evolution of resilience in protein interactomes across the tree of life. Proceedings of the National Academy of Sciences, 116(10), 2019.
Appendix A Efficient link Sampling
Our efficient link sampling strategy contains two subroutines – Online probability computation (Alg.2) and Iterative sampling (Alg.3). The Online probability computation subroutine Alg.2 essentially works online to assign each new incoming link with a pair of probabilities such that
(9) |
These probabilities will be used later in sampling (Alg.3) and do not need to be updated any more.
The Iterative sampling subroutine Alg.3 is an efficient implementation of step 5 in Alg.1. We may first show that the sampling probability of a link in is proportional to in Prop.A.1.
Proposition A.1.
Proof.
To show this, we first order the timestamps of links in between as where there exists an link for . Then, the probability to sample a link satisfies
which is exactly the probability that we need. ∎
We have the following Thm.A.2 with a weak assumption that evaluates the complexity of Alg.3. We assume that links come in by following a Poisson point process with intensity , which is a frequently used assumption to model communication networks (Lavenberg, 1983). This result indicates that if , for each node, we only need to record the most recent links to sample. This result is important as it means our method can do online training and inference with time and memory complexity that are not related to the total number of links.
Theorem A.2.
If the links that are connected to appear by following a Poisson point process with intensity . Then, the expected number of iterations of Alg.3 is bounded by .
Proof.
The number of iterations of Alg.3 is always bounded by . So we only need to prove that the expected number of interactions of Alg.3 is bounded by .
To show this, we order the timestamps of links in between as where there exists an link for . Further define we define for .
Due to the definition of Poisson process, we know that each time difference in follows i.i.d. exponential distribution with parameter . Therefore, are also i.i.d.. Let .
The probability that Alg.3 runs iterations is equal to the probability that the link with timestamp gets sampled. That is
Therefore, the expected number of iterations is
(10) |
where and . Next we will prove that each item in right-hand-side of Eq.10 satisfies
(11) |
If this is true, then
Appendix B Tree-structured sampling
We may further decrease the sampling complexity by revising Alg. 1 into tree-structured sampling. Alg. 1 originally requires to sample link times because we need to search links that connected to in the first step and then sample one link for each of the nodes in each following step. A tree-structured sampling strategy may reduce this number: Specifically, we sample links for each node in step but we make sure , which does not change the total number of walks. In this way, the times of link search decrease to . Suppose , , and , then the times of link search decrease to about . Though empirical results below show that tree-structured sampling achieves slightly worse performance, it provides an opportunity to tradeoff between prediction performance and time complexity.

We conduct more experiment to investigate this topic with Wikipedia and UCI datasets. The setup is as follows: first, we fix CAW sampling number and length , so that we always have ; next, we assign different values to , so that the shape of the tree changes accordingly; controlling other hyperparameters to be the optimal combination found by grid search, we plot the corresponding inductive AUC scores of CAW-N-mean on all testing edges in Fig. 7. It is observed that while tree-structured sampling may affect the performance to some extent, its negative impact is less prominent when the first-step sampling number is relatively large, and our model still achieves state-of-the-art performance compared to our baselines. That makes the tree-structured sampling a reasonable strategy that can further reduce time complexity.
Appendix C Additional Experimental Setup Details
C.1 Dataset Introduction and Access
We list the introduction of the six datasets as follows.
-
•
Reddit111 http://snap.stanford.edu/jodie/reddit.csv is a dataset of posts made by users on subredditts over a month. Its nodes are users and posts, and its links are the timestamped posting requests.
-
•
Wikipedia222http://snap.stanford.edu/jodie/wikipedia.csv is a dataset of edits over wiki pages over a month, whose nodes represent human editors and wiki pages and whose links represent timestamped edits.
-
•
Social Evolution333http://realitycommons.media.mit.edu/socialevolution.html is a dataset recording the detailed evolving physical proximity between students in a dormitory over a year, deterimined from wireless signals of their mobile devices.
-
•
Enron444https://www.cs.cmu.edu/~./enron/ is a communication network whose links are email communication between core employees of a cooperation over several years.
-
•
UCI555http://konect.cc/networks/opsahl-ucforum/ is a dataset recording online posts made by university students on a forum, but is non-attributed.
-
•
MOOC666http://snap.stanford.edu/jodie/mooc.csv is a dataset of online courses where nodes represent students and course content units such as videos and problem sets, and links represent student’s access behavior to a particular unit.
C.2 Baselines, Implementation and Training Details
C.2.1 CAW-N-mean and CAW-N-attn
We first report the general training hyperparameters of our models in addition to those mentioned in the main text: on all datasets, we train both variants with mini-batch size 32 and set learning rate = ; the maximum training epoch number is 50 though in practice we observe that with early stopping we usually find the optimal epoch in fewer than 10 epochs; our early stopping strategy is that if the validation performance does not increase for more than 3 epoch then we stop and use the third previous epoch for testing; dropout layers with dropout probability = 0.1 are added to the RNN module, the MLP modules, and the self-attention pooling layer. Please refer to our code for more details.
In terms of the three hyperparameters controlling CAW sampling, we discussed them in Sec 5.3. For all datasets, they are systematically tuned with grid search, whose ranges are reported in Tab.4.
Dataset | Sampling number | Time decay | Walk length |
---|---|---|---|
32, 64, 128 | {0.25, 0.5, 1.0, 2.0, 4.0} | 1, 2, 3, 4 | |
Wikipedia | 32, 64, 128 | {0.25, 0.5, 1.0, 2.0, 4.0} | 2, 3, 4 |
MOOC | 32, 64, 128 | {0.25, 0.5, 1.0, 2.0, 4.0} | 2, 3, 4, 5 |
Social Evo. | 32, 64, 128 | {0.25, 0.5, 1.0, 2.0, 4.0, 8.0} | 1, 2, 3 |
Enron | 32, 64, 128 | {0.25, 0.5, 1.0, 2.0, 4.0} | 1, 2, 3, 4 |
UCI | 32, 64, 128 | {0.6, 0.8, 1.0, 1.2, 1.4} | 1, 2, 3 |
Apart from the hyperparameters controlling CAW sampling, hidden dimensions of CAW-N mentioned in Sec.4.3, including that of the various encodings, MLPs, RNN, and attention projection matrices, are relatively less tuned. We select them based on two principles: 1) when node & link attributes are available, dimension of all these modules are set to have the same dimensions as baselines; 2) when node & link attributes are unavailable, the dimensions are picked from 32, 64, 128, whichever leads to a better performance.
C.2.2 Baselines
We list the introduction of the six baselines as follows:
-
•
DynAERNN (Goyal et al., 2020) uses a fully connected encoder to acquire network representations, passes them into LSTM and uses a fully connected network to decode the future network structures.
-
•
JODIE (Kumar et al., 2019) applies RNNs to estimate the future embedding of nodes. The model was proposed for bipartite graphs while we properly modify it for standard graphs if the input graphs are non-bipartite.
-
•
DyRep (Trivedi et al., 2019) also uses RNNs to learn node embedding while its loss function is built upon temporal point process.
- •
-
•
EvolveGCN (Pareja et al., 2020) uses a RNN to estimate the GCN parameters for the future snapshots.
- •
We introduce how we tune these baselines as follows.
DynAERNN. The model with code provided here is adapted into our evaluation pipeline. We follow most of the settings in the code. We tune the embedding size in {32, 64} and lookback in {2, 3, 5} to report the best performance.
JODIE. The model with code provided here is adapted into our evaluation pipeline. JODIE calculates the distances between the predicted item embedding to other items and uses the rankings to evaluate their performance. Here, we consider the negative distances as the prediction score. Based on the prediction score, we calculate mAP and AUC. We split the data according to the setting in section 5.1. The model is trained for 50 epoches. The dimensions of the dynamic embedding is searched in{64, 128} and the best performance is reported.
DyRep. The model with code provided here is adapted into our evaluation pipeline. We follow most of the settings in the paper. That is, we set the number of samples for survival to 5, gradient clipping to 100. And we tune the hidden unit size and embedding size in {32, 64} to report the best performance. The model uses likelihood based on point process to predict links and therefore we use these likelihood scores to compute AUC and AP.
VGRNN. The model with code provided here is adapted into our evaluation pipeline. We use several of its default settings: one-hot node features as input when node attributes are unavailable as suggested by the original paper (Hajiramezanali et al., 2019), one layer of GRU network as the history tracking backbone, a learning rate of 1e-2, and training for 1000 epochs. Its hidden dimension is searched in {32, 64} and the best performance is reported.
EvolveGCN. The model with code provided here is adapted into our evaluation pipeline. We utilize EvolveGCN-O version since it can capture more graph structural information. For most hyperparameters, we follow the default values. According to our setting, we sample an equal amount of negative links, which means we set negative_mult_training and negative_mult_test to 1. One central hyperparameter needs to be further tuned is number of previous snapshots used for training and testing. We search its optimal value in {3, 4, 6, 8, 10} when tuning the model for most datasets. Since Enron only contains 11 snapshots, we search its optimal value in {3,4,5,6}.
TGAT. The model with code provided here is adapted into our evaluation pipeline. We use several of their default settings. That is, we use product attention, set the number of attention heads to 2, set the number of graph attention layers to 2, and use 100 as their default hidden dimension. One central hyperparameter that needs to be further tuned is the degree of their neighbor sampling. We search its optimal value in {10, 20, 30} when tuning the model.
C.3 Evaluation of Snapshot-based Baselines
Wikipedia | MOOC | Social Evo. | Enron | UCI | ||
---|---|---|---|---|---|---|
total snapshots | 174 | 20 | 20 | 27 | 11 | 88 |
exact split | 122 / 26 / 26 | 14 / 3 / 3 | 14 / 3 / 3 | 19 / 4 / 4 | 7 / 2 / 2 | 62 / 13 / 13 |
referenced baseline | EvolveGCN | - | - | VGRNN | VGRNN | EvolveGCN |
We make the following decisions to evaluate snapshot-based baselines in a fair manner, so that their performances are comparable to those derived from the stream-based evaluation procedure. The first step we do is to evenly split the whole dataset chronologically into a number of snapshots. We determine the exact number of snapshots by referring the three snapshot-based baselines we use. For Wikipedia and MOOC dataset which are not used by any snapshot-based baseline, we split them into a total of 20 snapshots. Next, we need to determine the proportions of these snapshots assigned each to training, validation, and testing set. In doing this, our principle is that the proportions of these three sets should be close to 70:15:15 as much as possible, since that ratio is what we use for evaluating stream-based baselines and our proposed method. These decisions lead to our final splitting scheme summarized in Tab. 5.
Extra care should also be taken when testing snapshot-based methods. For a queried link in a snapshot, usually snapshot-based methods only make a binary prediction whether or not that link may exist at any time in that snapshot. They do not, however, take care of the case that the link may appear multiple times at different time points within that snapshot’s time range. This lead to a different evaluation scheme than stream-based methods, which do consider the multiplicity of links. Therefore, when testing snapshot-based methods, if a link appear in a certain snapshot for multiple times, we record the model’s prediction the same number of times before computing its performance metrics.
C.4 Choice of Evaluation Metric
When considering link prediction as a binary classification problem, the existing literature usually choose metrics from the following: Area Under the ROC Curve (AUC), Average Precision (AP), and Accuracy (ACC). The reason we do not use ACC is that a proper confidence threshold of decision is ill-defined in literature, which leads to unfair comparison across different works.
C.5 Computing Infrastructure
All the experiments were carried out on a Ubuntu 16.04 server with Xeon Gold 6148 2.4 GHz 40-core CPU, Nvidia 2080 Ti RTX 11GB GPU, and 768 GB memory.
Appendix D Additional Experimental Results
D.1 Performance in Average Precision
Task | Methods | Wikipedia | MOOC | Social Evo. | Enron | UCI | ||
---|---|---|---|---|---|---|---|---|
Inductive | new v.s. new | DynAERNN | 58.63 5.42 | 54.94 2.29 | 59.84 1.26 | 54.76 1.33 | 54.89 3.79 | 51.59 3.92 |
JODIE | 80.03 0.13 | 76.90 0.49 | 82.27 0.46† | 87.96 0.12† | 79.80 1.48† | 71.64 0.62 | ||
DyRep | 61.28 1.89 | 57.57 2.56 | 62.29 2.09 | 75.42 0.32 | 69.97 0.92 | 63.08 7.40 | ||
VGRNN | 60.64 0.68 | 52.55 0.82 | 65.44 0.82 | 67.83 0.53 | 67.93 0.88 | 67.50 0.92 | ||
EvolveGCN | 62.99 0.17 | 55.64 1.03 | 52.28 1.80 | 52.26 1.16 | 47.36 1.24 | 80.98 1.09† | ||
TGAT | 95.17 0.91† | 93.18 0.73† | 72.91 0.92 | 52.17 1.94 | 63.83 3.70 | 75.27 2.34 | ||
CAW-N-mean | 97.90 0.08∗ | 99.35 0.15∗ | 90.70 0.12∗ | 95.70 0.23∗ | 96.92 0.34∗ | 95.19 0.21∗ | ||
CAW-N-attn | 97.95 0.09∗ | 99.34 0.25∗ | 90.75 0.11∗ | 95.35 0.36∗ | 96.53 0.28∗ | 95.18 0.21∗ | ||
new v.s. old | DynAERNN | 66.59 2.90 | 63.76 2.82 | 82.02 1.59 | 52.54 0.22 | 55.50 2.07 | 57.29 2.52 | |
JODIE | 83.15 0.03 | 80.54 0.06 | 87.95 0.08 | 91.40 0.04† | 89.57 0.30 | 76.34 0.17 | ||
DyRep | 66.73 1.99 | 76.89 0.31 | 88.25 1.20† | 89.41 0.29 | 95.97 0.28† | 93.60 1.47† | ||
VGRNN | 52.84 0.66 | 60.99 0.55 | 62.95 0.58 | 69.20 0.52 | 67.93 0.88 | 67.50 0.92 | ||
EvolveGCN | 66.29 0.52 | 53.82 1.64 | 51.53 0.92 | 52.01 0.67 | 46.56 1.89 | 76.30 0.33 | ||
TGAT | 97.09 0.18† | 95.17 0.15† | 71.77 0.23 | 52.48 0.52 | 59.70 1.49 | 75.01 0.72 | ||
CAW-N-mean | 97.21 0.12∗ | 98.87 0.11∗ | 91.40 0.15∗ | 90.01 0.10 | 93.47 0.29 | 93.27 0.42 | ||
CAW-N-attn | 97.06 0.13∗ | 98.96 0.11∗ | 91.48 0.19∗ | 90.03 0.11 | 94.01 0.30 | 93.80 0.30 | ||
Transductive | DynAERNN | 85.58 2.12 | 76.58 1.41 | 89.29 0.49 | 66.58 1.84 | 60.90 2.70 | 84.95 2.13 | |
JODIE | 91.14 0.01 | 91.39 0.04 | 91.19 0.03† | 89.22 0.01 | 91.94 0.01 | 80.27 0.08 | ||
DyRep | 67.54 2.02 | 77.36 0.25 | 90.49 0.03 | 94.48 0.01† | 97.14 0.07† | 95.29 0.13† | ||
VGRNN | 50.87 0.81 | 67.66 0.89 | 83.70 0.56 | 78.66 0.67 | 94.02 0.52 | 82.23 0.56 | ||
EvolveGCN | 54.49 0.73 | 55.84 0.37 | 51.80 0.46 | 56.90 0.54 | 69.72 0.49 | 81.63 0.23 | ||
TGAT | 98.38 0.01† | 96.65 0.06† | 69.75 0.23 | 57.37 1.18 | 57.37 0.18 | 60.25 0.31 | ||
CAW-N-mean | 98.72 0.09∗ | 98.82 0.12∗ | 94.50 0.13∗ | 94.36 0.18 | 92.12 0.09 | 95.33 0.30∗ | ||
CAW-N-attn | 98.80 0.07∗ | 98.84 0.11∗ | 94.55 0.14∗ | 93.80 0.16 | 92.29 0.09 | 95.02 0.14 |
D.2 More Discussion on Stream-based vs. Snapshot-based Methods
Stream-based methods usually treat each temporal link as an individual training instance. In contrast, snapshot-based methods stack all the temporal links within a time slice into one static graph snapshot and do not further distinguish temporal order of links within that snapshot. In Tab. 2 and 6 we saw that stream-based methods generally exhibit better performance than snapshot-based methods. An important reason is that stream-based methods are able to access the few most recent interactions previous to the target link to predict. This makes them especially advantageous when used to model many common temporal networks whose dynamics are governed by some short-term laws. In contrast, snapshot-based methods are less able to access such immediate history, because they make prediction on all links within one future snapshot all at once. In principle they could alleviate this problem by making very fine-grained snapshots so that less immediate history is missed. However, this is not practical with real-word large temporal graphs whose temporal links come in millions, which leads to snapshot sequences of extreme length intractable to their recurrent neural structure. This issue was also observed by the recent work to model social interacting behaviors (Wang et al., 2020), although temporal convolutional networks may alleviate this issue to some extent. That said, snapshot-based methods usually has the advantage that they usually consume less memory and computation time. Stream-based methods on the other hand need to manage how they sample history very carefully to balance the efficiency and effectiveness. Our proposed algorithm on CAW sampling comes into the place in light of this to solve the problem.
Appendix E Visualizing CAWs and AWs
Here, we introduce one way to visualize and interpret CAWs. Our interpretation can also illustrate the importance to capture the correlation between walks to represent the dynamics of temporal networks, where the set-based anonymization of CAWs can work while AWs will fail. The basic idea of our intrepretation is to identify different shapes of CAWs via their patterns encoded in , and compare their contributions to the link-prediction confidence. The idea will be also used to interpret AWs so that we can compare CAWs with AWs.
First, we define the shapes of walks based on . Recall from Eq. 3 that encodes the number of times node that appears in different walk positions w.r.t source node . This encoding induces a temporal shortest-path distance between node and : . Note that in temporal networks, there is not a canonical way to define shortest-path distance between two nodes as there is no static structures. So our definition can be viewed the shortest-path distance between u and w over the subgraph that consists of walks in . Based on the way to define , we introduce the mapping from of the node to a coordinate of this node in the subgraph that consists of walks in : . This cooredinate can be viewed as a relative coordinate of node w.r.t. the source nodes , . Each walk can then represented as a sequence of such coordinates by mapping each node’s to a coordinate. The obtained sequence can be viewed as a shape of and we denote the obtained shape as . For instance, in the toy example shown by Fig. 2 right, the first CAW in , is mapped to a new coordinate sequence . The marks the setting that node , do not appear in .
Next, we score the contributions of CAWs with different shapes. We use CAW-N-Mean as enc() is simply mean over the encodings of sampled CAWs and further use linear projrepresented as a sequence of such coordinates by mapping each node’section to compute the final scalar logit for prediction. As the two operations mean and projection are commutative, the above setting allows each CAW contributing a scalar score to the final logit.

Fig. 8 lists the 3 highest-scored and the 3 lowest-scored shapes of CAW, which are extracted from the Wikipedia dataset with and . A law of general motif closure can be observed from the highest-scored CAWs: two nodes that commonly appear in some types of motif are more inclined to have a link in between. For example, the highest-scored shape of CAW, , implies that the nodes except the first in this CAW appear in the sampled common 3-hop neighborhood around the two nodes between which the link is to be predicted. Therefore, CAW-N essentially adaptively samples a temporal motif closure pattern that is very informative to predict this link. CAW-N does not explicitly enumerating or counting these motif patterns. In contrast, when CAWs do not bridge the two nodes, as shown in top-2 lowest-scored CAWs, very unlikely there will exist a link. Fig. 8 also displays each of the 6 CAW’s occurrence ratio with positive and negative links. The difference within each pair of ratios is an indicator of the corresponding CAW’s discriminatory power. We also see that the discriminatory power of CAWs is very strong: highest-scored CAWs almost never occur with negative links, and lowest-scored CAWs also seldom occur with positive links.

We further apply the similar procedure to analyze the AWs introduced in Sec. 4.1 and the model Ab.5 of Tab. 3 used for ablation study. Note that AW cannot establish the correlation between walks, each single AW itself, say , decides its own shape . We directly set with defined in Eq. 2. As the Wikipedia dataset is a bipartite graph, there are in total four different shapes when as listed align with the x-axis of Fig. 9. For illustration, we explain one shape of AW as an example, say : The corresponding walks have the second and the fourth nodes correspond to the same node, which the first, second and third nodes are different. As shown in Fig. 9, we can see that AW’s occurrence with positive versus negative links are highly mixed-up, compared to CAW’s. That suggests that AWs possess significantly less discriminatory power than CAWs. The main reason is that AWs do not have set-based anonymization, so they cannot capture the correlation between walks/motifs but CAWs can do that. This observation further gives a reason on why the model Ab.5 of Tab. 3 only achieves some performance on-par with the model Ab.2 where we totally remove the anonymization procedure: the anonymization adopted by AWs loses too much information of the structure and cannot benefit the prediction much. However, the original CAW-N well captures such information via the set-based anonymization.
Appendix F Comparison with the most recent work
Task | Wikipedia | MOOC | Social Evo. | Enron | UCI | |
---|---|---|---|---|---|---|
New vs. New | 98.25 0.14 | 98.73 0.12 | 86.94 0.12 | 95.37 0.04 | 70.51 0.21 | 75.47 0.08 |
New vs. Old | 97.33 0.07 | 97.06 0.03 | 89.13 0.79 | 93.35 0.07 | 78.12 0.23 | 80.53 0.16 |
Transductive | 98.68 0.03 | 98.39 0.05 | 90.83 0.17 | 95.44 0.08 | 87.56 0.10 | 78.91 0.08 |
Task | Wikipedia | MOOC | Social Evo. | Enron | UCI | |
---|---|---|---|---|---|---|
New vs. New | 98.13 0.07 | 98.72 0.12 | 84.96 0.04 | 93.81 0.11 | 72.81 0.01 | 75.47 0.08 |
New vs. Old | 97.45 0.08 | 97.29 0.07 | 87.68 1.44 | 90.79 0.03 | 77.05 0.03 | 80.53 0.16 |
Transductive | 98.70 0.08 | 98.47 0.11 | 89.14 0.05 | 93.54 0.06 | 85.19 0.08 | 78.91 0.08 |
Temporal Graph Networks(TGN) (Rossi et al., 2020) is a recent work that performs representation learning on temporal networks, which has been developed in parallel with this work. We were not aware of it until our paper was published. We would like to further conduct some experiments to compare TGN with CAW-N, in both inductive (New vs. New, New vs. Old) and transductive settings.
The implementation details of TGN experiments are as follows. We adapt the code of TGN code777https://github.com/twitter-research/tgn into our evaluation pipeline. We use the the default architecture and hyperparameter reported by the code: the backbone is TGN-attn variant with memory updater; the number of heads for graph attention is 2; minibatch size is 200; the sampling number of temporal neighbors is 10; the dimension of time embedding is 100. The model is trained for 50 epochs with early stopping, and the learning rate is 0.0001.
We report TGN’s AUC and AP scores in Tab.7, 8. Compared with previous baselines, we observe that TGN’s performance is generally better, and also shows smaller gaps between inductive and transductive settings. However, our CAW-N still outperforms TGN on most dataset, especially by significant margin on Enron and UCI.
In practice, we also note that TGN has the fastest execution speed among all baselines. It is also faster than CAW-N under its optimal hyperparameter setting. One possible reason is that TGN’s message function is simple and its graph attention module requires only the first-hop neighbors to be sampled. To further speed up the execution of our CAW-N method, we can parallelize the computation of encodings in each minibatch. Another direction is to reduce the complexity of CAW sampling: Our current implementation repeatedly samples a number of CAWs from scratch for each new temporal link. Alternatively, we may only maintain a dynamic list of sampled CAWs for each node as we proceed along the temporal link stream. Each time we encounter a new link that involves , we update by dropping outdated walks and by adding in newly sampled CAWs that start from . This would significantly reduce the redundancy of CAW sampling and thus speed up our method. Detailed implementation and evaluation are left for future study.
Appendix G Notes on Updated Experimental Results (Oct 16, 2022)
It has come to our notice there exists an error in our original implementation of as well as the attention Self-Att-AGG(). Fixing the error leads to numerical changes to some of the experimental results reported in the original paper: Tables 2, 3, 6, and Fig.8. Based on the correct implementation, we have updated the artifacts above and their relevant text . The correct implementation has also been updated to the GitHub repository at the original code link. Despite the change, the vast majority of the conclusions about CAWN remain unaffected.
We would also like to extend our gratitude to the authors of Online Graph Nets (https://openreview.net/forum?id=0IqFsR9wJvI), Hojin Kang, Jou-Hui Ho, Diego Mesquita, Jorge Pérez, Amauri H Souza, for helping us identify the aforementioned error.