EPNE: Evolutionary Pattern Preserving
Network Embedding
Abstract
Information networks are ubiquitous and are ideal for modeling relational data. Networks being sparse and irregular, network embedding algorithms have caught the attention of many researchers, who came up with numerous embeddings algorithms in static networks. Yet in real life, networks constantly evolve over time. Hence, evolutionary patterns, namely how nodes develop itself over time, would serve as a powerful complement to static structures in embedding networks, on which relatively few works focus. In this paper, we propose EPNE, a temporal network embedding model preserving evolutionary patterns of the local structure of nodes. In particular, we analyze evolutionary patterns with and without periodicity and design strategies correspondingly to model such patterns in time-frequency domains based on causal convolutions. In addition, we propose a temporal objective function which is optimized simultaneously with proximity ones such that both temporal and structural information are preserved. With the adequate modeling of temporal information, our model is able to outperform other competitive methods in various prediction tasks.
1 Introduction
Information networks are present everywhere due to its vivid depiction of relational data. Network data being sparse and irregular, network embedding algorithms [4], mapping nodes to vectors such that network properties are maintained, alleviate such drawbacks and further facilitate numerous prediction tasks. Most existing embedding methods generally consider local and global node proximity, among which models including DeepWalk [19] and GraphSAGE [7] are highly popular and efficient. However, one unrealistic assumption all of the above models make is that the network is static and all nodes have been sufficiently represented through the static network structure.
In reality, nevertheless, this is hardly the case. Hardly any networks in real life remain stagnant forever, with its structure constantly evolving over time [9]. Hence, in addition to structures, temporal evolutionary patterns for nodes, namely how local structures of a node change over time would serve as complements to static structures and are indicative of key properties of nodes. Complex as such patterns are, many of them are characterized by strong periodicity. For example, in social networks, colleagues interact more frequently during weekdays, while families are more active during weekends, where interaction patterns shed light on relationships underlying edges. There are also non-periodic patterns which still illustrate certain trends. For example, a user who is starting to join a community is generally expected to further join the community, where his evolution is monotonous but shows a trend he is inclined to follow. Therefore, modeling evolutionary patterns for nodes and incorporating them into learning node embeddings will be helpful to infer node labels and identify edge relations.
Temporal Network Embedding, reflecting the need for modeling such permanent change, came into life. Most existing temporal network embedding models either focus on certain network evolutionary processes, or are able to efficiently update their representation vectors so that they constantly embody up-to-date information. For example, DynamicTriad [27] and HTNE [29], both of which fall into the former category, model the triadic closure process and neighborhood formation process, respectively. However, both of them only focus on a relatively specific dynamic process of networks, and thus being unable to generalize to more complex evolutionary patterns.
While the latter category has received wide attention [11, 28, 5], there have been relatively few efforts into modeling network evolution. Therefore, to complement the scarcity of such works, we focus on the former category of temporal network embedding, hoping to capture intricate, both periodic and non-periodic patterns. Consequently, two major challenges arise:
-
•
Temporal features, which capture evolutionary patterns of nodes, are hard to learn. On one hand, node interactions are often noisy and sparse, making it difficult for us to extract information from. On the other hand, evolutionary patterns cover a wide range of contents, including neighborhood formation and proximity drifts, consisting of complex periodic patterns with multiple frequencies as well as diverse non-periodic ones.
-
•
Straightforward methods of incorporating temporal features to representations, such as concatenation and addition, compromise the quality of both. On one hand, they come from different spaces and cannot be easily aligned. On the other hand, such operations treat temporal and structural features as separate without utilizing one to complement the other.
To address these challenges, in this paper, we propose a temporal network embedding model, abbreviated EPNE, which combines both structural information and temporal features. The model consists of two components. First, we analyze that evolutionary patterns of nodes’ local structures consist of periodic and non-periodic patterns, for which we design methods to capture in time-frequency domains based on causal convolutions. What is more, we designed an objective function that is not only able to capture node proximity, but is also able to preserve the learned evolutionary patterns through representation vectors, enabling us to jointly optimize the model to capture both properties of networks.
To summarize, we make the following contributions:
-
•
We propose EPNE, a network embedding algorithm on temporal networks, which preserves both evolutionary patterns and topological structures of nodes.
-
•
We analyze the evolutionary patterns of the local structure of nodes, based upon which, a novel strategy is designed to learn temporal features for periodic and non-periodic patterns using causal convolutions.
-
•
We evaluate our model on several real-world networks for node and edge classification. The results demonstrate that our model is capable of preserving the temporal features of nodes and outperforms its counterparts.
2 Related Work
Static Network Embedding. With the advent of Skip-gram [15] models in natural language processing, similar models on graphs came into being, among which DeepWalk [19], Node2Vec [6] and LINE [21] are popular models, not only because they show impressive performance, but also because of the large number of algorithms derived upon them [23, 12]. Later, with the spread of deep learning, deep models [26, 22] including GraphSAGE [7] and GCN [10] were developed to perform deep learning for network embedding.
Temporal Network Embedding. Existing works mainly focus on two aspects illustrating network dynamics. On one hand, online update methods, through which representation vectors can be efficiently adjusted to reflect the latest changes in network topology, have been widely studied [11, 28, 5, 13], demonstrating comparable results and alleviating the need to retrain the model. On the other hand, network evolving processes and factors leading to them have also received extensive attention, typical examples of which are DynamicTriad, HTNE, CTDNE [17], tNodeEmbed [20], DynamicGCN [18] and most recently, HierTCN [24], and have all achieved outstanding performance in various tasks like link prediction, and visualizing dynamics.
DynamicTriad [27] models the triadic closure process ubiquitous in social networks, where two nodes sharing a neighbor in common are motivated to form links. HTNE [29] models the neighborhood formation sequences using Hawkes Process, through which the whole local structure of nodes is incorporated with temporally-aware weights. tNodeEmbed [20] present a joint loss that creates a temporal embedding of a node using LSTM to combine its historical temporal embeddings. EvolveGCN [18] adapts a GCN model along the temporal dimension and captures the dynamics of the graph sequence by using an RNN to evolve the GCN parameters. Yet elaborate as all of them are, only specific dynamic processes are taken into account, to which our model is trying to complement. Specifically, all of them fail to account for periodic temporal patterns, which we will show to be highly indicative.

3 Preliminaries
In this section, we define our problems in the context of temporal networks, followed by the introduction of causal convolutions, which will be used to learn temporal features of nodes in the next section.
3.1 Problem Definition
Definition 1
A Temporal Network is defined as
(1) |
where represents the network snapshot at time . represents the set of nodes and represents the set of edges at time . denotes the set of static edges, i.e. the set of edges that exist at least once in all time steps. indicates an (undirected) edge between and at time . We denote as the context of in a random walk at time .
Definition 2
Given a temporal network , the problem of Temporal Network Embedding aims to learn a mapping function
(2) |
where and preserves both local network structure and evolutionary patterns of node at time .
While DynamicTriad and HTNE capture temporal evolutions of nodes by modeling its interaction with neighbors, we consider it inadequate for two reasons. On one hand, higher-order proximity is generally modeled by network embeddings, while interactions with neighbors fail to take such higher-order information into account. On the other hand, interactions between nodes can be noisy and sparse, which compromises the quality of our learned features.
On the contrary, the representation vector of a node is able to higher-order structural information and consequently, its changes reflect a richer notion of temporal drifts. What is more, they do not suffer from sparsity and are generally less noisy. Therefore, we focus on the sequence of representation vectors in a fixed history length
(3) |
to learn the temporal features of a node.
3.2 Causal Convolutions
Recurrent Neural Networks (RNNs) are popular approaches for modeling sequential data, but they are typically slow to train due to recurrent connections. Hence we resort to causal convolutions [1], a comparably flexible but much more efficient architecture for modeling sequential data, whose computation also follows the sequence as provided by the data. In temporal networks, the causal convolution for sequence is
(4) |
where is the convolution kernel. Conventionally, the convolution kernel are trainable parameters, but aimed at our problem, we design fixed convolution kernels with prior analysis, so that our model can learn temporal features of nodes in temporal networks more efficiently.
4 Proposed Model
In this section, we describe our model for temporal network embedding, which preserves both evolutionary patterns and network topology in representation vectors, as illustrated in Figure 1. First, as evolutionary patterns of local structures consist of periodic patterns and non-periodic ones, we design a method to learn temporal features capturing evolutionary patterns in time-frequency domains. Second, in order to map both evolutionary patterns and network structure into the vector space, we design a temporal objective function and optimize it along with structural objective jointly.
4.1 Modeling Evolutionary Patterns
In this section, we introduce how temporal features representing evolutionary patterns are made up of, as well as how they are extracted and processed in our model. Primarily, we consider evolutionary patterns from two aspects.
-
•
Periodic evolutionary patterns. Local structures may emerge and disappear with strong regularity. For example, sports events like the Super Bowl is held annually, and related Reddit submissions will accordingly follow such regularity.
-
•
Non-periodic evolutionary patterns. Local structures and interactions may also follow non-periodic but gradual evolutionary patterns. For example, a user who has just watched a concert of Beethoven music will be inclined to view that of other musicians, such as Mozart and Haydn.
Aiming at modeling both patterns, we learn temporal features in frequency and time domains, respectively, before combining them to get the complete temporal features.
4.1.1 Frequency Domain Features
Considering node and one of its neighbors at time and , if , then a periodicity of between and is indicated. Periodic evolutionary patterns are critical in identifying relationships between nodes. For example, in social networks, the frequency of interactions between colleagues peak during weekdays and decline during weekends, while that between families and friends is completely the opposite, shrinking in weekdays and peaking during holidays. In co-author networks, experts from different domains may exhibit different patterns of cooperation due to different frequencies at which conferences are held.
Inspired by ideas in signal analysis, we extract periodic temporal features by decomposing them into basic patterns , where is a basic pattern and is the corresponding intensity which will serve as features in our model.
Based on the general ideas above, we define a generating function to generate a set of functions as basic patterns with different scales through translating and retracting
(5) |
where are the shifting and scaling factors respectively. We set the number of scales as . These basic patterns are regarded as convolution kernels, which are used to carry out causal convolutions to extract decomposed features as
(6) |
We concatenate the vectors on different scales and shifts to get the frequency domain features of a node
(7) |
For the choice of generating functions, one simple yet commonly used option is Haar wavelets, a sequence of rescaled square-shaped signals used in computer vision and time series analysis for feature extraction [16, 2]. Such common solutions prove to be sufficient by the experimental results for extracting general evolutionary patterns. For the shifting and scaling factors and , we adopt the setting of
(8) |
as stated in [3] to be generally adopted for Haar wavelets. In addition to common settings, our model allows for personalized designs that meet the properties of individual datasets. For instance, in social networks, we divide each time slice as a one-hour interval, and designed such that basic patterns thus generated would capture daily and weekly dynamics.
4.1.2 Time Domain Features
In addition to periodic patterns, the local structure of a node may follow a steady, monotonous trend instead of periodicity. We model non-periodic evolutionary patterns by learning features in time domains based on decay kernels. Generally for humans, more recent behaviors exert a stronger impact on the present, and so does decay kernels operate on vector sequences. We define a decay kernel , where is the decay rate, which is then fed into causal convolutions on sequences to capture non-periodic trends. Specifically, we have the time domain features similarly as
(9) |
We concatenate both periodic and non-periodic features to acquire the complete temporal features
(10) |
The challenge yet to overcome is that, how should the aforementioned features be incorporated to elevate the performance of the embedding algorithm.
4.1.3 Discussion
Here we show a brief discussion about our model and DynamicTriad, HTNE. It is evident that both DynamicTriad and HTNE only capture non-periodic temporal features in that, closed triads will not be open again and the intensity of Hawkes processes is monotonously decreasing. By comparison, we propose that periodic temporal patterns do shed light on relationships between nodes and capture them, which is the clear distinction between DynamicTriad, HTNE and our model. We will show how both periodic and non-periodic temporal features contribute to the performance of our model in the experiments.
4.2 Model Optimization
In this section, we introduce our objective function that is able to synthesize both structural proximity and evolutionary patterns, followed by our optimization scheme.
4.2.1 Preserving Evolutionary Patterns
In this section, we introduce the intuition underlying our temporal objective function, followed by its formal definition.
Since the representation vectors preserve pairwise proximity between nodes, we propose that relative positions between node pairs reflect relationships between the nodes. For example, if but , we would believe that node pairs and are identically adjacent in network space, yet edge and represent relationships that are scarcely correlated.
Since we propose that evolutionary patterns can implicitly reveal relationships between node pairs, as illustrated by the example of colleagues and families, we would hence derive that, evolutionary patterns are also indicative of relative positions between pairwise nodes, and vise versa.
The intuition introduced above can be implemented by an auto-encoding style objective function
(11) |
where implies the relative position between node pairs, is some function aiming to interpret relative positions between node pairs through temporal features, and measures a distance metric between two vectors. In particular, in our model, we set
(12) |
where denotes concatenation, is a trainable matrix, denotes the sigmoid function, and denotes cosine similarity between vectors.
4.2.2 Preserving Node Proximity
In this paper, we preserve proximity between pairwise nodes through an objective derived from DeepWalk, but it should also be noticed that EPNE is not restricted to specific models and can be generally incorporated to capture temporal information. We maximizes the likelihood that the context is observed conditional on the representation vectors and . Our objective preserving proximity, accelerated by negative sampling, is defined as
(13) | ||||
4.2.3 Overall Loss
We notice that the structural objective possesses the property of rotational invariance, i.e. the objective does not change regardless of how the vector space is rotated, which poses potential threats to the learning of our model. As we model temporal features in different vector spaces between time steps, it is important that those spaces should be aligned such that we do obtain evolutionary patterns instead of random rotations of vector spaces between time steps. To enforce such restriction, we propose a loss function imposing smoothness in a weighted manner between adjacent time steps
(14) |
where measures how the structure of node changes from time to time as defined in [8].
We have the overall loss function of our model by summing these objectives
(15) |
where being the temporal weight and being the smoothness weight are hyperparameters to be tuned.
We apply Stochastic Gradient Descent (SGD) to optimize the objective function. We learn representation vectors for nodes at each time step incrementally based on historical representation vectors, thereby restricting the number of parameters at each time step within . We summarize our learning algorithm in Algorithm 1.
5 Experiments
In this section, we evaluate our model on several real-world networks on node and edge classification tasks. We first introduce our experimental setups, followed by quantitative and qualitative results.
5.1 Experimental Setup
5.1.1 Datasets
We employ the following datasets with temporal information, whose statistics are listed in Table 1.
-
•
High School [14] is a social network collected in December, 2013. We consider students as nodes and active contacts collected by wearable sensors as edges. We split time steps by one-hour intervals. The labels of edges denote Facebook friendship.
-
•
Mobile. We build a social network from a mobile phone dataset collected in October 2010. We consider users as nodes and interactions as edges . We split each time step by a two-hour interval. The labels of edges represent colleague relationships between users.
-
•
AMiner [27, 25] and DBLP [29] are both co-author networks in which nodes represent researchers and edges represent co-authorship. For temporal settings, the datasets are split with four-year and one-year intervals respectively. The node labels represent research fields according to the conferences the author published his papers in. We take edge labels in AMiner as whether two researchers share the same field of interest.
5.1.2 Baselines
We compare our model with the following novel methods. Unless specified, these models are tested with their published codes as well as parameter settings mentioned in their original papers.
- •
-
•
Static Graph Neural Networks. We take GraphSAGE [7] as an example. We take unsupervised GraphSAGE without node features. We take 2-layer networks with a hidden layer sized 100. We adopt the neighborhood sampling technique with 20 neighbors to sample at each layer.
-
•
HTNE [29]. It is a dynamic network embedding method modeling sequential neighborhood formation processes using Hawkes Process.
-
•
DynamicTriad [27]. It is a dynamic network embedding method based on triadic closure processes and social homophily. We set and tested for optimal performances.
For static methods, we compress all temporal snapshots of the graphs into a “stacked” static graph as mentioned in Definition 1, just like what [29, 27] did for static methods.
For a fair comparison, the embedding dimensions are all set to . The number of paths for each node and the length of each path for DeepWalk, Node2Vec, and EPNE are all set to . It is worth mentioning that, although Perozzi et al. [19] took 80 paths per node, it was concluded redundant in our experiments as 10 paths per node have been capable of achieving similar performance. The window size for all random-walk based models is set to 5. In addition, we set for all four datasets. We set the history length and number of scales for High School, Mobile, AMiner and DBLP respectively.
Dataset | #temp. edges | # | ||
---|---|---|---|---|
High school | 327 | 5,818 | 188,508 | 45 |
Mobile | 2,985 | 54,397 | 701,030 | 12 |
AMiner | 11,056 | 70,217 | 308,886 | 32 |
DBLP | 28,085 | 150,571 | 236,894 | 27 |
5.2 Node Classification
We first employ two networks, AMiner and DBLP for node classification experiments. The embedding vectors are trained using the whole graph, which are split into training and test sets for classification, using Logistic Regression in sklearn package. We vary the size of the training set from 10% to 90% of the whole dataset, such that the results would be indicative of embedding consistency. We repeat all experiments for 10 times and report their mean Macro-F1 scores.
The results of node classification are shown in Table 2. It can be shown that temporal embedding models generally outperform their static counterparts, which underscores that temporal features carry rich information that helps infer user communities. In addition, our model EPNE generally achieves the best performance, outperforming DeepWalk and beating its temporal counterparts, indicating that temporal information can be better captured and leveraged using evolutionary patterns learned by our model. In addition, it is also demonstrated that our model can generate embeddings that are consistent and discriminative enough with arbitrary amounts of training data.
DBLP | AMiner | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
ratio of training | 10% | 30% | 50% | 70% | 90% | 10% | 30% | 50% | 70% | 90% |
EPNE | 0.6454 | 0.6454 | 0.6472 | 0.6505 | 0.6553 | 0.7799 | 0.7972 | 0.7976 | 0.8055 | 0.8189 |
DeepWalk | 0.5889 | 0.6053 | 0.6107 | 0.6110 | 0.6109 | 0.7427 | 0.7586 | 0.7731 | 0.7743 | 0.7811 |
LINE | 0.5437 | 0.5628 | 0.5658 | 0.5660 | 0.5701 | 0.6618 | 0.7172 | 0.7283 | 0.7325 | 0.7417 |
Node2Vec | 0.5769 | 0.6014 | 0.6089 | 0.6102 | 0.6106 | 0.7599 | 0.7819 | 0.7851 | 0.7842 | 0.7838 |
GraphSAGE | 0.5496 | 0.5859 | 0.5898 | 0.5893 | 0.5912 | 0.7035 | 0.7332 | 0.7365 | 0.7276 | 0.7430 |
HTNE | 0.6037 | 0.6188 | 0.6225 | 0.6248 | 0.6245 | 0.7757 | 0.7938 | 0.8028 | 0.8007 | 0.8071 |
DynamicTriad | 0.6072 | 0.6099 | 0.6213 | 0.6275 | 0.6284 | 0.7681 | 0.7889 | 0.7893 | 0.8046 | 0.8048 |
High school | AMiner | Mobile | ||||
Macro-F1 | Micro-F1 | Macro-F1 | Micro-F1 | Macro-F1 | Micro-F1 | |
EPNE | 0.5467 | 0.7039 | 0.6197 | 0.7234 | 0.8350 | 0.8722 |
DeepWalk | 0.5190 | 0.6829 | 0.4933 | 0.6857 | 0.8257 | 0.8629 |
LINE | 0.5157 | 0.6898 | 0.4962 | 0.6890 | 0.8323 | 0.8711 |
Node2Vec | 0.4970 | 0.6913 | 0.4417 | 0.6819 | 0.8268 | 0.8636 |
GraphSAGE | 0.3951 | 0.6241 | 0.4311 | 0.6783 | 0.7296 | 0.8035 |
HTNE | 0.5138 | 0.7012 | 0.4213 | 0.6766 | 0.8331 | 0.8734 |
DynamicTriad | 0.5102 | 0.6933 | 0.5098 | 0.6856 | 0.8290 | 0.8678 |
5.3 Edge Classification
We employ three networks, High School, AMiner and Mobile for edge classification, testing whether temporal information does help infer relationships between nodes, and whether our objective facilitates inferring such relationships. It should be noticed that we use inconsistent datasets for node and edge classification because there are certain datasets where we have no access to node or edge labels. The representation vector of an edge is obtained by concatenating representation vectors of its two end nodes. Logistic Regression in sklearn package is alike used. We split each dataset into training and test sets with a ratio of 7:3. We also evaluate each model 10 times with different seeds and show the mean results, in terms of Macro-F1 and Micro-F1.
The results of edge classification are shown in Table 3. As shown, our model is able to outperform all other baselines in High School and AMiner by and , respectively. It is thus indicated that due to our explicit modeling of temporal patterns, they are preserved through representation vectors which helps identify node relationships, while other temporal methods, incapable of preserving such relationships, barely outperform their static counterparts in such tasks. As for Mobile, we assume that the performance of our model is compromised due to insignificant temporal characteristics, which can also be inferred by DynamicTriad and HTNE’s inability to generate better representations.










5.4 Parameter Analysis
In this section, we analyze our model’s sensitivity to some parameters that are vital to the modeling of temporal information. We run our model on AMiner, in edge classification to show the performances. For brevity, we denote EPNE-t as our model with only non-periodic patterns, EPNE-f as our model with only periodic patterns, and EPNE as our complete model. All other parameters would be kept as default except those taken as parameters of the investigation.
-
•
Time-frequency Features: We carry out ablation studies on the temporal features we incorporate within our model. Specifically, we compare four models: one without any temporal features, or equivalently, DeepWalk, and three variants of our model: EPNE-t, EPNE-f, and EPNE. As shown in Figure 2(a), our model combining both features outperforms models with either, demonstrating the utility of both features. Besides, EPNE-f performs better than EPNE-t, which indicates that periodic patterns are more indicative of relationships between nodes.
-
•
Smoothness Term: We analyze the performance of our model concerning the weight of smoothness term . As shown in Figure 2(b), a distinctive gap between EPNE-no-smooth, the variant with and EPNE is observed, demonstrating the effectiveness of the smoothness term in keeping embedding spaces aligned across all time steps.
-
•
History Length: We study our model’s performance with respect to the history length . We set history length from 6 to 30 time steps. As shown in Figure 2(c), the performance of our model climaxes at = 12 (48 years) before taking a huge plunge henceforth. It should not be surprising because co-authorship over 48 years ago sheds little light on the present. We thus conclude that a careful selection of is needed to ensure good performance.
-
•
Temporal Weight: We also study the performance with respect to the temporal weight . We test three versions of our model with 0.001, 0.01, 0.1, 1, 10. As shown in Figure 2(d), poor performance is observed when the temporal weight is either negligible or too large, leaving only one feature at work. It is then concluded that our model performs the best with a moderate weight attached to the temporal objective, and as illustrated in the above dataset, a good empirical selection would be .
-
•
Number of Scales: We analyze the influence of the number of scales on our performance with different history length in Figure 2(e). We set ranging from 1 to 5, with set to 24, 48 and 72 years (or 6, 12 and 18 time steps) on AMiner. It can be observed that, when more scales are considered, better performance is generally ensured.
5.5 Network Visualization
We make qualitative evaluations of our models by visualizing embeddings learned from different models. In this section, we compare our model with a representative static model, DeepWalk, along with two temporal models, DynamicTriad and HTNE. We select 2000 researchers randomly from four research fields from the DBLP dataset, whose embedding vectors are then projected into two-dimensional space using PCA and plotted using scatter plots.
The plots for visualization are shown in Figure 3 where dots colored blue, orange, green and red represent researchers from the four fields. It can be shown that the static network embedding method, DeepWalk (Fig. 3(a)) fails to distinguish nodes with different labels, mixing red dots with green ones. By modeling temporal information in networks, temporal models are able to map nodes from different fields with distinctive boundaries. In addition, as shown in Figure 3(e), our model, EPNE, can project nodes with identical fields in a more condensed manner compared to DynamicTriad (Fig. 3(d)) with respect to red and green dots, while minimizing overlapping areas across red, green and blue dots for compared to HTNE (Fig. 3(c)).
6 Conclusion
In this paper, we analyze the temporal evolutionary patterns in real-world networks and demonstrated that such evolutionary patterns would contribute to obtaining more distinctive embedding vectors. We thus propose a novel network embedding model to learn representation vectors preserving not only static network structures, but also evolutionary patterns. As we observe that evolutionary patterns consist of periodic and non-periodic ones, different strategies are designed to learn time-frequency features, followed by a temporal objective to be jointly optimized along with structural objective. We conduct experiments on several datasets and the results in both node and edge classification affirmed our model’s ability to generate satisfactory embeddings with both structural and temporal information.
This work was supported by the National Natural Science Foundation of China (Grant No.61876006).
References
- [1] Shaojie Bai, J. Zico Kolter, and Vladlen Koltun, ‘An empirical evaluation of generic convolutional and recurrent networks for sequence modeling’, CoRR, abs/1803.01271, (2018).
- [2] Kin-Pong Chan and Ada Wai-Chee Fu, ‘Efficient time series matching by wavelets’, in Proceedings 15th International Conference on Data Engineering (Cat. No. 99CB36337), pp. 126–133. IEEE, (1999).
- [3] Charles K Chui, An introduction to wavelets, Elsevier, 2016.
- [4] Peng Cui, Xiao Wang, Jian Pei, and Wenwu Zhu, ‘A survey on network embedding’, CoRR, abs/1711.08752, (2017).
- [5] Lun Du, Yun Wang, Guojie Song, Zhicong Lu, and Junshan Wang, ‘Dynamic network embedding: An extended approach for skip-gram based network embedding.’, in IJCAI, pp. 2086–2092, (2018).
- [6] Aditya Grover and Jure Leskovec, ‘node2vec: Scalable feature learning for networks’, in Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 855–864. ACM, (2016).
- [7] Will Hamilton, Zhitao Ying, and Jure Leskovec, ‘Inductive representation learning on large graphs’, in Advances in Neural Information Processing Systems, pp. 1025–1035, (2017).
- [8] Mark Heimann, Haoming Shen, and Danai Koutra, ‘Node representation learning for multiple networks: The case of graph alignment’, CoRR, abs/1802.06257, (2018).
- [9] Petter Holme, ‘Modern temporal network theory: a colloquium’, The European Physical Journal B, 88(9), 234, (2015).
- [10] Thomas N Kipf and Max Welling, ‘Semi-supervised classification with graph convolutional networks’, arXiv preprint arXiv:1609.02907, (2016).
- [11] Jundong Li, Harsh Dani, Xia Hu, Jiliang Tang, Yi Chang, and Huan Liu, ‘Attributed network embedding for learning in a dynamic environment’, in Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pp. 387–396. ACM, (2017).
- [12] Qingqing Long, Yiming Wang, Lun Du, Guojie Song, Yilun Jin, and Wei Lin, ‘Hierarchical community structure preserving network embedding: A subspace approach’, in Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 409–418, (2019).
- [13] Jianxin Ma, Peng Cui, and Wenwu Zhu, ‘Depthlgp: learning embeddings of out-of-sample nodes in dynamic networks’, in Thirty-Second AAAI Conference on Artificial Intelligence, (2018).
- [14] Rossana Mastrandrea, Julie Fournet, and Alain Barrat, ‘Contact patterns in a high school: a comparison between data collected using wearable sensors, contact diaries and friendship surveys’, CoRR, abs/1506.03645, (2015).
- [15] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean, ‘Distributed representations of words and phrases and their compositionality’, in Advances in neural information processing systems, pp. 3111–3119, (2013).
- [16] Colm Mulcahy, ‘Image compression using the haar wavelet transform’, Spelman Science and Mathematics Journal, 1(1), 22–31, (1997).
- [17] Giang Hoang Nguyen, John Boaz Lee, Ryan A Rossi, Nesreen K Ahmed, Eunyee Koh, and Sungchul Kim, ‘Continuous-time dynamic network embeddings’, in Companion of the The Web Conference 2018 on The Web Conference 2018, pp. 969–976. International World Wide Web Conferences Steering Committee, (2018).
- [18] Aldo Pareja, Giacomo Domeniconi, Jie Chen, Tengfei Ma, Toyotaro Suzumura, Hiroki Kanezashi, Tim Kaler, Tao B. Schardl, and Charles E. Leiserson. Evolvegcn: Evolving graph convolutional networks for dynamic graphs, 2019.
- [19] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena, ‘Deepwalk: Online learning of social representations’, in Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 701–710. ACM, (2014).
- [20] Uriel Singer, Ido Guy, and Kira Radinsky. Node embedding over temporal graphs, 2019.
- [21] Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei, ‘Line: Large-scale information network embedding’, in Proceedings of the 24th International Conference on World Wide Web, pp. 1067–1077. International World Wide Web Conferences Steering Committee, (2015).
- [22] Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka, ‘How powerful are graph neural networks?’, arXiv preprint arXiv:1810.00826, (2018).
- [23] Cheng Yang, Zhiyuan Liu, Deli Zhao, Maosong Sun, and Edward Y Chang, ‘Network representation learning with rich text information.’, in IJCAI, pp. 2111–2117, (2015).
- [24] Jiaxuan You, Yichen Wang, Aditya Pal, Pong Eksombatchai, Chuck Rosenburg, and Jure Leskovec, ‘Hierarchical temporal convolutional networks for dynamic recommender systems’, in The World Wide Web Conference, pp. 2236–2246. ACM, (2019).
- [25] Yizhou Zhang, Guojie Song, Lun Du, Shuwen Yang, and Yilun Jin, ‘Dane: Domain adaptive network embedding’, in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pp. 4362–4368. International Joint Conferences on Artificial Intelligence Organization, (7 2019).
- [26] Jie Zhou, Ganqu Cui, Zhengyan Zhang, Cheng Yang, Zhiyuan Liu, and Maosong Sun, ‘Graph neural networks: A review of methods and applications’, arXiv preprint arXiv:1812.08434, (2018).
- [27] Lekui Zhou, Yang Yang, Xiang Ren, Fei Wu, and Yueting Zhuang, ‘Dynamic network embedding by modeling triadic closure process’, in Thirty-Second AAAI Conference on Artificial Intelligence, (2018).
- [28] Linhong Zhu, Dong Guo, Junming Yin, Greg Ver Steeg, and Aram Galstyan, ‘Scalable temporal latent space inference for link prediction in dynamic social networks’, IEEE Transactions on Knowledge and Data Engineering, 28(10), 2765–2777, (2016).
- [29] Yuan Zuo, Guannan Liu, Hao Lin, Jia Guo, Xiaoqian Hu, and Junjie Wu, ‘Embedding temporal network via neighborhood formation’, in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 2857–2866. ACM, (2018).