
new interpretations of the ‘closeness’ of relationships between
node 1 and other nodes and how these relationships evolve are
now possible.
To further develop this intuition, consider the following
concrete example. Suppose node 1 began publishing as a
Ph.D. student. Thus, papers produced during their candidature
were co-authored with a supervisor (node 3) and one of
the supervisor’s collaborators (node 2). After graduation,
the student (node 1) became a research scientist in a new
organization. As a result, a new collaboration with a senior
supervisor (node 4) is formed. Node 1 was introduced to node
5 through collaborations between the senior supervisor (node
4) and node 5, and later began working with node 6 after
the relationship with node 5 was formed. Similarly, after a
collaboration between node 1 and node 6, node 1 met node
7 because of the relationships between node 5 and node 8,
node 8 and node 7, and node 6 and node 7. Hence, as shown
in Figure 2, both the formation of the relationships between
node 1 and other nodes and the levels of closeness of such
relationships evolve as new connections are made between
node 1 and related collaborators.
Thus, the formation of each edge
(x, y)
changes not only the
relationship between
x
and
y
but also relationships between
nodes in the surrounding neighborhood. Capturing how and why
a network evolves (through edge formations) can produce better
signals in learned models. Recent studies capture the signal
by periodically probing a network to learn more meaningful
embeddings. These methods model changes in the network by
segmenting updates into fixed time windows, and then learn an
embedding for each network snapshot [
10
,
11
]. The embeddings
of previous snapshots can be used to infer or indicate various
patterns as the graph changes between snapshots. In particular,
CTDNE [
12
] leveraged temporal information to sample time-
constrained random walks and trained a skip-gram model
such that nodes co-occurring in these random walks produce
similar embeddings. Inspired by the Hawkes process [
13
] which
shows that the occurrence of current events are influenced by
the historical events, HTNE [
14
] leveraged prior neighbor
formation to predict future neighbor formations. In this paper,
we show how to more effectively embed fine-grained temporal
information in all learned feature representations that are
directly influenced by historical changes in a network.
How to effectively analyze edge formations is a challenging
problem, which requires: (1) identifying relevant nodes (not
limited to direct neighbors) which may impact edge formations;
(2) differentiating the impact of relevant nodes on edge
formations; (3) effectively aggregating useful information from
relevant nodes given a large number of useful but noisy signals
from the graph data.
To mitigate these issues, we propose a deep learning method,
which we refer to as Embedding via Historical Neighborhoods
Aggregation (EHNA), to capture node evolution by probing
the information/factors influencing temporal behavior (i.e.,
edge formations between nodes). To analyze the formation
of each edge
(x, y)
, we apply the the following techniques.
First, a temporal random walk is used to dynamically identify
relevant nodes from historical neighborhoods that may influence
new edge formations. Specifically, our temporal random walk
algorithm models the relevance of nodes using a configurable
time decay, where nodes with high contextual and temporal
relevance are visited with higher probabilities. Our temporal
random walk algorithm can preserve the breadth-first search
(BFS) and depth-first search (DFS) characteristics of traditional
solutions, which can then be further exploited to provide both
micro and macro level views of neighborhood structure for
the targeted application. Second, we also introduce a two-level
(node level and walk level) aggregation strategy combined
with stacked LSTMs to effectively extract features from nodes
related to chronological events (when edges are formed). Third,
we propose a temporal attention mechanism to improve the
quality of the temporal feature representations being learned
based on both contextual and temporal relevance of the nodes.
Our main contributions to solve the problem of temporal
network representation learning are:
•
We leverage temporal information to analyze edge forma-
tions such that the learned embeddings can preserve both
structural network information (e.g., the first and second
order proximity) as well as network evolution.
•
We present a temporal random walk algorithm which
dynamically captures node relationships from historical
graph neighborhoods.
•
We deploy our temporal random walk algorithm in a
stacked LSTM architecture that is combined with a
two-level temporal attention and aggregation strategy
developed specifically for graph data, and describe how
to directly tune the temporal effects captured in feature
embeddings learned by the model.
•
We show how to effectively aggregate the resulting
temporal node features into a fixed-sized readout layer
(feature embedding), which can be directly applied to
several important graph problems such as link prediction.
•
We validate the effectiveness of our approach using four
real-world datasets for the network reconstruction and link
prediction tasks.
Organization of this paper.
Related work is surveyed in
Section II and our problem formulation is presented in Sec-
tion III. Next, we describe the problem solution in Section IV.
We empirically validate our new approach in Section V, and
conclude in Section VI.
II. R
ELATED WORK
Static network embedding.
Inspired by classic techniques
for dimensionality reduction [
15
] and multi-dimensional scal-
ing [
16
], early methods [
15
,
17
,
18
,
19
] focused on matrix-
factorization to learn graph embeddings. In these approaches,
node embeddings were learned using a deterministic measure
of graph proximity which can be approximated by computing
the dot product between learned embedding vectors. Methods
such as D
EEP
W
ALK
[
3
] and N
ODE
2V
EC
[
1
] employed a more
flexible, stochastic measure of graph proximity to learn node
embeddings such that nodes co-occurring in short random
1118
Authorized licensed use limited to: Tsinghua University. Downloaded on June 01,2021 at 15:48:53 UTC from IEEE Xplore. Restrictions apply.
评论