Learning, Analyzing and Predicting Object Roles on Dynamic Networks Kang Li ,Suxin Guo

advertisement
2013 IEEE 13th International Conference on Data Mining
Learning, Analyzing and Predicting Object Roles on
Dynamic Networks
Kang Li∗ ,Suxin Guo† ,Nan Du‡ , Jing Gao§ and Aidong Zhang¶
Department of Computer Science and Engineering
The State University of New York at Buffalo
Emails: {kli22∗ ,suxinguo† , nandu‡ , jing§ and azhang¶ }@buffalo.edu
types of object roles of interest. [1] defines vulnerable nodes as
the nodes that the deletion of them could cause maximum network fragmentation. [2] assumes important objects should have
high PageRank scores. Generally, there are two drawbacks of
such methods. First, the topological properties in each task are
subjectively selected, thus critical information characterizing
the object roles in the networks may be missed. Even worse,
when object roles are complex, the aforementioned strategy
may not be able to characterize the object roles using any
existing topological property. Second, existing methods specified for static networks ignore the influence of the evolving
vertices and links in dynamic networks, thus are not applicable
for analyzing the dynamic patterns of object roles. To sum up,
these existing methods are not effective in mining the object
roles and their evolving patterns in dynamic networks.
Abstract—Dynamic networks are structures with objects and
links between the objects that vary in time. Temporal information
in dynamic networks can be used to reveal many important
phenomena such as bursts of activities in social networks and
human communication patterns in email networks. In this area,
one very important problem is to understand dynamic patterns
of object roles. For instance, will a user become a peripheral
node in a social network? Could a website become a hub on the
Internet? Will a gene be highly expressed in gene-gene interaction
networks in the later stage of a cancer? In this paper, we propose
a novel approach that identifies the role of each object, tracks
the changes of object roles over time, and predicts the evolving
patterns of the object roles in dynamic networks. In particular,
a probability model is proposed to extract latent features of
object roles from dynamic networks. The extracted latent features
are discriminative in learning object roles and are capable of
characterizing network structures. The probability model is then
extended to learn the dynamic patterns and make predictions
on object roles. We assess our method on two data sets on the
tasks of exploring how users’ importance and political interests
evolve as time progresses on dynamic networks. Overall, the
extensive experimental evaluations confirm the effectiveness of
our approach for identifying, analyzing and predicting object
roles on dynamic networks.
I.
In this paper, we propose a novel probability model to
address the problems of learning, analyzing and predicting
object roles in dynamic networks. Different from the aforementioned existing methods, the proposed model does not rely on
any subjectively selected topological properties, and is capable
of characterizing network dynamics. At the heart of our
framework is the latent feature representation of object roles,
which captures the structural information of each time point,
incorporates evolving information in dynamic networks, and
is discriminative in learning object roles. Specifically, at each
time point, objects of similar roles have close latent feature
representations, and the interactions of the latent object roles
can effectively reconstruct the observable network of the time
point. Moreover, the extracted latent feature representation can
well fit the existing supervised information in the proposed
model. Furthermore, to incorporate the evolving information in
the dynamic networks, the latent feature representation of each
time point is close to the prediction obtained from the previous
time points. This representation of object roles can be used
to build a variety of sophisticated analysis tools for dynamic
networks. We utilize the effective representation strategy for
exploring the evolution of object roles. This representation
method endues the proposed approach with simple and direct
visualizations that clearly show how the roles of individual
objects evolve as time progresses.
I NTRODUCTION
Dynamic networks exist in many different settings, such as
computer networks, social networks, biological networks and
sensor networks. During the formation and the dynamics of
these systems, objects usually play various types of changing
roles. For example, in Facebook, some users play as topic
hubs which are highly involved in various activities while some
other users play as peripheral objects who seldom participate
in discussions. As time goes by, highly involved users may
reduce their engagement with Facebook due to the increased
engagement with other social networks, and peripheral users
may become more active in the communities.
In the research of dynamic networks, a critical problem
is to understand object roles and their evolving patterns.
This problem is meaningful in many applications and has
attracted much attention. For instance, detecting users having
abnormal roles can be used to filter spammers in email
networks, analyzing the dynamics of customer engagement
in social networks can help improve the quality of service,
and predicting the informative genes in gene-gene interaction
networks is essential for preventing the onset of cancers.
Although mining dynamic networks using latent feature
representation on objects has been a hot topic recently, the proposed work in this paper significantly differs from the existing
work in both the method and the aimed task. In the existing
methods, exponential-family random graph models (ERGMs)
are the canonical way for representing observable networks
with latent variables. As discussed in [3], ERGMs usually
Nevertheless, existing methods for this problem usually
focus on static networks and make strong assumptions on the
relationships between specific topological properties and the
1550-4786/13 $31.00 © 2013 IEEE
DOI 10.1109/ICDM.2013.95
428
suffer from high computational and statistical cost thus hardly
work in practice. To avoid the drawbacks of ERGMs, several
alternative approaches [4]–[6] have been proposed in recent
years to use latent feature vectors as ”coordinates” to represent
the characteristics of each node. Compared with the proposed
probability model in this paper, these ”coordinates” methods
always assume that highly connected objects have close latent
feature representations and vice versa. Since objects of the
same roles are not necessarily highly connected or may be
even far away from each other, these existing methods can not
encode object roles as latent feature representations in many
applications and are not effective in the learning of object roles
in dynamic networks.
TABLE I: Notation
t
ni
r
c
G
Gi
Vni ×1
i
i
En
i ×ni
i
Yn ×c
i
Λni ×r
i
Hr×r
W
the Hadamard product of two matrices A and B of the same
size, and (A◦B)ij = Aij ·Bij . Similarly, (AB)ij = Aij /Bij
is the Hadamard division. Besides, a Gaussian distribution of
A with mean B and variance σ 2 is denoted as N (A|B, σ 2 ).
The proposed approach is built upon two bases: first, object
roles can be extracted from structural information; and second,
object roles should change gradually rather than abruptly. The
second base means that previous data can provide hints about
the current status of the network, and this assumption has
been widely used in dynamic network mining such as [4]–
[6]. In experiments, we investigate how people’s importance
and political interests evolve in dynamic networks. The results
well support these two bases.
We summarize the major variable matrices used in this
paper in Table I. Each dynamic network is assumed to be
periodically sampled into t snapshots, and denoted as G =
{Gi |i ∈ [1, t]}. Gi = {Vnii ×1 , Eni i ×ni } is the i-th snapshot,
where V i is the set of ni vertices, and E i is the set of links
between the vertices. In our context, each vertex is an object.
In the dynamic network G, we suppose there are c object
role classes, and the label matrix for the objects in Gi is Ynii ×c .
In the paper, we suppose labels of objects are provided in the
first snapshot, and focus on the following three tasks:
Overall, the contributions of this work include:
• In Section II-B, a Gaussian model is proposed to effectively extract latent features of object roles for both
supervised and unsupervised cases. A solution based
on variational Bayesian inference is then provided to
efficiently optimize the proposed model.
• In Section II-C, we extend the Gaussian model to incorporate dynamic information for learning and analyzing
object roles in dynamic networks.
• In Section II-D, we also provide the details of implementing the proposed model for predicting object roles
according to the learned evolving patterns.
• In Section III, we experimentally evaluate the proposed
methods on the tasks of mining people’s evolving importance and political interests in dynamic networks. The
overall performance well confirms the effectiveness of the
proposed model.
II.
number of snapshots in the dynamic network
number of objects in the i-th snapshot
number of features in latent feature representations of object roles
number of object role classes
the dynamic network
the i-th snapshot of the dynamic network
the set of vertices in the i-th snapshot
the set of links in the i-th snapshot
the label matrix for each node
the latent feature matrix of object roles in the i-th snapshot
the interaction matrix of object roles in the i-th snapshot
the coefficient matrix for learning object role classes
• Learning the role of each object at each snapshot of the
dynamic network;
• Analyzing how object roles evolve over time; and
• Predicting the object roles at the (t + 1)-th snapshot of
the dynamic network using the first to the t-th snapshots.
B. Detection on Static Networks
In this section, we propose a probability model for object
role detection on static networks, as a ground for the dynamic
object role analysis.
Different from the existing methods [4]–[6] that view
object roles as object ”coordinates”, we interpret the role of
each object as its properties that determine to what extent
and how the object impacts the other objects in the formation
and the dynamics of the network.
M ETHODOLOGY
In this section, we present the details of the proposed
LAP (Learning, Analyzing and Predicting) model for mining
object roles on dynamic networks. Specifically, we start from
detecting object roles on static networks, then extend the
developed model to dynamic cases. The prediction of object
roles is achieved through the extended model.
To better explain the intuition of this definition and its
difference from the ”coordinates” concept, we give an example
on the trade history of the ancient Tamil country [7] which is a
region in southern India. During the ancient time, people there
were frequently involved in both local and international, and
both inland and overseas trade. We conclude the rules of the
trade as follows.
A. Notation
1) The closer two objects were, the more likely they could
trade/interact. In the Ancient Tamil, most trade was by
barter, which was prevalent locally. As a result, more
trades were performed locally than internationally.
2) Objects with larger activity range had higher ability to
interact with other objects. As an evidence, the development of seamanship and the discovery of new routes significantly increased the trades between Tamil and Rome.
3) The properties of two objects determined how they interacted. For instance, on the trades between Tamil and
We first introduce the notation rule of this paper. Without
further notification, a scalar is denoted by a lower case letter,
e.g., a, b and λ; a matrix is denoted by an upper case letter such
as A, B and Λ. An×m represents that the matrix A contains
n rows and m columns. Besides, Ai,: and A:,j represent the
i-th row and the j-th column of the matrix A, respectively,
and Aij is the element at the i-th row and the j-th column.
In the formulation of our model, T r(U
s×s ) is the trace of
s
the square matrix Us×s and T r(Us×s ) = i=1 Uii . A ◦ B is
429
The posterior of Λ and H is then:
Rome, according to the different goods they could produce, Tamil exported pepper, ivory and gold, and imported
glass, coral and wine. Another example is that the changes
of the Emperor of Rome had significant impact on the
trade between Tamil and Rome.
p(Λ, H|E, σ 2 ) =
p(E|Λ, H, σ 2 )p(Λ)p(H)
.
p(E)
(1)
By the model in Eq.1, we can extract the latent feature
representation Λ of object roles as well as the role interaction
matrix H from an observable link matrix E of a static network.
In a network representation of the trades of the Ancient
Tamil, each node denotes an object that participated in the
trade, and each link represents a trade; whether there is a
link between any two objects is determined by the rule 1 and
the rule 2; and the weights of the links, which represent the
interaction types of the objects, are determined by the rule 3.
The Supervised Model on Learning Object Roles
In our experiment setting, we assume at the first time point,
there are several labeled objects to guide the learning of each
type of object roles and to maximize the margins between
different object roles.
In this process, there are two factors impacting the formation of the network: object coordinates and object roles. The
coordinates of objects, covering rule 1, determine the closeness
of each two objects. The roles of objects, covering rule 2 and
rule 3, determine to what extent and how each object impacts
the others. The objective of this paper is to explore how to
detect, analyze and predict such object roles.
For the first snapshot, the label matrix for the labeled
objects is Ŷm×c , in which m is the number of the labeled
objects. To extend the unsupervised model for supervised
cases, we introduce a feature coefficient matrix Wr×c which
measures the contribution of each feature in Λ to the object
role classes:
m c
N Yˆij |(Λ̂ · W )ij , σY2 .
(2)
p(Ŷ |Λ, W ) =
The Unsupervised Model on Learning Object Roles
i=1 j=1
Suppose object coordinates are a-dimensional and object
roles are r-dimensional. Let Ξn×a and Λn×r be the latent
coordinate matrix and latent role matrix for n objects, respectively. An observed link Eij from a source object i to a target
object j is generated as:
In Eq.2, Λ̂ denotes the related latent role features for the
labeled objects. σY2 is the noise variance in the Gaussian
distribution.
The priors for the feature
weighting
matrix−1W is:
2
c
tr(W CW W )
i
=
exp
−
.
p(W ) ∝ exp − i=1 W
2CWi
2
In the prior, W is column-wise independent and CW =
diag{CW1 , CW2 , ..., CWc } is the covariance matrix of the
prior.
Eij = Λi,: · M→ · Λj,: · M← · K(Ξi,: , Ξj,: ) + .
In the above equation, M→ measures how the role Λi,:
of the source object i impacts the link Eij ; and similarly,
M← measures how the role Λj,: of the target object j impacts
the link Eij . K(Ξi,: , Ξj,: ) is a closeness function measuring
how close two objects (object i and object j) are. The larger
K(Ξi,: , Ξj,: ) is, the more likely two objects i and j interact.
During the past few years, extensive work has been done to
estimate the latent closeness of two objects in networks. In this
paper, we use the truncated Katz kernel [8] for this purpose.
By this
rkernel, the closeness matrix K on a network E is
K = i=1 αi · E i , and K(Ξi,: , Ξj,: ) = Kij .
For the supervised cases, the posterior is:
p(Λ, H, W |E, Ŷ ) =
p(E|Λ, H)p(Ŷ |Λ, W )p(Λ)p(H)p(W )
p(E, Ŷ )
.
(3)
Given E and Ŷ , by maximizing Eq.3, we are able to obtain
the latent feature presentation Λ of object roles, the interaction
matrix H and the coefficient matrix W for learning object role
classes on Λ.
If we assume each Eij is independently generated
from a Gaussian distribution with mean Λi,: · M→ · Λj,: ·
M← · Kij and noise variance σ 2 , the conditional distribu2
tion
nthe observed link matrix E2 is: p(E|Λ, H, σ ) =
n of
i=1
j=1 N (Eij |(ΛHΛ ◦ K)ij , σ ), in which H = M→ ·
M← is the object interaction matrix measuring how each object
contributes to each link.
Solution of the Supervised Model
1) Variational Bayesian Inference: In this part, we apply
the Variational Bayesian technique [9], [10] to maximize Eq.3.
Suppose there is a trial distribution on the matrices
Λ,
H and W as Q(Λ, H, W ) which has the constraint
Q(Λ, H, W )dΛdHdW = 1. The free energy of the system
is:
Suppose the priors of the latent feature matrix Λ of the
object roles and the object interaction matrix H follow the
spherical Gaussian distributions and they
are column-wise
Λ 2
r
independent, we have: p(Λ) ∝ exp − f =1 2CfΛ
=
f
−1 2
tr(ΛCΛ Λ )
H r
, and p(H) ∝ exp − f =1 2CfH
=
exp −
2
f
−1
tr(HCH H )
.
exp −
2
F = EQ(Λ,H,W ) [log p(E, Ŷ , Λ, H, W ) − log Q(Λ, H, W )],
(4)
which can be further formulated into:
F =EQ(Λ,H,W ) [log p(Λ, H, W |E, Ŷ ) + log p(E, Ŷ )
− log Q(Λ, H, W )]
= log p(E, Ŷ ) − KL(Q(Λ, H, W )p(Λ, H, W |E, Ŷ ))
In the above priors, CΛ = diag{CΛ1 , CΛ2 , ..., CΛr } and
CH = diag{CH1 , CH2 , ..., CHr } are the prior variances of
those columns.
≤ log p(E, Ŷ ).
430
(5)
In Eq.5, Eq (f (x)) denotes the expectation of the role f (x)
q(x)
with respect to a distribution q(x). KL(qp) = q(x) p(x)
dx
represents the Kullback-Leibler (KL) divergence of a distribution p with regards to q. Since the value of a KL divergence is
always non-negative, in the end of the above formulation, we
can conclude that the lower bound for the evidence log p(E, Ŷ )
is F.
where ρ is a constant which is irrelevant to Λ. δ(Ŷ , i) is the
indicator function that equals 1 if the labeled set includes
object i and equals 0 otherwise.
Comparing with Eq.6, the derivation leads to the following
rules:
n
1
−1
exp − (Λi,: − Λi,: ) ΦΛ (Λi,: − Λi,: ) , (9)
Q(Λ) ∝
2
i=1
Through maximizing the free energy F, we can obtain the
optimum if and only if Q(Λ, H, W ) = p(Λ, H, W |E, Ŷ ). It
is usually intractable to approximate Q(Λ, H, W ) due to the
complexity caused by the high dimensionality of the interacted
members Λ, H and W . Therefore, we apply the variational
approximation Q(Λ, H, W ) = Q(Λ)Q(H)Q(W ).
where
Λi,:
By the variational Bayesian framework, the variational
posteriors of the matrices Λ, H and W can be iteratively
updated through the following rules:
1
exp EH,W log p(E, Ŷ , Λ, H, W ) ,
Q(Λ) =
zΛ
1
exp EΛ,W log p(E, Ŷ , Λ, H, W ) , (6)
Q(H) =
zH
1
exp EΛ,H log p(E, Ŷ , Λ, H, W ) ,
Q(W ) =
zW
Φ−1
Λi
The variational posteriors for the factor matrix H is computed in a similar way, by which we obtain:
r
1
exp − (H:,j − H :,j ) Φ−1
(H
−
H
)
,
Q(H) ∝
:,j
j,:
H
2
j=1
(11)
where
n
1 H :,j = ΦHj
Êij Λi,: ,
2
σE
i=1
(12)
n
1 −1
−1
ΦHj = CH + 2
(Λi,: Λi,: + ΦΛi ).
σE i=1
wherezΛ , zH and zW are constants which enforce the conditions Q(Λ)dΛ = 1, Q(H)dH = 1 and Q(W )dW = 1.
In the iterative solution, the conditional probability over the
observed link p(E|Λ, H) will introduce a term ΛHΛ ΛHΛ
which includes the quadratic multiplications of H and the
quartic integrations of Λ. This term may cause high time and
space complexities. Therefore, in each iteration, we utilize the
Λ∗ optimized in the last iteration as:
p(E|Λ, H) =
n r
2
N ((Ê)ij |Λi,: H:,j , σE
),
⎞
r
c
δ(
Ŷ
,
i)
1
=⎝ 2
Êij H :,j +
Ŷi,k W :,k ⎠ ΦΛ ,
σE j=1
σY2
k=1
⎛
r
1 = ⎝CΛ−1 + 2
(H :,j H :,j + ΦHj )
σE j=1
c
δ(Ŷ , i) +
(W :,k W :,k + ΦWk )) .
σY2
k=1
(10)
⎛
Similarly, the variational posterior of W is computed
through:
c
1
Q(W ) ∝
exp − (W:,k − W :,k ) Φ−1
(W
−
W
)
,
:,k
:,k
W
2
k=1
(13)
where
n
1 W :,k = ΦWk
δ(Ŷ , i)Ŷik Λi,: ,
σY2 i=1
(14)
n
1 −1
−1
Φ Wk = C W + 2
δ(Ŷ , i)Ŷik (Λi,: Λi,: + ΦΛi ).
σY i=1
(7)
i=1 j=1
in which Ê = E K/Λ∗ .
To verify the correctness of this simplification is to prove
that the optimal Λ and H obtained by this simplification
can also satisfy the optimization of the original conditional
probability. The proving process is straightforward. Due to the
limited pages of this paper, we skip the details here.
Plugging in the model for p(Λ, H, W |E, Ŷ ), we can compute the variational posterior Q(Λ) by:
EH,W log p(E, Ŷ , Λ, H, W )
⎡
⎛
r
n
1 ⎣
1 =−
Λi,: ⎝CΛ−1 + 2
EH (H:,j H:,j
)
2 i=1
σE j=1
c
δ(Ŷ , i) +
EW (W:,k W:,k ) Λi,:
(8)
σY2
k=1
⎡
⎛
r
n
1 1 ⎣
−2Λi,: ⎝ 2
−
Êij EH (H:,j )
2 i=1
σE j=1
c
δ(Ŷ , i) +
+ ρ,
Ŷi,k EW (W:,k )
σY2
By simply updates Λ, H and W through the above derivations iteratively until convergence, we can obtain the optimal
solution.
2
, σY2 ,
2) Parameter Setting: To set the hyper-parameters σE
CΛ , CH and CW , we take derivatives of the expectation of the
logarithm evidence EΛ,H,W log p(E, Ŷ , Λ, H, W ) with respect
to each of the hyper-parameters and set the derivatives to 0,
then we can obtain:
n
r
1 2
2
Êij − 2Êij Λi,: H :,j
=
σE
n · r i=1 j=1
(15)
+T r (ΦΛ + Λi,: Λi,: )(ΦH + H :,j H :,j ) ,
k=1
431
Ȧͳ
Ȧʹ
Ȧ–
ͳ
ʹ
–
ͳ
ʹ
ͳ
Since p(E i |Di ) ≡ p(E i |Λi , H i ), p(E i |Di ) can be well
approximated by Eq.7. The only thing remaining unsolved is
how to estimate p(Di |D1:i−1 ) ≡ p(Λi , H i |Λ1:i−1 , H 1:i−1 ).
We place Gaussian priors on Λi and H i as:
p(Λi |ΩiΛ , ΣiΛ ) =
–
j=1
p(H i |ΩiH , ΣiH ) =
–
ʹ
1
n c
δ(Ŷ , i) Ŷik2 − 2Ŷik Λi,: W :,k
Ŷ c i=1 k=1
+T r (ΦΛ + Λi,: Λi,: )(ΦW + W :,k W :,k ) ,
n
1 2
(ΦΛi )l,l + Λi,l
n i=1
r
1 2
(ΦHj )ll + H il
=
r j=1
C Wl
(16)
(19)
i
N (H:,j
|ΩiH:,j , ΣiHj I).
(20)
Let Λi for i ∈ [1, t] as a time series matrix independent of
other variables. When i = 2, since we only have Λ1 and no
change of Λ has been observed, the best prediction of Ω2Λ is
Λ1 . For i ≥ 3, we can approximate ΩiΛ through the MAR(1)
(Multivariate Autoregressive) model [11] as:
Λi = Λi−1 · AΛ + iΛ ,
(17)
(21)
where i is a Gaussian noise having zero mean and precision ΣiΛ . Suppose XΛi = [Λ1 , Λ2 , ..., Λi−2 ] and BΛi =
[Λ2 , Λ3 , ..., Λi−1 ], by maximum likelihood estimation, we
have:
(22)
AΛ = (XΛi XΛi )−1 XΛi BΛi ,
r
1 2
=
(ΦWl )ll + W hl
r
h=1
C. Detection and Analysis on Dynamic Networks
ΩiΛ = Λi−1 · AΛ ,
In this section, we extend the above probabilistic model for
detecting and analyzing object roles on dynamic networks.
ΣiΛ =
i
Let E denote the observed link matrix at the i-th sampled
time point. Suppose we have a sequence of such sampled
link matrices {E 1 , E 2 , ..., E t }, our goal of analyzing network dynamics is to obtain a sequence of low-rank matrices
{{Λ1 , H 1 }, {Λ2 , H 2 }, ..., {Λt , H t }}. The i-th pair {Λi , H i },
which approximates E i , captures the role of each object at
the i-th time point and the interaction patterns of these object
roles.
(23)
1
(B i − XΛi AΛ ) (BΛi − XΛi AΛ ).
i − 2 − r2 Λ
i
i
Similarly, XH
= [H 1 , H 2 , ..., H i−2 ] and BH
[H 2 , H 3 , ..., H i−1 ]. For H we have:
i
i −1 i
i
XH
) X H BH
,
AH = (XH
ΩiH
ΣiH =
The Object Role Model for Dynamic Networks
Let Di = {Λi , H i } denote the pair of the low-rank
matrices we wish to extract at the i-th time point. We present
the graphic model of our analyzing framework in Fig.1. The
observed network E i at the i-th snapshot is generated by
the latent parameters Di , and Di is determined by three
factors: the previous latent parameters D1:i−1 (D1 to Di−1 ),
the current object roles Λi and the current object interaction
pattern H i .
=H
i−1
· AH ,
1
i
i
i
(B i − XH
AH ) (BH
− XH
AH ).
i − 2 − r2 H
(24)
=
(25)
(26)
(27)
Since Λi and Hi are independent from each other, we
estimate p(Di |Di−1 ) as:
p(Di |Di−1 ) p(Λi |ΩiΛ , ΣiΛ )p(H i |ΩiH , ΣiH )
r
r
(28)
i
=
N (Λi:,j |ΩiΛ:,j , ΣiΛj I)
N (H:,j
|ΩiH:,j , ΣiHj I).
j=1
j=1
Thus, the evidence we wish to maximize is:
In the first snapshot, no previous latent parameters D1:i−1
exist. The object roles in E 1 can be effectively extracted
through the models proposed in Section II-B. In this paper,
we assume that labeled information exists in the first snapshot,
therefore we use the supervised model in Section II-B.
p(Λi , H i , E i , Di−1 ) p(E i |Λi , H i )p(Λi )p(H i )
p(Λi |ΩiΛ , ΣiΛ )p(H i |ΩiH , ΣiH ).
(29)
The Eq.29 is the model for learning object roles on
dynamic networks. Specifically, given previously extracted
feature D1:i−1 and the current observable network E i , we can
learn the current latent object roles Λi and their interaction
pattern H i .
i−1
has already been
At each snapshot i for i ∈ [2, t], D
calculated and E i is observable. With the assumption that
D1:i−1 and E i are independent, the posterior distribution of
on the parameter set Di is:
p(Di |D1:i−1 , E i ) = p(Di |D1:i−1 )p(Di |E i ).
N (Λi:,j |ΩiΛ:,j , ΣiΛj I),
In the above Gaussian priors, ΩiΛ and ΩiH are the
best estimations of Λi and H i based on D1:i−1 , respectively. ΣiΛ = diag{ΣiΛ1 , ΣiΛ2 , ..., ΣiΛr } and ΣiH =
diag{ΣiH1 , ΣiH2 , ..., ΣiHr } are the related variance matrices.
CΛ l =
C Hl
r
j=1
Fig. 1: The Graphical Model Representation
σY2 =
r
Solution of the Model for Dynamic Networks
(18)
432
Since E 1:t are all observable and Λt has already been extracted, by maximizing the logarithm of the above posterior
probability, we obtain Λt+1 = Λt · AΛ .
1) Variational Bayesian Inference: To optimize the objective in Eq.29, we follow an optimizing process which is almost
identical to Section II-B1, thus we skip the inference process
here. In this solution, we iteratively update:
r
r
i
1 i i ΩΛj,h
i
Λj,: =
ΦiΛ ,
Êjh H :,h +
2
σE
ΣiΛh
h=1
h=1
r
1
−1
i −1
i −1
i
(H :,h H :,h + ΦHh )
(Φ )Λj = CΛ + (ΣΛ ) + 2
σE
h=1
(30)
⎛
⎞
n
r
i
Ω
1
i
i
H
g,h
⎠,
H :,h = ΦiHh ⎝ 2
Ê i Λ +
σE j=1 jh j,: g=1 ΣiHh
(31)
n
1 i i
−1
i −1
i −1
i
(Φ )Hh = CH + (ΣH ) + 2
(Λ Λ + ΦΛj ).
σE j=1 j,: j,:
By sequential inference, we canpredict the object roles in
s
the (t + s)-th snapshot as Λt+s = j=1 Λt · AΛ .
With the predicted Λt+s , the object role probability in the
(t + s)-th snapshot is estimated as p(Fjt+s |Vit+s ) = Λt+s
i,: W:,j .
We can then obtain the role of each object following the
method at the end of Section II-C. .
III.
In this part, we experimentally evaluate the proposed LAP
algorithm on two real data sets: SocialEvolution [12] and
Robot.Net. The SocialEvolution data set is available upon
request1 and the Robot.Net data set is publicly available2 .
The experiments include three parts. On each data set, we
first evaluate the performance of LAP on the task of detecting
object role classes at each time point. Case studies are then
performed to evaluate the correctness of the extracted dynamic
patterns of object roles. To the end, we test the performance
of the proposed LAP algorithm on the task of object role
predictions. In the experiments, baselines are investigated
to quantitatively prove the superiority of the proposed LAP
algorithm.
2) Parameter Setting: By the derivatives similar to those
in Section II-B2, we can obtain:
n
r
1 i 2
i
i
2
(Êjh ) − 2Êjh Λj,: H :,h
=
σE
n · r j=1
(32)
h=1
i i
i
i
i
i
+T r (ΦΛ + Λi,: Λj,: )(ΦH + H :,h (H :,h ) ) ,
CΛi l =
i
CH
=
l
1 i
i
(ΦΛj )l,l + (Λj,l )2
n j=1
n
r 1
i
(ΦiHg )ll + (H gl )2
r g=1
A. Experiments on SocialEvolution Data Set
(33)
1) Dataset Description: We first evaluate LAP on the
SocialEvolution data set to demonstrate its capabilities in
analyzing the evolving patterns of people’s political interests
in dynamic networks. This data set was collected from October
2008 to May 2009, and it contains information of locations,
phone calls, music sharing logs, surveys on relationships,
political interests and etc.. In the experiments, we build the
dynamic networks of the SocialEvolution data set as: initially
the weights of links between the objects (people in the data set)
are set to 0 which stands for no link. If there is a phone call,
message or music share between two users at a time point, we
add the weight of the undirected link between the two users
by 1 for that time point. The constructed continuous dynamic
network is then divided into 5 snapshots which are denoted as
E 1 to E 5 , respectively.
Detection and Analysis on Object Role Classes
To make the final decision on the role label of each node,
we estimate p(Fj |Vi ), which is the conditional probability that
object Vi belongs to the role class Fj , through the estimation:
p(Fj |Vi ) = Ep(Λ),p(W ) Λi,: W:,j = Λi,: W :,j .
E XPERIMENTS AND A NALYSIS
(34)
By Eq.34, we can obtain the probability of each object
belonging to each object role class at each time point. For the
detection of object roles, each object is then assigned to the
class in which the object has the highest probability. For the
analysis of object roles, the trend of varying p(Fj |Vi ) over
different time points clearly reveals the dynamic patterns of
object roles. We explain more about how to use p(Fj |Vi ) in
the analysis of dynamic object roles in Section III-A3.
In this data set, surveys of some users’ political interests
at different time points have also been provided. According to
the surveys on whether people are interested in politics, we
divide the users into three classes: Indifferent, Moderate and
Enthusiastic. In the experiments, we suppose the surveys at the
first time point are known and use them in the training. The
surveys at the other time points are then used to numerically
evaluate the performance.
D. Prediction on Dynamic Networks
Based on the aforementioned object role detections and
analysis model, in this section, we develop the approach for
predicting object roles on dynamic networks.
2) Object Role Detection: In this part, we investigate the
performance of the proposed LAP algorithm on the task of
identifying object roles at each time point. For comparisons,
we also consider four baselines in this experiment as follows.
Suppose there are t + 1 periodically sampled snapshots
of a dynamic network, and among them, the first t snapshot
networks are observable. We predict the object roles at t +
1 based on the observable data by maximizing the posterior
probability below:
p(Λt+1 |E 1:t ) =
p(Λt+1 |Λt )p(Λt |E 1:t ).
(35)
As discussed in Section I, subjectively selected structural
properties are usually insufficient to cover the characteristics
1 http://realitycommons.media.mit.edu/socialevolution.html
2 http://www.trustlet.org/datasets/robots
Λt
433
net/
ROC curve
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
0.9
0.8
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
1
1
True Positive Rate
0.6
0.8
ROC curve
ROC curve
1
0.9
True Positive Rate
True Positive Rate
True Positive Rate
0.7
0
ROC curve
1
0.8
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
True Positive Rate
ROC curve
1
0.9
0.9
0.8
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
0.9
0.8
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
False Positive Rate
False Positive Rate
False Positive Rate
False Positive Rate
False Positive Rate
E1
E2
E3
E4
E5
(a)
(b)
(c)
(d)
(e)
0.9
1
Fig. 2: ROC curves for detecting Indifferent objects on SocialEvolution
TABLE II: Object Role Detection on SocialEvolution
of the aimed object roles. To prove it, we test two popular
topological properties PageRank [2] and Katz Centrality [13]
in the detection of object roles at each time point.
As explained in Section II-B, we assume that object roles
describe how and to what extent could the objects impact
the others. In the contrast, existing methods such as [4]–[6]
view object roles as the ”coordinates” of the objects, and they
assume objects of the same roles are highly connected. To
measure the impact of these two different assumptions, we
build a baseline named as Coordinates which is identical to
the proposed LAP model
H, σ 2 )
n that it defines p(E|Λ,
n except
2
as: p(E|Λ, H, σ ) = i=1 j=1 N (Kij |(ΛHΛ )ij , σ 2 ). By
this modification, highly connected objects tend to obtain close
latent feature representations in the results.
In the proposed LAP model, we predict the object roles at
time t base on the object roles at time 1 to time t − 1. The
major idea of the proposed model in utilizing the evolving
information is that the extracted object roles at time t should
be close to this prediction. However, in most existing methods
such as [6], the extracted object roles at time t are assumed
to be close to the object roles at time t − 1. To measure
the impact of these two different assumptions, we build a
baseline named as Prior which is identical to the proposed LAP
model except that it defines p(D i |Di−1 ) as: p(Di |Di−1 ) p(Λi |Λi−1 , CΛi−1 )p(H i |H i−1 , CH i−1 ). By this modification,
object roles at time t tend to be close to the learned object
roles at time t − 1.
E1
E2
E3
E4
E5
PageRank
0.537
0.581
0.534
0.520
0.572
E1
E2
E3
E4
E5
PageRank
0.586
0.503
0.606
0.552
0.549
E1
E2
E3
E4
E5
PageRank
0.622
0.724
0.661
0.600
0.502
Indifferent
Centrality
Coordinates
0.534
0.816
0.631
0.736
0.653
0.561
0.630
0.654
0.631
0.649
Moderate
Centrality
Coordinates
0.513
1.000
0.675
0.746
0.706
0.523
0.646
0.540
0.643
0.602
Enthusiastic
Centrality
Coordinates
0.540
1.000
0.709
0.649
0.688
0.536
0.597
0.594
0.575
0.505
Prior
1.000
0.890
0.774
0.721
0.795
LAP
1.000
0.890
0.863
0.827
0.882
Prior
1.000
0.811
0.766
0.675
0.741
LAP
1.000
0.811
0.784
0.700
0.728
Prior
1.000
0.888
0.543
0.569
0.540
LAP
1.000
0.888
0.875
0.843
0.840
performance is still good with the guidance of trained object
roles at E 1 . As time progresses, the guidance of the labeled
information at E 1 becomes weak, thus the performance of
Coordinates is very bad from E 3 to E 5 . The bad performance
on E 3 to E 5 indicates that Coordinates can not characterize the
object roles well without the guidance of labeled information.
This result supports that labeled information is critical for
extracting discriminant representations of object roles, and that
viewing object roles as ”Coordinates” does not work in these
cases.
For each investigated method, we test its performance on
the known labeled data with varying parameter values so as to
find the best parameter setting. In our model, we obtain that
the optimal number of latent features is 63.
An interesting finding in this experiment is that the fourth
baseline Prior performs significantly worse than the proposed
method LAP on detecting the Indifferent and Enthusiastic objects while achieving close performance on Moderate objects.
To seek the reason for this result, we analyze the data and
observe that the probabilities of Indifferent, Moderate and
Enthusiastic objects changing their levels of interests in politics
are on average 24.63%, 11.17% and 23.61%, respectively. This
observation indicates that Moderate objects are more consistent
in political interests while Indifferent and Enthusiastic objects
are more likely to change. Since Prior assumes the object
roles are close to previous object roles, it performs better
on Moderate objects that change less and performs worse on
Indifferent and Enthusiastic objects that change more.
Since the numbers of objects in different classes are highly
imbalanced, accuracy is not meaningful in evaluating the
performance. Therefore, we calculate the Receiver Operating
Characteristics (ROC) curve and use the Area Under the Curve
(AUC) to capture the quality of the ROC curve.
Table II summarizes the performance of all the investigated
methods on the SocialEvolution data set.
We first notice that PageRank and Centrality generally do
not perform well on detecting all the three object roles over
different time points, and in some cases the AUC scores are
even close to random guess. This bad performance indicates
that PageRank and Centrality are not able to capture sufficient
information from the structure to characterize the object roles.
Comparing with all these baseline methods, LAP achieves
the best performance on almost all the experiments. Compared
with the best baseline Prior, LAP improves the AUC scores by
up to 61.14% and on average 13.75%. To better illustrate the
advantage of the proposed LAP model, in Fig.2, we show the
ROC curves on detecting Indifferent objects (due to the space
limit, we only show this case).
For the third baseline Coordinates, the performance varies
significantly over different cases. On E 1 , since labeled information for each object role is provided, Coordinates can
reach very high AUC scores. On E 2 which is close to E 1 , the
434
Enthusiastic
Enthusiastic
Enthusiastic
Enthusiastic
(d) E 4
Moderate
TABLE III: Object Role Prediction on SocialEvolution
Indifferent
(c) E 3
Moderate
Indifferent
(b) E 2
Moderate
Indifferent
(a) E 1
Moderate
Indifferent
Indifferent
Enthusiastic
Moderate
(e) E 5
(d) E 4
Enthusiastic
Enthusiastic
Enthusiastic
Enthusiastic
E1
E2
E3
E4
E5
DyPageRank
0.533
0.606
0.552
0.549
E1
E2
E3
E4
E5
DyPageRank
0.558
0.661
0.600
0.502
Moderate
Indifferent
(c) E 3
Moderate
Indifferent
(b) E 2
Moderate
Indifferent
(a) E 1
Moderate
Indifferent
Indifferent
Enthusiastic
Fig. 3: Role Dynamics of Object 42 in SocialEvolution
Moderate
E
E2
E3
E4
E5
DyPageRank
0.518
0.534
0.520
0.572
1
(e) E 5
Fig. 4: Role Dynamics of Object 70 in SocialEvolution
3) Object Role Analysis: In this section, we do case studies
to demonstrate how we implement the proposed model to
analyze changing patterns of object roles. Without loss of
generality, we use 0, 1 and 2 to represent the classes of
Indifferent, Moderate and Enthusiastic, respectively. With the
class probabilities learned by the proposed LAP model, we
calculate the class number expectation of each object at each
time point. We view the class numbers as angles, scale them
to the range of [0, π], and demonstrate the dynamic patterns
in compasses as in Fig.3 and Fig.4.
Indifferent
DyCentrality
Coordinates
0.632
0.718
0.653
0.696
0.630
0.642
0.631
0.649
Moderate
DyCentrality
Coordinates
0.676
0.777
0.706
0.774
0.646
0.542
0.643
0.602
Enthusiastic
DyCentrality
Coordinates
0.709
0.749
0.688
0.738
0.597
0.566
0.575
0.505
Prior
0.888
0.817
0.739
0.806
LAP
0.888
0.826
0.818
0.908
Prior
0.804
0.769
0.675
0.747
LAP
0.804
0.776
0.642
0.688
Prior
0.867
0.749
0.586
0.561
LAP
0.867
0.765
0.835
0.803
performance is evaluated in AUC by comparing the prediction
with the ground truth. We summarize the results in Table III.
We first notice that the prediction performance of DyPageRank and DyCentrality is very close to the detection
performance of PageRank and Centrality in Table II. This
fact indicates that DyPageRank and DyCentrality are very
effective in learning the evolving patterns of the PageRank and
Centrality scores. Nevertheless, these two baselines perform
the worst among all the investigated methods on the task of
object role predictions. The bad performance verifies the fact
that PageRank and Centrality scores can not well capture the
characteristics of the object roles considered in this experiment.
In Fig.3, we show the changing pattern of object 42.
According to the surveys on political interests, this person
appeared to be highly interested in politics in the first three
time points which are around the date of the U.S. presidential
election at 2008. During the last two time points, this person
was less interested in politics and had doubts about the
president and the congress. In Fig.3, the learned object roles
of this person well reflect the dynamics of his/her interest in
politics.
As for Coordinates, it shows better performance in prediction than in detection. Notice that for object role detection,
Coordinates integrates two kinds of information: the prediction
based on previously learned object roles and the extraction
based on the current network. In contrast, in prediction, it
only uses the previously learned object roles. Therefore, the
extraction part, in which object roles are viewed as object
”Coordinates”, reduces the overall performance of the detection. This experiment again prove that viewing object roles
as ”Coordinates” is not effective in learning, analyzing and
predicting object roles.
In Fig.4, we show the changing pattern of object 70.
Similar to the object 42, this person appeared to be more
interested in politics during the first three time points than
the rest snapshots. This person claimed himself/herself as
slightly interested in politics around the presidential election
date and frequently switched his/her preferred party between
the Independent and the Democrat. During the last two time
points, this person showed no interest in politics at all and
expressed nothing about the president, the congress or the
economy policies. In Fig.4, the learned object roles well reflect
the dynamics of his/her interest in politics.
Similar to the results in Table II, in Table III, Prior
achieves similar performance with the proposed method LAP
on predicting Moderate objects but performs much worse
than LAP on the other object roles. This fact supports the
conclusion that Prior performs better on Moderate objects that
change less and performs worse on Indifferent and Enthusiastic
objects that change more.
Due to space limit, we only show the two cases in this
section. The excellent performance supports that the proposed
LAP model is effective in analyzing and visualizing the
dynamics of object roles.
4) Object Role Prediction: In this section, we investigate
the performance of LAP on the task of predicting object roles.
For comparison, we extend the baselines in Section III-A2
to this task. Specifically, we utilize the results of PageRank
and Centrality in autoregressive model for prediction, and
obtain two baselines DyPageRank and DyCentrality, respectively. Since the other two baselines Coordinates and Prior are
dynamic model, they can be directly used in the task of object
role prediction.
Compared to all the baselines, the proposed LAP model
achieves significantly better performance in most cases than
any baseline. The experiments confirm the effectiveness of
LAP in predicting object roles.
B. Experiments on Robot.Net Data Set
1) Dataset Description: The Robot.Net data set was
crawled daily from the website Robot.Net3 since 2008. This
In this experiment, we predict the object roles at the
second to the fifth time points using the previous data. The
3 http://robots.net/
435
ROC curve
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
0.1
0
1
0
0.1
False Positive Rate
0.2
0.5
0.6
0.7
0.8
0.9
ROC curve
0
0.1
0.2
0.3
0.4
0.5
0.6
0.2
0.7
0.8
0.9
1
False Positive Rate
(d) R4
0.4
0.5
0.6
0.7
0.8
0.9
1
(c) R3
ROC curve
1
0.9
0.8
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0.3
False Positive Rate
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
False Positive Rate
0.9
PageRank
0.511
0.564
0.624
0.540
0.575
0.521
R1
R2
R3
R4
R5
R6
PageRank
0.645
0.556
0.619
0.659
0.585
0.580
R1
R2
R3
R4
R5
R6
PageRank
0.718
0.564
0.654
0.630
0.623
0.585
R1
R2
R3
R4
R5
R6
PageRank
0.704
0.626
0.623
0.590
0.588
0.627
Master
(c) R3
Apprentice
Apprentice
Apprentice
0.7
0.6
0.5
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.3
0.2
0.1
0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
False Positive Rate
(f) R6
Fig. 5: ROC curves for detecting Observer on Robot.Net
TABLE IV: Object Role Detection on Robot.Net
R1
R2
R3
R4
R5
R6
Master
(b) R2
0.8
(e) R5
Observer
Centrality
Coordinates
0.549
0.899
0.537
0.793
0.530
0.502
0.535
0.517
0.537
0.531
0.541
0.511
Apprentice
Centrality
Coordinates
0.677
0.778
0.662
0.775
0.657
0.705
0.658
0.684
0.654
0.671
0.654
0.670
Journeyer
Centrality
Coordinates
0.511
0.911
0.519
0.809
0.511
0.504
0.518
0.522
0.527
0.524
0.528
0.545
Master
Centrality
Coordinates
0.680
0.646
0.682
0.685
0.678
0.726
0.668
0.709
0.665
0.698
0.665
0.745
Master
(a) R1
Observer
PageRank
Centrality
Coordinates
Prior
LAP
0.1
0.1
Observer
0.6
0.2
0
Observer
True Positive Rate
0.7
0.3
0.1
ROC curve
0.8
0.4
0.2
0
1
1
0.9
0.5
0.3
(b) R2
1
True Positive Rate
0.4
PageRank
Centrality
Coordinates
Prior
LAP
0.4
False Positive Rate
(a) R1
0
0.3
0.5
Journeyer
0.2
0.6
Journeyer
0.3
True Positive Rate
0
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.7
Journeyer
0.1
0.5
Apprentice
0.8
Journeyer
0.2
0.6
Apprentice
0.9
Journeyer
0.3
True Positive Rate
True Positive Rate
True Positive Rate
PageRank
Centrality
Coordinates
Prior
LAP
0.4
0.7
Observer
0.5
0.8
Observer
0.6
1
0.9
Observer
0.7
Apprentice
ROC curve
1
0.8
Journeyer
ROC curve
1
0.9
Prior
1.000
0.998
0.853
0.832
0.806
0.810
LAP
1.000
0.998
0.998
0.964
0.960
0.958
Prior
1.000
0.988
0.891
0.878
0.792
0.783
LAP
1.000
0.988
0.988
0.844
0.841
0.844
Prior
1.000
0.982
0.867
0.851
0.720
0.706
LAP
1.000
0.982
0.983
0.927
0.921
0.983
Prior
1.000
0.990
0.794
0.769
0.785
0.784
LAP
1.000
0.990
0.973
0.969
0.969
0.969
Master
Master
Master
(d) R4
(e) R5
(f) R6
Fig. 6: Role Dynamics of Object 99 in Robot.Net
TABLE V: Object Role Prediction on Robot.Net
data set contains the interactions among the users of the
website. For the experiments, we choose the first sampled
snapshot in each year during 2007 to 2012 and denote the
obtained six snapshots as R1 to R6 . In this website, each user
is labeled by the others as Observer, Apprentice, Journeyer
or Master according to his/her importance in the website.
Based on these labels, we divide the users into four object role
classes: Observer, Apprentice, Journeyer and Master. In the
experiments, we use the labeled information at R1 for training,
and test the performance of the proposed LAP on detecting,
analyzing and predicting the above object roles. Using varying
parameter values to test the performance on R1 , we obtain that
the optimal number of latent features is 70.
R1
R2
R3
R4
R5
R6
DyPageRank
0.524
0.517
0.518
0.517
0.518
R1
R2
R3
R4
R5
R6
DyPageRank
0.643
0.644
0.645
0.645
0.645
R1
R2
R3
R4
R5
R6
DyPageRank
0.681
0.678
0.678
0.677
0.677
R1
R2
R3
R4
R5
R6
DyPageRank
0.704
0.691
0.692
0.694
0.694
Observer
DyCentrality
Coordinates
0.556
0.896
0.556
0.793
0.555
0.502
0.555
0.517
0.555
0.531
Apprentice
DyCentrality
Coordinates
0.669
0.767
0.669
0.773
0.668
0.704
0.666
0.682
0.668
0.672
Journeyer
DyCentrality
Coordinates
0.534
0.907
0.534
0.808
0.534
0.504
0.535
0.520
0.535
0.524
Master
DyCentrality
Coordinates
0.670
0.857
0.671
0.791
0.671
0.570
0.672
0.573
0.672
0.575
Prior
0.997
0.998
0.852
0.832
0.806
LAP
0.997
0.998
0.997
0.964
0.960
Prior
0.988
0.988
0.890
0.878
0.791
LAP
0.988
0.986
0.986
0.844
0.841
Prior
0.983
0.983
0.867
0.853
0.712
LAP
0.983
0.981
0.983
0.927
0.921
Prior
0.990
0.989
0.870
0.854
0.772
LAP
0.990
0.989
0.989
0.912
0.907
SocialEvolution in Table II. PageRank and Centrality can not
capture the characteristics of the four object roles thus perform
the worst. Coordinates performs well when close to R1 and
perform much worse after several snapshots. Among all the
baselines, Prior performs the best.
Compared to these baselines, the proposed LAP model
achieves the best in most cases. It outperforms the best baseline
Prior by up to 39.24% and on average 11.51%. To better
illustrate the advantage of the proposed LAP model, in Fig.5,
we show the ROC curves on detecting the Observer objects
(due to the space limit, we show this case only).
3) Object Role Analysis: To evaluate the performance on
analyzing the dynamics of object roles, we do case study on
object 99. Specifically, we use 0 to 3 to represent Observer,
Apprentice, Journeyer and Master, respectively. The class
number estimated by expectation is scaled to the range of
2) Object Role Detection: Table IV summarizes the performance of all the baselines as well as the proposed LAP model
on the Robot.Net.
Overall, the results show similar trends to the results of
436
[0, 3π
2 ]. We summarize the results of the object 99 in Fig.6.
According to the votes, in the first snapshot, four users vote
this object as Apprentice and two vote the object as Journeyer.
Starting from the second snapshot, more and more users vote
the object 99 as Master. As we can observe from Fig.6, this
trend on votes is well reflected by the results of the proposed
LAP model. The case study supports that the LAP is effective
in capturing the dynamics of object roles.
for the proposed model. In experiments, we evaluated the
proposed model through the tasks of learning the dynamics
of people’s importance and political interests in two real
world data sets. Overall, the proposed LAP model significantly
outperforms the baselines on learning and predicting seven
types of object roles. Moreover, the dynamic patterns extracted
by the proposed LAP model well reflect the real changing
states of the object roles.
4) Object Role Prediction: In this part, we predict the
object roles at the second to the sixth time points using the
previous data, and summarize the results in Table V. The
results show similar patterns to those in Table III and Table
IV. Across the four baselines, DyPageRank and DyCentrality
are still not effective in characterizing the four different object
classes; Coordinates significantly suffers from its assumption
of viewing object roles as ”Coordinates”; and Prior performs
the best in the baselines.
VI.
The materials published in this paper are partially supported by the National Science Foundation under Grants No.
1218393, No. 1016929, and No. 0101244.
R EFERENCES
[1]
C. J. Kuhlman, V. S. A. Kumar, M. V. Marathe, S. S. Ravi, and D. J.
Rosenkrantz, “Finding critical nodes for inhibiting diffusion of complex
contagions in social networks,” Proc. of ECML’10, pp. 111–127, 2010.
[2] L. Page, S. Brin, R. Motwani, and T. Winograd, “The pagerank citation
ranking: Bringing order to the web,” World Wide Web Internet And Web
Information Systems, 1998.
[3] M. Handcock, G. Robins, T. Snijders, and J. Besag, “Assessing degeneracy in statistical models of social networks,” Journal of the American
Statistical Association, 2003.
[4] J. R. Foulds, C. DuBois, A. U. Asuncion, C. T. Butts, and P. Smyth,
“A dynamic relational infinite feature model for longitudinal social
networks,” Journal of Machine Learning Research - Proceedings Track,
pp. 287–295, 2011.
[5] C. Heaukulani and Z. Ghahramani, “Dynamic probabilistic models for
latent feature propagation in social networks,” Proc. of ICML’2013, pp.
275–283, 2013.
[6] Y.-R. Lin, Y. Chi, S. Zhu, H. Sundaram, and B. L. Tseng, “Analyzing
communities and their evolutions in dynamic social networks,” ACM
Trans. Knowl. Discov. Data, pp. 8:1–8:31, 2009.
[7] “Economy of ancient tamil country,” http://en.wikipedia.org/wiki/
EconomyofancientTamilcountry, accessed: 2013-06-18.
[8] Z. Lu, B. Savas, W. Tang, and I. S. Dhillon, “Supervised link prediction
using multiple sources,” Proc. of ICDM’10, pp. 923–928, 2010.
[9] S. Nakajima, M. Sugiyama, and R. Tomioka, “Global solution of variational bayesian matrix factorization under matrix-wise independence,”
Proc. of NIPS’10, 2010.
[10] C. Fox and S. Roberts, “A tutorial on variational Bayesian inference,”
Artificial Intelligence Review, pp. 85–95, 2012.
[11] W. Penny and S. Roberts, “Bayesian multivariate autoregressive models
with structured priors,” Proc. of Vision Image and Signal Processing,
pp. 33–41, 2002.
[12] A. Madan, M. Cebrian, S. Moturu, K. Farrahi, and A. Pentland,
“Sensing the ’health state’ of a community,” Pervasive Computing, pp.
36–45, 2012.
[13] L. Katz, “A new status index derived from sociometric analysis,”
Psychometrika, 1953.
[14] D. Chakrabarti, S. Papadimitriou, D. S. Modha, and C. Faloutsos, “Fully
automatic cross-associations,” in Proc. of KDD’04, 2004, pp. 79–88.
[15] E. M. Airoldi, D. M. Blei, S. E. Fienberg, and E. P. Xing, “Mixed
membership stochastic blockmodels,” J. Mach. Learn. Res., pp. 1981–
2014, 2008.
[16] C. Kemp, J. B. Tenenbaum, T. L. Griffiths, T. Yamada, and N. Ueda,
“Learning systems of concepts with an infinite relational model,” Proc.
of AAAI’06, pp. 381–388, 2006.
[17] P. Koutsourelakis and T. Eliassi-Rad, “Finding mixed-memberships in
social networks,” Proc. of AAAI Spring Symposium’08, 2008.
[18] K. Henderson, B. Gallagher, T. Eliassi-Rad, and H. Tong,
“Roix:structural role extraction & mining in large graphs,” Proc. of
KDD’12, 2012.
[19] R. Rossi and B. Gallagher, “Role-dynamics:fast mining of large dynamic networks,” Proc. of WWW’12, 2012.
The proposed LAP model significantly outperforms all
the baselines in most cases over different object roles. LAP
improves over the best baseline Prior by up to 22.69% and on
average 5.86%.
IV.
R ELATED W ORK
As reviewed in Section I, the proposed model significantly
differs from the existing studies in the area of dynamic network
mining. Besides these, there are also several other approaches
that are related to the task in this paper, thus we explicitly
discuss them in this section.
As for object role mining in graphs, existing methods [14]–
[17] usually adopt the assumption that two objects have the
same roles if they have the same relationships to all other
objects. This assumption heavily restricts the applicability of
these existing methods. For instance, suppose there are two
spammers in an email network, and they focus on spamming
users from different areas. In this case, the two spammers have
the same roles but totally different connections to other objects
in the email network, which contradicts with the aforementioned assumption used by the above existing methods.
Besides PageRank [2] and Centrality [13], several other
existing methods such as [18], [19] also use statistics in
topology to characterize object behaviors. The major drawback
of these methods is that the object roles are estimated through
subjectively selected topological features, thus these methods
may not be discriminative in learning many types of object
roles. Moreover, these methods do not incorporate the dynamic
information at each time point into the estimation of object
roles, which differs them from the proposed method.
V.
ACKNOWLEDGMENTS
C ONCLUSIONS
In this paper, we have introduced a probability model for
learning, analyzing and predicting object roles in dynamic
networks. The proposed model effectively integrates structural
information, dynamic information and supervised information
for extracting the latent feature representation of object roles.
The extracted object role representation is then used to identify
the role of each node, track the changes of object roles over
time, and predict the object roles in dynamic networks. We
have also provided the detailed variational bayesian inference
437
Download