>> Nikhil Devanur: It is my great pleasure to... of Microsoft's new faculty fellows. He has done great...

advertisement
>> Nikhil Devanur: It is my great pleasure to welcome Bobby Kleinberg. Bobby is one
of Microsoft's new faculty fellows. He has done great work on design analysis of
algorithms and applications to e-commerce, networking and many other things. And he's
going to talk about network formation.
>> Robert Kleinberg: Okay. Thanks Nikhil for inviting me to come give this talk. This
presentation will have sort of a Coals to Newcastle character to it in that it is some work
applying random graph theory and edge percolation to analyze network formation in the
presence of contagious risk. Much of the technical content of what I am going to present
while non-trivial for us might be sort of an exercise for some of the people in this room.
So I encourage you to think of this talk as an introduction to a provocative branch of
applied math about which very little is known and which, you know, there is perhaps no
more distinguished group of people all in one place to study these sorts of questions that I
have right here in this room.
>>: [inaudible] [laughter]
>> Robert Kleinberg: Hmm. As I was thinking about the types of questions that I am
going to get during this talk, that is not one that occurred to me. [laughter] A lot of us are
fascinated by the subject of complex networks and when researchers have looked at why
complex networks look the way they do, people take very different approaches toward
answering that question depending on whether they come from mathematics/physics or
from social sciences. So those who like to address complex networks from probabilistic
perspective will come up with models of network formation that say something like links
form according to the following simple rules. Nodes show up one at a time and link to
other nodes with probability proportional to varying degrees. Or with some probability
depending on your distance from other nodes in some post metric that is independent of
the network topology.
On the other hand, in the social sciences, you see models that don't have any explicit
probability built into them at all. So nodes form a network based on an endogenous
optimization process in which each of them constructs links strategically so as to
optimize some payoff that depends on the global structure of the network that they form.
An example of such a model just to introduce you to the aspect of a theory that may be
less familiar is this one from Jackson and Wolinsky. In this model there are two
parameters that define people's utilities for the network that gets formed. There is an
Alpha which is the cost of forming the link and a delta which is some decay parameter
that specifies how much of a preference you have for being close to nodes as opposed to
far away from them. And then your payoff is linear in the number of links you form so
you experience a negative payoff, a cost of Alpha every time you form a link. And then
for every other node in the network you have positive payoff which diminishes
geometrically with your distance from that node.
So for example, in a network where B and C are already linked together and A is
strategically deciding which subset of two possible links to form to them, it could either
do the empty set with a payoff of zero. One link with this payoff or two links with that
payoff and for different choices of Delta and Alpha you can get each of these three
behaviors to be utility optimizing for node A so that you can sort of draw a phase space,
you know, in the two-dimensional space of Alpha Delta settings and plot the graphs that
arise as optimum graphs at each point of the phase space.
There is an interesting issue involved in defining what it means for a graph to be
equilibrium of this game. I trust that you are all familiar with Nash equilibrium, but
generally Nash is not the right equilibrium concept for looking at network formation
games, because Nash considers what happens when individuals make unilateral
deviations, but the process of forming a link between two individuals should really be
thought of as a bilateral process where both of them must willingly opt into forming the
link. And accordingly game theorists analyze these games using stability conditions that
seem similar to Nash but are not exactly Nash. The stability condition that will be
relevant to this talk is the following one. We will assume that any node can unilaterally
secede from the entire network area and it can delete all of its links. And that any two
nodes can jointly decide to create a link between them if it benefits both of them and
benefits at least one strictly.
So the network is stable if neither of those two moves--if it is a local optimum under
those two types of moves. So there is no individual that can strictly improve its payoff by
seceding from the network and there is no pair that can each weekly improve their payoff
and at least one strictly by forming a link between them.
>>: [inaudible] individuals to cut off [inaudible]
>> Robert Kleinberg: We are not allowing individuals to cut off a subset of their nodes.
Good. So any network that is stable under that stronger definition is stable under our
definition. And so when we prove impossibility results for stable networks under our
definition, they automatically carry over to impossibilities for stable ones under the other
definition. But an existence theorem for stable networks under ours does not imply
existence for the other and so that is just something to bear in mind. Actually I claimed
that the more interesting result about stability of this talk would be the impossibility one
which does carry over but it is obviously a self-serving claim so you are free to form their
own opinion about which of the two is more interesting.
>>: But the reason for doing it that way is because of what you can do [inaudible]
because that is an interesting [inaudible]. I guess that is also a matter of [inaudible]
>> Robert Kleinberg: Speaking as a participant in the research process, the reason we
did this, we knew that a question of primary interest to us was going to be a price of
anarchy type question that was comparing the social optimum against the set of all Nash
equilibria. And so if we adopted the most permissive definition of stability and still got a
strong price of anarchy separation, that would imply the same for a less permissive
definition of stability. So given that in any such talk when I got to this slide there was
going to be controversy over what stability means, we wanted to make an initial choice of
stability definition that would resolve the controversy in a way that is most favorable to
saying the price of anarchy result is a strong one. That is what we did.
All right. Moving on. The Jackson-Wolinsky style payoff function that I just described
is--there are many other papers in the literature that propose different payoff functions
but nearly all of them have this mindset that there is a cost to link formation and the
benefit that you get indirectly from forming those costly links is that you make this
connected world and then we all benefit from being connected together in this network.
But there is a different type of network formation process that occurs quite often in
reality where links are beneficial and we pay a cost for all being connected together.
That will be the subject of this talk. So we originally got interested in this subject in
2zerozero9 because of financial contagion where these are banks making loans to each
other and at least in the popular consciousness about what happened in the fall of
2zerozero8, there is this sense that a tidal wave of failure propagated through the network
and the economy paid a large cost for being too densely connected in the presence of this
possibility of risk. So that although each individual business arrangement was beneficial
to the two parties involved, the overall outcome of constructing a densely interconnected
web got us trapped in this risky situation.
Okay, so here I have a picture of the loan network of the US federal banking structure.
This is taken from Beck and Atalay. The models that I am going to be talking about in
this paper are undirected graphs, whereas this one is it a directed graph whose edges
represent A makes a loan to B. At some level--the reason that failure propagates through
this network is because of something called counterparty risk. So these banks make
overnight loans to each other. It is literally the end of one day a bank that doesn't have
enough cash on hand to fulfill its minimum reporting requirement mandated by the Fed,
will ask to borrow a small amount from another bank so that in the morning when the Fed
has the right to inspect them and see if they have enough cash on hand, they will pass that
inspection. And then they will immediately give it back to the bank that loaned it to
them, unless they default. Which because these are giant banks there is almost no chance
that that will happen, but there is a small amount of risk any time money is given from
one party to another that it won't be given back.
So these risks of defaults were each individually viewed correctly as being
extraordinarily unlikely events. It is just that there were so many of them that when a lot
of them started happening at once unexpected results came out of that. Now that I have
told you about the general scheme of links benefit people, but being overly connected can
be to the detriment of the individuals, it is possible to find examples of that cropping up
all over the place. For example sexual contacts are usually treated by the participating
parties as being beneficial, but an excessive amount of linking carries attendance risks.
In some sense this one actually hues more closely to the story of my talk because this is
an undirected graph. You know the probability spread or failure is symmetric when a
link is formed between two individuals.
Another example would come from covert organizations, things like networks of
terrorists. The organization benefits from individuals being linked together because it
allows them to coordinate plans requiring more than one actor to execute the plan. And
the danger is that if one of these people is captured and coercively interrogated they may
divulge the identities of others and so the organization is at risk if it engages in an
excessive amount of linking and then one of the individuals is compromised.
There is another example that is not in the slides for some reason. Exchanging private
data could be another example of this. When I share my cell phone number with
somebody, that benefits me because that person, presumably it is somebody that I wanted
to share my phone number with, and it benefits me that they now have my personal
information and can call me easily. But if that person shares their address book with
someone else, who shares their address book with someone else and eventually one of the
people in this web turns out to be a spammer, then it was not beneficial to me that we
formed that highly interconnected network.
So now I am ready to present the basic framework that we use to qualitatively analyze all
of these situations although in each case it would be quite easy to find at least one
qualitative feature that differentiates this abstract model from the banking application or
the STD transmission application or which ever one you want to pick apart. So we have
nodes that are forming a network of max degree capital delta among each other. It is an
undirected graph. And there are two important parameters of the model. I think I should
write them down on the white board since they will only be on this slide for a minute but
they will be referenced throughout the talk. So Q is the probability that a node fails
spontaneously. And spontaneous failures are independent events, each with the
probability of Q.
In addition each edge makes an independent decision whether to be live or blocked. The
live edges are capable of transmitting failure and the blocked ones are not. And P is the
probability of an edge being live. So the way this determines the set of failed nodes in
the network is that first of all every node that fails spontaneously fails, and secondly
everybody that did not fail spontaneously also fails if they belong to a connected
component containing a failed node in the sub graph of live edges.
>>: [inaudible] time?
>> Robert Kleinberg: There is no time to mention. Now there is a notion that first the
network is formed and the failure is revealed, but those are the only two stages of the
game. So, yeah, if your application is STDs then you would want to have a time
dimension which is--that's the primary way that the STD application differs from the
abstract model, but that just is what it is.
Throughout the talk we will think of both spontaneous failures and transmission of long
edges as being rare events, so Q and P are both positive but they are less than 1,
significantly less than 1. So if a node has degree desivy and failure probability thesivy,
then its payoff is linear in the number of links it forms, so that is ADI; let me see if the
formula is on my next--good. So if you can see the slide then you can see the payoff
formula and if you can see this panel of the board then you will see the payoff formula
even when the slide is gone. You get a payoff of DI for every link you form, but only if
you don't fail. If you do fail you get a negative payoff of B, regardless of how many links
you formed.
>>: So the liveness otherwise on the edge only influences [inaudible] in failure?
>> Robert Kleinberg: Not the payoff, yeah, that is right. So you get benefit--if you don't
fail regardless of whether your edges were live or blocked you reap the benefits of it. We
are going to be comparing graphs now according to a parameter which is the minimum
welfare of any node in the network. This is another one of those points in the talk where
someone could interrupt me and say well why did you decide to--actually a philosophical
discussion I would love to have is why computer scientists are so bent on taking a graph
or an algorithm, summarizing it in a single number and then just saying mine is hundred
times better yours because on this one dimensional continuum mine came out 1zerozero
times higher than yours. But, okay, it is effective CS research that people seem bent on
taking a single number that measures, you know optimality and then ranking things
according to that number. And we choose min welfare. Now there are many other
natural choices we could have made. Average welfare would have been a very natural
one. We chose min welfare because at one point in this talk I will present an argument
that only works using min welfare and not the average welfare and I think I will
remember to tell you when that happens.
>>: Welfare in the top 1% [inaudible].
>> Robert Kleinberg: Yeah, that's right, that's right. So that I think would actually be a
pretty uninteresting parameter for analyzing this model. And as I hinted earlier, a
motivating question, not the only question that we will address but a motivating question
throughout the talk has to do with the price of anarchy. That is comparing this parameter
in a socially optimal graph, a graph that maximizes the min welfare versus the same
parameter in a stable graph under the stability definition that I showed you before. So
now it is useful to do some thought experiments to map out a region of parameter space
in which the question is interesting and distinguish it from regions in which the question
is completely uninteresting.
If A is less than BQP, what it means is that for any node A the risk of forming a link to B,
which is--okay, if B fails with probability Q and transmits that failure with probability P.
So anytime I form a link I get a negative BQP payoff from having formed the link. If that
is greater than the benefit of forming the link then we would just get a big independent
set and so A less than BQP is an uninteresting part of the parameter range. If there is
someone who I know for sure is going to fail, and A is greater than BP, then I would link
to them anyway despite knowing for certain that they are going to fail. So that is an
uninteresting part of the parameter range. Similarly if there is an edge that I know for
certain is going to transmit failure and A were greater than BQ, then I would still choose
to form this edge despite the fact that it is guaranteed to transmit failure. So based on
these thought experiments we want to be focusing on a range of parameters where A lies
between BQP and the min of BP and BQ.
>>: [inaudible] even the first one yet because [inaudible] attitudes or maybe because I
have some fellow [inaudible] so the cost of [inaudible] isn't as much of [inaudible]
because it doesn't increase the probability by that much.
>> Robert Kleinberg: Yes, I understand what you are saying.
>>: Something to do something that you can also disconnect [inaudible] isolated just
completely [inaudible]
>> Robert Kleinberg: It sort of--okay. So there is another--I forgot to write this down.
Another aspect of the model that I told you about is that people form a maximum of
capital Delta links. And we think of capital Delta as a constant that does not scale with
the network size. So I am going to get to your question but let me first talk about the
bounded max degree. You have unlimited max degree then when the number of
individuals in the world becomes astronomically large, you get into ridiculous solutions
where people are--that is not true. They lose their payoff if they fail. In an early version
of the model we didn't have the, you lose your payoff if you fail. Then you have the
pathology that people form tons of links because they know they are almost guaranteed to
fail but they are still getting a payoff of A for every link that they are forming.
I don't know how to answer your question, so definitely in a network of two nodes if A
were less BQP, they wouldn't form the link, but you are making a valid point that the
same argument does…
>>: I mean you've always got the option of disconnecting yourself [inaudible] BQP, that
is one thing that you would do. I mean given the [inaudible] you might just…
>> Robert Kleinberg: Well, that's true but if I have 1zerozero links and the first one
increases my failure probability by BQP but the second one by a little less than that and
so it is at least possible that for a very large degrees…
>>: But this assumes that Delta is bounded [inaudible] probably will [inaudible]
>> Robert Kleinberg: Yes. But the justification for Delta bounded was also not
completely airtight, okay. The paper makes a lot of assumptions to get into an interesting
range of parameters where I think an interesting phenomenon is observed. And I have
argued to forcefully and thereby gotten into a bit of trouble claiming that every other part
of the parameter’s face is uninteresting. I haven't even gotten to this last line which is an
important one actually and is entirely unjustified. So rather than merely assuming A is
between these two trivial upper bounds and lower bounds, I am going to assume that it
has a lot of room on either side. So it differs from the min BP and BQ by a factor Delta
which is small and exceeds the other one by Delta inverse. So the important one here
actually is the A less than Delta min of BP and BQ and it is important enough that I am
going to include it on my cheat sheet here and I think the cheat sheet is now complete,
which is good because I think if I wrote down here nobody could actually see it.
So now it is possible for me to tell you about our main results. It turns out that payoff A
over P which is the payoff you would get from forming 1 over P links if you are
guaranteed not to fail. It is a range in payoff space to where a critical transition takes
place. And we will see that as Delta tends to zero the optimal, the min welfare of a
socially optimal graph is A over P times a constant which is strictly greater than one by
an amount that is tending to zero with Delta. All of this suppresses the dependence on N,
the network size, because it turns out that, you know, the limit of the optimal min welfare
as N goes to infinity a constant not depending on N.
In qualitative terms what this is saying is that socially optimal graphs actually push up to
and slightly beyond a phase transition point in payoff space and one thing that we will see
is that in order to get payoffs beyond the space transition, you need to be able to know the
identities of the people you are linking to so in an anonymous model where each
individual just chooses their degree and then a random graph with that degree sequences
generated, it is not possible for them to achieve payoffs exceeding A over P by any
constant factor. We will also see that in stable graphs the largest min welfare achievable
is little o of A over P as Delta goes to zero. So it is bounded by G of Delta times A over
P where G of Delta tends to zero with Delta.
And in qualitative terms what is going on is compared to socially optimal graphs. Stable
graphs have higher average degree, which exposes their nodes to a greater amount of risk,
so much so that it wipes away all but a vanishingly small fraction of the benefit achieved
by doing that linking. And this is a form of tragedy of the commons where because of
the individuals do not feel most of the negative impacts that they generate on others, they
continue engaging in behaviors that are detrimental to society as a whole, while making
small increases in their own payoff until the general payoff shared by everybody is close
to zero. So I would like to talk now about how some of these results have proved.
So recall that forming a graph nodes spontaneously fail with probability Q and edges are
live with probability P. So this is edge percolation in a graph which is in endogenous to
the model. So in particular when people happen to form a complete graph, then the live
edge sub graph is a GNP. And as is well known in GNP where P is, you know, less than
1 over N by a constant factor all of the components are small so the probability of our
component size exceeding S is exponentially small NS and on the other hand when P is 1
plus X over N there is a giant component whose size is a constant fraction of N. The
constant depending on this X in enumerator and if you plot the dependence, you know, at
zero All the way up until the numerator is equal to 1 and then it immediately takes off
with a positive derivative, this continuity in the derivative at 1.So that will be important
in the talk also. The derivative here is positive but not infinite.
In analyzing the, you know, min welfare of arbitrary networks, we need to consider not
just GNP but edge percolation in general graphs of Max degree capital Delta. And I want
to start out by showing you that the min welfare cannot exceed A over P by very much. So
to do so let me remind you of a little bit of the existence proof of the giant component. So
if you just want to prove that there is a constant probability of forming a giant component
in GNP when P is 1+ X over N, you imagine performing breadth-first search starting from
any root node and in any step of this breadth-first search, you expose all of its adjacent
edges. Some number of additional nodes are found and placed into the back of the queue
and so the expected change in the queue size is 1+ X for the new nodes that show up -1 for
the node that was just inspected. So there is a positive drift in this random walk and for our
purposes the only thing that is necessary for our analysis is that there is a constant
probability that this random walk process continues beyond 1 over P steps. This is just a
picture of the edge exposure process of the random walk. So then if we have a graph with
minimum degree D and we do edge percolation where the min expected degree of vertices
is one plus X for some constant X greater than zero, what this argument establishes is that
for any node I there is a constant probability that it's live edge component contains at least
Absalom times D nodes.
If you think about the consequences in terms of the payoffs to nodes, that means that if
you want everybody to get payoffs 1+ X times A over P, you need to have minimum
degree at least 1 plus X over P, is there a…
>>: [inaudible] P times D-1 [inaudible]
>> Robert Kleinberg: We are thinking…
>>: We are thinking of 1 and then adding D-1 [inaudible]
>> Robert Kleinberg: Yeah, we are in the large D regime so…
>>: [inaudible] cheat sheet…
>> Robert Kleinberg: I should make a note for the next version of these slides. Yeah,
because that is right, you only expose D-1, the edge you came in on does not get reexposed. It will be fixed, in the next version of these slides. If everybody is getting a
payoff of 1+ X times A over P, they get a payoff of A per link so they must be forming 1
plus X over P links so they are getting expected degree 1 plus X. So that means that
everybody has a constant probability of being in a live edge component with this many
nodes. When that happens, their probability of failing is, let's see I am out of space to write
on my cheat sheet. I can just do this in the air, oh, there is, okay. The probability of
surviving is 1-Q to the power, Absalom D. And, okay we are taking D to be roughly A
over P and by A--okay, D is roughly A over P which is like Delta B, so my--I don't know
how to do this in this amount of space.
People have such a high failure probability that it washes away all of the benefit of
linking and they get a negative payoff. They get a positive payoff of A over P if they
survive. If they fail, they get a negative payoff of B which is like A over Delta P. so the
only way that this is beneficial for them is that their survival probability is like 1-Delta.
But it is much worse than that. So that shows that if X is fixed and I take Delta P and Q to
be small enough with respect to that fixed X then people cannot achieve a payoff of 1 plus
XA over P. On the other hand if we treat Delta as fixed and we allow X to be sufficiently
small, then I will show you that you can get payoffs that are strictly greater than A over P.
It is useful to think of two natural ways of achieving this. One is a bunch of disjoint
cliques, and in that case to get up to this level of payoff we are going to need to have the
degree be slightly more than 1 over P so the clique size has to be a little more than 1 over
P. And then what is going to happen is we are going to form a bunch of disjoint GNPs, for
this value of N. And the other way would be to make a gigantic high-growth graph which
basically looks like a, locally looks like a D regular tree in the realm of a very slightly
supercritical branching process.
Okay. So in the paper we do general P and Q sufficiently small in this next set of slides
focus on case P equals Q. And let's think about which of these two types of networks is
better. Starting from degree 1 over P should you increase your degree to 1 plus X over P
for a very small X? If you do you will gain AX over P performing those links. And you
lose B times the increase in your failure probability. And we know that your failure
probability depends in the GNP case on the giant component size and in the branching
process case on the size of the component of the, you know, branching process containing
the root.
So in the--here I have drawn the anonymous case, so that the random D regular graph on
N nodes, D is slightly more than A over P, so it has high girth and locally in the
neighborhood of the root the live edge sub graph is going to look like a super critical
branching process. So the size of the giant live edge component is going to be Tao of X
times N where Tao is another one of these functions that zero up until X hits 1 and then-sorry the way that I have parameterized it, is zero until X hits zero and then it takes off with
positive derivative.
So as you increase your degree from A over P to 1 plus X times A over P, what is
happening is your failure probability is increasing at a rate which is the derivative of Tao at
zero. So you are getting a benefit of AX over P from the links and your failure probability
is increasing by X Tao prime of zero, so this benefits you if A over P is greater than B Tao
prime of zero. But I am assuming that A over P is much less than B; A over P is less than
Delta B and Delta is sufficiently small parameter. So in an anonymous market it does not
benefit you to increase your degree at all beyond the thresholds of forming, of criticality.
Okay, that is what this hand wavy argument meant to convince you of. So now let's look at
the opposite case where you are forming cliques each of which is going to sample a live
edge component which is a GNP of expected degree slightly larger than 1. And let's do the
same thought experiment. You start with clique size 1 over P and you consider is it
beneficial to grow up to 1+ X times P.
So now what happens is the clique that node I belongs to is going to have a giant
component of size theta of X times 1 over P and in order for I to fail, the most likely cause
of failure is that it belongs to the giant component and somebody in the giant component
spontaneously fails. There are these other events that have truly tiny probability where I
belongs to one of the other components and somebody else in that tiny component fails.
These contribute negligibly. So, you know, in each of these three cases I does not fail in
this picture because it is in the giant component but the only failures happen outside. I
does not even belong to the giant component. The only case that is bad for I is that both it
ends the failed node. I am making the assumption by the way that there is only one failed
node here. It is a reasonable assumption because spontaneous failures happen with
probability P so the expected number of failures is 1 plus X, so the actual number of
failures is Poisson with expectation nearly 1, so maybe there is not one but a couple of
nodes that fail. A constant number of nodes are failing. I is in trouble if it belongs to the
giant component and at least one of the failures is also in there.
And so as you increase from 1 over P to 1+X over P, the benefit to I is again a gain in
payoff of AX over P and the detriment to I is a negative B times the derivative of theta
squared. So once again we are comparing AX over P versus 2BX theta of X theta prime of
zero. And growing beyond the face transition is beneficial if A over P is greater than 2B
theta of X theta prime of zero. And what happens now since we are thinking now of BEQ
Delta as being fixed and X is allowed to be as small as we want, we can always set X small
enough that A over P is greater than 2B theta of X theta prime of zero. And it means that if
we push past the phase transition by a sufficiently small amounts people's payoffs are still
growing.
So again the qualitative reason for this is that in that branching process if you are in a
giant component, it is so large that you are virtually guaranteed to fail. And if you are in a
bunch of small cliques then you need to events to happen, you need to be in giant
component and the failed guide needs to be in the giant component. And so it is really the
difference between theta itself which has positive derivative at zero and theta squared
whose derivative is zero. That is what determines the qualitatively different behavior
between growing cliques past the phase transition versus growing the degree of the
branching process past the phase transition.
So this slide is just summarizing what I have talked about so you can achieve
supercritical payoffs by forming disjoint cliques and if you are doing anonymous linking,
that is impossible, because giant components in this case are much larger than giant
components in this case and so as soon as you get past the phase transition the penalty of
going beyond that transition is too steep.
For the final part of this talk I am going to talk a little about stable graphs. Remember I
advertised this price of anarchy results that says there is a tragedy of commons where
additional linking washes out almost all of the positive benefits that people get from
forming these links. So in particular we just saw that a union of cliques of size just barely
beyond A over P gets a payoff which is actually reasonably high. So why is that graph not
stable already in and of itself? I mean after all…
>>: [inaudible]
>> Robert Kleinberg: I'm sorry, thanks. Not A over P, 1 plus X over P. So these cliques
are already getting pretty close to, you know, they are past the phase transition so if they
grow a little more the giant component is getting bigger reasonably fast so you might think
that this graph is already stable because there is not much room to grow the cliques, but if
these two guys form a single link between them, then they get a benefit of A. The cost of
forming that link is really just the probability that the node I linked to belongs to the giant
component in its graph and someone in that giant component failed. And that is small
enough that if these are optimally sized cliques these two nodes are going to benefit by
forming that link. So what is going on here is forming that link does not expose INJ to
very much failure, but because each of them is so highly connected to its own community,
that probability of failure multiplies over everybody that is linked over INJ so the total
amount of negative utility that they create for the world by forming that link is much
greater than the positive utility that they get. This is the form of the tragedy of the
commons where--Okay so the original tragedy of the commons story is there is a pasture,
everybody puts their cows out to graze on the pasture. The farmers have an infinite supply
of cows and they are just going to keep putting cows on the pasture as long as the farmer
gets additional benefit from that. And the model of the pasture is that it produces a net
benefit which is a function of number of cows grazing, a decreasing function of the number
of cows grazing it that eventually decreases to zero. And that net benefit is shared equally
among the cows.
Then from the standpoint of any individual farmer, if the benefit has not yet reached zero
they are going to put one additional cow on, because that cow will still get a positive
benefit from grazing and the farmers total share of the benefit is going up because the
farmer is increasing his number of cows. So you end up with this phenomenon that they
keep populating the pasture until the total benefit of everybody is zero. The same thing is
going to happen in our model so people will just keep linking and linking until they have
eaten up essentially all of the utility that they got from the linking. And that is the only
time that they will stop. So how do you actually analyze it? Let's consider two nodes I
and J in a graph where they have decided not to link to each other. If they decide not to
link, that means that each of them--there are two reasons for them to decide not to link.
One is that either I or J is already at its degree threshold, Capital Delta and the rules of
the game don't allow for it to form another link.
The other possibility is that one of them has a failure probability that is high enough that
the link does not benefit the other party. So the benefit to I if it forms the link is A and
the cost of forming the link is at most B times P times the failure probability of J. It
might be strictly less than this because maybe there is already another path from J to I in
the graph and so maybe I is already absorbing some fraction of the probability of J
failing. All we are going to need is that the cost is less than or equal to BPCJ. So if I
doesn't want to link that means that the benefit is less than or equal to the cost and so the
failure probability of J is already at least A over BP.
So now let's think about three types of nodes. Those are already at their max degree and
that is why they are not linking, those who have failure probability greater than A over
BP and that is why nobody is linking to them anymore. And the third type of nodes have
low degree and low risk. So if we are in a stable graph then every node must be one of
these three types. And furthermore the nodes of type III must form a clique because I
have just argued that if I and J both have degree below Capital Delta and they both have
risk less than A over BP then they are going to link to each other. So the nodes of type I
already have high degree. The nodes of type II…
>>: [inaudible] trying to form a clique and Delta blocks them, no?
>> Robert Kleinberg: Nodes of type III.
>>: Yes.
>> Robert Kleinberg: Because they are type III that means that they don't have type I so
their degree is not Capital Delta.
>>: [inaudible] many of them. They try to form a clique and [inaudible].
>> Robert Kleinberg: Then I would not call them type III. So consider a graph which is
already stable. Its nodes have types I, II and III. The nodes of type III by definition don't
have degree Delta and by the arguments on the preceding slide they are already linked
together in a clique, because by our assumption that the graph is stable there could not be
any missing edges between them. That also means there are not very many of them
because in a clique of Max degree Capital Delta, you know that you have at most Capital
Delta +1 nodes.
Suppose now for contradiction that everybody achieves a payoff greater than or equal to
epsilon A over P. It means that the type II nodes have a payoff which is at least epsilon
A over P and on the other hand type II means that your [inaudible] J is reasonably large A
over BP. These nodes are getting a payoff of ADJ minus A over P, which is at least
epsilon A over P. It implies that DJ is greater than or equal to 1 plus epsilon over P.
Good. So types I and II both have degrees greater than or equal to 1 plus epsilon over P.
I am silently here making the assumption that the Capital Delta is greater than 1+ epsilon
over P. But again that is the interesting part of the parameter space because we just saw
the sort of optimal cliques have size 1+ X over P, so if you're Max degree bound is
significantly less than 1 over P then the world is not even able to get up to the size of
GNPs where giant components start to emerge.
So if there were no nodes of type III we would be in a graph of min degree 1+ epsilon
over P. We would be doing edge percolation with probability P and already saw that that
process produces for every node a constant probability of belonging to, you know, a
component of size at least constant over P and that that eats up all of the positive benefit
from linking and gives them a negative payoff. So if nodes of type III did not exist, we
would already be done at this point. We would have reached a contradiction because we
assumed that everybody's payoff is at least epsilon A over P, and deduced that their
payoffs were negative. Since we have the type III nodes we have to do a little finesse to
finish the argument. So assume now that the number of nodes in the world is very, very
big, bigger than Delta to the power of Delta. We have this clique of type III nodes which
as I said has at most Delta +1 members in it. And that means within a hop distance of
Capital Delta you can only encounter about Delta to the power of Delta nodes. If N is
bigger than that, then there must be somebody who is far from this clique and within
Delta hops of this node, you see nothing but high degree nodes. So now run the same
breadth-first search argument starting with this node and you know that in the first Delta
steps of that breadth-first search you will never expose one of these type III nodes, so it is
as if they did not exist. And the first Capital Delta steps are already enough to get the
constant probability of being in a component of constant over P many nodes. So when N
is that large, we take one of these for nodes and deduce that it has a negative payoff
contradicting our assumption then that everybody's payoff is at least epsilon A over P.
So that is the story of what happens in stable graphs. This proves that, you know, in this
parameter range that I delineated people will do so much linking that they eat up almost
all of the utility gained from forming links.
So just a final slide of reflections and open questions before I end the talk, I guess the two
main qualitative conclusions of this research were first of all when we first started
thinking about this contagion model a natural guess was optimal graphs would consist of
cliques that go right up to but not beyond the point where the giant component emerges.
It is a natural guess because as soon as the giant component emerges, then there is a sharp
transition in the behavior of failure propagation and it is natural to think well, that is
eating up so much utility that you would push right up to the phase transition but not past
it. We have seen that that is actually not true. You push past the phase transition by a
small amount. You are able to get just barely supercritical payoffs but not, you know, a
factor that tends to zero with this Delta perimeter.
And the other qualitative insight that we saw is that when you form Nash stable graphs as
opposed to socially optimal ones they are much less benefit for their nodes as measured
by their min welfare function. So one of the open questions we would like to answer is is
it also the case that the average welfare in stable graphs is bad? And that translates pretty
crisply into a question about the expected sum of squares of the sizes of percolation
clusters in a graph of average degree 1 plus epsilon over P when you are doing edge
percolation with probability P. And it is possible actually that the answer to this question
is well known and we just don't know it.
>>: [inaudible]
>> Robert Kleinberg: And this is known for irregular graphs, and not just for regular
graphs?
>>: [inaudible]
>> Robert Kleinberg: Yes. The question is if I have a graph whose average degree is 1+
epsilon over P but it is not a D regular graph. It could have highly irregular degree
distribution. Is it still the case that the expected sum of percolation cluster sizes squared
diverges the way that it does [inaudible].
>>: [inaudible] you are saying that some of the degree is high?
>> Robert Kleinberg: Yes. We are assuming the average degree is high. So we don't
want to assume that the min degree is high.
>>: [inaudible]
>> Robert Kleinberg: So in some of our [inaudible] we assumed that the min degree is
large but then, you know, when I had to analyze stable graphs, the way I honestly
achieved that was I showed that stability implies that all of the low degree nodes are in a
clique over here and then I went over here. So I had to work for that min degree. And
the way that I was able to obtain it was--so this is why we are stuck on min welfare
instead of average welfare, right? The way I obtained this picture was by assuming that
everybody's payoff is at least epsilon A over P. So now you see why this question about
average welfare becomes a question about graphs of average degree D instead of min
degree D. It is possible that there is an easy answer to that question. We thought about it
a bit and did not see how to do it ourselves, but I would love it if one of you would
enlighten us.
We also don't know what the structure of socially optimal graphs is. So I showed you
that a union of cliques can get beyond A over P, and that nothing can get beyond A over
P by a factor much bigger than 1. But, you know, we have no strong reason to believe
that a union of cliques is literally the optimal graph and it might be interesting to prove or
disprove that conjecture, similar structural question about stable graphs. So in our paper
we show that there exists a clique size such that the union of cliques that size is stable.
But there might be sort of many other stable graphs. And an interesting style of question
is whether, you know, all nearly optimal graphs or all stable graphs must look like a
union of cliques with just a small number of edges added and deleted.
And finally for those in the room who are interested in modeling type questions as
opposed to this is a math problem that we couldn't solve that comes directly out of our
paper. So, you know, can you extend the model to, you know, change the payoff
structure of the game in some way that makes the users internalize the cost of linking so
that the outcome of playing the game to a stable outcome looks more like the social
optimum and less like the stable graphs where the tragedy of commons has set in and we
have eaten up all of the payoff. Okay, so thank you for your time and I am happy to stay
and take questions. [applause]
Download