>> Yuval Peres: Okay. Good afternoon everyone. ... Berry tell us about probabilistic aspects of minimum spanning trees.

advertisement
>> Yuval Peres: Okay. Good afternoon everyone. We are delighted to have Louigi AddarioBerry tell us about probabilistic aspects of minimum spanning trees.
>> Louigi Addario-Berry: Thanks very much Yuval. Thanks for the invitation. I'm starting with
this picture because it's a nice picture that's somewhat related to the subject of my talk. And so
since there was some interest, maybe I'll tell you quickly just what this is. So here we have a
hundred thousand uniformly random points in the square. We formed there a minimum
spanning tree and…
>>: [indiscernible]
>> Louigi Addario-Berry: Say again. The minimum spanning tree according to the Euclidean
distance. And we've picked some random root vertex which is somewhere around here. If I
have a laser I can do that from further away, and then colored all the vertices according to their
graft distance in the minimum spanning tree from that root vertex, so the closest vertices are
red. And then orange, yellow, green, blue, purple, violet codes, graph distance from the tree
from that root and you see all these interesting curves start to appear. So this is a…
>>: [indiscernible]
>> Louigi Addario-Berry: Uniformly at random.
>>: Uniformly at random.
>> Louigi Addario-Berry: Yeah. So this is a picture that no one can say anything about to date.
If we had used some version of this on the triangular lattice, then the people to ask would be
[indiscernible] who have a forthcoming paper on the subject.
>>: The black lines…
>> Louigi Addario-Berry: And the black lines, so the problem is there’s so many points here that
you can't see the, most of the edges of this tree, only the colors. So what I've done is drawn the
important edges and what important means is that, so an edge is drawn in bold here if on
either side of that edge lies at least 3% of the tree. So if the edge cuts at least 3000 vertices off
from the rest of the graph. So it gives you some idea of the, broadly of the structure of the
tree. Okay. But I won't be talking about two dimensions today. This is just sort of a
motivational picture. I'll be talking about minimum spanning trees in high dimensions. So just a
very brief reminder about what they are and how we build them, so if you trace back a bit you
find the earliest reference that I've been able to find is Otakar Boruvka and he asked the
question in the following sort of general form. So you are given the complete graph, so
endpoints with distinct edge weights, which we think of as distances between the points and
you're interested in forming a network so buying a network which is the unique connective
graph that minimizes the total length. Okay? So you want to connect up all of the cities and
you don't want, you want to pave a minimum number of miles of road. Okay. So we've chosen
all of the edge weights to be succinct so that there's a unique minimizer. That minimizer is
necessarily a tree. You're never going to buy a cycle if all you care about is a minimal
connectivity and this is the minimum spanning tree. So one of the standards, there's many easy
algorithms for building these things. One of the things I like about this problem is that it's
algorithmically very simple but probabilistically quite challenging. So here's one of the
methods, sort of simple greedy method often called Kruskal’s algorithm. It's really essentially
from this earlier paper. It just says order the edges of your network by increasing order of
weight. Then consider the edges one at a time in that order and add each edge unless doing so
would create a cycle. So we're always going to add the smallest weight edge in this example.
Okay, second smallest is 3. We add it. We don't add 6 because it would create a cycle. Add 7
and now we are already connected. We know we're not going to add any more. Don't add 8 or
9. Okay? So this is quite standard. I'm going over it quickly but I wanted to emphasize one
thing about this process, okay, which is the following. Anytime we decide not to add an edge,
that's because it would create a cycle, which means that its endpoints are already in the same
component. Okay? So that means that at every step if we're only interested in the connectivity
that has been formed by the edges that we've added, okay, so which, at a given stage, which
pairs of vertices are now connected up by some path, it doesn't matter whether we add all of
the edges or only the edges that this procedure adds, okay? So let's just convince ourselves of
that quite quickly. Here we add an edge, add an edge. The third edge we don't add, right, but
if we did and it we wouldn't gain any new connectivity. There was already a path between
those vertices. Okay. And so on. Here we add an edge and now any edge that we go, the last
edge we don't add, but that doesn't change the connectivity. Okay? So this gives you a
coupling between two processes. One where you only add edges that don't create cycles and
one where you add all the edges and that coupling is going to be basic for this talk. So I hope
it's clear to all. Okay. So what's…
>>: [indiscernible]
>> Louigi Addario-Berry: Say that again.
>>: That is not Boruvka’s That’s Kruskal’s.
>> Louigi Addario-Berry: This really is Kruskal’s, but I mean, in essence, in essence it's Boruvka’s
algorithm, okay? Kruskal, Kruskal really says one at a time, but Boruvka more or less says do
some greedy procedure with possibly larger clumps than single vertices, okay. So it's… I would
give the credit for this to Boruvka.
>>: [indiscernible] Boruvka does is every vertex is a minimal cost edge. And then you certainly
select all of those in your [indiscernible], it would be the same exponent as [indiscernible].
>> Louigi Addario-Berry: Right. So then, well, and it's not individual vertices, right? It's
components. It's every component…
>>: [indiscernible]
>> Louigi Addario-Berry: Yeah, yeah. But I mean it's essentially, it's essentially this sort of
global greedy procedure. It's slightly different from Kruskal; you're right but it's… Okay. So you
can think of various probabilistic models that you might impose if you wanted to understand
this very typical structure of these. I showed you one at the start. Given the sort of, the
description of the problem I just gave you, a natural probabilistic model it would just be to say
instead pick iid weight, iid weights or lengths, okay, and so that gives you sort of two different
probabilistic structures, one broadly Euclidean, one broadly iid or mean field, okay? And then
there's course, so that's sort of one axis of division that you can look at for research into
minimum spanning trees from a probabilistic perspective. Another is whether you study things
that actually have to do with the weights themselves or you study the sort of, metric or graph
structure of the object that you build. Okay. And there's a lot of research in both of those
camps. So on the weight functionals, so maybe the most famous classic results in this subject is
that the total weight of the minimum spanning tree if you use independent uniform weights on
the complete graph converges as 8 out of 3, so this is a famous result due to Enfreeze
[phonetic], okay but there's a lot more research in a similar direction. On the other hand, if you
want to study the global structure then there's also a huge body of work and there's all sorts of
questions you can ask. One natural probabilistic framework for asking those questions is to try
to study some sort of distributional convergence of your finite objects to some limit object
which might be an infinite volume limit ala global weak convergence or it might be some
compact limit, something to do with curves in the plane if you're in a Euclidean setting or…
There's a variety of possibilities and a lot of work on the subject including by people in this
room. Okay. So I've mostly been focused on trying to understand the iid picture and the metric
structure, okay? So there's a variety of ways you can do that. You can ask for local structure,
sort of global structure which you might get at by rescaling your tree to sort of maintain a
compact object that then gets large or you might try some other things. So now the question
you can ask is how, you know, if we manage to understand one object then, to what degree is it
a template for other objects? So if we can handle one sort of, some sorts of weight can we
learn to handle other weights or can we handle minimum spanning trees of other graphs and so
on. Okay. So that's, I'll call that the question of universality. Today I'm really going to focus on
talking about global structure. Okay. Global structure for iid edge weights, okay? And let me
just quickly make one observation that comes from the algorithm I just described. Okay. If I
take the complete graph on end vertices and I want to find, to understand the metric structure
of the minimum spanning tree, then as long as I use iid weights with some continuous
distribution, it doesn't matter what the weights are. Why is that? Well the algorithm I just
described starts by ordering the edges in increasing order of weight and as long as you have iid
continuous edge weights, then that's just starting from a uniformly random permutation of the
edges. Okay? And once you have that permutation, everything else about the procedure is
determined. So the distributional properties of the tree, if you don't care about the weights
themselves, only rely on this fact, so you're really free to choose the weights you want. We're
going to end up choosing uniform 0 1 weights to get a nice coupling, okay? So here's a,
description of the global structure of this tree. So if you take the minimum spanning tree which
I'll call Tn, so this is a random tree. And rescale it so that each edge has length n to the -1/3 in
mass 1 over n, so the total mass of the object is 1, and you call this rescaled metric space,
measured metric space Mn, then for some suitable notion of convergence and distribution Mn
converges in distribution to some random limiting metric space. Okay? So there's, to fully
explain this theorem I need to tell you sort of a few things, what this notion of convergence and
distribution is. There's sort of some work to make sense of what exactly this is telling you. For
today, I'd like to just focus on one thing about this theorem, which is this rescaling here, the
fact that the edges should be rescaled by n to the -1/3 so that the limiting object is compact,
okay? And so that statement is really equivalent to this statement. That if you look at the
diameter of the minimum spanning tree of the complete graph with these iid continuous edge
weights, then that diameter is of order n to the 1/3. So if you want to get a compact nondegenerate limit, you should rescale by n to the 1/3. So this is, this is some older work. I
should mention who I'm working with on this slide, Nicoli Broutin, Christina Goldschmidt and
Gregori Miermont and this theorem that I'm going to focus on is joint work with Nicoli Broutin
and Bruce Reed. Okay. So any questions about the result? Okay.
>>: 2006 and 2009.
>> Louigi Addario-Berry: So there was a conference paper where we described the results and
the journal version appeared in 2009. We proved it in 2006 and we proved it again in 2009.
[laughter]
>>: [indiscernible] expectation of the other [indiscernible]
>> Louigi Addario-Berry: Ah. That was another possibility.
>>: [indiscernible] [laughter]
>> Louigi Addario-Berry: Independently and simultaneously. Okay. So here's our setup. We're
going to have the complete graph, I said, now I'm making a choice from edge weights, uniform,
iid uniform edge weights. Okay. And remember the point I made a minute ago which is in the
process where you add the edges one it time, if you are only concerned about the connectivity
structure, the set of components, the set of vertices of each component, it doesn't matter
whether you add all of the edges or just a sub, or just a, the tree edges, okay, and that provides
us probabilistically now with a coupling with the classical error training random graph tree n, p.
So I'll remind you that the error training graph process uses these uniform 0 1 edge weights and
then for a parameter p between 0 and 1, you only keep the edges that weigh at most p. Okay?
And then every edge in the resulting graph, every edge is independently present and is
probability p. And we can do the same thing for the tree, just throw away all of the edges of
the tree with weight bigger than p. That gives us some forest, in general, right, because we've
taken a tree and thrown away some edges. The point is that the connective components of the
forest are exactly the same as the connective components of the tree. Okay. So if we want to
understand the global structure of this tree, then the time at which that global structure -- if we
view the p as a time parameter, then the time at which that global structure forms will be the
same time at which the sort of, at which Gnp becomes hooked up into one connective graph.
So in particular, it's classical that the connectivity threshold for Gnp is at log n over n. p is log n
over n, so we certainly don't need to go past log n over n to understand the final minimum
spanning through Tn. But in fact, the bulk of the structure was formed much earlier, around p
as 1 over n. Okay. And I'll make that precise. So I'm reading everything of interest that occurs
when p is around 1 over n, so here's a comic sketch of what I mean. If you look at edges just a
bit less than 1 over n at .99 over n then all the connective components of Gnp have logarithmic
size, okay? And so that will also be true for the tree. So if we're trying to build something of
size order n to the third, which you can at least believe -- I'm going to convince you of that, but
if that is indeed the case, then nothing really interesting has formed. Nothing really substantial
has formed on that scale at this time. On the other hand, if we go just a bit past 1 over n then
there's already unique component of linear size and all the other components have logarithmic
size. Okay, so you can somehow imagine that if all the other components have at most
logarithmic size, then plausibly they are not going to affect the diameter of the final tree very
much even after they hook on to, to that largest component. We know at this time then there's
going to be a component of this forest of linear size and all these logarithmic little pieces
plausibly won't change things very much. I'll try to convince you of that, but, this comic picture
should at least motivate why we want to study what's happening around p is equal to 1 over n.
This is really the time at which the structure of the MST is going to be built. I'm about to focus
in on what's happening around p equals 1 over n and describe to you how we're going to
understand the structure of Tnp at that time. Before I do that, I just want to make sure that
this picture is clear. So this is already an important thing to understand, right? When we're just
below 1 over n we have components of logarithmic size. Just above we have one component of
linear size. All the others are logarithmic. Let's look at what happens at that at that breakpoint
at 1 over n. Okay. So it turns out that at this point the components of Gn1 over n are pretty
treelike. This is good news for us if we're trying to understand the minimum spanning tree by
some sort of comparison with this graph process. In fact, the graph process is built, is building
of things that are already pretty close to being trees. If you look at the largest number of
edges, the largest number of edges more than a tree that any of these components has, hey,
that's some random quantity, but it stays bounded in probability as n gets larger. In other
words, it forms a tight sequence of random variables. That's, in fact, I mean there's a, the
expectation stays bounded of the largest number of cycles. Okay. So that's good, and most of
them in fact are already trees. So a component of Gnp that happens to be a tree is also exactly
a component of Tnp, right? It's the same set of vertices. It must be the same set of edges if it's
just tree. Okay. So how big are these things and what's their diameter? Well it turns out that
they have size around n to the 2/3. This was already asserted in the classical 1960 paper of
[indiscernible] on the subject, maybe 1959. But the fact that at 1 over n the size should be
around n to the 2/3 is certainly classical result. And more importantly for this talk, the diameter
of these components, of these largest components is around n to the 1/3. Okay?
>>: What is [indiscernible]
>> Louigi Addario-Berry: So it means that if you give me an epsilon, I can give you k such that
the probability is at least n to the 2/3 over k and at most kn to the 2/3 is at least 1 minus
epsilon. So you can bound it in an interval, a multiplicative interval around n to the 2/3 with
probability as close to 1 as you like.
>>: That means n probability. It doesn't mean [indiscernible]
>> Louigi Addario-Berry: No. [indiscernible] yeah. And I'm going to, I mean basically everything
I say in the talk will hold in the sense of improbability and also in expectation and I'm going to
sort of gloss that because everything that, in fact, happens with very high probability, so for the
most part I won't be too careful about that, but feel free to ask if something’s unclear. Okay.
So I'll sort of tried to justify this diameter into the 1/3 to you, but let me remark before I do so
that this already yields a lower bound on the diameter of the resulting tree. Okay? Because if I
take a component of the graph and it has diameter n to the 1/3, well, the tree has -- we know
that the forest has that same component, right, so the tree that spans, the minimum spanning
tree of that component, well what does it do? It takes all of the edges of the graph and then it
removes some of them and that can only increase the diameter, so that component of the tree
must also have diameter at least n to the 1/3, and on the other hand, that bit of tree is some
subtree of the final minimum spanning tree, so the minimum spanning tree must have diameter
at least n to the third. Okay. So this immediately yields the lower bound. Okay. I said that
already. Let me now explain where n to the 1/3 comes from. So let's suppose right here, here's
the vertices 1 up to n, right? And let's suppose you give me some particular set of vertices s
and you tell me that s is a component of Gn 1 over n, okay? Well we know that it's pretty
treelike. I said it has sort of at most they few edges more than a tree and it has a decent
chance of actually being a tree, okay? So suppose, so supposing s is a component of Gn 1 over
n and let's suppose that it actually is a tree. Then the symmetry of the model tells you that any
spanning tree of that set of vertices is equally likely to appear under that conditioning. So we
know that it's a tree. Which tree is it? Well for any particular tree to occur that gives you m-1
edges that have to occur and the rest of the edges have to not occur, plus no edges to the
complement. That means for any tree you are conditioning on the same number of, you're
asking for the same number of edges and non-edges and so any tree is equally likely. Okay. So
if I tell you that the components of the tree, it's a uniformly random spanning tree of its set of
vertices and uniform spanning trees, if I give you a uniformly random spanning tree of m
vertices that's the same as a uniformly random tree on m vertices, and it has m to root m,
right? That's classical. Okay. So then the diameter, that's probably too low to see, but it's
written on the slide as well. Diameter root m here m is equal to n to the 2/3; take the square
root you get n to the 1/3. That's where the n to the 1/3 comes from. And that's, that's true,
that's literally true if it happens to be a tree, if it has a few more edges than a tree then its
distribution is not too different from what you'd get if it really was a tree. That takes a little bit
of thought but it's pretty straightforward.
>>: [indiscernible] right there.
>>: [indiscernible] for…
>> Louigi Addario-Berry: It's a tree with probability, yeah, you can bound it away from 0 as n
goes to infinity. So that at least explains the lower bound plus some idea of where the scaling
comes, and the 1/3 comes from if you already believe this. Okay. So for an upper bound we
have to do some more work, right? We have, now we have, we said we have some
components of this size, right? And any given one of those components is going to have
diameter n to the 1/3 already. To get an upper bound we need to say that after all those
components look up and all the smaller component hookup, that the diameter doesn't get too
much bigger than this n to the 1/3. So you need to understand how these components are
going to connect together. Okay. So the sort of fundamental idea for how we do that is a
straightforward deterministic lemma that I'll tell you in a second. Before I tell it to you, since
we've already started talking about the structure of the critical random graph, I'd like to say a
little bit more about that because now we're going to start increasing the value of p. I told you
what this looks like at p exactly 1 over n, okay? And on the other hand my comic, what I called
the comic slide said that once we're at 1.01 over n, we've sort of already determined the global
structure, so we need to understand what's going on for p between 1 over n and 1.01 over n.
Okay, which is called the barely super critical phase. And the following is a picture of that. So
let's look at Gn1 plus epsilon over n, yeah. Epsilon’s small, and order the components by size,
and here's the picture. If epsilon is small enough then you can think of epsilon as being 0; we
are really at the critical phase. And small enough turns out to mean n to the minus 1/3. So if
epsilon is this small, then we are back on the picture of the previous page. All the components
have size around n to the 2/3 and diameter around n to the 1/3, okay; the largest components
have those properties. You can, you can easily convince yourselves that this value of epsilon is
at least plausible because if you take two components size n to the 2/3, then there is n to the
4/3 non-edges between them and once one of those non-edges comes along and becomes an
edge, they'll hook up and form a bigger component. So the characteristic time on which 2 of
the sort of largest components will hookup is around 1 over n to the 4/3 and that's precisely
what you get by plugging n to the minus 1/3 over n in there. So this is reasonable. Okay. Once
you get much bigger than n to the minus 1/3, but still smaller than n the picture is sort of in a
way much more uniform. It always looks, here we have sort of two lines to say that this is, you
know, more or less up to constant factors. Once epsilon is much bigger than n to the minus 1/3
a giant component, a unique largest component is already forming that has size around 2
epsilon n. It turns out this is sort of more recent, but also sort of well studied and well known
facts, the second largest component has size asymptotically 2 over epsilon squared log epsilon
cubed n. And so we don't, I'm not saying anything about the diameter of this one, but it is
classically known that the second largest component has this size and the second largest
diameter, which might not be the diameter of the second largest component, yeah, the largest
diameter outside of this giant guy over here is of order of 1 over epsilon log epsilon cubed n.
Okay. So you don't have to worry too much about these logs. They disappear when epsilon is n
to the minus 1/3 and for the purposes of this talk you're not going to be in bad shape if you
pretend they're not there and just treat this as one over epsilon squared and diameter one over
epsilon. So you can think of that kind of picture. I'll come back to some of this again in a
minute when I need it, but it's good at least to have seen it, right? Okay. So here's the
deterministic lemma that I promised that's going to tell us how to understand how these
components hookup. Okay? The picture that's worth drawing for you is that rather than just
looking at all of the connections as they form, we're going to do the following thing. We're
going to treat the largest component of the growing tree as special, okay? And we're gonna
look at how it increases in size and diameter in stages. So first we'll increase p a little bit and
see what new connections form that are not to this component, so that are between other
components, okay. And then once we've raised p a little bit and seen some new connections
coming over here, we will look at how those new components connect to this graph. So that's
the rough picture and that's the picture you should have in mind when I tell you this, this
lemma. Okay. So I'm defining some quantity which is just, it's LP for the longest path, the
largest number of vertices on any path in a connected graph. Okay. That's certainly an upper
bound on the diameter and it's essentially equal to the diameter if we are talking about trees.
The diameter is usually in terms of edges which is the reason for that minus 1. Okay. So if I
take, if I take some fixed graph G, yeah and 2 subgraphs of G but one bigger than the other,
okay? So I have 2 subgraphs and then spanning trees of each of those graphs. This is the
situation I have with my coupling between Gnp and Tnp, right? I got increase in graphs; that's
Gnp and spanning trees of those graphs and I say what's the diameter of T prime in terms of T?
Well an upper bound is whatever diameter I started with plus twice the length of any, the
longest length of any path outside of that tree. Okay? So let me, this is quite obvious when you
draw the picture; it's just saying that you had some spanning tree of H, right, now you have a
spanning tree of H prime. How much could the diameter have grown? Well at most you take
some path here, some path here, right, any path in T prime can be decomposed in this way and
there's some portion that's in H prime and then a path in H and then another path in H prime.
The fact that it's a tree guarantees you that you can't have, you can't have more bits than this.
Okay? So that means that that gives you this upper bound immediately. This twice the longest
path is just the length, an upper bound length of these 2 paths and this is an upper bound on
the length of the middle bit. Okay? Is that clear? Okay. So how are we going to apply that? So
let's go back to our coupling and order the components by size just like I said before. Then this
tells us that if when we increase from p to p prime for some values p and p prime, if which
component is the largest hasn't changed, right, then we can apply the lemma immediately to
get an upper bound of this form. We've got the new diameter is at most the old diameter plus
twice the longest path from out here. And now you can understand where, why I was caring
about the diameters of those, those small components. It's because they're going to come in as
some sort of a bound on this quantity. Okay? So I'll keep that. So now I want to explain in a
little more detail that picture of what's happening in the barely supercritical range that I just
gave to you. And this is, there's sort of a variety of references you can cite. One appropriate
reference is much of the work in this area is due to Thomas Luczak in the early to mid ‘90s. The
particular result that I'm saying here is essentially from a paper in 1994. So in this range that
we are interested in, okay, the largest component has size 2 epsilon n and if you remove that
component, okay, then what's left essentially looks like a sub critical random graph on the
remaining vertices. Okay? So here the parameter was one plus epsilon over n. After removing
the largest component what's left looks like a random graph with parameter one minus epsilon
over however many vertices are left. Okay? So if this, if we just remove, I'd like to explain this
a tiny bit. If rather than removing the largest component, I just removed some fixed set of k
vertices, well the independence, the model tells me exactly what the distribution of the rest is.
It's distributed like Gn-kp, right, and if you want to write p in this form then now you have to
use something over n-k instead of something over n. Okay? So if you just write, rewrite p in
terms of n prime you get essentially one minus epsilon instead of one plus epsilon when this is
the number of vertices that you remove. Okay? So that's the content of this is really that even,
so this would be true even if we removed some deterministic set of vertices and it's still true
even if you remove this particular random set of vertices. Okay? Good. The nice thing about
this is that that tells us that if we look at, if we remove the largest component what we have is
essentially a sub critical random graph, so we can even increase p a little bit on what's left and
we'll still have a sub critical random graph. So in particular, if I take p prime to be one minus
alpha p, okay, then what’s outside of the giant component now is going to look like a random
graph on the remaining vertices with this parameter, 1 minus alpha times what we had before.
Okay. In particular, if I take alpha small enough that I'm still subcritical, right, so alpha small
enough that one plus alpha times 1 minus epsilon is still less than 1 epsilon over 2, okay, then I
can apply the proceeding, the earlier bounds on the length of the longest path outside the giant
that came from that picture. Okay? So the point, again, what this is saying is you were already
at the point where, so this was my comic sketch. We're already to the point where there's one
big guy. Out here we know that the largest diameter was around one over epsilon with some
logarithmic correction that I told you not to worry about, and that would be true even after we
do the thing like I described which is scatter a few extra edges out here, okay? And these edges
come precisely from increasing the diameter from p to one plus alpha times p, okay? So even
after you scatter those the diameter is still small; still you get a bound of the same order.
Good. Okay. Well that more or less is all we need, in fact. Okay? And here's why. So let me
just, and this is really the key point, so let me keep it for a second. Okay? So here's what we're
going to do. We'll start, we know what's going on at one over n, right, I described to you that's
how we got a lower bound. We know what's going on at one over n. We even know what's
going on at one over n if we add in a little bit, okay, of this form. Okay? And to be careful you
need to care about how this lambda changes things, but you'll have to believe me that even if I
pick some large fixed lambda the diameter is still only n to the 1/3. And that you can prove
using classical facts. Okay. So we'll start from here and then we'll start increasing the value of
p in this way that I described. We'll add some edges out here and see how they connect to the
largest component. Add some more edges, see how they connect to the largest component.
And we'll do that in the geometric way. Okay? So this is my starting point epsilon 0, and then
I'll just let epsilon i be one plus alpha to the i times epsilon 0. Okay? So each time I multiply
epsilon by one plus alpha to get the next one. Okay? And I'll keep doing that almost all the way
up through the critical window. Okay. So I'm thinking when I say lambda large I want you to
think of that as saying large enough that we can start pretending that this picture is really the
correct picture, that there's one large component and all the rest are sort of moderately well
behaved. And that holds until we get almost to epsilon equals one. It's only really epsilon
equals little o of one, but let's pretend sort of that one over lambda, this large lambda is a proxy
for small, so we can continue until we get almost up to 1 .01. Our new 1 .01 is 1 plus 1 over
lambda, okay? So what happens, this fact in blue up here, plus the bound from the corollary on
the diameter outside tells you that when you go from stage i to stage i plus 1 the diameter
increases at most this quantity with the value epsilon i that we just wrote. Okay? Which as I
said if you want you can pretend it's just 1 over epsilon i. Okay, good, because we know what
we start with and we know what the difference is at each step, right? And I've just rewritten
that using the value of epsilon. It's n to the 1/3 and if you keep track of the log you get i over
one plus alpha to the i. Okay? So just sum over the whole range. This sum is just, this sum is
dominated by some geometric sum, right, and so even if we sum to infinity these terms we get
something bounded, and so that gets us almost all the way through the critical window and we
still haven't lost more than an extra factor n to the 1/3 in the diameter, so the diameter may
have increased by a constant, maybe even a large constant factor, but not by anything more.
Okay. Great. So I mean, this, if you believe this, it should already convince you that n to the
1/3 is the right answer because we said that at time 1.01, right, one plus one over lambda, all
the other components are already logarithmic, so they shouldn't, it'd be surprising if they had
any very large effect on the diameter, but let me really prove that to you as well.
>>: [indiscernible] order epsilon…
>> Louigi Addario-Berry: No, no. I mean alpha, I think you can take alpha as equal to 5/4, for
example. It's really some, all you need is that each time…
>>: Oh you mean [indiscernible]
>> Louigi Addario-Berry: I mean alpha is a quarter, yeah, so remember…
>>: Didn't you have to [indiscernible] alpha times one?
>> Louigi Addario-Berry: I had 1 plus alpha times 1 minus epsilon is less than 1 minus epsilon
over 2.
>>: [indiscernible] epsilon
>> Louigi Addario-Berry: Oh, sorry, yeah. So that was, this is what I meant.
>>: [indiscernible]
>> Louigi Addario-Berry: So here we have, so we have a random graph with p as around 1
minus epsilon over n, right, and so we can afford 1 minus epsilon over 2 over n, yeah. So I can
take alpha, yeah, I can take alpha like epsilon 1 over 2, yeah. And let me write this
>>: [indiscernible]
>> Louigi Addario-Berry: Say that again.
>>: [indiscernible]
>> Louigi Addario-Berry: Let me write this as 1 plus -- p is 1 over n plus lambda over n to the
4/3. Okay. Then, then to get from, yeah. So I can -- that's fine. So I can, so epsilon is -- so I can
take epsilon like lambda over 2 over n to the 4/3 at the -- so I can, I can increase, right, from 1
minus lambda over n to the 4/3 to 1 minus n minus lambda over 2 n to the 4/3 in the first step.
Yeah? And when I do that, the parameter p increases, right, so p prime then is like 1 plus 5
lambda over 4, 1 over n plus 5 lambda over 4 over n to 4/3. Right? Okay.
>>: [indiscernible]
>> Louigi Addario-Berry: So…
>>: [indiscernible]
>> Louigi Addario-Berry: So, I mean the point is that epsilon needs to geometrically increase,
right? Epsilon needs to geometrically increase so that this sum can geometrically decrease. So
I may have mis-parameterized alpha in this statement, but the point, but even if alpha is epsilon
over 2, right, then we still get a geometric increase in the parameter epsilon i if we add these, if
we add these epsilon over 2s at each step, right? So then epsilon i plus 1 is really epsilon i times
1 plus…
>>: [indiscernible] I don't see how it can work the way you wrote it at 1 plus alpha because I
know than the alpha [indiscernible] becomes too small [indiscernible]. Maybe I'm missing
something. But if you basically are choosing each side of your alpha to be the current epsilon
over 2…
>> Louigi Addario-Berry: Yes.
>>: Then only the first few alphas are small and then…
>> Louigi Addario-Berry: Yeah, and then they grow geometrically.
>>: [indiscernible] formula [indiscernible] because then you have a fixed alpha.
>> Louigi Addario-Berry: No. You're correct, so that came from hastily preparing the slides and
not looking at my old paper, but, no. You're right. So this looks, alpha needs to be small when
epsilon is small, but if you take alpha, I mean in the first step epsilon gets a constant factor
bigger than you can take a alpha constant factor bigger, right? And the point is that then these,
I mean the key point is that these epsilon i -- if you do that, if you start with if alpha i is of the
form constant greater than 2 to the i divided by n to the 4/3, right, then these, then the
corresponding epsilon i are still decreasing geometrically quickly. And so…
>>: [indiscernible]
>> Louigi Addario-Berry: Yeah, the epsilon [indiscernible] are increasing geometrically.
>>: [indiscernible] so now you instead of alpha, you want alpha i?
>> Louigi Addario-Berry: Yes.
>>: [indiscernible] so how do you [indiscernible] to finish?
>> Louigi Addario-Berry: So [indiscernible]
>>: [indiscernible]
>> Louigi Addario-Berry: Let's take that alpha i equals 5/4, 5/4 to the i, alpha and i. Let's…
>>: [indiscernible] increase geometrically [indiscernible]
>> Louigi Addario-Berry: So I think this is fine.
>>: [indiscernible] epsilon done to alpha?
>>: 2 alpha.
>> Louigi Addario-Berry: I mean, you know, and each…
>>: It's still [indiscernible] epsilon [indiscernible] is alpha i times [indiscernible]
>> Louigi Addario-Berry: Epsilon i plus 1 is 1 plus alpha i and that's [indiscernible]
>>: So that means that epsilon i is occurring exponentially and i squared [indiscernible]
>>: No, because of the 1 plus…
>> Louigi Addario-Berry: So sorry for the confusion, is the picture now clear? So let's, let's go
on from here though. I mean, so we can essentially repeat the same idea, right, starting now
from the barely super critical phase to go all the way up to the threshold for connectivity, okay?
So now, right, we want to go, for example, from 1 plus 1 over, 1 over n plus 1 over lambda n to
1 over n +5/4 over lambda n, right, and that increase in the parameter, the duality principle still
tells us that what is outside of the largest component looks like a subcritical random graph with
parameter 1 over n prime minus alpha over n prime. Okay? And the point is that once epsilon
is of the order of 1 is not much smaller than 1, okay, the sort of 1 plus epsilon 1 minus epsilon
dichotomy can't literally be true, because for epsilon greater than 1, it's sort of trivially,
obviously false. Okay? But there's still a duality between the supercritical, between the
supercritical graph and some sub critical random graph after you remove the largest
component, okay? And in that subcritical random graph all of the components will have
logarithmic size even after you increase the parameter a little bit. Okay? By some constant
factor and so now starting from 1 over lambda n we only need to increase the parameter by a
constant factor logarithmically many times to get all the way to 1, okay, so even if at each step
in what's left we, we do the worst thing which is increase the diameter by an additive factor log
n. At the and we've only increased things by at most log squared n. Okay? So I mean, you can
in fact write these 2 arguments in, these 2 steps in a unified way. I mean in both parts you're
just supplying this diameter bound for a judiciously chosen sequence of increasing sequence of
probabilities. Okay? But sort of conceptually things look a little different when you're in this
critical window and you're above the critical window so I prefer to present it like that. Okay. So
I think that's all I have prepared for today. Thank you. [applause]
>> Yuval Peres: Questions?
>>: Is the diameter of the -- the suppose I want to look at [indiscernible] spanning tree again
[indiscernible] very small?
>> Louigi Addario-Berry: Yeah. I think that you can get log or even log log n if you were willing
to spend a tiny bit more and I think there's some more on that.
>>: We should know [indiscernible] [multiple speakers]
>> Louigi Addario-Berry: There may even be a poster [indiscernible] [laughter]
>>: That wasn't [indiscernible] [laughter] [multiple speakers]
>>: [indiscernible] object. That was a flattery question. You subliminally absorbed it.
[indiscernible]
>> Louigi Addario-Berry: So this one? [laughter]
>> Louigi Addario-Berry: So this is like spring electron. I think it's called spring electrical
embedding. It's some mathematical package. The one on the first slide is just uniform points in
the in the plane. This one is really a [indiscernible] of the complete graph with some
embedding in the plane, so the colors in this picture represent edge weights. So all of the black
edges have weights at most one over n and then again, colors, this time colors code the weights
of edges starting from red and increasing to violet.
>>: And the edges have the [indiscernible]
>> Louigi Addario-Berry: The edge is not of the appropriate line. The edges have lines given by
embedding of the complete graph as [indiscernible].
>>: [indiscernible]
>> Louigi Addario-Berry: You'd have to ask [indiscernible] that.
>>: Do you also have concentration of the diameter around the 2/3?
>> Louigi Addario-Berry: So you… 1/3?
>>: [indiscernible]
>> Louigi Addario-Berry: So you won't have -- I don't expect you to have anything to good on
the lower tail, though I'd have to think about it.
>>: What did you mean by [indiscernible]
>> Louigi Addario-Berry: I mean, we can show, we can show in a different paper that in fact the
diameter divided by n to the 1/3 [indiscernible] distribution and that the limiting random
variable has a density, so it's not a, it's not a [indiscernible]
>>: What's the [indiscernible]
>> Louigi Addario-Berry: Yeah, I mean it's, we don't find an explicit formula for it or anything.
>>: [indiscernible]
>> Louigi Addario-Berry: I expect that if you were, you could get some sort of at least
exponential and possibly Gaussian upper tail bounds.
>>: [indiscernible]
>> Louigi Addario-Berry: And, he didn't explicitly state those, but I expect, I expect you could at
least extract exponential [indiscernible] so I mean the, the point is, you know, everything that
this -- one question for answering that is how, what sort of tail bounds do you have on the
diameter, the largest diameter outside of the largest component and there you can get quite
strong bounds which I believe are Gaussian, so you need to understand how these, how these
increments contribute to some final bound, but I expect with some work you could again get
Gaussian tail bound out of that.
>>: So maximum [indiscernible]
>> Louigi Addario-Berry: I'm not sure.
>>: [indiscernible]
>> Louigi Addario-Berry: No. I don't think [indiscernible] maximum…
>>: [indiscernible]
>>: [indiscernible]
>> Louigi Addario-Berry: I mean, there's some, there's some explicit formula for the degree of
the root which comes from some mixture of poisson distributions, and you should essentially
have independence between different parts of the tree. So a guess based on that is that it
should be like log n over log log n. But I haven't proved that, yeah.
>>: [indiscernible] has to be [indiscernible] has to be given [indiscernible]
>> Yuval Peres: Okay, let's thank Louigi again. [applause]
Download