17926 >> Kamal Jain: Hello. My name is Kamal... Oveis Gharan. And sorry if I mispronounce your name. ...

advertisement
17926
>> Kamal Jain: Hello. My name is Kamal Jain. Today we have a guest, Shayan
Oveis Gharan. And sorry if I mispronounce your name. He came here for an
interview for his fellowship candidate, but he thought he could give a talk since
he had some time. And I think it's a talk prepared on short notice. So people on
the camera, or later, so he proved recently a break-through result in a
long-standing problem which is asymmetric traveling salesman problem. And it
could be considered one of the center problem in approximation algorithms. So
Shayan will explain his new ideas here.
>> Shayan Oveis Gharan: Thank you. Okay. So today I'm going to talk about
the asymmetric traveling salesman problem. And I'll try to tell you some of our
main ideas for this problem. I will tell them in the talk. So that's a start.
So here's an outline of the talk. First I will define the traveling salesman problem,
and I will tell you a brief history about it. And then I will define the concept of thin
trees and its relation to ATSP. And, finally, I'll try to discuss some of our main
ideas for these approximation algorithms for ATSP.
Let's start with the problem definitions. Traveling salesman problem is general
form has a simple definition. Suppose like we have a set of cities. There's some
cost between them. And the traveling salesman who wants to visit each of these
cities exactly once. So, in fact, we want to find the shortest tool such that that
visit vertex is exactly one.
So this problem was initially proved to be implicated by [phonetic] Karp in 1972,
and it was one of the first problems to be proved to be in the complex. Traveling
salesman problem in its general form is very hard to deal with. So we'll relax it a
little bit.
So, in fact, we consider this a little relaxation. So in ATSP, we have a set V
vertices and a nonnegative cost function. We're looking to find the shortest
closest tool that visits each vertex at least once. We have substituted that
exactly once with at least once here.
For example, consider this graph. The cost of each edge is one in this graph.
And as you see, this tool I plotted in red is a tour of cost tool. And it visits some
vertices more than once. In terms that if you have this property you can assume
that the cost functions satisfies the triangle inequality. I.e., this inequality. So
you can do that by easily considering the shortest path distance of going from U
to V as the cost of U.
So, in fact, symmetric traveling salesman problem, which is well-known traveling
salesman problem, we'll call it STSP here, is a special case of ATSP. The main
difference is that in that problem the cost of UV, we call it from the cost of from
VG.
So there are quite a lot of work to solve this ATSP. This is, in fact, one
interesting instance. This tool is found on like over one million cities throughout
the world and the cost of cities only six person of the optimum. In fact, the
heuristics works very well for this problem. But here we'll focus on the
approximation algorithm.
So let me describe, TSP has a lot of applications in various areas. Here is one
application in genome sequencing. In genome sequencing there's a process
called shotgun sequencing which divides DNA into a set of short fragments. And
the question is you want to assemble these short fragments in order to obtain the
original DNA.
So if you think, you'll find out -- sorry, in fact, this problem is the optimization
problem. We want to find the original DNA to be as short as possible, be as
short as possible in length. So if you think you'll find out that this is in fact
minimum super sim problem. Super sim problem, you have a set of streams and
you're looking to find a super string with the smallest length such that given each
string, it's a sub string of the super string and you can convert that problem to the
ATSP problem.
So, in fact, here these five vertices correspond to five different streams. ATG
and C are the DNA basis. And we're looking to find like this super string, each of
these strings is a sub string of it. And the tour is the optimum tour.
One difference is that here what we're looking to is to find the maximum ATSP.
Note here that the cost of going, for example, from this vertex to this vertex is
maximum suffix of this, which is equal to the prefix of the other. So max ATSP
can be solved within constant factor. So this problem can be solved in constant
factor in approximation.
Let me briefly define what do I mean by approximation algorithm. Suppose that
we have a minimization problem. We say that an algorithm is off a factor, is
factor of optimization. If it only produces a solution, which is off a factor of the
optimum.
So the main difficulty is that when we're looking for this approximation algorithm,
it seems to be hard to find actual optimum, otherwise we'll find the optimum. In
fact, if it's hard to find the optimum, how can we compare the cost of our solution
with the optimum. So typically we will use a lower bound on optimum solution.
For that we'll use a linear programming relaxation. This linear program relaxation
always gives you the lower bound for the optimization problem.
So here are now the results for both symmetrics, asymmetric ATSP. You can
see there's a lot of work. I've only mentioned a small portion of them. So this is
the best constant factor approximation algorithm for STSP, which is tree
approximation by Christofides in 1976.
On the other hand, for ATSP, there are a huge number of works. Like four works
here. But as you see, only the factor in front of the log n has improved. So they
all give up an order of magnitude for general approximation.
But for symmetric here, see there are quite a lot of works. It's a special case
when our metrics is defined on a special graph. In particular, polynomial time
approximation is key, having found first for metrics defined on unweighted planar
graphs than for metrics defined on weighted planar graphs. And finally for
weighted bounded genus graphs.
>>: Are they run by STSP, too?
>>: Shayan Oveis Gharan: This is for TSP. There isn't any special results for
ATSP in specially bound [indiscernible] graphs. So here are our main results.
We finally broke the log n barrier. We give the first order of log n over log log n
approximation algorithm for general metric ATSP. This is joint work with
Asadpour and Saberi from Stanford and Mitchell Goemans Alexander Madry
from MIT. Moreover, we give the first constant approximation algorithm for
shortest path metrics defined unweighted bound genus graphs. This is joint work
with [inaudible].
So as you see, similar to the symmetric TSP, where the approximation factors
are much better for the special case, here you have sort of the similar thing for
ATSP where our approximation factor is constant for if it's a special case.
Okay? So before I tell you our approximation algorithm, I need to define the
Held-Karp relaxation because I want to compare our solution with the Held-Karp
relaxation. This is the definition of the Held-Karp relaxation, this Held-Karp
relaxation is proposed by Held and Karp in the 1970s. And it gives a lower
bound for the optimal solution of ASTP. So here's some notation for set S delta
minus of S is a set of edges that are pointing to this set; delta plus of S is those
pointing out of this set.
So, in fact, the fractional solution of these relaxation is the vectors where they
have two properties. The size that -- there are in fact a fractional one flow.
Because the minimum size of the cut is at least one, the size of each cut is
at least one, as seen in this inequality. The inequality says that the amount
of incoming flow to each vertex is equal to the amount of outgoing flow,
equal to one. So it is in fact flow as you see, because the amount of
incoming is equal to outgoing and it is one flow because the amount of
outgoing edges down to flow through these edges is one.
Also remember at L top S is the union of these two sets. It is all the edges
in this cut.
Okay. So in fact if you see, you'll find out that the optimal solution of ATSP
is in fact an integral one flow, not a fractional one flow. So it has both of
these properties, both of these inequalities integrally. This is why this
linear programming is relaxation for ATSP and therefore its optimal value is
a lower bound for the upper bound of ATSP. Moreover, optimum solution
of the relaxation can be computed efficiently using ellipsoid algorithm.
And also integrality gap is defined as the maximum ratio between integral
solution over a fractional solution.
So there are also quite a lot of work to find integrality gap both STSP and
ATSP. As you see, there is this gap here for symmetric TSP. And even
improve it by epsilon factor with greater result. For ATSP core
[indiscernible] conjecture that integrality is equal to four and equal to
symmetric TSP. But then [indiscernible] Goemans [indiscernible] found a
set of counter examples where the integrality gap is at least two where they
show that integrality is at least two. So it's open whether integrality gap is
constant or not. So here's a simple example that shows that integrality gap
is at least one and a half. Consider, this is the same graph I said before.
The cost of minimum tour is 12. But if you consider the half integral
solution in this graph, you'll see that it is in fact a fractional solution of
Held-Karp relaxation, its cut is eight. So the integrality gap instance is one
and a half. And if you play with it, you can find a set of examples where
integrality converges to 2 if N goes to infinity.
Okay. So let's try to work on the approximation algorithm. First I will go
through the thin trees and we'll define it and tell you its relation to ATSP.
So what's the thin tree? A thin tree is that spanning tree that has
fractionally small number of edges in each cut, small number of edges.
So in particular, we say that spanning tree is also comma sigma ten if it has
both of these inequalities. This says that the number of edges of our tree
in any cut is at most alpha times the size of that.
So remember that delta of S is all the edges of this cut. Except L top S is
the sum of the fractions of all the edges of this cut. So I want that my tree
has roughly small number of edges in this cut.
And, moreover, the cost of my tree is within sigma factor cost of X.
So if our tree has both of these properties, I say it's alpha comma sigma.
So is the definition clear? There is this observation which says if you have
an alpha comma sigma thin tree with respect to optimal solution of ATSP,
then it is possible to find an integral invariant tool, variant tool based on
this thin tree where its cost is at most two alpha plus sigma times cost of X.
Remember that X was an optimal solution of Held-Karp relaxation,
therefore its cut of lower bound ATSP, therefore the cost of -- therefore,
this tool is an alpha -- is an alpha plus sigma factor optimal solution of
ATSP. I will not prove this observation for you, but the proof is mainly
maximal mean cut argument and all you need to do is find the minimum cut
invariant augmentation of this tree.
And for that we will use this fractional solution of Held-Karp relaxation.
Okay. So here's an overview of our algorithm. So it has some similarities
with well-known Christofides algorithm for symmetric STP. In fact,
suppose that this is a fractional solution of Held-Karp relaxation and this is
our thin tree, these green edges. So what I need to do, I need to add some
edges to this tree such that the point of graphical formulary. So sometimes
I have to add multiple copies of an edge.
So here are corollaries. First of all, if we had a polynomial time algorithm
that can find constant comma, comma constant thin tree with respect to an
optimal solution of Held-Karp relaxation, it can be a constant factor
approximation algorithm for ATSP.
Moreover, even if we cannot find such a tree, when we can prove its
existence, this implies the constant integrality gap of ATSP. So it's a great
result. So even the existence of this thin tree is a great result. So let me
show you an example what these thin trees are.
Suppose we have this graph. So if you -- in this graph, the cost of each
edge is one. And suppose that I consider half integral solution in this case.
So, in fact, this will be a fractional solution of Held-Karp relaxation.
Suppose we have these two thin trees. It turns out that this has been a
better thinness. And why is that true? Because consider this cost, that's
the previous version from rest of the graph. This tree has two edges in this
cost. This has four edges. But let's see the size of the cost. Because the
fraction of each edge is a half. The size of this cut is two for both of them.
So this tree is 2 over 2-1, 1 comma 4 over 5 thin. But this is 2 comma 4
over 5 thin.
We call this 4 or 5 is related to the cost of our tree versus the cost of the
graph. In fact, this is alpha and this is sigma.
So if you see, this star-shaped tree has fractionally a small number of
edges in many cuts of the graph. So, for example, if you consider this cut
separate this vertex, this tree has two edges, but this one has only one
edge. But you can see there's a very bad cut for this tree, its thinness will
be much worse than the other, than this [indiscernible].
So if you see this example, you'll find that the duration of the edges doesn't
have anything to do with the thin tree. If I drop the direction, I will have the
same thinness, because the definition of the thin tree I've never used -- I
haven't used the direction.
So this is a great result. We can drop the direction with this thinness
property. So if you drop the direction, what you get is a fractionally two
connected graph. Because, remember, that X was a fractional one flow, the
amount of outgoing edges of each set was one. The amount of incoming
edges was one. If you drop the direction, you get fractionally two
connected graph.
Moreover, you can either consider our fractional graph or an integral graph.
So you can convert our fractional graph to integral graph simply by
multiplying the fraction of each edge to some large number of K. For
example, consider this graph, you multiply every fraction by 2, you'll get
this graph.
So, in fact, what we need to do is either to find an alpha comma sigma thin
tree in a fractionally two connected graph, or alpha over K comma sigma
over K in a KH connected graph.
So remember if we multiply everything by K, I get a KH connected graph
but instead I need to prove alpha over K comma sigma over K thin tree,
instead of alpha comma sigma thin tree because we multiply everything by
K.
Also we can do one more simplification. We can get rid of the cost
function. Why is that true? Let me define another distribution for these
thin trees. I say that a thin tree, spanning tree is alpha 10, not alpha
comma sigma. Only alpha 10. If it has only the first inequality. It has
fractions of any one of the edges of the graph. Suppose this is true. I'll
show you that we can get rid of the cost function. How can we do that?
Consider KH connected graph. Suppose it has like C over K thin tree for
some constant C. We can extract this thin tree from our KH connected
graph. And because this tree has small number of edges in any cut, the
graph will still be highly connected.
So we can do this again. We can consider again a thin tree in the new
graph and extract it. So if we do that, you'll find out after some time we'll
get a bunch of thin trees and we get the tree with the smallest cost. And
this will be a good sigma for it, for us. So all I need to do, I need to find
alpha thin tree. I need to have only this inequality for two to the N cuts.
So to recap, I need to do, either to find alpha thin tree in a fraction in a
connected two graph or alpha over K thin tree in a KH connected graph.
Both if I can solve either of these problems I will get alpha approximation
algorithm for ATSP. Okay. So from now I will only focus on how to find
these thin trees.
So let's think about it. For ATSP, first we have some problem in directed
domain. Then we try -- we find this notion of thin tree and we converted the
problem from directed domain to undirected domain. But if you think,
you'll find out that finding these thin trees is not an easy problem. In fact, if
you have a graph and you know the best thin tree in this graph, and you
know its thinness, proving this thinness to someone else is a difficult
problem.
Currently the best approximation algorithm for this we know is log n
approximation for finding the thinness. In fact, this thinness problem does
not seem to be an MP.
So it seems to be a very hard problem. In fact, at least ATSP was an NP.
So it seems to be much harder. But as you will see, in some cases we can
measure the thinness with some tools. And that leads us to approximation
algorithm for ATSP.
So here are our algorithm. First I will show you a deterministic algorithm.
This deterministic works for shortest metrics defined unweighted bound
genus graph. It gives constant over K thin tree with respect to KH
connected bounded genus graph. And for certificate I will use properties
of dual graphs.
Also when we're working on this algorithm, we found out that
mathematician called Louis Goddyn conjectured that any chain connected
graph contains constant of K sentry. If his conjecture is true, then the
integrality gaps of ATSP would be constant.
Then I will present a randomized algorithm. This will work for general
metrics ATSP. And it will find an order of log n over log log n thin tree with
respect to fractionally two connected graphs. I would find constant over K
thin trees that do KH connected graph. But here is fractionally, with
respect to two fractionally connected. Both of them will give us, so this will
give us constant trees -- constant approximation for ATSP will give us -- it
will give us order of log n over log log n approximation. So first I'll discuss
the deterministic algorithm and then randomness.
So okay. We're good. Next I will show you this constant factor
approximation algorithm for finding the thin trees unbounded graph.
So let me define what I mean by bounded genus graph, an orientable
surface with genus gamma is simply in sphere with gamma additional
hand-offs. For example, this surface is the surface for genus tree. And we
say that a graph has genus gamma if it can be embedded on genus gamma,
while edges don't cross each other, don't cross each other. For example, a
genus planar graph is zero, can be embedded on the surface and you'll find
genus of complete graph is an order of n to the 2.
For example, if you have a city with many one-way streets and a few
bridges or underpasses, its genus is bound, or genus is small because you
can easily embed it on a surface, genus embedding by each bridge or pass
by a handoff.
So this is our main theorem. We give a constant factor approximation
algorithm for ATSP metrics defined weighted bounded genus graph or
approximation factor is in order of square root of gamma log of gamma,
and gamma is the genus of the graph. So with that even if gamma is not a
constant but is sufficiently small will get good approximation. So in fact if
gamma is like order of log n we get order of square root of log n
approximation algorithm.
Okay. So truly our algorithm implies constantality gap of Held-Karp
relaxation for bounded genus graphs, prove integrality for thin trees. It will
have more -- it's a bit more general. In fact, even if the inputs instance of
ATSP do not have bounded genus but we can find an optimal solution of
Held-Karp relaxation where it can be embedded on a surface bounded
genus, then again we can use this algorithm. So we do not need the input
graph to have bounded genus, we need to have fractional solution to have
bounded genus.
So here is the overview of the proof. First, I will tell you an unsuccessful
attempt. Here I mainly prove our algorithm for planar graphs. I'll tell you
some of the ideas for extension for non-planar graphs. For planar graphs
I'll tell you unsuccessful attempt and I'll tell you our certificates of
measuring thin tree and I'll tell you our thin tree selection algorithm. For
that I will use the concept of trees defined. So let's use this by
unsuccessful attempt. Suppose we have this graph and we're looking to
find a thin tree in this graph. This is 5-H connected. So what's a
[indiscernible] certificate. So suppose we have this tree in green edges.
So one very simple idea is to chart it at with its parallel edges. And you
can prove that if this edge will -- well, will contain a cut if and only if the
power edges contain that cut too.
So this implies 1 over 5,1 thinness of this tree. Okay. So let's see if we can
generalize this. The answer is no. The answer is yes, but it will not work
for any planar graphs. So suppose we have this graph and consider this
cut and consider four edges in this cut. And because I'm looking to find
the spanning tree I need to select at least one edge from the edges of this
cut.
But none of the edges of this cut have any parallel edges. So I cannot use
the previous idea. So instead what I'm going to do, I'm going to first select
these green edges. The first step I will not have any problem. I can select
these green edges and charge them with the parallel edges. This is the
first. So an idea might be that if we select these edges, what if we contract
and try to solve the problem recursively? That seems to work. Seems to
work. So let's try to do that. So if we do that, we get this graph. Again, we
can select these green edges and charge them with the parallel edges and
then contract it and finally select this edge.
So if you see, you'll find out that these edges will make spanning tree.
Why? Because each time we contract a set of edges we select at least one
of them as an edge of our thin tree. So that planar graph will be spanning.
But let's see -- you might expect the final graph to be 1 over 14 because
each time we charge its edge to parallel edges and have three parallel
edges.
So unfortunately that's not the case. In fact, this is the final graph. And
this is the tree I selected. If you consider this cut, our tree has three edges
of this cut. But the size of the cut is 7. So at best it's 3 over 7, 3, 7, 10. And
why did that happen? The problem is that consider this edge. This is the
edge of selecting the last set of our algorithms. You charge this edge with
these three edges. But in this cut only this edge is contained. None of the
edges we charge is contained. So in fact our charging method will work
only in the first step. When we contract the edges it will not work.
But in the rest of this proof I will show you that the same algorithm will
work with a bit of modification.
Okay. So let's reach our certificate. Let me tell you how we are going to
measure the thinness of our thin tree. So here's an observation. We'll use
a simple observation planar graphs. In the planar graph, dual of any
minimal cut is a cycle. And vice versa. The dual of a cycle, the dual graph
is the minimal original graph. For example, in this graph the dual of this
cut in blue is this red cycle.
So if I want to prove the thinness on my tree, all I need to do is to prove its
thinness in the minimal cuts, just to measure its thinness in the minimal
cuts. For that, all I need to do is to measure its thinness in the cycle of the
dual graph. If my tree has fractionally small number of edges in each cycle
of a dual graph, it will be thin.
So how can I do that? Suppose that I can select a tree where the distance
of its edges in the dual graph is at least P. So this in fact implies 1 over
this thinness. And why is that true? Because our tree can have at most 1
over D fractions of the edges of each cycle in the dual graph. So, for
example, suppose that this is a cycle in the dual graph. This is equal to
tree. So the worst case my tree has at most two edges in the cycle.
>>: [indiscernible].
>> Shayan Oveis Gharan: The distance is the shortest path between the
edges in the tree. So the distance of this dual tree.
>>: Between the edges or?
>> Shayan Oveis Gharan: Between the edges of the dual. So I'm
considering my tree. Considering the edges of the dual graph. If they're
far apart, they're at a distance D with each other. At most my tree would
have 1 over D fractions of edges of n cycle, as you see, for example, on this
example.
So because it has at most one over D fractions in each cycle it will have at
most one over each fraction n cut of the original graph and one over
distance.
So all I need to do is to find a set of edges where there are far apart in the
dual. And also they make connected sub graph in the original graph
course. So, let's see, it's true in this example. Remember I proved this is 1
over 5 thin. So let's see our thinness certificate. If you see -- the dual is
plotted in red here. So these edges are selected. Have distance 5 in the
dual graph.
So this implies perhaps the thinness of our tree is 1 over 5. So next I will
tell you our thin tree selection algorithm. I will tell you how we can find
these far apart edges. So for that I will use tricks. What is a thread? A
thread is a path, you may think of it as a long path, where all of its vertices,
internal vertices has degree 2.
So if you see, you'll find that the dual of the thread is a set of parallel
edges. So suppose that I have a thread like this thread. And this edge, its
middle edge, is the only edge of this thread I've selected as an edge of my
thin tree.
So no matter what the other edges of my thin tree, the distance of this edge
from the rest of the edges will be at least D. At least D over 2. If D is the
length of the thread. Suppose length over thread is 5 here. This is the
distance of this edge from the rest of the edges will be at least 5 over 2.
So what will be an algorithm, if you want to? Each time select a thread dual
graph, pick the edges, the edges of our thin tree and then delete it. And do
it recursively.
And that's exactly our algorithm. These are relatively fine long threads
dual at single edges of the edges of our thin tree and deleted tree. Note
that interestingly the algorithm works with the dual graph. But will show
we'll select the thin tree in the original graph.
So let's go to the proof. First of all, what I need to show is that selected
edges are far apart. This is true by definition of thread, as I said before.
One more thing is that I need to prove that the selected edges would make
a connected spanning sub graph. And why is that true? The reason is so
simple. Consider the thread as the dual of the thread, a set of parallel
edges. So when I'm deleting these parallel edges, sorry, when I'm deleting
the thread edges, in fact I'm contracting these parallel edges in the original
graph.
So if you simulate our algorithm in the original graph, each time you select
a set of parallel edges, one of them as an edge of our thin tree and then
contract it. And we'll do it again. This is an algorithm I said before, but this
time I pick my edge of, pick the edge of my thin tree smartly. And, trivially,
the set of selected edges would make connected spanning sub graph as
before.
So it turns out, if you select it smartly like this, it will give you thin tree.
Also, there is one simpler proof, which says this algorithm will select at
least one edge from any cycle of a dual graph. If you think -- the proof is so
easy, because each time you pick a thread and all the internal methods of a
thread has degree two, the algorithm should select at least one edge of the
edge of dual graph. Now because the dual of any minimal cut is a cycle our
algorithm will select at least one edge from any cut.
Okay. Great. So the only thing I haven't proved yet is the existence of
these threads. And it turns out that they're real easy. So if you can prove
that if we have a planar graph of girth D it contains a thread of length G
over 5.
So this implies that, if you do the calculations, this implies that our tree will
be 1 over G starting. So once G is started you start the gear of the dual.
Remember our algorithm was working with -- we were trying find this
thread dual graph so it's 1 over GS starting thing, order of 1 over GS
starting.
So the only thing I need to show is that GS star is large, and this is the case
in fact. In fact, for planar graph high edge connectivity implies high dual
graph, because the dual of any minimal cut is a cycle. The dual of any
cycle is minimal cut because the minimal cuts have size K, the K of dual is
K. GS star is K and we get order of 1 over K thin tree.
And it is in fact 10 over K for planar graphs. So, yes, this is our thin tree
selection algorithm. So for extending this idea to nonplanar graphs I will
not tell you much about it. The main -- I will tell you the main idea. The
main idea is that for nonplanar graphs the dual can have very small cycles.
It's not the case that if our graph is highly connected the girth of the dual is
so large. In fact, it can contain very small cycle. But fortunately we can get
rid of these small cycles. And the idea is that we can somehow delete their
edges. The minimal algorithm is we consider the smaller cycles, delete
their edges from our original graph. If we do that, because they're small,
they will not disturb the connectivity of our graph and we can convert our
graph to planar graph and everything will be fine.
But there's some tricks. Okay. So this is the ideas of our some of our
concept approximation algorithm. And next I will focus on our order of log
n over log log n approximation algorithm.
Is there any questions? Okay. So let's move on. So again remember that
for in this case we are trying to give a randomized algorithm. It will work
for general metrics ATSP. It considers a fractionally two connected graph
as input and it's looking for an order of log n over log log n thin tree.
So let's start by simple observation. Suppose that this vector of y is the
vector of our fractionally two connected graph. But if you see, you will find
out that this vector Y is in fact not too far from the tree itself. In fact, if we
multiply by 1 minus 1 over n what we'll get is a fractional spanning tree. It's
a spanning tree that's a vector applies to top line vertices of our graph.
And why is that true? In fact, because this vector Y is fractionally two
connected if we multiply by 1 minus 1 over N, still the fractional size of the
edges any cut is bigger than 1, at least 1. Moreover, if you do the
calculations, you'll find out that some of the older fractions in y, some of
the older fractions of edges in y is n.
Because it turns out from the properties of Held-Karp relaxation and you
multiply size 1 minus 1 over n the sum of the fractions edges will be n
minus 1. It will be a fractional spanning tree.
So in fact all we need to do is to consider fractional spanning tree z and
find spanning tree t with respect to z. I mean, that is t with respect to z
across any cut.
Note that previously we were looking to find a tree that is t with respect to
y. But because z is roughly equal to y within some small factors if we find
a tree that's thin with respect to z everything is fine.
So let me start by unsuccessful attempt. Suppose that I round each edge
uniformly at random. Each edge of the -- with its probability. If it is z, if it
has a fraction of ze, I round it with probability z. So if I do that, the
expected number of edges of my tree across any cut, across any cut is
equal to the size of that cut.
So suppose that, for example, this is the cut. This is one of the cuts. So
what I need to have, I need to have a tree that is spanning. I need that this
selection algorithm selects at least one edge across any cut.
But if you want to have that, you need that the size of the cut, expected size
of the cuts to be at least order of log n. Otherwise there will be some cuts
which are not of the edges selected by the algorithm.
So but if you have this, the thinness of our tree cannot be better than order
of log n.
So what we're going to do instead of just uniformly round the edges, you
will select -- we have log n thinness here. So instead we will select a tree T
from some specific discrete distribution that preserves the marginal
probabilities.
So in fact the probability of selecting an edge is equal to z of e, where z is
the vector. So if you have such a distribution, what we'll get is that again
expected number of edges of our tree in any cut is equal to the size of that
cut. But here we will have something more.
Because we are sampling the spanning tree, it will have at least one edge in
any cut. It is connected. So we'll not have any problem for that. But the
only thing we need to prove is that we have concentration for this equality,
only in the right tail. We need to prove that the number of our edges of our
tree is not that large in any cut. So previously we need to -- we had to
prove its concentration in the left side, too, because our tree's spanning it
has at its edge one cut and we need to have concentration in the left tail,
and we need to have concentration on the right and we have 2 to the n
equations.
So this is why we can get better approximation if we want to measure -- if
we want to have concentration on the right tail, because the chain of bound
for the right tail gives us a better bound, comparing the share of bounds for
the left tail.
So let's see how we can get concentration. Previously, because we were
sampling the edges uniformly at random, independently uniformly at
random, we had independence, and we could use shared types of bound.
But here we cannot because we were sampling a tree and the edges are not
independent or are not selected independently. So we need to use
negative correlation.
So if we do not know the definition of the negative correlation, we say that
sampling procedure has negative correlation if the selection of an edge
into our, for example, tree can only decrease the probability of other edges.
In particular, for n is set of f of edges probably that all of these edges are
selected in our tree less than recall their marginal, the multiplication of
their marginals.
So remember that if you had negative correlation, then you would have
equality here. But here -- sorry if you have independence you would have
equality here. But here because you have negative correlation, it gives
inequality. But in fact this is all we need to have the chain of types of
bound. If we have negative correlation, then we can prove the correctness
of any general types of bounds.
And this is the idea. So in fact there are three possible, currently three
possible approaches to sample our random spanning tree while having
negative correlation.
And while having the probability of selecting edges equal to its marginals.
So this is the method maximum distribution used in our paper. It was
previously defined by Amin and Arash [phonetic] in selecting random
matching. There's also a pipeage rounding idea by the work of Calinescu
Checkuri Polin Vondrak. And in fact this -- I mean they proved -- they then
prove exactly this negative correlation, but they proposed a randomized
algorithm and this negative correlation can be found through their proofs
and be applied through their proofs.
Recently, after our paper, Checkuri Vondrak and Zenklusen found some
new approach called randomized swap rounding. This approach is purely
combinatorial. And its running time is better than others. In fact, it's more
general. I won't show you how it will generalize it. But however all of these
algorithms gives us negative correlation concentration bounds.
So after that, suppose that we have negative correlation concentration
bound. Let's see what will be our chain of bound. In fact, the probability
that our tree has more than log n over log log n the number of edges of our
tree is more than log n over log log n of the size of the cut, is less than or
equal to the n minus the constant times the size of that cut.
And this is in fact where the log n over log log n comes from. Because this
chains of types of bound are for the right tail. If you have log n over log log
n, you want to have log n over log log n deviation, the probability will go to
n minus constant.
But, yeah, so by this inequality we'll have concentration for one cut. So all
we need is to have concentration for all the 2 to the n cuts. For that we'll
use Karger formula. And Karger in fact proved that in any graph, through
at most n to the 2 K cuts of size at most K times the minimum cut.
Suppose we have a graph, related graph. The number of cuts of size k
times minimum cut is percent 2 K. We'll use this formula and we'll use it
through a union bound argument.
So in fact in our, as I said before, because we had fraction of 2 connected
graph, its minimum cut is roughly one or two. So the number of size of K,
say between K minus 1 and K is at most 2 to the K.
If you multiplied this n2K by this and N minus constant, it will be K if you
considered the cost of K to the minus K. These two will cancel out. It turns
out you can set this constant C to be large enough just by placing some
factor in front of the log n over log log n. So you can prove easily with high
probability our tree will be thin, will be log n over log log n thin in each of
the cuts of size of size L for L 1 and a half. So if you use union bound for
all the cuts, you'll find out that our tree will be log n over log log n 10 with
high probability.
So let's briefly -- let me briefly tell you where this log n over log log n come
from. I mean, is it -- show you a counter-example that we cannot do better
than log n over log log n. Example is simple. Consider complete graph n
vertices and suppose the fraction of each is 1 over n. So in this graph,
even if you -- if you select each edge uniformly and independently at
random, so suppose that we're not looking for proving anything for the left.
We already proved -- we're only looking to prove something for the right
tail. Suppose only we -- suppose that we select the edges uniformly at
random, but turns out this will be something like [indiscernible]. In fact, for
any edge, because the fraction of each edge is 1 over N, the size of this cut
is 1. So the probability that an edge of our tree will be selected will be
incident to this vertex will be 1 over N. So if you think, we'll select n edges
like n minus 1 over n edges and each time there is 1 over n probability that
this edge will be selected to one specific vertex. So it is like a balls and
bins problem where vertices are bins and edges are balls, and there's 1
over n probability that an edge is incident to a specific vertex.
So it's proved that in a balls and bins, n balls and n bins, there's a bin with
high probability that contains log n over log log n balls.
So in fact if we run this -- if you select our edges uniformly and
independently at random, there will exist a vertex with degree log n over
log log n with high probability. In fact, even if you only consider the cost
that separated the vertex from the rest of the graph you cannot do better
than log n over log log n. In fact, there are divorce costs in this example.
So this is a generalization by Checkuri Chenoff and Zenklusen. They
generalize the idea to any matrix. Not just the spanning tree. So they
proved that suppose that we had a fractional solution of matrix polytope,
and we're looking to sample a base from that matrix also by preserving
marginal probabilities. And they prove that we can do that while getting
negative correlation and therefore we have n type chain of bound for any
linear combination of the variables.
So in fact their algorithm works for a number -- for various problems, and
they give, they drive like log n over log nn approximation for number of
problems.
But always you cannot do better than log n over log log n because of the
right tail of general bounds.
But as I said before, the algorithm is purely combinatorial. You can find the
proof. But their proof is much smaller than comparing to us. You can find
it in Vondrak's home page.
Okay. So any questions? Let's conclude. So here first we consider the
problem in directed domain. Then we try to have some simplification. We
convert it to another problem undirected domain. But although the
probability in undirected seems to be much harder, but as you see we have
found good results for that.
So I mean trying to get rid of this direction would be a very good idea. Also
we've shown some interesting relation between ATSP and graph
embedding problems, dual graphs, planar graphs and these sorts of things.
And, finally, we prove that integrality gap ATSP is constant for shortest
path metric defined unbounded genus graphs.
So the interesting point is that the Checkuri Goemans and Karloff
integrality graph encountered example is planar. So in fact they proved
that integrality gap of ATSP on planar graphs is at least two. Therefore, in
fact at least for planar graphs we know that integrality gap is constant. It's
bounded low by 2, above by 22 as we prove. But for nonplanar graphs, if
there are bounded genus still everything is okay. But if they're not
bounded to genus, they're from general metrics is still open whether their
integrality gap is constant or not.
So let me tell you some often problems for future direction. So, first of all,
is it possible to generalize our ideas to find constant trees for general
metrics. Not only for rated bounded genus graphs. Not rated bounded
genus graphs. And, moreover, is it true, is it possible to get P task for
ATSP on planar graphs, like because we have P task for symmetric genus
planar graphs so it is hard to find P task for problems in directed domain.
There are a lot of difficulties.
But it's a nice problem to work on. Moreover, here we proved that this
guarded conjecture any KH connected graph contains a constant over K
thin tree implies constant integrality graph of the ATSP. Also, there is
some conjecture by Jaeger on number of zero flows. Does not tell you the
details, but this is a well-known conjecture number of zero flow. It says like
any 4-H connected graph contains 2 plus 1 over K number of flow. We can
search for it on the Web. But turns out that this conjecture implies some
version of the Jagger conjecture which has not been proved. This implies
that this conjecture seems to be a very hard problem. But anyway these
are the directions we know, but it's open whether any other direction is
true, whether, for example, we can prove Goemans conjecture from the
constant integrality gap of ATSP. If it's true, at least we know that although
this is a thin tree problem is a hard problem but it is something like ATSP.
And we lose nothing by reducing our problem to thin trees.
Also some directions between these two, these others are very interesting.
>>: [indiscernible] approach at the most you want to ->> Shayan Oveis Gharan: Yeah, there exists six numbers of flows for two
edge connected graphs. But the Jaeger conjecture is that if we increase
the connectivity to some very large number can we get 2 plus epsilon for
very small epsilon or not. 2 plus epsilon over 0 flow.
>>: Epsilon?
>> Shayan Oveis Gharan: 2 plus epsilon.
>>: What does epsilon mean?
>> Shayan Oveis Gharan: Pick any epsilon. Suppose you pick any epsilon.
And I should show that.
>>: No [indiscernible] integral.
>> Shayan Oveis Gharan: You can define for fractional graphs, too. By
setting the -- I mean, allowing the fraction of each edge to be between 1 and
1 plus epsilon. You can -- they are similar. This Jaeger conjecture asks if
you can find -- so, for example, if the graph contains a 2 plus epsilon,
number of zero flow, it is very near Eulerian graph.
>>: They're much bigger things, conjecture. So the types of conjecture s
that if there is a poor connected graph, you have at least three ->> Shayan Oveis Gharan: This top conjecture is special case of Jaeger
conjecture. For K equals 1.
>>: But you don't have to do epsilon. There's still a conjecture -- I led you
to the constant with the [indiscernible], can you show with a thousand
connected graph you can find three normal zero flow.
>> Shayan Oveis Gharan: This is still open.
>>: So you don't even have to talk about epsilon or anything.
>> Shayan Oveis Gharan: But at least for planar graphs you can do that or
for bounded genus graphs. If the connectivity is sufficiently large we can
get 2 plus epsilon.
>>: Planar graph.
>> Shayan Oveis Gharan: Our algorithm implies this. Also, there's some
various works, truly only on Jaeger conjecture that proves that Jaeger
conjecture is true for bounded genus graphs. But our algorithm is a bit
more general because it implies -- Goddyn conjecture, proves the Goddyn
conjecture which is more general than here. Okay. Thank you.
[applause]
Download