>> Konstantin Makarychev: Today, Tselil Schramm will tell us about this small set conjecture. Tselil is a grad year student at UC Berkeley and now she is interning at Microsoft research, so if you want to prove or disprove the conjectures it's the right time to do so. Talk to her after the talk about it. >> Tselil Schramm: Thanks for the very optimistic introduction, Konstantin. Hello everyone, thank you for coming to my talk. If at any point you have questions, please ask because I know the crowd is of varying backgrounds. I'll begin with a couple of definitions to level the playing field. I'll begin by defining the expansion of a graph or of a set in a graph. Basically, the expansion of a set in the graph is a measure of the connectivity of the graph. It's the number of edges leaving the set divided by the total number of edges inside the set in a d-regular graph. The notation of the definition is not that important, but please remember this notation because I'm going to be using it throughout. Here you can see an example, this set circled in red is relatively non-extending and the set circled in blue is relatively extending set. It's interesting to ask what is the minimum expanding set inside of a graph, so that is the parameter is the graph extension. You can ask okay, what is the minimum expanding set inside a graph. In this case, again, it's the set circled in red. You can kind of generalize this to ask what if I only want to find non-extending sets of a specific side. Say I want to limit sets of size 1/6 the total number of non-linear nodes in the graph, what is the minimum extending sets of Si’s. Here we can see in this example that although there might be a non-extending set inside the graph, it's possible that all sets of size 1/6 don't expand, or do expand a lot. This is an interesting extension of the problem. You can design the problem, uniform sparsest cut, which is given a graph, find the minimum expanding set. This is NP hard, but we have several approximation algorithms for this. One of them is Cheeger's algorithm and this algorithm is going to be important in the talk. How familiar, would anyone like a refresher of Cheeger's algorithm? Okay. I'll give you a refresher of Cheeger's algorithm. Cheeger's algorithm is basically an algorithm with the following guarantee. Say you give me a normalized adjacency matrix and a vertex and I will return a set that has expansion at most this quantity relating to the vertex and the adjacency matrix. This is like the Rayleigh quotient of the vector. Normalize meaning it's stochastic. >> You said a vertex but you meant a vector. >> Tselil Schramm: I said a vertex but I meant a vector, yes. And the way that the algorithm works is you take the vector and you order the vertices for the entries of the vector according to their size from least to greatest and then you output the set that maximizes this quantity. You look at the first vertex and if this minimizes then you take the first two vertices, third vertices and you output the set that minimizes this. It works by choosing a probability distribution over the index at which you break the set and you bound the expectation and that gives the desired approximation ratio. This is something that I will use later in the talk as well. There is also other approximations, for example, [indiscernible] and they are better in some regimes, but I won't discuss them today. Now there's the problem of approximating the sets of size at most delta, so this goes back to the expansion profile and the question is what can we do with this. There is some algorithms known, one of them due to someone sitting in this room, in part, or two people sitting in this room, and the point is we don't have quite the Cheeger's guarantee here. In the case of Ergamanister and Totali [phonetic] wear off by this factor which is okay if the delta is constant but it gets bad if delta is really small [indiscernible] exponential and here for the [indiscernible] Trevisan, there is a trade-off between how good the expansion approximation is and the size of the set. If you are willing to output a larger set you can get a pretty good approximation of the expansion. Why care about small set the expansion? I could give you reasons and in an industry like Facebook and I want to test a product in a small group, but actually, there are better heuristics than that, like, choose New Zealand, you know? Really, the reason we care is because of theory. Small set expansion is kind of embroiled in this important question in theoretical computer science about hardness of approximation. So you have probably heard of the Unique Games conjecture. If you have heard of it but don't remember it, or if you have never heard of it, don't worry. The point is, basically, that Unique Games conjecture asserts that a problem called Unique Games is NP hard to approximate within a certain ratio. If Unique Games is NP hard, we have reductions from it to vertex cover and to a bunch of constraint satisfaction problems, think like 3SAT and MAX-Cut, but this is really it, like a huge family of problems and we know that if Unique Games conjecture is true we have all of these optimal approximation algorithms for vertex cover and all the constraint satisfaction problems. Small set expansion is one of the few problems we know that reduces to Unique Games, so if we can show the small set expansion is hard, that would imply that all of these problems are hard to approximate and we have optimal bounds, and vice versa there's kind of a weaker reduction from Unique Games to small set expansion that has some assumption on the Unique Games instance that you have, namely, the small sets expand more and so showing that small set extension is easy to approximate within a certain ratio could have implications, I mean, it has implications that some version of Unique Games is not difficult throwing that on to the full picture. Now in order to understand the rest of my talk, you will have to understand this notion of gap problems. A gap problem is the following. Say that I am handing you a graph and I promise you that either there exists a set of size delta V that has expansion less than some parameter C, or I say all sets of size delta V has expansion of at least S. This is known as a gap CS problem. I model it with small set expansion because that's the one that we will need for today. Denote this problem in the following way, an algorithm that decides between a CS gap problem which tells you which scenario you're in gives you an S over C approximation algorithm. Does anyone have questions? >>: Can you go back to the previous slide? >> Tselil Schramm: Yeah. >>: [indiscernible] bigger then C? >> Tselil Schramm: Yes. In this case S is bigger than C. Yeah. It's a minimization problem. If you can distinguish between these two cases, you can say whether you can get an approximation for this vector. Gap amplification is when you take a gap problem and you reduce it to a different gap problem where the gap is wider. In essence, you are taking an approximation and saying it's hard to approximate within alpha and then saying okay. This reduces to approximating within beta where beta is an easier thing to approximate. Here you can see in the example a small set expansion. This would be like taking a set that kind of expands, but not that much and then read the widening the gap so a set that expands completely and a set that expands almost not at all. Gap amplification is a useful tool in improving hardness of approximation results. If you want to show that a problem is hard to approximate with a ratio of beta, what you can do is you can take a problem that is hard to approximate with a ratio alpha, say you know of some problem that is hard to approximate using alpha, and then you can do gap amplification until the ratio becomes beta. That implies that the problem that was hard to approximate within alpha reduces to the beta problem and so the beta approximation is also hard. Another reason you would do gap amplification is if you want to show that a problem is actually easy to approximate within some ratio. You can take a problem, the problem of approximating within ratio alpha and reduce this to approximating within the ratio beta via gap amplification and if you have a beta approximation algorithm, you have just disproved it's hard to approximate the problem within alpha. In our case, we would be happier with a result like this because we don't know a problem that is hard to reduce to Unique Games, but we could hope, or two small set expansion, but we could hope to take small set expansion with the ratio that we want, reduce it to a ratio that we already know because of the approximation algorithms and then show that it's easy. For the first reason it's important in these two settings. Label Cover is very important problem and also in Unique Games and gap amplification results are known for both of these problems. The way that you prove gap amplification results, sorry. The way that you prove gap amplification results for these is you do something like you take a graph times a power of the underlying instance. I will explain that a little more later. Our method is not that and as a nice benefit we don't have any growth of the vertex set and stuff like that, so it's kind of a novel gap amplification. Our result is gap amplification results for the small set expansion problem matching the known Unique Games gap amplification results. If you want the technical theorem, this is it. Basically, if we have a function that grows slower than root epsilon, we can reduce between epsilon and that function to any constant and beta. This is actually the main idea. Let me go back to this graph tensor power. Normally, to do a reduction, to do a gap amplification in Label Cover or in Unique Games, you take an underlying instance, which is a graph, and then you take its tensor power. Taking the tensor power essentially amounts to taking the Cartesian product of the vertex set and then adding edges between two vertices in the power graph if both of the endpoints have edges in the underlying graph as well. Here, you can see the tensor power of this edge, which is a simple graph. You would think since Unique Games and a small set expansion are kind of related, probably, that you could try to apply the same sort of idea in order to achieve gap amplification for small set expansion because this is the way that gap amplification is done for Unique Games. There's an obstacle when you are doing this. The obstacle is the following. Say I started with a gap that falls into the no case, falls into the case that every set with size delta has expansion that is too large. Say that there is some set with size delta that has expansion that is in the yes case, that has bounded expansion. Now I take the tensor product of the graph and I look at the product of the set with itself. This set will have size delta in the new graph, in the tensor power graph, but on the other hand the expansion doesn't change very much, so if it had expansion epsilon here, it has expansion about epsilon here and so that means that we can fail to reject graphs that we should have rejected. We can fail, in this case we can say totally this has a set that is the size at most delta that expands by at most whatever, and on the other hand here, we don't. This is a problem. What makes sense for gap amplification in small set expansion? Random walks seem like kind of a good idea. Sort of by definition of expansion, if a set is not expanding, then the probability maps of a random walk stays trapped inside the set for a very long time like here. On the other hand, if a set is expanding then that means that a random walk started from that set will spread really fast across the graph. We can try to think of taking the graph power as in the matrix power to do the reduction instead of taking the graph tensor power, which corresponds to the random walk. That's what we do. There's the reduction. We start with given a d-regular graph with the lazy random walk matrix m. Our graph, our output graph is going to be the TTH power of this matrix. Essentially, what that means is that we take the original vertex set and on it we put edges with the probability of starting at vertex A and ending at vertex B. Here is the reduction parameters are not very important. The main point is that if we choose the parameters carefully we can amplify the gap. Any questions at this point? >>: So do you have [indiscernible] for the matrix m? >> Tselil Schramm: Yeah. M is the matrix with self loop already added. >>: [indiscernible] >> Tselil Schramm: Exactly. If you start with 0 it's on and on. >>: There must be some limit as to how much you can amplify the gap. >> Tselil Schramm: There is. >>: Where is that limit? >> Tselil Schramm: We can get this gap amplification, but for some parameters it's meaningless. For example, when this becomes 1. >>: [indiscernible] corresponds to which [indiscernible] >> Tselil Schramm: I see. This is sets that expand by at most A and this is all sets that expand by at least B and it maps these sets that expand by at most 1/2ta or sets that expand by at least tb squared. >>: [indiscernible] >> Tselil Schramm: Right. Delta maps in new values and stuff depending on the -- yeah. So maybe this is actually a better way to think of the limiting. If you choose t too large then the set becomes really, really small. Other questions? >>: [indiscernible] specific [indiscernible] >> Tselil Schramm: Yes. Well, it's not just a specific [indiscernible]. It's also, it's a lower band. >>: I mean it doesn't [indiscernible] >> Tselil Schramm: No, no. It doesn't do anything else. Yes. >>: Is there some sort of spectral estimate on this? >> Tselil Schramm: Yeah, completely. It's 100 percent spectral. I will get to that. Okay. The proof proceeds as follows. First, we have to show that if you have a set that did not expand very much in the base graph, when you take the power of the random walk matrix it won't expand that much in the final graph. And then we also have to show that if the set expanded to begin with then after t steps of the lazy random walk it will expand even more, so we get upper and lower bounds for the expansion of the set after applying the graph power. This proof proceeds in the following way. First, what we do is we say, we perceive the contradiction and we say if the set did not expand enough in the final graph and after t steps of the lazy random walk then there must have been some step, at least one step in which it didn't expand enough along the way. Then what we want to do is we want to take the vector corresponding to that set and apply Cheeger's algorithm to it, but there's some technical difficulties because the support of the vector might be too big and there we use a method similar to Heer Bruckinster [phonetic] to reduce the support of the vector and get the set. For the first part in which we say that non-expanding sets don't expand too much, we can apply this bound which is known and Oveis Gharan and Trevisan have a particularly nice exposition of it when they are trying to prove something a little bit more sophisticated. Maybe I'll tell you guys a little bit about how this goes in case you don't know, but in the end, the point is that you can bound the expansion using a power of the expansion. What you essentially do is you turn this into a problem about positive semi-definite matrices. The lazy random walk matrix is positive semi-definite and for any positive semi-definite matrix p and Norm 1 vector x you can show that this quadratic form is larger than the product of these two quadratic forms, so essentially, if you look at the expansion after t steps of the random walk it should be greater than the, or it should be less than the expansion after one set than the expansion after t -1 sets. Then what you can do is you can turn this into a problem about eigenvalues. You can view this, which is actually this quadratic form, as an expectation over the lambda i for the k and then you can apply Chevy Chez Simon [phonetic] equality because the matrix is positive; and all of the eigenvalues are positive, Not negative. >>: You mean [indiscernible]? >> Tselil Schramm: No. Maybe you could apply [indiscernible] inequality? >>: If you wanted to use convexity of lambda i to the k [indiscernible] >> Tselil Schramm: Yeah. You can do that. It turns out to be the same. This is just an inequality that says if you have ordered a and b and you take a product like this it's positive, but I think you can use convexity. >>: If you don't have to, the [indiscernible] should not be important. >> Tselil Schramm: It's not. The positivity isn't worth it. >>: Is it? >> Tselil Schramm: Yeah. >>: Is the positivity important? Involved in all of the [indiscernible] >>: The positivity is important to this proof, but not to the result. >>: [indiscernible] if something is odd or even [indiscernible] >>: [indiscernible] this proof does depend on positivity but the basic inequality there that… >>: Is just reversibility. Just need reversibility. >>: Yes. >> Tselil Schramm: For this. >>: Forward [indiscernible] odd or even? >> Tselil Schramm: For this >>: X prime p to the tx is bigger than x prime ex to the power of 2 [indiscernible] >>: [indiscernible] works and then if you have [indiscernible] eigenvalues, but [indiscernible] it's a special case of [indiscernible] conjecture, which is not known, but that special case known from the [indiscernible] 40, 50. I once sent to Cheyenne like a whole list of reference and it keeps being rediscovered many times. [indiscernible] told me that we discovered inequality, wrote it up and then found it had, you know, ten earlier references. >> Tselil Schramm: Thanks. Maybe you can arrange it? Okay. In any case, this suffices for the upper bound on the expansion of a set in the target graph given the original expansion. In order to show that sets that expand in the original graph expand more in the target graph, what we do is we say the following. If by contradiction a set does not expand enough after t sets of the random walk, then we look at the probability mass of the set each of the different walks and we say okay. There must have been at least one step of the walk in which there wasn't enough expansion and we try, well, we don't try to identify that set but just the existence of that set is enough. Let's suppose by contradiction that there is some set S with bounded size such that at most the expansion is bounded by beta in the target graph. Let 1 sub s be the indicator vector for S and what we are going to do is we're going to try to extract from S a set that did not expand enough in the original graph and get from that a contradiction. We know that 1 minus beta is less than this quotient just by definition of expansion because one minus this is the expansion of the set S. We can translate that into a statement about norms of vectors. Now, what we do is we say if this product of the ratios, if we look at the product of the norm of V1 over V0 where Vi is the distribution over probability mass of vectors at the ith set, then if we look at the quotients of each of the sets and we look at the products what we end up with is the quotient from the beginning to the end. This is actually the quantity that we dealt with before which we have a bound on because of our assumption that the set did not expand enough. Then we know that at least one of these terms has to be bounded by 2 over TTH root. Now what we hope to do is we have this vector and we have a bound on the Rayleigh quotient and we know that Cheeger's algorithm can give us a set with expansion at most root that. The problem is that this vector might have really large support. This vector might have, you know, all n vertices have known 0 support and then Cheeger's algorithm might output a set that has size n over 2 and that’s something we want. We want to bound the number of vertices in the set. We can trouble shoot this issue in the following way. We take our vector and we pick some threshold theta and for every entry in the vector less than theta we just set it to 0 and four entries greater than or equal to theta we just subtract theta. Using some bounds on theta and the two norm of v which we get via the expansion, we can conclude that the Rayleigh quotient, this, which is using in Cheeger's algorithm to get the bound is bounded by the Rayleigh quotient of the vector before thresholding. In some sense reducing or thresholding the vector did not affect its Rayleigh quotient too much. Okay. Then we can just use our previous bound. We have an upper bound now on this Rayleigh quotient. From there we can apply Cheeger's algorithm and we get a bound on the expansion of our new set and we also have a bound on the size of the new set using Markov's inequality or what have you. And then essentially by selection of parameters we can get the result. We have the following result, small set expansion with sets less than or equal to A or all sets expanding by at least B if size delta reduces to an instance of SSE where for all sets in size of 4 delta over tb squared the sets of size half, there exists a set of size half ta or offsets are greater than tb squared. It's clearly polynomial time because all we do is apply t stuff to the random walk. And the way that we do this is first we showed that non-expanding sets don't expand too much and then we showed that expanding sets ended up expanding a lot by contradiction and basically what we did is we bound the expansion of the jth step of the random walk starting from S. We controlled the support size of the vector by thresholding and then we applied Cheeger's algorithm, so pretty simple. In order to get our kind of more human readable result, you can just set the parameters like this and you get this result which is analogous to the unique gains. It's the best-known gap application for Unique Games, but, you know, it's not like I think this isn't the main message. I think the main idea is that you can do gap amplification using these random walks. >>: In this human readable way, what is eta? >> Tselil Schramm: It is [laughter] it is any constant you want, so that for any eta that you want here, there exists some epsilon that you can reduce from. Epsilon the way that we… >>: I know what it says about eta, but I prefer the non-human readable. [laughter] >> Tselil Schramm: Me too. >>: [indiscernible] >> Tselil Schramm: I think the non-human readable is kind of like the victors write the history books kind of thing, yeah. >>: [indiscernible] [laughter] >> Tselil Schramm: I'd rather not say. Open problems from this, I think there are a lot of open problems, but a lot of them seem very difficult. The first open problem which was the original open problem for this work was can you get to more amplified gap amplification for small set expansion. And also it would be interesting to get a tensor powerbase gap amplification for small set expansion or random walk based gap amplification for unique gains. So a little bit more [indiscernible] to these problems. If we can amplify the gap from small set expansion with all sets, or with the existing set expanding by at most epsilon or all sets expanding by at least 1 minus epsilon to some gap version of SSE that we know is easy to approximate, then this would refute the hardness of more constrained version of Unique Games and that would be a very interesting result because it sort of sheds light on the Unique Games conjecture and maybe on the difference between small set expansion and Unique Games. We did not achieve this gap amplification. Another interesting thing to do would be to get a tensor based gap amplification for Unique Games or a random walk based, sorry a tensor based gap amplification for small set expansion or a random walk based gap amplification for Unique Games. The reason being that I think a lot of people think of these two problems as relatively equivalent, so it's strange that the two problems aren't amenable to the same kind of gap amplification. In applying tensor based gap amplification to small set expansion, one needs to overcome this issue with the completeness, the problem that, you know, you can get new non-expanding sets just by applying tensor products and I have thought a little bit about this but I haven't been able to resolve it. On the converse side, maybe it would be possible to get a random walk based reduction for Unique Games which I have also thought about a little bit. This would also be nice because the random walk based reduction does not amplify the size of the vertex set and so this could lead to better gap amplification because the size of the graph won't be a limitation as it is usually. But the obstacle that one runs into here is say that your base graph is disconnected, has several components and say that one of the components has a really, really good assignment and the other component has like a so-so assignment. Then when you, so what it would be natural to do would be to take the label extended graph and to apply the random walk to it until the problem gets harder, I mean easier. But the problem is for the perfect completeness component, when you apply the random walk it doesn't get anywhere. If you have the component of size a third which is perfect, that's the lower bound for how much you can amplify the gap. Whereas, in your other component things get worse and worse, it doesn't make much of a difference. Overcoming this problem also would be interesting. I thought about this a little. Maybe one could put some kind of expander on the whole instance, but that didn't seem to work out. If you have any ideas about this I would be happy to discuss them. Yes. Thank you. >> Konstantin Makarychev: Questions? >>: I guess one difference to the amplification for Unique Games is also so that, maybe it's just always said that usually you want to reduce the gap, you want to amplify the gap so that it is close to one. You said that you [indiscernible] to half. >> Tselil Schramm: Sorry. I think for the Unique Games it should have been one minus the corresponding. >>: I mean, for small set expansion you could hope for that you could amplify the gap to 1. >> Tselil Schramm: To 1. >>: Did you think about, [indiscernible] one [indiscernible] of that is it gets Cheeger's inequality. You usually use it as sort of is concentrates on expansion close to 0. But they are also variance is an inequality that given the whole range of expansions are close to 1 and close to 0. Just wondering if you try to increase the gap from half to close to 1? >> Tselil Schramm: No. We didn't. We thought about it a little bit, but we didn't. >>: So when you are raising random walks, are there other, is there any reason to suspect that some kind of [indiscernible] random walk or other kind of random walk with the other, that you could employ other [indiscernible] like inequalities that come with those random walks and give you some other issue? >>: Other random walks? >>: [indiscernible] random walk like you walk [indiscernible] over this amount of biasing by a degree or something. >> Tselil Schramm: I mean, all of the results that I presented today were for d-regular graphs, but actually we also have, I mean, we have analogous results for, you know, any degree distribution and, again, we do a wave during the walk, actually. >> Konstantin Makarychev: More questions? Let's thank the speaker again. [applause]