Maximizing Polynomials Subject to Assignment Constraints ∗ Konstantin Makarychev Microsoft Research Maxim Sviridenko University of Warwick Abstract We study the q-adic assignment problem. We first give an O(n(q−1)/2 )-approximation algorithm for the Koopmans–Beckman version of the problem improving upon the result of Barvinok. Then, we introduce a new family of instances satisfying “tensor triangle inequalities” and give a constant factor approximation algorithm for them. We show that many classical optimization problems can be modeled by q-adic assignment problems from this family. Finally, we give several integrality gap examples for the natural LP relaxations of the problem. 1 Introduction In 1963, Lawler [13] proposed the following problem, which he called the q-adic assignment problem: We are given a set [n] = {1, . . . , n} of indices and a 2q-dimensional array cu,v for u, v ∈ [n]q . The goal is to find a permutation π : [n] → [n] optimizing (maximizing or minimizing) the expression X cu,π(u) , (1) u∈[n]q where π(u) = (π(u1 ), . . . , π(uq )) for the vector u = (u1 , . . . , uq ). The maximization and minimization problems are called the maximum and minimum q-adic assignment problems respectively. This is a natural generalization of the quadratic assignment problem (QAP) originally defined by Koopmans and Beckman [12] in 1957 for which q = 2 and cu,v = wu dv for all u, v ∈ [n]2 . Thus, we call the problem with decomposable coefficients cu,v = wu dv the Koopmans–Beckman version of the q-adic assignment problem throughout the paper. Similarly, to the quadratic assignment problem, the general q-adic assignment problem has many practical applications and can be used to model variety of optimization problems. For example, Winter and Zimmerman [19] used the cubic assignment problem for scheduling. Burkard, Cela and Klinz [7] suggested applications of the 4-adic (or quartic) assignment problem in VLSI synthesis. The Koopmans-Beckman version of the problem can also be viewed as a variant of the hypergraph isomorphism problem studied by Babai et al. [3] where we would like to maximize the product of weights of hyperedges that are mapped into each other. (However, since techniques used in hypergraph or graph isomorphism algorithms heavily use the fact that all hyperedges must be mapped these techniques cannot be used for the q-adic assignment problem.) But despite enormous amount of applications, the problem remains highly intractable both from theoretical [15] and practical viewpoints even when q = 2. For larger q, e.g. for q = 4, the problem cannot be solved exactly in practice even if n = 14 (see [6]). ∗ an extended abstract of this paper appeared in Proceedings of ICALP2011 1 1.1 Overview of the Results In this paper, we study the maximization version of the problem with nonnegative coefficients. There are no known approximation algorithms that have a proven performance guarantee for the general case. The Koopmans–Beckman version of the maximum q-adic assignment problem was considered by Barvinok [4] who designed an approximation algorithm with performance guarantee εnq/2 and running time nO(q/ε) for any ε > 0. The best known approximation algorithm √ with performance guarantee O( n) for the special case of q = 2, i.e. the Maximum QAP, was given in [17] and [15]. In this paper, we present a new Oq (n(q−1)/2 )-approximation algorithm for the Koopmans–Beckman version of the maximum q-adic assignment problem thus improving upon the result of Barvinok [4] and matching the performance guarantee of [15] for the Maximum QAP. The Oq (·) notation hides terms dependent only on q. Note, that we cannot hope to get a poly-logarithmic approximation ratio, since such possibility was ruled out (under very plausible complexity assumptions) even for the maximum quadratic assignment problem in [15] (using the same method as in [15], one can show that under slightly stronger assumptions (see [5]), there exists δ > 0 and q such that the maximum q-adic assignment problem is NP-hard to approximate within a factor of nδ ). Thus, one of our main goals is to identify special families of instances that can be solved with a much better performance guarantee. Arkin, Hassin and Sviridenko [2] and, then Nagarajan and Sviridenko [17] showed how to get a constant approximation ratio in the Koopmans–Beckman version of the maximum quadratic assignment problem, if the coefficients dv satisfy the triangle inequality. We give a very general analog of this result for the general (i.e., not the Koopmans–Beckman version) maximum q-adic assignment problem. We further describe combinatorial optimization problems that can be expressed as special cases of this problem. Finally, we give Oq (n(q−1) )-approximation algorithm for the general case of the maximum q-adic assignment problem (improving on the trivial O(nq ) approximation algorithm). (q−1)/2 q−1 n for the We give an integrality gap examples of Ωq ln n for the general case, and Ωq n √ln n Koopmans–Beckman case of the maximum q-adic assignment problem. 1.2 Tensor Triangle Inequality Let K = {k1 , . . . , kq−1 } be a set of (q − 1) distinct numbers in [n], i.e. ka 6= kb for all a 6= b. For each vector v ∈ [n]q and each coordinate s ∈ [q], let Vs (v, K) be the set of (q − 1)! vectors obtained from v by replacing all coordinates except the coordinate s with indices k1 , . . . , kq−1 in an arbitrary order. For example for q = 3, s = 1, vector v = (v1 , v2 , v3 ) and K = {k1 , k2 } ⊂ [n] we get V1 (v, K) = {(v1 , k1 , k2 ), (v1 , k2 , k1 )}. Definition 1.1. We say that an instance of the maximum q-adic assignment problem satisfies the tensor triangle inequality if for every K = {k1 , . . . , kq−1 } and u, v ∈ [n]q , cu,v ≤ q X s=1 X v′ ∈V cuv′ . (2) s (v,K) Note that the total number of terms on the right hand side of (2) is q! and the total number of n different sets K is q−1 . We now discuss various special cases of the tensor triangle inequality: 1. For q = 2 the 4-dimensional tensor (c(i,j),(p,q) ) for i, j, p, q ∈ [n] satisfies the tensor triangle inequality if for any r ∈ [n] the following inequality holds c(i,j),(p,q) ≤ c(i,j),(p,r) + c(i,j),(r,q) . 2 (3) In the Koopmans–Beckman version of the problem, c(i,j),(p,q) = wij dpq and the condition (3) just becomes the triangle inequality dpq ≤ dpr + drq . If in addition dpq is symmetric, i.e. dpq = dqp we obtain the maximum quadratic assignment problem with the triangle inequality for which constant factor approximation algorithms are known [2, 17]. If wij is the weight of the edge (i, j) in a directed graph, dpq = 1 if p < q and dpq = 0 if p ≥ q, then the tensor triangle inequality is also satisfied, and the problem is exactly equivalent to the Maximum Acyclic Subgraph Problem. More generally, the condition (3) models the Koopmans–Beckman variant of the maximum quadratic assignment problem in which matrix dpq is not symmetric but satisfies the triangle inequality. There was no constant factor algorithm known for this problem before and the techniques from [2, 17] do not seem to generalize to this case. 2. In the Betweenness Problem, we are given the set T of triples (i, j, k), we would like to find a permutation π and maximize the number of triples such that either π(i) < π(j) < π(k) or π(i) > π(j) > π(k). This problem has applications in computational biology and was studied in [9, 14]. This problem is modeled by the q-adic assignment problem in the Koopmans– Beckman form with q = 3. We define c(i,j,k),(p,q,r) = w(i,j,k)d(p,q,r) such that w(i,j,k) = 1 for each (i, j, k) ∈ T , w(i,j,k) = 0, for each (i, j, k) 6∈ T and d(p,q,r) = 1 if p < q < r or p > q > r and d(p,q,r) = 0, otherwise. The condition (3) is satisfied since if dpqr = 1 then at least one of the term one right hand side of (3) must be one for any choice of k1 , k2 . We note however that the random assignment gives a 3 approximation for the Betweenness Problem, our algorithm gives a worse constant. 3. Finally, the condition (3) models the generalization of the Maximum Acyclic Subgraph Problem considered by Guruswami et al. [10]. In this problem for each customer we are given a list of preferences in the form i1 > i2 > · · · > iq the goal is to find a permutation π that maximizes the total number of satisfied customers, where a customer is satisfied if his preferences are satisfied, i.e. π(i1 ) > π(i2 ) > · · · > π(iq ). Guruswami et al. [10] show that assuming the Unique Games Conjecture there is no approximation algorithm for this problem with performance guarantee better than q! (note, that a bound of q! is trivial, because a random assignment gives a q! approximation). Since this problem can be modeled using condition (3) it shows that our e2 q! approximation for the maximum q-adic assignment problem is optimal up to a factor e2 . 2 Linear Programming Relaxation For any vector u ∈ [n]q , let us be the s-th coordinate of vector u. The maximum q-adic assignment problem can be formulated as the problem of maximizing the value of degree q multilinear (or tetrahedral) polynomial subject to assignment constraints: max X u,v∈[n] c q u,v q Y xus vs , (4) s=1 X i∈[n] 3 xij = 1, j ∈ [n], (5) X j∈[n] xij = 1, i ∈ [n], xij ∈ {0, 1}. (6) (7) We linearize the objective function (4) and relax the integrality constraint (7). Let Sq be the set of all permutations on q elements. For each u ∈ [n]q and σ ∈ Sq , let uσ denote the vectorv 1 2 3 with coordinates vs = uσ(s) i.e., σ permutes the coordinates of the vector u (e.g., if σ = , 2 1 3 u = (5, 17, 10), then uσ = (17, 5, 10)). In our linear programming relaxation we keep assignment variables xij for i, j ∈ [n] and we define new variables yu,v for u, v ∈ [n]q that have the value Q q s=1 xus vs if the variables xij are integral. Let R = {u ∈ [n]q : ui 6= uj , if i 6= j}. The fact that ourQobjective function (4) is a multilinear polynomial and assignment constraints (5)-(6) imply that qs=1 xus vs = 0 if u ∈ [n]q \ R or v ∈ [n]q \ R. Therefore, we can restrict our attention to the q-tuples from the set R. We now formally define our linear programming relaxation X max cu,v yu,v , (8) u,v∈R X yu,v ≤ xivs , v ∈ R, s ∈ [q], i ∈ [n], (9) u∈R:us =i X yu,v ≤ xus j , u ∈ R, s ∈ [q], j ∈ [n], (10) v∈R:vs =j yu,v = yuσ ,vσ , X xij = 1, i∈[n] X xij = 1, j∈[n] yu,v ≥ 0, xij ≥ 0, u, v ∈ R, σ ∈ Sq , (11) j ∈ [n], (12) i ∈ [n], (13) u, v ∈ R, (14) i, j ∈ [n]. (15) The integral solution of the maximum q-adic assignment problem P valid forQan Pconstraints (9) are q since u∈R:us =i yu,v = u∈R:us =i s′ =1 xus′ vs′ and we can iteratively apply the equations (12) for each coordinate of vector u except coordinate s. Analogously, we can show that Q constraints (10) are valid. The validity of constraints (11) follows from the fact that the product qs=1 xus vs does not depend on the order of factors. In Section 6, we show that our linear programming relaxation has integrality gap Ω(nq−1 / log n) for the general case of the maximum q-adic assignment problem. However, in the Koopmans– Beckman form, i.e. when cu,v = wu dv for all u, v ∈ [n]q the integrality gap is O(n(q−1)/2 √). We present an approximation algorithm in Section 5 and an almost matching Ω(n(q−1)/2 / log n)integrality gap example is presented in Section 7 of the paper. 3 General Case In this section, we show how to get Oq (nq−1 ) approximation in the general case. Note, that to get Oq (nq ) approximation we could simply let π : [n] → [n] to be a random permutation sampled uniformly from the set of all permutations. To get an improvement, we map some of the indices uniformly at random, and some according to the probability distribution given by the LP. We do it in such a way that for many u ∈ R, one and only one index i ∈ u is mapped according to the 4 LP and the rest (q − 1) indices in u are mapped uniformly at random. We analyze this rounding scheme term by term. We compare the expected profit we get from the term cu,v with the profit of the LP solution (i.e., cu,v yu,v ). We show that we lose only a constant factor by mapping the index i, and we lose a factor of n−(q−1) by mapping the rest of indices. We give the details of this rounding scheme in the following lemma. Lemma 3.1. Let (x∗ , y ∗ ) be an optimal solution of the linear program (8)–(15). There exists a randomized polynomial time algorithm that finds an integral solution to the linear program (8)–(15) with expected value at least 2(q−1) ! −1 X q 1 − 1q X n x∗us vs . cu,v q! q−1 s=1 u,v∈R Proof. The matrix X ∗ = (x∗ij ) for i, j ∈ [n] is doubly stochastic since it satisfies the constraints (12) and (13). Therefore, it can be represented as a convex combination of permutation matrices, i.e. X X∗ = λτ Π(τ ) (16) τ ∈[N ] PN where N ≤ n2 , τ =1 λτ = 1, λτ ≥ 0 for all τ ∈ [N ] and each Π(τ ) is a permutation matrix. Our randomized algorithm chooses one permutation matrix Π(τ ) with probability distribution given by coefficients λτ . We define the first random permutation ϕ(i) = j if Πij (τ ) = 1. Then, by (16), Pr[ϕ(i) = j] = x∗ij for any i, j ∈ [n]. We define the second random permutation ψ : [n] → [n] to be a random permutation on [n] chosen uniformly among all permutations. We now combine the mappings ϕ and ψ. Choose the set X ⊆ [n] at random such that each index i ∈ [n] belongs to X independently with probability 1/q. We define the mapping π as follows: if i ∈ X, then π(i) = ϕ(i); if i ∈ / X and ψ(i) ∈ / ϕ(X), then π(i) = ψ(i). All other elements (i ∈ /X and ψ(i) ∈ ϕ(X)) are mapped one-to-one to elements in [n] in an arbitrary way so that π is a permutation. We estimate probability, that π(u) = v for u, v ∈ R. For every s ∈ [q], let As (u, v) be the following event ^ As (u, v) = {ϕ(us ) = vs } {ψ(us′ ) = vs′ for all s′ ∈ [q], s′ 6= s}. Since the permutations ϕ and ψ are defined independently, we have −1 n ∗ . Pr[As (u, v)] = xus ,vs · (q − 1)! q−1 Let Bs (u, v) be the event Bs (u, v) = {us ∈ X} ^ / ϕ(X) for all s′ ∈ [q], s′ 6= s}. / X and vs′ ∈ {us′ ∈ Since each i ∈ [n] is chosen to the set X independently with probability 1/q and since | ∪s′ 6=s {us′ , ϕ−1 (vs′ )}| ≤ 2(q − 1) (always, and thus independently of As (u, v)) we have 1 Pr [Bs (u, v) | As (u, v)] ≥ q 5 1 1− q 2(q−1) . For every s ∈ [q], events As (u, v) and Bs (u, v) together imply the event {π(u) = v}, also events Bs (u, v) are disjoint for different s ∈ [q], therefore −1 q q X X x∗us ,vs n 1 1 2(q−1) Pr[As (u, v) ∧ Bs (u, v)] ≥ . Pr[π(u) = v] ≥ × 1− (q − 1)! q − 1 q q s=1 s=1 We are now ready to estimate the expected value of the solution. # " X X = cu,v Pr[π(u) = v] cu,π(u) E u∈R u,v∈R −1 q X x∗us ,vs n 1 2(q−1) 1 ≥ cu,v × 1− × (q − 1)! q − 1 q q s=1 u,v∈R 2(q−1) ! −1 X q 1 − 1q X n ∗ xus ,vs . cu,v = q! q−1 X u,v∈R s=1 The main result of this section, Theorem 3.2, easily follows from Lemma 3.1. Theorem 3.2. Let LP ∗ be the optimal value of the linear program (8)–(15). There exists a randomized approximation algorithm for the maximum q-adic assignment problem (1) that finds a solution n with expected value at least LP ∗ /(e2 (q − 1)! q−1 ), i.e. the approximation ratio of the algorithm is n e2 (q − 1)! q−1 = Oq (nq−1 ). Proof. The algorithm of Lemma 3.1 finds a solution with expected value 2(q−1) ! −1 X q 1 − 1q X n ∗ xus vs ≥ cu,v q! q−1 s=1 u,v∈R 2(q−1) −1 X −1 1 − 1q n n e−2 ∗ cu,v · q · yu,v ≥ LP ∗ . q! (q − 1)! q − 1 q−1 u,v∈R 4 Problems Satisfying the Tensor Triangle Inequality The following lemma provides us with an upper bound on the value of the linear programming relaxation for problems satisfying the tensor triangle inequality. Lemma 4.1. Let (x∗ , y ∗ ) be an optimal solution of the linear programming relaxation (8)–(15). If the input instance of the maximum q-adic assignment problem satisfies the tensor triangle inequality, i.e. the inequality (2), then for the optimal value LP ∗ of the linear programming relaxation (8)–(15) we have ! −1 X q X n x∗us vs cu,v LP ∗ ≤ q−1 s=1 u,v∈R 6 P ∗ and Proof. We apply the inequality (2) to each term in the expression LP ∗ = u,v∈[n]q cu,v yu,v average over different choices of the set K. We obtain q X X X X n −1 ∗ cuv′ yu,v LP ∗ ≤ q−1 u,v∈R K:|K|=q−1 s=1 v′ ∈Vs (v,K) −1 X q X X n ∗ = cu,v′ yu,v q−1 s=1 v∈R:vs =vs′ u,v′ ∈R ! −1 X q X n ∗ xus ,vs , cu,v ≤ q−1 s=1 u,v∈R where the last inequality follows from P constraints (10). The first equality is derived by considering ∗ generate the term c ′ y ∗ after applying the which terms in the expression LP ∗ = u,v∈R cu,v yu,v u,v u,v tensor triangle inequality. Combining Lemmas 4.1 and 3.1 we obtain the following theorem. Theorem 4.2. There exists a randomized approximation algorithm with approximation ratio e2 q! for the maximum q-adic assignment problem satisfying the generalized triangle inequality (2). 5 Koopmans–Beckman Case: The Problem with Decomposable Coefficients We now give an algorithm for the Koopmans–Beckman variant of the problem (an extension of our earlier work [15]). In this case, all weights cu,v have the form cu,v = wu dv . It is convenient to think that vectors u, v ∈ R are hyperedges and wu and dv are weights of these hyperedges. We write i ∈ u, if for some s, i = us . If u′ ∈ [n]q and u′′ ∈ [n]q do not have common vertices i ∈ [n], then we say that u′ and u′′ are disjoint. Theorem 5.1. There exists a polynomial-time randomized approximation algorithm for the Koopmans– Beckman version of the q-adic assignment problem with the approximation ratio O(q 2 n(q−1)/2 ). Proof. The input of the algorithm is a collection of weights {wu : u ∈ R} and {dv : v ∈ R}. The output is a permutation π : [n] → [n]. Fix the parameter ' & s ' & s n! n 1 1 = (q − 1)! . M= q−1 q q (n − q + 1)! Solve the linear programming relaxation (8)–(15). Denote the solution by (x∗ , y ∗ ) and the value of the objective function by LP ∗ . Now, greedily find a set of disjoint “heavy” (according to the LP) hyperedges E = {e ∈ R} using the following algorithm. 7 • Initialization: Let U and E be subsets of R. Initially, set U = R; E = ∅. • while (U = 6 ∅) – Find the most expensive edge e ∈ U with respect to the cost function X ∗ lp-cost(e) = we dv ye,v . (17) v∈R – Add e to the set E. – For every vertex i ∈ e, we denote e(i) = e. – Sequentially, for each i ∈ e and s ∈ {1, . . . , q}: we set Ai,s = {u ∈ U : us = i} and then remove elements of Ai,s from the set U . Thus, we partition all hyperedges in U that intersect with e into disjoint sets Ai,s (where i ∈ e, s ∈ {1, . . . , q}), such that for every u ∈ Ai,s , us = i. Note also that by removing hyperedges of Ai,s from U we ensure that each vertex i has at most one hyperedge e(i) associated with it. Moreover, each hyperedge belongs to exactly one set Ai,s . – Let Bi,s be the set of M hyperedges u from Ai,s with the maximum value of wu . If |Ai,s | < M , then let Bi,s = Ai,s . Note, that ∪i∈[n] ∪s∈{1,...,q} Ai,s = R and Ai,s ∩ Ai′ ,s′ = ∅ for (i, s) 6= (i′ , s′ ). We let I = ∪e∈E e = {i ∈ [n] : ∃e ∈ E s.t. i ∈ e}. By convention, we define Ai,s = ∅ if i does not belong to any edge e ∈ E. Note, that e(i) is defined for every i ∈ I, and Ai,s = ∅ if i ∈ / I. We have LP ∗ = XX ∗ wu dv yu,v = X X X ∗ wu dv yu,v i∈[n] s=1 u∈Ai,s \Bi,s v∈R | ∗ wu dv yu,v i∈[n] s=1 u∈Ai,s v∈R u∈R v∈R q X = q X X XX {z + q X X XX ∗ . wu dv yu,v i∈[n] s=1 u∈Bi,s v∈R } LPI∗ | {z ∗ LPII } ∗ . We now consider two cases. In Denote the first term in the sum LPI∗ and the second term LPII each case we run a separate approximation algorithm. Our final algorithm could be viewed as an algorithm that runs these two approximation algorithms and chooses the best solution. I. First, assume, that LPI∗ ≥ LP ∗ /2. Let mi,s = min{wu : u ∈ Bi,s }, if |Bi,s | = M ; and mi,s = 0, if |Bi,s | < M . Note, that max{wu : u ∈ Ai,s \ Bi,s } ≤ mi,s ≤ 8 1 X wu . M u∈Bi,s (18) We have LPI∗ = q XX X X ∗ wu dv yu,v i∈[n] s=1 u∈Ai,s \Bi,s v∈R ≤ X q X X X ∗ mi,s dv yu,v i∈[n] s=1 u∈R:us =i v∈R ≤ ≤ q XX X X ∗ mi,s dv yu,v i∈[n] s=1 u∈Ai,s \Bi,s v∈R q X XX mi,s dv x∗i,vs . i∈[n] s=1 v∈R In the last inequality, we used LP constraint (9). Then, using upper bound (18) on mi,s , we get q X q X XX XX X 1 mi,s dv x∗i,vs ≤ LPI∗ ≤ wu dv x∗i,vs M ≤ Hence, i∈[n] s=1 v∈R q XX 1 M i∈[n] s=1 v∈R X X wu dv x∗i,vs = i∈[n] s=1 u:us =i v∈R q X X s=1 u,v∈R 1 M u∈Bi,s q X X X wu dv x∗us ,vs . s=1 u∈R v∈R 1 wu dv x∗us ,vs ≥ M · LP ∗ . 2 By Lemma 3.1, there exists a randomized polynomial-time algorithm that finds a solution of cost ! −1 X q X (1 − 1q )2(q−1) n ∗ ALGI ≥ xus vs wu dv q! q−1 s=1 u,v∈R q n 1 LP ∗ 1 q (q − 1)! q−1 ∗ LP ≥ ≥ q−1 . n 2e2 q! q−1 2e2 q 2 n 2 ∗ ≥ LP ∗ /2. For every e ∈ E, select independently at random v ∈ R II. Now, assume, that LPII ∗ , and set φ(e) = v. Then, pick elements e ∈ E in a random order and for each with probability ye,v s = 1, . . . , q set π(es ) = φ(e)s , if the vertex φ(e)s does not have a preimage yet; and discard the vertex es , otherwise. In the end, map all discarded vertices in an arbitrary way. ∗ . We need to estimate the probability that π(e) = v. The probability that φ(e) = v equals ye,v ′ ′ Let N (v) be the number of e for which φ(e ) intersects v. We estimate E[N (v) | φ(e) = v] as follows X X X X E[N (v) | φ(e) = v] = 1 + ye∗′ v′ ≤ 1 + ye∗′ v′ = 1+ e′ ∈E\{e} v′ :v′ ∩v6=∅ q X q XX X e′ ∈E s=1 s′ =1 v′ :v′ ′ =vs s e′ ∈E v′ :v′ ∩v6=∅ ye∗′ v′ ≤ 1 + q X q XX e′ ∈E s=1 s′ =1 x∗e′ ′ vs . s Recall, P that all hyperedges e′ in E are disjoint. Thus the expression above can be bounded above n Pq by 1 + i=1 s=1 x∗ivs = 1 + q. This bound implies that ∗ yev 1 | φ(e) = v ≥ Pr[π(e) = v] = Pr[φ(e) = v] · E N (v) (q + 1), 9 where the last inequality follows from the Jensen’s inequality for the convex function x 7→ 1/x. Hence, the cost of the solution returned by the algorithm is at least ALGII ≥ XX we dv e∈E v∈R ∗ yev . q+1 On the other hand using that lp-cost(u) ≤ lp-cost(e(i)) for u ∈ Bi,s ⊂ Ai,s (see (17)), that |Bi,s | ≤ M , and that Bi,s = ∅ for i ∈ / I, we get ! ! q q n X X X X X X XX ∗ ∗ ∗ wu dv yu,v LPII = we(i) dv ye(i),v ≤ i=1 s=1 u∈Bi,s = q XX i∈I s=1 = qM |Bi,s | {z } lp-cost(u) X ∗ we(i) dv ye(i),v v∈R XX X e∈E i∈e i∈I s=1 u∈Bi,s v∈R | v∈R ∗ we dv ye,v ! ! 2 ≤ qM =q M X X i∈I X X e∈E v∈R | {z lp-cost(e(i)) ∗ we(i) dv ye(i),v v∈R ∗ we dv ye,v v∈R ! ! } . Combining the inequalities, we get ALGII ≥ ∗ LPII LP ∗ ≥ . (q−1) q 2 (q + 1)M 4n 2 q(q + 1) Thus, we have showed that the approximation ratio of the algorithm is O(q 2 n 6 (q−1) 2 ). Integrality Gap for the General Case We will use a variant of Bernstein inequality (see for example [16], Theorem 2.3 (b)). Lemma 6.1 (see e.g., [16], Theorem 2.3P (b)). Let the random variables X1 , . . . , Xm be independent, with 0 ≤ Xk ≤ 1 for each k. Let Sm = m k=1 Xk , let µ = E[Sm ]. Then for any ε > 0, 2 ε µ − 2(1+ε/3) Pr[Sm ≥ (1 + ε)µ] ≤ e Pr[Sm ≤ (1 − ε)µ] ≤ e−ε 2 µ/2 Theorem q−1 6.2. For each q ≥ 2, the integrality gap of the LP relaxation of the q-adic assignment is Ωq nln n . Proof. Consider the following random instance of the maximum q-adic assignment problem: For each u, v ∈ R, let cu,v = 1 independently with probability p= n ln n q! nq and cu,v = 0 with probability 1 − p. Consider any fixed integral solution xij , i, j ∈ [n] of the maximum q-adic assignment problem on that random instance. The total number of pairs u, v for 10 Q u, v ∈ R such that s∈[q] xus ,vs = 1 is exactly q! nq . Therefore, the expected value of the fixed integral solution on our random instance is exactly µ1 = pq! nq = n ln n. Let S be the value of our integral solution. By Lemma 6.1 with ε = 2, Pr[S > (1 + ε)µ1 ] ≤ e−6n ln n/5 . Now, using that the total number of possible integral solutions is n! < nn = en ln n , we get by the union bound that the value of all integral solutions for our random instance is at most (1 + ε)µ1 = 3n ln n with probability at least 1 − e−n ln n/5 . Let P ⊆ R be the random set of pairs (u, v) such that cu,v = 1 or cuσ ,vσ = 1 for some permutation σ : [q] → [q]. Define the following fractional solution of the linear programming relaxation (8)–(15): xij = 1/n for all i, j ∈ [n]; yu,v = y = 10q!n1 ln n for all (u, v) ∈ P; and yu,v = 0 for all (u, v) ∈ R×R\P. For convenience, we define an equivalence relation ∼ on pairs (u, v) ∈ R × R: Let (u′ , v ′ ) ∼ (u′′ , v ′′ ), if for some σ ∈ Sq , (u′ , v ′ ) = (u′′σ , vσ′′ ). Then yu′ ,v′ = yu′′ ,v′′ if (u′ , v ′ ) ∼ (u′′ , v ′′ ); and yu′ ,v′ and yu′′ ,v′′ are independent otherwise. We claim that this fractional solution is feasible with high probability. The constraints (12) and (13) are always satisfied, because xij = 1/n. The constraint (11) is satisfied, because if (u, v) ∈ P, then (uσ , vσ ) ∈ P for every σ ∈ Sk . So it remains to check the feasibility of constraints (9), (10). Consider one of these constraints. E.g., X yu,v ≤ xivs . u∈R:us =i The total number of yuv variables on the left hand side is (q − 1)! n−1 q−1 . These variables are independent, since (u′ , v) ∼ (u′′ , v) only if u′ = u′′ . For each yuv , Pr[yuv = y] ≤ q!p since there are q! pairs (u, v) ∈ R that could have cu,v = 1. Therefore the expected value of the left hand side is at most n−1 n−1 n ln n 1 1 (q − 1)! · q!p · y = (q − 1)! · n · = . q−1 q−1 10q!n ln n 10n q We apply Lemma 6.1 with ε = 9. We scale the values of random variables yu,v and the expectation by factor 1/y = 10q!n ln n, so that the values of random variables belong to {0, 1} and the expectation equals µ2 = q! ln n. The probability that the value of the left hand side is more than − ε2 µ2 81 (1 + ε)/(10n) = 1/n, i.e. this constraint is violated, is at most e 2(1+ε/3) = e−q! ln n 8 ≤ n−10q! . The total number of constraints (9) and (10) is 2qnq! nq ≤ n3q . Thus, by the union bound, all constraints (9), (10) are satisfied with probability at least 1 − n−7q . The expected value of this fractional solution is exactly X X X µ3 = E cu,v yu,v = E y cu,v = y Pr[cu,v = 1] u,v∈R u,v∈R u,v∈R 2 n n ln n 1 n 1 2 = · · q! = Ωq (nq ), = y · |R| · p = q 10q!n ln n 10 q q! nq By Lemma 6.1 with ε = 1/2, the value of the fractional solution is less than µ3 /2 with probability at most e−µ3 /8 . 11 By the union bound again for all bad events, we obtain that with probability at least 1 − n−7q − e−n ln n/5 − e−µ3 /8 > 0 the ratio between best fractional and best integral solutions for our random instance is at least q−1 1 n n 10 q = Ωq . 2n ln n ln n 7 Integrality Gap for the Koopmans–Beckman Case Theorem 7.1. The integrality gap LP relaxation of the Koopmans–Beckman case of the q of the nq−1 q-adic assignment problem is Ωq ln n . Proof. In the Koopmans–Beckman variant of the maximum q-adic assignment problem we have cu,v = wu dv for all u, v ∈ R. Consider the following random instance. For each u ∈ R we define wu = 1 independently with probability s n ln n p= q! nq and wu = 0 with probability 1 − p. Similarly, for each v ∈ R we define dv = 1 independently with probability p and dv = 0 with probability 1 − p. The analysis of the integrality gap for this random instance is very similar to the analysis in Section 6. The only difference is that now the coefficients in the objective function are not chosen completely independently, as a result the bound on the integrality gap decreases. Q Consider any fixed integral solution xij , i, j ∈ [n]. Let W = {(u, v) : u, v ∈ R, s∈[q] xus ,vs = 1}. I.e., (u, v) ∈ W , if u is mapped to v. We have |W | = q! nq . Therefore, the expected value of the integral solution on our random instance is exactly p2 q! nq = n ln n. We can also view the value of this specific solution as a sum of q! nq independent boolean random variables Y(u,v) for (u, v) ∈ W where Y(u,v) = wu dv . By Lemma 6.1 with ε = 2, the value of the integral solution is larger than 3n ln n with probability at most e−6n ln n/5 . The total number of possible integral solutions is n! < nn = en ln n . Thus, by the union bound, the value of all integral solutions for our random instance is at most 3n ln n with probability at least 1 − e−n ln n/5 . Let P ⊆ R be the random set of vectors u ∈ R such that wu = 1 or wuσ = 1 for some permutation σ : [q] → [q]. Let Q ⊆ R be the random set of vectors v ∈ R such that dv = 1 or dvσ = 1 for some permutation σ : [q] → [q]. Define the following fractional solution of the linear programming relaxation (8)–(15): xij = 1/n for all i, j ∈ [n]; yu,v = y = 10q!np ln n for all (u, v) ∈ P × Q; and yu,v = 0 for all (u, v) ∈ R × R \ P × Q. In other words, yuv = y · (maxσ wuσ ) · (maxσ dvσ ). We define an equivalence relation ∼ on vectors u ∈ R: Let u′ ∼ u′′ , if for some σ ∈ Sq , u′ = u′′σ . We then pick a subset R′ ⊂ R that contains exactly one representative from each equivalence class. For example, we may pick R′ = {(u1 , . . . , uk ) ∈ R : u1 < u2 < · · · < uq }. Then for every u ∈ R, there exists one and only one permutation σ such that uσ ∈ R′ . Observe, that for a fixed v ∈ R, yu′ ,v = yu′′ ,v if u′ ∼ u′′ , and that the variables yu,v , where u ∈ R′ , are independent given maxσ dvσ . 12 We claim that this fractional solution is feasible with high probability. As in the previous section, constraints (11), (12), and (13) are always satisfied. We now check the feasibility of constraints (9), (10). Consider one of these constraints for some v ∈ R, s ∈ [q] and value us ∈ [n]: X yu,v ≤ xivs . u∈R:us =i The left hand side can be rewritten as follows (below, we write i ∈ u if ut = i for some t ∈ [q]): X X X yu,v = yuσ ,v ′ u∈R:us =i u∈R :i∈u = X u∈R′ :i∈u = (q − 1)! X σ∈Sq :uσ(s) =i X yu,v σ∈Sq :uσ(s) =i yu,v X u∈R′ :i∈u = (q − 1)!y max dvσ · max wuσ σ u∈R′ :i∈u σ X ≤ (q − 1)!y max wuσ . ′ u∈R :i∈u σ ′ The total number of random variables maxσ wuσ in the sum above is n−1 q−1 . For each u ∈ R , Pr[maxσ wuσ = 1] ≤ q!p. Therefore, i hX n−1 −1 n − 1 q!p = p E max wuσ ≤ q!p2 u∈R′ :i∈u σ q−1 q−1 n ln n −1 n − 1 = p−1 q ln n. = p q! q−1 q! nq p √ Note that for q ≥ 2 (for q = 1 the problem is trivial), p−1 q ln n ≥ 2 n(n − 1)/(n ln n) ln n > n. We apply Lemma 6.1 with ε = 9: hX i √ −1 − n . Pr max w ≥ 10p q ln n ≤ e u σ ′ u∈R :i∈u σ √ Hence, with probability 1 − e− n , X X yu,v ≤ (q − 1)!y u∈R:us =i u∈R′ :i∈u max wuσ = (q − 1)! σ p 1 10p−1 q ln n = . 10q!n ln n n √ By the union bound, all 2qnq! nq constraints (9), (10) are satisfied with probability at least 1−e n/2 . Finally, we compute the expected value of the fractional solution. Note that if wu = 1 and dv = 1, then yu,v = y, hence the expected value equals # # " " X X X X y wu dv = yE dv yu,v wu dv = E wu · E E u,v∈R u,v∈R u∈R v∈R 2 n2 n ln n 3 1 n 2 2 = y pq! = (q!) n q 10q!n ln n q! q q s n q+1 1 1 n ln n q = Ωq (n 2 (ln n) 2 ). = 10 q! 13 P P Applying the Bernstein inequality (Lemma 6.1), we get that both sums u∈R wu and v∈R dv are concentrated around their expectations with high probabily. Hence, the integrality gap of the instance is ! ! r q+1 nq−1 n 2 (ln n)1/2 = Ωq . Ωq n ln n ln n References [1] W. P. Adams and T. A. Johnson, Improved Linear Programming-based Lower Bounds for the Quadratic Assignment Problem, DIMACS Series in Discrete Mathematics and Theoretical Computer Science, 16, 1994, 43-77. [2] E. Arkin, R. Hassin and M. Sviridenko, Approximating the Maximum Quadratic Assignment Problem, Information Processing Letters, 77, 2001, pp. 13-16. [3] L. Babai, P. Codenotti, Isomorphism of Hypergraphs of Low Rank in Moderately Exponential Time. In Proc. 39th Ann. IEEE Symp. on Theory of Computing (FOCS 2008), IEEE Comp. Soc. Press 2008, pp. 667-676. [4] A. Barvinok, Estimating L∞ norms by L2k norms for functions on orbits. Found. Comput. Math. 2 (2002), no. 4, 393–412. [5] M. Bellare, S. Goldwasser, C. Lund, and A. Russel, Efficient Probabilistically Checkable Proofs and Applications to Approximation. Proceedings of the Twenty-Fifth Annual ACM Symposium on Theory of Computing (STOC 1993), pages 294–304. [6] R.E. Burkard, E. Cela, Heuristics for biquadratic assignment problems and their computational comparison, European Journal of Operations Research v. 83 (1995), pp. 283-300. [7] R.Burkard, E. Cela abd B. Klinz, On the biquadratic assignment problem. Quadratic assignment and related problems (New Brunswick, NJ, 1993), 117–146, DIMACS Ser. Discrete Math. Theoret. Comput. Sci., 16, Amer. Math. Soc., Providence, RI, 1994. [8] E. Cela, The Quadratic Assignment Problem: Theory and Algorithms, Springer, 1998. [9] B. Chor and M. Sudan, A Geometric Approach to Betweenness, SIAM J. Discrete Math. 11(4): 511-523 (1998). [10] V. Guruswami, J. Håstad, R. Manokaran, P. Raghavendra and M. Charikar, Beating the Random Ordering is Hard: Every ordering CSP is approximation resistant. SIAM J. Comput. 40(3) (2011), pp. 878–914. [11] R. Hassin, A. Levin and M. Sviridenko, Approximating the minimum quadratic assignment problems, ACM Transactions on Algorithms 6(1), (2009). [12] T. C. Koopmans and M. Beckman, Assignment problems and the location of economic activities, Econometrica 25 (1957), pp. 5376. 14 [13] E. Lawler, The quadratic assignment problem. Management Science 9 1962/1963, pp. 586–599. [14] Y. Makarychev, Simple Linear Time Approximation Algorithm for Betweenness, Microsoft Research Technical Report MSR-TR-2009-74. [15] K. Makarychev, R. Manokaran and M. Sviridenko, Maximum Quadratic Assignment Problem: Reduction from Maximum Label Cover and LP-based Approximation Algorithm, ICALP 2010, pp. 594–604. [16] C. McDiarmid, Concentration. Probabilistic methods for algorithmic discrete mathematics, 195–248, Algorithms Combin., 16, Springer, Berlin, 1998. [17] V. Nagarajan and M. Sviridenko, On the maximum quadratic assignment problem, Mathematics of Operations Research 34(4), pp. 859-868 (2009). [18] M. Queyranne, Performance ratio of polynomial heuristics for triangle inequality quadratic assignment problems, Operations Research Letters, 4, 1986, 231-234. [19] T. Winter and U. Zimmermann, Real-time dispatch of trams in storage yards. Mathematics of industrial systems, IV (Valparaiso, 1996). Ann. Oper. Res. 96 (2000), 287–315. 15