New Error Correcting Codes from Lifting by Alan Xinyu Guo B.S., Duke University (2011) S.M., Massachusetts Institute of Technology (2013) Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Electrical Engineering and Computer Science at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 2015 c Massachusetts Institute of Technology 2015. All rights reserved. β Author . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Department of Electrical Engineering and Computer Science May 1, 2015 Certified by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Madhu Sudan Adjunct Professor Thesis Supervisor Accepted by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Leslie A. Kolodziejski Chair, Department Committee on Graduate Students 2 New Error Correcting Codes from Lifting by Alan Xinyu Guo Submitted to the Department of Electrical Engineering and Computer Science on May 1, 2015, in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Electrical Engineering and Computer Science Abstract Error correcting codes have been widely used for protecting information from noise. The theory of error correcting codes studies the range of parameters achievable by such codes, as well as the efficiency with which one can encode and decode them. In recent years, attention has focused on the study of sublinear-time algorithms for various classical problems, such as decoding and membership verification. This attention was driven in part by theoretical developments in probabilistically checkable proofs (PCPs) and hardness of approximation. Locally testable codes (codes for which membership can be verified using a sublinear number of queries) form the combinatorial core of PCP constructions and thus play a central role in computational complexity theory. Historically, low-degree polynomials (the Reed-Muller code) have been the locally testable code of choice. Recently, “affine-invariant” codes have come under focus as providing potential for new and improved codes. In this thesis, we exploit a natural algebraic operation known as “lifting” to construct new affine-invariant codes from shorter base codes. These lifted codes generically possess desirable combinatorial and algorithmic properties. The lifting operation preserves the distance of the base code. Moreover, lifted codes are naturally locally decodable and testable. We tap deeper into the potential of lifted codes by constructing the “lifted Reed-Solomon code”, a supercode of the Reed-Muller code with the same error-correcting capabilities yet vastly greater rate. The lifted Reed-Solomon code is the first high-rate code known to be locally decodable up to half the minimum distance, locally list-decodable up to the Johnson bound, and robustly testable, with robustness that depends only on the distance of the code. In particular, it is the first high-rate code known to be both locally decodable and locally testable. We also apply the lifted Reed-Solomon code to obtain new bounds on the size of Nikodym sets, and also to show that the Reed-Muller code is robustly testable for all field sizes and degrees up to the field size, with robustness that depends only on the distance of the code. Thesis Supervisor: Madhu Sudan Title: Adjunct Professor 3 4 Acknowledgments First and foremost, I thank my advisor, Madhu Sudan, for his patient guidance, support, and encouragement throughout my four years at MIT. Our shared taste in finding clean, general solutions to algebraic problems has led to many collaborations. I thank Scott Aaronson and Dana Moshkovitz for serving on my thesis committee. I am also indebted to Ezra Miller, who guided my undergraduate research at Duke, and to Vic Reiner and Dennis Stanton who mentored me at their REU in Minnesota. I thank my co-authors during graduate school: Greg Aloupis, Andrea Campagna, Erik Demaine, Elad Haramaty, Swastik Kopparty, Ronitt Rubinfeld, Madhu Sudan, and Giovanni Viglietta. I especially thank Ronitt and Piotr Indyk for their guidance early on, Erik for his excitement and willingness to entertain my more fun research ideas (i.e. hardness of video games), and Swastik for hosting my visit to Rutgers and teaching me about codes and PCPs. I also thank friends and faculty with whom I have shared conversations at MIT: Pablo Azar, Mohammad Bavarian, Eric Blais, Adam Bouland, Mahdi Cheraghchi, Henry Cohn, Matt Coudron, Michael Forbes, Badih Ghazi, Pritish Kamath, Ameya Velingker, Henry Yuen, and so many others. I thank my friends outside of the field for their friendship and the good times during the past four years. I especially thank Vivek Bhattacharya, Sarah Freitas, Henry Hwang, Steven Lin, Ann Liu, Lakshya Madhok, and Roger Que. I am grateful to my family. Without my parents, I would not exist, and neither would this thesis. Moreover, they always supported my education and encouraged me to pursue my dreams. I thank my younger sister Julia for encouraging me to set a good example. Finally, my biggest thanks goes to Lisa — my fianceΜe, best friend, and companion for the rest of my life. Graduate school was not always easy, but you were always there to support me and pick me up when I was down, and to share in my triumphs. Thank you for your unwavering love and loyalty, without which I do not believe I would have survived through graduate school. Only you know every twist and turn my journey has taken. Without hesitation, I dedicate this thesis to you. 5 6 Contents 1 Introduction 1.1 1.2 11 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 1.1.1 Error Correcting Codes . . . . . . . . . . . . . . . . . . . . . . . . . . 11 1.1.2 PCPs and Local Algorithms . . . . . . . . . . . . . . . . . . . . . . . 13 1.1.3 Affine-Invariance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 This Thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 1.2.1 Main Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 1.2.2 Lifting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.2.3 Robust Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.2.4 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 1.2.5 Organization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 2 Preliminaries 27 2.1 Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 2.2 Probability and Concentration bounds . . . . . . . . . . . . . . . . . . . . . 29 3 Error Correcting Codes 31 3.1 Classical Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 3.2 Local decoding, correcting, and list-decoding . . . . . . . . . . . . . . . . . . 32 3.3 Local testing and robust testing . . . . . . . . . . . . . . . . . . . . . . . . . 34 3.4 Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 3.4.1 36 Reed-Solomon code . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 3.4.2 Reed-Muller code . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Affine-invariance 36 39 4.1 Affine-invariant Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 4.2 Equivalence of Invariance under Affine Transformations and Permutations . . 40 5 Lifting Codes 47 5.1 The Lift Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 5.2 Algebraic and Combinatorial Properties . . . . . . . . . . . . . . . . . . . . 48 5.2.1 Algebraic Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 5.2.2 Distance of Lifted Codes . . . . . . . . . . . . . . . . . . . . . . . . . 50 Local Decoding and Correcting . . . . . . . . . . . . . . . . . . . . . . . . . 53 5.3.1 Local Correcting up to 1/4 Distance . . . . . . . . . . . . . . . . . . 53 5.3.2 Local Correcting up to 1/2 Distance . . . . . . . . . . . . . . . . . . 54 5.3.3 Local Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 Local Testing and Robust Testing . . . . . . . . . . . . . . . . . . . . . . . . 59 5.4.1 Local Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 5.4.2 Robust Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 5.3 5.4 6 Robust Testing of Lifted Codes 6.1 6.2 61 Robustness of Lifted Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 6.1.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 6.1.2 Robustness for Small Dimension . . . . . . . . . . . . . . . . . . . . . 62 6.1.3 Robustness of Special Tensor Codes . . . . . . . . . . . . . . . . . . . 66 6.1.4 Robustness for Large Dimension . . . . . . . . . . . . . . . . . . . . . 74 Technical Algebraic Results . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 6.2.1 Degree Lift . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 6.2.2 Analysis of Subspace Restrictions . . . . . . . . . . . . . . . . . . . . 85 8 7 Applications 7.1 91 Lifted Reed-Solomon Code . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 7.1.1 Relationship to Reed-Muller . . . . . . . . . . . . . . . . . . . . . . . 92 7.1.2 Rate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95 7.1.3 Global List-Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . 97 7.1.4 Local List-Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 7.1.5 Main Result: The Code That Does It All . . . . . . . . . . . . . . . . 102 7.2 Robust Low-Degree Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 7.3 Nikodym Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 A Algebra Background 107 A.1 Arithmetic over finite fields . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 A.2 Tensor codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 B Finite field geometry 111 B.1 Affine maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111 B.2 Affine subspaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114 9 10 Chapter 1 Introduction 1.1 1.1.1 Background Error Correcting Codes Error correcting codes arise as a solution to the problem of communicating over a noisy channel. The sender first encodes the message using an error correcting code, which adds redundancy to the message, into a codeword. The codeword is then sent over the noisy channel. The receiver receives a word which is a corruption of the codeword. The receiver then decodes the received word and hopefully retrieves the original message. The actual code is the set of possible codewords. Two classical parameters of interest are the rate and distance of the code. The rate is the ratio of the message length to the codeword length, and measures the efficiency of the encoding. The distance measures the error-correcting capability of the code. The Hamming distance between two strings is the number of symbols in which they differ. The distance of a code is the minimum Hamming distance between two distinct codewords. If the distance of a code is π, then in principle one can detect up to π − 1 errors and correct up to ⌊π/2⌋ − 1 errors: if the codeword has been corrupted in at most π−1 locations, then it cannot have been corrupted into a different codeword, so to detect errors one “merely” checks if the received word is a codeword; if the codeword has been corrupted in at most ⌊π/2⌋ − 1 locations, then 11 there is at most one codeword within Hamming distance ⌊π/2⌋ − 1 of the received word, so to correct errors one “merely” finds the nearest codeword to the received word. We often prefer to work with the relative distance of a code, which is simply its distance divided by the length of the codewords. There is a fundamental tradeoff between rate and distance, which is still not fully understood. In addition, there is the problem of designing codes which support efficient algorithms for encoding and decoding. Another notion of decoding is list-decoding. We just showed that if a code π has distance π, then for any word, there is at most one codeword within Hamming distance ⌊π/2⌋ − 1. If we wish to correct more than ⌊π/2⌋ − 1 errors, we cannot guarantee that there is a unique codeword. However, if the radius is not too large, then we can hope that there are not too many codewords within the radius from the received word. Instead of outputting the correct codeword or message, a list-decoding would output a list of potential codewords or messages. Another important feature of an error correcting code is the alphabet. Designing a code is easier using a larger alphabet. If the alphabet size is allowed to grow with the code length π, then the Singleton bound asserts that the rate π and relative distance πΏ satisfies π + πΏ ≤ 1 + 1/π. This bound is tight, as demonstrated by the Reed-Solomon code. The Reed-Solomon code is perhaps the most ubiquitous code in the literature. The idea is simple. Let π ≥ 1 and let π ≥ π be a prime power. Let Fπ be the finite field of size π. Each message π = (π0 , . . . , ππ−1 ) of length π over the alphabet Fπ is interpreted as a degree π − 1 polynomial π(π) = π0 + π1 π + · · · + ππ−1 π π−1 . Let πΌ1 , . . . , πΌπ be the elements of Fπ . The encoding of π is simply the evaluation of π at every point: (π(πΌ1 ), . . . , π(πΌπ )). The rate of this code is clearly π = ππ , and it follows from the Fundamental Theorem of Algebra that πΏ = 1 − π−1 , π so that π + πΏ = 1 + 1/π, meeting the Singleton bound. Furthermore, the Reed-Solomon code can be efficiently decoded up to half its distance using, for instance, the Welch-Berlekamp algorithm [WB86] (see [GS92] for an exposition). Guruswami and Sudan [GS99] showed that the Reed-Solomon code with distance πΏ > 0 can be list-decoded √ up to the “Johnson bound” (1 − 1 − πΏ fraction errors). More precisely, they gave an efficient algorithm which, on input a received word, outputs a list of π(1/π2 ) codewords that 12 √ are within (relative) distance 1 − (1 + π) 1 − πΏ of the received word. A related code is the Reed-Muller code. This code, parameterized by a degree π and a number π of variables, consists of polynomials π : Fπ π → Fπ of degree at most π (or rather, (οΈπ+π)οΈ their evaluations). Its message length is π = π and its code length is π = π π . When π = 1, this is simply the Reed-Solomon code. If π < π, it follows from the SchwartzZippel lemma that the distance of the Reed-Muller code is 1 − ππ . Although the Reed-Muller code’s rate-distance tradeoff is worse than that of the Reed-Solomon code, the Reed-Muller code offers locality, which we discuss in Section 1.1.2. Like the Reed-Solomon code, the Reed-Muller code can be list-decoded up to the Johnson bound [PW04]. 1.1.2 PCPs and Local Algorithms In the late 1980s and early 1990s, there was an explosion of interest in sublinear-time algorithms. Blum, Luby, and Rubinfeld [BLR93] showed that one can probabilistically test whether a given function π : {0, 1}π → {0, 1} is linear — that is, π (x + y) = π (x) + π (y) for every x, y ∈ {0, 1}π — by querying π at only 3 points. If π is linear, then their test accepts with probability one, while if π is “π-far” from linear (it disagrees with every linear function in at least π-fraction of the domain), then their test rejects with probability Ω(π). The space of linear functions π : {0, 1}π → {0, 1} forms an error correcting code with rate π/2π and distance 1/2 — this code is known as the Hadamard code. This was in fact the first code shown to be locally testable — that is, using a sublinear number of queries to the received word, one can verify membership in the code. It is also easy to show that the Hadamard code is locally correctable — one can correct any given symbol of the received word with high probability using a sublinear number of queries. To correct π at x ∈ {0, 1}π , select random y ∈ {0, 1}π and output the value π (x + y) − π (y). A related notion if local decodability — one can correct any given symbol of the message with high probability by making a sublinear number of queries to the received word. We will often consider codes π ⊆ Fππ that are linear (i.e. π forms a vector space over Fπ ). The Reed-Solomon code, Reed-Muller code, and Hadamard code are all linear codes. 13 Rather than thinking of words in Fππ as sequences of length π, we view them as functions from some fixed set π of cardinality |π| = π to the range Fπ . The structure of the set π and symmetries will play a role later. We use {π → Fπ } to denote the set of all such functions. We say a function π is π -far from π if πΏ(π, π) , minπ∈π πΏ(π, π) ≥ π . Given a code π ⊆ {π → Fπ } and integer β, an β-local tester π― is a distribution π on β (π β , 2Fπ ) with the semantics as follows: given oracle access to π : π → Fπ , the tester π― samples (π, π ) ← π, where π = (π1 , . . . , πβ ) ∈ π β and π ⊆ Fβπ , and accepts π if and only if π |π , (π (π1 ), . . . , π (πβ )) ∈ π . The tester is π-sound if π― accepts π ∈ π with probability one, while rejecting π that is πΏ-far from π with probability at least π · πΏ. We will also be interested in a stronger property of testers known as their robustness, formally defined by Ben-Sasson and Sudan [BSS06], based on analogous notions in complexity theory due to Ben-Sasson et al. [BSGH+ 04] and Dinur and Reingold [DR04]. The hope with a robust tester is that, while it may make a few more queries than the minimum possible, the rejection is “more emphatic” in that functions that are far from π typically yield views that are far from acceptable, i.e. if πΏ(π, π) is large, then so is πΏ(π |π , π ) for typicaly choices of (π, π ) ← π. Formally, a tester π is πΌ-robust if E(π,π )←π [πΏ(π |π , π )] ≥ πΌ · πΏ(π, π). Robustness can be a much stronger property than mere soundnesss since it allows for composition with other local testers. In particular, if there is an πΌ-robust tester for π with distribution π and if for every (π, π ) in the support of π, the property of being in π has an β′ -local tester that is π-sound, then π has an β′ -local tester that is πΌ · π-sound. The hope that membership in π has a nice local tester for every π in the support of π may seem overly optimistic, but for many symmetric codes (such as affine-invariant codes, to be discussed later), all the π ’s are isomorphic — so this is really just one hope. The interest in sublinear-time algorithms is obvious from a practical perspective. As the amount of data stored and transmitted by society continues its explosive growth, even lineartime algorithms may be too slow, and also unnecessary. Indeed, statisticians understood this decades ago when estimating the population averages. If we want an approximate answer, often π(1) queries suffice to give a good approximation with high confidence. Moreover, in 14 the context of error correction, if we have a very large encoded file and only wish to decode a small portion of it, then we need our code to be locally decodable. Surprisingly, sublinear-time algorithms for algebraic codes have played a prominent role in computational complexity theory. In particular, locally testable codes form the “combinatorial core” of probabilistically checkable proofs (PCPs). A PCP for a language πΏ is a protocol involving two parties, a Prover and a Verifier, and an input π₯, whereby the Prover supplies a proof, depending on π₯, in an attempt to convince the Verifier that π₯ ∈ πΏ. The Verifier makes a small number of random queries to the proof and then either accepts or rejects the proof based on what it sees. A valid PCP for πΏ satisfies the following: if π₯ ∈ πΏ, then there is some proof that the Prover can provide such that π accepts with probability 1; if π₯ ∈ / πΏ, then regardless of the proof provided by the Prover, the Verifier will reject with high probability (over its random queries). The celebrated PCP Theorem, proved in [AS98, ALM+ 98], characterizes the complexity class NP as the class of languages with PCPs where the Verifier makes only π(1) queries to the proof and uses only π(log π) bits of randomness in determining its random queries, where π is the length of the input π₯. Not only did this theorem elucidate the class of NP a bit more, but it also paved the way for proving that, for many combinatorial optimization problems, even approximating the solution is NP-hard. At the heart of the proof of the PCP theorem lies the problem of low-degree testing — the problem of testing whether a given function π : Fπ π → Fπ has total degree deg(π ) ≤ π or is far from any such function. Low-degree testing has been the most extensively studied algebraic property testing problem. First studied in the work of Rubinfeld and Sudan [RS96], low-degree testing and many variations have been analyzed in many subsequent works — a partial list includes [ALM+ 98, FS95, AS03, RS97, MR06, AKK+ 05, KR06, JPRZ09, BKS+ 10, HSS11]. When π βͺ π, low-degree tests making as few as π + 2 queries are known, that have 1/ poly(π)-soundness (see, for instance, Friedl-Sudan [FS95]). However, tests that make π(π) queries achieve constant soundness (a universal constant independent of π, π, π provided π is sufficiently larger than π), and even constant robustness. This constant robustness is central 15 to the PCP construction of Arora et al. [ALM+ 98]. In all cases with π βͺ π, low-degree tests operate by considering the restriction of a function to a random line, or sometimes plane, in the domain, and accepting a function if its restriction to the chosen subspace is a polynomial of degree at most π. Thus, the different restrictions π are different affine subspaces of low dimension (one or two) and the acceptable pattern π is the same for all π. In particular, the robust analysis of the low-degree test allows for low-query tests, or even proofs, of membership in π in constant dimensional spaces to be composed with the low-degree test in high dimensions to yield low-query PCPs. Robustness turns out to be much more significant as a parameter to analyze in these results than the query complexity of the outer test. Indeed, subsequent strengthenings of the PCP theorem in various senses (e.g. in [AS03, RS97, MR06]) rely on improving the robustness to a quantity close to 1, and this leads to PCPs of arbitrarily small constant, and then even π(1), error. 1.1.3 Affine-Invariance In an attempt to understand what exactly makes codes like the Reed-Muller code testable, Kaufman and Sudan set out to systematically study affine-invariant codes [KS08]. By then, it was known that symmetry in properties contributed to their testability, as in graph property testing. The hope was to find the analogous notion of symmetry for algebraic properties such as linearity or low-degree, and affine-invariance seemed to be a promising abstraction. A code π ⊆ {Fπ π → Fπ } is affine-invariant if, for every codeword π ∈ π, the codeword π β π΄ obπ tained by first applying an affine permutation π΄ : Fπ π → Fπ to the domain before evaluating π , is also a codeword. Kaufman and Sudan showed in [KS08] that any linear affine-invariant code that is “π-single-orbit characterized” is testable with π queries and soundness Ω(π −2 ). For example, the Hadamard code is 3-single-orbit characterized by π (x)+π (y)−π (x+y) = 0 for any x, y ∈ Fπ π (the characterization is by constraints on 3 points, and the constraints are all in a single orbit by the group action of affine permutations). Their proof simplifies, unifies, and generalizes the proofs found in [BLR93, RS96, AKK+ 05, KR06, JPRZ09] and unearths some of the fundamental underlying reasons for why testing works. Additionally, 16 Kaufman and Sudan initiated the systematic study of affine-invariant properties, and laid the groundwork for our structural understanding of affine-invariant properties. The hope was that eventually the study of affine-invariance would lead to constructions of new affineinvariant codes with desirable testability properties. While the locality properties (testability and correctability) of Reed-Muller codes are well-studied, they are essentially the only rich class of symmetric codes that are well-studied. The only other basic class of symmetric codes that are studied seem to be sparse ones (codes with few codewords). 1.2 This Thesis In this thesis, we use an algebraic operation, known as “lifting”, to construct new affineinvariant codes. This thesis includes results from [GKS13, GK14, GHS15], though we omit some results and generalize other results from these papers. [GKS13] is joint work with Swastik Kopparty and Madhu Sudan. [GK14] is joint work with Swastik Kopparty. [GHS15] is joint work with Elad Haramaty and Madhu Sudan. 1.2.1 Main Result The main result of our thesis is the construction of high-rate LCCs and LTCs. Theorem 1.2.1 (Main theorem, informal). ∀π, π½ > 0 ∃πΏ, πΌ > 0 such that for infinitely many π there exists π = π(π) = π(ππ ) and a linear code π ⊆ Fππ of distance πΏ and rate 1 − π½ such that β π is locally correctable and decodable from Ω(πΏ) fraction errors with π(ππ ) queries; √ β π is list-decodable from 1 − 1 − πΏ fraction errors in polynomial time, and is locally √ list-decodable from 1 − 1 − πΏ fraction errors with π(π3π ) queries; β π has an πΌ-robust tester using π(π2π ) queries. Theorem 1.2.1 is proved in Section 7.1.5, and is the culmination of the work in this thesis. 17 1.2.2 Lifting The lifting operation was first defined and used in [BSMSS11] to prove negative results — in particular, to construct “symmetric LDPC codes” that are not testable. The work of [GKS13] initiated the systematic study of lifting, and also was the first work to use lifting to prove positive results — in particular, to construct new high-rate locally correctable and locally testable codes. Our definition of lifting is more general and somewhat cleaner than that of [BSMSS11]. Starting from a base linear affine-invariant code π ⊆ {Fππ‘ → Fπ }, we π define the π-dimensional lift π π‘βπ ⊆ {Fπ π → Fπ } to be the code consisting of all π : Fπ → Fπ satisfying the following: π is in the lift if and only if, for every π‘-dimensional affine subspace π΄ ⊆ Fπ π , the restriction π |π΄ ∈ π. Lifting is a natural algebraic operation on affine-invariant codes, which is evident in the generic properties that lifted codes naturally possess. We show that if the base code π has (relative) distance πΏ, then the lifted code π π‘βπ has distance πΏ − π −π‘ , and in fact if π ∈ {2, 3}, then the lifted distance is also πΏ. So, lifting approximately preserves distance. Moreover, π π‘βπ is naturally π π‘ -single-orbit characterized by construction, and so, by [KS08], the natural π‘-dimensional test — choose a random π‘-dimensional affine subspace π΄ ⊆ Fπ π and accept if and only if π |π΄ ∈ π — is Ω(π −2π‘ )-sound. Finally, lifted codes are naturally locally correctable π — to correct π at a point x ∈ Fπ π , choose a random π‘-dimensional subspace π΄ ⊆ Fπ passing through x, use the correction algorithm for π to correct π |π΄ to a codeword π ∈ π, and then output π(x). 1.2.3 Robust Testing of Lifted Codes In [GHS15], we consider robust testing of lifted codes. We propose and analyze the following π‘β2π‘ test for π π‘βπ : Pick a random 2π‘-dimensional subspace π΄ in Fπ . π and accept if π |π΄ ∈ π Our main theorem relates the robustness of this test to the distance of the code π. Theorem 1.2.2. ∀πΏ > 0 ∃πΌ > 0 such that the following holds: For every finite field Fπ , for every pair of positive integers π‘ and π, and for every affine-invariant code π ⊆ {Fπ‘π → Fπ } satisfying πΏ(π) ≥ πΏ, the code π π‘βπ has a π 2π‘ -local test that is πΌ-robust. 18 Theorem 1.2.2 is proved in Section 6.1. As we elaborate below, Theorem 1.2.2 immediately implies a robust analysis for low-degree tests. Whereas almost all previous robust analyses of low-degree tests had more complex conditions on the relationship between the robustness, the degree, and the field size, our relationship is extremely clean. The dependence of πΌ on πΏ that we prove is polynomial but of fairly high degree πΌ = Ω(πΏ 74 ). We do not attempt to improve this relationship in this thesis and choose instead to keep the intermediate statements simple and general. We note that a significant portion of this complexity arises due to our desire to lift π‘-dimensional codes for general π‘, and here the fact that the robustness lower-bound is independent of π‘ is itself significant. Comparing with other testing results for lifted codes, there are only two prior works to compare with: Kaufman and Sudan [KS08] analyze a tester for a broader family of codes that they call “single-orbit” codes. Their result would yield a robustness of Θ(π −3π‘ ). (See Corollary 6.1.2.) Haramaty et al. [HRS13] also give a tester for lifted codes. They do not state their results in terms of robustness but their techniques would turn into a robustness of ππ · πΏ, where the ππ is a positive constant for constant π but goes to zero extremely quickly as π → ∞. Thus for growing π (and even slowly shrinking πΏ) our results are much stronger. Proof approach and some technical contributions In order to describe our test and analysis techniques, we briefly review the two main tests proposed in the literature for “low-degree testing”, when the field size is much larger than the degree. The most natural test for this task is the one that picks a random line in Fπ π and computes the proximity of the function restricted to this line to the space of univariate degree π polynomials. This is the test proposed by Rubinfeld and Sudan [RS96] and analyzed in [RS96, ALM+ 98, AS03]. A second low-degree test is somewhat less efficient in its query complexity (quadratically so) but turns out to have a much simpler analysis — this test would pick a random two-dimensional (affine) subspace in Fπ π and verify that the function is a bivariate polynomial of degree at most π on this subspace. This is the test proposed 19 by Raz and Safra [RS97] and analyzed in [RS97, MR06]. Both tests can be analyzed by first reducing the testing problem to that of testing constant variate functions (at most four variate functions) and then analyzing the constant dimensional problem as a second step. The first step is completely generic or at least it was sensed to be so. However there was no prior formalization of the fact that it is generic. The only class of functions to which it has been applied is the class of low-degree polynomials and a priori it is not clear how to even justify the claim of genericity. Here we show that the first step applies to all lifted codes, thus giving the first justification of the presumed genericity of this step, which we consider to be a conceptual contribution. For the second step, the robust analyses in [ALM+ 98, AS03] are quite algebraic and there seems to be no hope to use them on general lifted codes. The test and analysis of Raz and Safra [RS97] on the other hand feels much more generic. In this work we use their test, and extend it to general lifted codes and show that it is robust. Even the extension of the test is not completely obvious. In particular, to test low-degree polynomials they look at restrictions of the given function to 2-dimensional “planes”. When lifting π‘-dimensional properties, it is not immediate what would be the dimension of the restrictions the test should look at: Should it be π‘ + 1? Or 2π‘ or maybe 3π‘ − 1 (each of which does make logical sense)? We show that the 2π‘ dimensional tests are robust, with robustness being independent of π‘. Next we turn to our analysis. In showing robustness of their test, applied to generic lifted codes there is a major barrier: Almost all analyses of low-degree tests, for polynomials of degree at most π, attempt to show first that a function passing the test with high probability is close to a polynomial of degree twice the degree, i.e., at most 2π, with some additional features. They then use the distance of the space of polynomials of degree 2π and the additional features to establish that the function being tested is really close to a degree π polynomial. In extending such analyses to our setting we face two obstacles: In the completely generic setting, there is no nice notion corresponding to the set of degree 2π polynomials. One approach might be to consider the linear space spanned by products of functions in our basic space and work with them, but the algebra gets hairy to understand and 20 analyze. Even if we abandon the complete genericity and stick to the space of polynomials of degree π, but now allow π > π/2 we hit a second obstacle: The space of polynomials of degree 2π have negligible relative distance compared to the space of polynomials of degree π. Thus we need to search for a new proof technique and we find one by unearthing a new connection between “lifted codes” and “tensor product” codes. The tensor product is a natural operation in linear algebra and when applied to two linear codes, it produces a new linear code in a natural way. Tensor products of codes are well-studied in the literature on coding theory. The testing of tensor product codes was initiated by Ben-Sasson and Sudan [BSS06] and subsequently has been well-studied [DSW06, Val05, BSV09b, BSV09a, GGR09]. Specifically, a recent result of Viderman [Vid12] gives a powerful analysis which we are able to reproduce in a slightly different setting to get our results. In particular this is the ingredient that allows us to work with base codes whose distance is less than 1/2. Also, for the sake of the exposition we pretend that this test can test two-dimensional tensor products of one dimensional codes, with one-dimensional tests. (Actually, the test works with three dimensional tensors and tests them by looking at two-dimensional planes, but by suppressing this difference, our exposition becomes a little simpler.) To explain the connection between lifted codes and tensor product codes, and the idea that we introduce to test the former, we turn to the simple case of testing a bivariate lift of a univariate Reed-Solomon code. Specifically, let π be the family of univariate polynomials of degree at most π mapping Fπ to Fπ . Let π2 be the family of bivariate polynomials that become a univariate polynomial of degree at most π on every restriction to a line. The tensor product of π with itself, which we denote π ⊗2 corresponds to the set of bivariate polynomials of degree at most π in each variable. Clearly π2 ⊆ π ⊗2 but such subset relationships are not of immediate use in testing a code. (Indeed locally testable codes contain many nonLTCs.) To get a tighter relationship, now fix two “directions” π1 and π2 and let ππ1 ,π2 be the code containing all bivariate polynomials over Fπ that on every restriction to lines in directions π1 and π2 form univariate degree π polynomials. On the one hand the code ππ1 ,π2 is just isomorphic to the tensor product code π ⊗2 which is testable by the natural test, 21 by our assumption. On the other hand, we now have π2 = ∩π1 ,π2 ππ1 ,π2 so we now have a characterization of the lifted codes in terms of the tensor product. One might hope that one could use this characterization to get a (robust) analysis of the lifted test since it tests membership in ππ1 ,π2 for random choices of π1 and π2 , but unfortunately we do not see a simple way to implement this hope. Our key idea is look instead at a more complex family of codes ππ1 ,π2 ,π3 that consists of functions of degree π in directions π1 , π2 and π3 . (Of course now π1 , π2 , π3 are linearly dependent and so ππ1 ,π2 ,π3 is not a tensor product code. We will return to this issue later.) We still have π2 = ∩π1 ,π2 ,π3 ππ1 ,π2 ,π3 . Indeed we can even fix π1 , π2 arbitrarily (only requiring them to be linearly independent) and we have π2 = ∩π3 ππ1 ,π2 ,π3 . This view turns out to be more advantageous since we now have that for any π3 and π′3 we have ππ1 ,π2 ,π3 ∪ ππ1 ,π2 ,π′3 ⊆ ππ1 ,π2 which is a code of decent distance. This allows us to show that if the function being tested is close to ππ1 ,π2 ,π3 for many choices of π3 then the nearest codewords for all these choices of π3 are the same. An algebraic analysis of lifted codes tells us that a codeword of ππ1 ,π2 can not be in ππ1 ,π2 ,π3 for many choices of π3 without being a codeword of the lifted code and this lends promise to our idea. But we are not done, since we still need to test the given function for proximity to ππ1 ,π2 ,π3 and this is no longer a tensor product code so Viderman’s result does not apply directly. Fortunately, we are able to develop the ideas from Viderman’s analysis for tensor product codes [Vid12] and apply them also to our case and this yields our test and analysis. We note that this extension is not immediate — indeed one of the central properties of tensor product codes is that they are decodable from some clean erasure patterns and this feature is missing in our codes. Nevertheless the analysis can be modified to apply to our codes and this suffices to complete the analysis. In the actual implementation, as noted earlier, we can’t work with univariate tests even for the simple case above, and work instead by using a bivariate test for trivariate and 4variate functions. (This is similar to the reasons why Raz and Safra used a bivariate test.) This complicates the notations a bit, but the idea remains similar to the description above. Our task gets more complicated when the base code being lifted is π‘-dimensional for π‘ > 1. 22 The most natural adaptation of our analysis leads to dependencies involving πΏ (the distance of the base code) and π‘. We work somewhat harder in this case to eliminate any dependence on π‘ while working within the framework described above. 1.2.4 Applications Lifted Reed-Solomon Code The most interesting construction arising from lifting is the “lifted Reed-Solomon code”. As its name suggests, the lifted Reed-Solomon code is obtained by simply lifting the ReedSolomon code, a univariate code, to π dimensions. The lifted Reed-Solomon, by definition, contains the Reed-Muller code. However, if the degree π of the Reed-Solomon code is sufficiently large relative to the field size — in particular, if π ≥ π − π/π, where Fπ has characteristic π — then the lifted Reed-Solomon code contains polynomials of high degree as well. This fact follows from the characterization of low-degree polynomials as proven in [KR06]. In fact, using structural results about affine-invariance and lifts, we easily re-prove a special case of the characterization of low-degree polynomials in Theorem 7.1.6. The lifted ReedSolomon code is arguably more natural than the Reed-Muller code, the latter of which is an unnecessarily sparse subcode of the former. We therefore expect the lifted Reed-Solomon code to exhibit the same versatility as that of the Reed-Muller code, and indeed since the lifted Reed-Solomon code is a lifted code, it generically has good distance and is locally decodable and testable. We show that it is also in fact (locally) list-decodable up to the Johnson radius, just like the Reed-Muller code. What sets the lifted Reed-Solomon code apart from the Reed-Muller code is its vastly greater rate. If one insists on having positive distance πΏ > 0, then the π-variate Reed-Muller code has rate bounded by 1/π!, whereas the rate of the lifted Reed-Solomon code approaches 1 as πΏ → 0. The first family of high-rate locally correctable codes known were the multiplicity codes of Kopparty, Saraf, and Yekhanin [KSY14]. Kopparty [Kop12] showed that multiplicity codes are also locally list-decodable up to the Johnson radius. The only prior construction of high-rate codes that are robustly testable are the tensor product codes of Viderman [Vid12]. 23 Thus, the lifted Reed-Solomon code is the first high-rate code known to be locally correctable, locally list-decodable up to the Johnson radius, and robustly testable. This is the code in Theorem 1.2.1. Robust Low-Degree Testing An almost direct corollary of our robustness result for lifted codes is a π 4 -local robust lowdegree test for the setting π ≤ (1 − πΏ)π. To see why we get π 4 queries, note that when π ≥ π − π/π, the π-variate Reed-Muller code of degree π is not equal to the π-dimensional lift of the degree π Reed-Solomon code. But the latter turns out to be the π-dimensional lift of the bivariate Reed-Muller code of degree π. Applying our robust testing result to this lifted family yields a robust test making π 4 queries. But with some slight extra work we can get a better tester that makes only π 2 queries and this yields the following theorem. Theorem 1.2.3. ∀πΏ > 0 ∃πΌ > 0 such that the following holds: For every finite field Fπ , for every integer π ≤ (1 − πΏ)π and every positive integer π, there is a π 2 -query πΌ-robust low-degree test for the class of π-variate polynomials of degree at most π over Fπ . We note that previous works on low-degree testing worked only when π < π/2. This ratio seems to be achieved by Friedl and Sudan (see [FS95, Theorem 13]). Other works [RS96, ALM+ 98, RS97, AS03, MR06] seem to achieve weaker ratios for a variety of reasons discussed above. Nikodym Set Size Bounds One of the applications of lifted codes is to bounding, from below, the size of “Nikodym sets” over finite fields (of small characteristic). A set π ⊆ Fπ π is a Nikodym set if every point x has a line passing through it such that all points of the line, except possibly the point x itself, are elements of π. Nikodym sets are closely related to “Kakeya sets” — the latter contain a line in every direction, while the former contain almost all of a line through every point. A lower bound for Kakeya sets over finite fields was proved by Dvir [Dvi08] using the polynomial method and further improved by using the “method of multiplicites” by Saraf and 24 Sudan [SS08] and Dvir et al. [DKSS09]. Kakeya sets have seen applications connecting its study to the study of randomness extractors, especially [DS07, DW11]. Arguably, Nikodym sets are about as natural in this connection as Kakeya sets. Previous lower bounds on Kakeya sets were typically also applicable to Nikodym sets and led to bounds of the form |π| ≥ (1 − π(1))π π /2π where the π(1) term goes to zero as π → ∞. In particular, previous lower bounds failed to separate the growth of Nikodym sets from those of Kakeya sets. In our work, we present a simple connection that shows that the existence of the lifted Reed-Solomon code yields a large lower bound on the size of Nikodym sets, thereby significantly improving the known lower bound on the size of Nikodym sets over fields of constant characteristic. Theorem 1.2.4. For every prime π, and every integer π, there exists π = π(π, π) > 0 such that for every finite field Fπ of characteristic π, if π ⊆ Fπ π is a Nikodym set, then |π| ≥ π π − π (1−π)π . In particular, if π → ∞, then |π| ≥ (1 − π(1)) · π π . Thus, whereas previous lower bounds on the size of Nikodym sets allowed for the possibility that the density of the Nikodym sets vanishes as π grows, ours show that Nikodym sets occupy almost all of the space. One way to view our results is that they abstract the polynomial method in a more general way, and thus lead to stronger lower bounds in some cases. 1.2.5 Organization In Chapter 2, we establish notation and preliminary definitions. In Chapter 3, we formally define error correcting codes and relevant models: local decoding, correcting, list-decoding, and testing. In Chapter 4, we review some structural properties of affine-invariant codes. In Chapter 5, we formally define the lifting operation and prove generic properties of lifted codes. In Chapter 6, we prove that lifted codes are robustly testable with robustness parameter that depends only on the distance of the base code. In Chapter 7, we discuss applications of lifted codes: construction of the lifted Reed-Solomon code, robust low-degree testing, and new lower bounds on the size of Nikodym sets. 25 26 Chapter 2 Preliminaries 2.1 Notation Letters. We will typically use lower-case italic letters (e.g. π, π, π) to denote scalar values, lower-case bold letters (e.g. a, b, c) to denote vectors, and upper-case bold letters (e.g. A, B, C) to denote matrices. If A is a matrix, then we denote by Aπ* the π-th row of A and by A*π the π-th row of A. Sets and functions. For a set π and π ∈ N, let (οΈπ )οΈ π be the collection of subsets π ⊆ π with |π | = π. Let 2π be the collection of all subsets of π. For a positive integer π, define [π] , {1, . . . , π} and JπK , {0, 1, . . . , π − 1}. For sets π΄ and π΅, let {π΄ → π΅} denote the set of functions from π΄ to π΅. Vectors and Hamming distance. If a ∈ Nπ , let βaβ , ∑οΈπ π=1 ππ . If Σ is a finite set, π ∈ N, and s ∈ Σπ is a string, then let π π denote the π-th component of s, for π ∈ [π], so that s = (π 1 , . . . , π π ). For two vectors a, b ∈ Σπ , denote their Hamming distance by Δ(a, b) , #{π ∈ [π] | ππ ΜΈ= ππ } and their normalized Hamming distance by πΏ(a, b) , |Δ(a,b)| . π If π ⊆ Σπ is a set, then πΏ(a, π) , minb∈π πΏ(a, b) is the distance from a to π. If Σ is a field, then the (resp. normalized) Hamming weight of a ∈ Σπ is (resp. πΏ(a, 0)) Δ(a, 0). We will frequently think of functions π : π΄ → π΅ as vectors in π΅ π΄ and so we extend the vector notations to 27 functions as well. In particular, if π, π : π΄ → π΅, then Δ(π, π) , #{π₯ ∈ π΄ | π (π₯) ΜΈ= π(π₯)} and πΏ(π, π) , Δ(π,π) . |π΄| Note that πΏ(π, π) = Prπ₯∈π΄ [π (π₯) ΜΈ= π(π₯)]. Minkowski sums and affine subspaces. For sets π΄, π΅ ⊆ Fπ , their Minkowski sum is denoted π΄ + π΅ , {a + b | a ∈ π΄, b ∈ π΅}. For π΄ ⊆ Fπ , the span of π΄ is denoted ∑οΈ span(π΄) , { a∈π΄ πa · a | πa ∈ F}. For x ∈ Fπ and π΄ ⊆ Fπ , let (x, π΄) , {x} + span(π΄) be the affine subspace through x in directions π΄. Shadows. Let π, π, π ∈ N and π = ∑οΈ π≥0 π(π) ππ , π = ∑οΈ π≥0 π(π) ππ with π(π) , π(π) ∈ JπK for π ≥ 0. Then π is in the π-shadow of π, denoted by π ≤π π, if π(π) ≤ π(π) for π ≥ 0. If a, b ∈ Nπ , then a ≤π b means that ππ ≤π ππ for every π ∈ [π]. If a ∈ Nπ and π ∈ N, then a ≤π π means that ∑οΈ π×π and b ∈ Nπ , then A ≤π b means that π∈π ππ ≤π π for any subset π ⊆ [π]. If A ∈ N Aπ* ≤π b for every π ∈ [π]. Finite fields and polynomials. If π is a prime and π is a power of π, then Fπ denotes the finite field of size π. If X = (π1 , . . . , ππ ) are variables and d ∈ Nπ , then Xd denotes ∏οΈ ππ the monomial π π=1 ππ . Any polynomial β(X) ∈ Fπ [X] with such that β(X) = π (X) + ∏οΈ π π π(X) π π=1 (ππ − ππ ) for some π(X) ∈ Fπ [X] defines the same function Fπ → Fπ , since the polynomial π π − π is identically zero on Fπ . Observe that every function π : Fπ π → Fπ can ∑οΈ be expressed uniquely as a linear combination π (X) = d∈JπKπ πd ·Xd of monomials Xd with d ∈ JπKπ . Throughout this thesis, when we refer to a “polynomial” π : Fπ π → Fπ , we mean the unique π (X) defined above. The support of π is supp(π ) , {d ∈ JπKπ | πd ΜΈ= 0}. The degree of a polynomial π : Fπ π → Fπ is deg(π ) , max{βdβ | d ∈ supp(π )}. When we refer π to a polynomial π : Fπ π → Fπ in the context of codewords, we refer to the π -dimensional vector (π (x))x∈Fπ . π 28 Mod-star. For π ∈ N and π > 1, define the operation mod* by π mod* π = β§ βͺ β¨π π<π βͺ β©π mod (π − 1) π ≥ π * so that π π ≡ π π mod 2.2 π (mod π π − π). Probability and Concentration bounds If π is a random variable, we use E[π] and Var[π] to denote the expectation and variance of π, respectively. If the probability space of π is not clear from context, we use subscripts, e.g. if π = π(π, π), then Eπ [π] is the average over π with π fixed. Proposition 2.2.1 (Markov inequality). Let π ≥ 0 be a random variable with E[π] < ∞ and let π > 0. Then Pr[π ≥ π] ≤ E[π] . π Proposition 2.2.2 (Chebyshev inequality). Let π be a random variable with E[π] < ∞ and Var[π] < ∞, and let π > 0. Then Pr[|π − E[π]| ≥ π] ≤ Var[π] . π2 Proposition 2.2.3 (Hoeffding inequality). Let π1 , . . . , ππ ∈ [0, 1] be independent random β [οΈβ ]οΈ ∑οΈ variables and let π , π1 ππ=1 ππ . Let π > 0. Then Pr βπ − E[π]β > π ≤ 2 exp(−2ππ2 ). ΜοΈ Proposition 2.2.4. Let π, π : Fπ π → Fπ . Let πΏ(π, π) be the estimate of πΏ(π, π) by random sampling, i.e. independent uniformly random x ∈ Fπ π are chosen and the average of 1π (x)ΜΈ=π(x) is output. If Θ(ln(1/π)/π2 ) queries are used in the sample, then with probability at least 1−π, ΜοΈ π) has additive error at most π. the estimate πΏ(π, Proof. Let π ≥ ln(2/π)/(2π2 ). For each π ∈ [π], let ππ , 1π (xπ )ΜΈ=π(xπ ) , and define π , ∑οΈπ 1 ΜοΈ π=1 ππ so that πΏ(π, π) = π, and πΏ(π, π) = E[π]. By Proposition 2.2.3, π β [οΈβ ]οΈ βΜοΈ β Pr βπΏ(π, π) − πΏ(π, π)β > π ≤ 2 exp(−2ππ2 ) ≤ π. 29 30 Chapter 3 Error Correcting Codes 3.1 Classical Parameters Error correcting codes are schemes for encoding messages as codewords to protect them from noise. The code itself is the subset of valid codewords. The rate of a code is the ratio of the message length to the encoding length, and measures the efficiency of the encoding. The distance of a code measures the minimum distance between valid codewords, and indicates the error-correcting capability of the code. In this section, we formally define these notions. Definition 3.1.1 (Code). Let Σ be a finite set and let π be a natural number. A code over Σ of block length π is a subset π ⊆ Σπ . If there exists a set Σ0 , integer π ≤ π, and injective function Enc : Σπ0 → Σπ such that Enc(Σπ0 ) = π, then Enc is an encoding function for π. Definition 3.1.2 (Rate of a code). The rate of a code π ⊆ Σπ is log |π| . π log |Σ| Definition 3.1.3 (Distance of a code). The (normalized) distance of a code π ⊆ Σπ is πΏ(π) , minxΜΈ=y∈π πΏ(x, y). The following proposition shows that, algorithmic efficiency aside, any code supports unique decoding up to half its minimum distance. Proposition 3.1.4. For every r ∈ Σπ , there exists at most one c ∈ π with πΏ(r, c) < 31 πΏ(π) . 2 Proof. Suppose c, c′ ∈ π and πΏ(r, c), πΏ(r, c′ ) < πΏ(π) . 2 Then, by the triangle inequality, πΏ(c, c′ ) ≤ πΏ(c, r) + πΏ(r, c′ ) < πΏ(π), so c = c′ . Linear codes are codes whose alphabet is a (finite) field and whose codewords form a vector space over the field. Every code of interest to us in this thesis is a linear code. Definition 3.1.5 (Linear code). A code π ⊆ Σπ is linear if Σ = F is a field and π is a linear subspace of Fπ . Proposition 3.1.6. If π ⊆ Fπ is a linear code, then πΏ(π) is equal to the minimal normalized Hamming weight of nonzero c ∈ π. Proof. Let c ∈ π be nonzero of minimal Hamming weight. By definition, πΏ(π) ≤ πΏ(c, 0). On the other hand, for any two distinct x, y ∈ π, πΏ(x, y) = πΏ(x − y, 0) ≥ πΏ(c, 0), and so minimizing over x ΜΈ= y, we have πΏ(π) ≥ πΏ(c, 0). 3.2 Local decoding, correcting, and list-decoding In this section, we formally define the models of local decoding, local correcting, and local list-decoding. Intuitively, local decoding entails recovering a symbol of the original message using few queries, while local correcting entails recovering a symbol of the original codeword using few queries. Definition 3.2.1 (Local decoding). A code π ⊆ Σπ with encoding function Enc : Σπ0 → Σπ is (β, π, π)-locally decodable if there exists a randomized oracle π : [π] → Σ0 with oracle access to a received word r ∈ Σπ such that 1. ππ queries at most β symbols of r; 2. if there is m ∈ Σπ0 with πΏ(Enc(m), r) ≤ π , then Pr [ππ (π) = ππ ] ≥ 1−π for every π ∈ [π]. Definition 3.2.2 (Local correcting). A code π ⊆ Σπ is (β, π, π)-locally correctable if there exists a randomized oracle π : [π] → Σ with oracle access to a received word r ∈ Σπ such that 32 1. ππ queries at most β symbols of r; 2. if there is c ∈ π with πΏ(c, r) ≤ π , then Pr [ππ (π) = ππ ] ≥ 1 − π for every π ∈ [π]. If π is a linear code, then it is possible to encode π in systematically, i.e. such that the original message is part of the codeword. Of course, this is not algorithmically satisfying unless the systematic encoding function is explicit, i.e. computable in polynomial time. When π is linear, we can think of it as a space of functions {π → Fπ }. A systematic encoding is then equivalent to finding an interpolating set for π in π. Definition 3.2.3. If π ⊆ {π → Fπ } is linear, then πΌ ⊆ π is an interpolating set for π if, for every π : πΌ → Fπ , there exists a unique extension π ∈ π such that π|πΌ = π . Remark 3.2.4. If π ⊆ {π → Fπ } has an interpolating set πΌ ⊆ π, then |π| = π |πΌ| . In particular, if π is linear, then |πΌ| = dimFπ (π). Proposition 3.2.5. If π ⊆ {π → Fπ } is linear and has an explicit interpolating set πΌ ⊆ π, and π is a (β, π, π)-locally correctable code, then π is a (β, π, π)-locally decodable code. Proof. Let Enc be the map which takes π : πΌ → Fπ to its unique extension π ∈ π such that π|πΌ = π , guaranteed by the fact that πΌ is an interpolating set. Then the local correcting algorithm for π also serves as the local decoding algorithm, when restricted to πΌ. A local list-decoding algorithm outputs a list of oracles, such that each valid codeword within the given radius is computed by some oracle in the output list. Definition 3.2.6 (Local list-decoding). A code π ⊆ Σπ is (β1 , β2 , π, πΏ, π, π)-locally listdecodable if there exists a randomized algorithm π with oracle access to a received word r ∈ Σπ that outputs a list π1 , . . . , ππΏ : [π] → Σ of randomized oracles with oracle access to r, such that 1. ππ queries at most β1 symbols of r; 2. for each π ∈ [πΏ], ππ queries at most β2 symbols of r; 3. with probability at least 1 − π, the following holds: if there is c ∈ π with πΏ(c, r) ≤ π , then there is some π ∈ [πΏ] such that Pr[π π (π) = ππ ] ≥ 1 − π for every π ∈ [π]. 33 3.3 Local testing and robust testing In this section, we formally define the model of testing. Since we will be solely interested in the testing of linear codes, we only present the definition of local testing in the context of linear codes. We also define the notions of soundness and robustness, and prove some simple relationships between the two. Definition 3.3.1 (Local testing). A β-local tester for a code π ⊆ {π → Fπ } is a randomized algorithm π― with oracle access to a received word π : π → Fπ , which randomly samples β (π, π ) according to some distribution π on (π β , 2Fπ ), with π = (π1 , . . . , πβ ) ∈ π β and π ⊆ Fβπ and accepts if and only if π |π , (π (π1 ), . . . , π (πβ )) ∈ π . The tester is π-sound if π― accepts π ∈ π with probability one, and rejects π ∈ / π with probability at least π · πΏ(π, π). The tester is πΌ-robust if E(π,π )←π [πΏ (π |π , π )] ≥ πΌ · πΏ(π, π). Proposition 3.3.2. Let π― be an β-local tester for π. If π― is π-sound, then π― is (π/β)-robust. If π― is πΌ-robust, then π― is πΌ-sound. Proof. Suppose π― is π-sound. Observe that if π |π ∈ / π , then πΏ(π |π , π ) ≥ 1/β. Therefore, E(π,π )←π [πΏ (π |π , π )] = E(π,π )←π [πΏ (π |π , π ) | π |π ∈ / π]· ≥ (1/β) · Pr (π,π )←π Pr (π,π )←π [π |π ∈ / π] [π |π ∈ / π] (3.1) (3.2) ≥ (1/β) · π · πΏ(π, π). (3.3) Now suppose π― is πΌ-robust. Then Pr (π,π )←π [π |π ∈ / π ] ≥ E(π,π )←π [πΏ (π |π , π ) | π |π ∈ / π]· Pr (π,π )←π [π |π ∈ / π] (3.4) = E(π,π )←π [πΏ (π |π , π )] (3.5) ≥ πΌ · πΏ(π, π). (3.6) 34 Proposition 3.3.3. Let π―π be an β1 -local tester for π with distribution ππ , and suppose for every (π, π ) in the support of π, π has an β2 -local tester π―π with distribution ππ , and β2 ≤ β1 . 1. If π―π is πΌ1 -robust and π―π is πΌ2 -robust for every (π, π ) in the support of ππ , then π has an β2 -local tester that is (πΌ1 · πΌ2 )-robust. 2. If π―π is πΌ-robust and π―π is π-sound for every (π, π ) in the support of ππ , then π has an β2 -local tester that is (πΌ · π)-sound. Proof. Let π be the following distribution: choose (π, π ) ← ππ and then choose and output (π ′ , π ′ ) ← ππ . Let π― be the β2 -tester for π with distribution π. 1. If π―π is πΌ1 -robust and π―π is πΌ2 -robust, then [οΈ ]οΈ E(π′ ,π ′ )←π [πΏ (π |π′ , π ′ )] = E(π,π )←ππ E(π′ ,π ′ )←ππ [πΏ (π |π′ , π ′ )] (3.7) ≥ πΌ2 · E(π,π )←ππ [πΏ (π |π , π )] (3.8) = πΌ1 · πΌ2 · πΏ(π, π). (3.9) 2. If π―π is πΌ-robust and π―π is π-sound, then Pr (π ′ ,π ′ )←π 3.4 [οΈ ′ [π |π′ ∈ / π ] = E(π,π )←ππ Pr (π ′ ,π ′ )←ππ ]οΈ [π |π′ ∈ /π ] ′ (3.10) ≥ π · E(π,π )←ππ [πΏ (π |π , π )] (3.11) ≥ πΌ · π · πΏ(π, π). (3.12) Codes In this section, we present two of the most ubiquitous linear codes: the Reed-Solomon code and the Reed-Muller code. We will directly use the Reed-Solomon code in our constructions, 35 whereas the Reed-Muller code serves as a benchmark for comparison. 3.4.1 Reed-Solomon code The Reed-Solomon code consists of evaluations of low-degree univariate polynomials over a finite field Fπ . Definition 3.4.1. Let π be a prime power, and let π ∈ N. The Reed-Solomon code RS(π, π) of degree π over Fπ is the code RS(π, π) , {π : Fπ → Fπ | deg(π ) ≤ π}. Proposition 3.4.2. If π < π, then the Reed-Solomon code RS(π, π) has distance 1 − rate π π and π+1 . π In [GS99], Guruswami and Sudan showed that the Reed-Solomon code of distance πΏ > 0 √ can be efficiently list-decoded up to the Johnson radius 1− 1 − πΏ. We will use this algorithm as a subroutine in our list-decoding and local list-decoding algorithms for the lifted ReedSolomon code in Sections 7.1.3 and 7.1.4, respectively. Theorem 3.4.3 (Guruswami-Sudan list-decoding [GS99]). For every πΏ, π > 0, there is a polynomial time algorithm taking as input a function π : Fπ → Fπ and outputs a list β of size √ |β| = π(1/π2 ) satisfying the following: if π ∈ RS(π, (1 − πΏ)π) and πΏ(π, π) < 1 − 1 − πΏ − π, then π ∈ β. 3.4.2 Reed-Muller code The Reed-Muller code consists of evaluations of low-degree multivariate polynomials over a finite field Fπ . Definition 3.4.4. Let π be a prime power, and let π, π ∈ N. The π-variate Reed-Muller code RM(π, π, π) of degree π over Fπ is the code RM(π, π, π) , {π : Fπ π → Fπ | deg(π ) ≤ π}. Remark 3.4.5. It follows immediately from definitions that RM(π, π, 1) = RS(π, π). Proposition 3.4.6. If π < π, then the Reed-Muller code RM(π, π, π) has distance 1 − ππ and (π+π) rate πππ . 36 Reed-Muller codes play a prominent role in complexity theory due to their locality features. Reed-Muller codes are locally decodable/correctable, and are also list-decodable and locally list-decodable up to the Johnson radius [PW04, STV99, BK09]. Moreover, Reed-Muller codes are testable, even robustly [RS96, ALM+ 98, FS95, AS03, RS97, MR06, AKK+ 05, KR06, JPRZ09, BKS+ 10, HSS11]. Note that if we want a family of Reed-Muller codes with positive distance πΏ > 0, we need the degree π = (1 − πΏ)π. The rate is therefore roughly (1−πΏ)π π! < 1 . π! In particular, the rate never exceeds 12 . The multiplicity codes of [KSY14] were the first locally correctable codes with rate close to 1. The highlight of our work is the construction of codes with the same distance as that of the Reed-Muller code and same locality features (decodability, correctability, list-decodability, and testability), but with rate close to 1. 37 38 Chapter 4 Affine-invariance 4.1 Affine-invariant Codes In [KS08], Kaufman and Sudan examine the role of symmetry in algebraic property testing (testing of linear codes). The type of symmetry they focus on is affine-invariance. Viewing codewords as functions π : Fπ π → Fπ , where Fπ is an extension field of Fπ , one can permute π the symbols of π by applying a permutation π : Fπ π → Fπ to the domain, resulting in a new word π β π. Affine-invariance is simply the property of being closed under applying affine permutations to the domain, i.e. if π is a codeword, then π β π΄ is a codeword for any affine permutation π΄. Definition 4.1.1 (Affine-invariance). A code π ⊆ {Fπ π → Fπ } is affine-invariant if π βπ΄ ∈ π π whenever π ∈ π and π΄ : Fπ π → Fπ is an affine permutation. Affine-invariance appears to be the right abstraction of the low-degree property. In fact, when π = π is prime, then the only affine-invariant codes are the Reed-Muller codes [KS08]. However, when π is a prime power or if π is a power of π, then there is a richer collection of affine-invariant codes. Affine-invariant codes are particularly appealing because they possess rich structure. The main structural feature of affine-invariant codes we will use is that they are spanned by monomials. 39 π Definition 4.1.2 (Degree set). A code π ⊆ {Fπ π → Fπ } has a degree set Deg(π) ⊆ JπK if π = {π : Fπ π → Fπ | supp(π ) ⊆ Deg(π)}. The degree set Deg(π) is π-shadow-closed if, whenever d ∈ Deg(π) and e ≤π d, we have e ∈ Deg(π). Proposition 4.1.3 ([KS08]). If π ⊆ {Fπ π → Fπ } is linear affine-invariant, where Fπ has characteristic π, then π has a π-shadow-closed degree set. Proposition 4.1.4 ([BGM+ 11a]). If π ⊆ {Fππ → Fπ } is a Fπ -linear affine-invariant code, then dimFπ (π) = | Deg(π)|. 4.2 Equivalence of Invariance under Affine Transformations and Permutations In their work initiating the study of the testability of affine-invariant properties (codes), Kaufman and Sudan [KS07] studied properties closed under general affine transformations and not just permutations. While affine transformations are nicer to work with when available, they are not mathematical elegant (they do not form a group under composition). Furthermore in the case of codes they also do not preserve the code — they only show that every codeword stays in the code after the transformation. Among other negative features affine transformations do not even preserve the weight of non-zero codewords, which can lead to some rude surprises. Here we patch the gap by showing that families closed under affine permutations are also closed under affine transformations. So one can assume the latter, without restricting the class of properties under consideration. We note that such a statement was proved in [BGM+ 11b] for the case of univariate functions. Unfortunately their proof does not extend to the multivariate setting and forces us to rework many steps from [KS08]. Theorem 4.2.1. If π ⊆ {Fπ π → Fπ } is an Fπ -linear code invariant under affine permutations, then π is invariant under all affine transformations. 40 The central lemma (Lemma 4.2.2) that we prove is that every non-trivial function can be split into more basic ones. This leads to a proof of Theorem 4.2.1 fairly easily. We first start with the notion of a basic function. For π = π π , let Tr : Fπ → Fπ denote the trace function Tr(π₯) = π₯ + π₯π + · · · + π₯π π−1 . We say that π : Fπ π → Fπ is a basic function if π (X) = Tr(πXd ) for some d ∈ JπKπ . For π ⊆ {Fπ π → Fπ } and π ∈ π we say π can be split (in π) if there exist functions π, β ∈ π such that π = π + β and supp(π), supp(β) ( supp(π ). Lemma 4.2.2. If π ⊆ {Fπ π → Fπ } is an Fπ -linear code invariant under affine permutations, then for every function π ∈ π, π is either basic or π can be split. We first prove Theorem 4.2.1 from Lemma 4.2.2. Proof of Theorem 4.2.1. First we assert that it suffices to prove that for every function π ∈ π the function π˜ = π (π1 , . . . , ππ−1 , 0) is also in π. To see this, consider π ∈ π and π΄ : Fπ π → π π Fπ π which is not a permutation. Then there exists affine permutations π΅, πΆ : Fπ → Fπ such that π΄(X) = π΅(πΆ(X)1 , . . . , πΆ(X)π , 0, . . . , 0) where π < π is the dimension of the image of π΄. By closure under affine permutations, it follows π βπ΅ ∈ π. Applying the assertion above π−π times we have that π ′ (X) , (π β π΅)(π1 , . . . , ππ , 0, . . . , 0) is also in π. Finally π β π΄ = π ′ β πΆ is also in π. So we turn to proving that for every π ∈ π the function π˜ = π (π1 , . . . , ππ−1 , 0) is also in π. d d ˜ ˜(X) = ∑οΈ π X . Notice π d d|ππ =0 πd X . Writing π = π + π1 , we use d ∑οΈ Lemma 4.2.2 to split π till we express it as a sum of basic functions π = π π=1 ππ , where Let π (X) = ∑οΈ each ππ is a basic function in π. Note that for every ππ , we have supp(ππ ) ⊆ supp(π˜) or supp(ππ ) ⊆ supp(π1 ) (since the trace preserves ππ = 0). By reordering the ππ ’s assume the ∑οΈ first π ππ ’s have their support in the support of π˜. Then we have π˜ = π π=1 ππ ∈ π. We thus turn to the proof of Lemma 4.2.2. We prove the lemma in a sequence of cases, based on the kind of monomials that π has in its support. We say that d and e are equivalent (modulo π), denoted d ≡π e if there exists a π such that for every π, ππ = π π ππ mod* π. The following proposition is immediate from previous works (see, for example, [BGM+ 11b]). We include a proof for completeness. 41 Proposition 4.2.3. If every pair d, e in the support of π : Fπ π → Fπ are equivalent, then π is a basic function. Proof. We first note that since the Tr : Fπ → Fπ is a (π/π)-to-one function, we have in particular that for every π½ ∈ Fπ there is an πΌ ∈ Fπ such that Tr(πΌ) = π½. As an immediate consequence we have that every function π : Fπ π → Fπ can be expressed Tr βπ where π : Fπ π → Fπ . Finally we note that we can view π as an element of Fπ [X], to conclude that π = Tr βπ for some polynomial π. Now fix π : Fπ π → Fπ all of whose monomials are equivalent. By the above we can express π = Tr βπ for some polynomial π. By inspection we can conclude that all monomials in the support of π are equivalent to the monomials in the support of π . Finally, using the fact * that Tr(πΌXd ) = Tr(πΌπ Xπd mod π ) we can assume that π is supported on a single monomial and so π = Tr(πXd ) for some π ∈ Fπ . So it suffices to show that every function that contains non-equivalent degrees in its support can be split. We first prove that functions with “non-weakly-equivalent” monomials can be split. We say that d and e are weakly equivalent if there exists a π such that for every π, ππ = π π ππ (mod π − 1). Lemma 4.2.4. If β± ⊆ {Fπ π → Fπ } is an Fπ -linear code invariant under affine permutations and π ∈ π contains a pair of non-weakly equivalent monomials in its support, then π can be split. Proof. Let d and e be two non weakly-equivalent monomials in the support of π . Fix π and ∑οΈ ∏οΈ ππ π consider the function ππ (X) = a∈(F* )π π−π π (π1 π1 , . . . , ππ ππ ). We claim that (1) the π π support of ππ is a subset of the support of π , (2) π π d is in the support of ππ , (3) d is in the support of ππ only if for every π ππ = π π ππ (mod π − 1) and in particular (4) e is not in the support of ππ . Now let π = π(d) be the smallest positive integer such that π π ππ = ππ mod* π for every π. ∑οΈπ−1 Now consider the function π = π=0 ππ . We have that π ∈ π since it is an Fπ -linear 42 combination of linear transforms of functions in π. By the claims about the ππ ’s we also have that d is in the support of π, the support of π is contained in the support of π and e is not in the support of π . Expressing π = π + (π − π) we now have that π can be split. The remaining cases are those where some of coordinates of d are zero or π − 1 for every d in the support of π . We deal with a special case of such functions next. Lemma 4.2.5. Let π be a linear affine-invariant code. Let π ∈ π be given by π (X, Y) = Tr(Yd π(X)) where every variable in π(X) has degree in {0, π − 1} in every monomial, and d is arbitrary. Further, let degree of π(X) be π(π − 1). Then for every 0 ≤ π ≤ π and for every π ∈ Fπ , the function (π1 · · · ππ )π−1 Tr(πYd ) ∈ π. Note that in particular the lemma above implies that such π ’s can be split into basic functions. Proof. We prove the lemma by a triple induction, first on π, then on π, and then on the number of monomials in π. The base case is π = 0 and that is trivial. So we consider general π > 0. First we consider the case π < π. Assume without loss of generality that the monomial (π1 · · · ππ )π−1 is in the support of π and write π = π0 +π1π−1 π1 where π0 , π1 do not depend on π1 . Note that π1 ΜΈ= 0 and deg(π1 ) = (π − 1)(π − 1). We will prove that − Tr(Yd π1 (X)) ∈ π ∑οΈ and this will enable us to apply the inductive hypothesis to π1 . Let π(X, Y) = π½∈Fπ π (π1 + π½, π2 , . . . , ππ , Y). By construction π ∈ π. By linearity of the Trace we have β β π = Tr βYd β ββ ∑οΈ π0 + (π1 + π½)π−1 π1 β β = Tr(Yd (−π1 (X))), π½∈Fπ where the second equality follows from the fact that ∑οΈ π½∈Fπ (π§ + π½)π−1 = −1. Thus we can now use induction to claim (π1 . . . ππ )π−1 Tr(πYd ) ∈ π. Finally we consider the case π = π. Now note that since the case π < π is known, we can assume without loss of generality that π is homogeneous (else we can subtract off the lower degree terms). Now if π = π there is nothing to be proved since π is just a single monomial. So 43 assume π < π. Also if π has only one monomial then there is nothing to be proved, so assume π has at least two monomials. In particular assume π is supported on some monomial that depends on π1 and some monomial that does not depend on π1 . Furthermore, assume without loss of generality that a monomial depending on π1 does not depend on π2 . Write π = π1π−1 π1 + π2π−1 π2 + (π1 π2 )π−1 π3 + π4 where the ππ ’s don’t depend on π1 or π2 . By assumption on the monomials of π we have that π1 ΜΈ= 0 and at least one of π2 , π3 , π4 ΜΈ= 0. Now consider the affine transform π΄ that sends π1 to π1 + π2 and preserves all other ππ ’s. We have π = (οΈ )οΈ π β π΄ = Tr Yd (π1π−1 π1 + π2π−1 (π1 + π2 ) + (π1 π2 )π−1 π3 + π4 + π) where the π1 -degree of ∑οΈ every monomial in π is in [π − 2]. Now consider π ′ (x, y) = πΌ∈F* π(πΌπ₯1 , π₯2 , . . . , π₯π , y). The π ′ terms of π vanish in π leaving (οΈ (οΈ )οΈ)οΈ π ′ = −(π β π΄ − π) = Tr Yd −π1π−1 π1 − π2π−1 (π1 + π2 ) − (π1 π2 )π−1 π3 − π4 . Finally we consider the function π˜ = π + π ′ = Tr(Yd (−π2π−1 π1 )) which is a function in π of degree π(π − 1) supported on a smaller number of monomials than π , so by applying the inductive hypothesis to π˜ we have that π contains the monomial (π1 · · · ππ )π−1 . The following lemma converts the above into the final piece needed to prove Lemma 4.2.2. Lemma 4.2.6. If β± ⊆ {Fπ π → Fπ } is an Fπ -linear code invariant under affine permutations and all monomials in π ∈ π are weakly equivalent, then π can be split. Proof. First we describe the structure of a function π : Fπ π → Fπ that consists only of weakly equivalent monomials. First we note that the π variables can be separated into those in which every monomial has degree in [π − 2] and those in which every monomial has degree in {0, π − 1} (since every monomial is weakly equivalent). Let us denote by X the variables in which the monomials of π have degree in {0, π − 1} and Y be the remaining variables. Now consider some monomial of the form π = πXe Yd in π . Since π maps to Fπ we must π π have that the coefficient of (Xe Yd )π is ππ . Furthermore, we have every other monomial π ′ π ′ in the support of π is of the form π′ Yπ d Xπ₯e . Thus π can be written as Tr(Yd π(X)) where π−1 π(π1 , . . . , ππ ) = π˜(π1π−1 , . . . , ππ ). But, by Lemma 4.2.5, such an π can be split. 44 Proof of Lemma 4.2.2. If π contains a pair of non-weakly equivalent monomials then π can be split by Lemma 4.2.4. If not, then π is either basic or, by Lemma 4.2.6 is can be split. We also prove an easy consequence of Lemma 4.2.2. d Lemma 4.2.7. Let π ⊆ {Fπ π → Fπ } be affine invariant. If d ∈ Deg(π), then Tr(πX ) ∈ π for all π ∈ Fπ . Proof. We first claim that Lemma 4.2.2 implies that there exists π½ ∈ Fπ such that Tr(π½Xd ) is a non-zero function in π. To verify this, consider a “minimal” function (supported on fewest monomials) π ∈ π with d ∈ supp(π ). Since π can’t be split in π (by minimality), by Lemma 4.2.2 π must be basic and so equals (by definition of being basic) Tr(π½Xd ). Now let π = π(d) be the smallest positive integer such that π π d mod* π = d. If π = π π , note that π divides π and so one can write Tr : Fπ → Fπ as Tr1 β Tr2 where Tr1 : Fππ → Fπ is the function Tr1 (π§) = π§ + π§ π + · · · + π§ π π π−1 and Tr2 : Fπ → Fππ is the function Tr2 (π§) = π π§ + π§ π + · · · + π§ π/π . (Both Tr1 and Tr2 are trace functions mapping the domain to the range.) It follows that Tr(π½Xd ) = Tr1 (Tr2 (π½)Xd ). We first claim that Tr1 (π Xd ) ∈ π for every π ∈ Fππ . Let π = { ∑οΈ πΌ∈(F*π )π ππΌ ·πΌd | ππΌ ∈ Fπ }. We note that by linearity and affine-invariance of π, we have that Tr1 (Tr2 (π½) · πXd ) ∈ π for every π ∈ π. By definition π is closed under addition and multiplication and so is a π subfield of Fπ . In fact, since every π ∈ π satisfies π π = π (which follows from the fact π that πΌd = πΌπ d ), we have that π ⊆ Fππ . It remains to show π = Fππ . Suppose it is a strict subfield of size π π for π < π. Consider πΎ ππ for πΎ ∈ Fπ and π ∈ [π]. Since πΎ ππ ∈ π, π we have that πΎ ππ π = πΎ ππ for every πΎ ∈ Fπ and so we get πππ conclude that Xπ πd ππ π = ππ (mod πππ − ππ ). We = Xd (mod Xπ − X) which contradicts the minimality of π = π(d). We conclude that π = Fππ . Since Tr2 (π½) ∈ F*ππ , we conclude that the set of coefficients π such that Tr1 (π Xd ) ∈ π is all of Fππ as desired. Finally consider any π ∈ Fπ . since Tr2 (π) ∈ Fππ , we have that Tr1 (Tr2 (π)Xd ) ∈ π (from the previous paragraph), and so Tr(πXd ) = Tr1 (Tr2 (π)Xd ) ∈ π 45 46 Chapter 5 Lifting Codes 5.1 The Lift Operator First defined and used in [BSMSS11] to prove the existence of certain codes that are not locally testable, the lift operator takes short codes and creates longer codes from them. The original lift operator took codes defined over the domain Fπ and “lifted” them to codes defined over the domain Fππ . We generalize the definition, allowing one to lift codes over Fπ‘π to codes over Fπ π , for any π ≥ π‘ (in particular, π‘ does not need to divide π). Definition 5.1.1 (Lift). Let π ⊆ {Fπ‘π → Fπ } be a linear affine-invariant code. Let π ≥ π‘ be an integer. The π-dimensional lift of π, denoted by π π‘βπ , is the code β {οΈ }οΈ β π |π΄ ∈ π for every π‘-dimensional affine subspace π΄ ⊆ Fπ π π‘βπ , π : Fπ . π → Fπ π Proposition 5.1.2. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant, and let π ≥ π‘. If π ∈ π π‘βπ and π΄ : Fπ‘π → Fπ π is an affine map, then π β π΄ ∈ π. Proof. By Proposition B.1.1, there exists an invertible affine map π΄′′ : Fπ‘π → Fπ π and a linear map π΄′ : Fπ‘π → Fπ‘π such that π΄ = π΄′′ β π΄′ . By the definition of lift, π , π β π΄′′ ∈ π. Since π 47 is affine-invariant, it follows from Theorem 4.2.1 that π β π΄′ ∈ π. Therefore, π β π΄ = π β (π΄′′ β π΄′ ) = (π β π΄′′ ) β π΄′ = π β π΄′ ∈ π. 5.2 Algebraic and Combinatorial Properties We begin by exploring the algebraic and combinatorial (as opposed to algorithmic) properties of codes lifted from linear affine-invariant codes. First, we show that the lifting operator is a natural algebraic operation, preserving linearity and affine-invariance and composing well. We then present the degree set of a lifted code in terms of the degree set of the base code. Finally, we conclude by showing that lifting preserves distance. 5.2.1 Algebraic Properties Proposition 5.2.1. Let π‘ ≤ π be integers. If π ⊆ {Fπ‘π → Fπ } is linear affine-invariant, then π π‘βπ is linear affine-invariant. Proof. First, we show that π π‘βπ is linear. Let π, π ∈ π π‘βπ and let πΌ ∈ Fπ . If π΄ ⊆ Fπ π is a π‘-dimensional affine subspace, then, since π is linear, (πΌπ + π)|π΄ = πΌ · (π |π΄ ) + (π|π΄ ) ∈ π Next, we show that π π‘βπ is affine-invariant. Let π (X) ∈ π π‘βπ , and let M ∈ Fπ×π and π π×π‘ c ∈ Fπ and b ∈ Fπ π . Let π(X) = π (MX + c). Let Y = (π1 , . . . , ππ‘ ), let A ∈ Fπ π . Then π(AX + b) = π (M(AX + b) + c) = π ((MA)X + (Mb + c)) ∈ π. Since A, b were arbitrary, this implies that π ∈ π π‘βπ . Since M, c were arbitrary, this implies that π π‘βπ is affine-invariant. 48 Proposition 5.2.2 (Composition of lift). Let π‘ ≤ π ≤ π be integers. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant. Then (οΈ )οΈπβπ π π‘βπ = π π‘βπ . Proof. This follows immediately from the fact that choosing a π‘-dimensional affine subspace π΄ ⊆ Fππ is equivalent to first choosing an π-dimensional affine subspace π΅ ⊆ Fππ and then choosing a π‘-dimensional affine subspace π΄ ⊆ π΅. Proposition 5.2.3 (Degree set of lift). Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant with degree set Deg(π). If π ≥ π‘ is an integer, then the lift π π‘βπ has degree set Deg(π π‘βπ ) = {d ∈ JπKπ | E ≤π d =⇒ (βE*1 β, . . . , βE*π‘ β) mod* π ∈ Deg(π)} where E has rows [π] and columns Jπ‘ + 1K. Proof. Let π ∈ π π‘βπ . Let π· = Deg(π π‘βπ ). Let X = (π1 , . . . , ππ ) and let Y = (π1 , . . . , ππ‘ ). ∑οΈ d π×π‘ Write π (X) = and b ∈ Fπ π , it follows from d∈π· πd · X . For any matrix A ∈ Fπ Proposition A.1.4 that (οΈ )οΈ π‘ π‘ π ∏οΈ ∑οΈ (οΈ d )οΈ ∏οΈ ∏οΈ βE β π πππ π0 πππππ · ππ *π π (AY + b) = πd · E π=1 π=1 π=1 d∈π· E≤ d ∑οΈ π Since π (AY + b) ∈ π for any A, b, and π is affine-invariant, it follows by Proposition 4.1.3 that for every d ∈ π· and E ≤π d, with columns Jπ‘ + 1K, it holds that the monomial ∏οΈπ‘ βE*π β ∈ π, hence (βE*1 β, . . . , βE*π‘ β) mod* π ∈ Deg(π). Conversely, if d satisfies that π=1 ππ for every E ≤π d with columns Jπ‘ + 1K that (βE*1 β, . . . , βE*π‘ β) mod* π ∈ Deg(π), then it is easy to see that d ∈ π· by considering π (X) = Xd . 49 5.2.2 Distance of Lifted Codes Proposition 5.2.4 (Distance of lift). Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant, and let π ≥ π‘. The following hold: 1. πΏ(π π‘βπ ) ≤ πΏ(π); 2. πΏ(π π‘βπ ) ≥ πΏ(π) − π −π‘ ; 3. if π ∈ {2, 3} and πΏ(π) > π −π‘ , then πΏ(π π‘βπ ) = πΏ(π). We prove several lemmas, which in turn will help us prove Proposition 5.2.4. The first lemma proves Proposition 5.2.4 (1). Lemma 5.2.5. If π ⊆ {Fπ‘π → Fπ } is linear affine-invariant, and π ≥ π‘, then πΏ(π π‘βπ ) ≤ πΏ(π). Proof. The case π = π‘ is trivial, so assume π > π‘. By Proposition 5.2.2 and induction, it suffices to consider the case π = π‘ + 1. Let π ∈ π and let πΏ , πΏ(π, 0). Let X = (π1 , . . . , ππ‘ ). → Fπ defined by π(X, ππ‘+1 ) = π (X). Clearly, we have Consider the function π : Fπ‘+1 π be an affine map. Let πΏ(π, 0) = πΏ. We claim that π ∈ π π‘β(π‘+1) . Let π΄ : Fπ‘π → Fπ‘+1 π π΄′ : Fπ‘π → Fπ‘π be the affine map given by the projection of π΄ onto its first π‘ coordinates, i.e. π΄′ (X) = (π΄(X)1 , . . . , π΄(X)π‘ ). Since π ∈ π and π is affine-invariant, by Theorem 4.2.1 we have π β π΄′ ∈ π. Therefore, π β π΄ = π β π΄′ ∈ π. Since π΄ was arbitrary, this shows that π ∈ π π‘β(π‘+1) . The next lemma proves Proposition 5.2.4 (2), and uses a simple probabilistic argument. Lemma 5.2.6. If π ⊆ {Fπ‘π → Fπ } is linear affine-invariant, and π ≥ π‘, then πΏ(π π‘βπ ) ≥ πΏ(π) − π −π‘ . π‘ π Proof. Let π, π ∈ π π‘βπ be distinct. Let x ∈ Fπ π such that π (x) ΜΈ= π(x). Let π΄ : Fπ → Fπ be a random affine map such that π΄(0) = x, so that π β π΄, π β π΄ ∈ π (by Proposition 5.1.2) and are distinct. Then πΏ(π) ≤ Eπ΄ [πΏ (π β π΄, π β π΄)] (Proposition B.1.3) ≤ πΏ(π, π) + π −π‘ . 50 (5.1) (5.2) The next two lemmas will help in proving Proposition 5.2.4, for the cases π = 2 and π = 3 respectively. Lemma 5.2.7. For all π ≥ 2, if πΏ > 1 2π−1 and π : Fπ 2 → F2 such that 0 < πΏ(π, 0) < πΏ, then there exists a hyperplane π» ⊂ Fπ 2 such that 0 < πΏ(π |π» , 0) < πΏ. Proof. We proceed by induction on π. The base case π = 2 is straightforward to verify. Now suppose π > 2 and our assertion holds for π − 1. Let π»0 , π»1 be the affine subspaces given by ππ = 0 and ππ = 1, respectively. For π ∈ {0, 1}, let πΏπ , πΏ(π |π»π , 0). Note that πΏ > πΏ(π, 0) = (πΏ0 + πΏ1 )/2. If both πΏ0 , πΏ1 > 0, then by averaging we have 0 < πΏπ < πΏ, for some π ∈ {0, 1}, and so π» = π»π does the job. Otherwise, suppose without loss of generality that πΏ1 = 0. Note that 0 < πΏ0 < 2πΏ and 2πΏ > 1 . 2π−2 Thus, by the induction hypothesis, there exists an (π − 2)-dimensional affine subspace π»0′ ⊆ π»0 such that 0 < πΏ(π |π»0′ , 0) < 2πΏ. Let π»1′ , π»0′ + eπ , where eπ is the π-th standard basis vector. Note that πΏ(π |π»1′ , 0) = 0. Let ′ π» = π»0′ ∪ π»1′ . Then π» is an (π − 1)-dimensional subspace of Fπ 2 (spanned by π»0 and eπ ) )οΈ (οΈ such that 0 < πΏ(π |π» , 0) = πΏ(π |π»0′ , 0) + πΏ(π |π»1′ , 0) /2 < πΏ. Lemma 5.2.8. For all π ≥ 2, if π : Fπ 3 → F3 and πΏ(π, 0) ≥ 1 , 3π−1 then there exists a hyperplane π» ⊆ Fπ 3 such that 0 < πΏ(π |π» , 0) ≤ πΏ(π, 0). Proof. Let πΏ , πΏ(π, 0). We proceed by induction on π.For the base case, π = 2, πΏ ≥ 13 . Suppose π : F23 → F3 and for each π ∈ F3 , let ππ be the restriction of π to the hyperplane defined by π2 = π. If ππ ΜΈ= 0 for every π ∈ F3 , then by averaging there is some π ∈ F3 for which 0 < πΏ(ππ , 0) ≤ πΏ. Otherwise, without loss of generality, suppose π2 = 0. Further, without loss of generality, suppose π0 ΜΈ= 0 and π (0, 0) ΜΈ= 0. Now, if πΏ ≥ 32 , then the line π» = {(π₯, π¦) ∈ F32 | π₯ = 0} does the job, since 0 < πΏ(π |π» , 0) ≤ 2 3 ≤ πΏ. If πΏ < 23 , then there must exist some π, π ∈ F3 and π ∈ {0, 1} such that π (π, π) ΜΈ= 0 and π (π, 1 − π) = 0. Then the line π» = {(π, π), (π, 1 − π), (2π − π, 2)} does the job, since 0 < πΏ(π |π» , 0) = 1 3 ≤ πΏ. Now suppose π > 2 and the assertion holds for π − 1. For π ∈ F3 , let π»π be the hyperplane cut out by ππ = π and let πΏπ , πΏ(π |π»π , 0) for each π ∈ F3 . Then πΏ0 + πΏ1 + πΏ3 = 3πΏ. 51 If πΏπ > 0 for all π ∈ F3 , then by simple averaging, for some π ∈ F3 , we have 0 < πΏπ ≤ πΏ, so assume without loss of generality that πΏ2 = 0 and πΏ0 ≥ πΏ1 . First, suppose πΏ0 ≥ 1 . 3π−2 Then, by the inductive hypothesis, there exists an (π − 2)- dimensional affine subspace π» (0) ⊂ π»0 such that 0 < πΏ(π |π» , 0) ≤ πΏ0 . Suppose π» (0) is defined ∑οΈ π+1 . For by the linear equations π π=1 ππ ππ − π0 = 0 and ππ = 0 for some (π0 , . . . , ππ ) ∈ Fπ ∑οΈ (π) each π, π ∈ F3 , let π»π ⊂ π»π be the affine subspace defined by π π=1 ππ ππ −π0 = π and ππ = π. (οΈ )οΈ (1) (2) By averaging, for some π ∈ F3 , we have πΏ π |π» (1) , 0 ≤ πΏ1 . Take π» , π» (0) ∪ π»π ∪ π»2π . π Then 0 < πΏ(π |π» , 0) = Now, suppose πΏ0 < (οΈ )οΈ (οΈ )οΈ (οΈ )οΈ πΏ π |π» (0) , 0 + πΏ π |π» (1) , 0 + πΏ π |π» (2) , 0 1 , 3π−2 π 2π 3 so πΏ0 , πΏ1 ≤ 2 . 3π−1 subspace π» (0) ⊂ π»0 such that πΏ(π |π» (0) , 0) = ≤ πΏ0 + πΏ1 = πΏ. 3 There exists an (π − 2)-dimensional affine 1 . 3π−1 To see this, let a, b ∈ π»0 such that π (a), π (b) ΜΈ= 0, and suppose ππ ΜΈ= ππ , for some π ∈ [π]. Then take π» (0) to be the subspace (π) defined by ππ = ππ and ππ = 0. For π, π ∈ F3 , let π»π be the (π − 2)-dimensional subspace defined by ππ = ππ + π and ππ = π. Since πΏ1 ≤ Then, taking π» , π» (0) ∪ (1) π»π 0 < πΏ(π |π» , 0) = ∪ (2) π»2π , 2 , 3π−2 there is π ∈ F3 such that π |π» (1) = 0. π we have )οΈ (οΈ )οΈ (οΈ )οΈ (οΈ πΏ π |π» (0) , 0 + πΏ π |π» (1) , 0 + πΏ π |π» (2) , 0 π 3 2π = 1 < πΏ. 3π Now, we are ready to prove Proposition 5.2.4. Proof of Proposition 5.2.4. Parts 1 and 2 follow immediately from Lemmas 5.2.5 and 5.2.5, respectively, so we proceed with proving part 3. In light of part 1, it suffices to show that πΏ(π π‘βπ ) ≥ πΏ(π). We start with the π = 2 case. We proceed by induction on π − π‘. Indeed, the inductive (οΈ )οΈπ−1βπ step is straightforward since, by Proposition 5.2.2, we have π π‘βπ = π π‘βπ−1 , so by induction πΏ(π π‘βπ ) ≥ πΏ(π π‘βπ−1 ) ≥ πΏ(π). The main case is therefore the base case π = π‘ + 1. Suppose π ∈ π π‘βπ‘+1 such that 0 < πΏ(π, 0) < πΏ(π). By assumption, πΏ(π) > 52 1 , 2π‘ so we may apply Lemma 5.2.7 to conclude that there exists a hyperplane π» ⊂ Fπ 2 such that 9 < πΏ(π |π» , 0) < πΏ(π), contradicting the fact that π |π» ∈ π (since π ∈ π π‘βπ‘+1 ). Finally, we consider the π = 3 case. Again, we proceed by induction on π − π‘, and again the main case is the base case π = π‘ + 1. Suppose π ∈ π π‘βπ‘+1 such that 0 < πΏ(π, 0) < πΏ(π). If πΏ(π, 0) ≥ 1 , 3π−1 then, by Lemma 5.2.8, there exists a hyperplane π» ⊂ Fπ π such that 0 < πΏ(π |π» , 0) ≤ πΏ(π, 0) < πΏ(π), contradicting the fact that π |π» ∈ π. If πΏ(π, 0) < 1 , 3π−1 then there are at most two points a, b ∈ Fπ 3 such that π (a), π (b) ΜΈ= 0. Let π ∈ [π] such that ππ ΜΈ= ππ and let π» be the hyperplane defined by ππ = ππ . Then π |π» is nonzero only on a, so 0 < πΏ(π |π» ) = 5.3 1 3π−1 < πΏ(π), again contradicting the fact that π |π» ∈ π. Local Decoding and Correcting In this section, we explore some of the algorithmic decoding properties of lifted codes. In particular, for codes of distance πΏ > 0, we give simple algorithms for locally correcting up to πΏ/4 and then πΏ/2 fraction errors. We then show that all linear affine-invariant codes — in particular, lifted codes — have explicit interpolating sets, thereby showing that lifted codes are locally decodable as well. 5.3.1 Local Correcting up to 1/4 Distance The following algorithm is an abstraction of a simple algorithm for locally correcting ReedMuller codes of distance πΏ > 0 from πΏ/4 fraction errors. Theorem 5.3.1. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant with distance πΏ , πΏ(π), and (οΈ )οΈ let π ≥ π‘. Then, for every π > 0, the lifted code π π‘βπ is π π‘ , ( 14 − π)πΏ − π −π‘ , 12 − 2π -locally correctable. Proof. Let Corrπ be a correction algorithm for π, so that for every π : Fπ → Fπ that is πΏ/2close to some π ∈ π, Corrπ (π ) = π. The following algorithm is a local correction algorithm that achieves the desired parameters. 53 Local correction algorithm: Oracle access to received word π : Fπ π → Fπ . On input x ∈ Fπ π : 1. Choose uniform random affine map π΄ : Fπ‘π → Fπ π such that π΄(0) = x. 2. Set π , π β π΄. 3. Compute πΜοΈ , Corrπ (π ). 4. Output πΜοΈ(0). Analysis: Let π , ( 14 − π)πΏ. Fix a received word π : Fπ π → Fπ . Let π ∈ π be a codeword such that πΏ(π, π) < π − π −π‘ . Let π΄ : Fπ‘π → Fπ π be a random affine map such that π΄(0) = x. By Proposition B.1.3, the expected distance is Eπ΄ [πΏ (π β π΄, π β π΄)] ≤ πΏ(π, π) + π −π‘ < π , so, by Markov’s inequality, with probability at least 1 2 + 2π, we have πΏ(π β π΄, π β π΄) < 2πΏ . For such π΄, setting π , π β π΄, we have πΜοΈ = Corrπ (π ) = π β π΄, hence πΜοΈ(0) = π(π΄(0)) = π(x). 5.3.2 Local Correcting up to 1/2 Distance The following algorithm can locally correct lifted codes up to half the minimum distance. The basic idea is to decode along multiple lines and weight their opinions based on distance to the base code. This is a direct translation of the elegant line-weight local decoding algorithm for matching-vector codes [BET10] to the Reed-Muller code and lifted codes setting. Theorem 5.3.2. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant with distance πΏ , πΏ(π), and (οΈ )οΈ let π ≥ π‘. Then, for every π, π > 0, the lifted code π π‘βπ is π, ( 12 − π)πΏ − π −π‘ , π -locally correctable for π = π (ln(1/π)/π2 · π π‘ ). Proof. Let Corrπ be a correction algorithm for π, so that for every π : Fπ → Fπ that is πΏ/2close to some π ∈ π, Corrπ (π ) = π. The following algorithm is a local correction algorithm that achieves the desired parameters. 54 Local correction algorithm: Oracle access to received word π : Fπ π → Fπ . On input x ∈ Fπ π : ⌉οΈ ⌈οΈ 2 4 1. Let π = π2 ln π and choose affine maps π΄1 , . . . , π΄π ∈ Fπ π such that π΄π (0) = x for each π ∈ [π] independently and uniformly at random. 2. For each π ∈ [π]: (a) Set ππ , π β π΄π . (b) Compute π π , Corrπ (ππ ) and πΏπ , πΏ(ππ , π π ). (οΈ (c) Assign the value π π (0) a weight ππ , max 1 − 3. Set π , ∑οΈπ π=1 ππ . For every πΌ ∈ Fπ , let π€(πΌ) := 1 π )οΈ πΏπ ,0 πΏ/2 . ∑οΈ π:π π (0)=πΌ ππ . If there is an πΌ ∈ Fπ with π€(πΌ) > 21 , output πΌ, otherwise fail. Analysis: Let π , ( 12 −π)πΏ. Fix a received word π : Fπ π → Fπ . Let π ∈ π be a codeword such that πΏ(π, π) < π − π −π‘ . The query complexity follows from the fact that the algorithm queries π(ln(1/π)/π2 ) subspaces, each consisting of at most π π‘ points. Fix an input x ∈ Fπ π . We wish to show that, with probability at least 1 − π, the algorithm outputs π(x), i.e. π€(π(x)) > 12 . For each affine map π΄ : Fπ‘π → Fπ π , define the following: ππ΄ , πΏ(π β π΄, π β π΄) π π΄ , Corrπ (π β π΄) πΏπ΄ , πΏ(π β π΄, π π΄ ) (οΈ )οΈ πΏπ΄ ππ΄ , max 1 − ,0 πΏ/2 β§ βͺ β¨ππ΄ π π΄ = π β π΄ ππ΄ , βͺ β©0 π π΄ ΜΈ= π β π΄. Let π , Prπ΄ [π π΄ = π β π΄]. Note that if π π΄ = π β π΄, then πΏπ΄ = ππ΄ , otherwise πΏπ΄ ≥ πΏ − ππ΄ . Hence, if π π΄ = π β π΄, then ππ΄ ≥ 1 − ππ΄ , πΏ/2 otherwise ππ΄ ≤ 55 ππ΄ πΏ/2 − 1. Define πgood , Eπ΄ [ππ΄ | π π΄ = π β π΄] πbad , Eπ΄ [ππ΄ | π π΄ ΜΈ= π β π΄] πgood , Eπ΄ [ππ΄ | π π΄ = π β π΄] ≥ 1 − πbad , Eπ΄ [ππ΄ | π π΄ ΜΈ= π β π΄] ≤ πgood πΏ/2 πbad − 1. πΏ/2 Observe that Eπ΄ [ππ΄ ] ≤ 1 + (π − 1π )(π − 1) π ≤π Eπ΄ [ππ΄ ] = π · πgood Eπ΄ [ππ΄ ] = π · πgood + (1 − π) · πbad . We claim that π · πgood ≥ (1 − π) · πbad + 2π. (5.3) To see this, we start from (οΈ )οΈ 1 − π πΏ = π ≥ Eπ΄ [ππ΄ ] = π · πgood + (1 − π) · πbad . 2 Dividing by πΏ/2 yields 1 − 2π ≥ π · πgood πbad + (1 − π) · . πΏ/2 πΏ/2 Re-writing 1 − 2π on the left-hand side as π + (1 − π) − 2π and re-arranging, we get (οΈ )οΈ (οΈ )οΈ πgood πbad ≥ (1 − π) · − 1 + 2π. π· 1− πΏ/2 πΏ/2 The left-hand side is bounded from above by π · πgood while the right-hand side is bounded from below by (1 − π) · πbad + 2π, hence (5.3) follows. Consider the random affine maps π΄1 , . . . , π΄π chosen by the algorithm. Note that the ππ΄ 56 are defined such that π΄π contributes weight probability at least 1 − π, ππ΄π π to π€(π(x)), so it suffices to show that, with ∑οΈπ π 1 ∑οΈππ=1 π΄π > . 2 π=1 ππ΄π Each ππ΄ , ππ΄ ∈ [0, 1], so by Proposition 2.2.3, [οΈ β π β 1 ∑οΈ β Pr β ππ΄π − Eπ΄ [ππ΄ ] βπ π=1 [οΈ β π β 1 ∑οΈ β Pr β ππ΄π − Eπ΄ [ππ΄ ] βπ π=1 β ]οΈ β β ≤ 2 exp(−π2 π/2) ≤ π/2 β > π/2 β β ]οΈ β β ≤ 2 exp(−π2 π/2) ≤ π/2. β > π/2 β Therefore, by a union bound, with probability at least 1 − π we have ∑οΈπ π ∑οΈππ=1 π΄π π=1 ππ΄π Eπ΄ [ππ΄ ] − π/2 E[ππ΄ ] + π/2 π · πgood − π/2 = π · πgood + (1 − π) · πbad + π/2 (1 − π) · πbad + 3π/2 by (5.3) ≥ 2(1 − π) · πbad + 5π/2 1 . > 2 5.3.3 ≥ Local Decoding We show that all linear affine-invariant codes, in particular lifted codes, have explicit interpolating sets, which allows us to immediately translate the local correctability of lifted codes into local decodability. Finite field isomorphisms. Let Tr : Fππ → Fπ be the Fπ -linear trace map π§ β¦→ ∑οΈπ−1 π=0 π π§π . Let πΌ1 , . . . , πΌπ ∈ Fππ be linearly independent over Fπ and let π : Fππ → Fπ be the map π§ β¦→ (Tr(πΌ1 π§), . . . , Tr(πΌπ π§)). Since Tr is Fπ -linear, π is Fπ -linear, and in fact π is an isomorphism. Observe that π induces a Fπ -linear isomorphism π* : {Fπ π → Fπ } → {Fπ π → Fπ } defined by 57 π* (π ) = π β π. It is straightforward to verify that if π ⊆ {Fπ π → Fπ }, and π ⊆ Fπ π is an interpolating set for π* (π), then π(π) is an interpolating set for π. Theorem 5.3.3. Let π ⊆ {Fπ π → Fπ } be a nontrivial affine-invariant code with dimFπ (π) = π·. Let π ∈ Fππ be a generator, i.e. π has order π π − 1. Let π = {π, π 2 , . . . , π π· } ⊆ Fππ . Then π(π) ⊆ Fπ π is an interpolating set for π. * Proof. The map π induces a map π* : {Fπ π → Fπ } → {Fπ π → Fπ } defined by π (π ) = π β π. It suffices to show that π is an interpolating set for π ′ , π* (π). Observe that π ′ is affine-invariant over Fππ , and hence has a degree set Deg(π ′ ), by Proposition 4.1.3. By Proposition 4.1.4, dimFπ (π ′ ) = | Deg(π ′ )|, so suppose Deg(π ′ ) = {π1 , . . . , ππ· }. Every π ∈ π ′ ∑οΈ ππ is of the form π(π) = π· π=1 ππ π , where ππ ∈ Fπ π . By linearity, it suffices to show that if π ∈ π ′ is nonzero, then π(π§) ΜΈ= 0 for some π§ ∈ π. We have β‘ π π1 π π2 ··· π ππ· β€β‘ π1 β€ β‘ π(π) β€ β’ β₯ β₯β’ β₯ β’ β’ 2π1 β’ β₯ β’ 2 β₯ 2π2 2ππ· β₯ β’ β’π β₯ β₯ β’ π(π ) β₯ π ··· π π β’ β₯ β₯β’ 2β₯ = β’ β’ .. .. .. β₯ β’ .. β₯ β’ .. β₯ .. . β’ . . . β₯β’ . β₯ β’ . β₯ β£ β¦ β¦β£ β¦ β£ π(π π· ) π π·π1 π π·π2 · · · π π·ππ· ππ· and the leftmost matrix is invertible since it is a generalized Vandermonde matrix. Therefore, if π ΜΈ= 0, then the right-hand side, which is simply the vector of evaluations of π on π, is nonzero. Theorem 5.3.4. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant with distance πΏ , πΏ(π), and (οΈ )οΈ let π ≥ π‘. Then, for every π, π > 0, the lifted code π π‘βπ is π, ( 12 − π)πΏ − π −π‘ , π -locally decodable where π = π(ln(1/π)/π2 ). Proof. Follows immediately from Theorems 5.3.2 and 5.3.3 and Proposition 3.2.5. 58 5.4 5.4.1 Local Testing and Robust Testing Local Testing Many testing results may be viewed as robustness statements about properties. A characterization statement would say “an object π has a property π globally if and only if it has property π locally”. For example, degree π polynomials over Fπ of characteristic π are ⌉οΈ ⌈οΈ π+1 , characterized by the fact that, when restricted to π‘-dimensional subspaces, for π‘ = π−π/π they are degree π polynomials. This was proved in [KR06]. A testing statement would go further to say “if π does not have property π globally, then locally it often fails to have property π ”. For example, building on our previous example, a low-degree testing statement would say that if a polynomial over Fπ of characteristic π has degree greater than π, then on many π‘-dimensional subspaces, it has degree greater than π. This was also proved in [KR06]. A robust testing statement would go even further to say “if π is far from having property π globally, then locally its average distance from π is also far”. The lift operator is natural for many reasons, but primarily because it suggests such a natural test for lifted codes. The lifted code π π‘βπ , by construction, is locally characterized by the fact that any codeword, when restricted to π‘-dimensional subspaces, is a codeword of π. Along with the symmetry provided by affine-invariance, the work of Kaufman and Sudan [KS08] immediately imply that the natural π‘-dimensional test is Ω(π −2π‘ )-sound, i.e. if π∈ / π π‘βπ , then on at least Ω(π −2π‘ )-fraction of π‘-dimensional subspaces π΄, π |π΄ ∈ / π. Definition 5.4.1. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant, and let π‘ ≤ π ≤ π. The natural π-dimensional test for π π‘βπ is the π π -local tester with the following distribution: it π selects a uniformly random π-flat π΄ ⊆ Fπ π and accepts a function π : Fπ → Fπ if and only if π |π΄ ∈ π π‘βπ . Theorem 5.4.2 ([KS08]). Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant, and let π ≥ π‘. Then the natural π‘-dimensional test for the lifted code π π‘βπ is (π −2π‘ /2)-sound. 59 5.4.2 Robust Testing However, for our high-rate code constructions, we often take π to be constant and let π → ∞ to make our code longer. As such, a soundness of Ω(π −π(π‘) ) is insufficient as this quickly approaches zero. In fact, in Chapter 6, we prove that the natural 2π‘-dimensional (as opposed to π‘-dimensional) test is πΌ-robust, where πΌ is a polynomial in the distance of the code, but does not depend on π, π‘, or π. In particular, by Proposition 3.3.2, this implies that the 2π‘-dimensional test is πΌ-sound. Theorem 5.4.3. Let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant, and let π ≥ π‘. Then the natural (2π‘)-dimensional test for the lifted code π π‘βπ is πΏ(π)72 -robust. 2·1052 The proof of Theorem 5.4.3 is somewhat long and technical, so we defer it to Chapter 6. 60 Chapter 6 Robust Testing of Lifted Codes 6.1 Robustness of Lifted Codes In this section, we prove Theorem 6.1.18, which is simply a more precise restatement of Theorem 1.2.2. That is, we prove that the natural 2π‘-dimensional test for the π-dimensional lift of a π‘-dimensional code over Fπ is πΌ-robust, where πΌ depends only on the distance of the code and not on π‘ or π or the field size. Our approach is a standard one — we first analyze the test for low-dimensional settings (Theorem 6.1.4), and then use a general projection argument (“bootstrapping”) to get an analysis for all dimensions (Theorem 6.1.18). 6.1.1 Preliminaries We begin by presenting some basic definitions and results on robust testing. We define the robustness of a lifted code, specializing the definition to robustness with respect to subspace testers. We include the dimension of the testing subspace as a parameter in the robustness since this will be convenient later. Definition 6.1.1. Let π‘ ≤ π ≤ π. The code π π‘βπ is (πΌ, π)-robust if, for every π : Fπ π → Fπ , )οΈ [οΈ (οΈ )οΈ]οΈ (οΈ Eπ΄ πΏ π|π΄ , π π‘βπ ≥ πΌ · πΏ π, π π‘βπ 61 where the expectation is over uniformly random π-dimensional subspaces π΄ ⊆ Fπ π . When π is clear from context, we say the code is πΌ-robust. Observe that if π΄ is a random π1 -dimensional affine subspace and π΅ is a random π2 dimensional affine subspace, where π2 ≥ π1 , then )οΈ]οΈ )οΈ]οΈ]οΈ [οΈ (οΈ )οΈ]οΈ [οΈ [οΈ (οΈ [οΈ (οΈ ≤ Eπ΅ πΏ π|π΅ , π π‘βπ2 Eπ΄ πΏ π|π΄ , π π‘βπ1 = Eπ΅ Eπ΄⊆π΅ πΏ π|π΄ , π π‘βπ1 so if π π‘βπ is (πΌ, π1 )-robust, then it is also (πΌ, π2 )-robust. As a corollary to Theorem 5.4.2, the π-dimensional test (for π ≥ π‘) for π π‘βπ is π(π −3π‘ )robust. −3π‘ Corollary 6.1.2. If π ⊆ {Fπ‘π → Fπ } is linear affine-invariant, then π π‘βπ is ( π 2 , π)-robust for π ≥ π‘. −3π‘ Proof. It suffices to show that π π‘βπ is ( π 2 , π‘)-robust, which follows immediately from Theorem 5.4.2 and Proposition 3.3.2. Proposition 6.1.3 (Robustness composes multiplicatively). Let π‘ ≤ π1 ≤ π2 ≤ π and let π ⊆ {Fπ‘π → Fπ } be linear affine-invariant. If π π‘βπ is (πΌ2 , π2 )-robust and π π‘βπ2 is (πΌ1 , π1 )-robust, then π π‘βπ is (πΌ1 · πΌ2 , π1 )-robust. 6.1.2 Robustness for Small Dimension Throughout Sections 6.1.2 and 6.1.3, fix a linear affine invariant code π ⊆ {Fπ‘π → Fπ } with relative distance πΏ , πΏ(π). Let π ≥ 1 be an integer (we will use π = 3 or 4) and let π = ππ‘. Two codes that play a prominent role are the lift π π‘βπ of π to π dimensions, and the π-fold tensor product π ⊗π of π, which is also an π-dimensional code. We begin by giving a tester with robust analysis for π π‘βπ for this restricted choice of (οΈ π )οΈπ(1) π. We will show that the (π − π‘)-dimensional test is πΏπ -robust. (Note the robustness degrades poorly with π = π/π‘ and so can only be applied for small π). It is important, for Section 6.1.4, that there is no dependence on π‘. 62 Theorem 6.1.4. Let π ≥ 3 and π‘ ≥ 1 and set π = ππ‘. Then π π‘βπ is (πΌ0 , π − π‘)-robust for πΌ0 = πΏ 3π . 16(π2 +3π+2)3 Overview. For simplicity, we describe the proof idea for π‘ = 1. Suppose the average local distance to π 1β(π−1) on random hyperplanes is small. For a ∈ Fπ π , let πa be the code consisting of tensor codewords in π ⊗π whose restrictions to lines in direction a are also βοΈ codewords of π. Note that a πa = π 1βπ . Our main technical result (Theorem 6.1.11) of this (οΈ π )οΈπ(1) -robust. Now, observe that choosing a random hyperplane section shows that πa is πΏπ can be done by choosing π random linearly independent directions, choosing an additional random direction a that is not spanned by any π − 1 of the former, and choosing a random hyperplane spanned by π − 1 of these π + 1 random directions (call such a hyperplane “special”). Viewing the first π chosen directions as the standard basis directions, we see that the average local distance to π 1β(π−1) , and hence to π ⊗(π−1) , when restricting to special hyperplanes, is still small. Therefore, for most a, the average local distance to π ⊗(π−1) on special hyperplanes is small. By the robustness of πa , this implies that our received word is close to some codeword πa ∈ πa for most a. But these codewords πa are all codewords of π ⊗π and close to each other, so they must be the same codeword π ∈ π ⊗π . So we have shown that we are close to π ∈ π ⊗π . We proceed by showing that in fact π ∈ π 1βπ . Note that π ∈ πa for most a. Another technical result, Corollary 6.2.7, implies that in fact π ∈ πa for all a and we are done. To generalize to π‘ > 1, we replace dimension π subspaces with dimension ππ‘ subspaces throughout. Some work needs to be done to ensure that Theorem 6.1.11 still works, and also Corollary 6.2.7 must be generalized appropriately to remove the dependence on π‘. These issues will be discussed in the corresponding sections. (οΈ π )οΈ Definition 6.1.5. Let β ≤ π be an integer. A collection π· ⊆ Fπ‘π is β-proper if for every βοΈ π and every distinct π΄1 , . . . , π΄π ∈ π·, the union ππ=1 π΄π contains at least min {ππ‘, π − β} linearly independent vectors. Definition 6.1.6. For a set π· ⊆ (οΈFπ )οΈ π π , for every x ∈ Fπ π define π±π· (x) to be the collection π‘ 63 of subspaces through x in directions from π different sets from π·. More precisely, β (οΈ )οΈ}οΈ π β βοΈ π· β (x, π΄) β π΄ = π·π , {π·1 , . . . , π·π } ∈ β π {οΈ π±π·π (x) , π=1 Define π±π·π , βοΈ π±π·π (x). x∈Fπ π The testing subspaces through x are π―π· (x) , π±π·π−1 (x) and the decoding subspaces through x are ππ· (x) , π±π·1 (x). Similarly, the testing subspaces are π―π· , π±π·π−1 and the decoding subspaces are ππ· , π±π·1 . If π = (x, ∪ππ=1 π·π ) ∈ π±π·π , then the blocks of π are the sets π·1 , . . . , π·π . Two testing subspaces are adjacent if they differ in at most one block. Remark 6.1.7. If π· is β-proper for β ≤ π‘ then for any π < π we have that π±π·π consists of ππ‘-dimensional subspaces. π Definition 6.1.8. Define ππ· to be the code of all words π : Fπ π → Fπ such that π |π’ ∈ π for every decoding subspace π’ ∈ ππ· . π for any π·. If Remark 6.1.9. Observe that π π‘βπ is a subcode of ππ· βοΈ π· contains the π standard basis vectors, then ππ· is a subcode of π ⊗π . ⌊οΈ ⌋οΈ π log( 1πΏ )+log(π2 +3π+2)+1 Proof of Theorem 6.1.4. Define β , . We note that for the most log(π) interesting cases, where πΏ > 0 and π are fixed and π → ∞, β = 0. Our first step handles the less interesting cases (by appealing to a known result). Specifically, if β ≥ π‘ then by Corollary 6.1.2 we are done since (οΈ π −3π‘ π −3β π ≥ ≥ 2 2 −3 ( ) π log 1 +log(π2 +3π+2)+1 πΏ log(π) )οΈ = 2 πΏ 3π = πΌ0 . 16(π2 + 3π + 2)3 [οΈ (οΈ )οΈ]οΈ Now assume β < π‘ and let π , Eπ£ πΏ π|π£ , π π‘β(π−π‘) , where π£ ⊆ Fπ π is a uniformly random (π − π‘)-dimensional affine subspace. We will assume without lost of generality that π ≤ πΌ0 and in particular π ≤ πΏ 3π π+1 2 2 (π +3π+2) π−1 16( ) ≤ πΏ 2π π −β π+1 8(π−1 ) 2 64 . Observe that [οΈ (οΈ )οΈ]οΈ π‘β(π−π‘) π π = Eπ΄1 ,...,π΄π ∈(Fπ π Eπ΄∈ Fπ Eπ£∈π―{π΄ ,...,π΄π ,π΄} πΏ π|π£ , π 1 ) ( ) π‘ π‘ where π΄1 , . . . , π΄π are random sets such that their union is linearly independent, and π΄ is a random set such that {π΄1 , . . . , π΄π , π΄} is β-proper. Fix π΄1 , . . . , π΄π such that [οΈ (οΈ )οΈ]οΈ π‘β(π−π‘) Eπ΄∈(Fπ ≤ π. π Eπ£∈π―{π΄ ,...,π΄π ,π΄} πΏ π|π£ , π 1 ) π‘ Since π π‘β(π−π‘) ⊆ π ⊗(π−1) , [οΈ (οΈ )οΈ]οΈ ⊗(π−1) Eπ΄∈(Fπ ≤ π. π Eπ£∈π―{π΄ ,...,π΄π ,π΄} πΏ π|π£ , π 1 ) π‘ By affine-invariance, we may assume without loss of generality that π΄1 , . . . , π΄π form the (οΈFπ )οΈ π standard basis vectors for Fπ , let π·π΄ , {π΄1 , . . . , π΄π , π΄}. By Markov’s π . For any π΄ ∈ π‘ inequality, [οΈ [οΈ (οΈ )οΈ ]οΈ]οΈ 1 π ⊗(π−1) −π Pr Eπ£∈π―π·π΄ πΏ π|π£ , π ≥ 2πΏ π < πΏ . π΄ 2 So, for more than 1 − 12 πΏ π fraction of blocks π΄ such that π·π΄ is β-proper, we have a codeword (οΈπ+1)οΈ ⊗π −π π ⊆ π such that (by Theorem 6.1.11) πΏ(π, π ) < 2πΏ π < 12 πΏ π . For every ππ΄ ∈ ππ· π΄ π−1 π΄ two such blocks π΄, π΄′ , we have πΏ(ππ΄ , ππ΄′ ) ≤ πΏ(ππ΄ , π) + πΏ(π, ππ΄′ ) < πΏ π = πΏ(π ⊗π ), so there is some codeword π ∈ π ⊗π such that ππ΄ = π for every such π΄. For such π΄, it follows that for every b ∈ Fπ π , the restriction of π to the subspace (b, π΄) is a codeword of π, i.e. π|(b,π΄) ∈ π. (οΈ )οΈ π−π‘ π −π By Claim 6.1.17, for more than 1 − 21 πΏ π − π2 π−1 − π π−1 fraction of blocks π΄ (without the ⊗π requirement that π·π΄ be proper), π|(b,π΄) ∈ π for every b ∈ Fπ and π . In particular, π ∈ π (οΈ )οΈ −π‘ −β π π for every b ∈ Fπ / π for less than 12 πΏ π + π2 π−1 + π π−1 fraction of π΄. It sufficient to π , π|(b,π΄) ∈ (οΈ )οΈ π −π‘ π −π show that 21 πΏ π + π2 π−1 + π π−1 ≤ πΏ π − (π + 1)π −π‘ . Then it will follow from Corollary 6.2.7 65 that π ∈ π π‘βπ and since πΏ(π, π) ≤ 2πΏ −π π (οΈπ+1)οΈ π−1 we are done. Calculating: (οΈ )οΈ −π‘ (οΈ )οΈ −π π π π −π π π π −π π −π −π‘ +π + (π + 1)π ≤ +π + (π + 1) 2 π−1 π−1 2 π−1 π−1 π−1 (οΈ 2 )οΈ −π π + 3π + 2 π = π−1 2 π ππΏ ≤ 4(π − 1) 1 π πΏ . ≤ 2 The composability of robust tests immediately yields the following corollary where the test is now 2π‘ dimensional. 21 πΏ Corollary 6.1.10. π π‘β4π‘ is (πΌ1 , 2π‘)-robust, where πΌ1 ≥ 6·10 10 . (οΈ 12 )οΈ (οΈ 9 )οΈ πΏ πΏ Proof. By Theorem 6.1.4, π π‘β4π‘ is 432,000 , 3π‘ -robust and π π‘β3π‘ is 128,000 , 2π‘ -robust. There- fore, by composing, the 2π‘-dimensional robustness of π π‘β4π‘ is at least πΏ 12 432,000 · πΏ9 128,000 = πΏ 21 55,296,000,000 6.1.3 Robustness of Special Tensor Codes In this section, we prove the main technical result (Theorem 6.1.11) used in Section 6.1.2. (οΈ π )οΈ Theorem 6.1.11. Let π ≥ 3 and β ≤ π‘. Set π = ππ‘. Let π· ⊆ Fπ‘π be β-proper with |π·| ≥ π [οΈ (οΈ )οΈ]οΈ π π −β ⊗(π−1) blocks. Let π : Fπ . If π < πΏ |π·| 2 , then π → Fπ be a word with π , Eπ£∈π―π· πΏ π|π£ , π 4(π−1) (οΈ )οΈ |π·| π πΏ(π, ππ· ) ≤ π π−1 . Overview. Our analysis is an adaptation of Viderman’s [Vid12]. For simplicity, assume π‘ = 1. We define a function π : Fπ π → Fπ , which we show is both close to π and a codeword π of ππ· . Following Viderman’s analysis, we partition Fπ π into “good”, “fixable”, and “bad” points. Each hyperplane π£ ∈ π―π· has an associated codeword ππ£ ∈ π ⊗(π−1) , the nearest codeword to π|π£ , and an opinion ππ£ (x) about x. “Good” points are points for which any 66 hyperplane agrees with π. “Fixable” points are points for which hyperplanes agree with each other, but not with π. “Bad” points are points for which at least two hyperplanes disagree with each other. For good or fixable x, we naturally define π(x) to be the common opinion ππ£ (x) of any hyperplane π£ through x. Claim 6.1.13 implies that there are not many bad points, which immediately shows that π is close to π. So far, our proof has been a straightforward adaptation of Viderman’s. However, at this point, we are forced to depart from Viderman’s proof. A hyperplane is “bad” if it has more than 12 πΏ π−1 fraction bad points. Claim 6.1.12 shows that every bad point is in a bad hyperplane, and Claim 6.1.14 shows that there are less than 21 πΏπ bad hyperplanes. π and π―π· , this is In [Vid12], which analyses π ⊗π and axis-parallel hyperplanes instead of ππ· already enough, since this implies that in each axis-parallel direction, there are less than πΏπ bad hyperplanes, so the remaining points are all good or fixable and with a little bit more work, one can show that π can be extended uniquely to a tensor codeword using the erasure-decoding properties of tensor codes. Unfortunately, we do not have this structure. We say a line is “good” if it is contained in some good hyperplane, otherwise it is bad. We must further partition the bad points into merely bad and “super-bad” points, which are points such that either every hyperplane is bad, or there are two disagreeing good hyperplanes. For merely bad x, we define π(x) to be the common opinion ππ£ (x) of any good hyperplane π£ through x. For super-bad x, we pick any line π’ through x, take the restriction of π to the non-super-bad points on π’, and extend it to a codeword ππ’ ∈ π, and define π(x) , ππ’ (x). Two non-trivial steps remain: showing that π(x) is well-defined for super-bad π x, and showing that π ∈ ππ· . Claim 6.1.15 shows that, for any special plane, there are less than 1 πΏπ 2 lines in each direction that are bad (not contained in any good hyperplane) or contain a super-bad point. This is proved by exhibiting, for each such undesirable line, a bad hyperplane in a fixed direction containing the line. If there were too many undesirable lines, this would result in too many parallel bad hyperplanes, contradicting Claim 6.1.14. Finally, Claim 6.1.16 shows that if π’ is a line with no super-bad points, then π|π’ ∈ π is a codeword. 67 Now, we show that π is well-defined on super-bad x. Let π’1 , π’2 be two lines through x. Let π be the plane through x containing π’1 , π’2 . On this plane, by Claim 6.1.15, in each direction we have enough lines π’ with no super-bad points, for which π|π’ ∈ π (by Claim 6.1.16), so that we can uniquely extend π onto the entire plane (by Proposition A.2.5). This gives a well-defined value for π(x). π Finally, we show that π ∈ ππ· . Let π’ be any line. If π’ has no super-bad points, then π|π’ ∈ π follows from Claim 6.1.16. If π does have a super-bad point x, then π|π’ ∈ π by the way we defined π(x). This completes our analysis for the case π‘ = 1. To generalize to π‘ > 1, we replace lines with “decoding subspaces” (subspaces of dimension π‘), planes with subspaces of dimension 2π‘, and hyperplanes with “testing subspaces” (subspaces of codimension π‘). Some care must be taken when proving Claim 6.1.15, because the intersection of two decoding subspaces may have non-trivial dimension. We therefore require the notion of “β-properness” of π·, and must modify Claim 6.1.14 and also prove Claim 6.1.17 to accommodate this notion. Details follow. Proof of Theorem 6.1.11. For each testing subspace π£ ∈ π―π· , define ππ£ ∈ π ⊗(π−1) to be the closest codeword to π|π£ (break ties arbitrarily). We will partition Fπ π into three disjoint sets πΊ, πΉ, π΅ (good, fixable, and bad points, respectively) as follows: πΊ , {οΈ x ∈ Fπ π | ππ£ (x) = π(x) for every π£ ∈ π―π· (x) }οΈ }οΈ ′ x ∈ Fπ π | ππ£ (x) = ππ£ ′ (x) ΜΈ= π(x) for every π£, π£ ∈ π―π· (x) {οΈ }οΈ ′ π΅ , x ∈ Fπ π | ππ£ (x) ΜΈ= ππ£ ′ (x) for some π£, π£ ∈ π―π· (x) . πΉ , {οΈ Call a testing subspace bad if at least 21 πΏ π−1 fraction of its points are in π΅, and good otherwise. A decoding subspace is good if it is contained in some good testing subspace, and bad otherwise. Further, define the set π΅ ′ of super-bad points π΅ ′ , {x ∈ π΅ | every π£ ∈ π―π· (x) is bad or ∃ good π£, π£ ′ ∈ π―π· (x) such that ππ£ (x) ΜΈ= ππ£′ (x)}. 68 Claim 6.1.12. If π£, π£ ′ ∈ π―π· are adjacent good testing subspaces, then ππ£ |π£∩π£′ = ππ£′ |π£∩π£′ . In particular, every bad point is in a bad testing subspace. Proof. Suppose b ∈ π£ ∩ π£ ′ and ππ£ (b) ΜΈ= ππ£′ (b). Since π£, π£ ′ are adjacent, they have π − 2 blocks π΄1 , . . . , π΄π−2 in common. Let π£ have blocks π΄1 , . . . , π΄π−2 , π΄ and let π£ ′ have blocks π΄1 , . . . , π΄π−2 , π΄′ . Let π’ ∈ ππ· be the decoding subspace (b, π΄1 ). Since ππ£ |π’ , ππ£′ |π’ ∈ π disagree on b, they are distinct codewords and hence disagree on at least πΏπ π‘ points of π’, say x1 , . . . , xπΏππ‘ . For each π ∈ [πΏπ π‘ ], let π£π ∈ π―π· be the testing subspace (xπ , π΄2 ∪ · · · ∪ π΄π−2 ∪ π΄ ∪ π΄′ ). Since ππ£ (xπ ) ΜΈ= ππ£′ (xπ ), that means ππ£π disagrees with one of ππ£ , ππ£′ at xπ . Without loss of generality, suppose ππ£ disagrees with ππ£1 , . . . , ππ£πΏππ‘ /2 . We will show that π£ is bad, which proves the first part of the claim. For each π ∈ [πΏπ π‘ ], let π€π = (xπ , π΄2 ∪ · · · ∪ π΄π−2 ∪ π΄). Note that π’ ∩ π€π = {xπ } (since π· is β-proper, for β ≤ π‘), so all π€π are different parallel subspaces and hence disjoint. Since π€π ∈ π±π·π−2 , the restrictions ππ£ |π€π , ππ£π |π€π ∈ π ⊗π−2 are codewords and are distinct because they disagree on xπ , therefore, by Proposition A.2.4, they disagree on at least πΏ π−2 π π−2π‘ points in π€π , which are therefore bad. Thus, each π£π contributes πΏ π−2 π π−2π‘ bad points to π£, for a total of 21 πΏ π−1 π π−π‘ bad points since the π€π are disjoint. For the second part, suppose b ∈ π΅ is a bad point. We will show that b lies in a bad testing subspace. By definition, there are two testing subspaces π£, π£ ′ ∈ π―π· (b) such that ππ£ (b) ΜΈ= ππ£′ (b). Suppose π£ has blocks π΄1 , . . . , π΄π−1 and π£ ′ has directions π΄′1 , . . . , π΄′π−1 . Assume, without lost of generality, that if π΄π = π΄′π then π = π. Define π£0 , π£, and for π ∈ [π−1], define π£π ∈ π―π· to be the testing subspace through b in directions π΄′1 , . . . , π΄′π , π΄π+1 , . . . , π΄π−1 . Consider the sequence π£0 , π£1 , . . . , π£π−1 of testing subspaces. For each π, the testing subspaces π£π , π£π+1 are adjacent. Since ππ£0 (b) ΜΈ= ππ£π−1 (b), there exists some π such that ππ£π (b) ΜΈ= ππ£π+1 (b), and by the first part of the claim it follows that one of π£π , π£π+1 is bad. Claim 6.1.13. π ≥ |πΉ | ππ + |π΅| |π·| π π (π−1 ) 69 Proof. Observe that |π―π· | = π π‘ (οΈ |π·| )οΈ . Therefore, π−1 )οΈ]οΈ [οΈ (οΈ π = Eπ£∈π―π· πΏ π|π£ , π ⊗(π−1) = Eπ£∈π―π· [πΏ (π|π£ , ππ£ )] ∑οΈ 1 ∑οΈ 1 = 1ππ£ (x)ΜΈ=π(x) (οΈ |π·| )οΈ π π‘ π−1 π£∈π―π· π π−π‘ x∈π£ ∑οΈ 1 = #{π£ ∈ π―π· (x) | ππ£ (x) ΜΈ= π(x)} (οΈ |π·| )οΈ π π π−1 x∈Fπ π (οΈ )οΈ ∑οΈ ∑οΈ (οΈ |π·| )οΈ ∑οΈ 1 0+ + 1 ≥ (οΈ |π·| )οΈ π−1 π π π−1 x∈πΊ x∈πΉ x∈π΅ = |π΅| |πΉ | + (οΈ |π·| )οΈ . π π π π π−1 Claim 6.1.14. There are less than 21 πΏπ π‘−β bad testing subspaces. (οΈ |π·| )οΈ π Proof. By Claim 6.1.13, there are at most |π΅| ≤ π π−1 π bad points. Each bad testing (οΈ |π·| )οΈ subspace has at least πΏ π−1 π π−π‘ /2 bad points by definition. Each bad point has at most π−1 bad testing subspaces through it. Therefore, the number of testing subspaces is bounded by (οΈ )οΈ (οΈ )οΈ2 |π·| 2π 1 |π·| |π΅| ≤ π−1 π π‘ < πΏπ π‘−β . 1 π−1 π−π‘ · π−1 πΏ 2 π−1 πΏ π 2 π Now we proceed to prove the lemma. We construct a codeword π ∈ ππ· with πΏ(π, π) ≤ (οΈ |π·| )οΈ π π−1 in stages, as follows. First, for x ∈ πΊ ∪ πΉ , we define π(x) , ππ£ (x) for any testing subspace π£ ∈ π―π· (x). This is well-defined by the definition of πΊ and πΉ . Furthermore, since (οΈ |π·| )οΈ π(x) = ππ£ (x) = π(x) for x ∈ πΊ, we already guarantee that πΏ(π, π) ≤ |πΉ |+|π΅| ≤ π . π π−1 π For x ∈ π΅ β π΅ ′ , define π(x) , ππ£ (x) for any good testing subspace π£ ∈ π―π· (x), whose existence is guaranteed by the fact that x ∈ / π΅ ′ . This is well-defined because if π£, π£ ′ ∈ π―π· (x) are both good, then it follows from the fact that x ∈ / π΅ ′ that ππ£ (x) = ππ£′ (x). 70 Claim 6.1.15. Let π€ ∈ π±π·2 be a subspace in directions π΄1 , π΄2 ∈ π·. For each π ∈ {1, 2}, π€ contains less than 12 πΏπ π‘ decoding subspaces in direction π΄π which intersect π΅ ′ or are bad (not contained in any good testing subspace). Proof. By symmetry, it suffices to consider π = 2. Let π΄3 , . . . , π΄π ∈ π· be blocks in some other direction. Let π’1 , . . . , π’π ⊆ π€ be decoding subspaces in direction π΄2 such that, for each π ∈ [π], π’π intersects π΅ ′ or is bad. It suffices to exhibit, for each π ∈ [π], a bad testing subspace π£ ∈ π―π· containing π’π which has block π΄2 but not π΄1 . In this case we will show that |π£ ∩ π€| ≤ π π‘+β so each such bad testing subspace contain at most π β of the π’π -s. Since, by Claim 6.1.14, there are at most 12 πΏπ π‘−β such subspaces, we get that π ≤ 12 πΏπ π‘ . Indeed, since π· is β-proper, the subspace π’ + π£ ∈ π±π·π has dimension at least π − β. Therefore, dim(π£ ∩ π€) = dim(π£) + dim(π€) − dim(π£ + π€) ≤ π − π‘ + 2π‘ − (π − π) = π‘ + π and |π£ ∩ π€| ≤ π π‘+β . Fix π ∈ [π] and π’ , π’π and we will show that π’ is contained in a bad testing subspace. If π’ is bad, then we are done, since any testing subspace containing π’, in particular the testing subspace in directions π΄2 , . . . , π΄π , is bad. Now suppose π’ has a point x ∈ π’ ∩ π΅ ′ . Let π£ ∈ π―π· be the testing subspace (x, π΄2 ∪ · · · ∪ π΄π ). If π£ is bad, we are done. Otherwise, since x ∈ π΅ ′ , there exists another good hyperplane π£ ′ ∈ π·, in directions π΄′1 , . . . , π΄′π−1 , such that ππ£ (x) ΜΈ= ππ£′ (x). Without loss of generality, assume that if π΄π = π΄′π then π = π (in particular π΄1 ∈ / {π΄′2 , . . . , π΄′π−1 }). For each π ∈ [π − 1], if π΄2 = π΄′2 define π£π ∈ π―π· (x) to be the testing subspace (x, π΄′2 , . . . , π΄′π , π΄π+1 , . . . , π΄π ), and if π΄2 ΜΈ= π΄′2 define π£1 to be π£ and π£π to be (x, π΄2 , π΄′2 , . . . , π΄′π , π΄π+1 , . . . , π΄π−1 ). In any case define π£π , π£ ′ . For every π ∈ [π − 1], π£π and π£π+1 are adjacent. Note that for every π ∈ [π − 1], π£π contains the direction π΄2 and does not contain the direction π΄1 . We will show that π£π is bad for some π ∈ [π − 1]. Since ππ£1 (x) ΜΈ= ππ£π (x), there exists some π ∈ [π − 1] such that ππ£π (x) ΜΈ= ππ£π+1 (x), and therefore, by Claim 6.1.12, one of π£π , π£π+1 is bad. If π < π − 1, then π, π + 1 ≤ π − 1, and so we are done. If π = π − 1, then by assumption π£π = π£ ′ is good, so it must be that π£π−1 is bad. 71 Claim 6.1.16. If π’ ∈ ππ· is a decoding subspace and π’ ∩ π΅ ′ = ∅, then for every x ∈ π’ there is a codeword πx ∈ π defined on π’ such that πx (x) = π(x) and πΏ(πx , π|π’ ) < 2πΏ . In particular, π|π’ ∈ π. Proof. Fix x ∈ π’. Let π΄ = {a1 , . . . , aπ‘ } ∈ π· be the directions of π’. Since x ∈ / π΅ ′ , there is a good testing subspace π£ ∈ π―π· (x). Let π΄′ = {a′1 , . . . , a′π‘ } be some block in π£ not equal to π΄ and consider the subspace π€ = (x, π΄ ∪ π΄′ ) ∈ π±π·2 . For s, s′ ∈ Fπ‘π , let π€(s, s′ ) , ∑οΈ ∑οΈ x + π‘π=1 π π aπ + π‘π=1 π ′π a′π . Let π€(s, *) , {π€(s, s′ ) | s′ ∈ Fπ‘π } ∈ ππ· , π€(*, s′ ) , {π€(s, s′ ) | s ∈ Fπ‘π } ∈ ππ· . Let πΌ ⊆ Fπ‘π β {0} be the set of points s ΜΈ= 0 such that π€(s, *) intersects π΅ ′ or is bad. Similarly, let πΌ ′ ⊆ Fπ‘π β {0} be the set of points s′ ΜΈ= 0 such that π€(*, s′ ) intersects π΅ ′ or is bad. By Claim 6.1.15, |πΌ|, |πΌ ′ | < 12 πΏπ π‘ . Note that for each (s, s′ ) ∈ (Fπ‘π β πΌ) × (Fπ‘π β πΌ ′ ), we have π€(s, s′ ) ∈ / π΅ ′ : if s ΜΈ= 0 or s′ ΜΈ= 0, this follows from the definition of πΌ and πΌ ′ ; if s = s′ = 0, then π€(s, s′ ) = x ∈ / π΅ ′ . Thus π is defined on π€((Fπ‘π βπΌ)×(Fπ‘π βπΌ ′ )). Note that for each s ∈ Fπ‘π βπΌ, the decoding subspace π€(s, *) is good and hence contained in a good testing subspace π£s ∈ π―π· , therefore ππ£s |π€(s,*) ∈ π. Similarly, for each s′ ∈ Fπ‘π β πΌ ′ , the decoding subspace π€(*, s′ ) is contained in a good testing subspace π£s′ ∈ π―π· , hence ππ£s′ |π€(*,s′ ) ∈ π. Since |πΌ|, |πΌ ′ | < 12 πΏπ π‘ , by Proposition A.2.5, π can be extended uniquely into ππ€ ∈ π ⊗2 defined on π€. Define πx , ππ€ |π’ . Note that πx ∈ π since it is the restriction of ππ€ ∈ π ⊗2 to π’ = π€(*, 0). Also, if s ∈ Fπ‘π β πΌ, then π|π€(s,*) = ππ£s |π€(s,*) = ππ€ |π€(s,*) and in particular, π(π€(s, 0)) = ππ€ (π€(s, 0)) = πx (π€(s, 0)). So πΏ(π, πx ) ≤ |πΌ| ππ‘ < 2πΏ . Finally, since 0 ∈ / πΌ, πΌ ′ , we have π(x) = π(π€(0, 0)) = πx (π€(0, 0)) = πx (x). This proves the first part of the claim. For the second part (showing π|π’ ∈ π), fix some x0 ∈ π’. For each x ∈ π’, let πx be the codeword guaranteed by the previous part. Then, for every x ∈ π’, πΏ(πx0 , πx ) ≤ πΏ(πx0 , π|π’ ) + πΏ(π|π’ , πx ) < πΏ, therefore πx0 = πx . Moreover, for all x ∈ π’, πx0 (x) = πx (x) = π(x), so π|π’ = πx0 ∈ π. 72 We proceed to define π(x) for π₯ ∈ π΅ ′ . For such an x, pick any decoding subspace π’ ∈ ππ· (x), extend π|π’βπ΅ ′ to a codeword ππ’ ∈ π, and define π(x) , ππ’ (x). We now argue that this is well-defined. Suppose π’1 , π’2 ∈ ππ· (x) in directions π΄1 , π΄2 ∈ π·, respectively. We need to show that ππ’1 , ππ’2 are well-defined and that ππ’1 (x) = ππ’2 (x). Let π€ ∈ π±π·2 be the unique subspace containing π’1 , π’2 , so π€ = (x, π΄1 ∪ π΄2 ). By Claim 6.1.15, in each direction π΄1 , π΄2 , there are less than 21 πΏπ π‘ decoding subspaces in that direction in π€ which intersect π΅ ′ . In particular, this implies that π’1 , π’2 each contain less than πΏπ π‘ points from π΅ ′ . By what we just showed, there are sets π½1 , π½2 ⊆ Fπ‘π of size |π½1 |, |π½2 | > (1 − πΏ)π π‘ such that the “sub-rectangle” π , π€(π½1 × π½2 ) contains no points from π΅ ′ , and therefore π has already been defined on π . By Claim 6.1.16, on each decoding subspace π’ in π in either direction π΄1 or π΄2 , π|π’ ∈ π. Applying Proposition A.2.5, we see that π|π can be uniquely extended to a tensor codeword ππ€ ∈ π ⊗2 on π€, and this gives a way to extend π|π’π βπ΅ ′ to the codeword ππ’π , ππ€ |π’π ∈ π for π ∈ {1, 2}. Therefore, the extensions ππ’1 , ππ’2 agree on x since ππ’1 (x) = ππ€ (x) = ππ’2 (x), and moreover for each decoding subspace π’π this extension is unique since each decoding subspace has less than πΏπ π‘ points from π΅ ′ . Now that we have defined π : Fπ π → Fπ and have shown that πΏ(π, π) ≤ π (οΈ |π·| )οΈ π−1 , it only π . Let π’ ∈ ππ· be a decoding subspace. If π’ ∩ π΅ ′ = ∅, then remains to show that π ∈ ππ· π|π’ ∈ π follows from Claim 6.1.16. If π’ intersects π΅ ′ , by the way we defined π(x) for x ∈ π΅ ′ , we showed that for any decoding subspace π’ through x ∈ π΅ ′ , π|π’ ∈ π by extending π|π’βπ΅ ′ to a codeword. (οΈ π )οΈ Claim 6.1.17. Let β ≤ π‘ be a natural number and π΄1 , . . . , π΄π ∈ Fπ‘π be such that their (οΈ )οΈ π−π‘ (οΈ π )οΈ π −β union is linearly independent. Then at least 1 − π2 π−1 − π π−1 fraction of π΄ ∈ Fπ‘π satisfy that π΄, π΄1 , . . . , π΄π is β-proper. Proof. Let a1 , . . . , aπ‘ be the random elements comprising π΄. By a union bound, it suffices to (οΈ [π] )οΈ (οΈβοΈ )οΈ show for any π ∈ π−2 that π΄ ∪ π΄ is linearly independent with probability at least π π∈π )οΈ (οΈβοΈ (οΈ )οΈ −π‘ [π] π and for any π ∈ π−1 the probability that 1 − π−1 π∈π π΄π ∪ π΄ contains at least π − β −β π linearly independent elements is at least 1 − π−1 . (οΈ [π] )οΈ (οΈβοΈ )οΈ Fix π ∈ π−2 .For any π ∈ [π‘], the probability that aπ ∈ Fπ π is in the span of π∈π π΄π ∪ 73 {a1 , . . . , aπ−1 }, conditioned on the event that the latter is linearly independent, is (οΈβοΈ )οΈ π −2π‘+π−1 . So the probability that π∈π π΄π ∪ π΄ is linearly independent is π‘ ∏οΈ (οΈ 1−π −2π‘+π−1 )οΈ ≥ 1− π=1 π‘ ∑οΈ π π−2π‘+π−1 ππ = π −2π‘+π−1 π=1 ≥ 1−π −π‘ ∞ ∑οΈ π −π π=1 −π‘ = 1− Now fix π ∈ (οΈ [π] )οΈ π−1 π . π−1 . Similarly, The probability that ππ ∈ (οΈβοΈ π∈π )οΈ π΄π ∪ {a1 , . . . , aπ−1 }, condition on the event that the later linearly independent is π −π‘+π−1 . (οΈβοΈ )οΈ π∈π π΄π ∪ {a1 , . . . , aπ‘−β } are linearly independent is So we get that π‘−β π‘−β ∏οΈ ∑οΈ (οΈ )οΈ 1 − π −π‘+π−1 ≥ 1 − π −π‘+π−1 π=1 π=1 ≥ 1 − π −β ∞ ∑οΈ π −π π=1 −β = 1− 6.1.4 π . π−1 Robustness for Large Dimension In this section, we prove our main result of the chapter: Theorem 6.1.18. Let π , Eπ£ [πΏ(π|π£ , π π‘β2π‘ )], where π£ is a random affine subspace of dimension 2π‘. Let πΌ1 be the 2π‘-dimensional robustness of π π‘β4π‘ given by Corollary 6.1.10. If (οΈ )οΈ 3 1πΏ π < πΌ400 − 3π −π‘ , then π ≥ 1 − 4πΏ · πΏ(π, π π‘βπ ). In particular, π π‘βπ is (πΌ2 , 2π‘)-robust, where πΌ2 ≥ πΏ 72 . 2·1052 74 Notation. Throughout Section 6.1.4, fix the received word π : Fπ π → Fπ and define π , Eπ£ [πΏ(π|π£ , π π‘β2π‘ )], and we will assume that 0 < π < πΌ1 πΏ 3 400 − 3π −π‘ . The case where πΌ1 πΏ 3 400 > 3π −π‘ is easily dealt with at the end of the proof by using Corollary 6.1.2. Note that, since πΌ1 , πΏ ≤ 1, this implies π −π‘ ≤ πΏ . 1200 Throughout this section we will assume π ≥ 4π‘. If ^ π < 4π‘ we can pad the function π to get a function π^ : F4π‘ π → Fπ (by setting π (x, y) = π (x) 4π‘−π ) and applying our tester to π^. We will typically use for every x ∈ Fπ π and y ∈ Fπ π’, π£, π€ to denote affine subspaces of dimension π‘, 2π‘, and 4π‘ respectively. For any affine π‘βdim(π΄) subspace π΄ ⊆ Fπ be the codeword nearest to π|π΄ , breaking ties arbitrarily. π , let ππ΄ ∈ π Let ππ΄ , Eπ£⊆π΄ [πΏ(π|π£ , π π‘β2π‘ )], where the expectation is taken over uniformly random 2π‘dimensional subspaces π£ ⊆ π΄. Fix the following constants: πΌ1 πΏ 2 − πΌ1 π −π‘ 40 π + 2π −π‘ π , . πΎ πΎ , In particular, these constants are chosen so that the following bounds hold: πΏ 2 πΏ π ≤ . 10 20πΏ −1 (πΌ1−1 πΎ + π −π‘ ) ≤ Overview. This proof is a straightforward generalization of “bootstrapping” proofs originating in the work of Rubinfeld and Sudan [RS96] and which also appears in [ALM+ 98, AS03, Aro94]. Our writeup in particular follows [Aro94]. For simplicity, assume π‘ = 1. Our approach is to define a function π : Fπ π → Fπ and then show that it is both close to π and a codeword of π 1βπ . The definition of π is simple: for every x ∈ Fπ π , consider the opinion ππ’ (x) for every line π’ through x, and define π(x) as the majority opinion. We need to show that π is well-defined (the majority is actually a majority). Our main technical lemma (Lemma 6.1.21) of this section shows that most lines agree with each other, so π is well-defined. Lemma 6.1.21 uses Claim 6.1.19, which shows that for a 4-dimensional affine 75 subspace π€, if ππ€ is small, then for every x ∈ π€, most lines π’ ⊆ π€ satisfy ππ’ (x) = ππ€ (x). To prove Claim 6.1.19 we use the results of Section 6.1.2, in particular the robustness of the plane test in π = 4 dimensions (Corollary 6.1.10). Since the average πΏ(π|π’ , ππ€ |π’ ) over π’ through x is about πΏ(π|π€ , ππ€ ), by robustness this is less than πΌ1−1 ππ€ , which is small since ππ€ is small. Therefore, for most π’, πΏ(π|π’ , ππ€ |π’ ) is small and so it must be that ππ’ = ππ€ |π’ . Once we have shown that π is well-defined, showing that π is close to π requires just a bit of calculation. Showing that π ∈ π 1βπ involves more work. For each line π’, define π′π’ ∈ π to be the nearest codeword to π|π’ . Fix a line π’ and a point x ∈ π’. We want to show that π|π’ (x) = π′π’ (x). The idea is to show the existence of a “good” 4-dimensional π€ ⊇ π’ such that ππ€ is small and for more than 1 − 2πΏ fraction of points y ∈ π’ (including x) are “good” in the sense that π(y) = ππ’′ (y) for a non-negligible fraction of lines π’′ through y. Once we have such a π€, we show that for every good y ∈ π’, π(y) = ππ€ (y). Since π’ has more than 1 − 2πΏ fraction good points, this implies that πΏ(π|π’ , ππ€ |π’ ) < 2πΏ , hence π′π’ = π|π€ , so π′π’ (x) = π|π€ (x) = π(x), as desired. Claim 6.1.19. If π€ ⊆ Fπ π be a 4π‘-dimensional affine subspace with ππ€ ≤ πΎ, then for every x ∈ π€, at least 1 − πΏ 20 fraction of π‘-dimensional subspaces π’ ⊆ π€ satisfy ππ’ (x) = ππ€ (x). Proof. Fix x ∈ π€. Let π be the set of π‘-dimensional subspaces π’ containing x such that πΏ(π|π’ , ππ€ |π’ ) < 20πΏ −1 (πΌ1−1 ππ€ +π −π‘ ). By Corollary 6.1.10, Eπ’⊆π€ [πΏ(π|π’ , ππ€ |π’ )] ≤ πΏ(π|π€ , ππ€ )+π −π‘ ≤ π’∋x πΌ1−1 ππ€ +π −π‘ , so by Markov’s inequality, the probability that πΏ(π|π’ , ππ€ |π’ ) ≥ 20πΏ −1 (πΌ1−1 ππ€ +π −π‘ ) is at most −π‘ πΌ−1 1 ππ€ +π −1 −1 20πΏ (πΌ1 ππ€ +π −π‘ ) = πΏ . 20 For π’ ∈ π , since πΏ(π|π’ , ππ€ |π’ ) < 20πΏ −1 (πΌ1−1 ππ€ + π −π‘ ) ≤ πΏ 2 and ππ€ |π’ ∈ π, we have ππ’ = ππ€ |π’ and therefore ππ’ (x) = ππ€ (x). The following claim says that Eπ€ [ππ€ ] ≈ π, even if we insist that π€ contains a fixed π‘-dimensional subspace. −π‘ Claim 6.1.20. For any π‘-dimensional affine subspace π’ ⊆ Fπ π , Eπ€⊇π’ [ππ€ ] ≤ π + 2π , where π€ is a random 4π‘-dimensional affine subspace containing π’. In particular, for any point −π‘ x ∈ Fπ π , Eπ€∋x [ππ€ ] ≤ π + 2π . 76 Proof. Observe that )οΈ]οΈ [οΈ (οΈ π = Eπ£ πΏ π|π£ , π π‘β2π‘ )οΈ]οΈ [οΈ (οΈ ≥ Eπ£:π’∩π£=∅ πΏ π|π£ , π π‘β2π‘ · Pr[π’ ∩ π£ = ∅] π£ )οΈ )οΈ]οΈ (οΈ [οΈ (οΈ π‘β2π‘ · 1 − π −(π−3π‘) . (Lemma B.2.1) ≥ Eπ£:π’∩π£=∅ πΏ π|π£ , π Therefore, [οΈ [οΈ ]οΈ]οΈ Eπ€⊇π’ [ππ€ ] = Eπ€⊇π’ Eπ£⊆π€ πΏ(π|π£ , π π‘β2π‘ ) [οΈ ]οΈ β [οΈ ]οΈ π‘β2π‘ β ≤ Eπ€⊇π’ Eπ£⊆π€ πΏ(π|π£ , π ) π’ ∩ π£ = ∅ + Pr [π’ ∩ π£ ΜΈ= ∅] π£⊆π€ β [οΈ [οΈ ]οΈ]οΈ π‘β2π‘ β (Lemma B.2.1) ≤ Eπ€⊇π’ Eπ£⊆π€ πΏ(π|π£ , π ) π’ ∩ π£ = ∅ + π −π‘ = Eπ£:π’∩π£=∅ [πΏ(π|π£ , π π‘β2π‘ )] + π −π‘ π ≤ + π −π‘ −(π−3π‘) 1−π ≤ π + 2π −π‘ Lemma 6.1.21 (Main). For every x ∈ Fπ π , there is a collection π1 of at least 1 − πΏ 5 − πΏ 600 fraction of the π‘-dimensional affine subspaces through x, such that ππ’ (x) = ππ’′ (x) for every π’, π’′ ∈ π1 . Proof. Let π be the set of all π‘-dimensional affine subspaces π’ through x. Partition π into disjoint collections π1 , . . . , ππ with |π1 | ≥ · · · ≥ |ππ | according to the value of ππ’ (x). We will show that Prπ’∋π₯ [π’ ∈ π1 ] ≥ 1 − πΏ 5 − πΏ . 600 For every 4π‘-dimensional subspace π€, let ππ€ be the collection of π‘-dimensional subspaces π’ through x, guaranteed by Claim 6.1.19, satisfying 77 ππ’ (x) = ππ€ (x). Then Pr [π’ ∈ π1 ] ≥ π’∋x = (Lemma B.2.2) ≥ Pr [∃π π’, π’′ ∈ ππ ] π’,π’′ ∋x Pr [ππ’ (x) = ππ’′ (x)] π’,π’′ ∋x [ππ’ (x) = ππ’′ (x)] − π −(π−2π‘) π’∩π’ ={x} β‘ β€ Pr ′ = Eπ€∋x β£ Pr ′ [ππ’ (x) = ππ’′ (x)]β¦ − π −(π−2π‘) π’,π’ ⊆π€ π’∩π’′ ={x} β‘ β€ (Lemma B.2.2) ≥ Eπ€∋x β£ Pr [ππ’ (x) = ππ’′ (x)β¦ − π −2π‘ − π −(π−2π‘) ′ π’,π’ ⊆π€ π’,π’′ ∋x β€ β‘ πΏ [ππ’ (x) = ππ’′ (x)β¦ − ≥ Eπ€∋x β£ Pr ′ π’,π’ ⊆π€ 600 π’,π’′ ∋x β β€ β‘ β β πΏ ≥ Eπ€∋x β£ Pr [ππ’ (x) = ππ’′ (x)] ββ ππ€ ≤ πΎ β¦ · Pr [ππ€ ≤ πΎ] − ′ π€∋x π’,π’ ⊆π€ 600 β π’,π’′ ∋x β β‘ β€ β β πΏ ≥ Eπ€∋x β£ Pr [π’, π’′ ∈ ππ€ ] ββ ππ€ ≤ πΎ β¦ · Pr [ππ€ ≤ πΎ] − ′ π€∋x π’,π’ ⊆π€ 600 β π’,π’′ ∋x (οΈ )οΈ2 πΏ πΏ (Claim 6.1.19) ≥ 1− · Pr [ππ€ ≤ πΎ] − π€∋x 20 600 (οΈ )οΈ2 (οΈ )οΈ Eπ€∋x [ππ€ ] πΏ πΏ · 1− (Markov) ≥ 1− − 20 πΎ 600 (οΈ )οΈ2 (οΈ )οΈ πΏ π + 2π −π‘ πΏ (Claim 6.1.20) ≥ 1− · 1− − 20 πΎ 600 −π‘ πΏ π + 2π πΏ ≥ 1− − − 10 πΎ 600 πΏ πΏ = 1− −π− 10 600 πΏ πΏ ≥ 1− − 5 600 We are now ready to prove the main theorem. 78 Proof of Theorem 6.1.18. We will define a function π : Fπ π → Fπ and then show that it is close to π and is a codeword of π π‘βπ . For x ∈ Fπ π , define π(x) , Majorityπ’∋x {ππ’ (x)}, where πΏ the majority is over π‘-dimensional affine subspaces π’ through x. Since 5πΏ + 600 < 12 , it follows from Lemma 6.1.21 that π is well-defined. Next, we show that π is close to π. Indeed, π = Eπ£ [πΏ(π|π£ , ππ£ )] ≥ Eπ’ [πΏ(π|π’ , ππ’ )] [οΈ [οΈ ]οΈ]οΈ = Eπ’ Ex∈π’ 1ππ’ (x)ΜΈ=π(x) [οΈ [οΈ ]οΈ]οΈ = Ex Eπ’∋x 1ππ’ (x)ΜΈ=π(x) [οΈ [οΈ ]οΈ ]οΈ β ≥ Ex Eπ’∋x 1ππ’ (x)ΜΈ=π(x) β π(x) ΜΈ= π(x) · Pr[π(x) ΜΈ= π(x)] β [οΈ ]οΈ x β ≥ Ex Pr [ππ’ (x) = π(x)] β π(x) ΜΈ= π(x)] · πΏ(π, π) π’∋x )οΈ (οΈ πΏ · πΏ(π, π). (Lemma 6.1.21) ≥ 1− 4 Finally, we show that π ∈ π π‘βπ . Let π’ ⊆ Fπ π a π‘-dimensional affine subspace. We wish to show that π|π’ ∈ π. Let π′π’ ∈ π be the codeword of π nearest to π|π’ (not to be confused with ππ’ , the nearest codeword to π|π’ ). Let x ∈ π’. We will show that π′π’ (x) = π|π’ (x). For a 4π‘-dimensional affine subspace π€ ⊆ Fπ π , we say a point y ∈ π€ is good for π€ if Prπ’′ ⊆π€ [ππ’′ (y) = π’′ ∋y π(y)] > πΏ . 20 We will show, by a union bound, that there exists a 4π‘-dimensional affine subspace π€ ⊇ π’ such that 1. ππ€ ≤ πΎ; 2. x is good for π€; 3. more than 1 − πΏ 2 fraction of points y ∈ π’ are good for π€. Observe that for any y ∈ π’, picking a random 4π‘-dimensional π€ containing π’ and then picking a random π‘-dimensional π’′ ⊆ π€ through y that intersect π’ only on π¦ is equivalent to picking a random π‘-dimensional π’′ through y that intersect π’ only on π¦ and then picking a 79 random 4π‘-dimensional π€ containing both π’, π’′ . Therefore, for any fixed y ∈ π’ β‘ β€ Eπ€⊇π’ β£ Pr [ππ’′ (y) ΜΈ= π(y)]β¦ = E ′ [οΈ 1ππ’′ (y)ΜΈ=π(y) ≤ E [οΈ 1ππ’′ (y)ΜΈ=π(y) | π’ ∩ π’′ = {y} π€⊇π’ π’′ ⊆π€,π’′ ∋y π’ ⊆π€ π’′ ∋y π€⊇π’ π’′ ⊆π€,π’′ ∋y + Pr π€⊇π’ π’′ ⊆π€,π’′ ∋y ]οΈ ]οΈ [π’ ∩ π’′ ΜΈ= {y}] [οΈ ]οΈ (Lemma B.2.2) ≤ Eπ’′ ∋y 1ππ’′ (y)ΜΈ=π(y) | π’ ∩ π’′ = {π¦} + π −2π‘ [οΈ ]οΈ (Lemma B.2.2) ≤ Eπ’′ ∋y 1ππ’′ (y)ΜΈ=π(y) + π −(π−2π‘) + π −2π‘ πΏ πΏ πΏ πΏ πΏ + + 2π −2π‘ ≤ + ≤ . (Lemma 6.1.21 and definition of π) ≤ 5 600 5 300 4 Therefore, by Markov’s inequality, for any fixed y ∈ π’, [οΈ πΏ Pr [y is not good for π€] = Pr ′ Pr ′ [ππ’′ (y) ΜΈ= π(y)] ≥ 1 − π€⊇π’ π€⊇π’ π’ ⊇π€,π’ ∋y 20 ]οΈ πΏ 4 ≤ πΏ 1 − 20 5 ≤ · πΏ. 19 In particular, this applies for y = x. Further applying Markov’s inequality, we find that [οΈ ]οΈ 10 πΏ 5πΏ/19 Pr fraction of not good y in π’ ≥ = . ≤ π€⊇π’ 2 πΏ/2 19 Finally, since Eπ€⊇π’ [ππ€ ] ≤ π + 2π −π‘ (by Claim 6.1.20), we have π + 2π −π‘ πΏ Pr [ππ€ > πΎ] ≤ =π≤ . π€⊇π’ πΎ 10 Since πΏ ≤ 1 and 5 19 + 10 19 + 1 10 < 1, by the union bound such a desired π€ exists. Now that we have such a subspace π€, consider ππ€ . We claim that it suffices to prove that if y ∈ π’ is good, then ππ€ (y) = π(y). Indeed, since more than 1 − in π’ are good, we have πΏ(ππ€ |π’ , π|π’ ) < πΏ . 2 πΏ 2 fraction of points Therefore ππ€ |π’ = π′π’ , and since x is good, we 80 have π(x) = ππ€ (x) = π′π’ (x) as desired. It remains to prove that ππ€ (y) = π(y) for good y ∈ π’. By Claim 6.1.19, at least 1 − πΏ 20 fraction of π‘-dimensional π’′ ⊆ π€ through y satisfy ππ’′ (y) = ππ€ (y). Since y is good, more than πΏ 20 fraction of π‘-dimensional π’′ ⊆ π€ through y satisfy ππ’′ (y) = π(y). Therefore, there must be some π‘-dimensional π’′ ⊆ π€ through y which satisfies ππ€ (y) = ππ’′ (y) = π(y). Finally, for the robustness statement: if π −π‘ ≥ is at least 6.2 π −3π‘ 2 ≥ πΏ 72 . 2·1052 πΏ 24 , 1014 then by Corollary 6.1.2, the robustness Otherwise, the robustness is at least πΌ1 πΏ 3 57,600·400 − 3π −π‘ ≥ πΏ 24 . 2·1014 Technical Algebraic Results The purpose of this section is to prove Theorem 6.2.6 and its Corollaries 6.2.7 and 6.2.8. If we allow our robustness in Theorem 6.1.18 to depend on π‘, the dimension of the base code, then proving what we need for Theorem 6.2.6 is easy. However, removing the dependence on π‘ requires some new ideas, including the definition of a new operation (“degree lifting”) on codes, and the analysis of the distance of degree lifted codes. In Section 6.2.1, we define degree lifting and analyze the degree lifted codes (Proposition 6.2.4). In Section 6.2.2, we prove Theorem 6.2.6 and its corollaries. 6.2.1 Degree Lift In this section, we define the degree lift operation on codes with degree sets. The operation can be thought of as “Reed-Mullerization”, in the sense that the degree lift of the ReedSolomon code of degree π is the Reed-Muller code of degree π. This resembles the degree lift operation of Ben-Sasson et al. [BGK+ 13] who defined a “Reed-Mullerization” for algebraicgeometry codes (in contrast, we want to define it for codes over Fπ π spanned by monomials). Definition 6.2.1 (Degree lift). Let π ⊆ {Fπ π → Fπ } have degree set Deg(π). For positive integer π ≥ 1, define the π -wise degree lift π(π ) ⊆ {Fππ → Fπ } of π to be the code with π 81 degree set {οΈ Deg(π(π )) , (d1 , . . . , dπ ) ∈ {0, 1, . . . , π − 1}π×π β }οΈ π β ∑οΈ β dπ ∈ Deg(π) . β β π=1 Our goal with this definition is to prove Proposition 6.2.4, which says that the distance of π(π ) is nearly the same as the distance of π. One can show that πΏ(π(π )) ≥ πΏ(π) − ππ −1 To do so, we will use the following fact. Proposition 6.2.2. Let π‘, π ≥ 1 and let π = ππ‘. Let π0 ⊆ {Fπ‘π → Fπ } be linear affineinvariant and let π , (π0 )⊗π . For each π ∈ [π], let Xπ = (ππ1 , . . . , πππ‘ ). If π (X1 , . . . , Xπ ) ∈ π, and π΄1 , . . . , π΄π : Fπ‘π → Fπ‘π are affine transformations, then π (π΄1 (X1 ), . . . , π΄π (Xπ )) ∈ π. Proof. By linearity, it suffices to consider the case where π (X1 , . . . , Xπ ) = ∏οΈπ π=1 Xπ dπ is a monomial, where dπ = (ππ1 , . . . , πππ‘ ) ∈ {0, 1, . . . , π − 1}π‘ . Each Xπ dπ ∈ π0 , so by affineinvariance π΄π (Xπ )dπ ∈ π0 . Therefore, by Proposition A.2.2, π (π΄1 (X1 ), . . . , π΄)π (Xπ )) = ∏οΈπ dπ ∈ (π0 )⊗π = π. π=1 π΄π (Xπ ) Overview. To prove Proposition 6.2.4, we show, through Lemma 6.2.3, that there is a special subset of π-dimensional subspaces π΄, such that for any π ∈ π(π ), π |π΄ ∈ π. Then, we analyze the distance of from π to the zero function by looking at the distance on a random special π-dimensional π΄. This will yield a distance of πΏ(π(π )) ≥ πΏ(π) − π(1) as long as the special subspaces sample Fπ π well. However, we require the π(1) term to be (ππ −π‘ )π(1) , otherwise we would not be able to remove the dependence on π‘ in the robustness of Theorem 6.1.18. In order to do so, we need to further assume that π is the tensor product (π0 )π of some π‘-dimensional code π0 (which is satisfied by our use case). We now describe the special subspaces we consider in Lemma 6.2.3. Label the variables of ππ‘π Fππ by ππππ , where π ∈ [π], π ∈ [π‘], π ∈ [π ]. Let πππ , for π ∈ [π], π ∈ [π‘], be the variables π = Fπ parameterizing π΄. Note that an arbitrary subspace restriction corresponds to substituting, for each ππππ , an affine function of all of the variables π11 , . . . , πππ‘ . This is too much to hope for. However, if we substitute for ππππ an affine function of just ππ1 , . . . , πππ‘ , this works. 82 Lemma 6.2.3. Let π‘, π ≥ 1 and π = ππ‘. Let π0 ⊆ {Fπ‘π → Fπ } be linear affine-invariant and let π , (π0 )⊗π . Let π ≥ 1, and let π (X) ∈ π(π ), with variables X = (ππππ )π∈[π],π∈[π‘],π∈[π ] . Let π(π11 , . . . , πππ‘ ) be the π-variate polynomial obtained from π (X) by setting, for each π ∈ ∑οΈ [π], π ∈ [π‘], and π ∈ [π ], ππππ = π‘π=1 πππππ πππ + ππππ , for some πππππ , ππππ ∈ Fπ . That is, for all (π, π, π) ∈ [π] × [π‘] × [π ] ππππ is an affine function of ππ1 , . . . , πππ‘ . Then π ∈ π. Proof. By linearity, it suffices to consider the case where π (X) = ∏οΈπ ∏οΈπ‘ π=1 π=1 ∏οΈπ π=1 π πππππππ is a monomial, for some 0 ≤ ππππ ≤ π − 1. For each π ∈ [π ], define dπ , (π11π , . . . , πππ‘π ), so that ∑οΈ (d1 , . . . , dπ ) ∈ Deg(π(π )), i.e. π π=1 dπ ∈ Deg(π). Then π(π11 , . . . , πππ‘ ) = (οΈ π‘ π ∏οΈ π‘ ∏οΈ π ∏οΈ ∑οΈ π=1 π=1 π=1 )οΈππππ πππππ πππ + ππππ π=1 π ∏οΈ π‘ ∏οΈ π π‘ ∏οΈ ∑οΈ (οΈππππ )οΈ π ∏οΈ πππππ πππππ πππ0 = ππππ πππππ πππ eπππ π=1 π=1 π=1 e ≤π π=1 πππ = ∑οΈ (· · · ) eπππ ≤π ππππ ∀ π,π πππ π ∏οΈ π‘ ∏οΈ ∑οΈπ‘ πππ π=1 ∑οΈπ π=1 πππππ π=1 π=1 where the (· · · ) denotes constants in Fπ . So, it suffices to show that each monomial of the ∑οΈπ‘ ∑οΈπ ∏οΈ ∏οΈ πππππ form ππ=1 π‘π=1 πππ π=1 π=1 ∈ π, which we show in the remainder of the proof. Let β(π11 , . . . , πππ‘ ) be the π-variate polynomial obtained from π by substituting ππππ = ∏οΈ ∏οΈ ∏οΈ π πππ for each π ∈ [π], π ∈ [π‘], π ∈ [π ]. Then β(π11 , . . . , πππ‘ ) = ππ=1 π‘π=1 π π=1 πππ πππ = ∑οΈπ (οΈ∑οΈ )οΈ ∏οΈπ ∏οΈπ‘ ∑οΈπ ∑οΈπ π π=1 ππππ π is a monomial with degree π , . . . , π = π=1 π=1 ππ π=1 11π π=1 ππ‘π π=1 dπ ∈ Deg(π), hence β ∈ π. Now, consider applying an affine transformation as follows: for each ∑οΈπ‘ 1 ≤ π ≤ π and each 1 ≤ π ≤ π‘, substitute πππ ← π=1 πΌπππ πππ + π½ππ , and call the new 83 polynomial β′ . By Proposition 6.2.2, β′ ∈ π. On the other hand, β′ (π1 , . . . , ππ ) = (οΈ π‘ π ∏οΈ π‘ ∏οΈ ∑οΈ π=1 π=1 = )οΈ∑οΈπ π=1 ππππ πΌπππ πππ + π½ππ π=1 (οΈ π‘ π ∏οΈ π‘ ∏οΈ π ∏οΈ ∑οΈ π=1 π=1 π=1 π=1 π ∏οΈ π‘ ∏οΈ π ∏οΈ ∑οΈ )οΈππππ πΌπππ πππ + π½ππ )οΈ π‘ ∏οΈ ππππ πππππ πππππ ππππ0 = π½ππ πππ πΌπππ eππππ π=1 π=1 π=1 eπππ ≤π ππππ π=1 )οΈ (οΈ (οΈ )οΈ π π‘ (οΈ π‘ ∑οΈπ‘ ∑οΈπ ∑οΈ ∏οΈ ππππ )οΈ π ∏οΈ ∏οΈ ∏οΈ ∑οΈπ πππππ πππππ πππ0 π=1 = πππ π=1 π=1 π½ππ πΌπππ eππππ π=1 π=1 π,π,π π=1 e ≤ π πππ (οΈ π πππ ∀ π,π,π and since the πΌπππ and the π½ππ are arbitrary and π has a degree set Deg(π) = Deg(π0 )π , each ∑οΈπ‘ ∑οΈπ ∏οΈ ∏οΈπ‘ π=1 π=1 πππππ π ∈ π, as desired. monomial π− π=1 ππ π=1 Proposition 6.2.4. Let π‘, π ≥ 1 and π = ππ‘. Let π0 ⊆ {Fπ‘π → Fπ } be linear affine-invariant and let π , (π0 )⊗π . For any positive integer π ≥ 1, πΏ(π(π )) ≥ πΏ(π) − ππ −π‘ = πΏ(π0 )π − ππ −π‘ . Proof. Let π (X) ∈ π(π ) be a nonzero codeword with variables X = (ππππ )π∈[π],π∈π‘],π∈[π ] . For π ∈ [π], π ∈ [π], π ∈ [π ], π ∈ [π‘], let πππππ , ππππ ∈ Fπ , and let a , (πππππ )π∈[π],π∈[π‘],π∈[π ],π∈[π‘] and b , (ππππ )π∈[π],π∈[π‘],π∈[π ] . Let πa,b (π11 , . . . , πππ‘ ) be the π-variate polynomial obtained by setting ∑οΈ ππππ = π‘π=1 πππππ πππ + ππππ for each 1 ≤ π ≤ π and 1 ≤ π ≤ π‘. By linearity of π and thus of π(π ), it suffices to show that πΏ(π, 0) ≥ πΏ(π) − ππ −π‘ . Let b ∈ Fππ‘π be a point such that π (b) ΜΈ= 0. Consider choosing a uniformly at random. Then π πa,b ΜΈ= 0 since πa,b (0) = π (b) ΜΈ= 0. For fixed π¦11 , . . . , π¦ππ‘ , as long as for each π ∈ [π] there ∑οΈ is some π ∈ [π‘] such that π¦ππ ΜΈ= 0, then the points π‘π=1 πππππ π¦ππ + ππππ are independent and 84 uniform over Fπ . This occurs with probability at least 1 − ππ −π‘ . Therefore, πΏ(π) ≤ Ea [πΏ (πa,b , 0)] [οΈ [οΈ ]οΈ]οΈ = Ea Ey 1πa,b (y)ΜΈ=0 [οΈ [οΈ ]οΈ]οΈ = Ey Ea 1πa,b (y)ΜΈ=0 [οΈ ]οΈ ≤ ππ −π‘ + EyΜΈ=0 1πa,b (y)ΜΈ=0 ]οΈ [οΈ = ππ −π‘ + EyΜΈ=0 1π ((∑οΈπ‘π=1 πππππ π¦ππ +ππππ ))ΜΈ=0 = ππ −π‘ + πΏ(π, 0). 6.2.2 Analysis of Subspace Restrictions In this section we prove Theorem 6.2.6 and its corollaries. Overview. Corollary 6.2.7 says that if a codeword π of the tensor product π ⊗π of a π‘dimensional code π is not a codeword of π π‘βππ‘ , then on there is a point b such that on many π‘-dimensional subspaces π’ through b, the restriction π |π’ ∈ / π. We use this in the proof of Theorem 6.1.4 when arguing that if a tensor codeword π ∈ π ⊗π satisfies π ∈ πa (see overview) βοΈ for many a, then π ∈ a πa = π 1βπ . A special case of Corollary 6.2.8 says that if π is a lifted Reed-Solomon codeword but not a Reed-Muller codeword, then on many planes π is not a bivariate Reed-Muller code. The actual corollary merely generalizes this to arbitrary π‘ and codes π0 , π1 . Both Corollaries 6.2.7 and 6.2.8 are proved in a similar manner. Note that both are statements of the form “if π is in some big code but not in a lifted code, then on many subspaces it is not a codeword of the base code”. A natural approach is to write π out as a linear combination of monomials, restrict to an arbitrary subspace of the appropriate dimension, re-write the restriction as a linear combination of monomials in the parameterizing variables, and note that the coefficients of the monomials are functions in the parameterization coeffi85 cients. Since π is not in the lift, there is a monomial outside the base code whose coefficient (the “offending coefficient”) is a nonzero function. Then, one shows that these functions belong to a code with good distance, so for many choices of parameterizing coefficients, the offending coefficient is nonzero. Theorem 6.2.6 abstracts the above approach and shows that, in the case of Corollary 6.2.7, the offending coefficient is a codeword of the degree lift (π ⊗π )(π‘) of π ⊗π , and in the case of Corollary 6.2.8, the offending coefficient is a codeword of a lifted code. This necessitates the analysis of the distance of degree lifted codes, hence the need for Section 6.2.1. Lemma 6.2.5. Let π ⊆ {Fπ π → Fπ } be a linear code with a π-shadow-closed degree set. If π ∈ π, and (οΈ π π10 + π‘ ∑οΈ π1π ππ , . . . , ππ0 + π=1 π‘ ∑οΈ )οΈ πππ ππ π(π‘+1) πe (a) = πe (a) · Ye e∈{0,1,...,π−1}π‘ π=1 where a = (πππ )1≤π≤π;0≤π≤π‘ ∈ Fπ ∑οΈ = , then, for every e ∈ {0, 1, . . . , π − 1}π‘ , ∑οΈ d∈π· E≤π d βE*π β=ππ ∀π πd · )οΈ π (οΈ ∏οΈ ππ π=1 Eπ* πππ0π0 π‘ ∏οΈ π πππππ . π=1 In particular, 1. πe ∈ π(π‘ + 1), the (π‘ + 1)-wise degree lift of π (see Definition 6.2.1); 2. if π = (π0 )1βπ for some linear affine-invariant code π0 ⊆ {Fπ → Fπ }, then πe ∈ (π0 )1βπ(π‘+1) Proof. Let π· be the degree set of π. Write π (X) = 86 ∑οΈ d∈π· πd · Xd . Let π΄ : Fπ‘π → Fπ π be the (οΈ )οΈ ∑οΈ ∑οΈ affine map Y β¦→ π10 + π‘π=1 π1π ππ , . . . , ππ0 + π‘π=1 πππ ππ . Expanding, we get (π β π΄)(Y) = ∑οΈ d∈π· πd · π ∏οΈ (οΈ ππ0 + π‘ ∑οΈ π=1 )οΈππ πππ ππ π=1 β β π‘ ∑οΈ (οΈππ )οΈ ∏οΈ π π β = πd · πππ0π0 · πππππ ππ ππ β eπ π=1 π=1 d∈π· eπ ≤π ππ (οΈ π (οΈ )οΈ )οΈ π‘ π‘ ∑οΈ ∑οΈ ∏οΈ ππ ∏οΈ ∏οΈ βE β πππ ππ0 = πd · ππ0 πππ · ππ *π Eπ* π=1 π=1 π=1 d∈π· E≤π d )οΈ π (οΈ π‘ ∑οΈ ∑οΈ ∏οΈ ∏οΈ ππ π e = Y · πd · πππ0π0 πππππ E π* π‘ π=1 π=1 d∈π· ∑οΈ π ∏οΈ e∈{0,1,...,π−1} E≤π d βE*π β mod* π=ππ ∀π and therefore, for each e ∈ {0, 1, . . . , π − 1}π‘ , πe (a) = ∑οΈ d∈π· E≤π d βE*π β mod* π=ππ ∀π )οΈ π (οΈ π‘ ∏οΈ ∏οΈ ππ π ππ0 πd · ππ0 πππππ . Eπ* π=1 π=1 View the variables a = (πππ ) in the order (π10 , . . . , ππ0 , . . . , π1π‘ , . . . , πππ‘ ), and interpret E as (E*0 , . . . , E*π‘ ). If E ≤π d and d ∈ π· = Deg(π), then E ∈ Deg(π(π‘ + 1)). Therefore, πe ∈ π(π‘ + 1). Now suppose π = (π0 )1βπ for some linear affine-invariant π0 ⊆ {Fπ → Fπ }. It suffices to show that if d = (π1 , . . . , ππ ) ∈ Deg(π) and E ≤π d with entries πππ , π ∈ [π], 0 ≤ π ≤ π‘, then the length π(π‘ + 1) vector (E*0 , E*1 , . . . , E*π‘ ) ∈ Deg((π0 )1βπ(π‘+1) ). By Proposition 5.2.3, it ∑οΈ suffices to show that, if π’ππ ≤π πππ for every π ∈ [π] and 0 ≤ π ≤ π‘, then ππ π’ππ mod* π ∈ Deg(π0 ). Since d ∈ Deg(π) = Deg((π0 )1βπ ), this implies that if π′π ≤π ππ for π ∈ [π], then ∑οΈ ′ ∑οΈπ‘ * ′ π mod π ∈ Deg(π ). Set π , 0 π π π π=0 π’ππ . Observe that, since (ππ0 , ππ1 , . . . , πππ‘ ) ≤π ππ , this ∑οΈ ∑οΈ implies that π′π ≤π ππ . Therefore, ππ π’ππ mod* π = π π′π mod* π ∈ Deg(π0 ), as desired. Theorem 6.2.6. Let 1 ≤ π‘ < π. Let π1 ⊆ {Fπ‘π → Fπ } be a linear affine-invariant code, and π‘βπ let π2 ⊆ {Fπ . Then the π → Fπ } have a π-shadow-closed degree set. Suppose π ∈ π2 β π1 87 following hold: 1. if π2 = (π0 )⊗π for some linear affine-invariant code π0 ⊆ {Fπ‘π → Fπ }, where π = ππ‘, π −π‘ then there exists a point b ∈ Fπ fraction of π such that for at least πΏ(π0 ) − (π + 1)π / π1 ; π‘-dimensional affine subspaces π΄ ⊆ Fπ π passing through b, the restriction π |π΄ ∈ 2. if π2 = (π0 )1βπ for some linear affine-invariant code π0 ⊆ {Fπ → Fπ }, then for at least / π1 . πΏ(π0 ) − π −1 fraction of π‘-dimensional affine subspaces π΄ ⊆ Fπ π , the restriction π |π΄ ∈ Proof. Let π be the characteristic of Fπ . Let π΄ be parameterized by ππ = ππ0 + ∑οΈπ‘ π=1 πππ ππ , π×π‘ has full rank. Write where the matrix {πππ }π,π‘ π=1,π=1 ∈ Fπ π |π΄ (Y) = ∑οΈ πe (a) · Ye . e∈{0,1,...,π−1}π‘ / Deg(π1 ) such that πe ΜΈ= 0. Since π ∈ / π1π , there exists e ∈ 1. By Corollary A.2.3, π2 has a π-shadow-closed degree set. By Lemma 6.2.5 (1), πe ∈ π2 (π‘ + 1). For each b = (π1 , . . . , ππ ) ∈ Fπ π , let πe,b denote the polynomial πe with the variable ππ0 fixed to value ππ for each π ∈ [π] (i.e. insisting that π΄ passes through b). Observe that each πe,b ∈ π2 (π‘). Since πe ΜΈ= 0, there exists b ∈ Fπ π such that πe,b ΜΈ= 0. By Proposition 6.2.4, for at least πΏ(π2 (π‘)) ≥ πΏ(π0 )π − ππ −π‘ fraction of matrices {πππ }π∈[π];π∈[π‘] , we have πe,b ({πππ }π∈[π];π∈[π‘] ) ΜΈ= 0. Since, by Lemma B.2.2, at least 1−π π‘−π fraction of such matrices have full rank, we get that for at least πΏ(π0 )π − ππ −π‘ − π π‘−π ≥ (οΈ )οΈ πΏ(π0 )π −(π+1)π −π‘ of the full rank matrices satisfy πe,b {πππ }π,π‘ π=1,π=1 ΜΈ= 0 , and therefore π |π΄ (Y) ∈ / π1 . 2. By Proposition 4.1.3 it has a π-shadow-closed degree set. By Lemma 6.2.5 (2), πe ∈ (π0 )1βπ(π‘+1) , so πe (a) ΜΈ= 0 for at least πΏ((π0 )1βπ(π‘+1) ) ≥ πΏ(π0 )−π −1 fraction of choices a (including such that the corresponding matrix does not have full rank), and therefore, by Lemma B.2.2, π |π΄ (Y) ∈ / π1 for at least πΏ(π0 ) − π −1 − π π‘−π ≥ πΏ(π0 ) − 2π −1 . 88 Corollary 6.2.7. Let π‘, π ≥ 1 and let π = ππ‘. Let π ⊆ {Fπ‘π → Fπ } be a linear affineπ −π‘ invariant code. If π ∈ π ⊗π βπ π‘βπ , then there is a point b ∈ Fπ π such that for πΏ(π) −(π+1)π fraction of π‘-dimensional subspaces π’ through b, the restriction π |π’ ∈ / π. Proof. Follows immediately from Theorem 6.2.6 (1) with π0 = π1 = π, and π2 = π ⊗π . Corollary 6.2.8. Let 1 ≤ π‘ ≤ π. Let π0 ⊆ {Fπ → Fπ } be a linear affine-invariant code. Let π1 ( (π0 )1βπ‘ be a linear affine-invariant code that is a strict subcode of (π0 )1βπ‘ . If π ∈ (π0 )1βπ β (π1 )π‘βπ , then for at least πΏ(π0 ) − 2π −1 fraction of π‘-dimensional subspaces / π1 . π΄ ⊆ Fπ π , the restriction π |π΄ ∈ Proof. Follows immediately from Theorem 6.2.6 (2) with π2 = (π0 )1βπ . 89 90 Chapter 7 Applications 7.1 Lifted Reed-Solomon Code We begin by describing the central construction of the thesis — the lifted Reed-Solomon. As its name suggests, the lifted Reed-Solomon is simply obtained by lifting the Reed-Solomon code. With judicious choice of parameters, the lifted Reed-Solomon code is the first code to achieve a combination of parameters never achieved by one code before. For now, we describe our construction for an arbitrary choice of parameters. Let π be a prime power, let π ≥ 2 be an integer, and let π < π. Let RS , RS(π, π) and let π , RS1βπ . The code π is the lifted Reed-Solomon code. It automatically inherits distance, a well-structured degree set, local correctability, decodability, and robust testability by virtue of being a lifted code. Proposition 7.1.1. The distance of π is πΏ(π) ≥ 1 − π+1 . π Proof. By Proposition 5.2.4, πΏ(π) ≥ πΏ(RS) − π −1 , and by Proposition 3.4.2, πΏ(RS) = 1 − ππ , so πΏ(π) ≥ πΏ(RS) − π −1 ≥ 1 − π+1 . π Proposition 7.1.2. The code π is linear affine-invariant and d ∈ Deg(π) if and only if, for ∑οΈ * every e ≤π d, the sum π π=1 ππ mod π ≤ π. Proof. Follows immediately from Proposition 5.2.3. 91 (οΈ )οΈ Proposition 7.1.3. For every π, π > 0, the code π is π, ( 12 − π)πΏ − π −1 , π -locally correctable and decodable, where πΏ , 1 − π π and π = π(ln(1/π)/π2 ). Proof. The local correctability follows immediately from Theorem 5.3.2 while local decodability follows from Theorem 5.3.4. Proposition 7.1.4. Let πΏ , 1 − ππ . The code π has a π 2 -local tester that is πΏ 72 -robust. 2·1052 Proof. Follows immediately from Theorem 5.4.3. 7.1.1 Relationship to Reed-Muller We proceed by examining the relationship between the lifted Reed-Solomon code and the Reed-Muller code. It follows immediately from definitions that the Reed-Muller code is contained in the lifted Reed-Solomon code, i.e. RM(π, π, π) ⊆ RS(π, π)1βπ . Kaufman and Ron [KR06] proved the following characterization: a polynomial over Fπ π has degree π if and ⌈οΈ ⌉οΈ π+1 only if on every π‘-dimensional affine subspace its restriction has degree π, where π‘ = π−π/π and π is the characteristic of Fπ . This generalizes a result of Friedl and Sudan [FS95], which says that if π ≤ π − π/π and a polynomial has degree π on every line, then its global degree is π, and that this is not necessarily true if π ≥ π − π/π. Observe that, in the language of lifting, this statement simply says that RS(π, π)1βπ = RM(π, π, π) if and only if π < π − π/π. In Theorem 7.1.6, we give a more efficient proof of the above result of [FS95] using the technology of affine-invariance and lifting. In Theorem 7.1.8, we prove a specialization of the characterization of [KR06]: if π − π/π ≤ π < π, then polynomials of degree π are characterized by having degree π on planes, i.e. RM(π, π, 2)2βπ = RM(π, π, π), again using affine-invariance and lifting. First, we prove a handy lemma. Lemma 7.1.5. Let π ≥ π‘, let π be a prime power, and let π < π. If RM(π, π, π) = RM(π, π, π‘)π‘βπ , then RM(π, π − 1, π) = RM(π, π − 1, π‘)π‘βπ . Proof. Let π ∈ RM(π, π − 1, π‘)π‘βπ . Define π : Fπ π → Fπ by π(X) , π1 · π (X). Since π ∈ RM(π, π − 1, π‘)π‘βπ , by Proposition 5.1.2 it follows that deg(π β π΄) ≤ π − 1 < π − 1 92 for every affine π΄ : Fπ‘π → Fπ π Note that (π β π΄)(Y) = π(π΄(Y)) = (π΄(Y))1 · (π β π΄)(Y), so π‘βπ = deg(π β π΄) ≤ deg(π β π΄) + 1 ≤ π for every affine π΄ : Fπ‘π → Fπ π , hence π ∈ RM(π, π, π‘) RM(π, π, π), so deg(π) ≤ π. Therefore, deg(π ) = deg(π) − 1 ≤ π − 1. Theorem 7.1.6. Let π be a prime and let π be a power of π. Let π < π. Let π ≥ 2. Then RM(π, π, π) = RS(π, π)1βπ if and only if π < π − π/π. Proof. By Lemma 7.1.5, it suffices to show that equality holds for π = π − π/π − 1 and does not hold for π = π − π/π. Let π = π − π/π − 1 = (1 − π−1 )π − 1. This corresponds to the construction of π , RS(π, π)1βπ in Section 7.1 with π = 1. Let π ≥ 1 be such that π = ππ . ∑οΈ Let d = (π1 , . . . , ππ ) ∈ Deg(π). We will show that π π=1 ππ ≤ π = π − π/π − 1. Suppose, ∑οΈπ for the sake of contradiction, that π=1 ππ ≥ π − π/π. We will exhibit e ≤π d such that ∑οΈ π − π/π ≤ π π=1 ππ < π, which contradicts the fact that d ∈ Deg(π) by Proposition 5.2.3. If ∑οΈπ (π ) ∑οΈπ ∑οΈπ π=1 ππ . π=1 ππ ≥ π. Let π , π=1 ππ < π, then we are done by taking e = d, so assume (π ) By Claim 7.1.9, π ≤ π − 2. For each π ∈ [π], let ππ = ππ − ππ ππ −1 . Then π ∑οΈ ππ = π=1 π ∑οΈ ππ − π π −1 π=1 π ∑οΈ (π ) ππ ≥ (π − π)ππ −1 . π=1 ∑οΈ Let π ∈ [π] be the minimal integer such that ππ=1 ππ ≥ (π−π−1)ππ −1 . Since each ππ < ππ −1 , ∑οΈ this implies that ππ=1 ππ < (π − π)ππ −1 . Now, for π ∈ [π], define ππ , ππ , and for π > π, define (π ) ππ , ππ ππ −1 . By construction, e ≤π d. On the other hand, π ∑οΈ ππ = π=1 π ∑οΈ ππ + π π −1 π ∑οΈ (π ) ππ π=1 π=1 = π ∑οΈ ππ + πππ −1 π=1 which is at least (π − π − 1)ππ −1 + πππ −1 = (π − 1)ππ −1 = π − π/π and strictly less than (π − π)ππ −1 + πππ −1 = ππ = π. π−π/π Now, let π = π − π/π. Let π (X) = π1 π−π/π π2 93 . Clearly deg(π ) = 2(π − π/π) > π − π/π. We claim that π ∈ RS(π, π)1βπ . For any a, b ∈ Fπ π , we have π (aπ + b) = (π1 π + π1 )π−π/π (π2 π + π2 )π−π/π π/π π/π π/π (7.1) π/π = (π1 π π/π + π1 )π−1 (π2 π π/π + π2 )π−1 (7.2) (7.3) so, since π π = π , every monomial in π (aπ + b) is of the form π π·π/π for some 0 ≤ π ≤ π − 1, thus deg(π (aπ + b)) ≤ π − π/π = π. As a corollary, we see that, over prime fields, lifting the Reed-Solomon code simply yields the Reed-Muller code. Corollary 7.1.7. If π is a prime and π < π, then RS(π, π)1βπ = RM(π, π, π) for every π ≥ 1. Proof. If π = 1, then both codes are equal to RS(π, π), so there is nothing to prove. Assume π ≥ 2. If π = π − 1, then both codes are equal to everything, so again there is nothing to prove. If π < π − 1, then the result follows from Theorem 7.1.6. Theorem 7.1.8. Let π be a prime power, let π < π, and let π ≥ 2. Then RM(π, π, 2)2βπ = RM(π, π, π). Proof. By Lemma 7.1.5, it suffices to consider the case where π = π−1. It follows immediately from definitions that RM(π, π, π) ⊆ RM(π, π, 2)2βπ , so it only remains to show the reverse inclusion. We do so by showing that if d ∈ Deg(RM(π, π, 2)2βπ ), then βdβ ≤ π. Actually, we show the converse. By Proposition 5.2.3, it suffices to show that if βdβ ≥ π, then there exists E ≤π d such that (βE*1 β mod* π) + (βE*2 β mod* π) ≥ π where E has rows [π] and columns {0, 1, 2}. Without loss of generality, assume π1 ≥ π2 ≥ · · · ≥ ππ . Let π ≥ 2 be the smallest integer such that π2 + · · · + ππ ≥ π − π1 . Then π2 + · · · + ππ ≤ π − 1, for otherwise ππ ≥ π1 + 1, which is impossible. Construct E as follows. Define E1* , (0, π1 , 0) and for π ∈ {2, . . . , π}, define Eπ* , (0, 0, ππ ). Finally, for π > π, define Eπ* = (0, 0, 0). By construction, E ≤π d, 94 and moreover βE*1 β = π1 and βE*2 β = π2 + · · · + ππ . Since βE*1 β, βE*2 β ≤ π − 1, applying mod* π does not reduce them, hence (βE*1 β mod* π) + (βE*2 β mod* π) = βE*1 β + βE*2 β ≥ π1 + (π − π1 ) = π, as desired. 7.1.2 Rate Rate is the one key parameter that is not guaranteed by the lifting operator. However, the potential in lifted codes comes from the fact that it is possible to get extremely dense codes by lifting. From Section 7.1.1, we see that if π − π/π ≤ π < π, then there are lifted ReedSolomon codewords that are not Reed-Muller codewords. We show that, in fact, there are many such codewords. Our strategy is to lower bound the rate of the code. Observe that the lifted Reed-Solomon code, being a linear affine-invariant code, is spanned by monomials, i.e. has a degree set. Therefore, its dimension is equal to the size of its degree set. We use our knowledge of the structure of its degree set to lower bound the number of such degrees. In this section, we analyze the rate of π for special values of π. Let π be the characteristic of Fπ , let π ≥ 1 be such that π = ππ , let π ≤ π and let π = (1 − π−π )π − 1. We begin by re-interpreting what it means for a number π to be less than π, in terms of ∑οΈ the π-ary expansion of π. Through this section, for any π ∈ N, let π = π≥0 π(π) ππ be the π-ary expansion of π. Claim 7.1.9. If π ∈ JπK, then π ≤ π if and only if π(π) < π − 1 for some π − π ≤ π < π . Proof. Since π < π, π(π) = 0 for π ≥ π . Note that if π(π) = π − 1 for π − π ≤ π < π , then ∑οΈ ∑οΈπ−1 π π π −π π ≥ (π − 1) π −1 (π − 1) π=0 π = (1 − π−π )π, while if π(π) < π − 1 for some π=π −π π = π π − π ≤ π < π , this only decreases the value of π. Next, we use our knowledge of the degree set of the lifted Reed-Solomon code to provide a sufficient condition for a degree to be in the degree set. ⌈οΈ ⌉οΈ Claim 7.1.10. Let π = logπ π + 1 and let d ∈ JπKπ . If there is some π − π ≤ π ≤ π − π (π) such that ππ = 0 for every π ∈ [π] and every π ≤ π < π + π, then d ∈ Deg(π). 95 Proof. Let e ≤π d, and let π = ∑οΈπ π=1 ππ mod* π. We claim that π(π+π−1) = 0, which implies π ≤ π Claim 7.1.9 since π − π ≤ π + π − 1 < π . Note that π β¦→ π · π mod* π results in a cyclic permutation of the digits π(π) . So we may multiply d and e by an appropriate power of π, namely ππ −π−π , so that π = π − π. Therefore, we may assume without loss of generality that π = π − π, and we wish to show that π(π −1) = 0, i.e. π < ππ −1 . Note that since e ≤π d, for each (π) π ∈ [π] and π − π ≤ π ≤ π − 1 we have ππ ∑οΈπ π −π < ππ−1 ππ −π = ππ −1 . π=1 ππ < ππ = 0, i.e. ππ < ππ −π for every π ∈ [π]. Therefore Finally, we lower bound the rate. ⌈οΈ ⌉οΈ ππ Theorem 7.1.11. Let π = logπ π + 1 The rate of the code π is at least 1 − π−(π+π)/(ππ ) . Proof. Consider choosing d ∈ JπKπ uniformly at random. Let π = ⌊π/π⌋. For π ∈ [π], let πΈπ (π) be the event that ππ = 0 for every π ∈ [π] and π − ππ ≤ π < π − (π − 1)π. By Claim 7.1.10, ]οΈ [οΈβοΈ βοΈπ π π=1 πΈπ . We have π=1 πΈπ is sufficient for d ∈ Deg(π), so we wish to lower bound Pr Pr[πΈπ ] = π−ππ for every π ∈ [π], therefore [οΈ Pr π βοΈ ]οΈ πΈπ [οΈ = 1 − Pr π=1 π βοΈ ]οΈ πΈπ (7.4) π=1 = 1− = 1− π ∏οΈ π=1 π ∏οΈ [οΈ ]οΈ Pr πΈπ (7.5) (1 − Pr[πΈπ ]) (7.6) π=1 = 1 − (1 − π−ππ )π (7.7) ≥ 1 − (1 − π−ππ )(π+π)/π (7.8) ≥ 1 − π−(π+π)/(ππ 96 ππ ) . (7.9) 7.1.3 Global List-Decoding In this section, we present an efficient global list decoding algorithm for RS(π, π)1βπ . Define πΌ1 , . . . , πΌπ ∈ Fππ , π, and π* as in Section 5.3.3. Our main result states that RS(π, π)1βπ is isomorphic to a subcode of RS(π π , (π + π)π π−1 ) ⊆ {Fππ → Fπ }. In particular, one can simply list decode RS(π, π)1βπ by list-decoding RS(π π , (π + π)π π−1 ) up to the Johnson radius. We will use this algorithm for π = 2 as a subroutine in our local list decoding algorithm in Section 7.1.4. Theorem 7.1.12. If π ∈ RS(π, π)1βπ , then deg(π* (π )) ≤ (π + π)π π−1 . Proof. By linearity, it suffices to prove this for a monomial π (π1 , . . . , ππ ) = ∏οΈπ π=1 ππππ . We have ∑οΈ π* (π )(π) = (· · · )π ∑οΈ ππ πππ π π−π , (π11 ,...,π1π )≤π π1 .. . (ππ1 ,...,πππ )≤π ππ so it suffices to show that ∑οΈπ ∑οΈπ π=1 π=1 πππ π π−π mod* π π ≤ (π + π)π π−1 . By Proposition 7.1.2, ∑οΈ * for every ππ ≤π ππ , π ∈ [π], we have π π=1 ππ mod π ≤ π. Therefore, there is some integer ∑οΈ 0 ≤ π < π such that π π=1 ππ1 ∈ [ππ, π(π − 1) + π]. Thus, ππ π ≤ π π−1 ≤ π π−1 π ∑οΈ π=1 π ∑οΈ ππ1 + π ∑οΈ π ∑οΈ π=2 π=1 π ∑οΈ π ∑οΈ π−2 ππ1 + π π=1 ≤ (π(π − 1) + π)π and hence ∑οΈπ ∑οΈπ π=1 π=1 πππ π π−π πππ π π−π πππ π=2 π=1 π−1 π−1 + ππ (7.10) (7.11) (7.12) = π(π π − 1) + (π + π − π)π π−1 + π (7.13) ≤ π(π π − 1) + (π + π)π π−1 (7.14) mod* π π ≤ (π + π)π π−1 . Corollary 7.1.13. For every π ≥ 2 and ever π > 0, there is a polynomial time algorithm 2 that takes as input a function π : Fπ π → Fπ and outputs a list β of size |β| = π(1/π ) which 97 contains all π ∈ RS(π, π)1βπ such that πΏ(π, π) < 1 − √οΈ π+π π − π. ′ * Proof. Given π : Fπ π → Fπ , convert it to π = π (π), and then run the Guruswami-Sudan list decoder (Theorem 3.4.3) for RS , RS(π π , (π + π)π π−1 ) on π′ to obtain a list β of size √οΈ 2 ′ − π lies in β. |β| = π(1/π ) with the guarantee that any π ∈ RS with πΏ(π , π) < 1 − π+π π √οΈ We require that any π ∈ RS(π, π)1βπ satisfying πΏ(π′ , π) < 1 − π+π − π lies in β, and this π follows immediately from Theorem 7.1.12. 7.1.4 Local List-Decoding In this section, we present a local list decoding algorithm for LiftedRS(π, π, π), where π = √ (1 − πΏ)π which decodes up to radius 1 − 1 − πΏ − π for any constant π > 0, with list size π(1/π2 ) and query complexity π 3 . Theorem 7.1.14. For every π ≥ 2 and every πΏ, π > 0, setting π = (1 − πΏ)π, RS(π, π)1βπ is (οΈ (οΈ )οΈ)οΈ √ π, π 3 , 1 − 1 − πΏ − π, π(1/π2 ), 0.2 + πΏπ2 , π π61ππΏ -locally list-decodable. Proof. The following algorithm is the outer local list-decoding algorithm. Local list decoder: Oracle access to received word π : Fπ π → Fπ . 1. Pick an affine transformation β : Fπ → Fπ π uniformly at random. 2. Run Reed-Solomon list decoder (e.g. Guruswami-Sudan) on π β β from 1 − √ 1−πΏ− π 2 fraction errors to get list π1 , . . . , ππΏ : Fπ → Fπ of Reed-Solomon codewords. 3. For each π ∈ [πΏ], output Correct(π΄β,ππ ) where Correct is a local correction algorithm for the lifted codes for 0.1πΏ fraction errors given by Theorem 5.3.1, which has a failure probability of 0.2 + πΏπ2 , and π΄ is an oracle which takes as advice a univariate affine map and a univariate polynomial and simulates oracle access to a function which is supposed to be βͺ 0.1πΏ close to a lifted RS codeword. 98 Oracle π΄β,π (x): 1. Let π : F2π → Fπ π be the unique affine map such that π (π‘, 0) = β(π‘) for π‘ ∈ Fπ and π (0, 1) = x. 2. Use the global list decoder for RS(π, π)1β2 given by Corollary 7.1.13 to decode π β π √ from 1 − 1 − πΏ − 2π fraction errors and obtain a list β. 3. If there exists a unique β ∈ β such that β β β = π, output β(0, 1), otherwise fail. Analysis: For the query complexity, note that the local list-decoder makes π queries to π to output its oracles. For each oracle, Correct makes π queries to π΄β,ππ , which itself makes π 2 queries to π, so the entire oracle makes π 3 queries to π. To show correctness, we just have to show that, with high probability over the choice of √ β, for every lifted RS codeword π such that πΏ(π, π ) < 1 − 1 − πΏ − π, there is π ∈ [πΏ] such that Correct(π΄β,ππ ) = π , i.e. πΏ(π΄β,ππ , π ) ≤ 0.1πΏ. Fix such a function π . We will proceed in two steps: 1. First, we show that with high probability over β, there is some π ∈ [πΏ] such that π |β = ππ . 2. Next, we show that πΏ(π΄β,π ββ , π ) ≤ 0.1πΏ with high probability. √ For the first step, note that π |β ∈ {π1 , . . . , ππΏ } if πΏ(π β β, π β β) ≤ 1 − 1 − πΏ − 2π . By √ Proposition B.1.4, πΏ(π β β, π β β) has mean less than 1 − 1 − πΏ − π and variance less than √ 1/(4π). By Chebyshev’s inequality, the probability that πΏ(π |β , π|β ) ≤ 1 − 1 − πΏ − 2π is at least 1 − 1 . π2 π For the second step, we want to show that Prx∈Fπ [π΄β,π ββ (x) ΜΈ= π (x)] ≤ 0.1πΏ. First consider π the probability when we randomize β as well. Then π : F2π → Fπ π is a uniformly random affine map. We get π΄β,π ββ (x) = π (x) as long as π β π ∈ β and no other element β ∈ β has √ β β β = π β β. By Proposition B.1.4, πΏ(π β π, π β π ) has mean less than 1 − 1 − πΏ − π and variance less than 1/(4π 2 ). By Chebychev’s inequality, with probability at least 1 − π21π2 over √ β and x , we have πΏ(π |π , π|π ) ≤ 1 − 1 − πΏ − 2π and hence π |π ∈ β. For the probability 99 that no two codewords in β agree on β, view this as first choosing a random affine map ′ ′ ′ 2 π : F2π → Fπ π and then choosing random affine β : Fπ → Fπ and defining β , π β β . For any distinct π, π : F2π → Fπ , the probability over β′ that π β β′ = π β β′ is at most 2/π, by Proposition B.1.5. Since |β| = π(1/π2 ), there are at most π(1/π4 ) pairs of functions from β, so by the union bound, with probability at least 1 − π(1/(π4 π)), all the functions in β are distinct on β and hence π β π is the unique codeword in β which is consistent with π β β. (οΈ )οΈ Therefore, the probability, over β and x, that π΄β.π ββ (x) ΜΈ= π (x) is π π41π , and thus [οΈ ]οΈ Pr [πΏ(π΄β,π ββ , π β β) > 0.1πΏ] = Pr Pr[π΄β,π ββ (x) ΜΈ= π (x)] > 0.1πΏ β β x Prβ,x [π΄β,π ββ (x) ΜΈ= π (x)] ≤ (οΈ )οΈ0.1πΏ 1 = π 4 . π ππΏ √ So, for a fixed codeword π ∈ RS(π, π)1βπ such that πΏ(π, π ) < 1− 1 − πΏ−π, the probability (οΈ )οΈ 1 of success is π π4 ππΏ . Since there are π(1/π2 ) such codewords, the overall probability of (οΈ )οΈ success is π π61ππΏ . As a corollary, we get the following testing result. Theorem 7.1.15. For every πΏ > 0 and for any πΌ < π½ < 1 − √ 1 − πΏ, there is an algo- rithm which, given oracle access to a function π : Fπ π → Fπ , distinguishes between the cases where π is πΌ-close to RS(π, π)1βπ (where π , (1 − πΏ)π), and where π is π½-far, while making (οΈ )οΈ 4 π ln(1/(π½−πΌ))π queries to π. 4 72 (π½−πΌ) πΏ Proof. Let π , (πΌ + π½)/2, and let π , (π½ − πΌ)/8, so that πΌ = π − 4π and π½ = π + 4π. Set π , π(π2 ), where the constant is sufficiently large. Let π ′ be the π 2 -query Ω(πΏ 72 )-robust tester algorithm for π π(π, π)1βπ , which rejects words that are π-far with probability Ω(ππΏ 72 ), by Proposition 3.3.2. Let π be the π(ln(1/π)π 2 /(ππΏ 72 ))-query tester which runs π ′ repeatedly π(ln(1/π)π 2 /(ππΏ 72 )) times and accepts if and only if every iteration accepts, to increase the rejection probability for π-far words to 1 − π. Our algorithm is to run the local list-decoding 100 algorithm on π with error radius π, to obtain a list of oracles π1 , . . . , ππΏ . For each ππ , ΜοΈ ππ ) of the distance between π and the we use random sampling to compute an estimate πΏ(π, function ππ computed by ππ to within π additive error with failure probability π, and keep only the ones with estimated distance less than π + π. Then, for each remaining ππ , we run π on ππ . We accept if π accepts some ππ , otherwise we reject. The number of queries required to run the local list-decoding algorithm is π. For each ΜοΈ ππ ) is π(ln(1/π)/π2 ), π ∈ [πΏ], the number of queries to ππ we need to make for computing πΏ(π, by Proposition 2.2.4. Each query to ππ makes π 2 queries to π, for a total of π(ln(1/π)π 2 /π2 ) queries for each ππ and therefore π(ln 1/π)π 2 /π4 ) queries for the distance estimation step. The tester π makes π(ln(1/π)π 2 /(ππΏ 72 )) queries to each ππ , for a total of π(ln(1/π)π 4 /(ππΏ 72 )) queries to π for each π ∈ [πΏ] and therefore a grand total of π(ln(1/π)π 4 /(π3 πΏ 72 )) queries to π made by π . Therefore, the total number of queries to π made by our testing algorithm is (οΈ )οΈ (οΈ )οΈ 4 ln(1/π)π 4 π ln(1/π)π = π . π4 πΏ 72 π4 πΏ 72 If π is πΌ-close to RS(π, π)1βπ , then it is πΌ-close to some codeword π , and by the guarantee of the local list-decoding algorithm, there is some π ∈ [πΏ] such that ππ computes π . Since, ΜοΈ π ) ≤ πΌ + π < π, in which case ππ πΏ(π, π ) ≤ πΌ, with probability at least 1 − π we have πΏ(π, will not be pruned by our distance estimation. Since π is a codeword, this ππ will pass the testing algorithm π and so our algorithm will accept. The total failure probability is (οΈ )οΈ 1 therefore π + π π6 ππΏ . )οΈ (οΈ 1βπ 1 Now suppose π is π½-far from RS(π, π) . With probability 1 − π π6 ππΏ , the local listdecoding algorithm succeeds, so let us condition on its success. With probability 1−π(π/π2 ), all the distance estimations for all the ππ simultaneously succeed, and let us condition on this as well. Consider any oracle ππ output by the local list-decoding algorithm and not pruned by our distance estimation. Let ππ be the function computed by ππ . Then the estimated ΜοΈ ππ ) < π + π, so the true distance is πΏ(π, ππ ) < π + 2π. Since π is π½-far from any distance is πΏ(π, codeword, that means the distance from ππ to any codeword is at least π½ − (π + 2π) > π, and hence π will reject ππ with probability 1 − π. By the union bound, all the ππ not pruned will be rejected with probability 1 − π(π/π2 ). So the total failure probability in this case is 101 at most π(π/π2 ) + π (οΈ 1 π6 ππΏ )οΈ . In either case, the failure probability is at most some constant. 7.1.5 Main Result: The Code That Does It All The following is our main theorem, Theorem 7.1.16, a more precise restatement of Theorem 1.2.1, which is the culmination of the lifting technology results of Chapters 5 and 6. It states that there is a high-rate code that is simultaneously locally correctable and decodable, locally list-decodable, and robustly testable. Theorem 7.1.16. For every πΌ, π½ > 0 there exists πΏ > 0 such that the following holds: For infinitely many π ∈ N, there is π = π(π) = π(ππ½ ) and a linear code π ⊆ Fππ with distance πΏ(π) ≥ πΏ, rate at least 1 − πΌ, that is locally correctable and decodable up to πΏ 2 fraction errors with π(π) queries, locally list-decodable up to the Johnson bound with π(π 3 ) queries, and Ω(πΏ 72 )-robustly testable with π 2 queries. Proof. Let π = ⌈1/π½⌉ and let π = ⌈log2 π⌉ + 1. Let π = π2ππ · ⌈ln(1/πΌ)⌉ + π, and set πΏ , 2−π . For π ≥ π, set π = 2π and π , π π . Set π , (1 − πΏ)π − 1. Let π , RS(π, π)1βπ . By Proposition 7.1.1, πΏ(π) ≥ πΏ. By Theorem 7.1.11, the rate is at least 1 − πΌ. By Proposition 7.1.3, π is locally correctable and decodable up to πΏ 2 fraction errors with π queries. By Theorem 7.1.14, π is locally list-decodable up to the Johnson bound with π 3 queries. By Proposition 7.1.4, π is Ω(πΏ 72 )-robustly testable with π 2 queries. 7.2 Robust Low-Degree Testing In this section, we prove Theorem 7.2.2, which is simply a more precise restatement of Theorem 1.2.3. We do so by proving Theorem 7.2.1, a generalization from which Theorem 7.2.2 follows immediately. Theorem 7.2.1 replaces the Reed-Solomon code with an arbitrary univariate linear affine-invariant code π0 and replaces the bivariate Reed-Muller code with an arbitrary π‘-variate linear affine-invariant code π1 which is a strict subcode of (π0 )1βπ‘ . 102 Theorem 7.2.1. Let π‘ > 1 and let π ≥ 3. Let π0 ⊆ {Fπ → Fπ } and π1 ⊆ {Fπ‘π → Fπ } be linear affine-invariant codes such that π1 ( (π0 )1βπ‘ . Let πΏ , πΏ(π0 ). Fix π : Fπ π → Fπ . Let π , . Let πΌ2 be Eπ’ [πΏ (π|π’ , π1 )], where the expectation is taken over random π‘-dimensional π’ ⊆ Fπ {οΈ 2 (οΈ }οΈ π )οΈ the π‘-dimensional robustness of (π0 )1βπ . Then π ≥ min πΌ24πΏ , 2πΏ − 2π −1 · 2πΏ ·πΏ(π, (π1 )π‘βπ ). Theorem 7.2.2 (Robust plane testing for Reed-Muller). Let π ≥ 3. Fix a positive constant πΏ > 0 and a degree π = (1 − πΏ) · π. Let RM(π) be the π-variate Reed-Muller codes (οΈ 74 )οΈ πΏ of degree π over Fπ . Then RM(π) is 8·10 , 2 -robust. 52 Proof. Let RS be the Reed-Solomon code over Fπ of degree π. Let π be the characteristic of π. Let πΌ1 be the 2-dimensional robustness of RS1βπ . Then πΌ2 ≥ πΏ 72 2·1052 by Theorem 6.1.4 if π = 3, and by Theorem 6.1.18 if π ≥ 4. If π < π − π/π, then RM(π) = RS1βπ by Theorem 7.1.6, and so in this case the theorem follows immediately from Theorem 6.1.18. If π ≥ π − π/π and π ≥ 8πΏ , then RM(2) ( RS1β2 (by Theorem 7.1.6) but RM(π) = RM(2)2βπ (by Theorem 7.1.8), and so in this case the theorem follows immediately from Theorem 7.2.1, with π0 = RS, π‘ = 2, and π1 = RM(2). If π< 8 πΏ then the theorem follows from Corollary 6.1.2. Overview of Proof of Theorem 7.2.1. We illustrate the idea for the case where π‘ = 2, π0 is the Reed-Solomon code, and π1 is the bivariate Reed-Muller code of the same degree. The generalization to arbitrary π‘ and codes π0 , π1 is straightforward. If π is far from the lifted code, then on random planes π will be far from the bivariate lifted code and hence also from the bivariate Reed-Muller code. So the remaining case is when π is close to the lifted code. If the nearest function is a Reed-Muller codeword, then the theorem follows from the robustness of the lifted code. Otherwise, if the nearest function π is not Reed-Muller, then we show (through Corollary 6.2.8) that on many planes π is not a bivariate Reed-Muller codeword, and so π (being close to π) is not close to a bivariate Reed-Muller codeword (by the distance of the code). Proof of Theorem 7.2.1. Observe that, since (π1 )π‘βπ ⊂ (π0 )1βπ , we have πΏ(π, (π0 )1βπ ) ≤ 103 πΏ(π, (π1 )π‘βπ ). If πΏ(π, (π0 )1βπ ) ≥ min {οΈ }οΈ , then we are done since πΏ2 , πΏ(π, (π1 )π‘βπ ) 4 π = Eπ’ [πΏ (π|π’ , π1 )] [οΈ (οΈ )οΈ]οΈ ≥ Eπ’ πΏ π|π’ , (π0 )1βπ‘ ≥ πΌ2 · πΏ(π, (π0 )1βπ ) }οΈ {οΈ 2 πΏ π‘βπ , πΏ(π, (π1 ) ) ≥ πΌ2 · min 4 πΌ2 πΏ 2 · πΏ(π, (π1 )π‘βπ ). ≥ 4 1βπ Therefore, suppose πΏ(π, (π0 ) ) < min {οΈ }οΈ πΏ2 , πΏ(π, (π1 )π‘βπ ) 4 πΏ2 . 4 1βπ‘ codeword to π, so that π ∈ / (π1 )π‘βπ and πΏ(π, π ) < which π |π’ ∈ / π1 , then, since π1 is a subcode of (π0 ) . Let π ∈ (π0 )1βπ be the nearest If π’ is a π‘-dimensional subspace for , πΏ(π|π’ , π1 ) ≥ πΏ − πΏ(π|π’ , π |π’ ). Since πΏ2 Eπ’ [πΏ (π|π’ , π |π’ )] = πΏ(π, π ) < , 4 by Markov, [οΈ ]οΈ πΏ πΏ Pr πΏ(π|π’ , π |π’ ) ≥ ≤ π’ 2 2 By Corollary 6.2.8, Pr [π |π’ ∈ π1 ] ≤ 1 − πΏ + 2π −1 . π’ By the union bound, it follows that for at least 2πΏ −2π −1 fraction of the π‘-dimensional π’ ⊆ Fπ π , it holds that πΏ πΏ(π|π’ , π1 ) ≥ πΏ − πΏ(π|π’ , π |π’ ) ≥ . 2 Therefore, (οΈ π = Eπ’ [πΏ (π|π’ , π1 )] ≥ 104 )οΈ πΏ πΏ −1 − 2π · . 2 2 7.3 Nikodym Sets In this section, we use the lifted Reed-Solomon code to improve upon the polynomial method to show that Nikodym sets are large. π Definition 7.3.1. A subset π ⊆ Fπ π is a Nikodym set if for every x ∈ Fπ there exists a nonzero a ∈ Fπ π such that x + π‘a ∈ π for every nonzero π‘ ∈ Fπ . Before we prove our lower bound, let us recall how the polynomial method yields a good lower bound in the first place. Let π be a Nikodym set. Let π be the rate of the code RM(π, π − 2, π), which is approximately π ≈ 1 . π! Now, assume for the sake of contradiction that |π| < π π π . Then there exists a nonzero π with deg(π ) ≤ π − 2 that vanishes on every point of π. This is because the coefficients of π provide π π π degrees of freedom, but we only have |π| linear constraints. Now, we claim that π actually vanishes everywhere, contradicting the fact that it is nonzero. To see this, let x ∈ Fπ π be an arbitrary point. By the Nikodym property of π, there is a line πΏ through π such that (πΏ β {x}) ⊆ π. Consider the univariate polynomial π |πΏ . It vanishes on π − 1 points πΏ β {x} but has degree deg(π |πΏ ) < π − 1, so π |πΏ is identically zero. But then π (x) = π |πΏ (x) = 0. Therefore |π| ≥ π π π ≈ ππ . π! Observe that, in the above argument, the only property we actually used was the fact that, on every line πΏ, the restriction π |πΏ has degree deg(π |πΏ ) < π − 1. So pulling π from the Reed-Muller code was needlessly restrictive. We could use a larger code if possible, and in fact the largest code satisfying this property is none other than the lifted Reed-Solomon code! Theorem 7.3.2. Let π be a prime and let π = ππ . If π ⊆ Fπ π is a Nikodym set, then (οΈ )οΈ ⌈οΈ ⌉οΈ ππ |π| ≥ 1 − π−(π +π)/(ππ ) · π π where π = logπ π + 1. In particular, if π is fixed and π → ∞, then |π| ≥ (1 − π(1)) · π π . Proof. Let π = 1−π−(π +π)/(ππ ππ ) . Suppose, for the sake of contradiction, that |π| < π π π . Let π = RS(π, π − 2)1βπ . This is the construction of Section 7.1 with π = π . By Theorem 7.1.11, dimFπ (π) ≥ π π π > |π|, so there exists a nonzero π ∈ π such that π(x) = 0 for every x ∈ π. This follows from the fact that each equation of the form π(x) = 0 is a linear constraint on 105 π, and there are |π| linear constraints on π which has dimFπ (π) > |π| degrees of freedom. We proceed to show that π = 0, a contradiction. Let x ∈ Fπ π . Since π is a Nikodym set, there is a nonzero a ∈ Fπ π such that x + π‘a ∈ π for every nonzero π‘ ∈ Fπ . Define πa (π‘) , π(x + π‘a). Since π ∈ π, we have deg(πa ) ≤ π − 2. However, πa (π‘) = 0 for every π‘ ΜΈ= 0, so πa vanishes on π − 1 > deg(πa ) points, so πa = 0. In particular, π(x) = πa (0) = 0. 106 Appendix A Algebra Background This appendix contains well-known facts about arithmetic over finite fields and tensor codes that are used in the thesis. A.1 Arithmetic over finite fields Much of our work involves manipulating polynomials over finite fields. Expanding multinomials over a finite field, which has positive characteristic, is different from expanding over a field of zero characteristic, because binomial coefficients may vanish due to the characteristic of the field. For example, over a field of characteristic 2, the binomial (π + π )2 expands to π 2 + 2ππ + π 2 = π 2 + π 2 . The key fact we will use is (a generalization of) Lucas’ Theorem, which tells us what a multinomial coefficient looks like modulo a prime π. Theorem A.1.1 (Generalized Lucas’ Theorem). Let π0 , π1 , . . . , ππ ∈ N and let π be a prime. ∑οΈ (π) (π) Write ππ = ππ=1 π(π) · ππ where ππ , ππ ∈ JπK for every 0 ≤ π ≤ π and 0 ≤ π ≤ π. Then (οΈ π0 π1 , . . . , π π )οΈ ≡ (π) π (οΈ ∏οΈ π0 (π) π=0 )οΈ (π) π1 , . . . , π π Proof. Let π = {1, . . . , π0 }. Partition π into ππ,π 107 (mod π). [οΈ ]οΈ (π) ⊂ π, for each π ∈ Jπ + 1K and π ∈ π0 , [οΈ ]οΈ (π) π (π) of size |ππ,π | = ππ . For each π ∈ Jπ + 1K and π ∈ π0 , let πΊπ,π = Zππ0 act on ππ,π be cyclic β¨οΈπ β¨οΈπ0(π) permutation. Let πΊ = π=1 πΊπ,π , which acts on π. Let πΆ be the collection of all π=0 (οΈ π0 )οΈ π-colorings π : π → [π] such that |π−1 (π)| = ππ . Observe that |πΆ| = π1 ,...,π , and the π action of πΊ on π naturally induces an action of πΊ on πΆ, as follows: if π ∈ πΆ, and π ∈ πΊ, then for π₯ ∈ π, (ππ)(π₯) = π(π −1 π₯). Since |πΊ| is a power of π, so is the size of any orbit of πΆ (οΈ π0 )οΈ under πΊ. So, to compute π1 ,...,π (mod π), it suffices to count the number of fixed points π of πΆ under πΊ. Since πΊ fixes each ππ,π , it is easy to see that a coloring is fixed under πΊ if and only if each ππ,π is monochromatic. It therefore suffices to show that for each π ∈ Jπ + 1K (π) and each π ∈ [π], there are exactly ππ values for π such that ππ,π has color π. Fix π ∈ [π]. ∑οΈ (π) Observe that there are ππ = ππ=1 ππ elements of color π in total, and each set ππ,π of color π contributes ππ elements. The claim then follows easily by induction on π − π. In particular, it characterizes which multinomial coefficients are nonzero modulo π. π π1 ,...,ππ (οΈ Corollary A.1.2. If π, π1 , . . . , ππ ∈ N, then )οΈ ΜΈ≡ 0 (mod π) only if (π1 , . . . , ππ ) ≤π π. This allows us to expand multinomials over a finite field and know which terms of the usual expansion disappear. Corollary A.1.3. Let π be a prime and let F be a field of characteristic π. Let π ∈ N and let π₯1 , . . . , π₯π ∈ F. Then (οΈ π ∑οΈ π=1 )οΈπ π₯π = ∑οΈ (π1 ,...,ππ )≤π π (οΈ )οΈ ∏οΈ π π π₯π π π1 , . . . , ππ π=1 π Expanding multinomials over finite fields is particularly important for us since we frequently look at the restricting polynomials to affine subspaces, which entails composing with an affine function. Proposition A.1.4. Let π be a prime and let F be a field of characteristic π. Let X = (π1 , . . . , ππ ) and let Y ∈ (π1 , . . . , ππ‘ ). Let π (X) ∈ F[X], and let A ∈ Fπ×π‘ and b ∈ Fπ π . If 108 π (X) = ∑οΈ d∈π· πd · Xd , then π (AY + b) = ∑οΈ d∈supp(π ) (οΈ )οΈ π‘ π π‘ ∑οΈ (οΈ d )οΈ ∏οΈ ∏οΈ ∏οΈ βE β π πd · πππ π0 πππππ · ππ *π E π=1 π=1 π=1 E≤ d π where E in the summation is a π × (π‘ + 1) matrix with rows indexed by [π] and columns indexed by Jπ‘ + 1K. Proof. π (AY + b) = ∑οΈ πd · (AY + b)d d∈π· = ∑οΈ d∈π· πd · (οΈ π‘ π ∏οΈ ∑οΈ π=1 πππ ππ + ππ π=1 )οΈ π‘ ∏οΈ ππ π π ππ0 (Corollary A.1.3) = πd · ππ πππππ ππ ππ ππ0 , ππ1 , . . . , πππ‘ π=1 (ππ0 ,ππ1 ,...,πππ‘ )≤π ππ π=1 d∈π· )οΈ (οΈ (οΈ )οΈ π‘ π π‘ ∏οΈ ∑οΈ d ∏οΈ ∏οΈ ∑οΈ βE β π ππ *π πd · πππ π0 πππππ · = E π=1 π=1 π=1 E≤ d d∈π· ∑οΈ π ∏οΈ )οΈππ ∑οΈ (οΈ π A.2 Tensor codes The tensor product is a natural operation in linear algebra that, when applied to two linear codes, produces a new linear code in a natural way. There are many equivalent ways to define the tensor product of two codes. Since in this thesis we think of codes as linear subspaces of functions in {Fπ π → Fπ }, we define the tensor product in this context. Definition A.2.1. Let π ≥ 2, let π‘1 , . . . , π‘π ≥ 1 and π = ∑οΈπ π=1 π‘π , and for each π ∈ [π], let the code ππ ⊆ {Fπ‘ππ → Fπ } be linear and let ππ,a ⊆ Fπ π be the π‘π dimensional subspace consisting of all points where the π-th block (of π‘π coordinates) is free and all the [π] β {π} ∏οΈ π‘ blocks are fixed to a ∈ πΜΈ=π Fππ . The tensor product code π1 ⊗ · · · ⊗ ππ ⊆ {Fπ π → Fπ } is the 109 code β }οΈ β ∏οΈ β π : Fπ Fπ‘ππ π → Fπ β π |ππ,a ∈ ππ for every π ∈ [π] and a ∈ β {οΈ π1 ⊗ · · · ⊗ ππ , πΜΈ=π π Define π ⊗π β β , π ⊗ · · · ⊗ π. The following characterization of tensor product codes will be helpful. Proposition A.2.2. Let π ≥ 2, let π‘1 , . . . , π‘π ≥ 1 and π = ∑οΈπ π=1 π‘π , and for each π ∈ [π], let the code ππ ⊆ {Fπ‘ππ → Fπ } be linear, and let Xπ = (ππ1 , . . . , πππ‘π ) be variables. Then π1 ⊗ · · · ⊗ ππ = spanFπ {οΈ π ∏οΈ π=1 β }οΈ β β ππ (Xπ ) β ππ ∈ ππ β Corollary A.2.3. If π ⊆ {Fπ‘π → Fπ } has a degree set Deg(π), and π ≥ 1, then π ⊗π has degree set Deg(π ⊗π ) = Deg(π)π . In particular, if π is linear affine-invariant, and Fπ has characteristic π, then π ⊗π has a π-shadow-closed degree set. Proposition A.2.4. Let π1 and π2 be codes with distance πΏ1 and πΏ2 repectively. Then πΏ(π1 ⊗ π2 ) is at least πΏ1 πΏ2 . In particular, πΏ (π ⊗π ) ≥ πΏ(π)π . The following is a statement about the erasure decoding properties of tensor product codes. {οΈ }οΈ Proposition A.2.5. Let π = π1 ⊗ . . . ππ ∈ Fπ and π ⊆ Fπ π → Fπ π be a subset such that ∏οΈ π‘ for every π ∈ [π] and a ∈ πΜΈ=π Fππ satisfy |π ∩ ππ,a | ≥ (1 − πΏ(ππ ))π π‘π . Let π : π → Fπ be such ∏οΈ π‘ that for every π ∈ [π] and a ∈ πΜΈ=π Fππ satisfy that π|π∩ππ,a can be extended into a codeword of ππ on ππ,a . Then there exists a unique π′ ∈ π such that π′ |π = π. 110 Appendix B Finite field geometry This appendix describes some basic structural facts about affine maps, as well as basic geometry of affine subspaces over finite fields. B.1 Affine maps The following proposition allows us to decompose an arbitrary affine map (not necessarily injective) into a composition of a linear map and an injective affine map. Proposition B.1.1. Let π‘ ≤ π. For every affine map π΄ : Fπ‘π → Fπ π , there exists a linear ′′ ′ map π΄′ : Fπ‘π → Fπ‘π and injective affine map π΄′′ : Fπ‘π → Fπ π such that π΄ = π΄ β π΄ . π ′′ ′ Proof. If π΄ : x β¦→ π΄πΏ (x) + b for some linear π΄πΏ : Fπ‘π → Fπ π and b ∈ Fπ , and if π΄πΏ = π΄πΏ β π΄ ′ π‘ π‘ ′′ ′ for some injective affine π΄′′πΏ : Fπ‘π → Fπ π and linear π΄ : Fπ → Fπ , then π΄ = π΄ β π΄ where π΄′′ : x β¦→ π΄′′πΏ (x) + b. Therefore, we reduce to the case where π΄ is linear, i.e. π΄(0) = 0. Fix a basis e1 , . . . , eπ‘ ∈ Fπ‘π . For π ∈ [π‘], let vπ , π΄(eπ ). After re-labeling, we may assume without loss of generality that, for some 0 ≤ π ≤ π‘, the vectors v1 , . . . , vπ are linearly independent and vπ+1 , . . . , vπ‘ ∈ span{v1 , . . . , vπ }. There exist (unique) πππ ∈ Fπ , for π ∈ [π] ∑οΈ and π + 1 ≤ π ≤ π‘, such that vπ = ππ=1 πππ vπ . Define π΄′ : Fπ‘π → Fπ‘π as the unique linear map ∑οΈ such that π΄′ (eπ ) = eπ for π ∈ [π] and π΄′ (eπ ) = ππ=1 πππ eπ for π + 1 ≤ π ≤ π‘. For π ∈ [π], set 111 wπ , vπ , and extend w1 , . . . , wπ to a set of linearly independent vectors w1 , . . . , wπ‘ ∈ Fπ π . ′′ Define π΄′′ : Fπ‘π → Fπ π to be the unique linear map such that π΄ (eπ ) = wπ . By construction, π΄′′ is injective since its image is span{w1 , . . . , wπ‘ }. Moreover, for every π ∈ [π], (π΄′′ β π΄′ )(eπ ) = π΄′′ (π΄′ (eπ )) = π΄′′ (eπ ) = wπ = vπ = π΄(eπ ) and for every π + 1 ≤ π ≤ π‘, (οΈ ′′ ′ ′′ ′ (π΄ β π΄ )(eπ ) = π΄ (π΄ (eπ )) = π΄ ′′ π ∑οΈ )οΈ πππ eπ = π=1 π ∑οΈ ′′ πππ π΄ (eπ ) = π=1 π ∑οΈ πππ vπ = vπ = π΄(eπ ), π=1 so, by linearity, π΄′′ β π΄′ = π΄. The next three propositions describe the behavior of random affine maps. We are particularly interested in how πΏ(π β π΄, π β π΄) behaves, where π, π are distinct functions and π΄ is a random subspace. Proposition B.1.3 covers the case where π΄(0) is fixed, while Proposition B.1.4 allows π΄ to be truly free. π π‘ Proposition B.1.2. Let π‘ ≤ π, let x ∈ Fπ π , and let π΄ : Fπ → Fπ be a random affine map such that π΄(0) = x. For any fixed nonzero y ∈ Fπ‘π , π΄(y) is a uniformly random point in Fπ π . Proof. We can choose π΄ by choosing πππ ∈ Fπ for π ∈ [π], π ∈ [π‘] independently and uniformly at random, and setting β‘ β€ π · · · π1π‘ β’ 11 β₯ .. β₯ β’ .. . . π΄(y) = x + β’ . . . β₯y β£ β¦ ππ1 · · · πππ‘ The π-th coordinate of π΄(y) is therefore π₯π + ∑οΈπ‘ π=1 πππ π¦π which is uniformly random in Fπ , and the coordinates of π΄(y) are independent since the πππ are independent. π‘ π Proposition B.1.3. Let π‘ ≤ π, let x ∈ Fπ π , and let π΄ : Fπ → Fπ be a random affine map −π‘ such that π΄(0) = x. If π, π : Fπ π → Fπ , then Eπ΄ [πΏ(π β π΄, π β π΄)] ≤ πΏ(π, π) + π . 112 Proof. We have [οΈ [οΈ ]οΈ]οΈ Eπ΄ [πΏ(π β π΄, π β π΄)] = Eπ΄ Ey∈Fπ‘π 1π (π΄(y))ΜΈ=π(π΄(y)) [οΈ [οΈ ]οΈ]οΈ = Ey∈Fπ‘π Eπ΄ 1π (π΄(y))ΜΈ=π(π΄(y)) [οΈ [οΈ ]οΈ]οΈ ≤ π −π‘ + Ey∈Fπ‘π β{0} Eπ΄ 1π (π΄(y))ΜΈ=π(π΄(y)) (B.1) (Proposition B.1.2) = π −π‘ + πΏ(π, π). (B.4) (B.2) (B.3) Proposition B.1.4. Let π‘ ≤ π, and let π΄ : Fπ‘π → Fπ π be a uniformly random affine map. If −π‘ π, π : Fπ ≤ π → Fπ and πΏ , πΏ(π, π), then πΏ(π β π΄, π β π΄) has mean πΏ and variance πΏ(1 − πΏ)π π −π‘ /4. π‘ π Proof. Observe that for any fixed y ∈ Fπ π , if π΄ : Fπ → Fπ is uniformly random, then π΄(y) is a uniformly random point in Fπ π , and for distinct y, the π΄(y) are pairwise independent. Therefore, 1π (π΄(y))ΜΈ=π(π΄(y)) has mean πΏ and variance πΏ(1 − πΏ). Since πΏ(π β π΄, π β π΄) = ∑οΈ π −π‘ y∈Fπ‘π 1π (π΄(y))ΜΈ=π(π΄(y)) , it follows that the mean is πΏ and variance is πΏ(1−πΏ)π −π‘ ≤ π −π‘ /4. Finally, we prove that if π, π are distinct polynomials on a plane, then they cannot agree on too many lines. Proposition B.1.5. Let π, π : F2π → Fπ be distinct. For any a, b ∈ F2π , define πa,b , πa,b : Fπ → Fπ by πa,b (π ) , π (aπ + b) and similarly πa,b (π ) , π(aπ + b). Then πa,b = πa,b for at most 2π 3 pairs (a, b) ∈ F2π × F2π . Proof. Let β = π − π and βa,b = πa,b − πa,b . We have four cases to consider. 1. π1 , π2 ΜΈ= 0: if βa,b = 0, then the polynomial β(π, π ) is divisible by π − ππ21 π − ππ21π1 − π2 . There are π(π − 1) pairs (a, b) which correspond to this factor (π − 1 choices for a and then π choices for b given a). 2. π1 ΜΈ= 0, π2 = 0: if βa,b = 0, then the polynomial β(π, π ) is divisible by π − π2 . There are π 2 pairs corresponding to this factor (π choices for π1 and π choices for π1 ). 113 3. π1 = 0, π2 ΜΈ= 0: same as previous case, by symmetry. 4. π1 = π2 = 0: there are at most π 2 such pairs (π 2 choices for b). Say a pair (a, b) is bad if βa,b = 0. Trivially, deg(β) ≤ 2(π − 1), so β has at most 2(π − 1) linear factors. Each factor from cases 1, 2, and 3 corresponds to at most π 2 pairs (a, b), resulting in at most 2π 2 (π − 1) bad pairs (a, b) from those cases, and there are at most π 2 bad pairs from case 4. So, the total number of bad pairs is at most 2π 2 (π − 1) + π 2 ≤ 2π 3 . B.2 Affine subspaces The next two lemmas analyze the behavior of random affine subspaces in Fπ π . Lemma B.2.1 shows that a low-dimensional subspace is disjoint from almost all low-dimensional affine subspaces. Lemma B.2.1. Let π‘ ≤ π < π. Let π’ ⊆ Fπ π be a fixed affine subspace of dimension π‘, and let π£ ⊆ Fπ π be a uniformly random affine subspace of dimension π. Then Prπ£ [π’ ∩ π£ ΜΈ= ∅] < π −(π−π−π‘) . Proof. By affine symmetry, we may assume that π£ is fixed and π’ is random. Furthermore, we can assume that π£ = (0, π΅), where π΅ is a basis and hence |π΅| = π. We choose π’ by choosing random x ∈ Fπ π , random basis π΄ = {a1 , . . . , aπ‘ }, and setting π’ , (x, π΄). Let πΈ be the event that π’ ∩ π£ ΜΈ= ∅. Re-arrange a1 , . . . , aπ‘ so that for some 0 ≤ π ≤ π‘ − 1, aπ ∈ span(π΅) if and only if π ≤ π . Note that πΈ holds if and only if there exist π1 , . . . , ππ‘ ∈ Fπ π such that x + π1 a1 + · · · + ππ‘ aπ‘ ∈ π−π span(π΅). Let π : Fπ be the linear map that projects onto the last π−π coordinates. π → Fπ Note that ker(π ) = π΅. For each π ∈ [π‘], let a′π , π aπ ∈ Fπ−π . Then πΈ holds if and only if π π x ∈ span(a′π +1 , . . . , a′π‘ ). Therefore, there are at most π π‘ choices for π x, hence at most π π+π‘ choices for x, out of π π total choices for x, so Pr[πΈ] ≤ π π+π‘ ππ = π −(π−π−π‘) . Lemma B.2.2 shows that if π βͺ π, then π random vectors are likely to be linearly independent, and in particular two random low-dimensional subspaces through a fixed point 114 x are likely to intersect only at x. Lemma B.2.2. Let π < π and a1 , . . . , aπ ∈ Fπ π be uniformly chosen vectors. Then the probability that {aπ }ππ=1 are linearly independent is at least 1 − π −(π−π) . In particular, the probability that two π‘-dimensional subspaces through a point x ∈ Fπ π will intersect only on x is at least 1 − π −(π−2π‘) . Proof. The probability that aπ+1 ∈ / span {a1 , . . . , aπ } given that the latter are linearly independent is 1 − π −(π−π) . Therefore the probability that all of them are independent is π−1 ∏οΈ (1 − π π=0 −(π−π) )≥1− π−1 ∑οΈ π −(π−π) =1−π −(π−π) π ∑οΈ π −π ≥ 1 − π −(π−π) . π=1 π=0 For the last part, observe that choosing two π‘-dimensional subspaces through x is equivalent to choose 2π‘ basis vectors, given that each π‘ are linearly independent. So the probability that they intersect only on x, is the same as that those vectors are linearly independent. Hence, by the first part, this probability is at least 1 − π −(π−2π‘) . 115 116 Bibliography [AKK+ 05] Noga Alon, Tali Kaufman, Michael Krivelevich, Simon Litsyn, and Dana Ron. Testing Reed-Muller codes. IEEE Transactions on Information Theory, 51(11):4032–4039, 2005. [ALM+ 98] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and the hardness of approximation problems. Journal of the ACM, 45(3):501–555, May 1998. [Aro94] Sanjeev Arora. Probabilistic checking of proofs and the hardness of approximation problems. PhD thesis, University of California at Berkeley, 1994. [AS98] Sanjeev Arora and Shmuel Safra. Probabilistic checking of proofs: A new characterization of NP. Journal of the ACM, 45(1):70–122, January 1998. [AS03] Sanjeev Arora and Madhu Sudan. Improved low degree testing and its applications. Combinatorica, 23(3):365–426, 2003. Preliminary version in Proceedings of ACM STOC 1997. [BET10] Avraham Ben-Aroya, Klim Efremenko, and Amnon Ta-Shma. Local list decoding with a constant number of queries. In 51st IEEE Symposium on Foundations of Computer Science (FOCS), 2010. [BGK+ 13] Eli Ben-Sasson, Ariel Gabizon, Yohay Kaplan, Swastik Kopparty, and Shubhangi Saraf. A new family of locally correctable codes based on degree-lifted algebraic geometry codes. In Dan Boneh, Tim Roughgarden, and Joan Feigenbaum, editors, Symposium on Theory of Computing Conference, STOC’13, Palo Alto, CA, USA, June 1-4, 2013, pages 833–842. ACM, 2013. [BGM+ 11a] Eli Ben-Sasson, Elena Grigorescu, Ghid Maatouk, Amir Shpilka, and Madhu Sudan. On sums of locally testable affine invariant properties. In APPROXRANDOM, pages 400–411, 2011. [BGM+ 11b] Eli Ben-Sasson, Elena Grigorescu, Ghid Maatouk, Amir Shpilka, and Madhu Sudan. On sums of locally testable affine invariant properties. Electronic Colloquium on Computational Complexity (ECCC), 18:79, 2011. 117 [BK09] K. Brander and S. Kopparty. List-decoding Reed-Muller over large fields upto the Johnson radius. Manuscript, 2009. [BKS+ 10] Arnab Bhattacharyya, Swastik Kopparty, Grant Schoenebeck, Madhu Sudan, and David Zuckerman. Optimal testing of Reed-Muller codes. In FOCS, pages 488–497. IEEE Computer Society, 2010. [BLR93] Manuel Blum, Michael Luby, and Ronitt Rubinfeld. Self-testing/correcting with applications to numerical problems. Journal of Computer and System Sciences, 47(3):549–595, 1993. [BSGH+ 04] Eli Ben-Sasson, Oded Goldreich, Prahladh Harsha, Madhu Sudan, and Salil Vadhan. Robust PCPs of proximity, shorter PCPs and applications to coding. In Proceedings of the 36th Annual ACM Symposium on Theory of Computing, pages 1–10, New York, 2004. ACM Press. [BSMSS11] Eli Ben-Sasson, Ghid Maatouk, Amir Shpilka, and Madhu Sudan. Symmetric LDPC codes are not necessarily locally testable. In IEEE Conference on Computational Complexity, pages 55–65. IEEE Computer Society, 2011. [BSS06] Eli Ben-Sasson and Madhu Sudan. Robust locally testable codes and products of codes. Random Structures and Algorithms, 28(4):387–402, 2006. [BSV09a] Eli Ben-Sasson and Michael Viderman. Composition of semi-LTCs by twowise tensor products. In Irit Dinur, Klaus Jansen, Joseph Naor, and JoseΜ D. P. Rolim, editors, APPROX-RANDOM, volume 5687 of Lecture Notes in Computer Science, pages 378–391. Springer, 2009. [BSV09b] Eli Ben-Sasson and Michael Viderman. Tensor products of weakly smooth codes are robust. Theory of Computing, 5(1):239–255, 2009. Preliminary version in Proc. APPROX-RANDOM 2008. [DKSS09] Zeev Dvir, Swastik Kopparty, Shubhangi Saraf, and Madhu Sudan. Extensions to the method of multiplicities, with applications to kakeya sets and mergers. In FOCS, pages 181–190. IEEE Computer Society, 2009. [DR04] Irit Dinur and Omer Reingold. Assignment testers: Towards a combinatorial proof of the PCP-theorem. In Proceedings of the 45th Annual IEEE Symposium on Foundations of Computer Science, pages 155–164, Loc Alamitos, CA, USA, 2004. IEEE Press. [DS07] Zeev Dvir and Amir Shpilka. An improved analysis of linear mergers. Computational Complexity, 16(1):34–59, 2007. [DSW06] Irit Dinur, Madhu Sudan, and Avi Wigderson. Robust local testability of tensor products of LDPC codes. In Josep DΔ±Μaz, Klaus Jansen, JoseΜ D. P. Rolim, 118 and Uri Zwick, editors, APPROX-RANDOM, volume 4110 of Lecture Notes in Computer Science, pages 304–315. Springer, 2006. [Dvi08] Zeev Dvir. On the size of Kakeya sets in finite fields. Journal of the American Mathematical Society, (to appear), 2008. Article electronically published on June 23, 2008. [DW11] Zeev Dvir and Avi Wigderson. Kakeya sets, new mergers, and old extractors. SIAM J. Comput., 40(3):778–792, 2011. [FS95] Katalin Friedl and Madhu Sudan. Some improvements to total degree tests. In Proceedings of the 3rd Annual Israel Symposium on Theory of Computing and Systems, pages 190–198, Washington, DC, USA, 4-6 January 1995. IEEE Computer Society. Corrected version available online at http://people.csail.mit.edu/madhu/papers/friedl.ps. [GGR09] Parikshit Gopalan, Venkatesan Guruswami, and Prasad Raghavendra. List decoding tensor products and interleaved codes. In Proceedings of the 41st Annual ACM Symposium on Theory of Computing, STOC 2009, Bethesda, MD, USA, May 31 - June 2, 2009, pages 13–22, 2009. [GHS15] Alan Guo, Elad Haramaty, and Madhu Sudan. Robust testing of lifted codes with applications to low-degree testing. Electronic Colloquium on Computational Complexity (ECCC), 22:34, 2015. [GK14] Alan Guo and Swastik Kopparty. List-decoding algorithms for lifted codes. CoRR, abs/1412.0305, 2014. [GKS13] Alan Guo, Swastik Kopparty, and Madhu Sudan. New affine-invariant codes from lifting. In Robert D. Kleinberg, editor, Innovations in Theoretical Computer Science, ITCS ’13, Berkeley, CA, USA, January 9-12, 2013, pages 529– 540. ACM, 2013. [GS92] Peter Gemmell and Madhu Sudan. Highly resilient correctors for multivariate polynomials. Information Processing Letters, 43(4):169–174, September 1992. [GS99] Venkatesan Guruswami and Madhu Sudan. Improved decoding of ReedSolomon and algebraic-geometric codes. IEEE Transactions on Information Theory, 45:1757–1767, 1999. Preliminary version appeared in Proc. of FOCS 1998. [HRS13] Elad Haramaty, Noga Ron-Zewi, and Madhu Sudan. Absolutely sound testing of lifted codes. In Prasad Raghavendra, Sofya Raskhodnikova, Klaus Jansen, and JoseΜ D. P. Rolim, editors, Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques - 16th International Workshop, APPROX 2013, and 17th International Workshop, RANDOM 2013, Berkeley, 119 CA, USA, August 21-23, 2013. Proceedings, volume 8096 of Lecture Notes in Computer Science, pages 671–682. Springer, 2013. [HSS11] Elad Haramaty, Amir Shpilka, and Madhu Sudan. Optimal testing of multivariate polynomials over small prime fields. In Rafail Ostrovsky, editor, IEEE 52nd Annual Symposium on Foundations of Computer Science, FOCS 2011, Palm Springs, CA, USA, October 22-25, 2011, pages 629–637. IEEE, 2011. [JPRZ09] Charanjit S. Jutla, Anindya C. Patthak, Atri Rudra, and David Zuckerman. Testing low-degree polynomials over prime fields. Random Struct. Algorithms, 35(2):163–193, 2009. [Kop12] Swastik Kopparty. List-decoding multiplicity codes. Electronic Colloquium on Computational Complexity (ECCC), 19:44, 2012. [KR06] Tali Kaufman and Dana Ron. Testing polynomials over general fields. SIAM Journal of Computing, 36(3):779–802, 2006. [KS07] Tali Kaufman and Madhu Sudan. Algebraic property testing: The role of invariance. Electronic Colloquium on Computational Complexity (ECCC), 14(111), 2007. [KS08] Tali Kaufman and Madhu Sudan. Algebraic property testing: the role of invariance. In STOC, pages 403–412, 2008. [KSY14] Swastik Kopparty, Shubhangi Saraf, and Sergey Yekhanin. High-rate codes with sublinear-time decoding. J. ACM, 61(5):28, 2014. [MR06] Dana Moshkovitz and Ran Raz. Sub-constant error low degree test of almostlinear size. In Proceedings of the 38th Annual ACM Symposium on Theory of Computing, Seattle, WA, USA, May 21-23, 2006, pages 21–30, 2006. [PW04] Ruud Pellikaan and Xin-Wen Wu. List decoding of q-ary reed-muller codes. IEEE Transactions on Information Theory, 50(4):679–682, 2004. [RS96] Ronitt Rubinfeld and Madhu Sudan. Robust characterizations of polynomials with applications to program testing. SIAM Journal on Computing, 25(2):252– 271, April 1996. [RS97] Ran Raz and Shmuel Safra. A sub-constant error-probability low-degree test, and a sub-constant error-probability PCP characterization of NP. In Proceedings of the Twenty-Ninth Annual ACM Symposium on Theory of Computing, pages 475–484, New York, NY, 1997. ACM Press. [SS08] Shubhangi Saraf and Madhu Sudan. Improved lower bound on the size of Kakeya sets over finite fields. ArXiv e-prints, August 2008. 120 [STV99] Madhu Sudan, Luca Trevisan, and Salil Vadhan. Pseudorandom generators without the XOR lemma. Proceedings of the 31st Annual ACM Symposium on Theory of Computing, pages 537–546, 1999. [Val05] Paul Valiant. The tensor product of two codes is not necessarily robustly testable. In Chandra Chekuri, Klaus Jansen, JoseΜ D. P. Rolim, and Luca Trevisan, editors, APPROX-RANDOM, volume 3624 of Lecture Notes in Computer Science, pages 472–481. Springer, 2005. [Vid12] Michael Viderman. A combination of testability and decodability by tensor products. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques - 15th International Workshop, APPROX 2012, and 16th International Workshop, RANDOM 2012, Cambridge, MA, USA, August 15-17, 2012. Proceedings, pages 651–662, 2012. [WB86] Lloyd R. Welch and Elwyn R. Berlekamp. Error correction of algebraic block codes. US Patent Number 4,633,470, December 1986. 121