Switching Network Model Aaron H. Potechin

Analyzing Monotone Space Complexity Via the
ARCHIVES
Switching Network Model
MASSACHUrTS INSTITUTE
OF TECH' OLOLGY
by
JUN 3 0 2015
Aaron H. Potechin
Submitted to the Department of Mathematics
in partial fulfillment of the requirements for the degree of
LIBRARIES
Doctor of Philosophy in Mathematics
at the
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
June 2015
Massachusetts Institute of Technology 2015. All rights reserved.
A uthor ........
Signature redacted
.................
Department of Mathematics
May 8, 2015
Certified by..........
Signature redactedJonathan A. Kelner
Associate Professor
Thesis Supervisor
Accepted by ...................
Signature redacted
Wichel X. Goemans
Chairman, Department Committee on Graduate Theses
2
Analyzing Monotone Space Complexity Via the Switching
Network Model
by
Aaron H. Potechin
Submitted to the Department of Mathematics
on May 8, 2015, in partial fulfillment of the
requirements for the degree of
Doctor of Philosophy in Mathematics
Abstract
Space complexity is the study of how much space/memory it takes to solve problems. Unfortunately, proving general lower bounds on space complexity is notoriously hard. Thus, we instead consider the restricted case of monotone algorithms,
which only make deductions based on what is in the input and not what is missing. In this thesis, we develop techniques for analyzing monotone space complexity
via a model called the monotone switching network model. Using these techniques,
we prove tight bounds on the minimal size of monotone switching networks solving
the directed connectivity, generation, and k-clique problems. These results separate
monotone analgoues of L and NL and provide an alternative proof of the separation
of the monotone NC hierarchy first proved by Raz and McKenzie. We then further
develop these techniques for the directed connectivity problem in order to analyze
the monotone space complexity of solving directed connectivity on particular input
graphs.
Thesis Supervisor: Jonathan A. Kelner
Title: Associate Professor
3
4
Acknowledgments
This thesis is based on work supported by a National Science Foundation Graduate
Research Fellowship under grant No. 0645960.
The author would like to thank his advisors Jonathan Kelner and Boaz Barak and
his collaborator Siu Man Chan for their guidance on this project. The author would
also like to thank his family and friends for their support during his PhD studies.
5
6
Contents
1
2
Introduction
13
1.1
Known Results on Space Complexity . . . . . . . . . . . . . . . . . .
13
1.2
Approaches to Space Lower Bounds . . . . . . . . . . . . . . . . . . .
14
1.3
Known Results on Monotone Complexity . . . . . . . . . . . . . . . .
15
1.4
Overview of Our Results . . . . . . . . . . . . . . . . . . . . . . . . .
17
Background Material
19
2.1
Switching Networks: From Practice to Theory . . . . . . . . . . . . .
20
2.2
Turing machines and space complexity
. . . . . . . . . . . . . . . . .
21
2.3
Switching-and-rectifier networks and nondeterministic space complexity 23
2.4
Branching programs, switching networks, and deterministic space complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
25
3 Bounds on monotone switching networks for directed connectivity
29
3.1
D efinitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
30
3.2
Certain Knowledge Switching Networks . . . . . . . . . . . . . . . . .
34
3.2.1
The certain knowledge game for directed connectivity . . . . .
35
3.2.2
Adapting the certain knowledge game for monotone switching
netw orks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
36
3.2.3
A partial order on knowledge sets . . . . . . . . . . . . . . . .
37
3.2.4
Certain knowledge switching networks and reachability from s
39
3.2.5
An upper bound on certain-knowledge switching networks
41
3.2.6
A lower size bound on certain-knowledge switching networks
7
.
44
3.4
Elementary results on monotone switching networks . . . . . . . . .
48
3.3.1
The knowledge game for directed connectivity . . . . . . . .
49
3.3.2
A partial order on states of knowledge
50
3.3.3
Knowledge description of monotone switching networks
3.3.4
Reduction to reachability from s . . . . . . . . . . . . . . . .
55
3.3.5
Reduction to certain knowledge switching networks . . . . .
57
.
47
.
.
.
.
.
60
3.4.1
Function descriptions of sound monotone switching networks
.
61
3.4.2
Fourier analysis . . . . . . . . . . . . . . . . . . . . . . . . .
64
3.4.3
Invariants and a quadratic lower bound . . . . . . . . . . . .
65
3.4.4
General lower bounds . . . . . . . . . . . . . . . .
68
3.4.5
Conditions for a good set of functions . . . . . . .
72
A Superpolynomial Lower Bound . . . . . . . . . . . . .
75
.
.
.
.
.
. . . . . . . . . . . . . . . . . . . . . . . . .
3.5.1
From certain knowledge decriptions and knowledge d escriptions
. . . . . .
76
3.5.2
A criterion for E-invariance
. . . . . . . . . . . .
. . . . . .
78
3.5.3
Choosing Fourier coefficients via dot products . .
. . . . . .
81
3.5.4
Proof of Theorem 3.5.1 . . . . . . . . . . . . . . .
. . . . . .
82
.
.
.
.
to function descriptions . . . . . . . . . . . . . . .
An 001g) lower size bound
. . . . . . . . . . . . . . . .
. . . . . .
84
3.7
Monotone analogues of L and NL . . . . . . . . . . . . .
. . . . . .
89
3.8
Finding paths with non-monotone switching networks . .
. . . . . .
90
.
.
.
3.6
Lower bounds on monotone switching networks for gen eration and
95
4.1
Circuit Depth and Switching Network Size . . . . . . . .
95
4.2
Our Results and Comparison with Previous Work . . . .
97
4.3
Lower bounds on monotone switching networks for Generation
99
4.4
Lower bounds on monotone switching networks for k-clique . .
.
k-clique
.
.
.
4
54
Fourier Analysis and Invariants on Monotone Switiching Networks For
Directed Connectivity
3.5
. . . . . . . . . . . .
.
3.3
Simplified bounds on certain-knowledge switching networks .
.
3.2.7
8
104
5
Upper Bounds: Dealing with Lollipops via Parity
5.1
Motivation for considering particular input graphs for directed connectiv ity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
6
109
5.2
Lollipops and certain knowledge switching networks . . . . . . . . . . 110
5.3
A general upper bound . . . . . . . . . . . . . . . . . . . . . . . . . .111
Improved Lower Bound Techniques for Monotone Switching Net123
works for Directed Connectivity
7
109
6.1
Permutations on input graphs, cuts, functions, and Fourier coefficients 124
6.2
Calculating Permutation Averages . . . . . . . . . . . . . . . . . . . .
125
6.3
Lower bounds from permutation averages . . . . . . . . . . . . . . . .
132
6.4
Equations on sum vectors
6.5
Freedom in choosing a base function
6.6
Building up a base function
. . . . . . . . . . . . . . . . . . . . . . . . 134
. . . . . . . . . . . . . . . . . .
139
. . . . . . . . . . . . . . . . . . . . . . .
145
6.7
Constructing a base function . . . . . . . . . . . . . . . . . . . . . . .
155
6.8
The cost of extending a base function . . . . . . . . . . . . . . . . . . 159
6.9
Putting everything together: A lower bound . . . . . . . . . . . . . .
165
167
Conclusion and Future Work
169
A Proof of Lemma 3.2.32
9
10
List of Figures
3-1
A monotone swiitching network solving directed connectivity . . . . .
31
3-2
A sound but not complete switching network . . . . . . . . . . . . . .
32
3-3
A complete but not sound switching network . . . . . . . . . . . . . .
33
3-4
A non-monotone switching network solving directed connectivity . . .
34
3-5
A certain knowledge switching network G' solving directed connectivity
on V(G)
3-6
=
{s, t, a, b} together with a certain knowledge description of it 38
A certain knowledge switching network G' solving directed connectivity
on V(G) ={s, t, a, b, c} together with a certain knowledge description
ofit .........
3-7
=
{s, t, a, b, c} together with a certain knowledge description
ofit .........
....................................
55
An illustation of the partial reduction of monotone switching networks
to certain knowledge switching networks
3-9
40
A monotone switching network G' solving directed connectivity on
V(G)
3-8
....................................
. . . . . . . . . . . . . . . .
60
Representing cuts of V(G) . . . . . . . . . . . . . . . . . . . . . . . .
63
3-10 A certain knowledge switching network and the reachability function
description for it. . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
64
3-11 A monotone switching network together with the reachability function
description for it and the Fourier decomposition of each function . . .
65
3-12 An illustration of the linear independence argument . . . . . . . . . .
68
3-13 An illustartions of the ideas used to prove Lemma 3.5.12 . . . . . . .
80
11
12
Chapter 1
Introduction
The central topic of this thesis is space complexity, i.e. how much space/memory
it takes to solve various problems. We are particularly interested in the space complexity of the directed connectivity problem, which asks if there is a path from from
some vertex s to another vertex t in a directed input graph. Determining the space
complexity of the directed connectivity problem would resolve the fundamental open
question of whether the complexity class L (Logarithmic Space) is equal to the complexity class NL (nondeterministic logarithmic space). Unfortunately, proving general
lower bounds on space complexity is notoriously difficult and is currently beyond our
reach. Nevertheless, if we place restrictions on what our algorithms can do we can obtain bounds on the space complexity of these algorithms, which gives insight into the
general problem. In this thesis, we use a model called the switching network model to
analyze the space complexity of monotone algorithms, which make deductions based
on what is present in the input but not what is absent from the input.
1.1
Known Results on Space Complexity
Although no non-trivial lower space bounds are known for directed connectivity and
most other problems, we do have considerable knowledge about deterministic and
nondeterministic space complexity. It is folklore that the directed connectivity problem is NL-complete. This is the basis for the statement above that to resolve the
13
L versus NL problem, which asks whether non-determinism helps in space-bounded
computation, it is sufficient to understand the space complexity of the directed connectivity problem. In 1970, Savitch [23] gave an upper bound with an O(log 2 n)-space
deterministic algorithm for directed connectivity. This proved that any problem that
can be solved nondeterministically using g(n) space for some function g(n) can be
solved deterministically in O(g(n)2 ) space. In 1988, Immerman [11] and Szelepcsenyi
[26] independently gave an O(log n)-space nondeterministic algorithm for directed
non-connectivity. This proved the surprising result that for any problem which can
be solved nondeterministically in g(n) space, it's complement can also be solved nondeterministically in g(n) space. In particular, NL=co-NL.
For the problem of undirected connectivity (i.e. where the input graph G is undirected), a probabilistic logarithmic space algorithm was shown in 1979 using random
walks by Aleliunas, Karp, Lipton, Lovisz, and Rackoff [1]. In 2004, Reingold [22] gave
a deterministic O(log n)-space algorithm for undirected connectivity, showing that
undirected connectivity is in L. Trifonov [27] independently gave an 0 (lg n lg lg n) algorithm for undirected connectivity. As discussed in Chapter 2, Reingold's result was
an inspiration for our choice of the switching network model (which is an undirected
graph whose edges depend on the input) for analyzing space complexity.
1.2
Approaches to Space Lower Bounds
For the problem which we are considering, proving space lower bounds, there have
been two general approaches so far. One approach is to use the branching program
model, which is an intuitive way to represent the different branches an algorithm
can take depending on the input. Bounds are then proved on restricted branching
programs/algorithms.
For example, one common restriction is that the branching
program/algorithm must look at the bits of the input in order and may only pass
through the input k times for some constant k. For a survey of some of the results
on branching programs, see Razborov [20].
Another approach which is specialized to analyzing the directed connectivity prob14
lem is to use the JAG (Jumping Automata on graphs) model or variations thereof.
The JAG (Jumping Automata on Graphs) model was introduced by Cook and Rackoff [7] as a simple model for which we can prove good lower time and space bounds but
which is still powerful enough to simulate most known algorithms for the directed connectivity problem. This implies that if there is an algorithm for directed connectivity
breaking these bounds, it must use some new techniques which cannot be captured by
the JAG model. Later work in this area has focused on extending this framework to
additional algorithms and more powerful variants of the JAG model. Two relatively
recent results in this area are the result of Edmonds, Poon, and Achlioptas [8] which
shows tight lower bounds for the more powerful NNJAG (Node-Named Jumping Automata on Graphs) model and the result of Lu, Zhang, Poon, and Cai [16] showing
that Reingold's algorithm for undirected connectivity and a few other algorithms for
undirected connectivity can all be simulated by the RAM-NNJAG model, a uniform
variant of the NNJAG model.
In this thesis, we take a third approach, using the switching network model to
analyze space complexity. There are two main reasons for this choice. First, unlike
the branching program model and the JAG model we may restrict switching networks
to be monotone, only using what is present in the input and not what is absent from
the input. This means that the switching network model may be used to analyze
monotone space complexity. Second, switching networks have the important property
that every step which is made on them is reversible. This property plays a crucial
role in our lower bound techniques.
1.3
Known Results on Monotone Complexity
Since we are analyzing monotone algorithms and switching networks, our work is
part of monotone complexity theory. We now describe some of the known results in
monotone complexity theory and how our work fits in.
Looking at time complexity rather then space complexity, in 1985 Razborov [21]
used an approximation method to show that any monotone circuit solving the k-clique
15
problem on n vertices (determining whether or not there is a set of k pairwise adjacent
vertices in a graph on n vertices) when k
proving that mP
#
=
[19--] must have size at least N'), thus
mNP. This method was later improved by Alon and Boppana
[2] and by Haken [10].
Looking at monotone circuit depth instead of monotone circuit size, Karchmer and
Wigderson [12] used a communication complexity approach to show that any monotone circuit solving undirected connectivity has depth at least Q((lg n) 2 ). Recalling
that monotone-NC is the class of problems which can be solved by a monotone circuit
of depth O((lg n)'), this proved that undirected connectivity is not in monotone- NC
and separated monotone-NC and monotone-NC 2 . Raz and McKenzie [19] later sep-
#
arated the entire monotone NC hierarchy, proving that for any i, monotone-NC'
monotone-NCi+l.
Our results are closely related to the results of Karchmer, Wigderson, Raz, and
McKenzie, as there is a close connection between monotone circuit depth and the
size of monotone switching networks. As discussed in Chapter 4, any problem which
can be solved by a monotone circuit of depth O(g(n)) can be solved by a monotone
switching network of size 2 0(g(n)), so our lower size bounds on monotone switching
networks prove corresponding lower bounds on monotone circuit depth. Conversely,
any problem which can be solved by a monotone switching network of size
be solved by a monotone circuit of depth
Q(g(n)
2
2 0(g(n))
can
), so the previous lower bounds on
monotone circuit depth imply corresponding lower bounds on the size of monotone
switching networks.
That said, while our results are closely related to previous work on monotone
circuit depth, our approach is very different and thus gives further insight into the
problem. Also, as discussed in Chapter 4, comparing our bounds with previous work,
our size lower bounds on monotone switching networks are stronger and our corresponding depth lower bounds on monotone circuits are equally strong. Thus, our
work strengthens the previous results of Karchmer, Wigderson, Raz, and McKenzie.
16
1.4
Overview of Our Results
We now give an overview of the remainder of this thesis. In Chapter 2 we present
background material on switching networks and their relationship to space complexity.
In chapter 3 we prove an no(lgn) lower size bound on monotone switching networks
solving the directed connectivity problem, separating monotone analogues of L and
NL. In chapter 4 we apply the techniques developed in chapter 3 to the generation
problem and the k-clqiue problem. For the generation problem, we show that any
monotone switching network solving the generation problem on pyramid graphs of
height h < n' where e <
must have size
nQ(h).
This gives an alternative proof of the
separation of the monotone-NC hierarchy. For the k-clique problem, we show that any
monotone switching network solving the k-clique problem for k < nE where 6 <
must
have size n.(k). In chapters 5 and 6 we return to the directed connectivity problem,
focusing on the difficulty of using monotone switching networks to solve directed
connectivity on particular input graphs.
In chapter 5 we prove an upper bound,
showing that monotone switching networks can easily solve directed connectivity on
input graphs in which the vast majority of vertices are either easily reachable from
s or can easily reach t. Finally, in chapter 6 we further develop the lower bound
techniques in chapter 3 to show that monotone switcing networks are no more efficient
than Savitch's algorithm for input graphs where the vertices are not too connected
to each other.
17
18
Chapter 2
Background Material
In this chapter, we present background material on switching networks and their connection to space complexity. Along the way, we present background material on two
closely related models, switching-and-rectifier networks and branching programs, and
their connection to nondeterministic space complexity and space complexity respectively. The definitions in this chapter for switching networks, switching-and-rectifier
networks, and branching programs are slightly modified from Razborov [201.
Intuitively, a switching network is an undirected graph with distinguished vertices
s' and t' where each edge is switched on or off depending on one bit of the input.
Given an input, a switching network outputs 1 if there is a path from s' to t' for that
input and 0 otherwise. The precise definition is as follows.
Definition 2.0.1. A switching network is a tuple ZG', s', t', P') where G' is an undirected multi-graph with distinguished vertices s', t' and /t' is a labelingfunction that associates with each edge e' C E(G') a label p(e') of the form xi = 1 or xi = 0 for some i
between 1 and n. We say that this network computes the function f : {0, 1} -+ {0, 1},
where f(x)
=
1 if and only if there is a path from s' to t' such that each edge of this
path has a label that is consistent with x.
We take the size of a switching network to be
{0, 1}
-+
|V(G')|,
and for a function f
{0, 1}, we define s(f )(n) to be size of the smallest switching network com-
puting f.
19
Definition 2.0.2. We say that a switching network is monotone if it has no edge
labels of form xi = 0.
The main result of this chapter is the following theorem, which says that any
program using g(n) space can be simulated by a switching network of size
2 0(g(n)).
This means that if we can show a lower size bound on switching networks solving a
given problem, this gives us a corresponding space lower bound for that problem.
Theorem 2.0.3. If f E DSPACE(g(n)) where g(n) is at least logarithmic in n,
then s(f)(n) is at most 2 0(g(n))
In the remainder of this chapter, we first give a brief history of the switching
network model. We then present proofs of this theorem and related results.
2.1
Switching Networks: From Practice to Theory
Although the switching network model has fallen into relative obscurity, it has a rich
history. Switching networks were widely used in practice in the early 20th century.
In the words of Shannon [24],
"In the control and protective circuits of complex electrical systems it is frequently
necessary to make intricate interconnections of relay contacts and switches. Examples of these circuits occur in automatic telephone exchanges, industrial motor control
equipment and in almost any circuits designed to perform complex operations automatically."
Since switching networks were widely used in practice, the early research on switching networks was devoted to finding ways to construct efficient switching networks.
A key observation of Shannon [24], [25] is that if we can write a function f as a
short boolean algebra equation such as
f=
(x 1 V x 2 ) A (-,(x 1 A
x3 )
V X 3 ) then it can
be computed with a corresponding series-parallel switching network. An alternative
technique for constructing switching networks was later found by Lee [13]. Lee [13]
observed that we can construct switching networks from branching programs.
20
However, with the development of the boolean circuit model, which is vastly more
powerful, switching networks became obsolete. As switching networks became obsolete, branching programs were no longer needed for constructing switching networks.
Nevertheless, branching programs found a theoretical use when Masek [17] discovered
that branching programs could be used to analyze space complexity.
Just as the branching program model is now used theoretically, the switching network model can also be used theoretically. As noted later in this chapter, combining
Masek's observation that branching programs can be used to analyze space complexity with Lee's construction of switching networks via branching programs shows that
switching networks can also be used to analyze space complexity. As noted in Chapter
4, Shannon's series-parallel construction of switching networks allows us to analyze
circuit depth with switching networks as well.
2.2
Turing machines and space complexity
Since we are analyzing space complexity, we use Turing machines which have two
tapes, a read-only input tape and a work tape. A formal definition is given below.
Definition 2.2.1. A deterministic Turing machine T = (Q, F, {, , -1}, }, , qo F)
is a 7-tuple where
1.
Q is the
set of states.
2. qo E Q is the initial state.
3. F C
4.
Q is
the set of final states.
F is the set of alphabet symbols.
5. 0 E F is the zero symbol, HE F marks a left boundary, and -iE F marks a right
boundary.
6. E C F \ {H, -1} is the set of input symbols.
21
7. 6 : (Q \ F) x F2
-
Qx
F x { L, S, R} 2 is the transitionfunction which takes the
Turing machine's current state and the symbols currently under the input tape
head and the work tape head and determines which state the Turing macine goes
to next, what symbol is written on the work tape, and whether the input tape
head and the work tape head shift left, stay stationary, or shift right.
Additionally, we assume the following:
1. A Turing machine head cannot shift left when it is on a left boundary and cannot
shift right when it is on a right boundary. Left and right boundaries cannot be
overwritten with anything else.
2. The input tape initially has H at position 0, the input symbols at positions 1
through n, and H at position n + 1. The work tape initially has F- at position 0
and 0 everywhere to the right of it.
3. Both the input tape head and the work tape head start at position 0.
nondeterministic Turing machines are defined in the same way as deterministic Turing machines except that 6(q, 71, '72) is now a subset of
a single element of
Q
Q x F x { L, S, R} 2
rather than
x F x {L, S, R} 2 . This represents the fact that at any given
point there can be several possibilitiesfor what the Turing machine does next.
Definition 2.2.2. We say that a deterministic Turing machine T computes the function
f
: {0, 1}
-+ {0, 1} in space g(n) if
1. E = {0, 1} and F = {ACCEPT, REJECT}.
2. For any input, T halts.
3. Whenever f(x) = 1, the Turing machine ends at ACCEPT on input x.
4.
Whenever f(x) = 0, the Turing machine ends at REJECT on input x.
5. For any input x E {0, 1}, the work tape never moves to the right of position
g(n).
22
-
We say that a nondeterministic Turing machine T computes the function f: {0, 1}"
{0, 1} in space g(n) if
1. E
{O, 1} and F = {ACCEPT, REJECT}.
2. Whenever f(x) = 1, the Turing machine has a nonzero probability of ending at
ACCEPT on input x.
3. For any input x
c {0,
1}", the work tape never moves to the right of position
g(n).
2.3
Switching-and-rectifier networks and nondeterministic space complexity
Before analyzing switching networks, we first consider a closely related model, switchingand-rectifier networks, and show how they capture nondeterministic space complexity.
Definition 2.3.1. A switching-and-rectifier network is a tuple (G, s,t, p) where G
is a directed multi-graph with distinguished vertices s, t and A is a labeling function
that associates with some edges e E E(G) a label p(e) of the form xi = 1 or xi = 0
for some i between 1 and n.
{0, In
_,
We say that this network computes the function f :
{0,1}, where f(x) = 1 if and only if there is a path from s to t such that
each edge of this path either has no label or has a label that is consistent with x.
We take the size of a switching-and-rectifiernetwork to be
f : {0,1}"
|V(G)
, and for a function
{0, 1}, we define rs(f)(n) to be size of the smallest switching-and-
rectifier network computing f.
Remark 2.3.2. Note that switching networks are the same as switching-and-rectifier
networks except that all edges are undirected and we cannot have edges with no label.
However, allowing edges with no label does not increase the power of switching networks, as we can immediately contract all such edges to obtain a smaller switching
network computing the same function where each edge is labeled.
23
The following theorem is folklore.
Theorem 2.3.3. If f c NSPACE(g(n)) where g(n) is at least logarithmic in n,
then rs(f)(n) is at most 2 0(g(n))
Proof. Let T be a nondeterministic Turing machine computing
f
using g(n) space.
To create the corresponding switching-and-rectifier network, first create a vertex vj
for each possible configuration cj of T, where a configuration includes the state of
the Turing machine, where the Turing machine head is on the input tape, where the
Turing machine head is on the read/write tape, and all of the bits on the read/write
tape. Now add edges in the following way:
For each configuration cj1 of T, the head on the input tape is on some bit xi.
All information except for the input is included in the configuration and the Turing
machine T only sees the bit xi, so which configurations T can go to next depends
only on xi. For each other configuration cj, of T, add an edge from v, to vj, with
no label if the T can go from cj, to c12 regardless of the value of xi and add an edge
from vj, to v32 with label xi = b if T can go from cj, to cj, if and only if xi = b, where
b E {0, 1}.
After we have added all of these edges, we obtain our switching-and-rectifier network G by merging all vertices corresponding to an accepting configuaration into one
vertex t. Now if there is a path from s to t in G whose edge labels are all consistent
with the input x, there must have been a nonzero chance that T accepts the input
x. Conversely, if there is a nonzero chance that T accepts the input x then there
must be a path from s to t in G whose edge labels are all consistent with the input
x. Thus, G computes
f.
Now we just need to bound the size of G. There are a constant number of states
of T, there are 29(n) possibilities for what is written on the work tape, there are n + 2
possibilities for where the head is on the input tape and there are g(n) + 1 possibilities
for where the head is on the read/write tape. Thus, there are at most
2 0(9(n))
possible
configurations of T which implies that JV(G)J is at most 20(9(n)), as needed.
l
24
Proposition 2.3.4. Directed connectivity can be solved nondeterministicallyin logarithmic space.
Proof. It is easily verified that the following algorithm (which just takes a random
walk from s) solves directed connectivity nondeterministically in logarithmic space:
1. Start at s
2. Keep track of the current location v, the destination t, and the number of steps
taken so far. At each point, choose randomly where to go next. If we ever arrive
at t, accept. If we get stuck or take at least JV(G)J steps without reaching t, then
D
reject.
Corollary 2.3.5. Directed connectivity is NL-complete with respect to AC0 reductions.
Proof. Looking at the construction in the proof of Theorem 2.3.3, determining whether
a given edge in the switching-and-rectifier network is there or not and what its label
is can be computed by unbounded fan-in circuits of constant depth (where the depth
needed depends on the Turing machine T). This implies that the directed connectivity problem is NL-hard with respect to AC 0 reductions.
By Proposition 2.3.4,
directed connectivity is itself in NL so directed connectivity is NL-complete.
E
Remark 2.3.6. Essentially, the space needed to conpute a function f nondeterministically is log(rs(f)). The only reason this is not quite true is that Turing machines
are a uniform model while switching-and-rectifiernetworks are a nonuniform model.
2.4
Branching programs, switching networks, and
deterministic space complexity
We now define branching programs and show how branching programs and switching
networks capture deterministic space complexity.
Definition 2.4.1. A branchingprogram is a switching-and-rectifiernetwork (G, s, t, it)
which satisfies the following conditions:
25
1. G is acyclic
2. Every vertex either has outdegree 0 or 2. For each vertex with outdegree 2, one
edge has label xi = 0 and the other edge has label xi = 1 for some i.
3. t has outdegree 0.
We take the size of a branchingprogram to be
|V(G)|,
and for a function f:
{0, 1}
-+
{0, 1}, we define b(f)(n) to be size of the smallest branching program computing f.
Theorem 2.4.2. If f C DSPACE(g(n)) where g(n) is at least logarithmic in n,
then b(f)(n) is 20
Proof. By definition, there is a deterministic Turing machine T computing
f in
space
O(g(n)). First take the switching-and-rectifier network G given by the construction
in the proof of Theorem 2.3.3. Since T is deterministic, this switching-and-rectifier
network will satisfy conditions 2 and 3 of Definition 2.4.1 but may not be acyclic.
To fix this, let m = JV(G)j and take m copies of G. Now for each i, take the
edges in the ith copy of G and do the following. Have them start at the same vertex
but end at the corresponding vertex in the (i + 1)th copy of G. This preserves the
structure of G while making it acyclic. Finally, merge all copies of t into one vertex
t and take s to be the first copy of s.
For a given input x, if G has no path from s to t then our modified switching-and-
rectifier network won't either. If G has a path from s to t then it has length at most
m - 1 so we can follow this same path in our modified switching-and-rectifier network
to go from s to t. Thus, our modified switching-rectifier network computes f. It still
satisfies properties 2 and 3 and is now acyclic as well, so it is a branching program.
It has size at most JV(G)
2
and JV(G)j is 20(9(n)), so our branching program has size
20(9(n), as needed.
E
We now give a proof of Lee's [131 observation that we can construct switching
networks from branching programs.
Lemma 2.4.3. If we take a branching program and make all of its edges undirected,
we will obtain a switching network computing the same function.
26
Proof. Let G be a branching program. For each input x, consider the directed graph
GX where V(Gx)
=
V(G) and E(Gx) is the set of edges of G whose labels are consistent
with x. The vertex t corresponding to the accepting state has outdegree 0 and since
G is a branching program, each other vertex of Gx has outdegree at most 1.
Now note that we make the edges of G, undirected, it does not affect whether
or not there is a path from s to t in G,. To see this, note that if there was a path
from s to t in Gx, there will still be a path in Gx from s to t after we make all of the
edges of Gx undirected. For the other direction, if there is a path P from s to t in
Gx after we make all of the edges of Gx undirected, look at whether the edges of P
are going forwards or backwards (with respect to the original direction of the edges
in Gx). Each vertex of Gx has outdegree at most 1 so if we had a backwards edge
followed by a forwards edge this would just get us back to where we were. This is
a contradiction as P cannot have repeated vertices. This implies that P must be a
sequence of forward edges followed by a sequence of backwards edges. But P cannot
end with a backwards edge as t has outdegree 0. Thus P consists entirely of forward
edges and thus P was a path from s to t in G to begin with.
This implies that if G' is the switching network constructed by making all of the
edges in G undirected, then for all inputs x there is a path from s' to t' in G' consistent
with x if and only if there is a path from s to t in G consistent with x. Thus G' and
E
G compute the same function, as needed.
Corollary 2.4.4. For any function
f
: {0, 1} -+ {0, 1}, s(f)(n) < b(f)(n).
Theorem 2.0.3 now follows immediately.
Theorem 2.0.3. If f c DSPACE(g(n)) where g(n) is at least logarithmic in n,
then s(f)(n) is 20(9
Proof. By Theorem 2.4.2, b(f)(n) is
2 0(g(n)).
By Corollary 2.4.4, s(f)(n) < b(f)(n)
E
so s(f)(n) is 20(g(n))
Remark 2.4.5. Looking at the constructions in the proof of Theorem 2.0.3, determining whether a given edge in the switching network is there or not and what its label
27
is can be computed by unbounded fan-in circuits of constant depth (where the depth
needed depends on the Turing machine T). This implies that the directed connectivity
problem is L-hard with respect to AC0 reductions. Reingold [22] proved that undirected connectivity is itself in L so this gives a proof that the undirected connectivity
problem is L-complete with respect to AC0 reductions.
Remark 2.4.6. Essentially, the space needed to conpute a function f deterministically is O(log(s(f)(n))). The only reason this is not quite true is that Turing machines
are a uniform model while switching networks are a non-uniform model.
28
Chapter 3
Bounds on monotone switching
networks for directed connectivity
In this chapter, we present our original work [18] on monotone switching networks
for the directed connectivity problem. The techniques developed here form the foundation for all of our later results. The main result of this chapter is the following
theorem.
problem has size n
)
Theorem 3.0.7. Any monotone switching network solving the directed connectivity
We now give an overview of the remainder of this chapter. In Section 3.1 we
develop definitions specifically for analyzing the directed connectivity problem. In
Section 3.2 we define and prove bounds on a restricted case of monotone switching
networks called certain knowledge switching networks. In Section 3.3 we partially generalize the techniques and results of Section 3.2 to all monotone switching networks.
In Section 3.4 we introduce a very different way of analyzing monotone switching networks, Fourier analysis and invariants. In Section 3.5 we show how all of the previous
material can be synthesized to give a superpolynomial lower size bound on monotone
switching networks solving directed connectivity. Finally, in Section 3.6 we carry out
our analysis more carefully to prove Theorem 3.0.7.
After proving Theorem 3.0.7, we discuss two limitations of this result. As discussed
29
in Section 3.7, there is some ambiguity in how to define monotone-L, so instead of
saying that we have separated monotone-L from monotone-NL, we say that we have
separated monotone analogues of L and NL. In Section 3.8 we note that non-monotone
switching networks can easily solve input graphs consisting of just a path from s to
t and no other edges. Thus, our monotone bounds for solving directed connectivity
on these input graphs cannot be extended to non-monotone switching networks and
bounds for other input graphs are needed to have a chance at proving non-monotone
lower bounds. We return to this question in Chapters 5 and 6.
3.1
Definitions
Since we will be working extensively with the directed connectivity problem, it is
convenient to develop definitions specifically for directed connectivity.
Definition 3.1.1. A switching network for directed connectivity on a set of vertices
V(G) with distinguished vertices s, t is a tuple < G', s', t', pa' > where G' is an undirected multi-graph with distinguished vertices s',t' and p' is a labeling function giving
each edge e' c E(G') a label of the form v1
G
v 2 or
-(V
1
-+
v 2 ) for some vertices
V(G) with v1 -A v 2
.
v 1 ,v 2
-+
Remark 3.1.2. For the remainder of this chapter, all switching networks G' will be
assumed to be switching networks for directed connectivity on a set of vertices V(G)
with distinguised vertices s, t, so we will just write switching network for brevity.
Definition 3.1.3. Whenever we consider a switching network, we assume that our
input graphs G all have vertex set V(G) (though they may have different edges).
We take Vred(G) = V(G)
\
{s,t} and we take n = |Vred(G)|.
We exclude s,t when
considering the size of V(G) because it makes our calculations easier.
Definition 3.1.4.
1. We say that a switching network G' accepts an input graph G if there is a path
P' in G' from s' to t' such that for each edge e' E E(P'), p/(e') is consistent
30
with the input graph G (i.e. of the form e for some edge e C E(G) or
-e
for
some e 0 E(G)).
2. We say that G' is sound if it does not accept any input graphs G which do not
have a path from s to t.
3. We say that G' is complete if it accepts all input graphs G which have a path
from s to t.
4. We say that G' solves directed connectivity if G' is both complete and sound.
5. We take the size of G' to be n' =IV(G')|.
6. We say that G' is monotone if it has no labels of the form -(vi
-
v 2 ).
A monotone switching network solving directed connectivity on V(G) = {s, t, a, b}
s -+ t
Figure 3-1: In this figure, we have a monotone switching network G' solving directed
connectivity, i.e. there is a path from s' to t' in G' whose labels are consistent with the
input graph G if and only if there is a path from s to t in G (where V(G) = {s, t, a, b}).
For example, if we have the edges s -* a, a -- b, and b -+ t in G, so there is a path
from s to t in G, then in G', starting from s', we can take the edge labeled s
-+
a,
then the edge labeled a -+ b, then the edge labeled s -+ a, and finally the edge labeled
b -+ t, and we will reach t'. On the other hand, if G has the edges s -+ a, a -+ b,
b -+ a, and s -+ b and no other edges, so there is no path from s to t, then in G' there
is no edge that we can take to t', so there is no path from s' to t'.
31
A monotone switching network for directed connectivity on
V(G) {s, a, b, t} which is sound but not complete
0,
G1
Figure 3-2: In this figure, we have another monotone switching network G'. This G'
accepts the input graph G if and only G contains either the edge s -+ t, the edges
s
-4
a and a
-
t, or the edges s
-4
b and b
-
t. Thus, this G' is sound but not
complete.
Remark 3.1.5. Figures 3-2 and 3-3 illustrate why solving directed connectivity on a
set of vertices V(G) requires a larger switching network. Roughly speaking, the reason
is that we are trying to simulate a directed graph with an undirected graph.
Remark 3.1.6. It is natural to ask where the examples of Figures 3-1 and 3-4 come
from. As we will see, there are many different ways to construct switching networks
solving directed connectivity on a set of vertices and we will give the particular constructions leading to the switching networks in Figures 3-1 and 3-4 later on. For now,
the readershould just make sure that he/she understands Definition 3.1.1. That said,
it is a good exercise to verify that these switching networks have the claimed properties
and to try and figure out what they are doing.
In this thesis we analyze monotone switching networks. However, rather than
looking at all possible input graphs, we focus on particular sets of input graphs. To
do this, instead of assuming that the switching networks we analyze solve directed
connectivity, we only assume that these switching networks solve the promise problem
where the input graph G is guaranteed to either be in some set I of input graphs
which contain a path from s to t or to not contain a path from s to t.
Definition 3.1.7. Given a set I of input graphs which all contain a path from s to t,
let m(I) be the size of the smallest sound monotone switching network which accepts
32
A monotone switching network for directed connectivity on
V(G) = {s, a, b, t} which is complete but not sound
I -- + b b -- +a
8
a
G
Figure 3-3: In this figure, we have another monotone switching network G'. This G'
accepts the input graph G whenever G contains a path from s -+ t (where V(G) =
{ s, t, a, b}), so it is complete. However, this G' is not sound. To see this, consider the
input graph G with E(G) = {s -+ a, b -+ a, b -+ t}. In G', we can start at s', take
the edge labeled s - a, then the edge labeled b -+ a, then the edge labeled b -+ t,
and we will reach t'. Thus, this G' accepts the given input graph G, but G does not
contain a path from s to t because the edge from b to a goes the wrong way.
all of the input graphs in I.
In this chapter, we focus on input graphs which contain a path from s to t and no
other edges, as they are the minimal YES instances and are thus the hardest input
graphs for a monotone switching network to accept. We have the following definitions.
Definition 3.1.8. If V(G) is a set of vertices with distinguished vertices s, t,
1. Define PV(G),l to be the set of input graphs G such that E(G) = {vo V2, * - - ,v 1 _1
-+
v 1} where v0 = s, vi = t, and vo,
v..
,vi
v 1 , vi
are distinct vertices of
V(G).
2. Define PV(G),<l = U j=1PV(G),j
3. Define PV(G)
=
1
-
PV(G),j
Proposition 3.1.9. A monotone switching network solves directed connectivity on
V(G) if and only if it is sound and accepts every input graph in PV(G)Corollary 3.1.10. The size of the smallest monotone switching network solving directed connectivity (on V(G)) is m(PV(G))-
33
A non-monotone switching network solving
directed connectivity on V(G) = {s, a, b, t}
b__'-( a t
t
G'
Figure 3-4: In this figure, we have a non-monotone switching network G' solving
directed connectivity. Note that the edge with label -,(b - t) and the edge with label
-i(a -+ t) are necessary for G' to be complete. To see this, consider the input graph
G with E(G) = {s -+ a, a -+ b, b -+ t}. To get from s' to t' in G' we must first take
the edge labeled s -- a, then take the edge labeled a -- b, then take the edge labeled
-(a -+ t), and finally take the edge labeled b -÷ t.
In chapters 5 and 6 we will analyze the difficulty of directed connectivity for a
given input graph G. To do this, we consider the promise problem where the input
graph is guaranteed to either be isomormphic to G or to have no path from s to t.
Definition 3.1.11. Given an input graph G, let m(G) = m(I) where I = {G 2
G 2 is obtained from G by permuting the vertices of Vred(G)}
We give this definition here because it will make it easier for us to state some of
the results of this chapter.
3.2
Certain Knowledge Switching Networks
In this section, we define certain knowledge switching networks, a subclass of sound
monotone switching networks for directed connectivity which can be described by
a simple reversible game for solving directed connectivity.
We then prove upper
and lower bounds on the size of certain knowledge switching networks, showing that
certain-knowledge switching networks can match the performance of Savitch's algorithm and this is tight. The main result of this section is the following theorem.
34
Definition 3.2.1. Given a set I of input graphs all of which contain a path from s
to t, let c(I) be the size of the smallest certain-knowledge switching network which
accepts all of the input graphs in I.
Theorem 3.2.2. Let V(G) be a set of vertices with distinguished vertices s, t. Taking
n
Vre(G)J, if I > 2 and n > 2(l - 1)
1
n
then
1)_u
) F* 1 <- C(PV(G),1) <- C( V(G),<1) < nF11
2.
3.2.1
2
lgn+
c(PV(G))
-2
The certain knowledge game for directed connectivity
We can use the following simple reversible game to determine whether there is a path
from s to t in an input graph G.
Definition 3.2.3 (Certain knowledge game). We define a knowledge set K to be a
set of edges between vertices of V(G). An edge u
that there is a path from u to v in G.
-4
v in K represents the knowledge
We do not allow knowledge sets to contain
loops.
In the certain knowledge game, we start with the empty knowledge set K
=
{}
and
use the following types of moves:
v 2 e E(G), we may add or remove v 1
v 2 from K.
1. If we directly see that v1
-
2. If edges v 3
v 5 are both in K and v 3 -4 v 5 , we may add or remove
V3 -+ v 5
-+
v 4 , v4
-+
-
from K.
We win the certain knowledge game if we obtain a knowledge set K containing a path
from s to t.
Proposition 3.2.4. The certain knowledge game is winnable for an input graph G if
and only if there is a path from s to t in G.
35
3.2.2
Adapting the certain knowledge game for monotone switching networks
Intuitively, certain knowledge switching networks are switching networks G' where
each vertex v' E V(G') corresponds to a knowledge set Kv, and the edges between
the vertices of G' correspond to moves from one knowledge set to another. However,
there are two issues that need to be addressed. First, if G' is a switching network,
I',
v', w' c V(G'), and there are edges with label e between u' and v' and between v'
and w', then we may as well add an edge with label e between u' and w'. This edge
now represents not just one move in the game but rather several moves. Thus an
edge in G' with label e should correspond to a sequence of moves from one knowledge
set to another, each of which can be done with just the knowledge that e E E(G).
The second issue is that the basic certain knowledge game has many winning
states but a switching network G' has only one accepting vertex t'. To address this,
we need to merge all of the winning states of the game into one state. To do this,
we add a move to the game allowing us to go from any winning state to any other
winning state.
Definition 3.2.5 (Modified certain knowledge game). In the modified certain knowledge game, we start with the empty knowledge set K
=
{}
and use the following types
of moves:
1. If we directly see that v1
-+*v 2
2. If edges v 3
v 5 are both in K and v 3 f v 5 , we may add or remove
v3
-+
3. If s
-+
v4 , v 4
-+
E E(G), we may add or remove v1
-
v 2 from K.
v 5 from K.
-4
t E K then we can add or remove any other edge from K.
We win the modified certain knowledge game if we obtain a knowledge set K containing
a path from s to t.
Remark 3.2.6. In the modified certain knowledge game, an edge v 1
.-
+
v 2 in K now
represents knowing that either there is a path from v1 to v 2 in G or there is a path
from s to t in G.
36
Proposition 3.2.7. The modified certain knowledge game is winnable for an input
graph G if and only if there is a path from s to t in G.
Definition 3.2.8. We say a monotone switching network G' is a certain knowledge
switching network if we can assign a knowledge set Kv, to each vertex v' G V(G') so
that the following conditions hold:
1. Ks' = {}
2. KtI
contains a path from s to t (this may or may not be just the edge s
3. If there is an edge with label e = v1
-
-
t)
v 2 between vertices v' and w' in G', then
we can go from Kv, to K&' with a sequence of moves in the modified certain
knowledge game, all of which can be done using only the knowledge that v1
v 2 c E(G).
We call such an assignment of knowledge sets to vertices of G' a certain knowledge
description of G'.
Proposition 3.2.9. Every certain knowledge switching network is sound.
Proof. If there is a path from s' to t' in G' which is consistent with an input graph G
then it corresponds to a sequence of moves in the modified certain knowledge game
from K,, =
{}
to a K,, containing a path from s to t where each move can be done
with an edge in G. This implies that the modified certain knowledge game can be
won for the input graph G, so there is a path from s to t in G.
3.2.3
D
A partial order on knowledge sets
There is a partial order on knowledge sets as follows. This allows us to restate the
definition of certain knowledge switching networks more cleanly and will be important
in Section 3.3
Definition 3.2.10. Define the transitive closure R of a knowledge set K to be
37
a}
K={s
s
'
K =}
a
K=ese
a
b
an
a, s - b})
b
b
t
a
K =Is -+
b
K
b}
-+fs
=
tK
t
Figure 3-5: In this figure, we have a certain knowledge switching network G' solving directed connectivity (on V (G) = {s, t, a, b}) together with a certain knowledge
description for it.
k.
=
{V1
-4 V2 : V1,V2 c
are all in K} if s--4 t
2. K = {vji
V2 : V1,iV2 E
V (G), v1
V2,
there is a path from v1 to V2 whose edges
K
V(G), vi
V2}
if s -+ t E K
1. We say that K 1 < K 2 if K 1 C K
.
Definition 3.2.11.
2. We say that K 1 = K 2 if K1 = K
2
.
2
Remark 3.2.12. Intuitively, K describes all knowledge which can be deduced from
K. The statement that K1 < K2 means that all of the information that K1 represents
about G is included in the information that K 2 represents about G.
Proposition 3.2.13. If K 1 , K 2 , K3 are knowledge sets for V(G), then
1. K1 < K1 (reflexivity)
2. If K1 < K 2 and K2 < K1 then K1
K2
(antisymmetry)
3. If K1 < K 2 and K 2 < K3 then K1 < K3 (transitivity)
Proposition 3.2.14.
38
1. For knowledge sets K1 , K 2 , there is a sequence of moves from K1 to K2 in the
modified certain knowledge game which does not require any information about
the input graph G if and only if K1
K2.
2. For knowledge sets K1, K2 and a possible edge e of G, there is a sequence of
moves from K1 to K 2 in the modified certain knowledge game which only requires
the information that e E E(G) if and only if K1 U {e}
1K2U {e}.
Corollary 3.2.15. We can restate the definition of certain knowledge switching networks as follows. A monotone switching network G' is a certain knowledge switching
network if we can assign a knowledge set Kv, to each vertex v' e V(G') so that the
following conditions hold:
1. KSI
2. Kt,
=
{}
{s -+ t}
3. If there is an edge with label e = v1
-
v 2 between vertices v' and w' in G', then
Kv, U {e} = Kw U {e}
3.2.4
Certain knowledge switching networks and reachability
from s
While certain knowledge switching networks can consider all paths in G, most of
our examples only consider reachability from s. For these examples, the following
definition and lemma are extremely useful.
Definition 3.2.16. For each subset V C Vred(G), define Kv = {s
Lemma 3.2.17. For any v 1 ,v 2 E Vred(G) and V
C
-
Vred(G) with v1
v : v E V}
c V there is a
sequence of moves in the modified certain knowledge game from Kv to Kvu{v 2} which
only requires the knowledge that v 1
Proof. Starting from K
1. Add v,
-*
=
-+
v2
G
E(G)
Kv, take the following sequence of moves:
v 2 to K.
39
2. Add s
-+
v 2 to K (we already have s -- v, and vi
3. Remove v,
--
v 2 in K)
v 2 from K.
-
El
We are now at K = KvuV{ 2}
Remark 3.2.18. Note that there is no sequence of moves in the modified certain
knowledge game from Kvl to KIV 2} which only requires the knowledge that v1 -+
V2
E E(G). The reason is that this is not reversible. We are forgetting the fact that
there is a path from s to v1 in G or a path from s to t in G and forgetting information
is irreversible. Also, while we can deduce that there is a path from s to v 2 in G from
the fact that there is a path from s to v1 in G and v 1
-+
v 2 e E(G), we cannot deduce
that there is a path from s to v1 in G from the fact that there is a path from s to v 2
in G and v 1
v2
C
E(G).
s-t
s
b
at
Kf a,bj
Kja)
i
s +b
Kjbj
a
t
C
t
tl
Kfa,c)
Ct
K{c)j
K(b,c)
Figure 3-6: In this figure, we have a certain knowledge switching network G' solving
directed connectivity (on V(G) = {s, t, a, b, c}) together with a certain-knowledge
description for it. By default, we take K., = {} and Kt, = {s -+ t}.
Remark 3.2.19. Figure 3-6 shows how removing old information can help in finding
a path from s to t with the modified certain knowledge game. If G is the input graph
40
with V(G) = { s, t, a, b, c} and E(G) = {s
-+
a, a
-+
s' to t' in G', we must first take the edge labeled s
b, b -+ c, c -+ t}, then to get from
-
a to reach K = K{a}, then take
the edge labeled a -+ b to reach K = K{a,b}, then go "backwards" along the edge labeled
s
a to reach K =
-
K{b},
then take the edge labeled b -s c to reach K = K{b,,} and
finally take the edge labeled c -+ t to reach t'.
3.2.5
An upper bound on certain-knowledge switching networks
While the certain knowledge condition is restrictive, certain knowledge switching
networks nevertheless have considerable power. In particular, the following certain
knowledge switching networks match the power of Savitch's algorithm.
Definition 3.2.20. Given a set of vertices V(G) with distinguished vertices s, t, let
G'(V(G), r) be the certain knowledge switching network with vertices t' U {v' : V C
< r} and all labeled edges allowed by condition 3 of Definition 3.2.8,
each v'v has knowledge set Kv, s' = v',, and K, = {s - t}.
Vred(G), V|
where
Definition 3.2.21. Define G'(V(G)) = G'(V(G), n)
Example 3.2.22. The certain-knowledge switching network shown in Figure 3-5 is
G'({s, t, a, b}) = G'({ s, t, a, b}, 2). The certain-knowledge switching network shown
in Figure 3-6 is G'({s, t, a, b, c}, 2).
Theorem 3.2.23. For all 1 > 1, c(Pv(G),<l)
() + 2
(f
Proof. Consider the switching network G' = G'(V(G), [lg 11). G' is a certain-knowledge
switching network with IV(G') =
(n) +2,
so to prove the theorem it is sufficient
to show that G' accepts all of the input graphs in PV(G),<l.
Consider an input graph G - Pv(G),<l. E(G) = {vi -÷ vj
vo
=
s', vj
=
t', and
j
: 0 < i < J - 1} where
< 1. We will show that G' accepts G by showing that we can
win at the following reversible pebble game without placing more than [lg 1] pebbles
on the vertices vi,. -,
vj_ 1 . This pebble game was introduced by Bennet
time/space tradeoffs in computation.
41
[3
to study
Definition 3.2.24. In the reversible pebble game for directed connectivity, we start
with a pebble on s.
We win the game if we can place a pebble on t.
We are only
allowed to use the following type of move:
1. If there is a pebble on a vertex v1 C V(G) and an edge v1
-+
v 2 C E(G) where
v 1 # v 2 , we may add or remove a pebble from the vertex v 2
Remark 3.2.25. Note that after putting a pebble on a vertex we do NOT remove the
pebble on the previous vertex. This is a key difference between this game and the JAG
model.
Definition 3.2.26. Let G be the input graph with vertices {vo, v 1 , v 2 , -
{Vi -+
v 1 : i
>
} and edges
0}. Assume that we start with a pebble on s = vo.
1. Let f (i) be the minimal number m such that we can place a pebble on vi without
ever having more than m pebbles on the vertices v 1 , - - - , vi1.
2. Let g(i) be the minimal number m such that we can reach the game state where
there is a pebble on v 0 and vi and no other pebbles without ever having more
than m pebbles on the vertices v 1 ,- -- ,vi.
We need to show that f(l) < [log 1] for all 1 > 1. This follows immediately from
the following lemma. This result was first proved by Bennet [31 and we present a
proof here for convenience.
Lemma 3.2.27. For all integers i > 1,
1.- f (i <- [log ( I
2. g(i) < [log (i)] + 1
Proof.
Proposition 3.2.28. For all i > 1, g(i) < f(i) + 1
Proof. We have a sequence of moves allowing us to place a pebble on vi while placing
at most f(i) pebbles on the vertices v, ... , vi_ 1 . After completing this sequence, run
42
the sequence in reverse except that we do not remove the pebble on vi. When we are
done, we only have pebbles on vo and vi and at all times we have at most f(i) + 1
pebbles on the vertices vi, -.. , vi, as needed.
Proposition 3.2.29. For all i, j > 1, f (i + j) < max {g(i), f (j) + 1}
Proof. We first take a sequence of moves allowing us to reach the state where there
are pebbles on vo and vi and no other pebbles without ever having more than g(i)
pebbles on vi,. -, vi. We then take the sequence of moves allowing us to put a pebble
on vj without ever having more than f(j) pebbles on v 1 , -- - , vj_1 except that we shift
the sequence of moves to the right by i. This sequence now allows us to start from
the state where there are pebbles on vo and vi and no other pebbles and put a pebble
on vi+j without ever having a pebble on v,, - -- , vi_ 1 or having more than f(j) + 1
pebbles on vi, - -- , vi+j-1. Composing these two sequences of moves gives a sequence
of moves putting a pebble on vi+j while never having more than max {g(i), f(j) + 1}
pebbles on vi, -- - , vi+-1
Applying Proposition 3.2.28 to Proposition 3.2.29 we obtain that for all i, j, f(i+
j)
< max {f(i), f(j)} + 1. Clearly f(1) = 0 so we can easily show by induction that
for all i > 1, f(i) 5 [log (i)]. Applying Proposition 3.2.28 again we obtain that for
E
all i > 1, g(i) < [log (i)] + 1, as needed.
By Lemma 3.2.27 there is a sequence of moves for the reversible pebble game
on G allowing us to win without placing more than [ig 1] pebbles on the vertices
vi, - -- , vj_ 1 . We now translate this winning sequence of moves into a sequence of
moves in G'. If we have pebbles on a set of vertices V C {vO,
...
, vj_ 1} with vi E V
and our move is to place a pebble on vi+ 1 , this corresponds to moving from Ky to
Kvulm,
} in G' along an edge labeled vi
-
vj+1 (we know such an edge exists because
of Lemma 3.2.17). Similarly, if we have pebbles on a set of vertices V C {vo, ** , vi _1 }
with vi, vj+1 E V and our move is to remove a pebble from vi+1, this corresponds to
moving from Kvyf,2 1 } to Kv in G' along an edge labeled vi -* vi+1 . Finally, if we
have pebbles on a set of vertices V C {vo, *** , vj1} with vj_ 1 E V and our move is
43
to place a pebble on vj = t, this corresponds to moving from Ky to Kt, in G' along
an edge labeled vy_1 -+ t. The entire sequence of moves corresponds to a walk from s'
to t' in G' whose edge labels are all consistent with G so G' accepts the input graph
G, as needed.
3.2.6
A lower size bound on certain-knowledge switching networks
In this subsection, we prove lower bounds on c(PV(G),<l).
Definition 3.2.30. For a knowledge set K such that K 0 Kt,, define
V(K) ={v : v C Vre(G),-]w C V(G) : w # v,v -+w E E(G) or w -+v c E(G)}
)
Proposition 3.2.31. If K1 = K2 * Kt, then V(K1 ) = V(K 2
The heart of our lower bound argument is the following lemma, which says that if
we have an input graph G containing a path P from s to t and no other edges, then
for any walk W' from s' to t' in G' whose edge labels are all in E(P), there is one
vertex v' on W' such that V(Kv') contains many vertices of P and no vertices not in
P, which gives a lot of information about P.
Lemma 3.2.32. Let G' be a certain knowledge switching network. For any certain
knowledge description of G' and any path P = s -+ v 1
+ -..
--+ v1 _ 1
-
t, if G is the
labels are all in G from a vertex v'tart with K,
- Ks, to a vertex v' n with K,
Kt, then W' passes through a vertex v' such that Kv' 0 Kt,, V(Kv,)
C
{v 1,
, 1
-
input graph with vertex set V(G) and E(G) = E(P), if W' is a walk in G' whose edge
-
and |V(Kv')| > lg(l)|.
However, the proof of this lemma is long, so we relegate it to the appendix.
Instead, we show here that this lemma holds if all Kv' are of the form Kv where
V
C
Vred(G) or are equivalent to KtI. This is equivalent to proving the following
result about the reversible pebbling game.
44
Lemma 3.2.33. f(l) > Flg 1]
This result was proved by Li and Vitanyi [15]. We give a short alternative proof
of this result here which emphasizes the role of reversiblity.
Proof. We have that f(1)
=
0, f(2)
=
1, and f(i) is an nondecreasing function of i,
so this follows immediately from the following claim.
Lemma 3.2.34. For all i> 2, f(2i - 1) > f(i) + 1
Proof. Note that if we merge all of the vertices vi,--. , vi_
1
playing the reversible pebbling game with the vertices s, vi, ...
with s
,
=
vo, we are
v2 -1
1 and trying to
pebble v 2i-1. By definition, this requires placing at least f(i) pebbles on the vertices
vi ...
,
v 2i- 2 at some point.
If at this point, we have any pebble on the vertices v 1 , - -,
vi_ 1 , then we have
used f(i) + 1 pebbles. Thus, we may assume that we have no pebbles on the vertices
v 1 , ** , vi_ 1 . Now let vj be the leftmost vertex that is pebbled and run the sequence
of moves we used to reach this state in reverse. At some point, we must remove a
pebble on vj so that we can reach the initial state of no pebbles anywhere. However,
to do this, we must have first had f(j)
f(i) pebbles on the vertices vi, -- - , vji.
Moreover, at this point we still had a pebble on vj so we had a total of at least f(i) + 1
D
pebbles placed, as needed.
This completes the proof of Lemma 3.2.32 when all Kv' are of the form Ky where
D
V Q; Vred(G) or are equivalent to Kt,.
The other part of our lower bound proof is finding a large collection of paths such
that each pair of paths has very few vertices in common.
Lemma 3.2.35. If m, k1 , k2 are non-negative integers and there is a prime p such
that k 2 < k1 < p and m > pk 1 , then there is a collection of pk2+1 subsets of [0, m - 1]
such that each pair of subsets has at most k 2 elements in common.
Proof. To obtain our collection, first take the set of all polynomials in Fp[x] of degree
at most k 2 where Fp is the integers modulo p. There are pk
45
2
+1 such
polynomials. For
each such polynomial f(x), let Sf
polynomials fi and
f2
=
{(x, f(x)) : 0 < x < ki}. For any two distinct
of degree at most k 2 , S OSf2 = {(x,
-f 2 (x) = 0}.
is a non-zero polynomial in Fp[x] of degree at most k 2 , so there are at most
-
k2
x E Fp such that fi(x) - f 2 (x) = 0. Thus,
Isf, n sf2I
k2
-
fi
f2
fi(x) : fx(x)
We now translate these sets into subsets of [0, pki - 1] by using the map
#:[0,pkl
O(x)
- 1] --+ [0, ki - 1] x Fp
([x], (x mod p))
Taking our subsets to be
#F1 (Si)
for all f E lF[x] of degree at most k 2 , each such
subset of [0, pki -- 1] ; [0, m-1] has ki elements. Since
any distinct fl, f2 c Fp[x] of degree at most k 2 ,
# is injective
J4O-(Sf 1 )n#- I(S
2
and surjective, for
)1 = I|f1OSf2|I
k2
D
and this completes the proof.
then there is a collection of at least
( n )k 2 +1
'
Corollary 3.2.36. If n > 2 and k 1 , k 2 are non-negative integers with k 2 < k1
paths from s to t of length k 1 + 1 on
the vertices V(G) such that each pair of paths has at most k2 vertices in common
(excluding s and t).
Proof. We prove this result using Lemma 3.2.35 and a suitable prime number p chosen
using Bertrand's postulate.
Theorem 3.2.37 (Bertrand's postulate). For any integer m > 3 there is a prime p
such that m <p < 2m - 2
Corollary 3.2.38. For any real number m > 1 there is a prime p such that m < p <
2m.
Proof. By Bertrand's postulate, for any real number m > 3 there is a prime p such
that m < rm] < p < 2[rm] - 2 < 2m. If rn c [1, 2] then m < 2 < 2m. If m E [2, 3]
then m < 3 < 2m.
By Corollary 3.2.38 we can take a prime p such that
that n > pk1 and since k, <
we have that k 2
The result now follows from Lemma 3.2.35.
<
<p <
. We now have
which implies that k, <
-
< p.
1
We now prove the following lower bound on certain knowledge switching networks:
46
Theorem 3.2.39. If 1 > 2 and n > 2(1 - 1)2, then c(P,1 ) > (2(1-
Proof. Taking ki = I - 1 and k2
collection of at least
(2 n)k 2 +1
=
Flg 1] - 1 and using Corollary 3.2.36, we have a
( 2 (l 1 ) F11 paths of length 1 from s to t on the set of
vertices V(G) such that each pair of paths P, P has at most k 2 vertices in common
(excluding s and t). However, by Lemma 3.2.32, for any certain knowledge switching
network G' which accepts all of the input graphs in P,,, we can associate a vertex
v' in G' to each path Pi in our collection such that IV(K,,)l > k 2 and V(K,,) is a
subset of the vertices of P. This implies that we cannot have v' = v for any i = j
as otherwise we would have that IV(Kv')l = IV(Kv')l > k 2 and IV(K,')I is a subset
of the vertices of both P and P, which is impossible as any two distinct paths in
our collection have at most k 2 vertices in common. Thus, IV(G')
> (2(1n-1)) [g
E
needed.
3.2.7
1, as
Simplified bounds on certain-knowledge switching networks
We now use Theorems 3.2.23 and 3.2.39 to prove Theorem 3.2.2.
Theorem 3.2.2. Let V(G) be a set of vertices with distinguished vertices s, t. Taking
n = |Vred(G)|, if 1 > 2 and n > 2(l - 1)2 then
1. (12(2.
1 ))
lgl]
n41L"-12
C(PV(G),l)
c(PV(G)) <
2
+- ~11911
c(PV(G),<1)
gn+
2
Proof. For the first statement, the lower bound is just Theorem 3.2.39. To prove the
upper bound, note that by Theorem 3.2.23 we have that c(P",
If [lgl] =1 then 1 = 2 so
n > 2(l
1)2 >
zY41
()
-
n
2[lg 1]. This implies that E
g1 1]
1
1
) < E 1
(n)
+ 2.
= n. If [lgl] > 1 then 1 > 2 so
(n)
< [lgl1 ([Inl])
< W1911, as needed.
For the second statment, the upper bound follows immediately from the upper
bound of the first statement.
For the lower bound, taking 1 = []
47
by Theorem
3.2.39 we have that
C(PV(G))
ng-4
c(PV(G),i)
0
S gn-
n
2(1 - 1)
(-
DO
>
(
jy~ ig-
2,1
g
-1gn
Remark 3.2.40. As noted in Remark 2.4.6, a size bound of E(s(n)) on switching
networks solving a problem roughly corresponds to a space bound of 9(lg (s(n))) on
algorithms solving that problem. Thus, the size bounds of Theorem 3.2.2 correspond
to a space bound of E( lg 1] Ig n) for finding all paths of length at most 1 and a space
bound of E((lg n) 2 ) for finding all paths, which is exactly the performance of Savitch's
algorithm.
3.3
Elementary results on monotone switching networks
In this section, we see which of our techniques and results about certain knowledge
switching networks can be generalized to all monotone switching networks.
The most important thing to note is that for general sound monotone switching
networks, at any given vertex v' we may not be certain of which paths are in G.
Instead, we will have several possibilities for which paths are in G and will only
know that at least one of them holds.
We begin by defining a knowledge game
for directed conectivity which takes this difficulty into account. We show that any
sound monotone switching network can be described using this game. We then show
a nice simplification of sound monotone switching networks which can be done by
increasing their size by a factor of at most n. Finally, we show a partial reduction
of sound monotone switching networks to certain knowledge switching networks. By
itself, this reduction is not strong enough to prove good lower size bounds. However,
it will play an important role in Section 3.5.
48
3.3.1
The knowledge game for directed connectivity
Definition 3.3.1. A state of knowledge J is a multi-set {K1 , - - - , K,} of knowledge
sets (we can have duplicates in J). In the knowledge game for directed connectivity, J
represents knowing that for at least one i E [1, m] the knowledge about G represented
by Ki is true.
Example 3.3.2. If J = { K{a}, K{b}, K{} then J represents knowing that either
there is a path from s to a in G, a path from s to b in G, a path from s to c in G, or
a path from s to t in G.
Definition 3.3.3. In the knowledge game for directed connectivity, we start at J.,
{{}}
and we win if we can get to J, = {{s
--
t}}. We are allowed to use the following
types of moves. If J = { K1 , - - - , Kn} then
1. If we directly see that an edge v1 -+ v 2 is in G we may add or remove v1 -+ v 2
from any K.
2. If v 3
3. If s
v4 ,v 4
-+
-
-+
v5
e Ki andV3 : v 5 we may add or remove v 3 --+ v 5 from K.
t G Ki we may add or remove any other edge from Ki.
4. If i, j C [1,m], i
#
j, and Ki C K then we may remove K from J.
5. If K is a knowledge set such that Ki
C
K for some i E [1, m] then we may add
K to J.
Remark 3.3.4.
The knowledge game for directed connectivity is a generalization of
the certain knowledge game for directed connectivity. The moves which are new are
the moves of types 4 and 5. Moves of type 4 make sense because if we know that
Kj implies Ki and have the statement that Ki OR K is true then this statement is
equivalent to the statement that Ki is true. Moves of type 5 are the inverse of moves
of type
4 so we still have reversibility.
Proposition 3.3.5. It is possible to win the knowledge game for directed connectivity
for an input graph G if and only if there is a path from s to t in G.
49
3.3.2
A partial order on states of knowledge
We have a partial order on states of knowledge as follows.
Definition 3.3.6.
1. We say thatJi = {K 11 ,-
, Kimi} < J2 ={K2 1 ,-
,Kim 2 } if for all j E [1, m 2 ]
there is an i E [1,m1 ] such that K1 i < K 2
We say that J
= J2 if J1
<
J2 and J2
<
J1
.
2.
Remark 3.3.7. Intuitively, the statement that J1 < J2 means that all of the information that J1 represnents about G is included in the information that J2 represents
about G.
Proposition 3.3.8. If Ji, J2 , J3 are states of knowledge then
1. J1 < J1 (reflexivity)
2. If J1 < J2 and J2 < J1 then J1 = J2 (antisymmetry)
3. If J1 < J2 and J2 < J3 then J1 < J3 (transitivity)
We now show the connection between this partial order and the knowledge game
for directed connectivity.
Proposition 3.3.9. For states of knowledge J1 , J2 , if we can go from J1 to J2 in the
knowledge game for directed connectivity with no information about the input graph
.
G then J1 = J2
Proof. By transitivity, to show that if we can get from J1 to J2 in the knowledge
game for directed connectivity with no information about the input graph G then
Ji - J 2 it is sufficient to show that if we can get from J1 to J2 in the knowledge
game for directed connectivity with a single move then J - J2 . J can be written as
Ji = {Ki,
...
,
Kim} and we have the following cases:
1. If we use a move of type 2 or 3 altering some knowledge set Kj to reach J2 then
J2 = {K 2 1 ,- .. , K 2m} where K2 j = Kli for all i
Kii =-K2j so J 1 _ J2
50
# j and Kjj = K 2j. For all i,
2. If we use a move of type 4 to delete some knowledge set Kj from Ji then
J2 =
{K 2 1 , ... , K 2(-1), K 2(j+1), -- - , K2m} where for all i # j, K 2i
there exists a
J1
<
12
#j
J2 . For all i #
=
K1l and
such that K 2 i 2 = K1 j2 < K1 j. For all i z j, Ki < K 2i so
j,
K 2 i < K1i and
K2i2
< Ki so J2 < J1 . Thus, J - J2 as
needed.
3. If we use a move of type 5 to add some knowledge set K to J then
J2 =
{K2,
, K2m,
K2(m+1)} where for all i E
[1, m], K 2i
=
Kli and there is
a j E [1, m] such that Ki < K2 (m+) = K. For all i E [1, m], K 2 < Kii so
J2 < Ji. Kij < K 2 (m+l) and for all i E [1, m], Kii < K2 so J,
<
J2 . Thus,
J, - J2 as needed.
To show the converse to Proposition 3.3.9, we use the following lemma.
Lemma 3.3.10. If J = {K 11 ,... ,Kimi} = J2 = {K2 1 ,.--- ,K2m 2 } then there is a
set I C [1, mi], a set
a function
f2
12 C
: [1, M 2 ] \
[1,
12 -+
n 2]
of equal size to 1, a function fi : [1, m1 ] \ I1
> 11,
I2 such that
I2, and a perfect matching < : I1 -
1. For all i c [1, in
1 ] \ 1, Klf1i() < Ki
2. For all j
G [1, iM 2] \ I,
K 2f 2 (,)
<
K 23
3. For all i E 1, K1l = K20(i).
Proof. Consider the graph formed as follows.
The vertices of this graph will be
the knowledge sets {Kii, i E [1, m1]} U {K 2j, j E [1, m 2]}. Since J1 < J2 , for each
j
E [1, M 2] there is an i E [1, m1 ] such that K1l < K 2j. Draw a directed edge from
each K 23 to the corresponding Kii (if there are more then one possible i, just choose
one of them). Since J2 < J 1, for each i C [1, m1 ] there is a
j
E [1, M 2 ] such that
K 23 < K1 i. Draw a directed edge from each Ki to the corresponding K 23 (if there
are more then one possible
j, just choose
one of them).
After adding all of these edge, we have a bipartite graph where each vertex has
outdegree 1. This graph must have the structure of a set of cycles along with paths
51
leading into the cycles. Choose 11 and I2 such that for each cycle C there is exactly
one ic E
AI
and exactly one Jc C 12 such that Klic is in C and K 2jc is in C. Then
#(ic)
for all cycles C set
= jc. We know we can do this because there cannot be a
cycle consisting entirely of vertices of the form Kli or a cycle consisting entirely of
vertices of the form K2,. We then choose the functions fi and
f2
as follows.
1. For all i E [1, m1 ] \ I, there is a cycle C such that there is a path from Kli to
We take fi(i) = ic.
C and thus a path from Kii to Ki.
2. For all j C [1, M 2] \ 12 there is a cycle C such that there is a path from K 2 j to
C and thus a path from K2j to K 2 jc. We take
f2(j) =
jc.
Now note that an edge from a knowledge set K, to a knowledge set K 2 implies that
K2
< K 1 . By transitivity, a path from a knowledge K1 to a knowledge set K 2 also
implies that K 2 < K1 . This implies that for any cycle all knowledge sets in the cycle
are equivalent. The result now follows immediately because
1. For all cycles C, Klic and K 24(ic) = K2jc are in the same cycle.
2. For all i E [1, mi]
\
3. For all j E [1, iM2]
\
1 there is a path from K1 2 to Klf 1 (i)
there is a path from K 2 to K 2f 2 (j)
2
Corollary 3.3.11. For states of knowledge Ji, J2 , we can go from J1 to J2 in the
knowledge game for directed connectivity with no information about the input graph
.
G if and only if J1 = J2
Proof. The only if part is just Proposition 3.3.9. For the if part, assume that J1
{K 1 1 , ...
, KimI}
- J2 = {K 2 1 , ...
,
K2m2 }.
By Lemma 3.3.10 there is a set I, C
[1, m1 ], a set I2 C [1, m 2] of equal size to I1, a function fi : [1, ml] \ I, -
function
f2 :
[1, M 2 ]
\ 12
-
12,
and a perfect matching 0 : 1 -+ 12 such that
1. For all i c [1, mi 1 ] \11, KIf 1(i) < Ki
'2,
K 2f 2 (j) < K 2
,
2. For all j E [1, M 2 ] \
52
=
1, a
3. For all i E I,, Ki
=-
K2+()
We will go from J= J to J2 in the knowledge game for directed connectivity using
the following steps.
1. Use moves of type 2 and 3 to replace each Kli with K1 i.
2. For all i E [1, m 1 ] \ I,, Kif,(j) < Kli which implies that Kif(j) C K12 . We can
thus use moves of type 4 to delete K 1i for all i E [1, Mi] \ I,.
i E 1}. For all i E A1, Kii - K24(j) so
3. We are now at J = {Ki2
Thus, J
=
{k
: j E
2
12}.
For all
j
,\
E [1,iM 2
12, K 2f2(j)
=
kii
k24(j).
K 2j which
implies that R 2f 2 (j) C K 2j. We can thus use moves of type 5 to add K 2j for all
j E [1, m 2 ] \ 124. We finish by using type 2 and 3 to replace each K 21 with K 2j and obtain J = J2
We have similar results when we directly see that an edge e is in the input graph
G.
Definition 3.3.12. For a state of knowledge J = {K 1 , - -- , K}
and an edge e, define
J U {e} = {K1 U {e}, ... , Km U {e}}
Lemma 3.3.13. For states of knowledge J1 , J2 , we can go from J1 to J2 in the
knowledge game for directed connectivity with the information that v 1
if and only if J1 U {v 1
Proof. If J1 U {v 1
J1 U {v 1
-
-*
-
v2 }
J1 U {v 1
v2}
v 2} to J2 U {v 1
J2 U {v 1
-
-*
-*
-
v 2 C E(G)
v2}
v 2 } then by Lemma 3.3.10 we can go from
v2 } in the knowledge game for directed connectivity with
no information about the input graph G. With the information that v 1
we can go from J1 to J1 U {v 1
-+
v 2 } and from J 2 U {vi
-÷
-
v 2 E E(G)
v 2 } to J 2 in the knowledge
game for directed connectivity using moves of type 1. Thus, we can go from J1 to J2
in the knowledge game for directed connectivity, as needed.
53
For the converse, note that for any states of knowledge J1 , J2 , for any sequence of
moves in the knowledge game for directed connectivity to go from J = J to J= J2,
if we replace J with J U {v l
v 2} at each step we will still have a correct sequence
of moves. Moreover, all moves of type 1 now correspond to doing nothing. This
implies that we can get from J U {vi
-a
v 2} to J2 U {vi
-
v 2 } in the knowledge game
for directed connectivity without knowing anything about the input graph G so by
Proposition 3.3.9 we have that J1 U {vI
v2}
-+
J2 U {vi
-
v 2 }, as needed.
D
Knowledge description of monotone switching networks
3.3.3
In this subsection, we show that all sound monotone switching networks can be described in terms of the knowledge game.
Definition 3.3.14. If G' is a monotone switching network, we call an assignment of
states of knowledge J,, to vertices v' of G' a knowledge description if the following
conditions hold:
1. J',{{}}
2. Jt/
{{s
-+
t}} or Jt, =
{}
3. If there is an edge with label e = v 1
-
v 2 between vertices v' and w' in G' then
Jv' U {e} - Jw, U {e}.
Remark 3.3.15. It is impossible to reach the state of knowledge J
{{}}
in the knowledge game for directed connectivity. If J,, =
{}
=
{}
from J,=
this says that the
vertex v' is impossible to reach from s' regardless of the input graph G.
Proposition 3.3.16. A monotone switching network G' has a knowledge description
if and only if it is sound.
Proof. If G' has a knowledge description then it is sound because we can only win
the knowledge game for directed connectivity if the input graph G has a path from s
to t. Conversely, given a sound monotone switching network G', set J
is a walk from s' to v' in G' whose edge labels are all in E}.
54
=
{E : there
If there is an edge with label e between vertices v' and w' in G' then for every
K E J', K U {e} E Jw/.
Obviously, K U {e} < K U {e} so this implies that
Jw, U {e} < Jv, U {e}. By a symmetrical argument, we also have that Jv,
<
Jw, U {e}.
Thus, Jv, U {e} = J, U {e}, as needed.
{{}} and Jt,
Now we just need to check that J,,
{}
{{s
-+
t}} or J
{{}}. Since G'
J,, and can be used to delete everything else in J,,. Thus J,,
is sound, for every K E Jt', K contains a path from s to t so K -{s
moves of type 2 and 3 we can transform every K in Jt' into K
can use moves of type 4 to delete all but one copy of {s
had Jt =
{}
or we are left with {{s
-
-+
t}}. Thus Jtl
{s
-
= {}.
-+
--+ t}. Using
t} and then we
t}, so either we originally
{{s
-
t}} or J,=
{}, as
D
needed.
s
t
Kj.1
K(a,b)
K~b,,)
Kja,c)
Kg,1
K, c,a)
Figure 3-7: A monotone switching network that solves directed connectivity on
V(G) = {s, a, b, c, t} together with a knowledge description of it. The label inside
each vertex gives the J for that vertex, with each line corresponding to one of its K.
By default we take J,, {{}} and J, = {{s - t}}.
3.3.4
Reduction to reachability from s
In this subsection, we prove the following theorem which shows that there is little
loss in only considering monotone switching networks G' which only make deductions
based on reachability from s.
55
Theorem 3.3.17. If (G', s', t', pi') is a sound monotone switching network, then there
is a sound monotone switching network (G', s', t', /p') such that G' accepts exactly the
same inputs as G', |V(G')| :
(n + 1)IV(G')j, and G' has a knowledge description
where for any vertex v' of G', for any K in Jv,, K
{Kv : V
C
Vred(G)} U { Kt}
Proof. We construct G' by taking n +1 copies of G' and making the s' of each copy
equal to the t' of the previous copy. We take s' for G' to be the s' of the first copy
of G' and t' for G' to be the t' of the last copy of G'. Clearly, G' accepts exactly the
same inputs as G' and we have that IV(G') < (n + 1)1V(G')|.
Now for a vertex v' we construct J, as follows. For each walk W' from s' to v' in
G'2, create a K for that walk as follows:
1. Start with the set Xo = {s} of vertices in G.
2. Let ei = vi -+ wi be the edge in G which is the label of the ith edge in W'. Take
Xi = Xi_
1
if vi 0 Xi
1
and take Xi = Xi_ 1 U {wi} if vi E Xi- 1 . Let X be the
set obtained after taking the final edge in W'.
3. Set K = Uvex\{sf{s --4 V}.
Now take Jo, to be the set of all such K.
Following similar logic as was used to prove Proposition 3.3.16, it can be verified
that this assignment of states of knowledge to vertices of G' satisfies condition 3 of
Definition 3.3.14 and that J,
Jt, =
{{}}. We just need to show Jt,
{{s --+ t}} or
{}.
To show that Jt,
{{s
-+
t}} or Jt, = {}, consider a given walk W' from s' to
t' in G' and look at how the set {Xi} changes as we go along W'. Let Y be the set
of vertices we have when we first reach the vertex t which was the t' of the ith copy
of G'. Yo
=
{s}. Since G' is sound, if t 0 Y then the portion of W' from t to t+
must have at least one edge which crosses the cut between Y and V(G) \ Y. If ek is
the first edge on this portion of W' crossing this cut, then Y 9 Xk_1
C
X
C
Yj+l-
Thus either t E Y C Y+1 or Y C Y+1. There are only n vertices except for s and t
so this implies that t E Y,+1 . Thus for all K c Jt', s -+ t c K. Using the same logic
as before, Jt,
{{s -÷ t}} or Jt,= {}, as needed.
56
Reduction to certain knowledge switching networks
3.3.5
Finally, we prove a theorem that shows that in some sense, monotone switching
networks can be reduced to certain-knowledge switching networks.
Although this
theorem is not strong enough to prove any lower size bounds, the reduction used in
this theorem is deep and will play a crucial role in Section 3.5.
nectivity together with a knowledge description and a path P' = {s' -÷ v',v'
,...
,v
, 2
v,
-+
1 ,v, 1
-+
t'}
-
Definition 3.3.18. Given a sound monotone switching network G' for directed con-
from s' to t' in G', define the certain-knowledge
switching network H'(G', P') as follows:
{{}}
First, if we do not already have that J8, =
{{}}
and Jt,
k E [0, l'], Jv, = { K, 1 ,.--
sets K,,
,
K
.
{{s
=
, Kv,}
t}}. Now let v= s' and let v',
=
J,,
=
: k c [0,l'], S
{{}}
t'. For each
for some positive integer mk and some knowledge
For each non-empty subset S of [1, Mk] let Kvs
We take V(H'(G'P')) = {
knowledge set Ks.
-+
and J,=
C
{{s
=
UjEsK,
.
take J,=
and Jt, = {{s -+ t}}, then
[1,mk], S f 0} where each w',Ss has
-+ t}} so we take s' = w,
and
in H'(G', P'). We take all possible edges which are allowed by condition 3
t'= w',/
of Definition 3.2.8.
Theorem 3.3.19. If G' is a sound monotone switching network for directed connectivity with a given knowledge description and P'
V/,_),V/,_1
{s' -+ v',v'
-+
v,--- ,v-
2
-+
_ t'} is a path from s' to t' in G', then it is possible to take a subset of
the edges of H'(G', P') and assign a direction to each edge to obtain a directed graph
Hried(G', F') for which the following is true:
1. He(G', P') consists of a directed path from s' to t' and directed cycles.
2. Every vertex in H'ed(G', P') is on a path or cycle.
3. For all vertices w',
(a) If w/,/s
74
where |SI is odd,
s' then the incoming edge for w'IS has the same label as the
edge from v_ 1 to v' in P' and its other endpoint is either of the form
57
where |T| =Sj or the form w's where S2 is obtained by adding
W',/
or deleting one element from S.
#
(b) If w',,
t' then the outgoing edge for w's has the same label as the
edge from v' to v'k
in P' and its other endpoint is either of the form
1
where |T| = |S| or the form w's where S2 is obtained by adding
w'it
or deleting one element from S.
4.
For all vertices w's where
|S|
is even,
(a) If wis 4 t' then the incoming edge for w't s has the same label as the
edge from v' to v'k
1
in P' and its other endpoint is either of the form
where S2 is obtained by adding
where |T\ = \S| or the form wits2
kS
w'Vk+lT
or deleting one element from S.
(b) If w'i' s f s' then the outgoing edge for w', s has the same label as the
edge from v-_
to v' in P' and its other endpoint is either of the form
where S 2 is obtained by adding
where |T| = |S| or the form w',
w'/
or deleting one element from S.
Proof. For all k, letting ek be the label of the edge from v' to v'+1 we apply Lemma
and Jv,k+1 U {ek}. This gives us a set Iki
3.3.10 to the states of knowledge Jv' U {ek}
k
[1, Mk],
a set Ik2
C
[1, mk+1] of equal size to Iki, a function fkl
function fk2 : [1, mk+1] \ 1 k2
k2,
-4
[1, imk] \ Ik1 -4 Iki,
and a perfect matching #k :ki
Ik2
1
C
a
such that
1. For all i C [1, nk] \ 41, Kv'f,,(i) U {ek} < Kv,'i U {ek}
2. For all j C [1, mnk+1] \
3. For all i C
'ki,
12, Kv'+lfk 2 (j) U
Kvj U {ek} - K
{ek}
10k(i)
K
U {ek}
U {ek}.
Proposition 3.3.20.
1. For all S C
Ik1,
Kv, s U {ek} = K
2. For all S C [1, m 1 ] and i
{ek} and if fkl(i)
c
4
(s) U {ek}
S\Ik1, if fkl(i)
C S then Kv's U
{ek}
58
S then Kv SU{ek}
Kv'(s\{fk1(i)}) U {ek}
KV (Su{fk1(i)})U
3. For all T C [1,m 2 ] and j e T \ Ik2, if fk2(j)
Kv/+i(TUffk 2 (j)}) Ufek}
and if fk2(j) E T then Kvi
T then K,,
TU
{ek}
T U {ek}
Kv'(T\{fk
2 (j)})U
{ek}
We now choose the edges of H ed(G', P') and assign directions to them as follows. For
each vertex w', s,
then take the edge with label ek between w', and w'V
to~ ~ ~ ~ ~
~I W" IfIIi vnte aetis
then have this
S|
is
even
to
w'
(.If
edge
go
from
w',
odd then have this
1. If S C
Ikl
to /s
tos)
edge go from w,
2. If S Z Iki then take the first i E S \
between w',s and w'/ s
and SA{fkl(i)}
=
(i)})
'kl
and take the edge with label
where SA{fkl(i)} = S U {fi(i)} if
S \ {fki(i)} if fkl(i)
fkl(i) 0
c S. Have this edge go from
S
w', S to
W/(sa{fkl(i)}) if ISI is odd and have this edge go from w'(sa{fkl(i)}) to w' S
ISI
ek
if
is even.
For each vertex w',Vk+1 T'
1. If T C Ik2 then take the edge with label ek between w',iT and w', _q. If ITI
to w'Vk~lT*. If ITI is even then have
is odd then have this edge go from w', _q(T
Vko-'(T)
this edge go from w',
+1T
to WI,
_
VkO-
2. If T Z Ik2 then take the first
WT
TA{fk
to w'"
j2
(T)'
E T\Ik2 and take the edge with label ek between
and we',(TA{fk 2 (j)}) where TA{fk 2 (j)}
2
(j)} = T\{fk 2 (j)} if fk2(j)
T U {fk2(j)} if fk2(j)
c T. Have this edge go from
T and
w',i(TA{fk 2 (j)})
odd and have this edge go from w', (TAIfA (j)}1) to w v'
if ITI is
Vk+ITvk+
2
1
IT
if
ITI is even.
Conditions 3 and 4 of Theorem 3.3.19 are now satisfied by the edges we have chosen.
All vertices have indegree one except for s' and all vertices have outdegree one except
for t'. This implies that Hred(G', P') consists of a path from s' to t' and directed
cycles and that every vertex is on a path or cycle, as needed.
59
D
c'
b/
:S)a
Kj}
K(b
Kfb)
K{c}
d
K{
K
t')
cd
K{,I
K1
K{f}'
Kja}
K{!}>
S
s-a
a
Kt')
Kjc,d}:
K{,}
b'
a-+b
b-*c
c'
c-*d
dl
d-*t
t'
Figure 3-8: This is an illustration of the ideas used in the proof of Theorem 3.3.19.
Above, we have the path P' from s' to t' in G', where the J for each vertex is
given below that vertex with each line corresponding to one of its K. Below, we
have the arrows between all of the knowledge sets from the argument used to prove
correspond to going along a bidirectional
Lemma 3.3.10. Here the functions {0}
edge. The functions {fkl} and {fk2} correspond to going along a unidirectional
edge and then going the opposite direction along a bidirectional edge. To get from
s' to t' in H ed(G', P') we have the following sequence (not shown): K.1 11} = {},
Ka,{i} = {s -+ a}, Ka'{1,2} = {s -+ a, s - b}, Ka'{2} = {s - b}, Kb'{1} = {s -+ b},
Kb/{1,2} = {s
=
{s
b, s
{s
-+
-+ c}, Ka'{2,3} = {s
a, s -+ c}, Ka'{3}
-+ c, s -+ d}, Kt'{1} = {s
C}, Ka'{1,3} =
Kd{1}i
4
b, s -4 c}, Ka'{1,2,3} = {s -
-
c}, Kb'{2}
= {s-
= {S -+
a, s
-+
b, s
-+
-+
c},
c}, Kc'{ 1} = {s
t}.
Corollary 3.3.21. If G' is a sound monotone switching network for directed connectivity with a given knowledge description and P = {s
-4
v1 , v1
-+
v 2 ,... , v1
1
-+
t}
is a path from s to t in G, then any path P' in G' from s' to t' using only the
edges of P must pass through at least one vertex a' such that Ja' /
Jt' and if
Ja' = { Ka'i, - - - , Kaim} then V = UT V(Ka'i) contains at least [lg 11 of v 1 , , 1 _Proof. This follows immediately from Theorem 3.3.19 and Lemma 3.2.32.
3.4
El
Fourier Analysis and Invariants on Monotone Switiching Networks For Directed Connectivity
Unfortunately, Theorem 3.3.19 is not sufficient to prove a superpolynomial lower size
bound on monotone switching networks solving directed connectivity.
60
To prove a
good lower size bound, more sophisticated techniques are needed. In this section,
we introduce a very different way of analyzing the problem: Fourier analysis and
invariants.
While these techniques are less direct, they allow us to give lower size bounds on
all monotone switching network for directed connectivity, not just certain knowledge
switching networks. In this section, we use Fourier analysis and invariants to prove
a quadratic lower size bound and show how more general lower size bounds can be
obtained.
3.4.1
Function descriptions of sound monotone switching networks
The following tautology is trivial yet illuminating:
For any yes/no question, the
answer is yes if and only if it is not no.
Before, we analyzed each vertex of the switching network in terms of how much
progress has been made towards showing directly that the answer to the question is
yes. Here, we will analyze each vertex of the switching network in terms of which
NO instances have been eliminated. For monotone switching networks, we only need
to consider maximal NO instances, as once these have been eliminated all other NO
instances must have been eliminated as well. For directed connectivity, the maximal
NO instances correspond to cuts. Thus, we will analyze each vertex of the switching
network in terms of which cuts have been crossed. We make this rigorous below.
Definition 3.4.1. We define an s-t cut (below we use cut for short) of V(G) to be
a partition of V(G) into subsets L(C), R(C) such that s e L(C) and t E R(C). We
say an edge v1 -s v 2 crosses C if vi C L(C) and v 2 c R(C). Let C denote the set of
all cuts C of V(G).
Definition 3.4.2. We define a function description of a monotone switching network
to be an assignment of a function hv to each vertex v' E V(G') such that
1. Each h', is a function from C to {0, 1}.
61
2. VC c C,s'(C) =1 and t'(C)
=
0.
3. If there is an edge e' e G' with label e between vertices v' and w' in G', for all
C E C such that e does not cross C, v'(C) = w'(C).
For convenience we identify each vertex v' with its associatedfunction h', i.e. we take
v'(C) = h',(C) for all C E C
Proposition 3.4.3. Any monotone switching network which has a function description is sound.
Proof. Assume that G' has a function description yet accepts some input graph G
which does not have a path from s to t. Then there is some path P' in G' from s' to
t' whose labels are all in E(G). Now let C be the cut such that L(C) = {v E V(G) :
there is a path from s to v in G}. Note that E(G) cannot have any edge crossing C
as otherwise there would be a path in G from s to some vertex in R(C). This implies
that for any two adjacent vertices v' and w' in P', v'(C) = w'(C). But then we must
have that s'(C) = t'(C), contradicting the fact that s'(C) = 1 and t'(C) = 0.
E
We now show the converse to this proposition, that every sound monotone switching network has a function description.
Definition 3.4.4. For a cut C, define the input graph G(C) to be the graph with
vertex set V(G) and edge set
E(G(C)) ={e : e does not cross C}
Definition 3.4.5. Define the reachabilityfunction descriptionfor a sound monotone
switching network G' to be the assignment of the function hv, to each vertex v' C V(G')
where
hV, : C -+ {0, 1}
hv,(C) = 1 if there is a path from s' to v' in G' whose edge labels are all in E(G(C))
and 0 otherwise.
Proposition 3.4.6. For any sound monotone switching network G', the reachability
function description is a function description of G'.
62
Proof. Consider the reachability function description for G'. For all C E C, s'(C)
=
1.
Assume that t'(C) = 1 for some C E C. If so, there must be a path P' in G' from s'
to t' such that no edge label in P' crosses C. If so, then since G(C) contains all edges
which do not cross C, all edge labels in P' are contained in E(G(C)) so G' accepts
G(C) and is thus not sound. Contradiction. Thus, t'(C) = 0 for all C.
To see that the third condition for a knowledge description holds, assume that it
does not hold. Then there is an edge e' in G' with endpoints v' and w' and a cut
C such that the label e of e' does not cross C but v'(C)
#
w'(C). Without loss of
generality, v'(C) = 1 and w'(C) = 0. But then there is a path P' from s' to v' such
that none of the labels of its edges cross C. w' is not on P' as otherwise we would
have that w'(C) = 1. But then taking P2 to be the path P' with the edge e' added
at the end, P2 is a path from s' to w' such that none of the labels of the edges of W2
cross C, so we should have w'(C)
=
D
1. Contradiction.
Remark 3.4.7. Reversibility is crucial here. If we defined a similar reachability
function description for a switching-and-rectifier network H and had an edge with
label e from vertex v to vertex w in H where e does not cross some cut C, we could
have v(C) = 0 but w(C) = 1 if there is a no path from s to v in H whose edges are
all in G(C) but there is a path from s to w in H whose edges are all in G(C).
a
s t
b
C1
t C3 C4
C2
C3
C4
Figure 3-9: Here we show how to represent all of the cuts of V(G) simultaneously
when V(G) = {s, a, b, t}. The column determines whether a is with s or t and the
row determines whether b is with s or t.
63
S-*t
K1.1
s
8 at
s 1
'1
a
b
a
-t
0
0
KK.,bb
b
b8b
t
1
1
8 1 0
t 0 0
Kjbj
Sa
tt
1t
bS 0 0
t 0 0
t
Figure 3-10: This is the certain-knowledge switching network G'({s, a, b, t}, 2) together with its certain-knowledge description and the reachability function description
for it.
Fourier analysis
3.4.2
The space of functions from C to R is a vector space which we will use Fourier analysis
to analyze. We have a dot product, a Fourier basis, and Fourier coefficients as follows:
Definition 3.4.8. Given two functions f, g : C -4 R, f . g = 2-n
cec f (C)g(C)
Proposition 3.4.9. If G' is a sound monotone switching network for directed connectivity with a given function description, then for any vertex v' E G', ||v'|| = Vv' -v' <
1.
Definition 3.4.10. Given a set of vertices V C Vre(G), define ev : C -+ R by
ev(C) = (-l)|VnL(C)I
Proposition 3.4.11. The set {ev, V
C
Vred(G)} is an orthonormal basis for the
vector space of functions from C to R.
Definition 3.4.12. Given a function f : C -+ R and a set of vertices V C Vred(G),
define
fv
= f - ev.
Proposition 3.4.13 (Fourier Decomposition and Parseval's Theorem). For any function f : C
-+
R, f = XVCVred(G) fvev and fV f =
64
VVred(G)fV
t
S -4
s
a
a
b
sat
t
s
t
s
s
-e{)
e{}
-
(efa + etb} + e{,b})
a-*t
ba
sb
t
a
sa+tb--
at
1(e(
-
etb})
t
(e(
-
t
e(.1 - e{b} + eaa
)
0
Figure 3-11: In this figure, we have the monotone switching network solving directed
connectivity on V(G) = {s, a, b, t} shown in Figure 3-1 together with the reachability
function description for it and the Fourier decomposition of each function.
Remark 3.4.14. The switching network shown in Figures 3-1 and 3-11 determines
whether or not there is a path from s to t by checking all of the cuts one by one
to determine if there is an edge in G crossing that cut. This can be generalized to
give a monotone switching network of size 2' solving directed connectivity. While the
size of such a switching network is enormous, this extreme construction is interesting
because the switching network is analyzing all of the possible paths from s to t at the
same time.
3.4.3
Invariants and a quadratic lower bound
We now introduce functions and invariants which we will use to analyze sound monotone switching networks.
Definition 3.4.15. Given a switching network G', a directed path P' from s' to t' in
G', and a set of edges E, recall that pu' is the labeling function for edges in G' and
define
AE(P)
(nd
(FI
e)
-
vtart(P', e))
e'EE(P'):p'(e')E
where v'tart(P', e') and v'nd(P', e') are the vertices in G' such that e' goes from V'tat(P', e')
to V'end(P', e') in P'.
Lemma 3.4.16. Let G' be a sound monotone switching network for directed connectivity on V(G), let G be an input graph on the set of vertices V(G), let (E1 , E2 ) be a
65
partition of E(G), and let P' be a directed path from s' to t' in G' whose edge labels
are all in E(G).
-
1. If C E C is a cut which cannot be crossed by any edge in E1 , then (A E1 (P')
AE 2 (P))(C)
-
1.
-
2. If C E C is a cut which cannot be crossed by any edge in E2, then (ZAE 1 (P')
AE 2 (P))(C)
=
-1.
Proof. For any cut C,
(ZAE 1 (P) + AE 2 (P')) (C)
-1.
To see this, note that since
all edge labels of P are in E1 U E2,
(AE 1 (P') + AE2
(P'))(C)
e')
vend (P',
=
- V'tart(P', e'))(C) =
(t
- s')(C) =
-1
e'EE(P')
Now if C E C is a cut which cannot be crossed by any edge in El, then AE1 (P') (C)
0 so
(AE 1 (P')
A (P', E2 ))(C) = -(AEI(P') + AE 2
-
(P'))(C)
1
Similarly, if C E C is a cut which cannot be crossed by any edge in E 2 , then
AE 2 (P')(C)
=
0 So
(A E1 (P)
-
AE 2
(P))(C)
(AEI(P')
AE 2 (P))(C)
We now use the fourier analysis and invariant approach to prove a quadratic lower
bound using a linear independence argument.
Theorem 3.4.17. Let V(G) be a set of vertices with distinguishedvertices s, t. Taking
n
= Vred(G)|, m(PV(G),<2) =
n + 2 and m(PV(G), 3)
Proof. The upper bounds on m(PV(G),
2)
=
(n) +
and m(PV(G),
3)
n + 2.
follow immediately from
Theorem 3.2.23. We prove the lower bounds using the following proposition:
Proposition 3.4.18. IV(G')l > dim(span{V(G')}) + 1
66
Proof. t' = 0 so IV(G')l = IV(G') \ {t'} + 1 > dim(span{V(G') \ {t'}}) + 1 =
E
dim(span{V(G')}) + 1
If P is a path of length 2 in G from s to t, then P has the form P = s
Take E= {s
-
v -- t.
v} and E2 = {v -+ t}. For any cut C,
-
1. If v C L(C) then C cannot be crossed by any edge in E1 so by Lemma 3.4.16,
(AE 1 (P')
- EE
2 (P))
1
(C)
2. If v E R(C) then C cannot be crossed by any edge in
(AE, (P')
This implies that AE 1
AE 2 (P')) (C)
-
AE 2 (P)
(P') -
=-
2
so by Lemma 3.4.16,
1
Note that AE 1 (P/)
-v}.
-
AE 2 (P') is
a linear combination of vertices of G' so if G' accepts all inputs in PV(G),2 then
e{v} E span{V(G')} for all v C Vred(G).
e{} = s' E span{V(G')} as well so by
Proposition 3.4.18, IV(G') \ {s', t'}I > n + 1 + 1
=
n + 2.
If P is a path of length 3 in G from s to t, then P has the form P = s v2
t. Take Ej={s -v
1
,v 2
v,
-- t} and E2 = {vi --+ v 2}. For any cut C,
1. If v1 ,v2 E L(C) then C cannot be crossed by any edge in E2 so by Lemma
3.4.16,
(AE, (P)
-AE
-1
2 (PT))(C)
2. If vi E L(C), v 2 c R(C) then C cannot be crossed by any edge in El so by
Lemma 3.4.16,
(ZAE 1 (P) -
AE 2 (PT))(C)
=
1
3. If v 2 E L(C), v1 E R(C) then C cannot be crossed by any edge in E2 so by
Lemma 3.4.16,
(AE 1 (P)
-
IAE 2 (PT))(C)
=
1
67
4. If v 1 , v 2 E R(C) then C cannot be crossed by any edge in E2 so by Lemma
3.4.16,
(L\E 1 (P')
-
AE 2 (PT(C)
=-
By direct computation, this implies that (AE 1 (P
AE 2 (e
V1}+e{V2
e{ 1,V 2 }). Thus, we have that if G' accepts all inputs in PV(G),<; 3 , since we already have
that ell = t' E span{V(G')} and e{v} E span{V(G')} for all v E Vred(G), we also
have that for all
E Vred(G), e{V 1 ,V 2} E span{V(G')}.
Vi, V 2
By Proposition 3.4.18,
IV(G')I > (n) + n + 1 + 1 = (n) + n + 2, as needed
(2 a -+ t
s at
11
s at
e{)
-
s at
00
0
E(P) = {s -+ a, a -+t}
El = {s -+ a}, E2 = {a-+t}
s at
-
P/s
A(P', E1 ) -
b -+t
Ssat
bs
b
b
e)
A(P', E2) = -eta}
a-+ b
s-*a
P/
E]
(e{} -
e{.}) 1(e{
-
sats
0bS
e{a) -
E(P) = {s -+ a, a -+ b, b -+ t}
E= {s -+ a, b-+ t},E2 ={a -+b}
at
0
efb} + e{a,b})
bs
s at
1 1
A(P', E1 ) - A(P', E2) = !(-e}
- e{a) + e{b) - e(.,bl)
Figure 3-12: This figure demonstrates the reasoning used to prove Theorem 3.4.17.
Edges whose label is in E1 are colored blue and edges whose label is in E2 are colored
red.
3.4.4
General lower bounds
Unfortunately, the linear independence argument breaks down for longer paths. The
problem is that for paths P of length greater than 3, we can no longer find a non-trivial
partition (E1 , E2 ) of the edges of P such that AE1 (P') and AE2 (P') are invariant over
all sound monotone switching networks G' and paths P' from s' to t' whose edge
labels are all in P. Thus, for longer paths we need a more sophisticated approach.
68
For this approach, we partition the edges of E(G) into several sets {E} and look
at the dot product of vertices v' E V(G') with a carefully chosen set of functions
These functions are chosen so that for all i,
gG,Ei
-s' = 1 and whenever there
is an edge between vertices v' and w' in G' with label e E Ej,
V'
gG,Ei
= WI
9G,E
-
{gG,Ei}.
We now imagine the following game. There are several players, one for each set of
vertices E. At each vertex v', the ith player has value gG,Ei -v'.The players are trying
to go from all having value 1 at s' to all having value 0 at t' in a sound monotone
switching network G' while only taking edges in G' whose labels are in E(G). While
doing this, they are trying to keep their values as close to each other as possible.
However, since every edge the players take has label in Ej for some i, for any
given move there will be some player whose value remains fixed. This means that
their values cannot all change at the same time so there will be some point where
there is a significant discrepency between their values. This corresponds to a vertex
v' and i, j such that v' - (gG,Ej
9G,Ei)
-
is non-negligible, which we can use to prove
our lower bounds. We make this intuition rigorous below.
Definition 3.4.19.
1. We say a function g : C -- R is e-invariant for some pos-
sible edge e if g(C) = 0 for any cut C which can be crossed by e.
C --+ R is E-invariant for a set of edges E if g is e-
2. We say a function g
invariantfor every e
c
E.
Proposition 3.4.20. If P' is a path from s' to t' in G' and E is a set of edges, then
if g is an E-invariantfunction, zAE(P')
g
=
0
Theorem 3.4.21. Let I = {Gj} be a set of input graphs on V(G) all of which
contain a path from s to t and let M be a constant. If for each j we have a partition (E1 , ... , Ekj) of the edges of Gj, functions gGj,E 1 '' ,gGj,Eq,
constants {zj} and { Mj} such that
1. For all j, YgGj,Ei 9 is Eij -invariantfor i C {1, qj}
2. For all j and all i E [1, q]
,
gGj,Eij - e{} = Zj
69
> 0
C -
R, and
#
(gGjE 1i3
4. For all i, j, ||gGj,Eij
j2 and all i 1 ,i 2
,
3. For all J1, j2 where j1
gGjEj ) * (YGj 2 ,Ei 2 j 2
-
|
- YGj,E1j
-
gG 32 ,E1
0
2
<; M
Then
m(I)2
j2
Proof.
Lemma 3.4.22. Let G be an input graph containing a path from s to t. If we have
a partition (E1 ,--- , Eq) of the edges of G and functions gG,Ei such that gG,Ei
is
Ei-
invariant for all i then for any sound monotone switching network G', for any path
P' in G' from s' to t' whose edge labels are all in E(G),
q
AE(G)\E (P')
(G,Ei
-
(q
9G,E 1 )
-
2)(gG,E 1 -e})
9G,Ei
-
i=2
e{}
i=2
Proof. Let P' be a walk from s' to t' in G' whose edge labels are all in E(G). Since
9G,Ei
is Ei-invariant,
Vi, AE(G)\Ei(P)
9G,Ei
=
AE(G)(P)
gG,Ei =
9G,Ei (t'
-
s') = -gG,Ei
e{}
Since gG,E 1 is E1 -invariant,
k
q
S
AE(G)\EjP)
gG,E1
= ((q - 2) 5AE,(Y) + (q
i=2
=
((q
=
(q
-
'AE1
' gG,E 1
i=2
- 2)
-
AE(P))
2)(gG,E 1
(t'
-
g G,E 1
s')) =
-(q
- 2 )(gG,E 1 '
e{})
Putting all of these equations together gives the needed equality.
70
Dl
Corollary 3.4.23. Let G be an input graph containing a path from s to t. If we
have a partition (E1 ,--- , Eq) of the edges of G and functions gG,Ej such that gG,Ei
Zs
Ei-invariantfor all i and YG,Ei e{} is the same for all i, then for any sound monotone
switching network G' which accepts G, taking z = gG,E1 - e{},
E
I|V' - (gG,Ei - gG,E 1 ) >Z
i=2 V'eV(P')
In particular, there must be some i E [2, q] such that
E
v'EV(G')
|v'
(gG,Ei -g G,E1 )
>
q-i
Proof. This follows immediately from Lemma 3.4.22 and the fact that for all i,
|v' - (gG,E,
-
9G,E 1) I
AE(G)\Ei(F') - (G,Ei
-
9G,E1
)
E
v'EV(P')
because AE(G)\Ej (P') is a linear combination of the vertices in P' where each vertex
has coefficient -1,
0, or 1..
We now prove Theorem 3.4.21 using Corollary 3.4.23, an orthogonality argument,
and the Cauchy-Schwarz inequality.
Proposition 3.4.24. If {gj} is a collection of nonzero orthogonal functions from C
to R, then for any function h : C -+ R where
Ih|
=
h h < 1,
Z
(g
1 h) 2
< 1
Proof. If {gj} is a collection of nonzero orthogonal functions, we can extend it to
an orthogonal basis {gj} U {f } for the vector space of functions from C to R. Now
h=E
gf+f , so
1 > h - h = J7jjl9 2 +
i.h
Z(fh)
as needed.
>,,
Now let G' be a sound monotone switching network which accepts all of the inputs
in I = {C3 }. By Corollary 3.4.23, Vj, ]i :Zv'V(G') (gG,EZ
Using the Cauchy Schwarz inequality (EZ, f(v')g(v'))
71
2
< E,
- gGEi 3 )
' > J
f(v')
2
E,
g(v')
2
with
f
(v') = 1 and g(v')
=
(gG,,Ejj
gGj,Elj) - v',
-
((gGj,Eij -
Vj,
we have that
gGj,E 1 j)
')
((qj-1Z,)2
2
v'cV(G')
This implies that
z((gG,Eij
-
i
v'cV(G')
H9GJEi
2
YGj,E 1 ) v')
-gG 3 ,E
(z
q(G-1
>
--
2
2V
)2
(G')|
However, by Proposition 3.4.24,
zz((gqGj,Ej
G,E 1
I G,Eij- - gGj,Elj)
j
v'CV(G')
1 r2 2
v)
HGJEiJ~-YGj,E1
v'<V(G')
(
Putting these inequalities together, IV(G')I >
Z3
)2
2
,U1(G)
so
IV(G') I>
M3 i
as needed.
3.4.5
0
Conditions for a good set of functions
The simplest way to use Theorem 3.4.21 is to take one input graph G, find a set of
functions {gG,E} and then obtain the other input graphs and sets of functions by
symmetry. We now give conditions which are sufficient to ensure that we can do this
and deduce that if such sets of functions exist for paths P of arbitrary length then any
monotone switching network solving directed connectivity must have superpolynomial
size.
Theorem 3.4.25. Let V(G) = {s, t, V 1i, - , Vm}.
If there is a partition E1 , - - - , Eq
of the edges of G, functions {gG,Ej}, a value z > 0, a value M, and a value r K m
such that:
1. gG,Ei is E-rinvariantfor i E [1, q]
2. For all i G [1, q] and all V C Vred(G) with |V| < r, (gG,Ei
3.
gG,E1
e{} = Z
72
-
gG,E 1 )
- eV
=
0
4. For all i e [1, q], |gG,E - 9G,E
1
< M
then for all n > 2m 2 , if W is a set of vertices such that V (G) 9 W and |W\{s, t}I = n
then letting H be the input graph with V(H) = W and E(H) = E(G), m(H) >
2
(q-1l)M
Proof. We first show that we can add additional isolated vertices to the input graph G
while still keeping the same functions (expressed in terms of their Fourier coefficients).
Proposition 3.4.26. For any U, V C
symmetric difference function, i.e. VAU
Vred(G),
=
euev = evAu where A is the set-
(U U V) \ (U n V)
Proposition 3.4.27. For all v, w E Ved(G), for all C E C,
1. (e{} + e{ W)(C)
=
2 if w c R(C) and 0 if w E L(C).
2. (e{} - e{v)(C) = 2 if v E L(C) and 0 if v E R(C).
3. ((e{} - e{fv)(e{} + e{w))(C) = 4 if v G L(C) and w G R(C) and 0 otherwise.
Corollary 3.4.28.
+
1. If e = s -+ w for some w E Vred(G) then g is e-invariant if and only if (e{}
efw})g = 0. Equivalently, g is e-invariant if and only if .vu{w} = -- $v whenever
w c V.
t for some v E Vred(G) then g is e-invariant if and only if (e{}
-
-
2. If e = v
e{v})g = 0. Equivalently, g is e-invariant if and only if .vu{v}
V.
3. If e = v -+ w for some v, w
E Vred(G)
then g is e-invariant if and only if (e{}
-
v
- =v whenever
e{v})(e{} + e{w})g = 0. Equivalently, g is e-invariant if and only if jvu{v,w}
-jvu{v}
+
Yvu{w}
We now write
+ jv whenever v, w $ V.
ZVCVed(G)
9G,E
graph H and take gH,Ej
=
=
VCVred(G)
civev. By Corollary 3.4.28 if we have the input
CiveV
then all conditions of Theorem 3.4.25 are
still satisfied by {gH,Ei}. Moreover, for all i and all V Z Vred(G), gH,Ei ' ev = 0.
We now take our set of input graphs I = {Hj} so that
73
1. Each H is obtained from H by applying some permutation oa to the vertices
W \ {s, t}.
2. For all distinct ji and j2,
Qj1
(Vred(G))
n ogi (Vred(G)) < r
By Corollary 3.2.36, we can take at least (,)r such graphs.
Proposition 3.4.29. If we take Eij = oj(Ei ) and gHj,Eij
Vrd(G)
=
CiVeoj(v) then
1. For all j, gHj,Eij is Ejj-invariantfor i c [1, q]
2. For all j and all i G [1,q]
3. For all i, j, (gGj,Eij
4.
For all i, j, I|gGj,Eij
-
,
e{} = z
gGj,Ei
gG,,E 1 )
0 whenever VI <
' ev
- gGj,E1IH
r or V
Z Jj(Vred(G))
M
Proof. This follows immediately from the properties of the functions {gH,Ej} and the
1:
fact that for all i, j and all V, gHj,E, - eo. (y) = 9H,E, - ey
Now note that since (gHj,E- gHj,Eij)-eV= 0 whenever IVI < r or V
and for all distinct ji and
il, i2, J, j2
where j
1
j2,
i(Vred(G))
n Uj(Vrecj(G))
Z
O (Vred(G))
< r, we have thatfor all
J2,
(gHj1,Eijj
--
YHjl ,Ei3 )
9H 2
(g2,Ei2j2
,E132 )
0
Applying Corollary 3.4.21,
z
)
m(H) > m(I) >
( Z 1)2
ZM
(q
2
-
1)M 2m
D
as needed.
...
-+ vi_ 1
-+
1,
t} and let P be the path s -+ v1
-
Corollary 3.4.30. Take V(P) = {s,v1, - - - , vi
t. If n > 2(1- 1)2 and we can find a partition {E 1 , -- - , Eq} of the edges
of P, functions {gPEj}, values z, M, and a value r < 1 such that:
1. gP,E is Ei-invariantfor i G [1, q]
74
2. (gP,
-
gPE1 )
3. gPE1 e{}
-ev = 0 for all i C [1,q] and all V C Vred(G) with |VI < r
=Z>
4. For all i, |YgP,E
0
M
~- 9P,E1
'
then m(PV(G),l) > (q-i~)M 2(1-1)
El
Proof. This follows immediately from Theorem 3.4.25.
Example 3.4.31. For 1 = 2 and r
=
1 we can take P = s
--
v 1 }, E2
-
e{ 1}), and
(e{
9P,E2 -
=
{v 1
-+
t}, gP,E1
(e{}
=
{s
-+
+ e{v 1 }) This gives
gPE1 = -ev 1 . Using Proposition3.4.28 it can be verified directly that YP,E
E-invariantfor i E {1, 2}. 1lgp
for all n > 2, m(P,1) >
{s
=
1 and z = gP,E1 - e{} =
v 1 , v 2 -+ t}, E 2
=
3 and r
{vI
-+
=
so by Theorem 3.4.25,
2 we can take P = s
v 2 }, gPE2
=
4(e{}
-
v 1 --
V2
t,
-+
e{V 1} I e{V 2 } -F 3e{ v,V 2 }), and
-
e{- 1 } - e{v 2 } ~ e{i,V 2 }) . This gives 9P,E2
,-
Zs
j
Example 3.4.32. For 1
E=
=
YP,E 2
v, -+ t, El =
-
9P,
=
-e{V
1
,V2}.
Using
Proposition 3.4.28 it can be verified directly that gp,E is El-invariantfor i c {1, 2}.
gp
=
1 and z
9PE1 - e{} =
4 so by
Theorem 3.4.25, for all n > 8, mPf,2) >
8
Remark 3.4.33. These bounds are around the square root of the bounds obtained
from the linear independence argument. This square root comes from the Cauchy-
Schwarz inequality and so far we have not found a way to avoid having this square
root. Nevertheless, getting a lower bound for m(PV(G)) which is around m(PV(G))c
for some c > 0 is good enough for our purposes and unlike the linear independence
argument, we can use these techniques for longer paths.
3.5
A Superpolynomial Lower Bound
While the Fourier analysis and invariant approach of Section 3.4 is powerful, we need
to actually find suitable functions {gpEI}. There are several possibilities for how we
could do this. One possibility is to look directly at the values gPE (C) for all C E C.
75
However, this gives us very little control over the Fourier coefficients of each gPEi. A
second possibility is to work directly with the Fourier coefficients of each gp,E. This
approach is viable, but it would involve analyzing how to satisfy many equations for
Ei-invariance simultaneously. Here we take a third approach. We look at the dot
products of each 9PE, with the vertices of the certain knowledge switching network
It turns out that all of the conditions of Corollary 3.4.30 correspond to
G'(V(G)).
simple conditions on these dot products. Furthermore, we have complete freedom in
choosing these dot products, which enables us to construct suitable {gp,E}, and thus
prove the following theorem.
Theorem 3.5.1. For all 1 > 2, if we have V(G) = {s,t,v1,--- ,v 1 1} and let P be
v1 ---
...- 4 v1_1 -+t
path
l| ands taking the, partition
then takingthe
r = [lg
Ej =
{v-_
vi} of the edges of E (where vo = s and v, = t), we can find functions
{gP,EjI such that:
1. 9P,E, is Ei-invariantfor all i
2. (gP,E
-
9PE1 )
ev for all i and all V
C
Vred(G) with
|VI < r
3. YP,E l e{} > 0
Combined with Corollary 3.4.30, this immediately proves the following theorem,
which implies a superpolynomial lower bound on monotone switching networks solving directed connectivity on n vertices.
Theorem 3.5.2. For any integer 1 > 2, there is a constant ci such that if V(G) is a set
of vertices with distinguishedvertices s, t and n =|Vred(G)
3.5.1
then m(Pv(G),l) > cin
FIg 11
2
From certain knowledge decriptions and knowledge descriptions to function descriptions
For the proof of Theorem 3.5.1, we adapt results from Sections 3.2 and 3.3 to the
Fourier analysis and invariants approach.
We begin by taking certain knowledge
descriptions and knowledge descriptions and giving a corresponding function description.
76
Definition 3.5.3. For a given knowledge set K, define the function
K :C
{O, 1}
K(C) = 1 if there is no edge in K which crosses C and 0 otherwise.
Proposition 3.5.4. If we can get from K1 to K 2 in the certain knowledge game using
only the knowledge that some edge e is in G and e does not cross some cut C then
K2(C) = K1(C).
Proof. This follows immediately from the fact that if e does not cross C, then for any
knowledge set K, no individual move on K in the certain knowledge game which can
be done with only the knowledge that e is in G changes the value of K(C).
F]
Corollary 3.5.5. If a monotone switching network G' has a certain-knowledge description {K,,} where each vertex v' is assigned the knowledge set Kv, then if we
assign each vertex v' the function Kv,, this is a function description of G'.
Definition 3.5.6. For a given state of knowledge J, define the function
J: C
{0, 1}
J(C)
0 if there is no K in J such that K(C) = 1 and 1 otherwise.
Proposition 3.5.7. If we can get from J1 to J2 in the knowledge game for directed
connectivity using only the knowledge that some edge e is in G and e does not cross
some cut C then J2 (C) = J1 (C).
Proof. This follows immediately from the fact that if e does not cross C, then for any
state of knowledge J, no individual move on J in the knowledge game for directed
connectivity which can be done with only the knowledge that e is in G changes the
l
value of J(C).
Corollary 3.5.8. If a monotone switching network G' has a knowledge description
where each vertex v' is assigned the state of knowledge Jv, then if we assign each v'
the function Jr', we have a function description of G'.
Remark 3.5.9. If we take the knowledge description given in the proof of Proposition
3.3.16 and take the correspondingfunction description we will obtain the reachability
function description.
77
3.5.2
A criterion for E-invariance
Now that we have translated certain knolwedge descriptions and knowledge descriptions into function descriptions, we give and prove the following criterion for Einvariance.
Theorem 3.5.10. If g is a function from C to R and E is a set of edges between
vertices in V(G) then g is E-invariant if and only if g - v' = g - v' whenever v', v'
are vertices of G'(V(G)) such there is an edge between v' and v' in G'(V(G)) whose
edge label is in E.
Before proving Theorem 3.5.10, we need a few preliminary results.
Lemma 3.5.11. If g is a function from C to R and E is a set of edges between
vertices in V(G) then g is E-invariant if and only if g -J1 = g -J2 whenever J1 , J2 are
states of knowledge such that J1 U {e} = J2 U {e} for some e E E and all knowledge
sets in J1 and J2 are either equal to Kt, or have the form Kv where V C Vred(G).
Proof. The only if direction follows immediately from Proposition 3.5.7. To prove the
if direction, assume that g - J1 = g - J2 whenever J1 , J2 are states of knowledge for
V(G) such that all knowledge sets in J1 and J2 are either equal to Kt, or have the
form Ky where V
C
Vred(G) and there is an e e E for which it is possible to go from
J1 to J2 in the knowledge game for directed connectivity using only the knowledge
that e is in G. Given a cut C which can be crossed by an edge e c E, if R(C) # 0,
take J, = {KL(c)} and J 2 = UvR(C){KL(c)u{v}}. If R(C) = 0, take J=
J2
= Jt'. We have that J2 (C)
=
{KL(c)},
0, Ji(C) = 1, all knowledge sets in J1 and J2 are
either equal to Kt, or have the form Ky where V C Vre(G), and J U {e} = J2 U {e}.
Now consider any cut C2 C C.
If L(C) n R(C2 ) is nonempty then Ji(C2 )
J2 (C2 ) = 0. If R(C2 ) C R(C) then J1 (C2 )
Ji(C2 )
=
J2 (C2 ).
J 2 (C2 )
together, 0 = g - J2- g -
-
=
J2 (C 2 )
-
.
By our assumption, g - J, = g - J2
= 1. Thus, if C2 # C then
J1 (C2 ) # 0 if and only if C2 = C. Putting everything
=
24(J2 (C) - Ji(C))g(C) so g(C) = 0. Thus, g(C) = 0
for any C which can be crossed by an edge e c E, as needed.
78
E
Lemma 3.5.12. Let J1 = {K 11 ),.
{e}
-
J2 U
, Kim1 }
and let J2
=
{K2 1 ,- -
,
K2m 2 }- If J1 U
{e} for some possible edge e then we may write J2 - J1 as a sum of terms of
the form K 2 -K
1
where K1U{e} = K 2 U{e} and both K1 and K2 are either of the form
U3-sK1j where S C [1,mi], S # 0 or the form
UkGETK2k
where T C [1,m 2 ],T
# 0.
Proof. We first give a proposition which allows us to express J2 - J in terms of these
knowledge sets.
Proposition 3.5.13. If J = {K 1 , K 2 , J(C) =
E
,
Km} where m # 0, then for any C E C,
(-1)Is l+((UicsKi)(C))
Proof. This is just the inclusion-exclusion principle. Note that J(C) = 0 if Ki(C) = 0
for every i and 1 otherwise. If Ki(C) = 1 for some i, then we can add or remove i
from S without affecting (UiesKi)(C).
But then all terms in the sum on the right
cancel except Ki(C), which is 1.
If Ki(C) = 0 for all i, then for all non-empty subsets S of [1, m], (UiesKi)(C) = 0.
Choosing an arbitrary i, we can add or remove i from S without affecting (UicsKi)(C),
so we again have that everything cancels except Ki(C), which is 0.
E
Lemma 3.5.12 now follows directly from Theorem 3.3.19. We can easily create a
sound monotone switching G' which has a path P' from s' to t' such that there are
vertices v', V+
vo+
1
1
oni P' with
J,ij+1= J1 and J,
i+1
= J2 and there is an edge e' from v' to
with label e. By Proposition 3.5.13 we have that
J2 - J1=
(-1)Il+1((UjeTK 2j)(C))
TC[1,m 21,
(-1)sl+1((uicsKii)(C))
SC[1,m1],
S/4J
TO0
By Theorem 3.3.19, if we let Ee, be the set of directed edges corresponding to e' in
Hr' (G', P'),7
A(e')
ke /EE.
=
E
(--1)'S+1((UiSsKi)(C
(-1)+l(UJETK 2 )(C)) SC[,m1],
TC[1,M2]
Soo
T=AO
79
where if e' goes from w' to w' in Hrd(G', P') then A(e')
Thus, J2
-
J=
w-
w.
ZCE., A(ek) and the result follows.
e2
1e
,
I
V3
e'
/q
(G
' S'
s a K eb, 0
t2
8Kb~
I
et
-*
/
e
e, }
V/I
V/
Hj
K{} e2
V1
P1
Figure 3-13: In this figure, we illustrate the ideas used in the proof of Lemma 3.5.12.
It can be verified that A(e'1) = A(e'4) + A(e' 0 ), A(e) = A(e's) + A(e'7) + A(e'g), and
e(ee)
A(e') =
+ Z(e').
With these results, we can now prove Theorem 3.5.10.
Theorem 3.5.10. If g is a function from C to
and E is a set of edges between
vertices in V(G) then g is F-invariant if and only if g -
=
g -v; whenever vi, v;
are vertices of G'e(V(G)) such there is an edge between vi and v6 in G'e(V (G)) whose
edge label is in F.
Proof. The only if direction follows immediately from Proposition 3.5.4. To prove
the if direction, assume that g(C)
0 for some cut C which can be crossed by an
edge e E E. By Proposition 3.5.11, there exist states of knowledge J1 , J2 for V(G)
such that J U {e}
2
U 1J
{e} and all knowledge sets in J1 and J2 are either equal
to K or have the form K where V e Vre(G), and g- J1
Lemma 3.5.12, we may write J2
K1 U {e}
g- J2 . But then by
as
fJi a sum of terms of the form K 2
-
K 1 where
K2 U {e} and both K 1 and K 2 are either Ke' or of the form Ky where
V G Vred(G). Since g -(J2
that g- (K2
-
#
-
K1)
-
Ji) # 0, there must be at least one such pair K1 , K 2 such
0. But then taking vi and v to be the corresponding vertices
in G'e(V(G)), there is an edge with label e between v and v and g v($g - v, as
needed. T
80
3.5.3
Choosing Fourier coefficients via dot products
In this subsection, we prove the following theorem relating the Fouerier coefficients
{v} of a function g to the dot products of g with the functions {Kv}
Theorem 3.5.14. For any set of values {av : V C
function g : C -+ R such that for all V C
r, v = 0 for all V such that
|V| <
Vred(G),
Vred(G)},
g -Kv
there is a unique
av. Furthermore, for any
r if and only if av =g - Kv = 0 for all V such
that |V| < r.
Proof.
Proposition 3.5.15. For any V
Vred(G), Ky =
2v
EZ
v (-1)Iuleu
Proof. Note that Kv(C) = 1 if V C L(C) and 0 otherwise. Now for all cuts C E C,
E(-1)Ue (C) = I E (-1)IUI(-1)UnL(C)I
I
S
Ucv
Ucv
If V Z L(C) then all terms will cancel so
2v
EUv (-1)UWeu(C) = 0. If V C L(C)
then
(-1)IUI(-1)j nL(C)1
Ucy
This implies that Kv
1
Ucy
EUCV (-1)UIeu, as needed.
C
Corollary 3.5.16.
1. For all V C
Vred(G),
ev - Kv
# 0
2. For all subsets U, V of Vred(G), if U g V then eu - Kv = 0.
To see the first part of Theorem 3.5.14, pick an ordering {V} of the subsets
V C Vred(G) such that if i <
j
then V/
(t V4. We now pick the Fourier coefficients
.v, in increasing order of i. By statement 2 of Corollary 3.5.16, for all subsets U, V
of Vred(G), if U t V then eu Kv = 0. This means that for each i, once we pick
this determines the value of g Kv
=
av, as for any
j
> i, V9
v,,
Vi so ev, -KV, = 0. By
statement 1 of Corollary 3.5.16, for all i, ev, -Kv, $ 0. This means that we always have
a unique choice for each coefficient
v, which gives g - Kv, = av. Putting everything
81
together, there is a unique function g : C --+ R such that for all V C Vred(G),
g - Ky = av, as needed.
Now we just need to show that if av
v
#
=
0 for all V such that IVI < r then
0 for all V such that IVI < r. To see this, assume it is false. Take a minimal
=
subset V of Vred(G) such that v
v
g -Kv
#
0 and au = g - Ku = 0 for all U C V. Then
0, U = 0 for all U C V, and g - KU = 0 for all U C V. However, by Corollary
3.5.16, if v f 0 and 'u = 0 for all U C V then g- Kv
3.5.4
#
0. Contradiction.
D
Proof of Theorem 3.5.1
We are now ready to construct the functions
{gP,Ei}
and prove Theorem 3.5.1, which
we recall below for convenience.
Theorem 3.5.1. For all 1 > 2, if we have V(G) = {s, t, v 1 ,-- ,vi
1
} and let P be
the path s -
v - --
Ei = {vi_
vi} of the edges of E (where vo = s and v, = t), we can find functions
1
-
->
vi_1
-+ t then taking r =lg 1] and taking the partition
{gP,Ei} such that:
1. gPEi is Ei-invariantfor all i
2. (gPEi - 9P,E1 ) ey for all i and all V C Vred(G) with
|VI
< r
3. gPE1 e{} > 0
Proof. Note that by Theorem 3.5.10 and Theorem 3.5.14, the three conditions of
Theorem 3.5.1 are equivalent to the following three conditions:
1. For all i, gP,E 'U
9PEi vVfor
any vertices u', v' of G' (V(G)) which have an
edge between them whose label is ei.
2. gPEi KV
=
gPE1 - Kv for all i and V C Vred(G) with IVI < r = [lg l
3. gP,E1 ' KfI > 0
By Theorem 3.5.14, we can choose the values {gp,Ei - Ky : V C Vred(G)} freely, so it
is sufficient to give a function b' : V(G'(V(G))) x {[1, l]} -4 R such that
82
1. If there is an edge between vertices u' and v' whose label is in E then b'(u', i)=
b'(v', i)
2. b'(v', i) = b'(v', 1) for all i whenever K E {KV : V C
Vred(G),
V| < r} U {Kt'}
3. b'(s', 1) = 1 and b'(t', 1) = 0
We choose the values {b'(v', i)} by looking at connected components of certain graphs
which are closely related to V(G'(V(G)). Let H' be the graph with
1. V(H')
{v' E V(G'(V(G))) : v' has knowledge set K c {K
: V C Ved(G),
IVI < r} U {K}}
2. E(H') = {(U', v') : u', v' E V(H'), there is an edge between u' and v' whose
label is in E(P)}
For each i let Hj be the graph with
1. V(Hj) = V(G'(V(G)))
2. E(Hl) = E(H) U {e' E E (G' (V(G))) :up'(e')
E}
Proposition 3.5.17. If u',v' E V(H') and u' and v' are in the same connected
component of Hj for some i then u' and v' are in the same connected component of
H'.
Proof. Assume that we have u' and v' which are in different components of H' but
are in the same component of Hj for some i. If so, choose u' and v' to minimize
the length of the shortest path in Hj from u' to v'. Note that there cannot be any
W' E V(H') on this path, as otherwise w' is either in a different component of H'
than u' in which case we could have taken the shorter path from u' to w' instead or
w' is in a different component of H' than v' in which case we could have taken the
shorter path from w' to v' instead. Thus all edges of the path between u' and v' in
Hj are not edges of H' and thus must have label ej. But then since G'(V(G)) has all
allowable edges, there must be an edge between u' and v' with label ej so u' and v'
are actually in the same connected component of H'. Contradiction.
83
El
This proposition implies that we may first choose any set of values {b'(v')} such that
b'(u') = b'(v') whenever u' and v' are in the same connected component of H' and then
choose any set of values {b'(v', i)} such that if v' E V(H') then b'(v', i)
all i and b'(u', i)
=
b'(v') for
b'(v', i) whenever u' and v' are in the same connected component
of Hj. One way to do this is to take b'(v', i)
=
b'(u') whenever v' is in the same
connected component of Hi as u' for some u' E V(H') and take b'(v', i)
v' is not in the same connected component as any u'
E V(H').
=
0 whenever
This is guaranteed to
satisfy the first two conditions of Theorem 3.5.1.
For the third condition, we need to check that s' and t' are in different connected
components of H' so that we may choose b'(s') = 1 and b'(t') = 0. To check this,
assume that s' and t' are in the same connected component of H'. Then there is a path
from s' to t' in H'. However, this is impossible by Lemma 3.2.32. Contradiction.
D
Remark 3.5.18. In choosing the values b'(v') for v' E H', we are essentially picking
the Fourier coefficients v : V < r for a "base function" g which we then extend to
an Ei-invariantfunction gi for every i. The crucial idea is that if we only look at the
Fourier coefficients (gi)v for |VI < r, all of the gi look identical to g and thus look
identical to each other.
3.6
An ONQ(") lower size bound
In this section, we prove an nn(lgn) lower size bound on monotone switching networks solving directed connectivity by explicitly finding the functions {gPEj
given
by Theorem 3.5.1 and then modifying them by "cutting off" high Fourier coefficients.
Theorem 3.6.1.
m(Pv(G),l) >
n
1)2)
64(l -_1
1 1 n_3
-
m(PV(G)) >
_n16
FlT
2
4
Proof. The first step in proving this lower bound is to gain a better understanding of
the functions given by Theorem 3.5.1.
84
Definition 3.6.2. For all V C
\ {s})
1. gv(C) = 0 if (L(C)
2. gv(C)
Vred(G),
define the function gv : C -> R so that
V
if (L(C) \ {s}) C V
= 2 n(--)IV\L(C)l
The most important property of these functions is as follows.
Lemma 3.6.3. If V1 , V2 C Vred(G), Kv1 - gV2
=
1 if V
=
V2 and 0 otherwise.
Proof. We have that Ky(C) = 1 if V C L(C) and Kv(C) = 0 otherwise. Now note
that
Ky, - gV2
=
(-1)IV2\L(C)l
CEC:V C(L(C)\{s})CV2
Ths implies that K, -gv2 = 1 if V = V2 and 0 otherwise, as needed.
C
We can now construct the functions {gpE} in terms of the functions {gv} and
analyze their Fourier coefficients.
Lemma 3.6.4. If {gp,I are the functions given by Theorem 3.5.1 then we have that
1.
gPPE
2.
YP,Ei
=VCVred(G) b(V,
-ev =
EVc
i)gv
b(U, i)(gu . ev) = Euv b(U, i)(-2)UJ
Proof. The first statement follows from the definition of b(V, i) and Lemma 3.6.3. For
the second statement, we use the following proposition.
Proposition 3.6.5. For all U, V C Vred(G), gu - ev = (- 2 )uI if U C V and 0
otherwise
Proof.
gu * ey
(-1)U\L(C) (_,IlvnL(C)l
=
CEC:(L(C)\{s})CV
If there is some i E U \ V, then shifting i from L(C) to R(C) or vice versa changes
(-i)U\L(C)(_))VnL(C)
by a factor of -1.
Thus, everything cancels and we have
gU * ey =
Es (-1)U\L(C)(-1)IvnL(C)
CeC:(L(C)\{s})9U
85
-
gu ev = 0. If U C V then
E
CEC:(L(C)\{s})CU
(_1)IUI
= (- 2 )IUI
F
The completes the proof of Lemma 3.6.4
El
Corollary 3.6.6. If we take each b(V, i) to be 0 or 1 when choosing the functions
} then for all V C
|P,Ei - ev
Vred(G),
< 2 2v
If we take the functions {gP,Ei} given by Theorem 3.5.1 directly, then
YP,E 1
YgPE
-
{gP,Ei
may be very large. The key observation is that as shown below using Corollary
3.4.28, we can cut off all of the Fourier coefficients ev where IVI > r
Lemma 3.6.7. Taking r = [lg
11,
[lg 1].
there exist functions gi such that
1. For all i, gi is Ei-invariant.
2. For all i and all V such that |V| < r, (gi - gi) -ev =0
3. g 1 - ell = 1
||gi - gjiH < (1 - I)F191--1 2 41g 1+2
4. For all i,
Proof. We repeat Corollary 3.4.28 here for convenience.
Corollary 3.4.28.
=
s
-*
w for some w E Vred(G) then g is e-invariant if and only if (e{}
+
1. If e
elw})g = 0. Equivalently, g is e-invariant if and only if
.vu{w}
= -jv whenever
w V V.
-
2. If e = v -+ t for some v E Vred(G) then g is e-invariant if and only if (e{}
efv})g = 0. Equivalently, g is e-invariant if and only if
.vu{v}
= .v whenever
v V V.
3. If e = v -+ w for some v, W G Vred(G) then g is e-invariant if and only if
(e{} - e{ })(e{} + e{w})g
jvu{v,w}
= -gvu{v}
+
=
0. Equivalently, g is e-invariant if and only if
Yvu{w}
+ .v whenever v, w $ V.
Definition 3.6.8. Define the functions {gi} so that
86
1. gi eV = gPE - eVifV < r
2. gi - ev = 0 if |V| > r
3. If i = 1 (so that E= {s
-+
v1}), and |VI =r then
(a) If vi E V then g- ev = -gi - ev\{v1 } if v1 E V
(b) If v1 V V then g -ev = 0
4. If i
= 1 (so that E = {v1 _ 1
(a) If vi_
1
C V then gi
(b) If vi_1
(a) If vi_ 1 , vi
(c) If vi
c
t}) and |V| = r then
ev = gi - ev\{fvl}
V then gi ev = 0
5. If i V {1, l} (so Eji
(b) If vi
-
{vi_1 -+ vi}) and |V|
e V then gi - ev =
=
r then
ev\{lv,,vi} - gi
V and vi_ 1 $ V then gi ev = -gi
ev\{Vp} + gi -v\{Vj
1}
ev\{v,}
V then gi - ev = 0
Proposition 3.6.9. gi is Ej invariantfor all i.
Proof. We can show that gi is Ej invariant using Corollary 3.4.28. When vi_ 1 , vi
and IV U {vi_ 1, vi} \ {s, t}
VV
;> r we can check directly that the associated equation
in Corollary 3.4.28 holds. When vi- 1 , vi V V and IV U {vi_ 1 , vi} \ {s, t}I < r we use
the fact that the associated equation in Corollary 3.4.28 must hold for gp,E and thus
holds for gi as well.
E
Corollary 3.6.10. The functions {gi} have the following properties:
1. For all i, gi is Es-invariant.
2. For all i and all V C Vred(G) where
|V|
< r, gi - ev = 91 - ev
3. For all i, gi ev
=
0 whenever V C Vred(G) and |V| > r
4. For all i, gi - e{}
=
1
87
5. For all i, gi , ev
6. |gi - ev| < 3-
#
0 for at most
2 2 [g 1--2
<
22[ 1g 1
V with |VI = r
for all V with |V| = r
Proof. The first statment is just Proposition 3.6.9. The second, third, and fourth
statements all follow from the definition of the functions {gj} and the properties of
the functions {gP,}Ei. For the fifth statement, note that every time we fix a nonzero
Fourier coefficient gi - ev where IVI = [ig l1 we use Fourier coefficients of the form
gi ew where IWI < r to determine its value. Moreover, we never use the same Fourier
coefficient twice, so the number of nonzero Fourier coefficients gi -ev where IVI = r is
at most (-_).
Finally, the sixth statement follows from the definition of gi and our
bounds on the Fourier coefficients of the functions {9P,Ei}-
Note that when we look at gi2 - g 1 , all Fourier coefficients with V < r cancel.
From this, it follows that for any i, IIg-gi
2
< 21gi 12 +21g11
2
(1_ I)r-124r+2.
D
We now prove Theorem 3.6.1. By Corollary 3.4.30, taking M = (1 - 1)r2122r+1
-
1)2,
1n
'M(PV(G),1)>)2>51+(
- M(l -
n2
1) 2(l - 1)
2 5 2+ 1 (
-
-
+.
)
for all n > 2(1
Recalling that r = [lg 11 and using the fact that 2r > l, we may reexpress this bound
as
I 1l1
[lgl]
m(Pv(G),l) >
>
' 26g%+1(/_iii
-
1)
64(1 -1)2
- 2
Note that we may ignore the condition that n > 2(1
because the bound is trivial
-1)2
if n < 2(l - 1)2. Taking 1 = [inl], we have that
M (PV(G)
1
>
1Ign-3
2
--
2
as needed.
1
-n
2
ign3
6
1
4
E
88
3.7
Monotone analogues of L and NL
We would love to say directly that monotone L does not equal monotone NL. Unfortunately, as discussed in Gringi and Sipser [9], while it is relatively straightforward
to define monotone NL, it is not at all clear how to define monotone L. In this
section, we note that whether or not we have separated monotone L from monotone
NL depends on how we define monotone L.
We define the following uniform and non-uniform monotone analogues of NL,
which are the same as the ones given by Gringi and Sipser
[91.
Definition 3.7.1. A nondeterministic Turing machine T is monotone if whenever T
is examining an input bit xi and can make a given transition when xi = 0, it can also
make that transition when xi = 1.
Definition 3.7.2. Define mNSPACE(g(n)) to be the class of all problems solvable
by a monotone nondeterministic Turing machine in O(g(n)) space.
Theorem 3.7.3. If f
E
mNSPACE(g(n)) where g(n) is at least logarithmic in
n, then there are monotone switching-and-rectifiernetworks of size at most
2 0(9(n))
computing f.
Proof. This can be proved by applying the exact same construction we used to prove
Theorem 2.3.3 and noting that if the initial nondeterministic Turing machine T is
monotone then so is the resulting switching-and-rectifier network.
E
By Theorem 3.7.3 and the fact that the directed connectivity problem is in mNL, we
can define a non-uniform monotone analogue of NL as follows:
Definition 3.7.4. Define mNLswitching-and-rectifiernetworks to be
{f
: {0, 1}*
-
{0, 1} : f can be computed by monotone switching-and-rectifiernet-
works of polynomial size}
Unfortunately, we do not know how to define a nice uniform monotone analogue
of L. However, we can mimic Definition 3.7.4 and define a non-uniform monotone
analogue of L as follows:
89
Definition 3.7.5. Define mLswitching networks to be
{ f : {0,
1}*
--
{0, 1} : f can be computed by monotone switching networks of polyno-
mial size}
Theorem 3.7.6. mLswitching networks
C mNLswitching-and-rectifier networks
Proof. This follows immediately from Theorem 3.6.1 and the fact that the directed
connectivity problem is in mNL.
D
However, Gringi and Sipser [9] define a different non-uniform monotone analogue
of L. They define a monotone analogue of L as follows:
Definition 3.7.7. A boolean circuit is monotone if it has no NOT gates.
Definition 3.7.8. A boolean circuit is leveled if we can partition the gates into levels
such that the outputs of the gates at each level depend only on the inputs and the
outputs of the gates at the previous level. Define the width of a leveled boolean circuit
to be the maximum number of gates in a level.
Definition 3.7.9. Define mLcircuits to be
{f : {0, 1}* -+ {0, 1} : f can be computed by a monotone leveled circuit of polynomial
size and logarithmic width}
As far as we know, our analysis does not imply any lower bounds on the minimum
width and size of monotone leveled circuits solving the directed connectivity problem,
so we are currently unable to prove that mLcircuits $ mNLswitching-and-rectifier
networks-
Thus, although we have separated monotone analgoues of L and NL, we cannot say
directly that monotone L is not equal to monotone NL.
3.8
Finding paths with non-monotone switching networks
Unfortunately, proving lower size bounds on all switching networks solving the directed connectivity problem is much harder than proving lower size bounds on monotone switching networks solving the directed connectivity problem. Non-monotone
90
switching networks can use the information that edges are not there in the input
graph, which can be very powerful.
In this section, we note that there are small
sound non-monotone switching networks for directed connectivity on n vertices which
accept all of the inputs in PV(G), so the bound of Theorem 3.6.1 does not hold for
non-monotone switching networks.
Definition 3.8.1. Given a set I of input graphs on a set V(G) of vertices with
distinguished vertices s, t where each graph in I contains a path from s to t, let s(I)
be the size of the smallest sound switching network for directed connectivity on V(G)
which accepts all of the input graphs in I.
Theorem 3.8.2. For all n, if V(G) is a set of vertices with distinguished vertices s, t
and |Vred(G)J = n then s(PV(G)) < n _+2.
Proof. The intuitive idea is as follows. If the input graph consists of just a path from
s to t, it is easy to find this path; we just have to follow it. If we are at some vertex
v, and see that there is an edge from v, to v 2 and no other edges going out from vi,
then we can move to v 2 and we can forget about v, because the only place to go from
v 1 is v 2. We can thus easily follow a path as long there are no branches in the path.
While doing this, we only need to remember around 3 ig n bits of information. We
need to remember what v, and v 2 are and we need to remember how many other
possible edges going out from v, we have confirmed are not in G. We now give a
rigorous proof:
Definition 3.8.3. Define Gpathfinder(V(G)) to be the non-monotone switching net-
work for directed connectivity on V(G) constructed as follows:
1. Start with G'(V(G), 2)
2. For each pair of distinct ordered vertices v 1 , V 2
n - 1 between v',1, and v
V/
and vj'
1 .
G Vred(G),
add a path of length
in parallel to the edge labeled v1
Give the edges in this path the labels {,(v 2
{ s, vi, v 2 }}
91
-
--
U)
v 2 between
: u C V(G)\
n3 + 2
Proof. There are n vertices of the form v, where v c Vred(G).
there are n - 1 added paths which start at v'f
vertices v'
adds n -2 vertices. Thus, IV(Gathfinde,(V(G)))
V(G' (V(G), 2))l
n+2=n 3
= n(n-1 +n+2
+n+2<n
_n2
3
-
For each of these
and each of these paths
V(G' (V(G), 2))I = n(n-1)(n-2).
so IV(Gpathfinder(V(G)))= n(n -1)(n--2)+n(n21)
+2asneeded.
+
Proposition 3.8.4. 1V(Gathfinde,(V(G)))
E
Proposition 3.8.5. G'athfinder(V(G)) accepts all input graphs in PV(G).
Proof. If G is an input graph with vertex set V(G) and edges E(G)
i
=
{vi
-*
vi+1
E [0, 1 - 1]} where vo = s and v, = t then we have a path from s' to t' whose edges
are all consistent with G as follows.
2. If we are at v
vi
-
for any i E [1, 1- 2] then go to o
-+
v1
along the edge labeled
vi+1.
3. If we are at v
vi+1
along the edge labeled s
.
1. If we are at s' then go to v'
-e
U
for any i C [1, 1 - 2], then for all u c V(G)
\
{s, vi, vi+}
E(G). Thus we can go to vjov 1 } along the path between vvg and
Vt
{vi,vi+1}.
4. If we are at v'
then go to t' along the edge labeled v 1 _1
-
t
Lemma 3.8.6. G'pthfinder(V(G)) is sound.
Proof.
Definition 3.8.7. Given an input graph G, create an input graph Ga as follows. Let
Ea
=
{v
-*
Take V(Ga)
w : V,W E Vred(G),V
=
w,Vu c V(G)\{s,v,w},w -÷ u $ E(G)}.
V(G), E(Ga) = E(G) u Ea-
Proposition 3.8.8. If Grathfinder(V(G)) accepts an input graph G then G'(V(G), 2)
accepts the corresponding input graph Ga.
92
Using Proposition 3.8.8, to prove Lemma 3.8.6 it is sufficient to show that for any
input graph G there is a path from s to t in Ga only if there is a path from s to t
in G. To show this, assume that there is no path from s to t for some input graph
G. Then let V be the set of all vertices v such that there is a path from v to t in G.
Since there is no path from s to t in G, s V V. Let C be the cut with R(C) = V.
Note that there cannot be any edge in G that crosses C.
Assume there is an edge in Ga which crosses C. Then it must be an edge v
-
w in
Ea and we must have v E L(C), w E R(C). This implies that there is a path from w
to t. However, by the definition of Ea, w y t and Vu E V(G)\{s, v, w}, (w, u) V E(G).
Thus, any path from w to t must go through v, so there must be a path from v to
t and we should have that v c R(C). Contradiction. There is no edge in Ga which
crosses C, so there is no path from s to t in Ga, as needed.
D
Theorem 3.8.2 follows immediately from Proposition3.8.4, Proposition 3.8.5, and
D
Lemma 3.8.6.
Example 3.8.9. The switching network in Figure 3-4 is G'athfinder({st,a,b}) with
some edges removed.
93
94
Chapter 4
Lower bounds on monotone switching
networks for generation and k-clique
The material in this section comes from joint work between Siu Man Chan and the
author [5] on applying the techniques developed in Chapter 3 to other problems,
in particular the generation problem and the k-clique problem. This work gives an
alternate proof of the monotone circuit depth lower bounds and separation of the
monotone NC-hierarchy proved by Raz and McKenzie.
4.1
Circuit Depth and Switching Network Size
We begin by showing the connection between circuit depth and switching network
size mentioned in the introduction and Chapter 2.
Theorem 4.1.1. If a function f can be computed by a (monotone) circuit of depth
d, then it can be computed by a (monotone) switching network of size at most
2d
+ 1.
Proof. This is proved using Shannon's [24], [25] construction of a series parallel switching network from a boolean formula/circuit. This construction is recursive and works
as follows.
1. In the base case where the circuit is a single input bit b or -b, take the switching
network G' to have the vertices s' and t' and a single edge with label b = 1 or
95
b = 0 respectively between s' and t'.
2. If the final gate of the circuit is an OR gate, take the two switching networks
G' and G' correspnding to the two inputs of this gate. Now merge the s' of
G' and G' together and let this be the new s'. Also merge the t' of G' and
G' together, letting this be the new t'. This effectively places G' and G' in
parallel.
3. If the final gate of the circuit is an AND gate, take the two switching networks
G'1 and G'2 correspnding to the two inputs of this gate. Take s' to be the s' of
G1, merge the t' of G' with the s' of G', and take t' to be the t' of G'. This
effectively places G' and G'2 in series.
4. If the final gate of the circuit is a NOT gate, if there are two NOT gates in
a row then just delete both of them. Otherwise, use De Morgan's laws that
-,(A V B) = -,A A -B and -,(A A B) = -A V -B to reduce to one of the two
cases above.
Now we just need to check the size of the resulting switching network. We check this
using induction. For the base case, IV(G')I = 2 = 20 + 1. For the other cases, using
the inductive hypothesis, IV(G')j < IV(G')l + IV(G') - 1 <
2d
+ 1, as needed.
E
Theorem 4.1.2. If a function f can be computed by a (monotone) switching network
of size n', then it can be computed by a (monotone) circuit of depth O((lgn') 2 ).
Proof. This is proved using Savitch's algorithm. Savitch's algorithm checks whether
there is a path of length at most 1 from s to t as follows. If 1 = 1 then just check
whether there is an edge from s to t in G. Otherwise, for each possible midpoint m,
recursively check whether there is a path of length at most
path of length at most
[j]
[]
from s to m and a
from m to t. If the answer is yes for any midpoint m,
accept. Otherwise, reject.
Savitch's algorithm was used implicitly in the switching networks G'(V(G), [lg 1])
from Section 3.2. Savitch's algorithm can be implemented with circuits as follows.
For each m, recursively create the circuit checking for a path of length at most []
96
from s to m and the circuit checking for a path of length at most [12] from m to t. Put
an AND gate between these two outputs. Now take the OR of all of the ooutputs of
these circuits. This can be done with [lg n] additional levels, where n is the number
of vertices of the graph.
This means that each recursive level adds depth at most [ign] + 1 to the circuit.
There are at most [ig n] levels of recursion in Savitch's algorithm, so the total circuit
depth is at most ([lg n'])([lg n'] + 1).
Now given any problem which can be solved by a (monotone) switching network
of size n', take the circuit corresponding to Savitch's algorithm on this switching
network and this will be a (monotone) circuit of depth O((lg n') 2 ).
4.2
E
Our Results and Comparison with Previous Work
Raz and McKenzie [19] proved the following lower depth bounds on monotone circuits
(the first bound was first proved by Karchmer and Wigderson [12]).
Theorem 4.2.1.
1. Any monotone circuit solving undirected connectivity has depth Q((lgn) 2 ).
2. For some e > 0, any monotone circuit solving the generation problem on pyrmamids of height h < nE has depth Q(hlgn).
3. For some E > 0, any monotone circuit solving the k-clique problem where k < nE
has depth Q(k lg n)
Using Theorem 4.1.2, this implies the following lower size bounds on monotone
switching networks.
Corollary 4.2.2.
1. Any monotone switching network solving undirected connectiivty has size n"N'.
2. For some e > 0, any monotone circuit solving the generation problem on pyrmamids of height h < nE has size n"
97
1n.
3. For some e > 0, any monotone circuit solving the k-clique problem where k < n'
has size n" Vg.
In this chapter, we prove the following bounds.
Theorem 4.2.3.
1. Any monotone switching network solving directed connectiivty has size n"(n).
2. For some e > 0, any monotone circuit solving the generation problem on pyrmamids of height h < nE has size nQ(h).
3. For some c > 0, any monotone circuit solving the k-clique problem where k
<
nE
has size nQ(k).
Using Theorem 4.1.1, this implies the following depth lower bounds on monotone
circuits.
Corollary 4.2.4.
1. Any monotone circuit solving directed connectivity has depth Q((lgn) 2 ).
2. For some c > 0, any monotone circuit solving the generation problem on pyrmamids of height h < nE has depth Q(h lg n).
3. For some e > 0, any monotone circuit solving the k-clique problem where k < n
has depth Q(klgn)
Thus, our bounds for directed connectivity, generation, and k-clique are a strengthening of the results of Raz and McKenzie. For the undirected connectivity problem,
the bound of Karchmer, Wigderson, Raz, and McKenzie is tight for both monotone
circuit depth and monotone switching network size as undirected connectivity can
be solved by a monotone circuit of depth O((lg n)2 ) and trivially by a monotone
switching network of size O(n). This example shows that Theorem 4.1.2 cannot be
strengthened and that bounds on monotone circuit depth cannot always be strengthened to a corresponding bound on monotone switching network size. This in turn
98
implies that considerably different ideas are used in the communication complexity
approach for proving lower bounds on monotone circuit depth and our approach for
proving lower size bounds on monotone switching networks.
4.3
Lower bounds on monotone switching networks
for Generation
The generation problem is a generalization of the directed connectivity problem.
Definition 4.3.1. In the generation problem on a set of vertices V(G) with distinguished vertices s,t, we have an axiom s and a set of implications of the form
V =-
w where V C V(G)
\ {t}
and w E V(G)
\ (V U {s}).
We have a YES instance
if we can deduce t from the axiom s and we have a NO instance otherwise.
The YES instances we will be considering for this problem are pyramid instances.
Definition 4.3.2. A pyramid P of height h consists of vertices {vij : i
C [0,
h],j E
[0, h-i]} where vij has children vi_ 1,j and vi-1,j+1. We call VhO the root of the pyramid.
Definition 4.3.3. Given a pyramid P where V(P) C Vred(G), we define the YES
instance G(P) so that
1. V(G(P)) =V(G)
2. E(G(P)) = {s =
[0, h - i]} U{vh0
vo0
j E [0, h]} U {{vi-1j, vi-, +}
-- > vij :
[1, h],j E
4 t}
All of the techniques we used in analyzing directed connectivity apply to this
problem with only a few small changes. First, the certain knowledge game becomes a
logical proof system except that we are not allowed to delete information arbitrarily.
The reversible pebbling game changes slightly as follows.
Definition 4.3.4 (Reversible Pebbling Game for Generation). In the reversible pebbling game for generation, we start with a pebble on s and can make the following
types of moves
99
1. If we have an implication V
(V U
-- > w where V C V(G) \ {t} and w E V(G) \
{s}) and we have pebbles on all vertices in V, we can add or remove a
pebble from w.
2. If we have a pebble on t we can add or remove any other pebble.
We win the reversible pebbling game if we have a pebble on t.
Just as the heart of our lower bound for directed connectivity was a lower bound
on the number of pebbles needed to win the reversible pebbling game for a path
graph, the heart of our lower bound here is a lower bound on the number of pebbles
needed to win the pebbling game on a pyramid of hieght h. This bound was proved
by Cook [61 and we present the proof here for convenience.
Lemma 4.3.5. At least h + 1 pebbles are required to win the reversible pebbling game
on a pyramid of height h.
Proof. In order to win the reversible pebbling game, we must reach a state where
every path from the top of the pyramid to the bottom of the pyramid contains a
pebble. Consider the first time we reach such a state. Note that moves where we
pebble or unpebble a vertex when its children are pebbled cannot affect whether there
is a pebble on these paths, so must have just done a move where we pebbled a vertex
at the bottom of the pyramid. We now have that
1. For every path P from the top of the pyramid to the bottom of the pyrmaid,
there is a pebble on one of its vertices.
2. There exists a path P from the top of the pyramid to the bottom of the pyrmaid
which only has a pebble on its bottom vertex
We now show that these conditions require having h + 1 pebbles on the pyrmamid,
which immediately implies the result.
Lemma 4.3.6. Assume we have a pyramid and pebbles on the pyramid such that
1. For every path P from the top of the pyramid to the bottom of the pyrmaid,
there is a pebble on one of its vertices.
100
2. There exists a path P from the top of the pyramid to the bottom of the pyrmaid
which only has a pebble on its bottom vertex.
then we must have at least h + 1 pebbles on the pyramid.
Proof. We prove this lemma by induction. The base case h = 0 is trivial. If h > 0
then let P be a path from the top of the pyramid to the bottom of the pyrmaid
which only has a pebble on its bottom vertex. Without loss of generality we may
assume that P starts at VhO and goes to V(h_1)O. Now note that if we consider the
subpyramid rooted at V'(h1)o, this pyramid has height h - 1 and must satisfy both
conditions given in the lemma, so by the inductive hypothesis there must be at least
h pebbles on this pyramid. Furthermore, if we consider the path P, which goes along
the vertices Vi(h-i), this path must have at least one pebble on it, giving a total of at
1I
least h + 1 pebbles, as needed.
l
This completes the proof of Lemma 4.3.5.
Now if we say that an implication V
=>
w crosses a cut C if V C L(C)
and w c R(C), all of our techniques can be applied in exactly the same way! The
only remaining difference is that the analysis is slightly different when we explicitly
compute the Ei-invariant functions.
Lemma 4.3.7. Taking r = h + 1 and E= {e=} where e 1 ,
, eIE(G(P))l are
the edges
of G(P), there exist functions {gi} such that
1. For all i, gi is Ei-invariant.
2. For all i and all V C Vrd(G) such that |V| < r = h+ 1, (gi - g) -ev = 0
3. gi -ell = 1
4.
For all i, gi
-
gill < h2h24h+8
Proof. Taking the functions {gP,Ej given by the analogue of Theorem 3.5.1 for the
generation problem, when we take all of the b'(V) to be 0 or 1,
1. gP,E is Ei-invariant
101
2. gP,-
3.
e{} -= 1
(gPE --
,E
1
4. |gP,E - ev| <
) -
ev = 0 whenever IVI < r = h + 1
2 21vI
Corollary 3.4.28 must be adjusted as follows.
Proposition 4.3.8.
1. If e = s
>
w for some w G Vred(G) then g is e-
invariant if and only if (e{} + e{f})g = 0. Equivalently, g is e-invariant if and
only if .vu{w}=
2. If e
(e{}
=
-
v
-
v whenever w
$
V.
t for some v C Vred(G) then g is e-invariant if and only if
=>
eqv})g = 0.
Equivalently, g is e-invariant if and only if
vu{v} =
v
whenever v $ V.
3. If e = {u, v}
w for some u, v, w
==>
Vred(G) then g is e-invariant if and
only if
(e{} - e{u})(e} - e{t})(e} + e{w})g = 0
Equivalently, g is e-invariant if and only if
9Vu{uVw}
= -VU{u,v}
whenever u, v, w
+
YVU{u,w}
+ Vu{v,w} +
jvu{u}
V.
Definition 4.3.9. Define the functions {gi} so that
1. gi - ev = gP(G),i'ev if IVI <h + 1
2. gi - ev = 0 if |V| > h + 1
3. If ei is of the form s ==> v and |V| = h+1 then
(a) If v E V then gi - ev = -gi - eV\{V}
(b) If v V V then gi ev = 0
4. If ei is of the form v
=->
t and |V| = h + I then
102
+
9VU{v} - jvu{w} - jV
(a) If v E V then gi ev = gi - ev\{v1_ 1}
(b) If v V V then gi ev = 0
5. If ej is of the form u,v -- >w and |VI = h + 1 then
(a) If U),v,w G V then
(gi)v
=
+ (gi jyfV) + (gi j yx, + (gi y ,
-(giV\{w}
+ (Yi)V\{UUW - (9 )V\u,vl - (g )yg{s,,,,
(b) Ifv,wE V anduVV then
(Yiyv
-(gi)VU{U}\V,W}
(c) If w G V and u, v
(d) If w
-
(9i)V\{W} + (gi)V\{v
+(
V then (gi)v = -- (92)V\{w}
V oruC V butv
V then (g)
=O
Proposition 4.3.10. gi is Ei-invariantfor all i.
Proof. We can show that gi is ei-invariant using Proposition 4.3.8. When u, v, w $ V
and IV U {u, v, w} \ {s, t}| ;> r we can check directly that the associated equation in
Proposition 4.3.8 holds. When u, v, w
V and |V U {u, v, w} \ {s, t}j < r we use
the fact that the associated equation in Corollary 3.4.28 must hold for gP,E, and thus
U
holds for gi as well.
Corollary 4.3.11. The functions {gi} have the following properties:
1. For all i, gi is Ei-invariant.
2. For all i and all V C Vred(G) where V| < h + 1, (Ei)v
3. For all i, (9i)V =0 whenever V
C Vred(G)
and |V| > r=h+1
4. For all i, (9i)f} = 1
5. For all i, (gi)v # 0 for at most ('2)
103
(Y1)y
V with |V| = h +1
6. |(gi)v| < 7.
2
(h+1)-2
<
2 2h+3
for all V with |V| = h + 1
Proof. The first statment is just Proposition 4.3.10. The second, third, and fourth
statements all follow from the definition of the functions {ge} and the properties of
the functions {gp,,}.
For the fifth statement, note that every time we fix a nonzero
Fourier coefficient (gi)v where IVI = h + 1 we use Fourier coefficient(s) of the form
gi - ew where IWI = h to determine its value. Moreover, we never use the same
Fourier coefficient twice, so the number of nonzero Fourier coefficients gi - ev where
VI = h +1 is at most
(l).
Finally, the sixth statement follows from the definition
of gi and our bounds on the Fourier coefficients of the functions {gPE I
Note that when we look at gi-g, all Fourier coefficients with V < r = h+1 cancel.
From this, it follows that for any i, I i - g,112 < 21|g11 2 + 21|g,1i
2
((2 )) 2 4h+8 <
h2h24h+8.
Applying Theorem 3.4.25 with q
M = h 2 2h+4, whenever r ; 2(+2
1
2
Choosing h
4.4
(h+1) + 1, r = h + 1, z = 1, m = (h+1), and
we get a lower bound of
n
1
h
n
(h+1)hh2 2h+4 (h+1)
-1 we get a lower bound of
2
h+1
1)
- 4(h + 1) 8(h
2
Lower bounds on monotone switching networks
for k-clique
We can also use this framework to prove lower bounds on monotone switching networks for the k-clique problem. Although we do not have a certain knowledge or
pebbling game that we can use as a guide, the symmetry of the problem allows us to
give and analyze the sets of functions directly.
For the k-clique problem, we analyze yes instances containing a clique
Q of size
k. Our NO instances are k - 1 colorings C of the vertices Vred(G). Letting C be the
set of all such colorings, we have an inner product and Fourier coefficients as follows:
104
Definition 4.4.1. Given f, g : C -+ C, define (f, g)
Ecc f(C)g(C).
||fil =
(f, f)
Definition 4.4.2. Given a function U :
WZvEVred(G)
C(v)U(V)
where w
=
Vred(G)
[0, k - 2], define eu(C) =
-
ek-i
Proposition 4.4.3. The functions {eu} form an orthonormal basis for the space of
functions from C to C.
Definition 4.4.4. Given a function f : C -+ C and a function U : Vred(G)
+
[0, k-2],
define fu = (f, eu).
Proposition 4.4.5 (Fourier Decomposition and Parseval's Theorem). For any function f : C
C, f
-+
=
E
U:Vred(G)-+[O,k-2]
fueu and (f, f)
=
U 12
ZU:Vre d(G)-+[O,k-2]
Lemma 4.4.6. For any vo G Q there exist functions {g, : v C Q} such that
1. For all v C Q, gv is Ev-invariant where Evv is the set of edges of
Q
incident
with v.
2. For all e and all U:
k -2
Vred(G) -*
[0, k - 1] such that U(v) is nonzero for at most
v E Q, (gv-g,eu) =0
3. For all v c Q, (gv, 1) = 1
4. For all v c Q, ||gv - gvn|| < 4k
Proof.
Definition 4.4.7. Define q(k)
=
(k-
_k-2.
Definition 4.4.8. For every vertex v in a k-clique Q, we define gv(C) to be #(k)(-1)t(t2)! if C restricted to
Q contains a monopoly of size t not
Lemma 4.4.9. (1, gv) = 1
105
containing v and 0 otherwise.
Proof. There are (k - 1) (k- ) -
choices for how to color the vertices of
Q so that
there is a monopoly of size t not containing v. Thus,
1
k - 1) (k -2)!
k-i
1)k
(k
- (k)(--1)'(t - 2)!
(t - 2)!
-
(1, gV) = (k
t=2
1)
0(k) (k - I)! k(1) t ( k
(k - 1
t
t=2
Note that
k-1
k-1
(lI)t (k
(-1)t (k t 1)
t=2
Thus, (1 g)
(2)!(k-2)
=
1 + k - 1 = (1 + - 1 )k-1 + k - 2 = k - 2
t
t=o
1, as needed.
D:
k
Lemma 4.4.10. jjvj < k-2
Proof. Recall that 0(k)
Again,
there are (k
Agiteeae(
(k 1)k(k-2)!(k-2)*
how to color the vertices of
Q so that
-
1)
choices for
_-2
(k-1)
there is a monopoly of size t not containing v
Thus,
=
(k - 1)k (k -k)
(k
-
(k -
1 )1
2) 2
k-
k -1)
t
- 1
(k
t
(t -
-
2)'
)!
-
(ge,Ig9)
(t - 2)!
(k - 2)!
Note that (k-1) (t-2)! < k - I, so this is at most (k - 1)k < kk.
Thus,
|gV1
=
(gv, g) < k , as needed.
Lemma 4.4.11. For all distinct v, w E Q and any function D :
" ECCC:VuCQ\{w},C(u)=D(u)9V(C)
Q \ {w}
-
[0, k - 2],
is 0$(k) if D maps every vertex in Q\{w} to a different
color and 0 otherwise.
Proof. Consider the positions of the vertices in
Q \ {w}.
If some color has a monopoly
of size m > 2 on these vertices, then if w has this color as well, then for this cut C,
ge(C)
= 0(k)(-1)m+l(m - 1)!.
Otherwise, there are m - 1 choices for the color
106
w which give w its own color. For these C, g,(C) = 0(k)(-1)'(m - 2)!. For the
remaining choices for the color of w, at least two colors will have at least two vertices,
so g(C) = 0. Adding everything up, we get 0 as claimed.
If no color has a monopoly on these vertices, all choices for the color of w pair it
with one other vertex, so g,(C) =
E
(k).
Corollary 4.4.12. Whenever U(v)
#
0 for at most k - 2 v E
Q,
for all v E
Q,
(gv)U = (gvo)u.
Proof. Choose some v 2 such that U(v 2 )
such that w
(9V)
U
=
z
v, w / v 2 , and U(w)
#
0. Now for any v E
Q,
there must be a w
0. From the above lemma, this implies that
(gV 2 )u. This holds for any v E Q so (gv)u = (gv2)u = (gv) u
D
Putting everything together, Lemma 4.4.6 follows.
Applying Theorem 3.4.25 with q = k, r = k - 1, z = 1, m = k, and M= 4k,
whenever n > 2k 2 we get a lower bound of
()
k 21)
8(k-1)k-2
Choosing k=
L2
we get a lower bound of
1--)2 >
8(k - 1)
n 2k2
107
-2
4n
4
108
Chapter 5
Upper Bounds: Dealing with
Lollipops via Parity
5.1
Motivation for considering particular input graphs
for directed connectivity
We have two reasons for analyzing the difficulty of particular input graphs for directed
connectivity. First, the problem of what makes an input graph hard for monotone
switching networks/algorithms to solve directed connectivity on is interesting in its
own right. Second, obtaining such bounds is a necessary first step for obtaining nonmonotone bounds. As discussed in Section 3.8, non-monotone switching networks can
easily solve input graphs which consist of an isolated path from s to t and no other
edges, so we cannot use such input graphs to prove non-monotone lower bounds.
Thus, we must consider other input graphs and a necessary first step is to prove
monotone lower bounds for these graphs.
In chapter 6 we will develop the techniques in chapter 3 further to prove such lower
bounds. Here we first prove an upper bound which shows that the minimal length of
a path from s to t in an input graph G is not the only factor which makes it easy or
hard for monotone switching networks/algorithms to solve directed connectivity on
G.
109
Definition 5.1.1.
We call a vertex v E Vred(G) a lollipop if there is an edge from s
to v in G or an edge from v to t in G but not both.
In this chapter, we show that adding lollipops to an input graph G does not significantly increase the difficulty of solving directed connectivity on G using monotone
switching networks.
5.2
Lollipops and certain knowledge switching networks
As a warmup, we show that lollipops can help even certain knowledge switching
networks solve directed connectivity.
Theorem 5.2.1. If G is an input graph such that
1. There is a path P in G from s to t of length 1
2. For all v E V(G)
\ V(P),
s -+ v
c
E(G)
then c(G) < l!nlg n + 2
Proof.
Definition 5.2.2. Given an ordering v 1 ,---
,vn
of the vertices of Vred(G),
define
P'(v1,--- ,vn) to be the certain knowledge switching network with vertices s',v',--and all allowed edges where K,= {s -+ t}, K,, = Ku
, and Kt,
{s
-+
,v',
t}.
Lemma 5.2.3. If G 2 is an input graph such that
2. For all v C V(G)
\
V(P), s
- - --+v
then P'(v1, - - - , v,) accepts G 2 if u 1 ,
u 1 _1 -+ t in G from s to t of length l
E E(G 2
)
1. There is a path P = s - u 1
, ul-1 appear in order in v 1 ,-
,Vn
Proof. Take uo= vo = s, U 1 = v =t, v = s', and v'+1 = t'- If ul,-
, ul-1 appear
j
. To see this,
in order in vi,-
,
vn then for all
E [0, n] we can go from K , to K,
note that we have the following cases
110
t'
1. If vj+1 V U'_ 1 {ui} then s
2. If v3 +1 =
v1 ,-
Uk+1
-
vj+1 E E(G 2) so we can go from K,' to K,'
for some k c [0, 1 - 1] then since u1, - -, u, appear in order in
-,Vnwe know that
Uk
E U
U
0
{vi}. Thus, since Uk
-
Uk+1 E
E(G2 ) we
can go from K,, to K,,+i, as needed.
This implies that P'(vi,
) accepts G 2 , as needed.
,
E
We now construct G' by taking q random orderings of the vertices Vred(G), taking
the corresponding P', merging all s' together, and merging all t' together. IV(G')=
nq + 2.
By Lemma 5.2.3, for a given permutation G 2 of G, when we choose each P' the
probability that it accepts G 2 is at least (1'.
Thus, the probability that no P'
accepts a given permutation G 2 of G is at most (1-
)q.
There are at most ni-1
different permutations of G, so by the union bound the probability that G' does not
accept all permutations of G is at most n1-1(1
-
1
)
)q.
1 )] then n'-'(1 - (
If we take q = [(l - 1)! ig (n-'
1 1 ))
< 1 so there is a nonzero
chance G' accepts all permutations of G. Thus, since [(l - 1)! lg (ni- 1)] < 1! lg n,
c(G) < l!n lg n + 2.
5.3
A general upper bound
We now show a general upper bound on input graphs G in which the vast majority
of vertices are lollipops.
Theorem 5.3.1. For any input graph Go containing a path from s to t, if we can
win the reversible pebbling game on Go using at most k pebbles then letting m
V(G0 )
\ {s, t},
=
for any input graph G such that G includes Go and all other vertices
of G are lollipops, m(G) is nO(1)mO(k). More precisely,
m(G) <
2
3
2
k+2 m k+l
n(n+ 2 2k+1mk+1)(gn)
Proof.
111
2
Definition 5.3.2. We call a vertex v E Vred(G) even if there is a path from s to v in
G. and odd if there is a path from v to t in G.
Proposition 5.3.3. If any vertex v C Vred(G) is both even and odd then there is a
path from s to t in G.
Definition 5.3.4. We define a parity knowledge set K to be a set of pairs (V, b) where
V C Vred(G) and b E {0, 1}.
(V, 0) represents the knowledge that either there is a
path from s to t in G or the number of odd vertices in V is even. (V, 1) represents the
knowledge that either there is a path from s to t in G or the number of odd vertices
in V is odd.
Definition 5.3.5. In the parity knowledge game, we can make the following types of
moves on a parity knowledge set K
1. If we directly see the edge e = s -+ v in G then we may add or remove ({v}, 0)
from K
2. If we directly see the edge e =v -+ t in G then we may add or remove ({v}, 1)
from K
3. If ({v}, 0) E K and we directly see the edge e = v -+ w in G then we may add
or remove ({w},0) from K
4. If ({w}, 1) E K and we directly see the edge e = v -+ w in G then we may add
or remove ({v}, 1) from K
5. If (V, b 1), (W, b 2 ) E K then we may add or remove (VAW, b1 + b 2 mod 2) from
K
6. If ({s}, 1) C K or ({t}, 0) C K then we may add or remove anything else from
K.
We win the parity knowledge game if we reach a parity knowledge set containing
({s}, 1) or ({t}, 0)
112
Proposition 5.3.6. We can win the parity knowledge game for an input graph G if
and only if there is a path from s to t in G.
Just as we did for certain knowledge switching networks, we can use this game to
construct switching networks where each vertex v' corresponds to a parity knowledge
set Kv, and each edge with label e between vertices v' and w' in G' corresponds to a
sequence of moves in the parity knowledge game between Kv, and K"1, all of which
can be done using only the knowledge that e E E(G).
We first note that given a set of vertices V
C
V(G) and a distinguished vertex v,
if all vertices in V \ {v} are lollipops then it is relatively easy to recover the parity of
v from the parity of V.
Definition 5.3.7. If all vertices U C Vred(G) are lollipops then we take b(U) = 0 if
there are an even number of u C U such that there is an edge from u
-
t and we take
b(U) = 1 otherwise.
Lemma 5.3.8. For any V
C
Vred(G), we can construct a switching network G' such
that
1. G' has a vertex with parity knowledge set {(V,0)} and a vertex with parity
knowledge set {(V, 1)}
2. For all v c V, G' has a vertex with parity knowledge set {({v}, 0)} and a vertex
with parity knowledge set {({v}, 1)}
3. If G is an input graph such that for some v E V, all vertices u E V \ {v} are
lollipops, then for each b G {0, 1} there is a path P' in G' from {(V, b+b(V\{v})
mod 2)} to {(v, b)} whose edge labels are all in E(G) and of the form s -+ u or
u -+ t where u e V \ {v}
4.
The number of vertices in G' excluding the ones mentioned above is at most
Proof. We prove this lemma by induction. Let f(x) be the number of additional
vertices needed if IVI = x. The base case IVI = 1 is trivial. For larger IVI, split
113
V into two halves, VL and VR. If v
C
then taking an ordering of wi, -, wIVRI of
VL
the vertices in VR, take one vertex for each of the additional parity knowledge sets
K
=
(VL U {wi : i < j}, b) where j E [0, IVRI
-
1] and b E {0, 1}.
If we are at the parity knowledge set (VL U {wi : i < j}, b) and j > 1, we may
use the following sequence of moves to reach the parity knowledge set (VL U {w i<
j
-
1},b+b({wj})
mod 2)
1. Add ({w}, bw,) to K
2. Since (VL U Iwi : i < j}, b) C K and ({wj}, bw,) c K, add (VL U {wi :
j-
1}, b + bj,
<
mod 2) to K. Then remove (VL U{wi : i < J}, b) C K from K.
3. Remove ({w 3 }, be,) from K
+
Applying this process repeatedly, if we start at the parity knowledge set (V, b
b(V \ {v}) mod 2) we will reach the parity knowledge set (VL, b + b(V \ {v}) + b(VR)
mod 2) = (VL, b + b(VL \ {v}) mod 2). We can now use the inductive hypothesis on
VL.
Following similar logic, if v E VR we can taking an ordering w, ... , wiye of
the vertices in VL, take one vertex for each of the additional parity knowledge sets
K = (VR U {wi : i
j}, b)
where
j
c [0, VL
-
1] and b C {0, 1}, and then use the
inductive hypothesis on VR. Putting everything together,
f (|V|) < 21VLI
f(IVL
)+
21VR + f (|VR|) < 2|VL|(1 + [Ig |VLfl) + 21VRI(1 + [lg VRf)
< 2|VLI [Ig VJ+ 21VRI[lgJVJ]
< 2|V||Ig|IV|]
as needed.
El
We also note that if we have subsets V, W and all vertices in VAW are lollipops,
it is easy to obtain the parity of W from the parity of V. We will use this as follows.
If we have a distinguised vertex v and subsets V, W such that v E V n W and all
114
other vertices of V and W are lollipops, then we can switch from storing infromation
about v in the parity of V to storing information about v in the parity of W.
Lemma 5.3.9. For any V, W C Vred(G) such that V
nW
#
0, we can construct a
switching network G' such that
1. G' has a vertex with parity knowledge set {(V, b)} for b E {0, 1}
2. G' has a vertex with parity knowledge set {(W, b + b(VAW)
mod 2)} for b e
{o, 1}.
3. If G is an input graph such that all vertices in VAW are lollipops, then for
each b E {0, 1} there is a path P' in G' from the vertex with parity knowledge
set {(V, b)} to the vertex with parity knowledge set {(W, b + b(VAW)
4.
mod 2)}
The number of vertices in G' is at most 2(V| + |W|).
Proof. Letting vi, - -- , v,, be the vertices of Vred(G), take one vertex for each distinct
parity knowledge set of the form ((U -j+ 1vi
n V)
all allowable edges. Note that starting from j
=
U (Uj~vvi n W), b) and then add
1, each new
j
gives distinct parity
knowledge sets if and only if vj C VAW. Thus, we have at most 2(fVAW| + 1) <
2(V + WI) vertices. Also note that whenever vj E VAW, since vj is a lollipop we
can go from {((Ug7igvj n V) U (U _ Vi n W), b + b ((VAW) n (Uj_-Ivi))
{((U-j+1 vi nV) U (uivi nW), b+ b ((VAW) n (uj v.))
from
j
= 0 to
j
mod 2)} to
mod 2)} Using this to go
D
= n gives the result.
Corollary 5.3.10. For any subsets V1 ,all i G [1, r], Vi n Wi
#
1. For all values b 1 ,-
, Vr,
W1 , ...
, Wr C
Vred(G) such that for
0, we can construct a switching network G' such that
br, G' has a vertex with parity knowledge set {(V, bi) : i e
[1, r]}
+
2. For all values b 1,-- br, G' has a vertex with parity knowledge set {(Wi, bi
b(VXAWj)
mod 2): i E [1, r]}
115
3. If G is an input graph such that for all i, all vertices in V1AW are lollipops,
then for each set of values b 1 , - , b, there is a path P' in G' from the vertex with
parity knowledge set {(Vi, bi)
{(W, bi + b(VIAW)
i C [1, r]} to the vertex with parity knowledge set
mod 2): i E [1, rl}
4. The number of vertices in G' is at most 2' Z, ( ViI + |Wi) < 2r+ln.
Proof. We prove this corollary by induction on r. The base case r = 1 is Lemma
5.3.9. If r > 1, for each possible set of values b 2 , - , b, create a copy of the G' given
by Lemma 5.3.9 between the vertices with parity knowledge sets {(Vi, bi) : i E [1, r]}
and new vertices with parity knowledge sets {W 1, bi + b(VAW 1 ) mod 2} U {(V, bi)
i E [2, r]}. The total number of vertices in these copies is at most 2(1I+IW1|). Now
for b1 = 0 and b, = 1, we use the inductive hypothesis on the vertices with parity
knowledge sets {W 1, bi + b(V1 /AW1 )} U {(Vi, bi) : i C [2, r]} and the vertices with
parity knowledge sets {(W, bi + b(VAWj) mod 2) : i E [1, r]}. The total number of
vertices required is 2'
2r
E-2
( Vi +
IWi 1).
Summing these up gives an upper bound of
E (1Vi + IWi), as needed.
We now define switching networks which are effective at solving inputs with lots
of lollipops.
Definition 5.3.11. Given disjoint sets of vertices {Vi : i C [1, q], j E [1, m]} where
for all i, Vi1,..-
Vi
is a partition of Vred(G), we define the switching network
G,,k({Vij}) so that
1. s' has parity knowledge set {} and t' has parity knowledge set {(t, 0)}
2. For every i, every I C [1, m] such that |I| < k, and every set of values {bj : j C
I}, there is a vertex in G' with parity knowledge set UjEI{(Vij, bj)}.
3. For every i, every I C [1,m] such that
Va C
Via, and every set of values
{bj : j
\I\
< k, every a E I, every vertex
C I}, there is a vertex in G' with parity
knowledge set UjcI\{a}{(Vij, bj)} U {({Va, ba})}.
116
4. For every i, every I
|I
C [1,m] such that
< k, every a E I, and every set
of values {bj : j E I \ {a}}, there is a copy of the switching network given
by Lemma 5.3.8 between the vertices with parity knowledge set of the form
UjcI\{a}{(Vij, bi)} U
{({ Via, b})} where b e {0, 1} and the vertices with parity
knowledge set of the form UjIe\{a}{(Vi, b)}} U {({Va, b})}
where Va e Via and
bE {0, 1}
5. For every i, every I C [1, m] such that I|
k, every pair of distinct a, c e I,
every vertex Va G Via, every vertex v, E Vc, and every set of values {b:
j C I}, there is a vertex in G' with parity knowledge set UjzI\{a,c}{(Vi, bj)} U
{({Va, ba})} U {({ve, bc})}.
6. For every i, every I C [1, m] such that II < k, every distinct a, c E I, every
vertex Va E Va, and every set of values {bj : j e I \ {c}}, there is a copy of the
switching network given by Lemma 5.3.8 between the vertices with parity knowledge set of the form UjE1\{a,c}{(Vij, bj)} U {({Va, ba})} U
{({Vi,
{0, 1} and the vertices with parity knowledge set of the form
{({Va, ba})}
b})} where b c
UiEI\{a,c}{(Vi
,
b )}U
U {({ve, b})} where v, E Vi, and b E {O, 1}
k and for every i1 ,i 2 E [1,q] such that
7. For every I C [1,m] such that I|
i1 < i2 and for all j G I, V(i,)j n V( 1 )i
$
0, we have a copy of the switching
network given by Corollary 5.3.10 between the vertices with parity knowledge
sets Ujys{(V(i1 )i bj)} and the vertices with parity knowledge sets
UjEI(Vi2 ),
bi +
b(V(i)jAV(i 2 )j)
mod 2)}
Proposition 5.3.12. For any disjoint sets of vertices V1 ,.--
,Vm
C Vred(G), the
switching network GWk({Vij}) is sound.
Proof. Assume that there is a path P' from s' to t' in G,k(V1, V2 , -
, Vm)
whose edge
labels are all in E(G) for some input graph G. Note that any such path corresponds
to a sequence of moves in the parity knowledge game from K = {} to K = {({t}, 0)}
where each move can be done using an edge in E(G). Since we can only win the
117
parity knowledge game when there is a path from s to t in G, there must be a path
from s to t in G. Thus, G,,(V1, V 2 , -
, Vm)
is sound.
Lemma 5.3.13. Let Go be an input graph with distinguished vertices s, t such that
IV(Go)\{s, t} = m and it is possible to win the reversible pebbling game on Go without
placing more than k pebbles. Letting v 1,-.
,vm
be the vertices of V(Go)
\{s, t},
if G
is an input graph such that
1. E(Go) C E(G)
2. All vertices in V(G)
\
V(Go) are lollipops
and {Vij : i E [1,q], j C [1,m]} are sets of vertices such that
1. For all i, Vi1,. -vim
I is a partitionof Vred(G)
2. For all I C [1,m] such that I| < k, there is an i e [1,m], such that v E Vi for
all j C I and Va
Vij
whenever a V I and j E I.
then Gk({Vi }) accepts the input graph G.
Proof. If we have pebbles on vertices vj: j E I in Go then this corresponds to being
at a vertex with parity knowledge set {(Vi, b(Vij
v3 C Vi for all j E I and Va V Vi whenever a V
\ {vj}))
I and j
: j E I} for an i such that
c I. We now match each
reversible pebbling move on Go with a sequence of moves in G,,
fVi
1.
Proposition 5.3.14. If vj E V(1), n V(,)j and all other vertices of V(ij 3 and V(,)j
are lollipops then
b(V(l)j \ {vj}) + b(V( 1 )jAV(i 2 )j) mod 2 = b(V( 2 )j \ {v 3 })
To add a pebble on
Vb
for some b E [1, m], we first use the copy of the switching
network given by Corollary 5.3.10 to shift to an i such that vj E Vi for all j E I U {b}
and Va 0 Vi2 whenever a V I U {b} and
j
c I U {b}. We now have the following cases
118
1. If we have an edge s -
Vb
G,,({Vij}) we first add
in E(GO) and are using it to add a pebble on
({Vb},
Lemma 5.3.8 to transform
({Vb,
in
0) to K. We then use a copy of the G' given by
0}) into
(Vib, b(Vi
2. If we have a pebble on va and an edge va
add a pebble on vb, in
Vb,
G,,k({Vi})
-* Vb
\ {Vb})).
in E(GO) and are using it to
we first use a copy of the G' given by Lemma
5.3.8 to transform (Via, b(Via \ {va})) into ({Va}, 0) in K. We then use the edge
Va --
Vb
to add ({vb}, 0) to K. Finally, we use a copy of the G' given by Lemma
5.3.8 to transform ({Va}, 0) into (Via, b(Via
\
{va})) and then use a copy of the
G' given by Lemma 5.3.8 to transform ({vb}, 0) into (Vib, b(Vi \ {Vb})).
To remove a pebble on vb, we run these steps in reverse except that we do not need
to shift to a different i.
If we have a pebble on va and an edge va -+ t in E(G) which we use to win, in
G,,({Vij}) we first use a copy of the G' given by Lemma 5.3.8 to transform (Via, b(Via\
{Va}))
into ({Va}, 0) in K. We then use the edge vi -
t to add ({t}, 0) to K and
win.
D
We now analyze the size of G
Lemma 5.3.15.
IV(G,,k(IVij}))
k({Vij}).
2qn2 g n(2m) k + 2q 2 n(2m)k
Proof. We count how many of each type of vertex G({Vi})
1. There are at most q E1k
has.
2r (') possible parity knowledge sets of the form
Uj I{(Vi, bj)}.
2. There are at most qrt n
k
UjCI\{a}{(Vi, bj)} U{({Va,
{bj} and
2r (m) possible parity knowledge sets of the form
ba})} as we have the same choices for I and the values
are now choosing one vertex Va-
3. For each fixed a, there are at most q E_- 2'(7) copies of the switching network given by Lemma 5.3.8 between vertices with parity knowledge set of the
form UjCI\{a}{(Vij, b 3 )} U
{({ Via, b})}
and vertices with parity knowledge set
of the form UjGI\{a}{(Vij, bj)} U {({Va, b})}.
119
Each of these copies has at most
2 Va lg n additional vertices so the total number of additional vertices is at most
21Val ig n Ek- 2r (m). Summing this up over all a gives 2qn lg n
4. There are at most qn 2
UicI\{ac}{(Vij bj)} U
2 ' (in)
k_ 2r(7) possible parity knowledge sets of the form
{({Va,
ba )U
I{({vc b})}
as we have the same choices for
I and the values {bj} and are now choosing two vertices Va and
5. For each fixed c, there are at most qn j _- 2
(7)
vc.
copies of the switching network
given by Lemma 5.3.8 between vertices with parity knowledge set of the form
UjEI\{a}{(Vij,
bj)}U{({Va, ba})}U{({Vc, b})}
set of the form UjEI\{a}(
j
bj)} U
{({a,b
and vertices with parity knowledge
Each of these
a})} U {({vc, b})}.
copies has at most 2 Vc Ig n additional vertices so the total number of additional
vertices is at most 2qnIVc lg n j$k- ' 2r (7). Summing this up over all c gives
2qn2 lg n E _- 2r(r)
6. There are at most
(q)
_1=Z
(7) copies of the switching network given by Corol-
lary 5.3.10. Each such switching network has at most 2r+ln vertices giving us
a total of at most q2
k
2'(7) vertices
Summing up all of these numbers and adding s' and t' gives
k
2
q(1+n+n +qn)
k-i
2
+ 2q(n2+n)lgnE2
+ 2
r=O
r=1
If n > m > 4, this is at most 2qn 2 lg n(2m)k + 2q2 n(2m)k, as needed.
D
Finally, we put an upper bound on how large q has to be in order to have a nonzero
chance that for all possible v, - --vm, for all I C [1, m] such that III < k, there is an
i E [1, m], such that vj E Vi for all j E I and va
Vi whenever a
I and j E I.
To do this, we choose each Vil, -.- , Vin by taking each v E V(G) and randomly
selecting which Vpj it will be part of. For a given vi, -
, Vm
and I of size k, the
probability of success is (y)k (ink)m-k. Now
mn
-)k
k
(m
M- k)
k\ m\
>-(
> m-k
1 -- --
120
2k
> mn-k 2--2k __(Tn -k
Thus, for a given vi, - - , vm and I of size k, the probability that we fail for all i is at
most (1 - ( 4 m) k) . There are at most mknm possibilities for vi,- - , vm and I, so
the expected number of v,- - , vm and I which fail is at most
mknm (1 - (4m)-k)q
If q > (4m)k ig
(ik
nm) then this is less than 1 so there must be some chance of success.
-
Thus we may choose q so that q < 2m(4m)k ig n. Plugging this into 2qn 2 ig n(2m)k
2q2 rn(2m)k we get an upper bound of
2
23 k+ 2 m 2 k+ rn2 (lg n)
+
2 5k+ 3 m 3k+2 n(lg n) 2
as needed
23 k+ 2 m 2 k+lrn(n
+
2
2
k+lmk+l)
(lg n) 2
D
121
122
Chapter 6
Improved Lower Bound Techniques
for Monotone Switching Networks for
Directed Connectivity
In this chapter, we prove the following theorem, which shows that if G is a directed
acyclic input graph, we get a lower bound on m(G) based on the length of the shortest
path from s to t as long as the vertices of G aren't too connected to each other.
Theorem 6.0.16. If r and m are constants such that m < 2000r4 and G is a directed
acyclic input graph such that
1. There is no path of length at most 2'-1 from s to t.
2. For any vertex v G V(G), there are at most m vertices w E V(G) such that
either there is a path of length at most
2r-2
from v to w in G or there is a path
of length at most 2 r-2 from w to v in G
then
m (G) >
(9mn)
4
201 E(G) V/2(r+ 1)! 9m
The key idea in proving this theorem is as follows. In chapter 3 (see Theorem
3.4.25), we found functions gG,E, corresponding to each input graph G such that
whenever G' accepts G it must contain a vertex v' such that gG,Ei
123
-
9G,E1
V'
is large.
We then obtained a lower bound based on an orthogonality argument. Unfortunately,
the orthogonality argument depended on the input graphs consisting almost entirely
of isolated vertices.
To prove Theorem 6.0.16 we instead argue that if we take a
random permutation o- of Vred(G),
9G,E1 ) - v']
E[(gG,Ei -
is small for any possible v'
and this is sufficient to prove lower bounds.
6.1
Permutations on input graphs, cuts, functions,
and Fourier coefficients
We first define what we mean by permutations of functions, cuts, and input graphs
and show how permutations affect the Fourier coefficients.
Definition 6.1.1. We take
SVred(G)
to be the set of permutations c- of V(G) such that
u(s) = s and -(t) = t.
Definition 6.1.2. For each permutation - E
1. Given an edge e = v
-
SVred(G),
w, define o-(e) = o-(v)
-
o-(w).
2. Given an input graph G with vertices V(G), define -(G) to be the graph with
vertices V(G) and edges {ou(e), e
e
E(G)}.
3. Given a cut C E C, define -(C) to be the cut such that L(o-(C)) = o-(L(C)) and
R(u(C)) = o-(R(C)).
4. Given a function g : C -+ R, define the function -(g) : C
o-(g)(a-(C))
=
g(C), i.e. a-(g)(C)
=
-+
R so that
g(a--1 (C)).
Proposition 6.1.3. For all - C SVed(G),
1. An edge e crosses a cut C if and only if o-(e) crosses o-(C).
2. g : C -÷ R is e-invariant for some edge e if and only if o-(g) is o-(e)-invariant.
3. For allf,g :C->R, -(f)-o(g) =f g
124
4. For all V
C
Vred(G), o(ev) = eo(v)
5. For all g :C -+ R, for all V c Vred(G), OY),(A) =A
Proof. Statements 1 and 3 are clear, statement 2 follows from statement 1, and statement 5 follows from statements 3 and 4. Thus we only need to show statement 4. To
see statment 4, note that
-(ev)(C) = ev(o- 1 (C)) =
(-1)-,(oa(C)nv1
(_
1
)7-1 (L(C))nVi
-
(_i)L(C)no(V)
_
D
eo(V) (C)
Finally, we recall the definition of m(G).
Definition 6.1.4. For an input graph G containing a path from s to t, m(G) =
m({-(G) : - E SVde,(G)})
Remark 6.1.5. The function m(G) is a complexity measure on G which measures
how hard it is for sound monotone switching networks to solve directed connectivity
on G.
6.2
Calculating Permutation Averages
For our lower bounds, we will need bounds on expressions of the form EGSVred(G)
-(g)] 2 where
f, g
[f
are functions from C to R and c-(g) is a permutation of g. Here
we obtain bounds on EacsVred(G)[f . -(g)] 2 in terms of the norm I fII of f and certain
sums associated to the function g. This will be extremely useful because the function
f
will correspond to a vertex in an arbitrary sound monotone switching network G'
so we will have no control over anything except for I f 1. However, we will have a
large degree of control over g and will thus be able to adjust the values of many of
the sums associated with g to give us the bounds we need. We now state our bound
on
EEsV
d(G)[(f
Definition 6.2.1. Given functions gi,92 : C -+ R, define
Sk,u1,u 2 (Y1, 92 ) =
E
A,B,CCVred(G):AI=kIBI=ul,IC=u2,
AnB=AnC=BlC =
125
g1AUBg2AUC
Theorem 6.2.2. For any function g : C < - 1, for any function f : C ->
IVR,
R such that v is only nonzero when
;
EEsVe(G)[(f
o(g)) 2 ] <
+ U2)!
-u)!(k
-V'(k +
4
n
k,ui,u2
.
Sk,ul,u2 (gg)
k+u1+u2
2
lfHl
Proof.
Lemma 6.2.3. For any functions f, g : C - R,
k!ui!u 2!(n - k - ui
2)!
-
-
EovsVred(G)[(f . o(g)) 2 ]
k,ui
U2
Proof.
EaESV,-d(G)
2]=
.
F
(Z
I
f
or (g) V
)
EO'ESVre(G)[(f
Vi CVred(G)
E
fV 2 a (g) V2
V2 CVred(G)
Taking A =v nV2 , B = V1 \ V2 , C = V2 \ V, this is equal to
jAUB
ETSsVred(G)
k,u2,Uz
.
AUC
9 gAUB
S
AUC
A,B,CCVred(G):
LA|=k,\B|=u1,BC =u2,
AnB=AnC=BnC=O
which is equal to
Y,
k,ui,u 2
EAUBACEEESVred(G) U(Y)AUB
(g) AuC]
(6.1)
A,B,CCVred(G):
|A|=k,|B|=u1,|C|=U2,
AnB= AnC=BnC=O
Now by statement 5 of Proposition 6.1.3, for any A, B, C C Vred(G) such that JAI
|B|= u1 , |C|= u2 , and AnB=AnC= BnC=0,
EassVreS(G) U(g)AUB
()AUC]=
EaCsVreS(G) [-
1(AUB)Y-1(AUC)]
k!ui!u2 !(n - k - ui- U 2 )!
--
126
I
sk,ui,U2
(g, g)
=
k,
Plugging this into (6.1) we deduce that
Eo-csV, ed(G)
[(f
k!ui!u 2 !(n -
=
k - u1
-
2 )!
k,ui,u2
which completes the proof of Lemma 6.2.3.
E
We now wish to bound how large sk,ul,U2
(f,
f)
can be in terms of I|f 1l.
Definition 6.2.4. Given a function g : C --+ R and a subset A C Vred(G), define
SA,u 1 (9) =AUB
BCVred(G):
IB|=u1,AnB=O
Definition 6.2.5. Given functions 9 1 ,92 : C -+ R and a subset A C V,d(G), define
SA,u1,u2
(91, 92)
=
E
g1AUB2AUC
B,CCVred(G):
IBju1,ICI=u2,
AnB=AnC=BnC=0
Proposition 6.2.6. For all A,
8A,ui (91)SA,u 2 (g 2
SB,ui-U,u 2 -u(91,
92)
u>O ACBCVred(G),
|BI=AI+u
,
Corollary 6.2.7. For all k, u1 , and u 2
Z:
A:IAI=k
SA,ui (91)SA,u 2 (92)
=
5 (kI- )Sk+u,ui
-U,U2 -U
(91, 92)
u>O
Proof. This follows immediately when we sum Proposition 6.2.6 over all A of size
D
k.
Surprisingly, Corollary 6.2.7 has an almost identical inverse formula.
127
,
Lemma 6.2.8. For all k, u1 , and U2
Sk,u,u2
(- 1)U (k
(91, 92)
SA,ui -u (91) SA,u 2 -u (92)
U
u>O
A:|AI=k+u
Proof. This lemma corresponds to the fact that the inverse of a matrix like
(k+1)
(k+2)
(k+3)
o
1
(k+2)
(k+3)
0
0
1
(k+3)
0
0
0
1
is
(k+1)
1
k+2)
2
(k+3)
(k+3)
(k+2)
0
1
0
0
1
0
0
0
(k+3)
1
Calculating directly, every entry in the product of these matrices has the form
(-1)u
U=O
where
j
+ u)
U
(j+2)
m-U
is k plus the row number and m is the column number minus the row number.
Now if m > 1,
m
(-1)"
(Ij +
+ Mn
U)
j!
U=0
m)'
_(j+
-(I
f!m!
U=0
(-1 )"
+ (-1))"
1
=0
If m = 0 then
m
1
E
-1
UO-
E
To bound how large ISk,u,U 2 (f, f) can be in terms of If 11, we just need bounds
128
for how large expressions of the formZAjAjk SA,uj(g)sA,u 2 (g) can be.
,
Proposition 6.2.9. For all k, ul, u 2
>
sA,ui(9)sA,u 2 (9)
A:IAI=k
1:
(SA,ui (_9))2
A:IAI=k
(A,U2
(g) 2
A:IAI=k
Proof. This follows immediately from the Cauchy-Schwarz inequality.
D
Proposition 6.2.10. For all k,u1 and all A with A| = k,
(SA,ui(g))
k)
2
I:
)2
( AUB
BCVred(G):IBI~ul,
AnB=O
Proof. This follows from the Cauchy-Schwarz inequality
(BCVed(G):IBI=
2
I
<
f(B)h(B)
)
E3
BCVred(G) IBI=ui,
AnB=O
AnB=0
E3
(f(B)) 2
(h(B)) 2
BCVred(G):IBI=ui,
AnB=0
El
with f (B) = 1 and h(B)= .AUB.
Corollary 6.2.11. For all k, ul,
(SA,u 1 (g)) 2
I:
A:IAI=k
k)
k+u1)
U1
(
B:IBI=k+ui
<
k +1
n- k
2
(yB1)
Proof. This follows immediately when we sum Proposition 6.2.10 over all A of size
El
k.
,
Corollary 6.2.12. For all kU 1 , U2
A:IA|=k
SA,ui (9)SA,u
k) n -k)
(n
2
(9
U1
U2
k
(k +U1
U1
)
(
I:
+U2
U2)
gI
Proof. This follows immediately from Proposition 6.2.9 and Corollary 6.2.11
129
E
f
is a function f : C -
S
(k
U>0
f)
k
-
U)
U
("U1i-
+
Sk,u,U 2 (,
,
R, we have that for all k, u 1 , u 2
-
u
U
)
Corollary 6.2.13. If
n - k- u
k+u1
k+U2
U2-U )(U -U||U2-|
1l
Proof. This follows immediately from Corollary 6.2.12 and Lemma 6.2.8 with k replaced by k + u, u, replaced with u, - u, and u 2 replaced with a2
-
D
U.
Corollary 6.2.14. If n > k + max{ui, u 2} + 2(max{ui, u 2}) 2 then
(f, f)l
< 2
(k+u)
Ui (n)
U2
(n)
U1
k +U2
U2 lfil
)
ISk,ulu2
Proof. The idea is to show that each term in the sum in 6.2.13 is at most half of the
previous term. This follows from the equations
1
k+u+1
-_____
+1
)k+u
'
U
2U-1
u2-u-1
(ul-u)(U2-u)
2
(n-k-u)
n-k-
UjUk+u 1 1 U2
4(ul-u)(u2-u)
n2 k-u_
k+u 2
(ui -U) (U2 -u)
(k+u+1)2
-
32.
1
-
FD
Theorem 6.2.2 now follows easily. By Lemma 6.2.3,
EgcsVred(G)[(f . o(g)) 2 ]
k!ui!U2 !(n - k
=
ni!
k,u1 ,U2
130
-
ui-
U2)!
Sk,u2 (f i
)S
,u2 (9
g)9
Assuming that n > 2k + u -+u 2 , plugging Corollary 6.2.14 into this gives
Eas
~Sd
(G)
[(f
g)2
.
(k-+ i- l!kU2!(n
!k+9)
<2<2S
E -(+
(n - u n -U2)!
kulu2
V(k +-2-k+21+u2
-u)!(k
+ U2)!
<2
u2
<2
-
U2)!
k
V
4<4
ISk,ui,U2 (g,
k+U1 +U2
n
ku,2(gg)IIfII
sku1 ,u2 (g, g)U
'f
+
2---2n
k,u,,U2
as needed.
2
(ui+u 2 +k)(k+u1+2)
k,un
(k +U2)!
k
vl-f(k + u 1)! (k +
ki 2 u
9
g)I-II
2
To check that the bounds on n hold when needed, note that by our
assumption that .v is only nonzero when IV I
k + max{ui,
2
}>
-1.
- -- 1, we may ignore all terms where
Thus, for all of our terms, n
2(k+max{ui,u 2}
+ 1)2
This is bigger than both k + max{ti, U2} + 2(max{ui, u 2}) 2 and 2k + ui + U2, as
needed.
E
For the functions g we will be looking at, it is difficult to bound |sk,ul,u2 (9, g)1
directly. We would like a bound in terms of the sums SA,u(g). To obtain such a
bound, we apply Lemma 6.2.8 and Proposition 6.2.9 to Theorem 6.2.2.
Corollary 6.2.15. For any function g : C -+ R such that v is only nonzero when
IV I
r for some r < -
EUESVred(G)[(f
- 1, for any function
O(g))
2]
f
: C -+ R,
< 4(r + 1)f 11E 2'
k,u
(SAu(g)) 2
(k + u)!
E
k+u
A:IAI=k
Proof. By Lemma 6.2.8 and Proposition 6.2.9,
(k +-i)!(k +
n
ku1,U2
<SE
ku ,U 2
u:!min {U 1 ,u 2 }
k+u1+u2
2
(
k~
k--\
2)!
sk,ui,u 2 (9 9)
k k+ul)!(k
ul~u2 +
n
2
(5A,ul -u(9)) 2
U22)!
A:IAI=k+u
,
2(k +ui)!
-
f II
ISku172(1911
Isk,ui,u 2 (9, g)| '|f|
-
ku1 ,u 2 (n
u1 - U2 - k)! sk,ul,U2 (9, 9)1 -1
131
>1?
E
A:IAIu~k+u
(AU
(
g-())
2
Replacing k + u with k, u1 - u with u 1 , and u 2 - u with u 2 on the right hand side,
z (k +
,U
uI:
n
k,ui,U2
ui)!(k+ U 2 )!
k+
Sk,u 1 ,u 2 kyg)
2
1U
2
+ui)! (k + U2)!(g)
k)(k
A:|AI=k
(sA,u 1 (9)) 2
S
(sA,u 2 (g)) 2
A:|AI=k
(
k,u1,2
_
n k+u_ 2 u2
2
n
u1
k
2
A:IAI=k
2
u)!g))2
1)+
(r+
sA,ul
+
k
= ((
u
n
i
+1)! 2S~u (g)
AI (A|=k
2
2
k,u1
A:A|I=k
Plugging this in to Theorem 6.2.2 gives the desired result.
6.3
D
Lower bounds from permutation averages
In this section we show how we can obtain lower size bounds on monotone switching
networks using Corollary 6.2.15.
Definition 6.3.1. For an input graph G with a path from s to t, we say that FG =
{g : e e E(G)}
is a set of invariant functions for G if
1. For all e c E(G), ge is e-invariant.
2. For all e E E(G), g, - e
= 1
Theorem 6.3.2. Let G be an input graph containing a path from s to t and let
FG= {g, : e E E(G)} be a set of invariant functions for G. If for all e E E(G) we
have that 4 ev is only nonzero when
|V| <
r for some r <
_-
1, then for any edge
e o c E(G),
m(G) >
1
1max
|E(G)j -I1
e(EE(G)\{e,,}
(k +
+ u)!
4(r + 1)5 k 2k(k
k
132
n k+u
AI)
A:A
(sA,u(ge - geo))2
k
Proof.
Lemma 6.3.3. Let G be an input graph containing a path from s to t and let FG
=
{g, : e c E(G)} be a set of invariant functions for G. For any edge eo E E(G), for
any sound monotone switching network G', for any path P' in G' from s' to t' whose
edge labels are all in E(G),
1.
Zv',Gp') ZeE E(G)\{eo}
2.
Ev'CV(G),ecE(G)\{eo} 1V'
3. EvCV(G'),eEE(G)\{eo}[IV'
K'
(ge
(g -
-
geo)f > 1
g)
>
(ge - Yeo) 2 -
1)V(G)
(lE(G)I-1) 2 1V(G')
2
Proof. The first statement follows from Corollary 3.4.23. The second statement follows immediately from the first statement. The third statement follows immediately
D
from the second statement and the fact that for any X, E[X 2 ] > (E[X]) 2 .
Now if G' is a monotone switching network accepting all of the input graphs
{o-(G) : - G
SVred(G)}
EorSVred(G)
then by statement 3 of Lemma 6.3.3,
[Ev'EV(G')\{s',t},eE(G)\{eo}[V - U(ge
(E(G)
- ge 0 ) 2]]
1
-
1)2V(G')2
Applying Corollary 6.2.15 to g, - ge0 and v' for each e E E(G) \ {eo} and v' E V(G'),
since IIv'I < 1 for all v' E V(G'),
|VG'|
EeCEG\eO [4(r + 1)
> E(G)I1V(G')l
- I EnEG\e}ATk+u
S 2k(k + u)!
k,u
()2
(sA,u fge -
eo
2
A:IA|=k
0
Theorem 6.3.2 now follows immediately.
Theorem 6.3.2 says that for our lower bound, we want to find a set of invariant
functions FG
the sums ZA:A(
{eg : e E E(G)} such that for some eo C E(G), for all e E E(G) \ {eo}
k (sA,u(ge - geo)) 2 are as small as possible.
133
6.4
Equations on sum vectors
Rather than choosing the functions {ge : e E E(G)} directly, we choose the sums
{SA,u(ge)
A C Vred(G),u > 0,e E E(G)} and have these sums determine the
functions {ge : e
c
E(G)}. However, our choices are not entirely unrestricted. In this
section, we show which equations on the sums {SA,u(ge) : A C Vred(G), u
0, e E
E(G)} we need to obey. These equations are most easily expressed in terms of vectors
with sums as their coordinates.
Definition 6.4.1. For a function g and k, u > 0, define Sk u,g to be the vector with
one coordinate for each A C Vred(G) of size k such that
(Sk,u,g)A = sA,u(g)
Proposition 6.4.2. For any function g : C -+ RI, ISk,u,g|1
Vred(G), AI = k}
k + 1}. Take
Pk
-A:IAI=k
(sA,u(g))
2
be the matrix with rows correspondingto the subsets { A C
and columns corresponding to the subsets {B C Vred(G), B|
(Pk)AB =
-
Definition 6.4.3. Let
2
1 if A C B and 0 otherwise.
Proposition 6.4.4. For any function g, sk,u,g
Sk+1,u-1,g
Proof.
(Sk,u,g)A = SA,u()
Z
C
C:ACC,
IC|=IA|+u
( PkSk+1,u-1,g)A
Z
=
= 1
sB,u-l(g
:
B,C:ICI=IAI+u,
B:|B|=|A+1,
IB|=IAI+1,
ACB
C
C
C:ACC,
|C|=A|+u
ACBCC
Definition 6.4.5. For any collection of vectors {Sku,g : k, u > 0}, for all k and all
U>
1, define the error vector
1
ek,ug = Sk,u,g
-
-Pk Sk+1,u-1,g
As we will show, to guarantee that we have an actual function g it is sufficient
to have all of the error vectors be equal to 0. However, we also need to ensure that
134
we get an e-invariant function for each e c E(G). e-invariance gives us another set
;> 0}. We repeat Corollary 3.4.28 here for
of equations on the vectors {ku,9 : k,
convenience.
Corollary 3.4.28.
e{w})g
w
-
w for some w E Vred(G) then g is e-invariant if and only if (e{}
+
1. If e = s
0. Equivalently, g is e-invariant if and only if
=
.vu{w}
= -V
whenever
V.
-
2. If e = v -+ t for some v E Vred(G) then g is e-invariant if and only if (e{}
efv})g = 0. Equivalently, g is e-invariant if and only if .VU{v}
v
= .v
whenever
V.
3. If e = v -
w for some v, w
C
Vred(G) then g is e-invariant if and only if
(e{} - e{v})(e{} + e{fw)g = 0. Equivalently, g is e-invariant if and only if
gvu{v,w}
+ vu{w} + v whenever v, w
= -vu{v}
V.
Lemma 6.4.6. If g is an e-invariantfunction for some e E E(G), A ; Vred(G), and
|=
k then
1. If e= s -
w for some w C Vred(G) and w C A then for all u,
(Skug)A -
2. If e =v
(s'k,u-1,g)A
t forsomev G Vrd(G) and v E A then for all u,
(Sk,u,g)A =
3. If e = v
-
-(k_1,u,g)A\{w}
-
(8
_.1,,,g)A5(vj -
(Sk,u_1,g)A
w for some v,w E Vred(G) and v,w G A then for all u,
(Sk-1,u,g)A\{v}
(Sk,4u,g)A
--
-
(-k-1,u,
1,u_1,g)Axv1 -
135
9
)A\fwj +
(Sk-1,u_1,g)A\{w
(Wk-2,u,g)A\{v,w}
+
(IS,u-2,g)A
Proof. To show statement 1, note that
E
=
E
B
+
(k-1,u,g)A\{w}
B: ACB,
|B|=|AI+u-1
YB
B: A\{w}CB,
wVB,|B|=|AI+u-1
E
-
= (Sk,u-1,g)A
YBU{w}
B:A\{w}CB,
w O B , B| =j A|+u -1
5YB
-
= (Sk,u-1,g)A
B:ACB,
IBI=IAI+u
= ('k,u-1,g)A
('k,u,g)A
-
Rearranging now gives the desired statement. Similarly, to show statement 2, note
that
S
B
+
5
-:
BBAC
B:AQB,
|BI=|A|+u-1
vIB,|B|=|AI+u-1
E
+
(Sk,u-1,g)A
v
YBU{v}
B:A\{v}CB,
B,|B|=|A|+u-1
(Sk,u-1,g)A
59B
+
=
YB
B:A\{v}CB,
B:ACB,
IBI=IAI+u
=
(Sk,u-1,g)A
+
(Sk,u,g)A
Rearranging now gives the desired statement. The proof for statement 3 is more
complicated but uses similar ideas. In particular, note that
S
(*k-2,u,g)A\{v,w}
B -+-
IBI=IAI+u-2
E
YB+
B:ACB,
B:A\{v}CB,
B+
B:A\{w}CB,
viZB,|B|=|A\+u-2
w-B,IBI=IAI+u-2
YB
I:
B:A\{v,w}CB,
v,wOB,IBI=IAI+u-2
Now let's consider these terms one by one.
1-
5
B = (8k,u-2,g)A
B:ACB,
IBI=IAI+u-2
2.
E
B:A\{v}CB,
v VB,IB I= IAI+u-2
YB
=
BB:A\{v}CB,
IBI=IAI+u-2
E
B:A\{v}CB,
vEB,|B|= AI+u-2
136
YB =
k-,u-1,g)A\{v}-
k,u-2,g)A
3.
I
9B =
B:A\{w}QB,
w i B,|B| IA +u-2
YB
=
(jBU{v,w}
+
YB-
k-,u-,g)A\{w}-Sk,u--2,g)A
B:A\{w}QB,
B:A\{w}CB,
wEB,IB|=|A|+u-2
IB=IAI+u-2
4. By statement 3 of Corollary 3.4.28,
YB =
.BU{v}
-
YBU{w})
B:A\{v,w}CB,
v,wOB, B|=|IA|+u-2
v,wOB,|BI=|A +u-2
5B
E
E
-+
B:ACB,
YB
B
-
B:A\{v,w}CB,
B:A\{v}CB,
B:A\{w}gB,
wVB,IBI=IAI+u
\BI=\Ai-+u
gvB,IBI=\AI+u-1
-1
We now have that
1.
YB
(kug)A
B:ACB,
IBI=iA|+u
2.
Y
YB
=
5
(8k-1,u,g)A\{w}-
iB|=|
=
(SBk-1,u,g)A\{w}-- (k,u-1,g)A
wEB,IBI=IAI+u-1
YB
=
(8k-1,u,g)A\{v}
E
-
3.
YB
B:A\{w}CB,
B:A\{w}CB,
A|+u -1
w$B ,
YB
=
(Sk-1,u,g)A\{v} -
('k,u-1,g)A
B:A\{v}CB,
B:A\{v}CB,
vOB,|B|=|AI+u-1
vEB,IBI=IAI+u-1
Putting these three statements together,
I:
YB
=
(Sk,u,g)A
+
( k-1,u,g)A\{w} -
(k--1,u,g)A\{v}
B:A\{v,w}CB,
v,wOB,|B =|AI+u-2
Putting everything together,
(Sk,u,g)A
(8k-2,Ug)A\{v,w}
-
+
(Sk-1,u,g)A\{w}
(Sk-1,u-1,g)A\{v}
+
Rearranging now gives the desired result.
-
(8k-1,u,g)A\{v}
(k-1,u-1,g)A\{w}
-
('k,u-2,g)A
El
For each possible edge v --> w, we define difference vectors which show far a
collection of vectors {Sk,u,g : k
> 0} is from representing a (v -+ w)-invariant
function.
137
;> 0}, define the vectors
Definition 6.4.7. Given a collection of vectors {kug : k,
{Akug,v-_w : k,u > 0, v, u)
1. If w - A then
Otherwise,
# wv # tw =
V(G),v
(Ak,,,gs*w )A = (S-k,u,g)A -
s,v
s
w
-4
((s'k,u-1,g)A -
-4
t}
as follows
(sk-1,u,g)A\{w})
(Ak,u,g,s-+w)A = 0
2. If v c A then
(Ak,u,g,v-+t)A = (Sk,u,g)A -
Otherwise,
= 0
(Ak,u,g,v-+t)A
((8k-1,u,g)A\{v}
-
(Wk,u-1,g)A)
3. Ifv,wEA then
(k,u,g,v-÷w)A
=
--
Otherwise,
( k,u,g)A
--
(Sk-1,u-1,g)A\{v}
(1Ak,u,g,v-*w)A
-
(Sk-1,u-1,g)A\{w}
+
+
(k-2,u,g)A\{v,w}
(sk,u-2,g)A)
= 0
Lemma 6.4.8. For any edge e = v
and v -+ w z
(k-1,u,g)A\{w}
(Qk-1,ug)A\{v} -
-
w where v C V(G) \ {t}, w c V(G) \ {s, w},
k U > 0} corresponds to an
s -+ t, the collection of vectors {SI, ,:
e-invariantfunction g if and only if the error vectors ek,u,g and the difference vectors
Ak,u,g,e are all 0.
Proof. The only if statement follows from Proposition 6.4.4 and Lemma 6.4.6. For
the if statement, we define g by taking
A =
(sAIo,g)A.
The fact that EjAIOge
=
0
now corresponds precisely to the criteria for e-invariance in Corollary 3.4.28 so g is
e-invariant. Using the fact that 'k,u,g = 0 for all k and u, we show by induction on
u that for all k, u and all A with JA = k, ('k,U, 9 )A
case u
=
= ZB:ACBB|=AI+U5B-
The base
0 is trivial. For u > 0, assuming the result holds for u - 1 and following the
same logic as before,
I(Pkk+1u-1,g)A -
SB,u-(g)
B:|B|=IA+1,
ACB
5
=C
B,C:IC|=|A|+u,
IB=|AI+1,
C
C:ACC,
ICI=IAI+u
ACBCC
Since
(ek,u,g)A = 0, (Sk,u,g)A =
(Pk4k+1,u-1,g)A
138
EC:ACC,IC|=|A|+uC,
as needed.
D
6.5
Freedom in choosing a base function
In this section we analyze what freedom we have in choosing a base function g which
we will use to choose our functions {ge}. In turn, this will determine what freedom
we have in choosing the coordinates of our vectors {sk,,}.
Recall the graph H' used in the proof of Theorem 3.5.1. Adapting the definition
of H' to our current probelm, let H' be the graph such that
1. V(H')
{v' E V(G'(V(G)))
: v' has knowledge set K E {Kv : V
C Vred(G),
IVI < r}U {Kt}}
{(u', v') : u', v' c V(H'), there is an edge in G'(V(G)) between u' and
2. E(H')
V' with label in E(G)}
By the reasoning used in the proof of Theorem 3.5.1 we may choose the values of g-Kv
on each connected component of H' except for the connected component containing
t' where we must have that g - Kt,
=
0. We now determine what these connected
components are by finding a representative for each connected component of H'.
In doing so, our reasoning is very close to the reasoning about reachable pebbling
configurations in Li and Vitanyi [151 and in Li, Tromp, and Vitanyi [141.
Definition 6.5.1. Given a fixed r and a set of vertices V C Vred(G), we say an
w is relevant for V if it is from a vertex v E V U {s} to a vertex
edge e = v
-4
w C V U {t}
\ {v}
and there is a path in G from v to w of length at most 2 r-IVI-1
Definition 6.5.2. For a fixed r, we say that a set of vertices V C V(G) \ {s} is
irreducible if V = {t'} or V
C
Vred(G) and there are no relevant edges for V.
Theorem 6.5.3. Each connected component of H' has exactly one vertex v' such that
if V is the set of vertices such that K, = Ky (where we take KtI = Kt}) then V is
irreducible.
Proof. If V
C
Vred(G)
and there is a relevant edge e = v
w for V where w
-
$
t
then we may go from Ky to KV\{w} along edges in H'. This effectively reduces V to
V \ {w}. If V
C
Vred(G)
and there is a relevent edge e = v
139
-
t for V then we may
go from Kv to Kt, along edges in H'. This effectively reduces V to {t}. We will now
show that the process of reducing a set of vertices V
C
Vred(G) until we obtain an
irreducible set of vertices U always yields the same U.
Lemma 6.5.4. If there is a relevant edge e for V of the form v -s t then when we
reduce V to an irreducible set of vertices U, we always obtain U = {t}.
Proof. If we ever use an edge v
-
t which is relevent, we will reach U = {t}. Thus,
it is sufficient to show that no matter what other reductions we do, we always have a
relevant edge v
-+
t for V. To show this, note that if we have a relevant edge v
-
t
for V, the only way to make this edge irrelevant is to remove v from V. We can only
do this by reducing V with a relevant edge of the form u
-
v where u E V U {s} \ {v}.
However, once we do this, since we reduce the size of V by 1 the edge u
->
t must
now be relevant so we still have a relevant edge u -+ t for V.
E
Lemma 6.5.5. In the process of reducing V by relevant edges, if we can reach an
irreducible set of vertices U by starting with one relevant edge, then we can reach U
regardless of which relevant edge we start with.
Proof. We prove this result by induction on JVJ. The base case IVI
Assume that there is a relevant edge v 1
-+
=
0 is trivial.
v 2 such that if we start with v, -+ v 2 we
may reach U. We show that for any other relevant edge v 3
-+
v 4 , we may reach U if
we start with it. To show this, we first show that if we start by reducing V to V
using vi
->
\{v2 }
v 2 , we can then reduce V \ {v 2} to V \ {v 2 , v 4} by using another relevant
edge. Since we can reach U from V \ {v 2}, by the inductive hypothesis we can reach
U regardless of which edge we use next from V \ {v 2 }, so we must be able to reach
U from V \ {v 2 , v4 }. We then show that we can also reach V \ {v 2, v4 } by starting
with the relevant edge v 3
-
v 4, so we can reach U by starting with the relevant edge
v 3 -+ v 4 , as needed.
The proofs for these two parts are below. Note that we do not have to consider
the case where v 2 or v 4 equals t because in this case by Lemma 6.5.4 we always end
up at {t}.
140
1. If we start by using the edge v,
the edge V3
-*
v 2 to reduce V to V \ {v 2 }, we can then use
-+
v 4 to reduce V \ {V2} to V \ {v 2 , V4} unless v 3 = V2. However, in
this case, the edge v, to V4 is relevant for V
\ {V2}
so we can use it to reduce
V \ {v 2} to V \ {v 2 , v 4 }. Thus, we can always reach V \ {v 2 , V4} by starting with
v1
--
v 2 . Note that we cannot have vI
= 04
because G is acyclic.
2. Following similar logic, if we start by using the edge v 3
V \ {V4}, we can then use the edge v
->
V2
-
V4
to reduce V \
{V4}
to reduce V to
to V \ {V2,
unless vl= V4. However, in this case, the edge v 3 to V2 is relevant for V \
so we can use it to reduce V \
V\
{v 2 , V4} by starting with V3
to V \ {V2,
{V4}
-+
V4.
V4}.
V4}
{V4}
Thus, we can always reach
Note that we cannot have v 2 = v 3 because
G is acyclic.
Corollary 6.5.6. Regardless of how we choose to reduce V, we always end up at the
same irreducible set of vertices U.
Proof. Lemma 6.5.5 implies that the possible sets of vertices we end with is unchanged
by our first reduction and thus by any reductions that we do. However, once we reduce
to some irreducible set of vertices U, U is the only possible set of vertices to end with.
Thus, we must have always had that the only possible end set of vertices was U, as
needed.
E
We now prove Theorem 6.5.3 using Corollary 6.5.6.
Corollary 6.5.6 says that
we may define a mapping R which takes each set of vertices V and maps it to the
irreducible set of vertices that it reduces to.
Proposition 6.5.7. For all V, R(V) is in the same component of H as V
Proof. This follows from the fact that every reduction step can be carried out using
a path in H'.
Corollary 6.5.8. Every connected component of H has at least one irreducible set of
vertices.
141
Proposition 6.5.9. If V and W are adjacent in H then R(V) = R(W)
Proof. If V is adjacent to W in H then either we can obtain V from a reduction step
on W or vice versa. Without loss of generality we may assume that we can obtain
V from a reduction step on W. This immediately implies that R(W) = R(V), as
D
needed.
This means that R is invariant on all connected components of H, so there cannot
be a connected component of H with two irreducible sets of vertices U1 and U2 as
otherwise we would have R(U1 ) = U1 and R(U2 )
=
U2 which would mean that R
would not be invariant on this connected component.
E
By Theorem 6.5.3 and the reasoning used in the proof of Theorem 3.5.1, we
may freely choose the Fourier coefficients (P)v where V is irreducible and this will
determine the other Fourier coefficients
v where IV| < r. We now analyze what
freedom this gives us in choosing the coordinates {Sk,,g,}.
Definition 6.5.10.
We say that e = v -4 w is relevant for ('k,u, g )v if v G V U {s},
w E V U {t}, and there is a path of length at most 2 r-IVI-u-1 from v to w in G.
Definition 6.5.11.
1. We call a coordinate ('ko,g)v free if IVI = k < r and V is irreducible. Otherwise,
we say that (Sko,g)v is fixed.
2. We call a coordinate ('k,u, g )v free if there are no relevant edges for ('k,u,g)v.
Otherwise, we say that (Sku,g)v is fixed.
We now show that we may choose the free coordinates ('k,o,g)v and then determine
all of the other coordinates (Sku,g)v accordingly.
Theorem 6.5.12. If we choose the coordinates {(Sku,g)A} in increasing order of k+u
up to k + u = r - 1 where coordinates with equal k + u are chosen in increasing order
of u then if:
142
1. We choose the fixed coordinates {(4,u,)A} by choosing any relevant path v
for
(Sk,,,,g)A
and setting
(Ak,u,g,v-w)A
-+
w
= 0.
2. We choose the free coordinates { (,O,g)A}
arbitrarily.
3. We choose the free coordinates {(Sk'u,g)A} where u > 0 so that ( K,u,)A
=
0
then
1. For all edges e e E(G) there exists a function g, with these coordinates which
is e-invariant.
2. For any function g with these coordinates, for all V such that |Vt < r, g -Ky
=
g * KRMv)
Proof. We prove this theorem by induction on k + u. The base case k + u = 0 is
trivial. Assume that we have already chosen the values for k + u <
V such that IVt <
j, g -Kv
j
and that for all
= g -KR(v). Now consider the coordinates with k +u =
First assume that we choose the coordinates {(Sk,o,g)V
: |VI
j.
= j} so that for
any g with these coordinates, g - Ky = g - KR(v) (this choice is arbitrary for the
free coordinates). For any e = v
-+
w such that there is a path of length at most
2r-3-1 from v to w in G, whenever IVI < j -
1 and v E V U {s} we have that
R(V U {w}) = R(V) so g - Kvu{w} = g - Ky = g - KR(v). Using the same reasoning
used to prove Theorem 3.5.1, there is a function g with these coordinates which is
e-invariant.
For each such e, choose such a function g,. Now note that since g, is e-invariant,
(Ak,u,ge,e)V =
0 for all k, U, V. In particular,
implies that if we choose these coordinates,
(Ak,u,g,,e)V
=
(Ak,u,g,e)v=
0 whenever k+u < j. This
0 whenever e is relevant for
(Sk,u,g,e)V.
Now note that when we choose a fixed coordinate ( k,o,g)v, we set (Eko,,,-_w)v
where v
--
0
w is relevant for (sKo,g)v. There is a unique (Sk,o,g)v for which this is true.
From the above, we know that choosing (Sk,o,g)v so that g - Ky = g - KR(v) makes
(Ak,o,g,v_w)y =
all IV I
j, for
0. Thus, this choice must be our actual choice. This implies that for
all g with these coordinates, g -Kv
143
=
g - KR(v).
At this point, we just need to show that the choices for ('k,u,g)v where u > 0
correspond to our functions {gej, which we show with the following lemma.
Lemma 6.5.13. For all of the functions g, described above, if we choose the coordinates {(S,,g)v} as described above then ('k,u,g)v
= (s,u,g,)v whenever k + u < j.
Proof. We prove this by induction on u. Note that all of the functions {g,} have the
same coordinates for k + u < j so we just have to prove equality for a single e. The
base case u
=
0 is the fact that all of these g, have the coordinates {(Sko,g)v}. For
the free coordinates
that (,)
u,g)v
with u > 1, since we are choosing these coordinates so
(S4ku,g)v
0v =(0P(k+1,u-1,g)v = (Pk
k+1,u_1,g,)v = (su,g,)v. For
the fixed coordinates (k,g)v with u > 1, we chose (sk,u,g)v so that (A,,,,e)v = 0
for some e which is relevant for (sk,u,g)v. Now note that
(Ak,u,ge,e)v
the inductive hypothesis, ('k,,g)v = (sku,g)v whenever IVI +u <
that
(aku,g)v
= (SkU,g)V,
= 0 as well. By
j, so
we must have
El
as needed.
l
This completes the proof of Theorem 6.5.12.
Theorem 6.5.12 says that we may choose the coordinates (sk,u,g)v in increasing
lexicographic order of k + u and u and may choose the free coordinates
arbitrarily. However, this does not allow us to control the vectors Skug for
To control the vectors
,
(Sko,g)v
u
> 1.
for u > 1 we now show that we may instead choose the
coordinates ('ku,g)v in increasing lexicographic order of k + u and k and choose all
of the free coordinates (,u,g)v
as long as we make sure that the error vectors are 0
on the free coordinates (they will automatically be 0 for the fixed coordinates).
Corollary 6.5.14. If we choose the coordinates {(k,,g)v} in increasing order of
k + u up to k + u
=
r - 1 where coordinates with equal k + u are chosen in decreasing
order of u then if:
1. We choose the fixed coordinates { Sug)v} by choosing any relevant path v
for
(Sk,u,)A
and setting
(Akug,'vW)A
=
-4
w
0.
2. We choose the free coordinates {(4,,g)v} so that ('k-1,u+1,g)v = 0 whenever
(Wk-1,u+1, 9 )V
is free.
144
then
1. For all edges e G E(G) there exists a function g, with these coordinates which
is e-invariant.
2. For any function g with these coordinates, for all V such that |V| < r, g -Kv
g - KR(v)
Proof. This is exactly the same as Theorem 6.5.12 except that we are choosing the
coordinates in decreasing order of u rather than increasing order of u. This has no
effect on our choices for the fixed coordinates as all of our choices only depend on
coordinates ('ku,g)v where k +u < j. For the free coordinates, note that in Theorem
6.5.12 the free coordinates (sko,g)v are defined by the coordinates (8k,o,g)v (which
are arbitrary) and the requirement that (sk-1,u+1,g)v
0 whenever (k-lU+l,
9
)V
is
free. As long as these properties are preserved, the order in which we choose the
free coordinates (Sku,g)v does not matter. Thus, we may choose these coordinates in
decreasing order of u as long as we ensure that (Sk-1,u+1,g)v
=
0 whenever (sk-l,u+l, g )v
E
is free.
6.6
Building up a base function
Corollary 6.5.14 allows us to construct a base function g while having considerable
control over all of the sum vectors {Sk,g,}. This is extremely useful because we will
try to make II,,ug I small for every k and u. However, in order to use Theorem 6.5.12
we need a way to choose the free coorindates of the sum vectors {
kug}
to make these
vectors small while also making the error vectors 0. In this section, we show how to
do this. For this section, we are always working with collection of sum vectors {k,u,}
where we have already decided which coordinates of these vectors are fixed and which
coordinates of these vectors are free.
Definition 6.6.1. For each k, U,
1. Define
l7k,ufixed
to be the projection which projects any vector in the same vector
space as Sk,,,, onto the fixed coordinates of Sk,u,g.
145
2. Define rk,u,free to be the projection which projects any vector in the same vector
space as Sk,u,g onto the free coordinates of sk,,,.
3. Define Sk,u,g,fixed
=
4. Define Sk,u,,free
=Fk,u,free(sk,u,g)
Proposition 6.6.2.
7k,u,fixed(Sk,u,g)
S'k,u,g
=
Sk,u,g,fixed
+
Sk,u,,,free
Proposition 6.6.3. The free coordinates of the error vectors are all 0 if and only if
Sk,u,g,free =
Wk,u,free
(Pksk+1,u-1,g)
U
whenever u > 0
Definition 6.6.4. Let Pk,u,fr,,ee be the matrix Pk except that we make all of the rows
corresponding to a set of vertices A such that
(Sk,u,g)A
is fixed 0 and we make all of
the columns corresponding to a set of vertices B such that (Sk+1,u-1,g)B is fixed 0.
Definition 6.6.5. When u > 0 define the adjustment vector
ak,u,g
Proposition 6.6.6.
7
~-k,u,free(Pk Sk+1,u-1,g,fixed)
+
k,u,free(Pks'k+1,u-1,g) = -k,u,g
Pk,u,freeSk+1,u1,gfree
Corollary 6.6.7. The free coordinates of the error vectors are all 0 if and only if
Pk,u,freeSk+1,u-1,g,free = ak,u,g
+
Usk,u,g,free
whenever u > 0
Proof. This follows immediately from Proposition 6.6.6 and Proposition 6.6.3.
If we let P be the matrix
Pk,u,free
E
after we delete all the rows and columns
corresponding to fixed coordinates (which were all 0 rows and columns by definition), let X' be the vector dkn,g + USk,u,g,free after we delete all fixed coordinates
146
of -k,ug (which were also all 0) and let ' be the vector sk+1,u-1,g,free after deleting all fixed coordinates of Sk+l,u-1,g (which were also all 0), then we now have the
matrix equation PF? =
If P does not have full row rank then it may be im-
?.
possible to find a ' such that Py' =
. We will take care to avoid this case by
showing that PPT has no zero eigenvalues. If P does have full row rank then we
want to find the Y with smallest norm such that P/ = ? and then find a bound on
Sk+1,u-1,g,free1| =
we can get is
MY
2
1Y 2
in terms of
- USk,u,g,free 2 -
Hakd,,
The best bound
2.
{
; maxsyj{minv
We have now reduced our problem to a problem of the following form. For a given
real matrix P with full row rank, obtain bounds on maxyoo{ming.p#g
{
}}
Lemma 6.6.8. If P is a real matrix with full row rank then letting { A} be the
eigenvalues of PTP,
2
2
.r
max
m mmg_x+O 1 y:Py=x
max {-}
}}= i:,\i=4o
1:#0k Ai
Proof. Note that PT has full column rank, so PI = X if and only if pTp.
-
PTE
Thus, for a given i we are looking for the ' with smallest norm such that PTPYI
PT-X.
Since pTp is a positive semidefinite real symmetric matrix, if
j
is the number
of columns of P then there is an orthonormal basis of Ri consisting of eigenvectors
el, ..
- -ej
of pTp with eigenvalues Ai,
Proposition 6.6.9. If PT
-
=
, Aj. We now write pT
cie' then ci
=
=
_ cii.
0 for every zero eigenvector ei of
pTp
Proof. Note that if pTP
= 0 then -pTpi = Pei - Pei = 0 so Pei = 0. But then
eWPT=c = X-T P- =0
E
Writing
=
Zi=
iei and solving the equation PTP= pTj
pTp -pTp(d.
=
i=1
Zidis
i=1
147
=
Zcis
i=1
we obtain that
This says that di =-
whenever Ai # 0 and di can be anything when Ai =0 (as ci = 0).
To minimize the norm of ', we take y= ZI
2 = SpTpO
=
Ei:AiO
$ ge'. Now
2
2 and
(C,)2
Comparing term by term, we have that
ZA.O
(A)2
(maxA=o
{&})
Z
1
and this inequality is an equality when ci is nonzero for this particular i and 0 for all
other i, so maxj:o{minVppy
}
=
maxo
{y}
Thus, we are interested in the smallest nonzero eigenvalue of pTp. Since P is real
with full row rank, the eigenvalues of ppT are precisely the nonzero eigenvalues of
pTp, so we can find the smallest nonzero eigenvalue of pTP by finding the smallest
eigenvalue of ppT.
We give an exact answer when P =
Pk
and give bounds for many other P.
Theorem 6.6.10. If k < 1 then PkPi has eigenvalues (n - k - i)(k + 1 - i) with
multiplicity (n)
-
(i_2) for i E [0, k]
Proof. Given V C Vred(G) where
lVi
< k, let s'v be the vector such that for any
A C V if B C Vred(G), |B| = k, and BnV = A then ('v)B =
(-)|k,
5
where BnV = A have total weight
In other words, we make the coordinates ( V)B
(-l)IAI
where the weight is spread evenly among these coordinates.
Lemma 6.6.11. If V
C Vred(G)
and |V|
=
m < k then 5 v is an eigenvector of PkPk'
with eigenvalue (n - k - m)(k + 1 - m)
Proof. Consider where the weight is in PkPkx'v. Letting m = JVJ, the coordinates
where IBn VI = j have total weight (-i)i().
Multiplying by PkP[ multiplies
and shifts this weight as follows. For a given B with IB n Vl =
m+
j)j ways to
j
there are (n - k
-
(Xv)B
add a vertex to B then remove a vertex and obtain a set of vertices
C such that IC n VI = j - 1. There are (m - j)(k - j) ways to add a vertex to B then
remove a vertex and obtain a set of vertices C such that IC n VI = j + 1. Finally,
since there are (n - k)(k + 1) total ways to add a vertex to B and then remove a
vertex from B there must be (n - k) (k +1) - (n - k - m+ j)j - (m - j)(k - j) ways
to add a vertex to B then remove a vertex and obtain a set of vertices C such that
148
C n V=
j.
Thus, from the original total weight of (-I)i(') on the coordinates
( V)B where B n Vl = j, we get
1. A total weight of (-1)i(7)(--k -m+j)j on the coordinates (PkPkj 'V)c where
|CnV= j-l
)(m - j)(k -j) on the coordinates (PkP[?v)c where
2. A total weight of (-1)j
|CfnV =j+1
((n - k)(k +1) - (n - k - m +j)j - (m - j)(k - j))
3. A total weight of (-1)()
on the coordinates (PkP,[yv)c where |C n V = j
Turning this around, in PkPTzv we have the following contributions to the total
weight of the coordinates (PkPkTxj)c where IC n VI = j:
1. A contribution of (-1)j+'
(TV)B
where B n VI
(iN) (n -
j+
k - m+ j+ 1)(j + 1) from the coordinates
1
2. A contribution of (-1)--' (g)(m-- J+1)(k - j+1) from the coordinates ('v)B
where IBnVl=j-1
3. A contribution of (-1)j(')((n - k)(k + 1) - (n - k - m+j)j - (m - j)(k - j))
from the coordinates ('V)B
where B
nV
=
j
Summing these contributions together, we get the following total weight for the co149
ordinates (PkPk,7jv)c where |C n V| = j
m
+ (-1)
(((n-k)(k+1)-(n
+
\\j)
-
j+1)(j+1)-
+ I(n -k-rn
:+1m-+
(m
-
j+
1)(k - j + 1)
-k-m+j)J-(m-j)(k-j))
(_)
(-)
rn
(n - k - m + j + 1)(j + 1) + (1)3-1)(
Ir+ 1)(kr--j
+(m-j+1)(k--j+1)
+
(-i)i(
+(n--k)(k+ 1)- (n-k-m+J)j-(m-j)(k-j))
=
+ (n-k)(k+1)
=(-1)()(
j+
j)(n - k - m +j + 1) - (n - k - m + j)j - j(k -
(-1)-
1)
(m-j)(k- j))
-m(n--k-m+j
+1)+j-jk+J2 -j+nk+n-k
2
-k-mk
+jk+jm-j2)
=(-1)j (7)(rmn+M2
=
(-1)
-rm+nk+n-k 2 -k)
( (n-k-m)(k+1 -im)
By symmetry the weight on the coordinates (PkPkiv)c where IC n VI =
j
will be
+
spread evenly. Thus, 'v is an eigenvector of PkP[ with eigenvalue (n - k - M) (k
D
1-rm) where m = IVI, as needed.
Now we need to check that the vectors {v : IVI = m} span the eigenspace of
PkP[ with eigenvalue (n - k - m)(k +1 - m) and that these are the only eigenvalues.
We also need to find the dimension of these eigenspaces.
Definition 6.6.12. For each V C Vred(G) where IV| < k, take /v to be the vector
on subsets B C Vred(G) of size k where (fv)B
=
1 if V C B and 0 otherwise.
Lemma 6.6.13. For all j < k, the vectors {Yv
V|
=
are a basis for the vector
space spanned by the vectors {V : V| - j}
Proof. We first check that each 'y with V
vectors { v: IV <j}.
150
=
j is
in the vector space spanned by the
Lemma 6.6.14. For any V C Vred(G), the vectors {A:ACVIAj
are a basis for the vector space of vectors & such that ()A
C
XA
only depends on
[0, V|]}
|A
V|.
Proof. Note that all of these vectors are nonzero and live in different eigenspaces
of PkP[, so they are all linearly independent.
(EA:ACV,IAI=j
XA)B
only depends on |V
By symmetry, for all j c [0, VI],
n B| so all of these
vectors are in this vector
space. This vector space has dimension IV + 1 so these vectors must be a basis for
this vector space, as needed.
D
Corollary 6.6.15. For all V C Vred(G) of size at most k, ,v E span{ZA.ACVIAj
j
aE[0, Vl]}
We now check that the vectors {
the vectors {y : VI <
VI =
j} span
the vector space spanned by
j}.
Lemma 6.6.16. The span of the vectors {v : V C Vred(G),IV I= j} contains all of
the vectors {sy
j}
V C Vred(G), V|
Proof. For all V C Vred(G) where VI
i c [0, V
1}.
j, consider the sums {ZA:ACVIAnvI=i
YA
By symmetry, the values of these sums on a coordinate B will depend
only on IB n VI. Moreover, this sum is 0 on a coordinate B if IB n VI < i and
nonzero if IB n VI = i. Thus, if we take a linear combination of these sums by
choosing the coefficients in increasing order of i, each choice determines the value on
the coordinates B where IB n VI = i and does not affect the value on the coordinates
B where B n VI < i, so we may arbitrarily determine all of these values. This implies
that s'v is a linear combination of these sums, as needed.
Finally, we check that the vectors {V
independent. First note since k <
the vectors {
,
pT k has
: V
C
Vred(G),VI
D
=
}
are linearly
full column rank. To see this, note that
:V = k} are a basis for the vector space which PkPkj acts on. From
the above, this vector space is spanned by the vectors {V : VI < k} so the only
eigenvalues of PkP are {(n - k - m)(k + 1 - m) : m E [0, k]}. In particular, PkPkT
has no zero eigenvalues so
PTk
has full column rank. Now assume that there are
coefficients {cV : V C Vred(G), Vl
=
j} such that ZV:VVre(G),IVI=j cvv = 0. If we
151
let ' be the vector with value cv in every coordinate V,
since k < a2 all of the matrices Pk-1'
k-1
-
PC
0. However,
PT have full column rank and thus '= 0.
I'
All cv are 0 so the vectors {f/v : V C Vred(G), V
-
j}
are linearly independent, as
needed.
E
Using Lemmas 6.6.11 and 6.6.13, for every
j, the
sum of the dimensions of the
eigenspaces of PkP[ with eigenvalues (n-k-m)(k+ 1-rm) where m < j is equal to the
number of vectors {yV : V C Vred(G), IVI = j} which is ('). This immediately implies
that the dimension of the eigenspace of PkPkj with eigenvalue (n - k - j)(k + 1 - j)
is
(n)
-
E
(j).
For our results we are not dealing with the matrix Pk itself but rather matrices
which are close to Pk. We now bound the smallest eigenvlue of such matrices.
Theorem 6.6.17. Let P be a matrix obtained from Pk as follows: Take a collection
of bad vertices V and bad pairs of vertices E. Delete the rows and columns containing
bad vertices and pairs of vertices from
Pk.
If
1. |V| < n
2. There is an m <
vertices w
2QgJok
such that for every vertex v
V, there are at most m
V with {v,w} c E
then the smallest eigenvalue of ppT is at least n2
Proof. The proof idea is as follows. The smallest eigenvalue of ppT is miny {I
12}
This quantity can only be reduced when we subtract a symmetric positive semidefinite
matrix M from PpT. Thus, we will choose M so that M is a symmetric positive
semidefinite matrix and ppT
-
M has a large minimum eigenvalue, and this will give
the claimed bound. We construct M as follows:
Definition 6.6.18. Let U, W be subsets of Vred(G)
\V
such that |UI = k - 1 and
U n W = 0. Define Mu,w so that MU,w has the same rows and columns as ppT and
1. If no pair of vertices in U U W is bad then (MU,W)A 1 A2 = 1 if U C A 1 C U U W
and U C A 2 C U U W and 0 otherwise.
152
2. If a pair of vertices in U U W is bad then Mu,w = 0
Lemma 6.6.19. If we take M =
EU,W:U,WCVred(G)\V,
|U=k-1,|W|=j,UnW=0
MUw and have |A 11 = |A 2 | = k
then
1. If A 1 U A 2 has no bad vertices or pairs of vertices and A 1 n A 21 = k - 1 then
n -|JV|
2mj( j + k)
- k - 1
<_~A
-|V|
_n
j-
j -2
n
- k - 1
2
2. If A has no bad vertices or pairs of vertices then
+ k))k n -- V - k
(-2mJ(j
MAA
k n -V
I-
- I
n
- k
I
3. If A 1 or A 2 has any bad vertices or pairs of vertices, MA 1 A2 is undefined. Otherwise, if A 1 U A 2 has any bad pairs of vertices or JA1 n A 2 | < k - 1 then
MA 1 A 2
=
0
Proof. Statement 3 follows immediately from the definitions. The upper bounds in
statements 1 and 2 follow by noting that these are the maximal possible number of
Muw which have
(MUW)A
1
A2
or
(MUW)AA
equal to 1. For the lower bounds, for any
A 1 , A 2 such that A 1 U A 2 has no "bad" vertices or pairs of vertices and IAinA 21 > k-1,
take random U, W with U, W
C
Vred(G) \ V, |UI = k - 1, JW| = j, U C A 1 C UU W,
UnW = 0 and U C A 2 C UU W. To do this, first randomly choose U. After choosing
U, start with Wo = A 1 U A 2 \ U and add vertices to W one at a time. We add at
most j vertices and the probability that each new vertex adds a bad pair of vertices
is at most
m(j+k)
n-IVj-(k+j)
Thus by the union bound the probability that (MU,W)A
1 A 2 is
(U)l2i
mj&j+k)
0 rather than 1 is at most n-IVI-(k+j)
<
2mj(j+k)
n4
-
as IV| < i and kj << n.
El
Corollary 6.6.20. If we instead take
M
_n-_V-k-) ZU,W:U,WC Ve(G)\VIUI=k-1,w|=j,Unw=0 MUw
=
1. If A 1 U A 2 has no bad vertices or pairs of vertices and
0
< (ppT
-
M)AM)AjA2
1 A 2 <-
2mj(j+k)
n
153
then
|A1 n A 21 =
k - 1 then
V|
-
2. If A has no bad vertices or pairs of vertices then (ppT - M)AA > n - k km - k(n-k-IVI)
1
j-
3. If A 1 U A 2 has any bad vertices, (ppT
M)A 1 A2 is undefined. Otherwise, if
-
A 1 UA 2 has any bad pairs of vertices or |A 1 nA 2 | < k-1 then (PPT -M)AA
2
= 0
-
Proof. This follows immediately from Lemma 6.6.19, the fact that (PPT)AA > n
k - IVI - km whenever A has no bad vertices or bad pairs of vertices, and the fact
that
VI
n-
-1
j
(n-IVI-k-1)
-
-
k
1
The final step is as follows. For each column, we subtract all of the non-diagonal
elements from the diagonal element and then set all non-diagonal elements to 0. This
will not increase the minimum eigenvalue as
(1
0
is a positive semidefinite matrix. Note that each column has at most k(n - k - IVI)
nonzero non-diagonal elements, so by Corollary 6.6.20, all of the diagonal elements
are still at least
n- k
-
V|
-
km
-
k(n-k-IVI)
1
J-
_
2mj(j+k)k(n-k-IVI)
n
k-VI)_ 2mj(j+k)k(n-k-IVI)
n then n-k-|V|-kmTaking j = 8k, if m <
- 2000k0 then n
Thus, the minimal eigenvalue of ppT is at least
n,
as claimed.
>
2
E
Corollary 6.6.21. If G is an input graph such that
1. There are at most 14 vertices v such that there is a path from s to v or a path
from v to t in G of length at most 2'-1
2. For any vertex v E Vred(G) there are at most 2O3
W G Vred(G)
such that there
is a path from v to w or a path from w to v in G of length at most 2 r2
154
then for all k, u such that k + u < r, given any Sk,u,g,free and aku,
Sk+,u-1,g,free
we can choose
so that
1-
PkSk+1,u-1,g,free
2.
Sk+1,u-1,gfree
I I J ,12 <4U2
11 12
+411- 2
USk,u,g,free+ ak,u,g
12
2
USk,u,g,free +
ak,u,g
Proof. Recall from before that the equation
-
n
-
Sk,ugfree
Pkk+1,u-1,g,free
k,,2
Usk,u,g,free
+ akug
is
equivalent to the equation Pj = i where P is the matrix Pk,u,free after we delete all the
rows and columns corresponding to fixed coordinates, - is the vector
after we delete all fixed coordinates of
and gyk,u,g,
is the vector
dk,u,g+tsk,u,gfree
Sk+1,u-1,g,free
after
deleting all fixed coordinates of Sk+1,u-1,gNow for a given k and u, let V be the set of vertices V such that s
is relevant for (Sk,u,g)A when v
w
-
E
6.7
v or v -+ t
A. Let E be the set of (v, w) such that v -+ w or
v is relevant for (Sk,u, 9 )A when v, w
6.6.17 we may take ' so that
-
c
A. Applying Lemma 6.6.8 and Theorem
yfl I I -I|I and the result follows.
D
Constructing a base function
We now use the tools we have developed to construct our base function g.
Theorem 6.7.1. If r and m are constants such that m
<
2000r4
and G is a directed
acyclic input graph such that
1. There is no path of length at most 2'-1 from s to t.
2. For any vertex v
e V(G), there are at most m vertices w E V(G) such that
either there is a path of length at most 2r-2 from v to w in G or there is a path
of length at most 2 r-2 from w to v in G
then there is a function g such that
1. yf
2.
= 1
'v = 0 whenever
|V|
> r
155
3. For every e E E(G) there is a function g, such that
(a) ge is e-invariant.
(b)
4.
4 ev
= Yv whenever
(9mn)
Isk,u,gI2
|V| <
r
2
Proof. We construct this function g by using Corollary 6.5.14 and Corollary 6.6.21 to
choose the vectors sk,u,g in increasing lexicographic order in (k + u, k). For k + u = 0
we take
=
1I0,Og
and for k + u
{}
j
=
> 0 we start with s0 ,j,g = 0. Corollary
6.5.14 guarantees that when we do this, every statement except the last one will hold,
so we just need to show that skU,g2 5 (9mn)
2
. This follows immediately from the
following lemma.
Lemma 6.7.2. If k + u > 1 then
<
1. ISku,g,fixedI' K(9mn)
2.
Sku,g,free2 <
!(9mm)
2
k
Proof. We prove this by induction. We prove bounds on each term in terms of previous
terms and then verify that if the bound holds for the previous terms it holds for the
current term as well.
Lemma 6.7.3. For any k > 1 and any u,
Sk,u,g,fixed112
< 2s
_,u_1,g 12
+ 61 [
k2g
Proof. By our definitions, for every term
(4m + 6km) IIu',
9
+ 6km|Sekl,u_1,g
2+
2 +
(Sk,u,g)A
,12
3nm|k- 2,u,g| 2
which is fixed one of the following
is true
1. w E A and there is a relevant path from s to w. In this case,
-
(sk-1,u,g)A\{w1
2(('k1,u,g)A\{w})
This implies that
2
156
((Sk,ug)A)
2
< 2((s',_1,g)A) 2
+
(Sk,u-1,g)A
=
(S'k,u,g)A
-
(Wk_1,ug)A\{v
(_k,_u1,g)A
implies that
This
In this case,
((4k, ,,)A)
2
=
('k,u,g)A
2(Sk_1, ,,)A\{V1)
2
+
2. v E A and there is a relevant path from v to t.
2(Sku--,g)A)2
3. v, w E A and there is a relevant path from v to w. In this case,
= (Sk_1,u,g)A\f{v
(Sk,u,g)A
-
-
(8_k-1,u,g)A\{w
-
(Sk-1,u_1,g)A\{v
-
(Sk-1,u_1,g)A\{w
(Sk-2,u,g)A\{v,w}
+ (k,u-2,g)A
This implies that
6((k-1,u,g)A\{w})2
+ 6((k-2,u,g)A\{v,w}) 2(8k,u,g)A
2
- 6((k-1,u,g)A\{v})
+
k
6(('k-1,u_1,g)A\jv1 ) 2
+
6((k_1,_1,g)A\fw
)2
+ 6(('k,u-2,g)A)
To bound I|Sk,u,gfixed 12 we sum the above inequalities over every A such that
2
(Sk,,,g)A
is fixed. Consider how many times each term can appear in this sum. Terms of the
form
((1k,
1,g)V)
2
and
(('k,.-2,g)V)
2
will only appear once (with the appropriate
constant in front) as they only occur when A = V. Terms of the form ((Sk_1,,,
g
)v)2
will appear up to 2m times from cases 1 and 2 (these terms can occur when A = V Uv
where v is (r - k - u - 1)-linked to s or t) and up to km times from case 3 (these
terms can occur when A = V U v where v is (r - k - u - 1)-linked to a vertex in A).
Terms of the form
((Sk-1,u-1,g)A)
2
cannot occur from cases 1 and 2 but will appear
up to km times from case 3 (these terms can occur when A = V U v where v is
(r - k - u - 1)-linked to a vertex in A). Finally, terms of the form
appear up to
n- 2
((8k-2,u,g)A)
2
will
times (these terms can occur when A = V U {v, w} where v and w
are (r - k - u - 1)-linked to each other). Putting everything together,
Sk,u,g,fixed 112
< 21 Sku_1,g 12 + 611
k,1-2,g|2
+ (4m + 6km)|IIk _1,u,g 12
+ 6km|Skl,u_1,g
2
+ 3mSk-2,u,g
F157
Corollary 6.7.4. For a given k > 1 and any u, if the bounds hold for earlierterms
then ||Sk,u,g,fixeds|2 < I(9mn)
Proof. If k = 0 then Sk,u,gfixed = 0. By Lemma 6.7.3 and the bounds of Lemma 6.7.2,
2(9mn)
k+u-1
2
6km(9n))
2
<
+ 6(9mn)
_ __
22
+
-v/9mn
k+u-2
2
+ (4m + 6km)(9mn)
k+u-1
2
) k+ u - 2
2
+ 3nm(9mn)
6
100k 2 m 2
9mn
9mn
+
2
Sk,u,g,fixed 2
Since n > 2000mr4 this is less than - (9mn)
+
6km
1
+9mn3
k+u
(9mn
2
2-u
Lemma 6.7.5. For any k > 1 and any u,
k-
1,u+1,g 2
< m(k + 1)k Ik,u,g,fixed
2
Proof. Recall that ak-1,u+1,g is the projection of -Pisk,u,g,fixe
of Sk-1,u+1,g. Now
For the A such that
is free, there are at most (k - 1 + 2)m = (k + 1)m B such that A C B,
(Sk-1,u+1,g)A
= k, and
=BI
(Pk-1Sk,u,g,fixed)A = ZACBB~k ('ku,g,ixed)B.
onto the free terms
('k,u,g)B
is fixed. Thus, for all such A, the sum ZACB,IBI=k
(Sk,u,g,fixed)B
has at most (k + 1)m terms so using the Cauchy-Schwarz inequality,
((P-18k,u,g,fixed)A)
2
A
(k + 1)m
((Sku,g,fixed)B)
2
ACBjBI=k
Summing over all such A, each term ((sk,u,g,fixed)B)
I*k-1,u+1,g|
< m(k +
2
appears at most k times so
1)k||I ,u,g,fixed I2
E
Corollary 6.7.6. For any k > 1 and any u,
Sk
I 2 <4u
ug,f
ree
1
2
sk-1,u+1,g,free
n Vn
158
2
+ 4m(k + 1)k
I k,u,g,fixed
2
Proof. By Corollary 6.6.21,
2
2k<u4U2free
2+4
+
2
II -nk-1,u+1,g,free
2
+
skUgjree
44
k-1,u+1,g 2
Pluggin in Lemma 6.7.5,
r-
112 <4Uu2
sk,u,g,free
g,free skugfixed
2
4m(k + 1)k
112
2
E
as needed.
Corollary 6.7.7. For any k > 1 and any u if previous bounds hold then
ISk,u,g,free
2
<
1(9mn)
.
Proof. This follows immediately from Corollary 6.7.6 and the fact that n > 2000mr4
This completes the proof of Lemma 6.7.2 and thus Theorem 6.7.1
6.8
The cost of extending a base function
We now have our base function g which we know can be extended to an e-invariant
functions {g.} for every e E E(G) (we assume that s -+ t
E(G) and E(G) has no
edges of the form v -+ s or of the form t -+ v). However, while we were careful to
make the sum vectors Sk,u,g as small as possible for k + u < r, we have little control
over the sum vectors
,
for k + u = r. In this section, we show that when we
extend g to ge, the sum vectors sk,u,g, are not too large.
For all of these lemmas, let r be a constant and let g be a function such that
v = 0 whenever IV ;> r.
Lemma 6.8.1. If e is of the form s
-+
w and .vu{w}
V| < r - 1, if we take g, so that
159
= -v
whenever w
$
V and
1. g'ev = v whenever
2. gev = -PV\{w}
|V|
r and
if w E V, V| = r and 0 if w V V,|V| = r
then g, is e-invariant. Moreover,
1. If w $ A and |A| + u
=
r,
2. If w E A and A| + u = r,
SA,ut(ge) = -SA,u-1(g)
SA,u(ge) = -SA\{w},u(g)
+ SAu{w},u-2(9)
+
SA,u_1(g)
Proof. The fact that g, is e-invariant follows from Corollary 3.4.28. For the moreover
statement, note that if w $ A and JAI + u = r,
SA,u (e)
=
4eBU{w}
E
B:ACB,|B|=r-1,w B
5
-
YB
S
B:ACB, BI=r-1
YB
B:ACB,|B|=r-1,weB
:
+
-SA,u-1
5:
B
+
-
YB
B :AUjw}CB,IBjz~r-1
=
-SA,u-1(g)
+ SAU{w},u-2(9)
If w C A and IAI + u = r then by Lemma 6.4.6,
SA,u(ge) = -SA\{w},u(ge)
+
SA,u-1(ge) = -SA\{w},u(g)
Lemma 6.8.2. If e is of the form v -+ t and .vu{v}
=
#
r and
2. gev = gv\{v} if v G V,|V| = r and 0 if v
then g, is e-invariant. Moreover,
160
SA,u-1(9)
v whenever v 0 V, |V| < r -1,
if we take g, so that
1. gev = gv whenever |V|
+
V,|V| = r
1. If v 0 A and A| +u
=
r,
2. If v G A and AI +u = r,
SA,u(ge) = SA,u-1(g) -
SA,u(ge)
SAu{v},u-2(g)
SA\{v},u(g) -
SA,u-1(Y)
Proof. The fact that ge is e-invariant follows from Corollary 3.4.28. For the moreover
statement, note that if v
A and JAI + u = r,
SA,u(e)
B
YeBU{v}
S
YB
B:ACB,JBr-l,v B
=
B:ACB,|B|=r-1,vgB
=
B
-
YB
YB
B:ACB,|B|=r-1
B:ACB,IBI=r-l,vCzB
SA,u-1()
S
--
YB
B?:Aufv}CB,jBj=r-1
= SA,u-1(9) -
SAu{v},u-2(9)
If v E A and IAI + u = r then by Lemma 6.4.6,
SA,u(ge) = SA\{v},u(ge)
-
SA,u-1(9e) =
SA\{v},u(9)
Lemma 6.8.3. If e is of the form v -+ w and .vu{v,w}
4
ev =
Yv whenever
2. 4ev = -.
v\{w}
3. ev = -v\{w}
4.
ev=OifwVIV|
+
|VI
.V\{v}
# r
+
Yv\{v,w}
if v, w G V,
if v V V, w c V|IV| = r
=r
then ge is e-invariant. Moreover,
161
|VI
SA,u-1(Y)
= -vu{v}
whenever v,w $ V,| V < r - 2, if we take ge So that
1.
-
= r
+
Yvu{w}
+ Yv
1. If v,w
A and |A +u
SA,u(ge)
r,
-SA,u-1(9)
+
2SAu{w},u-2(9)
SAU{v},u-3(9) -
-
SAU{w},u-3(g)
-
SAU{v,w},u-3 (9)
+
SAU{v,w,u-4(9)
+
SA,u-2(9)
2. If w E A,v $ A and JA| +u = r then
SA,u(ge)
3. If w
+
=
-SA\{w},u(g)
-
SA\{w}U{v},u-2(9)
2SA,u-1(9) -
SA,u-2(g)
SAU{v},u-2(9) + SA\{w},u-1(9)
+ SAu{v},u-3(9)
A,v ( A and AI+u-=r,
SA,u(Ye)
=
SAU{w}\{v},u-1(9) - SA,u-1(9)
+ SA\{v},u-1(9)
-
SAU{w}\{v},u-2(9) -
+ SAu{w},u-3(Y)
SA,u-2(g)
4. Ifv,w c A and IA +u==r,
SA,u(Ye)
SA\{v},u(9) -
SA\{w},u(9)
+
SA\{v,w},u(g)
SA\{,u_1(g) - SA\{w,u_1(9)
+ SA,u-2(g)
Proof. The proof of this lemma is similar but more complicated. Again, the fact that
ge
is e-invariant follows from Corollary 3.4.28. For the moreover statement, note that
ifv,w
A and |A|+u=r,
SA,u(9e)
z
4eBU{w}
BACB,
|B|=r-1,v,wOB
+
z
B:ACB,
IB|==r-2,v,w B
162
9eBU{v,w}
Let's consider each term separately. For the first term,
YB
gB
E
BACB,
B:ACB,
IB|=r-1,veB
|BI=r-1
-SA,u-1(9)
-
YB
B
I:
-
BACB,
|B|=r-1,V,w
+
E eBU{w}
B:ACB,
IB|=r-1,v,
w B
B:ACB,
Bj =r-1,v,weB
B:ACB,
IB|=r-1,wEB
+ SAU{v},u-2(9)
+ SAu{w},u-2(9)
-
SAU{v,w},u-3(9)
For the second term,
4eBU{v,w}
E
B:ACB,
lBI=r-2,v,wVB
(.BUfw} -
YBU{v}
+
YB)
B:ACB,
|BI=r-2,vw B
Consider each of these terms separately. By the same logic as above,
YB
SA,u-2(g)
-
SAU{v},u-3(9) -
SAU{w},u-3(9) + SAU{v,w},u-4(9)
B:ACB,
|BI=r-2,v,w B
For the first term,
E
E
YBU{w}
B:ACB,
YB
B:AU{w}CB,
|BI=r-2,
IBI=r-1,
v,w B
vVB
~
--
B
B:AU{w}CB,
B|=r-1
=
sAU{w},u-2(g)
B
B:AU{w}CB,
vEB,
IB =r-1
-
SAU{v,w},u-3(9)
Following the same logic,
5
YBU{v}
SAU{v},u-2(9)
BACB,
B\=r-2,v,wVB
163
-
SAU{v,w},u-3(g)
B
Putting everything together,
SA,u(ge)
=
+ SAU{v},u-2(9) + SAU{w},u-2(Y)
(-SA,u-1(9)
+ (SA,u-2(g)
SAU{v},u-3(9)
-
+ (SAU{w},u-2(g)
-
SAU{v},u-3(g)
2SAU{w},u-2(g)
-
SAU{w},u-3(g)
SAU{v,w},u-3(Y))
-
+
-sA,u-1(9)
-
SAU{w},u-3(9)
-
-
-
SAU{v,w},u-3(9))
+ SAU{v,w},u-4(g))
(sAU{v},u-2(9)
-
SAU{v,w},u-3(g))
SAU{v,w},u-3(9) + SA,u-2(g)
+ SAU{v,w},u-4(Y)
Now that we have done this calculation, note that if w E A, v
A and JAl + u = r
then
SA,u(Ye) = SA\{w},u+1(ge)
-SA\{w},u(Y)
-
SA,u-2(g) -
+
2
sA,u-1(9) - SAU{v},u-2(g)
+ SA\{w},u-1(9)
SA\{w}U{v},u-2(g) + SAU{v},u-3(9)
For the fourth part, by Lemma 6.4.6, if v, w E A and IA + u = r then
SA,u(se) = SA\{v},u(g)-SA\{w},u(g)+SA\{v,w},u(Y)-SA\{v},u-1(9)-SA\{w},u-1(9)-+SA,u-2
Finally, note that if w
SA,u(ge)
=
(g)
A, v E A and IAl + u = r then
SAU{w},u-1(9e)
+ SA\{v},u-1(g)
=
SAU{w}\{v},u-1(Y)
-
SA,u-1(9)
-
SAU{w}\{v},u-2(Y)
-
SA,u-2(g) + SAU{w},u-3(g)
Corollary 6.8.4. If g, is defined according to Lemma 6.8.1, Lemma 6.8.2, or 6.8.3
then for all k,
Z:
A:|A|=k
(SA,r-k(ge))
2
{
< 200 max max
(SB,r-k 2 -1(g))
B:IBI=k2
164
2
(SB,r-k 2 -2 (g)) 2
,
B:IBI=k2
Proof. Consider the equation in the first moreover statement of Lemma 6.8.3:
SA,u(9e)
-SA,u-1(9)
2sAU{W},u-2(g)
SAU{v},u-3(Y)
-
SAU{w},u-3(g)
-
-
SAU{v,w},u-3(Y)
+ SA,u-2(g)
+ SAU{v,w},u-4(g)
Applying a Cauchy-Scwarz inequality to this,
(SA,u(ge))
10((SA,u-1(g)) 2
2
-
(SAU{v},u-3 (g) 2
(AU{w},u-2(g) )
-
2
(g) )2
(SAU{W},u-3
+ (SAU{v,w},u-3(g))
2
+ (sA,u-2(g))
2
(sAU{v,W},u-4 (g)) 2)
Similarly, we can apply a Cauchy-Scwarz inequality to all of the other equations.
In this way, each sum
of the form
each term
(SB,u(ge) ) 2
A:IAI k (SA,rk(ge))
(SB,u(ge)) 2
2
can always be bounded by a sum of terms
where IBI+-u
= r - 1 or r - 2 and I(IBI - k)|I
must come from an A with JA|
=
2. Moreover,
k, AAB C {v, w}. Thus, for
each B there are at most two A which will give a term with that B. This implies
that the coefficient for each term
(SB,u(9e) ) 2
has magnitude at most 2, so for all k we
have the inequality
S
(SA,r-k(ge) )
A:IAI=k
2
20
k 2 :|k-k 2 <2
(SB,r-k
B:IBI=k
2 -1
(g))
2
2
(SB,r-k 2 -2(9))2
+
B:IBI=k 2
The result now follows immediately as there are 5 possible k 2 with Ik - k2<
< 2. E
6.9
Putting everything together: A lower bound
We now put everything together and prove our lower bound on m(G)
Theorem 6.0.16. If r and m are constants such that m < 2000r4 and G is a directed
acyclic input graph such that
1. There is no path of length at most 2'-1 from s to t.
2. For any vertex v G V(G), there are at most m vertices w
165
c
V(G) such that
either there is a path of length at most 2r-2 from v to w in G or there is a path
of length at most
2 r-2
from w to v in G
then
(
(9mn)
(
nr
)
m(G) ;>
201 E(G) IVr (r + 1)! 9m
Proof. We obtain the base function g from Theorem 6.7.1 and obtain the e-invariant
functions {ge} using Lemmas 6.8.1, Lemma 6.8.2, and 6.8.3. Now by Corollary 6.8.4,
for all e E E(G), if k + u = r then IIsk,u,geI1
all el, e 2 E E(G), IIsku,9ge2-ge 1
12
2
< 200(9mn)
9
. This implies that for
< 800(9mn)2 1 if k + u = r and is 0 otherwise.
By Theorem 6.3.2,
m(G) >
E
-
JEGj- 1
{4(r
2k (k + u)I
(
1
ma
(\eGE(G)\{eo}
k,u
nk+u
2
sk,u,ge -geo 12
With some algebra, noting that we only need to consider terms where k + u = r, this
implies that
m (G) >
(9mn)
(n7
)4
(30mE(G) I2r(r + 1)! 9m
as needed.
EI
166
Chapter 7
Conclusion and Future Work
In this thesis, we developed a powerful new framework for analyzing monotone space
complexity via the switching network model. Using this framework, we separated
monotone analogues of L and NL and gave an alternative proof of the separation of
the monotone-NC hierarchy. We also investigated the monotone space complexity
of directed connectivity more deeply, proving upper and lower bounds on m(G) for
particular types of input graphs G. To the best of our knowledge, this focus on
particular types of input graphs is novel.
However, this work leaves many open questions. Most importantly, can this framework be extended to prove non-monotone lower bounds? Such an extension would
have to consider all NO instances, not just maximal NO instances. This will almost
certainly be extremely difficult and require many new techinques, but if successful, it
would be a major breakthrough.
Even for monotone space complexity, there are several open problems remaining.
One open problem is to determine the relationship between mLcircuits and
mLswitchingnetworks.
In other words, what is the relationship between the power of
monotone switching networks of polynomial size and the power of leveled monotone
circuits with polynomial size and logarithmic width? Answering this question would
help resolve the question of how monotone space complexity should be defined. Another open problem is to prove tight bounds on the size of monotone switching networks solving other problems such as the k-matching problem which asks whether or
167
not a graph has a matching of size k. Finally, for the directed connectivity problem,
can we prove better bounds on m(G)? In particular, can we prove tight or almost
tight bounds on m(G) when G is an acyclic directed graph? Brakensiek and the
author [4] have shown partial results in this area by proving almost dight bounds on
m(G) when G is an acyclic directed tree, but problem is still wide open for general
acyclic G.
Thus, it is our hope that this thesis represents not only an introduction of a new
framework for analyzing monotone space complexity and a summary of what has
been achieved so far, but also a foundation for further research on the fundamental
problem of space complexity.
168
Appendix A
Proof of Lemma 3.2.32
Lemma 3.2.32. Let G' be a certain knowledge switching network. For any certain
knowledge description of G' and any path P = s -+ vi -+
.
--
vi_ 1
-
t, if G is the
input graph with vertex set V(G) and E(G) = E(P), if W' is a walk in G' whose edge
labels are all in G
from
a vertex v'tar with Kv'
K
to a vertex v'nd with K,
Kt, then W' passes through a vertex v' such that K,, i Kt1, V(Kv') C {v 1 ,---
=
,-1
and IV(Kv)| ;> [lg(l)].
Proof. In this proof, we will split the path P in two and use induction on each half.
This will require projecting onto each half of P in two different ways.
Definition A.0.1.
1. Call the vertices L = {V1,
2. Call the vertices R = {v
, vi 2 1 } the left half of P.
I.
,- -1]1
, vi1} the right half of P.
Definition A.0.2.
1. We say an edge e = u
-+
v is a left edge if u,v G L U {s}
2. We say an edge e = u
-+
v is a right edge if u,v G R U {t}
3. We say an edge e = u
-
v is a left-jumping edge if u = s and v E R. Note that
t c R.
169
4. We say an edge e = u
that s
-*
v is a right-jumping edge if u =G L and v = t. Note
L.
Out first projections focus on the progress we have made towards showing that
there is a path from s to R U {t} and L U {s} to t, respectively.
Definition A.0.3. Given a vertex v E V(G),
1. Define pi(v) =v
if v $ R and pi(v) = t if v ER.
2. Define pr(v) =v if v
L and pi(v) = s if v
L.
Definition A.0.4. Given an edge e =u -+ v where u, v E V(G),
1. Define pi(e)
pi(u)
2. Define pr(e)
Pr(U)
-+
pi(v).
Pr(V).
Definition A.0.5. Given a knowledge set K,
1. Define pi(K) = {pi(e) : e G K,pi(e)
# t -*
2. Define pr(K) = {pr(e) : e E K,p,(e)
#s-
t}.
s}.
Definition A..6. Given a certain knowledge switching network G' together with
a knowledge description of G', define pi(G') to be the certain knowledge switching
network formed from G' with the following steps:
1. Replace all edge labels e with pl(e)
2. Replace all knowledge sets Kr, in the certain knowledge description with p,(Kvi).
3. Contract all edges in the switching network which now have label t
-
t. When
contracting an edge e' with endpoints v' and w', we may choose either Kv, or
Kw, to be the knowledge set for the resulting vertex.
Similarly, given a certain knowledge switching network G' together with a knowledge
description of G', define pr(G') to be the certain knowledge switching network formed
from G' with the following steps:
170
1. Replace all edge labels e with p,(e)
2. Replace all knowledge sets K,, in the certain knowledge description with p,(Kv').
3. Contract all edges in the switching network which now have label s -+ s. When
contracting an edge e' with endpoints v' and w', we may choose either Kv' or
K., to be the knowledge set for the resulting vertex.
Proposition A.0.7. Given a certain knowledge switching network G' for directed
connectivity on V(G),
1. pi(G') is a certain knowledge switching network for directed connectivity on
V(G)
\ R.
Furthermore, for any vertex w' E V(pi(G')), for all of the vertices
V' E V(G') which were contracted into w', K,' = Pi(Kv,).
2. p,(G') is a certain knowledge switching network for directed connectivity on
V(G)
v'
c
\ L.
Furthermore, for any vertex w' E V(p,(G')), for all of the vertices
V(G') which were contracted into w', Kw,
pr(Kv,).
Proof. We prove the first claim, the proof for the second claim is similar. To prove
the first claim, it is sufficient to show the following.
1. p, (K,,) = Ks,
2. p, (Kt) = Kt,
3. For any knowledge sets Ku,, K,, and any possible edge e which is not a right
edge, if Ku/ U {e} = K,, U {e} then pi(Kn') U {p1 (e)} - pi(Kg) U {p1 (e)}.
4. For any knowledge sets Ku,, Kv, if K,, = K,, or Ku, U {e} - K,, U {e} for some
right edge e then p,(Ku,)
p(Kvi)
The first two statements are trivial. For the third and fourth statements, we consider
the effect of pi on each type of move in the modified certain knowledge game.
1. If we originally added or removed an edge e from K after directly seeing e, if e
was not a right edge then we now add or remove p,(e) from p,(K) after directly
seeing pl(e). If e was a right edge then we now do nothing.
171
2. If we originally added or removed an edge v 3
v3
-4
v4 , v 4
-+
or remove pi(v
v5 from K after noting that
-+
v5 c K, then if p1(v3 ),p 1 (v4 ),p 1 (v5 ) are all distinct we now add
3
-+
v 5 ) from p1 (K).
If p(v 3 ), p(v 4 ), p(v 5 ) are not all distinct
two of them must be equal to t. In all of these cases we now do nothing.
If p(v) = pi(v 4 ) = t then p1(v3 --+ v5 ) = pi(v 4 -+ v5 ). This means that
pi(K U {v 3
-
V5 })
= pi(K) = pi(K \ {v 3
-+
Similar logic applies if
v 5 }).
pi(v 4 ) = p(v) = t. Finally, if p(v 3) =p(v 5 ) = t then p1(v 3
we again have that pl(K U {v 3
-
V5 }) =
pi(K)
=
p1 (K \ {v
v5 ) = t
-+
3 -+V5)
3. If we originally added or removed an edge e f s -+ t after noting that s
K, if e was not a right edge we now add or remove an pl(e)
t so
-
#
s
-
-
tE
t after noting
that s -+ t C pl(K). If e was a right edge then we now do nothing.
Using Proposition 3.2.14, statements 3 and 4 follow directly from these observations.
El
We now define a slightly different projection to each half. These projections will
help us look at the progress towards removing obsolete information after obtaining a
left-jumping or right-jumping edge.
Definition A.0.8. Given a knowledge set K,
1. Define p*(K) = {pi(e) e E K, pi(e)
t
2. Define p*(K) = {p,(e)
s
e E K, p,(e)
-
t, pi(e)
#
s
-
t}.
s, p,(e) 7 s -+ t}.
Definition A.0.9. Given a certain knowledge switching network G' together with
a knowledge description of G', define p*(G') to be the certain knowledge switching
network formed from G' with the following steps:
1. Delete t' and all other vertices v' such that Kv, = Kt from G'
2. Delete all edges e' such that e' has an endpoint v' and label e and Kv U {e} = Kt
3. Replace all edge labels e with pi(e)
4.
Replace all knowledge sets Kv, in the certain knowledge descriptionwith p*(K,).
172
5. Contract all edges in the switching network which now have label t
-
t. When
contracting an edge e' with endpoints v' and w', we may choose either Kv' or
K&' to be the knowledge set for the resulting vertex.
6. Add the vertex t' to G', assign it the knowledge set Kt= {s
-
t}, and add all
labeled edges with endpoint t' to G' which are allowed by condition 3 of Definition
3.2.8.
Similarly, given a certain knowledge switching network G' together with a knowledge
description of G', define p*(G') to be the certain knowledge switching network formed
from G' with the following steps:
1. Delete t' and all other vertices v' such that Kv, - Kt, from G'
2. Delete all edges e' such that e' has an endpoint v' and label e and K,'U{e} = Kt,
3. Replace all edge labels e with p,(e)
4.
Replace all knowledge sets Kv, in the certain knowledge description with p*(Kv).
5. Contract all edges in the switching network which now have label s -+ s. When
contracting an edge e' with endpoints v' and w', we may choose either K1 or
K&' to be the knowledge set for the resulting vertex.
6. Add the vertex t' to G', assign it the knowledge set Kt, = {s -+ t}, and add all
labeled edges with endpoint t' to G' which are allowed by condition 3 of Definition
3.2.8.
Proposition A.0.10. Given a certain knowledge switching network G' for directed
connectivity on V(G),
1. p*(G') is a certain knowledge switching network for directed connectivity on
V(G)
\
R. Furthermore, for any vertex w' G V(p*(G')), for all of the vertices
V' E V(G') which were contracted into w', K& -= p*(Kv).
173
2. p*(G') is a certain knowledge switching network for directed connectivity on
V(G)
\
L. Furthermore, for any vertex w' e V(p*(G')), for all of the vertices
V' E V(G') which were contracted into w', Kw, - p*(K,,).
Proof. We prove the first claim, the proof for the second claim is similar. To prove
the first claim, it is sufficient to show the following.
1. p*(Ks,) = Ksl
2. For any knowledge sets Ku,, K,, and any possible edge e which is not a right
edge, if K' U {e}
Kv U {e}
K then p*(K') U {p1 (e)}I
3. For any knowledge sets Ku', Kvr, if KI
Kt, for some right edge e then p*(K,,)
p*(Kv,) U {p 1 (e)}.
Kv, 0 K, or K,' U {e} = K, U {e}
p*(K,)
The first statement is trivial. For the second and third statements, we consider the
effect of p* on each type of move in the modified certain knowledge game.
1. If we originally added or removed an edge e from K after directly seeing e, if e
was not a right edge then we now add or remove p,(e) from pi(K) after directly
seeing pl(e). If e was a right edge then we now do nothing.
2. If we originally added or removed an edge v3
v3
-
V4 V 4
-+
V5
K, then if p1(v3),p 1(v4 ),pj(v 5 ) are all distinct we now add
V5
or remove pi(v 3
v5 from K after noting that
-4
-
v 5 ) from pi(K). Note that we cannot have v3 = s and
= t because we are assuming that we never have a knowledge set K such
that K = Kti. If p(v 3 ), p(v 4 ), p(v 5 ) are not all distinct two of them must be
equal to t. Following the same logic as before, in all of these cases we now do
nothing.
3. We do not have to consider moves where s
-4
t
C
K because we are assuming
that we never have a knowledge set K such that K = Kt,.
Using Proposition 3.2.14, statements 2 and 3 follow directly from these observations.
E174
Now that we have defined these projections, we give two more useful definitions
and then prove Lemma 3.2.32.
Definition A.0.11.
1. We say a vertex v' on a walk W' satisfies the lemma for the left half if K,, # Ktl,
V(Kv') C {v 1 , --- ,vi_ 1}, and |V(Kv') n L| > [lg(l)] - 1.
%
2. We say a vertex v' on a walk W' satisfies the lemma for the right half if K,,
Ktl, V(Kv ) C {v 1 ,-
,vi_}, and |V(Kvi) n R| > [lg(l)] - 1.
Definition A.0.12.
1. We say a knowledge set K is left-free if K 0 Kt, and V(K) n L = 0.
2. We say a knowledge set K is right-free if K # Kt, and V(K) n R = 0.
We now prove Lemma 3.2.32 by induction. The base case 1 = 2 is trivial. If 1 > 2
then given a walk W' from v'tr to V',n whose edge labels are all in E(P), first modify
W' and G' slightly as follows. Let u' be the first vertex on W' such that if e is the
label of the edge after u' then Ku, U {e} - Kt,. If u' is not the vertex immediately
before v'fd then add an edge from u' to v',d in G' with label e and replace the portion
of the path from u' to v' n with this single edge. Note that if the lemma is still
satisfied now then it was satisfied originally. This modification ensures that we do
not have to worry about moves in the modified certain knowledge game where we
have s -÷ t E K.
We now show that W' must have at least one vertex v' which satisfies the lemma
a
K81
for the left half. To see this, apply the projection p, to G' and W'. pl(K
and pl(Kvi)l
)
Kt,, so by the inductive hypothesis there must be some vertex w'
on p1 (W') such that V(K,,) C L and IV(Kw,) I
[Ilg l - 1]. Choose a v' which was
contracted into w' by pi. V(Kv') C {v 1 , - -- , vi-1 } and IV(K,) n L|I
[lg l - 1], so
v' satisfies the lemma for the left half, as needed. Following similar logic, W' must
also contain a vertex satisfying the lemma for the right half.
Now take b' to be the last vertex on W' which either satisfies the lemma for the
left half or satisfies the lemma for the right half. Without loss of generality, we may
175
assume that b' satisfies the lemma for right half. We may also assume that Kb, is
left-free, as otherwise b' satisfies Lemma 3.2.32. There are now two cases to consider.
Either Kb, contains a left-jumping edge, or it does not.
If Kb, does not contain a left-jumping edge, then apply pi to the portion of W'
between b' and t'. p(Kb,) =
{}
and p,(Kt,) = Kt, so using the inductive hypothessis
there must be a vertex a' on this portion of W' which satisfies the lemma for the left
half. However, this constradicts the definition of b'.
If Kb, does contain a left-jumping edge then choose a sequence of moves in the
certain knowledge game for going along W'. Let K be the first knowledge set we
obtain such that k conatins a left-jumping edge and for every K 2 after K but before
Kb', K 2 contains a left-jumping edge. K occurs when we are transitioning between
some vertices v' and w' in W' along an edge e' with label e.
Note that p*(K) = Kt,. This implies that p*(Kv') Up,(e) = p*(K.,) Upi(e) = Kt,.
Now take the portion of p*(W') from p*(v') to p*(b') and replace p*(v') with t'. Since
Kb, is left-free, p*(Kb,) = K,,. Using the inductive hypothesis, there must be a vertex
a' between v' and b' on W' which satisfies the lemma for the left half. Ka, occurs
beetween K and Kb', so Ka, also contains a left-jumping edge. Thus, a' satisfies
Lemma 3.2.32, as needed, and this completes the proof.
176
El
Bibliography
[11 Romas Aleliunas, Richard M. Karp, Richard J. Lipton, Liszl6 Lovisz, and
Charles Rackoff. Random walks, universal traversal sequences, and the complexity of maze problems. pages 218-223. Proceedings of the 20th Annual Symposium
on Foundations of Computer Science.
[2] Noga Alon and Ravi B. Boppana. The monotone circuit complexity of boolean
functions. Combinatorica, 7(1):1-22, 1987.
13]
Charles H. Bennett. Time/space trade-offs for reversible computation.
Journal on Computing, 18(4):766-776, 1989.
SIAM
Bounds on the size of sound
[4] Joshua Brakensiek and Aaron Potechin.
monotone switching networks accepting permutation sets of directed trees.
arXiv:1301.3780v1.
[5] Siu Man Chan and Aaron Potechin. Tight bounds for monotone switching networks via fourier analysis. Theory of Computing, 10(15):389-419, 2014.
[6] Stephen A. Cook. An observation on time-storage trade off. Journal of Computer
and System Sciences, 9:308-316, 1974.
[7] Stephen A. Cook and Charles W. Rackoff. Space lower bounds for maze threadability on restricted machines. SIAM Journal on Computing, 9(3):636-652, 1980.
[8] Jeff Edmonds, Chung K. Poon, and Dimitris Achlioptas. Tight lower bounds for
st-connectivity on the NNJAG model. SIAM Journal on Computing, 28(6):22572284, 1999.
[9] Michelangelo Grigni and Michael Sipser. Monotone complexity. In Michael S.
Paterson, editor, Proceedings of LMS workshop on boolean function complexity,
Durhan, 1990. Cambridge University Press.
[10] Armin Haken. Counting bottlenecks to show monotone P # NP. pages 3640. Proceedings of the 36th Annual Symposium on Foundations of Computer
Science.
[11] Neil Immerman. Nondeterministic space is closed under complementation. SIAM
Journal on Computing, 17:935-938, 1988.
177
[12] Mauricio Karchmer and Avi Wigderson. Monotone circuits for connectivity require superlogarithmic depth. pages 539-550. Proceedings of the 20th Annual
Symposium on the Theory of Computing.
[13] C.Y. Lee. Representation of switching functions by binary decision programs.
Bell Systems Technical Journal, 38:985-999, 1959.
[14] Ming Li, John Tromp, and Paul Vitanyi. Reversible simulation of irreversible
computation by pebble games. Physica D, 120:168-176, 1997.
[15] Ming Li and Paul Vitanyi. Reversibility and adiabatic computation: Trading
time and space for energy. Proceedings of the Royal Society of London. Series A:
Mathematical, Physical and Engineering Sciences, 452(1947):769-789, 1996.
[16] Pinyan Lu, Jialin Zhang, Chung K. Poon, and Jin Y. Cai. Simulating undirected
st-connectivity algorithms on uniform JAGs and NNJAGs. 16th International
Symposium on Algorithms and Computation.
[17] William J. Masek. A fast algorithm for the string editing problem and decision
graph complexity. Master's thesis, Department of Electrical Engineering and
Computer Science.
[18] Aaron Potechin. Bounds on monotone switching networks for directed connectivity. pages 553-562. Proceedings of the 51st Annual Symposium on Foundations
of Computer Science.
[19] Ran Raz and Pierre McKenzie. Separation of the monotone NC hierarchy. pages
234-243. Proceedings of the 38th Annual Symposium on Foundations of Computer Science.
[20] Alexander Razborov. Lower bounds for deterministic and nondeterministic
branching programs. pages 47-60. Proceedings of the 8th FCT, Lecture Notes
in Computer Science vol. 529.
[21] Alexander Razborov. Lower bounds for the monotone complexity of some
boolean functions. Soviet Mathematics Doklady, 31:354-357, 1985.
[22] Omer Reingold. Undirected st-connectivity in log-space. pages 234-243. Proceedings of the 37th Annual Symposium on the Theory of Computing.
[23] Walter J. Savitch. Relationship between nondeterministic and deterministic tape
classes. Journal of Computer and System Sciences, 4:177-192, 1970.
[24] Claude Shannon. A symbolic analysis of relay and switching networks.
actions of American Institute of Electrical Engineers, 57:713-723, 1938.
Trans-
[25] Claude Shannon. The synthesis of two-terminal switching circuits. Bell Systems
Technical Journal, 1949.
178
[26] R6bert Szelepcsenyi. The method of forcing for nondeterministic automata. Acta
Informatica, 26(3):279-284, 1988.
127] Vladimir Trifonov. An o(log n log log n) space algorithm for undirected stconnectivity. pages 626-633. Proceedings of the 37th Annual Symposium on the
Theory of Computing.
179