Lecture 30: Recall: defns of Φ, f , Z .

advertisement
Lecture 30:
Recall: defns of Φ, f Φ, ZnΦ.
Recall: Proposition: (a special case of Ruelle, Theorem 3.4) For
n.n. Zd SFT M and a n.n. potential Φ,
P (Φ) = P (σ|M , f Φ) = lim (1/nd) log ZnΦ
n→∞
Recall: Perron-Frobenius Theorem:
Let A be an irreducible matrix with spectral radius
λA := max eigenvalues λ of A |λ|. Then
1. λA > 0 and is an eigenvalue of A.
2. λA has a (strictly) positive eigenvector and is the only eigenvalue
with a nonnegative eigenvector.
Proposition: For a TMC MC (i.e., n.n. SFT with d = 1 defined
by 0-1 matrix C) with C irreducible and a n.n. potential Φ (defined
on the non-forbidden configurations on edges), define the matrix C Φ
by:
(C Φ)ij = Cij e−Φ(ij)
Then
P (Φ) = P (σ|MC , f Φ) = log λC Φ
Corollary: h(σ|MC ) = log λC .
Proof:
ZnΦ = 1(C Φ)n−11
So,
P (Φ) = lim (1/n) log ZnΦ = lim (1/n)1(C Φ)n1
n→∞
n→∞
P-F Theorem guarantees a positive (right) eigenvector v corresponding to λ = λC Φ .
1
Claim: There exist constants K1, K2 > 0 s.t.
K1λn ≤ 1(C Φ)n1 ≤ K2λn
Proof of Claim:
1(C Φ)n1 ≤ 1(C Φ)n(
1(C Φ)n1 ≥ 1(C Φ)n(
v
1
) = (1 · v)
λn
min(v)
min(v)
1
v
) = (1 · v)
λn
max(v)
max(v)
Take log, divide by n, take limn, to get:
P (Φ) = P (σ|MC , fΦ) = log λC Φ
Example: Golden Mean SFT, Φ = 0.
C=
1 1
1 0
√
Eigenvalues are log(1 ± 5)/2.
√
h(σ|MC ) = P (σ, 0) = log λC = log(1 + 5)/2.
Example: 1D Ising (ferromagnetic, no external field, coupling
strength = 1, inverse temperature = β)
Here Φ(00) = −β = Φ(11) and Φ(01) = β = Φ(10) and
1 1
C=
1 1
and so
β −β e e
CΦ =
e−β eβ
Eigenvalues are eβ ± e−β .
2
P (Φ) = log λC Φ = log(eβ + e−β ).
Theorem (Variational Principle for irreducible TMC) Let C be
irreducible, Φ be nn potential, and M = M(σ|MC ). Then
Z
P (Φ) = log λC Φ =
sup
hµ(σMC ) + f Φdµ
µ∈M(σ|M )
C
and the sup is achieved uniquely by the stationary Markov chain:
vj
Pij = CijΦ
λC Φ vi
where C Φv = λC Φ v.
Special case: Φ = 0: topological entropy is the sup of measuretheoretic entropies.
Idea of Theorem: the measure µΦ that achieves the sup is a staΦ
tionary measure that satisfies µΦ(w) ∼ e−U (w) for configurations
w on finite subsets of Zd (by µΦ(w), we mean the measure of the
cylinder set corresponding to w). So, if Φ = 0, µΦ should be the
“most uniform” stationary measure with support contained in MC .
Example: Golden Mean, Φ = 0.
v = [λ, 1]T . So,
P =
1/λ 1/λ2
1
0
where λ is the golden mean.
Example: Ising Model:
v = [1, 1]T . So,
P =
1
eβ + e−β
3
β
−β
e e
e−β eβ
Lecture 31:
Recall:
Theorem (Variational Principle for irreducible TMC) Let C be
irreducible, Φ be nn potential, and M = M(σ|MC ). Then
Z
hµ(σMC ) + f Φdµ
P (Φ) = log λC Φ =
sup
µ∈M(σ|M )
C
and the sup is achieved uniquely by the stationary Markov chain:
vj
Pij = CijΦ
λC Φ vi
where C Φv = λC Φ v.
Special case: Φ = 0: topological entropy is the sup of measuretheoretic entropies.
Proof of Variational Principle:
Step 1: For all µ ∈ M(σ|MC ),
Z
hµ(σMC ) + f Φdµ ≤ log λC Φ = P (σ|MC , f Φ).
Lemma:
−
max P
(p1 ,...,pn ):pi ≥0, i pi =1
X
pi log pi +
i
X
X
piai = log(
e ai )
i
with equality iff pi = eai /K, where K =
i
ai
ie .
P
Proof of Lemma: Let
eai
pi K
αi = , xi = a
K
ei
Then
X
αi = 1,
X
i
i
4
αixi = 1.
By Jensen applied to φ(x) = x log x,
X
X
pi K
0 = φ(1) ≤
αiφ(xi) =
pi log a
ei
i
i
X
X
piai + log K
pi log pi −
=
i
i
Proof of Step 1: Let α = {Ai} be the time-0 partition: Ai = {x ∈
MC : x0 = i}. Then
Z
X
n−1
Φ
Hµ(α0 ) + Snf dµ ≤
−µ(A) log µ(A) + µ(A)Snf Φ(xA)
A∈α0n−1
(where xA achieves the max of Snf Φ on A). By the Lemma, this is
X
Φ
≤ log(
eSnf (xA))
A∈α0n−1
So
Z
hµ(σMC ) +
f Φdµ = lim (1/n)(Hµ(α0n−1) +
Z
n→∞
≤ lim (1/n) log(
n→∞
X
eSnf
Φ (x
A)
Snf Φdµ)
)
A∈α0n−1
= P1(σ|MC , f Φ) = P (σ|MC , f Φ)
Step 2:Let µ ∈ M(σ|M ). Let S = {i : µ(Ai) > 0} (recall
Ai = {x ∈ MC : x0 = i}). Let
Aij = {x ∈ MC : x0 = i, x1 = j}
and
Pij =
µ(Aij )
, i, j ∈ S,
µ(Ai)
5
πi = µ(Ai), i ∈ S
(here, P is a transition probability matrix, not pressure). Then P
defines a first-order stationary Markov chain ν with stationary vector
π and h(µ) ≤ h(ν) with equality iff µ = ν.
ν is called the Markovization of µ.
Proof: Check πP = π
−i
−1
hµ(σC ) = Hµ(α | ∨∞
i=1 σ (α)) ≤ Hµ (α|σ (α))
= Hµ(σ −1(α) ∨ α) − Hµ(σ −1(α)) = Hµ(σ −1(α) ∨ α) − Hµ(α)
X
µ(Aij )
−1
µ(Aij ) log(
= Hµ(σ (α)|α) = −
)
µ(A
)
i
ij
X
=−
πiPij log Pij = hν (σC )
ij
−i
By entropy inequality 10, equality holds iff α ⊥σ−1(α) ∨∞
i=2 σ (α).
i.e., µ is Markov, equivalently µ = ν.
Step 3: Unique measure that achieves the sup.
From Steps 1 and 2, we know that the sup is ≤ log λ and the
only invariant measures that can achieve the sup must be first-order
stationary Markov.
Step 3: The sup is log λ, it is achieved by the the first order
stationary Markov chain µ defined by
vj
Pij := CijΦ ,
λvi
and it is achieved uniquely.
Proof: Recall that v is a right eigenvector for λ = λC Φ . Let w be
a strictly positive left eigenvector for λ, normalized s.t. w · v = 1.
Then π, defined by πi = wivi , is the stationary vector for P :
X
X
vj
(πP )i =
πiPij =
wiviCijΦ
λvi
j
j
6
= wi
X
CijΦ
j
vj
= w i vi = π i
λ
Note that
− log Pij − Φij = − log CijΦ
vj
− Φij = log λ + log vi − log vj
λvi
provided Cij > 0. Thus, for any invariant measure ν on MC ,
Eν (− log Pij −Φij ) = log λ+Eν (log vi−log vj ) = log λ+Eν (g◦σ−g) = log λ
where g(x) = − log vx0
(Note that since ν is a measure on MC , whenever ν(Aij ) > 0, we
have Cij > 0).
In particular
Z
X
X
Φ
hµ(σ|M )+ f dµ = −
πiPij log Pij −
Φ(ij)πiPij = Eµ(− log Pij −Φi
ij
ij
and so µ achieves the sup.
If ν is another first order stationary Markov chain on MC , defined
by P 0, π 0, which achieves the sup, then
Z
Eν (− log Pij0 − Φij ) = hν (σ|M ) + f Φdν = log λ
So,
Eν (− log Pij0 − Φij ) = log λ = Eν (− log Pij − Φij )
So,
Eν (
log Pij
) = 0.
log Pij0
But
X
X
P
0 0
0
log Eν 0 = log
πiPij Pij /Pij = log
πi0 Pij = 0.
P
ij
ij
7
Thus,
P
P
=
0
=
log
E
ν
P0
P0
By Jensen, equality happens only if P 0 = P for all i, j s.t. Pij0 > 0.
But Pij0 > 0 iff Pij > 0:
if Pij0 > 0, then Pij > 0 since ν is a measure on MC ;
P
P
1 = j:P 0 >0 Pij0 = j:P 0 >0 Pij and so if Pij0 = 0, then Pij = 0.
ij
ij
0
So, P = P and ν = µ. Eν log
8
Lecture 32:
General Variational Principle (Saff): Let T be a continuous Zd+
action on a compact metric space M and f : M → R be a continuous
function. Then
Z
P (T , f ) = sup hµ(T ) + f dµ
µ∈M(T )
Last time: special case where the supremum is achieved uniquely
by an explicit, simple µ (1st-order stationary Markov chain).
Example: There exists T , f s.t. the sup in the variational principle
is not achieved.
T will be a Z-action, i.e. continuous map compact metric M and
f = 0. So, P (T, f ) = h(T ).
For i = 1, 2, . . ., let Ti be a sequence of continuous maps on compact metric spaces Mi, with corresponding metrics ρi, s.t. h(Ti) is
strictly increasing to a limit L < ∞.
We can assume that the ρi-diameter of each Mi is 1 because the
diameter of each Mi is finite and so we can rescale each metric to
have diameter 1. Continuity of each Ti is not affected by rescaling.
Let M be the disjoint union of the Mi and one more point M0 =
{?}. Define T : M → M by T |Mi = Ti and T (?) = ?. Endow M
with metric ρ:
For i 6= 0 and x, y ∈ Mi, ρ(x, y) := (1/i2)ρi(x, y).
P
For 0 < i < j, x ∈ Mi, y ∈ Mj , ρ(x, y) := jk=i(1/k 2).
P
2
For i 6= 0, x ∈ Mi, d(x, ?) = ∞
k=i (1/k ).
Fact: T : M → M is continuous (because Mi → ?, T (Mi) = Mi)
and T (?) = ?.
9
Proposition: Let T : M → M be a continuous map and for
i = 1, 2, . . ., µi ∈ M(T ). Let p be a countably infinite probability
P
vector: p = (p1, p2, . . .), with each pi ≥ 0 and i pi = 1. Then
X
P
pihµi (T ).
h i piµi (T ) =
i
Proof of Proposition: more general version (replace {µi} with a
continuum of measures) in Theorem 4.3.7 of Keller; more special
version (replace {µi} with just two measures) in Theorem 8.1 of
Walters. See below. Proof that the sup in the variational principal is not achieved:
Let µi := (µ|Mi )/(µ(Mi)) if µ(Mi) 6= 0. Then
µ=
X
pi µi
{i:µ(Mi )6=0}
where pi = µ(Mi). So,
X
hµ(T ) =
pihµi (T ) ≤
X
pih(Ti) < L
{i:µ(Mi )6=0}
{i:µ(Mi )6=0}
Since Ti = T |Mi, each h(T ) ≥ h(Ti) for all i. Thus, h(T ) ≥ L.
Thus,
hµ(T ) < h(T ).
(in fact, one can show that h(T ) = L, but we don’t need that here).
Idea of preceding: There are invariant measures µi on each Mi
s.t. hµi (Ti) → L, but the sequence of supports of these measures
converges to a single point, which cannot carry any entropy.
Note: In fact, the argument above can be modified to only use
the Walters-Theorem 8.1 version (i.e., an average of two measures)
of the Proposition above, which we now prove:
10
hpµ+(1−p)ν (T ) = phµ(T ) + (1 − p)hν (T ).
Proof:
Let α be a finite measble partition. Since φ(x) = −x log x is
concave-down (φ00 < 0),
X
−(pµ(Ai)+(1−p)ν(Ai)) log(pµ(Ai)+(1−p)ν(Ai))
Hpµ+(1−p)ν (α) =
i
≥ p(
X
−µ(Ai) log µ(Ai)) + (1 − p)(
X
−ν(Ai) log ν(Ai))
i
i
= pHµ(α) + (1 − p)Hν (α).
So,
Hpµ+(1−p)ν (α) ≥ pHµ(α) + (1 − p)Hν (α).
Now,
[(pµ(A)+(1−p)ν(A)) log(pµ(A)+(1−p)ν(A))]−[pµ(A) log µ(A))]
−[(1 − p)ν(A) log ν(A)]
= pµ(A)[log(pµ(A) + (1 − p)ν(A)) − log(pµ(A))]
+(1 − p)ν(A))[log(pµ(A) + (1 − p)ν(A)) − log(1 − p)ν(A))]
+pµ(A)[log(pµ(A))−log µ(A)]+(1−p)ν(A)[log((1−p)ν(A))−log(ν(A))]
≥ 0 + 0 + pµ(A) log p + (1 − p)ν(A) log(1 − p)
(since log is increasing) So,
Hpµ+(1−p)ν (α) − (pHµ(α) + (1 − p)Hν (α)) ≤ log 2.
So,
hpµ+(1−p)ν (T, α) = phµ(T, α) + (1 − p)hν (T, α)
So,
hpµ+(1−p)ν (T ) ≤ phµ(T ) + (1 − p)hν (T )
11
Let > 0. Choose α1, α2 s.t.
hµ(T, α1) > hµ(T ) − , hν (T, α1) > hν (T ) − ,
Then,
hpµ+(1−p)ν (T, α1 ∨ α2) = phµ(T, α1 ∨ α2) + (1 − p)hν (T, α1 ∨ α2)
≥ phµ(T, α1) + (1 − p)hν (T, α2) > phµ(T ) + (1 − p)hν (T ) − Argue differently if entropy is infinite. Theorem: If a G-action T (where G is Zd+ or Zd) is expansive and
f is continuous, then the sup in the variational principle is achieved.
Cor: For the shift action σ|M on a (two-sided) shift space M and
continuous f , E(σ|M , f ) 6= ∅.
Outline of proof of Theorem:
Let α = {A1, . . . , Ak } be a finite Borel partition of M .
Defns: α is δ-fine if for each i, diam(Ai) < δ
Defn: For a Borel probability measure µ, α is µ-good if for each i,
µ(∂Ai) = 0.
Step 1(Theorem 4.5.6, Keller): Let δ be an expansive constant for
T . For a δ-fine partition of M , we have ∨g∈GT −g (α) = B, t he Borel
σ-algebra.
Step 2(Theorem 4.2.4, Keller): Let µ ∈ M(T ). If α is µ-good,
then the function M(T ) 7→ R, given by µ 7→ hµ(T , α) is upper
semi-continuous at µ, i.e., whenever µn → µ weakly, then
lim sup hµn (T , α) ≤ hµ(T , α)
n→∞
12
R
R
Weak convergence means: for all continuous f , f dµn → f dµ;
equivalently for all Borel A with µ(∂A) = 0, µn(A) → µ(A).
Step 3: (Lemma 4.4.10, Keller): For a compact metric space M , a
Borel probability measure µ on M , and δ > 0, there exists a δ-fine,
µ-good partition of M .
Note: Step 3 involves topology and measure theory, but has nothing to do with the dynamics.
Proof of Theorem:
Let µn ∈ M(T ) s.t. hµn (T ) +
R
f dµn → P (T , f ).
Note that M(T ) is compact in the weak topology: the space X
of all Borel probability measures is a weakly compact metric space,
and M(T ) is a weakly closed subset of X.
So, there is a subsequence ni s.t. µni converges weakly to some
µ ∈ M(T ).
Let δ be an expansive constant for T . By Step 3, there is a δ-fine,
µ-good partition α of M .
By Step 1 and the Sinai generator theorem, for each ν ∈ M(T ),
hν (T , α) = hν (T ). In particular hµ(T , α) = hµ(T ) and each
hµni (T , α) = hµni (T )
Then by Step 2,
Z
P (T , f ) = lim sup(hµni (T )+
i→∞
Z
f dµni ) = lim sup(hµni (T , α)+
i→∞
Z
≤ hµ(T , α) +
Z
f dµ = hµ(T ) +
13
f dµ
f dµni )
Download