Applications of free probability and random matrix theory Øyvind Ryan December 2007

advertisement
CMA 2007
Applications of free probability and random matrix
theory
Øyvind Ryan
December 2007
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Some important concepts from classical probability
Random variables are functions (i.e. they commute w.r.t.
multiplication) with a given p.d.f. (denoted f )
Expectation (denoted E ) is integration
Independence
Additive convolution (∗) and the logarithm of the Fourier
transform
Multiplicative convolution
Central limit law, with special role of the Gaussian law
∗n
Poisson distribution Pc : The limit of 1 − nc δ(0) + cn δ(1)
as n → ∞.
Divisibility: For a given a, find i.i.d. b1 , ..., bn such that
fa = fb1 +···+bn .
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Can we find a more general theory, where the random variables
are matrices (or more generally, operators), with their
eigenvalue distribution (or spectrum) taking the role as the
p.d.f.?
What are the analogues to the above mentioned concepts for
this theory?
What are the applications of such a theory?
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Free probability
Free probability was developed as a probability theory for random
variables which do not commute, like matrices
The random variables are elements in a unital ∗-algebra
(denoted A), typically B(H), or Mn (C).
Expectation (denoted φ) is a normalized linear functional on A.
The pair (A, φ) is called a noncommutative probability space.
For matrices, φ will be the normalized trace trn , defined by
n
trn (a) =
1X
aii .
n
i =1
For random matrices, we set φ(a) = τn (a) = E (trn (a)) is
defined by.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
What is the "central limit" for large matrices?
We will attempt to make a connection with classical probability
through large random matrices. We would like to define random
matrices as "independent" if all entries in one are independent from
all entries in the other.
Assume that X1 , ..., Xm are n × n i.i.d. complex matrices, and
τn (Xi ) = 0, τn (Xi2 ) = 1. What is the limit when m → ∞ in
X1 + · · · + Xm
√
?
m
If Xi = √1n Yi where Yi has i.i.d. complex standard Gaussian
entries, then
X1 + · · · + Xm
√
∼ X,
m
where X = √1n Y and Y has i.i.d. complex standard Gaussian
entries. Therefore, matrices with complex standard Gaussian entries
are central limit candidates.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The full circle law
What happens when n is large? The eigenvalues converge to what
is called the full circle law. Here for n = 500.
1
0.8
0.6
0.4
0.2
0
−0.2
−0.4
−0.6
−0.8
−1
−1
−0.5
0
0.5
1
plot(eig( (1/sqrt(1000)) * (randn(500,500) +
j*randn(500,500)) ),’kx’)
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The semicircle law
35
30
25
20
15
10
5
0
−3
−2
−1
0
1
2
3
A = (1/sqrt(2000)) * (randn(1000,1000) +
j*randn(1000,1000));
A = (sqrt(2)/2)*(A+A’);
hist(eig(A),40)
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The Marc̆henko Pastur law
What happens with the eigenvalues of N1 XX H when X is an n × N
random matrix with standard complex Gaussian entries?
The eigenvalue distribution converges to the Marc̆henko
Pastur law with parameter Nn , denoted µ Nn .
Let f µc be the p.d.f. of µc . Then
p
(x − a)+ (b − x)+
,
(1)
2πcx
√
√
where (z)+ = max(0, z), a = (1 − c)+ and a = (1 + c)+ .
f
µc
1
(x) = (1 − )+ δ(x) +
c
The matrices N1 XX H occur most frequently as sample
covariance matrices: N is the number of observations, and n is
the number of parameters in the system.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Four different Marc̆henko Pastur laws µ Nn are drawn.
1.6
c=0.5
c=0.2
c=0.1
c=0.05
1.4
1.2
Density
1
0.8
0.6
0.4
0.2
0
0.5
1
Øyvind Ryan
1.5
2
2.5
3
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Derivation of the limiting distribution for
√1 XXH
N
When x is standard complex Gaussian, we have that
E |x|2p = p!.
A more general statement concerns a random matrix √1N XXH ,
where X is an n × N random matrix with independent standard
complex Gaussian entries. It is known [HT] that
p 1 X k(π̂) l(π̂)
1
H
√ XX
N
n ,
= p
τn
N n
N
π∈Sp
where π̂ is a permutation in S2p constructed in a certain way from
π, and k(π̂), l (π̂) are functions taking values in {0, 1, 2, ...}.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
One can show that this equals
p X
X ak
1
H
√ XX
.
=
1+
τn
N 2k
N
π̂∈NC
k
2p
The convergence is "almost sure", which means that we have very
accurate eigenvalue prediction when the matrices are large.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Motivation for free probability
One can show that for the Gaussian random matrices we
considered, the limits
φ Ai1 B j1 · · · Ail B jl = lim trn Ain1 Bnj1 · · · Ainl Bnjl
n→∞
exist. If we linearly extend the linear functional φ to all polynomials
in A and B, the following can be shown:
Theorem
If Pi , Qi are polynomials in A and B respectively, with 1 ≤ i ≤ l ,
and φ(Pi (A)) = 0, φ(Qi (B)) = 0 for all i, then
φ (P1 (A)Q1 (B) · · · Pl (A)Ql (B)) = 0.
This motivates the definition of freeness, which is the analogy to
independence.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Definition of freeness
Definition
A family of unital ∗-subalgebras (Ai )i ∈I is called a free family if


aj ∈ Aij


⇒ φ(a1 · · · an ) = 0. (2)
i1 6= i2 , i2 6= i3 , · · · , in−1 6= in


φ(a1 ) = φ(a2 ) = · · · = φ(an ) = 0
A family of random variables ai is called a free family if the algebras
they generate form a free family.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The free central limit theorem
Theorem
If
a1 , ..., an are free and self-adjoint,
φ(ai ) = 0,
φ(ai2 = 1,
supi |φ(aik )| < ∞ for all k,
√
then the sequence (a1 + · · · + an )/ n converges in distribution to
the semicircle law.
In free probability, the semicircle law thus
√ has the role of the
1
Gaussian law. it’s density is density 2π 4 − x 2
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Similarities between classical and free probability
1
Additive convolution ⊞: The p.d.f. of the sum of free random
variables. The role of the logarithm of the Fourier transform is
now taken by the R-transform, which satisfies
Rµa ⊞µb (z) = Rµa (z) + Rµb (z).
2
The S-transform: Transform on probability distributions which
satisfies Sµa ⊠µb (z) = Sµa (z)Sµb (z)
3
Poisson distributions have their analogy in the free Poisson
distributions: These are given by the Marc̆enko Pastur laws µc
with parameter c, which also can be written as the limit of
⊞n
as n → ∞
1 − cn δ(0) + nc δ(1)
4
Infinite divisibility: There exists an analogy to the Lévy-Hinčin
formula for infinite divisibility in classical probability.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Main usage of free probability in my papers
Let A and B be random matrices. How can we make a good
prediction of the eigenvalue distribution of A when one has the
eigenvalue distribution of A + B and B? Simplest case is when
one assumes that B is Gaussian (Noise). What about the
eigenvectors?
Assume that we have the eigenvalue distribution of
1
H
N (R + X )(R + X ) , where R and X are n × N random
matrices, with X Gaussian. If the columns of R are
realizations of some random vector r , what is the covariance
matrix E (ri rj∗ )?
Have use for multiplicative free convolution with the
Marc̆henko Pastur law. This has an efficient implementation.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Channel capacity estimation
The following is a much used observation model in MIMO systems:
1
Ĥi = √ (H + σXi )
n
(3)
where
n is the number of receiving and transmitting antennas,
Ĥi is the n × n measured MIMO matrix,
H is the n × n MIMO channel and
Xi is the n × n noise matrix with i.i.d zero mean unit variance
Gaussian entries.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Channel capacity estimation
With free probability we can estimate the eigenvalues of n1 HHH
based on few observations Ĥi . This helps us estimate the channel
capacity:
The capacity of a channel with channel matrix H and signal to
noise ratio ρ = σ12 is given by
C
=
=
1
1
H
log det I + 2 HH
n
nσ
n
1X
1
log(1 + 2 λl )
n
σ
(4)
(5)
l=1
where λl are the eigenvalues of n1 HHH .
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Observation model
Form the compound observation matrix
i
1 h
Ĥ1...L = √ Ĥ1 , Ĥ2 , ..., ĤL ,
L
from the observations
1
Ĥi = √ (H + σXi ) ,
n
(6)
Using free probability, one can with high accuracy estimate the
eigenvalues of n1 HHH from the eigenvalues of Ĥ1...LĤH
1...L .
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Free capacity estimation for channel matrices of various rank
4.5
4
Capacity
3.5
True capacity, rank 3
Cf, rank 3
True capacity, rank 5
Cf, rank 5
3
True capacity, rank 6
Cf, rank 6
2.5
2
1.5
0
10
20
30
40
50
60
Number of observations
70
80
90
100
Figure: The free probability based estimator for various number of
observations. σ 2 = 0.1 and n = 10. The rank of H was 3, 5 and 6.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Application areas
digital communications,
nuclear physics,
mathematical finance
Situations in these fields, can often be modelled with random
matrices. When the matrices get large, free probability theory is an
invaluable tool for describing the asymptotic behaviour of many
systems.
Other types of matrices which are of interest are random unitary
matrices and random Vandermonde matrices.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
List of papers
Free Deconvolution for Signal Processing Applications.
Submitted to IEEE Trans. Inform. Theory.
arxiv.org/cs.IT/0701025.
Multiplicative free Convolution and Information-Plus-Noise
Type Matrices. Submitted to Ann. Appl. Probab.
arxiv.org/math.PR/0702342.
Channel Capacity Estimation using Free Probability Theory.
Submitted to IEEE. Trans. Signal Process.
arxiv.org/abs/0707.3095.
Random Vandermonde Matrices-Part I: Fundamental results.
Work in progress.
Random Vandermonde Matrices-Part II: Applications to
wireless applications. Work in progress.
Applications of free probability in finance. Estimation of the
covariance matrix itself (not only it’s eigenvalue distribution).
2008.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
References
[HT]: "Random Matrices and K-theory for Exact C ∗ -algebras". U.
Haagerup and S. Thorbjørnsen. citeseer.ist.psu.edu/114210.html.
1998.
This talk is available at
http://heim.ifi.uio.no/∼oyvindry/talks.shtml.
My publications are listed at
http://heim.ifi.uio.no/∼oyvindry/publications.shtml
Øyvind Ryan
Applications of free probability and random matrix theory
Download