Document 10504914

advertisement
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
Random Matrices: Theory and Applications
c World Scientific Publishing Company
Computing with Beta Ensembles and Hypergeometric Functions
VESSELIN DRENSKY
Institute of Mathematics and Informatics, Bulgarian Academy of Sciences
1113 Sofia, Bulgaria
drensky@math.bas.bg
ALAN EDELMAN
Department of Mathematics, Massachusetts Institute of Technology,
77 Massachusetts Avenue, Cambridge, MA 02139, U.S.A.
edelman@math.mit.edu
TIERNEY GENOAR
Department of Mathematics, San José State University,
1 Washington Square, San Jose, CA 95192, U.S.A.
tgenoar@gmail.com
RAYMOND KAN
Joseph L. Rotman School of Management, University of Toronto
105 St. George Street, Toronto, Ontario, Canada M5S 3E6
kan@chass.utoronto.ca
PLAMEN KOEV
Department of Mathematics, San José State University,
1 Washington Square, San Jose, CA 95192, U.S.A.
koev@math.sjsu.edu
Received (January 25, 2014)
Revised (Day Month Year)
Keywords: Wishart; Jacobi; Laguerre; Hermite; MANOVA; eigenvalue; hypergeometric
function of a matrix argument
Mathematics Subject Classification 2000: 15A52, 60E05, 62H10, 65F15
We survey the empirical models for the Hermite, Wishart, Laguerre, Jacobi, and
MANOVA beta ensembles. The eigenvalue distributions of these ensembles are expressed
in terms of the hypergeometric function of a matrix argument. We present a number of
identities for this function that have been instrumental in computing these distributions
in practice. These identities include a number of new results.
1
January 25, 2014
2
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
1. Introduction
Multivariate statistical analysis relies heavily on the classical random matrix
ensembles—Hermite, Wishart, Laguerre, Jacobi, and MANOVA—and the distributions of their eigenvalues in order to infer information about multivariate datasets.
Muirhead’s text [29] provides an in-depth analysis of the real (β = 1) case. Most of
the results have complex (β = 2) analogues [31] and the extension to quaternions
(β = 4) is straightforward although lacking on applications thus far.
In the past 10–15 years generalizations of these ensembles to any β > 0 have
been introduced. This suggests that most of the classical results (and perhaps most
of random matrix theory) also generalize to any β > 0 [12]. Advances in computational tools have also made these theoretical results very important in practice.
The theoretical importance of the distributions of the extreme eigenvalues is also
well documented [20,21].
The goal of this paper is to present in one place the empirical models, the eigenvalue distributions, and the identities that have made computing those distributions
possible. We include several new results.
This paper is organized as follows. In section 2 we introduce the random matrix
ensembles, their empirical models as well as the hypergeometric function of a matrix
argument and the various combinatorial objects that we will need later on. In
section 3 we present the identities for the hypergeometric function and the Jack
function. The known distributions and densities of the extreme eigenvalues of the
beta ensembles are in 4. We prove all new results in section 5. The numerical
experiments are in section 6. We finish with open problems in section 7.
2. Preliminaries
In this section we survey the classical definitions from combinatorics and multivariate polynomials. We refer to Stanley [34,35] for a more detailed treatment of the
material in this section.
For an integer k ≥ 0 we say that κ = (κ1 , κ2 , . . .) is a partition of k (denoted
κ ` k) if κ1 ≥ κ2 ≥ · · · ≥ 0 are integers such that κ1 + κ2 + · · · = k. The quantity
|κ| = k is called the size of κ.
All partitions can be partially ordered: we say that µ ⊆ λ if µi ≤ λi for all
i = 1, 2, . . .. Then λ/µ is called a skew shape and consists of those boxes in the
Young diagram of λ that do not belong to µ. Clearly, |λ/µ| = |λ| − |µ|.
The skew shape κ/µ is called a horizontal strip when κ1 ≥ µ1 ≥ κ2 ≥ µ2 ≥
· · · [35, p. 339].
The upper and lower hook lengths at a point (i, j) in a partition κ (i.e., i ≤
κ0j , j ≤ κi ) are, respectively,
h∗κ (i, j) ≡ κ0j − i +
2
(κi − j + 1)
β
and
hκ∗ (i, j) ≡ κ0j − i + 1 +
2
(κi − j).
β
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
3
The products of the upper and lower hook lengths are denoted, respectively, as
Y
Y
Hκ∗ =
h∗κ (i, j)
and
H∗κ =
hκ∗ (i, j).
(i,j)∈κ
(i,j)∈κ
Hκ∗ H∗κ .
Their product is denoted as jκ ≡
For a partition κ = (κ1 , κ2 , . . . , κn ) and β > 0, the Generalized Pochhammer
symbol is defined as
Y i−1
(β)
β+j−1
(2.1)
(a)κ ≡
a−
2
(i,j)∈κ
κi n Y
Y
i−1
a−
=
β+j−1
2
i=1 j=1
n Y
i−1
a−
=
β
,
2
κi
i=1
where (a)k = a(a + 1) · · · (a + k − 1) is the rising factorial.
For a partition κ = (k) in only one part (a)(β)
κ = (a)k is independent of β.
The multivariate Gamma function of parameter β is defined as
m
Y
m(m−1)
i−1
m−1
β
4
Γ(β)
(c)
≡
π
Γ
c
−
β
for <(c) >
β.
(2.2)
m
2
2
i=1
Definition 2.1 (Jack function).
The Jack function Jκ(β) (X) = Jκ(β) (x1 , x2 , . . . , xm ) is a symmetric, homogeneous
polynomial of degree |κ| in the eigenvalues x1 , x2 , . . . , xm of X. It can be defined
recursively [34, Proposition 4.2]:
Jκ(β) (x1 , x2 , . . . , xn ) = 0, if κn+1 > 0;
(β)
J(k)
(x1 ) = xk1 1 + β2 · · · 1 + (k − 1) β2 ;
X
Jκ(β) (x1 , x2 , . . . , xn ) =
Jµ(β) (x1 , x2 , . . . , xn−1 )x|κ/µ|
βκµ ,
n
n ≥ 2,
(2.3)
µ⊆κ
where the summation in (2.3) is over all partitions µ ⊆ κ such that κ/µ is a
horizontal strip, and
Q
∗
κ
hν (i, j), if κ0j = µ0j ;
(i,j)∈κ Bκµ (i, j)
ν
,
where Bκµ (i, j) ≡
βκµ ≡ Q
(2.4)
µ
hν∗ (i, j), otherwise.
(i,j)∈µ Bκµ (i, j)
Definition 2.2 (Hypergeometric function of matrix arguments). Let p ≥ 0
and q ≥ 0 be integers, and let X and Y be m × m complex symmetric matrices
with eigenvalues x1 , x2 , . . . , xm and y1 , y2 , . . . , ym , respectively. The hypergeometric
function of two matrix arguments X and Y , and parameter β > 0 is defined as
(β)
p Fq
(a1 , . . . , ap ; b1 , . . . , bq ; X, Y )
∞ X
(β)
X
1 (a1 )(β)
Cκ(β) (X)Cκ(β) (Y )
κ · · · (ap )κ
≡
·
·
. (2.5)
(β)
k! (b1 )(β)
Cκ(β) (I)
κ · · · (bq )κ
k=0 κ`k
January 25, 2014
4
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
Beta ensemble (m × m)
Joint eigenvalue density
π
Hermite
m(m−1)β
4
(2π)
m
2
·
(Γ(1 + β/2))m
(β)
e−
tr(Λ2 )
2
dµ(Λ)
Γm (1 + β/2)
n
Wishart (n DOF, cov. Σ)
n
|Σ|− 2 β
|Λ| 2 β−r · 0 F0(β) (− β2 Λ, Σ−1 )dµ(Λ)
(β) n
Km
( 2 β)
Laguerre (with param. a)
tr(Λ)
1
|Λ|a−r e− 2 β dµ(Λ)
Km (a)
Jacobi (with param. a, b)
MANOVA (param. n, p
and covariance Ω)
(β)
(β)
Sm
1
|Λ|a−r |I − Λ|b−r dµ(Λ)
(a, b)
(β)
pβ
p
p
(n + p)
Km
|Ω| 2 |Λ| 2 β−r |I − Λ|− 2 β−r
(β)
(β)
Km (n)Km (p)
−1
× 1 F0(β) ( n+p
, Ω)dµ(Λ)
2 ; Λ(Λ − I)
Table 1. Joint eigenvalue densities of the beta ensembles (where r ≡
m−1
β
2
+ 1).
For one matrix argument X, we define
(β)
p Fq
(a1 , . . . , ap ; b1 , . . . , bq ; X) ≡ p Fq(β) (a1 , . . . , ap ; b1 , . . . , bq ; X, I).
(2.6)
2.1. A new measure—the Vandermonde determinant
The Vandermonde determinant to power β is a part of every multivariate integral
we consider. To avoid it appearing everywhere, we conveniently incorporate it into
a new measure µ
dµ(X) =
Y
|xi − xj |β dx1 dx2 · · · dxm ,
i<j
where X = diag(x1 , x2 , . . . , xm ).
2.2. Beta random matrix ensembles
The Beta random matrices are defined in terms of their joint eigenvalue distributions. A matrix is from a particular beta ensemble if its (unordered)
eigenvalues λ1 , λ2 , . . . , λm have joint density as in Table 2.2. We denote Λ =
diag(λ1 , λ2 , . . . , λm ) and for any matrix A we denote its determinant as |A|.
We require that the parameters above be such that β > 0 and a, b > m−1
2 β. We
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
5
define
m!( β2 )ma
(β) m
Γ(β)
m (a)Γm 2 β
,
Km (a) ≡ m(m−1) ·
m
Γ β2
π 2 β
m!Γ(β) m β
Γ(β) (a) Γ(β)
m (b)
(β)
Sm
(a, b) ≡ m(m−1)m 2 m · m (β)
.
β
Γ
(a
+
b)
β
m
Γ
π 2
(β)
(2.7)
(2.8)
2
The expression (2.8) is the value of the Selberg Integral [32].
We refer to [10] (Hermite), [8,17] (Wishart), [15] (Laguerre), [11] (Jacobi),
and [7] (MANOVA).
2.3. Empirical models for the classical random matrix ensembles
The following are the empirical models for the beta–ensembles above which are
valid for any β > 0.
• Hermite [10]:

N (0, 2) cm−1

 cm−1 N (0, 2)
1 
..
H=√ 
.
2



..
.
..
. c2
c2 N (0, 2) c1
c1 N (0, 2)



,



where ci ∼ χiβ , i = 1, 2, . . . , m − 1, are independent and all elements on
the diagonal are i.i.d. N (0, 2).
• Wishart [8,17]: The procedure is iterative. Let the eigenvalues of the covari(β)
(β)
ance Σ be d21 , d22 , . . . , d2m . The Wishart model is W = β1 (Zm
(n))T Zm
(n),
where


τ1
χβ di
 ..

..


.
.
Zi(β) (n) = 
.


τi−1
χβ di
χ(n−i+1)β di
(β)
Here τ1 , τ2 , . . . , τi−1 are the singular values of Zi−1
(n). The base (m = 1)
(β)
case is Z1 (n) = χnβ d1 .
Note than the number of degrees of freedom n need not be integer.
• Laguerre [10,33]: L ≡ β1 BB T , where

χ2a

 χβ(m−1) χ2a−β

β


B=
 , a > (m − 1).
.
.
..
..
2


χβ χ2a−β(m−1)
January 25, 2014
6
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
• Jacobi [14] (see also [23,26]):

T
J ≡ Z Z,
where
cm −sm c0m−1


.

cm−1 s0m−1 . .
,

..
0
. −s2 c1 
c1 s01


Z≡


with
ck ∼
q
Beta a −
m−k
2 β, b
−
m−k
2 β
,
q
c0k ∼ Beta k2 β, a + b − 2m−k−1
β ,
2
sk =
s0k =
q
1 − c2k ;
q
1 − c02
k.
• MANOVA [7] Let Λ be m × m beta–Wishart with n degrees of freedom
and covariance Ω and let M be m × m Wishart with p degree of freedom
and covariance Λ−1 . Then
(M −1 + I)−1
is m × m beta–MANOVA with parameters n, p and covariance Ω.
3. Identities involving p Fq(β) (a1:p ; b1:q ; X, Y )
Let
r≡
m−1
2 β
+ 1;
X ≡ diag(x1 , x2 , . . . , xm );
Y ≡ diag(y1 , y2 , . . . , ym );
X̄ ≡ diag(x2 − x1 , x3 − x1 , . . . , xm − x1 );
Ȳ ≡ diag(y2 − y1 , y3 − y1 , . . . , ym − y1 );
Y
∆(X) ≡
(xi − xj );
i<j
∆(Y ) ≡
Y
i<j
(yi − yj ).
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
7
3.1. Identities for p Fq(β)
(β)
0 F0
(β)
0 F0
(X) = etr(X) ;
tr(X)
(3.1)
(β)
0 F0 (X, Y − I)
y1 tr(X)+x1 tr(Y )−mx1 y1
(X, Y ) = e
=e
(β)
1 F0
(a; X) = |I − X|
−a
,
(3.2)
(β)
1 F1
m
( m−1
2 β; 2 β; X̄, Ȳ
);
(X < I);
(3.3)
(β)
−a
· 1 F0(β) (a; X(I − X)−1 , Y − I);
1 F0 (a; X, Y ) = |I − X|
Q
(β) m
−β
2 ;
1 F0 ( 2 β; X, Y ) =
i,j (1 − xi yj )
(β)
1 F1 (a; c; X)
(β) β m
1 F1 ( 2 ; 2 β; X)
(β) β
m
2 F1 ( 2 , n; 2 β; X)
(3.6)
β m
t
1 F1 ( 2 ; 2 β; X − tI) · e ;
(1 − t)n · 2 F1(β) ( β2 , n; m
2 β; (1
(3.7)
=
=
(β)
− t)X + tI), (t 6= 1);
Γm (c)Γm (c − a − b)
;
(β)
Γ(β)
m (c − a)Γm (c − b)
(1)
(see [19] for the definition of A);
2 F1 (a, b; c; xI) = Pf(A)
(β)
(β)
c − a, b; c; −X(I − X)−1 · |I − X|−b
2 F1 (a, b; c; X) = 2 F1
(β)
2 F1
(a, b; c; I) =
c−a−b
= 2 F1 (c − a, c − b; c; X) · |I − X|
(a, b; c; X) = 2 F1(β) a, b; a + b + r − c; I − X
× 2 F1(β) (a, b; c; I),
(β)
p Fq
( β2 , a2:p ; b1:q ; xI)
(2)
p Fq
=
(a1:p ; b1:q ; X) =
(3.8)
(β)
(β)
(β)
(3.5)
= etr(X) · 1 F1(β) (c − a; c; −X);
(β)
2 F1
(3.4)
(3.9)
(3.10)
(3.11)
;
(a or b ∈ Z≤0 );
m
p Fq ( 2 β, a2:p ; b1:q ; x);
(3.12)
(3.13)
m m−j
xi p Fq (a1:p − j + 1; b1:q − j + 1; xi ) i,j=1 ;
(3.14)
∆(X)
Qq
m
Y
(i − 1)! j=1 (bj − m + 1)m−i
(2)
Q
p
p Fq (a1:p ; b1:q ; X, Y ) =
j=1 (aj − m + 1)m−i
i=1
m p Fq (a1:p − m + 1; b1:q − m + 1; xi yj ) i,j=1 ×
;
(3.15)
∆(X)∆(Y )
mt
X
X C (β) (X −1 )
Γ(β)
(β)
m (r + t)
κ
t
F
(r,
−t;
−X)
=
|X|
, (t ∈ Z≥0 ); (3.16)
2 0
k!
Γ(β)
m (r)
k=0 κ`k, κ1 ≤r
(β)
p Fq
(a1:p ; b1:q ; X) = lim
c→∞
(β)
p+1 Fq
(a1:p , c; b1:q ; 1c X)
(β)
= lim p Fq+1
(a1:p ; b1:q , c; cX).
c→∞
(3.17)
The references for the above identities are:
(3.1)
(3.2)
(3.3)
(3.4)
(3.5)
[16, (13.3), p. 593] (also [29, p. 262] for β = 1 and [31, p. 444] for β = 2);
[2, section 6] (also [27] for β = 1);
[16, p. 593, (13.4)] (also [29, p. 262] for β = 1 and [31, p. 444] for β = 2);
[27, (6.29)];
[27, (6.10)];
January 25, 2014
8
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
(3.6) [16, p. 596, (13.16)] (also [29, (6), p. 265] for the case β = 1);
(3.7) New result, see Theorem 5.2. For m = 3, β = 1 this is due to Bingham [3,
Lemma 2.1];
(3.8) New result, see Theorem 5.2;
(3.9) [16, (13.14), p. 594] (also [5, p. 24, (51)] for the case β = 1);
(3.10) [19];
(3.11) [16, Proposition 13.1.6, p. 595]; see also [29, p. 265, (7)] for the case β = 1;
(3.12) [16, Proposition 13.1.7, p. 596]. The condition a or b ∈ Z≤0 implies that
the series expansion for 2 F1(β) terminates.
(3.13) New result, see Theorem 5.3. See also [18, (5.13)] for β = 2;
(3.14) [30, (33), p. 281];
(3.15) [18, Theorem 4.2] (there is a typo in [18, (4.8)]: βn−1 should be βn ), see also
[30, (34), p. 281];
(3.16) [13];
−|κ|
(3.17) [16, (13.5)]. Trivially implied by limx→∞ (x)(β)
= 1.
κ ·x
3.2. Integral identities
Define
−
c(β)
m ≡π
m(m−1)
β
4
m!
m
Y
Γ( i β)
2
i=1
and recall that r =
(β)
p+1 Fq
m−1
2 β
+ 1. Then
(a1:p , a; b1:q ; Y )
Z
1
= (β) (β)
e−tr(X) p Fq(β) (a1:p ; b1:q ; X, Y )|X|a−r dµ(X);
cm Γm (a) Rm
+
(β)
p+1 Fq+1 (a1:p , a; b1:q , b; Y
=
Γ( β2 )
1
(β)
Sm
(a, b − a)
Cκ(β) (Y −1 ) =
)
Z
(β)
[0,1]m
(3.18)
p Fq
(a1:p ; b1:q ; X, Y )|X|a−r |I − X|b−a−r dµ(X);
|Y |a
(β) (β)
cm Γm (a) · (a)(β)
κ
Z
R+
m
(β)
0 F0
(−X, Y )|X|a−r Cκ(β) (X)dµ(X).
(3.19)
(3.20)
The references for the above identities are:
(3.18) [27, (6.20)] (see also [13]);
(3.19) [27, (6.21)] (see also [13]);
(3.20) This is a conjecture of Macdonald [27, Conjecture C, section 8], proven by
Baker and Forrester [2, section 6], with Dubbs and Edelman [7] correcting
a typo in [2].
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
9
Define T ≡ diag(t1 , t2 , . . . , tn ). The following identities are due to Kaneko [22]
Z
m,n
Y
(xi − tk )
[0,1]m i,k=1
m
Y
(β)
xai (1 − xi )b dµ(X) = Sm
(a + r, b + r)
i=1
(4)
× 2 F1 β
Z
m,n
Y
β
(1 − xi tk )− 2
[0,1]m i,k=1
−m, β2 (a + b + n + 1) + m − 1; β2 (a + n); T ;
m
Y
(3.21)
(β)
xai (1 − xi )b dµ(X) = Sm
(a + n + r, b + r)
i=1
×
4
(β
)
2 F1
β
− mβ
2 , 2 (m
− 1) + a + 1; β(m − 1) + a + b + 2; T , (3.22)
(β)
is the value of the Selberg integral (2.8).
where Sm
3.3. Identities involving the Jack function
We will predominantly use the “C” normalization of the Jack function, but there
are other normalizations, Jκ(β) , Pκ(β) , and Q(β)
κ , related as in (3.25). The properties
of these normalizations are that Jκ(β) (X) is such that for |κ| = n the coefficient of
x1 x2 · · · xn in Jκ(β) (X) is n! [34, Theorem 1.1]. The zonal polynomial is Cκ(1) (X) and
Pκ(2) (X) = Q(2)
κ (X) = sλ (X) is the Schur function.
X
Cκ(β) (X) = (trX)k ;
(3.23)
κ`k
Jκ (xIm ) = (x β2 )|κ|
Jκ(β) (X) =
(β)
m
2β κ
jκ
C (β) (X)
2 |κ|
( β ) |κ|! κ
(3.24)
κ (β)
= Hκ∗ Q(β)
κ (X) = H∗ Pκ (X);
(β)
Jκ(β) (X) = |X| · J(κ
(X)
1 −1,κ2 −1,...,κm −1)
m
Y
(m − i + 1 + β2 (κi − 1)),
(3.25)
(3.26)
i=1
where κm > 0;
X κ C (β) (X)
Cκ(β) (I + X)
σ
=
;
σ Cσ(β) (I)
Cκ(β) (I)
σ⊆κ
Pµ̂(β) (X) = |X|N Pµ(β) (X −1 ),
where µ1 ≤ N and µ̂i = N − µm+1−i , i = 1, 2, . . . , m
(i.e., the partition µ̂ is the complement of µ in (N m ));
(3.27)
(3.28)
January 25, 2014
10
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
For partitions κ = (k) in only one part,
Ck(β) (I) =
β −1
m
2β k
2
k
X
;
(3.29)
k
(β)
k Cs (X)
s Cs(β) (I)
s=0
k
m X
tk−s k
Cs(β) (X);
=
β
m
s
2
k
β
s
s=0 2
Ck(β) (X + tI) = Ck(β) (I)
tk−s
(3.30)
(3.31)
For partitions κ = (1k )
(β)
(β)
2 k m
J(1
k ) (Im ) = ( β ) ( 2 β)(1k ) =
k−1
Y
(m − i);
(3.32)
i=0
(β)
C(1
k ) (Im ) =
k−1
k−1
Y m−i
1 2 k Y
(m − i) =
( β ) k!
;
jκ
iβ + 1
i=0
i=0 2
for β = 2 this is
m
k
. (3.33)
The references for the above identities are:
(3.23) This is the definition of the normalization Cκ(β) (X) [34, Theorem 2.3];
(3.24) [34, Theorem 5.4];
(3.25) [22, (16)], [28, (10.22), p. 381]. Then Pκ(1) = Q(1)
κ = sκ , the Schur function
[34, Proposition 1.2];
(3.26) [34, Propositions 5.1 and 5.5];
(3.27) [22, (33)]. This also serves as a definition for the generalized binomial coefficient σκ ;
(3.28) This is a result of Macdonald [27, (4.5), (6.22)]: Both sides have xµ̂1 1 · · · xµ̂mm
as the leading term and for the scalar product h· , ·i0n defined in [28,
(10.35)]), h|X|N Pµ (X −1 ), |X|N Pν (X −1 )i0n = hPµ (X), Pν (X)i0n , which is 0
for µ 6= ν [28, (10.36)].
(3.29) Directly from (3.24);
(3.31) Directly from (3.27);
(3.32) Directly from (3.24);
(3.33) Directly from (3.25).
3.4. Multivariate Gamma function identities
We found the following identities useful in simplifying expressions involving the
multivariate Gamma function.
Γ(x)
Γ(β)
m (x)
.
=
Γ x− m
Γm x − β2
2β
(β)
Γ(β)
m (x + a)
= 1.
ma
x→∞ Γ(β)
m (x)x
lim
(3.34)
(3.35)
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
11
The first identity (3.34) is due to Pierre-Antoine Absil in the case β = 1. In general,
from the definition (2.2)
m
Y
Γ x − i−1
Γ(x)
Γ(β)
m (x)
2 β
.
=
=
(β)
i
β
Γ x− m
Γ x − 2β
Γm x − 2
2β
i=1
The second identity, (3.35)], follows directly from the definition (2.2). See also [13].
4. Distributions of the extreme eigenvalues
In this section we review the known explicit formulas for the densities and distributions of the extreme eigenvalues of the beta ensembles.
4.1. Wishart
These results are from [13]. Recall again that r = m−1
2 β + 1. Then
x −1 n2 β
βΣ Γ(β) (r)
· 2tr( x βΣ−1 ) 1 F1(β) r; n2 β + r; x2 βΣ−1 ;
P (λmax (A) < x) = (β) mn
Γm 2 β + r
e 2
and when t ≡
n
2β
(4.1)
− r is a nonnegative integer, then
x
P (λmin (A) < x) = 1 − etr(− 2 βΣ
−1
)
X
κ⊆(mt )
1 (β) x −1
C ( βΣ ).
|κ|! κ 2
(4.2)
It is an open problem to obtain an expression that does not have the requirement
for t to be a nonnegative integer—see section 7.
For the trace we have
∞
nβ x X
dens trA (x) = x2 βΣ−1 2 e− 2z β
k=0
×
X
κ`k
(β)
n
2β κ
1
·
Γ( nm
2 β + k)
1
·
· Cκ(β) (I − zΣ−1 ),
k!
β
2z
k−1
x
2z β
(4.3)
where z is arbitrary. Muirhead [29, p. 341] suggests the value z = 2σ1 σm /(σ1 +σm ),
where σ1 ≥ · · · ≥ σm > 0 are the eigenvalues of Σ.
The second sum in (4.3) is the marginal sum of 1 F0(β) ( n2 β; I − zΣ−1 ) thus the
truncation of (4.3) for |κ| ≤ M can be computed using mhg as follows. If s is the
vector of eigenvalues of Σ, these marginal sums (for say k = 0 through M ) are
returned in the variable c by the call:
[f ,c]=mhg(M ,α,n/α,[],1-z./s),
where α = 2/β, making the remaining computation trivial.
Empirically, the trace can be generated off the Wishart model from [8], which
upon closer observation reveals that
1
tr(A) ∼ (χ2nβ σ1 + χ2nβ σ2 + · · · + χ2nβ σm ).
β
January 25, 2014
12
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
4.2. Laguerre
The eigenvalues of the Laguerre ensemble are Wishart distributed with Σ = I and
n = a β2 .
Interestingly, we have explicit expressions for the densities of the largest and
smallest eigenvalue which are not related in an obvious way to the distributions.
In these expressions the matrix argument is of size m − 1 making them faster to
compute (fewer partitions to sum over) than the straightforward derivatives of the
distributions.
N EW ST U F F
(4.4)
Γm ( m−1
2 β
(β)
dens λmin (L) (x) =
m
2β
·
+ 1)
e−
2
m
2 β + 1, −t; − βx Im−1 .
Γ(β)
m (a)
× 2 F0(β)
tm
·
x
2β
mx
2 β
(4.5)
The density (4.4) is new; it generalizes the β = 1 result of Sugiyama [36]—see
Theorem ??. The result for the smallest eigenvalue (4.5) is a generalization of the
same result of Krishnaiah and Chang [25] for β = 1. Dumitriu [9, Theorem 10.1.1,
p. 147] established (4.5), but did not provide the scaling constant. We give full
derivation in Theorem 5.5.
Forrester [15] used (3.21) to obtain the following expressions for the smallest
eigenvalue
P (λmin < x) = 1 − e−
mxβ
2
4
)
(β
(−m; 2t/β, −xIt );
1 F1
(β)
dens λmin (L) (x) = mxr e−
mxβ
2
Km−1 (a + β)
1 F1
(β)
Km
(a)
4
(β
)
(4.6)
(−m + 1; 2t/β + 2; −xIt ), (4.7)
(β)
with Km
(a) as in (2.7).
(4)
The expression (4.6) is identical (as a function of x) to (4.2)—the 1 F1 β function
in (4.6) and the truncated 0 F0(β) function in (4.2) are the same polynomial in x of
degree mt.
4.3. Jacobi
Let the m × m matrix C be Jacobi distributed with parameters a, b. Define
(β)
Dm
(a, b) ≡
and recall that r =
m−1
2 β
(β)
Γ(β)
m (a + b)Γm (r)
(β)
Γm (a + r)Γ(β)
m (b)
+ 1. Then [11]
(β)
P (λmax (C) < x) = Dm
(a, b) · xma · 2 F1(β) (a, r − b; a + r; xI);
(4.8)
b−r ma−1
(β)
dens λmax (C) (x) = Dm (a, b) · ma(1 − x)
(β)
× 2 F1 (a −
β
2,r
x
− b; a + r; xIm−1 ).
(4.9)
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
13
When t ≡ b − r is a nonnegative integer, the above expressions are finite and (3.12)
yields the following alternatives:
mt
X
(β)
(a)(β)
κ Cκ ((1 − x)I)
;
k!
k=0 κ`k, κ1 ≤t
Γ a + b − m−1
β Γ(b + β2 Γ β2
2
dens λmax (C) (x) =
m
Γ b − m−2
2 β Γ(t + 1)Γ 2 β Γ(a)
P (λmax (C) < x) = xma
X
× (1 − x)t xma−1 2 F1(β) (a − β2 , −t; −t − β; xIm−1 ).
(4.10)
(4.11)
0
For the smallest eigenvalue we use that the matrix C = I − C is Jacobi distributed with parameters b, a:
P (λmin (C) < x) = 1 − P (λmax (C 0 ) < 1 − x)
dens λmin (C) (x) = dens λmax (C 0 ) (1 − x).
(4.12)
(4.13)
The distribution (4.8) is due to Dumitriu and Koev [11]. The density result (4.9)
is new (see section 5.3). For β = 1, the above results are due to Constantine [6,
(61)] (eq. (4.8)), Absil, Edelman, and Koev [1] (eq. (4.9)), and Muirhead [29, (37),
p. 483] (eq. (4.10)).
5. New results
We now prove the identities (3.2), (3.7), (3.8), and (3.13) (section 5.1), and prove
the formulas for the densities of the smallest eigenvalue of the Jacobi ensemble,
(4.9) (section 5.3) and the Laguerre ensemble (4.5) (section 5.4).
5.1. New identities for p Fq(β)
Theorem 5.1. With the notation from Section 3.1 the identity
(β)
0 F0
(X, Y ) = etrX 0 F0(β) (X, Y − I)
m
= ey1 trX+x1 trY −mx1 y1 1 F1(β) ( m−1
2 β; 2 β; X̄, Ȳ )
(5.1)
holds for any β > 0.
Proof. The first part is from Baker and Forrester [2, sec. 6]. Let Ik be the identity
matrix of size k. Since Cκ(β) (X − x1 Im ) = Cκ(β) (X̄) and analogously for Y ,
(β)
0 F0
(X, Y ) = ey1 trX · 0 F0(β) (X, Y − y1 Im )
= ey1 trX+x1 trY −mx1 y1 · 0 F0(β) (X − x1 Im , Y − y1 Im )
∞ X
X
1 Cκ(β) (X − x1 Im )Cκ(β) (Y − y1 Im )
·
= ey1 trX+x1 trY −mx1 y1 ·
k!
Cκ(β) (Im )
= ey1 trX+x1 trY −mx1 y1 ·
k=0 κ`k
∞ X
X
k=0 κ`k
1 Cκ(β) (Im−1 ) Cκ(β) (X̄)Cκ(β) (Ȳ )
·
·
k! Cκ(β) (Im )
Cκ(β) (Im−1 )
m
= ey1 trX+x1 trY −mx1 y1 · 1 F1(β) ( m−1
2 β; 2 β; X̄, Ȳ ).
January 25, 2014
14
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
The following two theorems use the fact that ( β2 )(β)
κ = 0 for partitions κ in more
than one part.
Theorem 5.2. Let the matrices X and I be m × m. Let t and n be real numbers
and let β > 0. The following identities hold:
β m
2 ; 2 β; X + tI)
(β) β
m
2 F1 ( 2 , n; 2 β; X)
(β)
1 F1
β m
t
2 ; 2 β; X) · e ;
t)n 2 F1(β) ( β2 , n; m
2 β; (1
= 1 F1(β)
= (1 −
(5.2)
− t)X + tI),
t 6= 1.
Proof. We transform the left hand side of (5.2) using (3.29) and (3.31):
β
∞
X
2
1
(β) β m
· m k · Ck(β) (X + tI)
1 F1
2 ; 2 β; X + tI =
k!
2β k
k=0
∞
X
1 Ck(β) (X + tI)
·
k!
Ck(β) (I)
k=0
∞
k
X
1 X k−s
k
C (β) (X)
t
·
· s(β)
=
s
k! s=0
Cs (I)
k=0
∞
∞
(β)
X C (X) X tj−i j i
=
·
·
(β)
j!
i
C
i (I)
i=0
j=i
β
∞
∞
X
X
2
1
tj−i
=
· m k · Ck(β) (X) ·
i!
(j − i)!
2β k
i=0
j=i
=
= 1 F1(β)
β m
2 ; 2 β; X)
· et .
We use (3.31) again to obtain (5.3):
(1 − t)n 2 F1(β) β2 , n; m
2 β; (1 − t)X + tI
β
∞
(n)k
X
2
1
k = (1 − t)n
·
· Ck(β) ((1 − t)X + tI)
m
k!
β
2
k
k=0
(β)
∞
k
X (n)k X
k Cs ((1 − t)X)
= (1 − t)n
tk−s
k! s=0
s
Cs(β) (I)
k=0
∞
∞ j−i X
X
j
Ci(β) (X)
t
n
i
= (1 − t)
(1 − t)
(n)j
(β)
j!
i
C
(I)
i
i=0
j=i
=
∞
∞ l
X
X
1 Ci(β) (X)
t
· (β)
(n)i · (1 − t)n+i
(n + i)l
i! Ci (I)
l!
i=0
l=0
|
{z
}
1
= 2 F1(β) ( β2 , n; m
2 β; X).
(5.3)
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
15
Theorem 5.3.
(β)
p Fq
( β2 , a2:p ; b1:q ; xI) = p Fq ( m
2 β, a2:p ; b1:q ; x).
Proof. Using (3.29),
(β) β
p Fq ( 2 , a2:p ; b1:q ; xI) =
=
∞
β
X
1 ( 2 )k (a2 )k · · · (ap )k (β)
·
Ck (xI)
k!
(b1 )k · · · (bq )k
k=0
∞
X
k=0
β
β)k
1 ( 2 )k (a2 )k · · · (ap )k k ( m
·
x 2β
k!
(b1 )k · · · (bq )k
( 2 )k
= p Fq ( m
2 β, a2:p ; b1:q ; x).
5.2. Density of the largest eigenvalue of the Jacobi ensemble
We repeat the argument in [1] to extend the result for the density of the smallest
eigenvalue of the Jacobi ensemble to any β > 0 (see equations (4.9) and (4.13)). We
start with the joint density of the eigenvalues of the Jacobi ensemble from Table 2.2:
m
Y
Y
1
λa−r
(1 − λi )b−r
|λi − λj |β dλ1 · · · dλm .
(β)
i
Sm (a, b) i=1
i<j
dens(λ1 , λ2 , . . . , λm ) ≡
where, as before, r ≡
m−1
2 β
+ 1. Let x ≡ λ1 ≥ λ2 ≥ · · · λ1 .
Z
dens(x) = m
dens(x, λ2 , . . . , λm )dλ2 · · · dλm
Z
Y
a−r
b−r
|λi − λj |β
x
(1 − x)
[0,x]m−1
=
m
Sm (a, b)
(β)
×
m
Y
[0,x]m−1 1<i<j
|x − λi |β λa−r
(1 − λi )b−r dλ2 · · · dλm .
i
i=2
We change variables λi = xti , i = 2, 3, . . . , m.
dens(x) =
m
xam−1 (1 − x)b−r
(β)
Sm
(a, b)
×
Z
|ti − tj |β
[0,1]m−1
m−1
Y
(1 − xti )b−r ta−r
(1 − ti )β dt2 · · · dtm .
i
i=1
From (3.19) we have for T ≡ diag(t2 , t3 , . . . , tm−1 ), a matrix of size m − 1,
Z
Y
1
(β)
β
m+1
r−b, a+ 2 ; a+ 2 β+1; xIm−1 =
|ti −tj |β
2 F1
(β)
β
β
[0,1]m−1 i<j
Sm−1
+
r,
b
−
2
2
×
m
Y
i=2
(1 − xti )b−r ta−r
(1 − ti )β dt2 · · · dtm−2 .
i
January 25, 2014
16
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
Qm
(Note that 1 F0(β) (r − a, xIm−1 , T ) = 1 F0(β) (r − a, xT ) = i=2 (1 − xti )a−r . )
Therefore
(β)
β
β
Sm−1
+
r,
b
−
2
2
dens(x) = m
x(a−r)m (1 − x)mb−1
(β)
Sm
(a, b)
× 2 F1(β) −a + r, β2 + r; b + r; x−1
I
,
m−1
x
which using (3.11) gives
(β)
Sm−1
β
2
+ r, b −
β
2
x(a−r)m (1 − x)mb−1
(β)
(a, b)
Sm
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 x−(a−r)(m−1)
(β)
β
β
Sm−1
+
r,
b
−
2
2
xa−r (1 − x)mb−1
=m
(β)
(a, b)
Sm
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 ,
dens(x) = m
(β)
= mb · Dm
(b, a) · xa−r (1 − x)mb−1
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 ,
which yields (4.9) and (4.13).
5.3. Density of the smallest eigenvalue of the Jacobi ensemble
We repeat the argument in [1] to extend the result for the density of the smallest
eigenvalue of the Jacobi ensemble to any β > 0 (see equations (4.9) and (4.13)). We
start with the joint density of the eigenvalues of the Jacobi ensemble from Table 2.2:
dens(λ1 , λ2 , . . . , λm ) ≡
m
Y
Y
1
λa−r
(1 − λi )b−r
|λi − λj |β dλ1 · · · dλm .
(β)
i
Sm (a, b) i=1
i<j
where, as before, r ≡ m−1
2 β + 1. We integrate all but the smallest eigenvalue (call
it x) out of the above density:
Z
dens(x) = m
dens(λ1 , . . . , λm−1 , x)dλ1 · · · dλm−1
[x,1]m−1
Z
Y
m
= (β)
xa−r (1 − x)b−r
|λi − λj |β
Sm (a, b)
[x,1]m−1 i<j≤m−1
×
m−1
Y
i=1
|λi − x|β λa−r
(1 − λi )b−r dλ1 · · · dλm−1 .
i
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
17
We change variables λi = (1 − x)ti + x, i = 1, 2, . . . , m − 1.
Z
m
x(a−r)m (1 − x)mb−1
|ti − tj |β
dens(x) = (β)
Sm (a, b)
[0,1]m−1
a−r
m−1
Y β
x−1
×
ti
(1 − ti )b−r dt1 · · · dtm−1 .
ti 1 −
x
i=1
From (3.19) we have for T ≡ diag(t1 , t2 , . . . , tm−1 ), a matrix of size m − 1,
Z
Y
1
(β)
β
|ti − tj |β
r − a, 2 + r; b + r; zIm−1 =
2 F1
(β)
β
β
[0,1]m−1 i<j
Sm−1
+
r,
b
−
2
2
×
m−1
Y
(1 − zti )a−r tβi (1 − ti )b−r dt1 · · · dtm−1 .
i=1
Qm−1
(Note that 1 F0(β) (−a + r, zIm−1 , T ) = 1 F0(β) (−a + r, zT ) = i=1 (1 − zti )a−r . )
Now, for z = x−1
x
(β)
β
β
Sm−1
+
r,
b
−
2
2
x(a−r)m (1 − x)mb−1
dens(x) = m
(β)
(a, b)
Sm
× 2 F1(β) −a + r, β2 + r; b + r; x−1
x Im−1 ,
which using (3.11) gives
(β)
Sm−1
β
2
+ r, b −
β
2
x(a−r)m (1 − x)mb−1
(β)
(a, b)
Sm
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 x−(a−r)(m−1)
(β)
β
β
Sm−1
2 + r, b − 2
=m
xa−r (1 − x)mb−1
(β)
Sm
(a, b)
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 ,
dens(x) = m
(β)
= mb · Dm
(b, a) · xa−r (1 − x)mb−1
× 2 F1(β) −a + r, b − β2 ; b + r; (1 − x)Im−1 ,
which yields (4.9) and (4.13).
5.4. The density of the smallest eigenvalue of the Laguerre
ensemble
We will obtain this result as a limiting argument from the density of the smallest
eigenvalue of the Jacobi ensemble. The connection is that if λi are the eigenvalues
2bλi
of a Jacobi matrix, then λ̄i ≡ limb→∞ β(1−λ
are the eigenvalues of a Laguerre
i)
matrix.
January 25, 2014
18
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
Proposition 5.1. Let L(β)
m (a; Λ) be the joint eigenvalue density of the Laguerre
(β)
ensemble and Jm (a, b; Λ) be the joint eigenvalue density of the Jacobi ensemble as
defined in Table 2.2. Then
1
(β)
lim 2 m · Jm
a, b; Λ( β2 bI + Λ)−1 = L(β)
m (a; Λ).
b→∞ ( b)
β
Proof.
(β)
Jm
a, b; Λ( β2 bI + Λ)−1
( β2 b)m
=
1
π
=
1
· (β)
(a, b)
( β2 b)ma Sm
m
Y a− m−1 β−1
×
λi 2
1
i=1
m(m−1)
β
2
Γ
( β2 )ma m!Γ(β)
m
×
m
Y
β
2
+
2−a−b
β
2b λi
Y
|λj − λk |β
j<k
m
Γ(β)
m (a + b)
(β)
m
ma
Γ(β)
m (b)b
2 β Γm (a)
a− m−1
2 β−1
λi
1+
·
2−a−b
λi
2b β
i=1
Y
|λj − λk |β .
j<k
We then take the limit as b → ∞ and use (3.35) to obtain the desired result.
Next, we derive the density of the largest eigenvalue of the Laguerre ensemble.
Theorem 5.4. The density of the largest eigenvalue of the Laguerre ensemble is
nmβ
(β)
2 −1
nmβ 2 Γm (r) − mβx βx
(β) mβ
2
e
+ 1; nβ
+ r; xβ
Im−1 .
(5.4)
1 F1
2
2
2
(β) nβ
2
4Γm
+r
2
Proof. The density of the maximum eigenvalue of the Jacobi ensemble with parameters a and b is given by
(β)
(β)
β
Γm (a + b)Γm (r)
(β)
b−r ma−1
a
−
f (x) = (β)
ma(1
−
x)
x
F
,
r
−
b;
a
+
r;
xI
2 1
m−1
(β)
2
Γm (a + r)Γm (b)
(β)
(β)
Γm (a + b)Γm (r)
β
x
(β)
b−1−a+ β
2 xma−1 F
= (β)
ma(1
−
x)
a
−
,
a
+
b;
a
+
r;
−
I
(5.5)
2 1
m−1
(β)
2
1−x
Γm (a + r)Γm (b)
where r = (m − 1)β/2 + 1. Let y = (2b/β)x/(1 − x), we have
1
βy
= 1+
,
1−x
2b
x=
dx =
1
βy
2b
+ βy
2b
,
β(1 − x)2
dy,
2b
(5.6)
(5.7)
(5.8)
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
19
and the density function of y is given by
(β)
(β)
β
βy
Γm (a + b)Γm (r) maβ
(β)
b+1−a− β
2 xma−1 F
(1
−
x)
a
−
,
a
+
b;
a
+
r;
−
I
2 1
m−1
(β)
(β)
2
2b
Γm (a + r)Γm (b) 2b
ma−1 −(m−1)a−b− β2
(β)
(β)
b−ma Γm (a + b)Γm (r) maβ βy
βy
=
1+
(β)
(β)
2
2
2b
Γm (b)Γm (a + r)
βy
β
(β)
× 2 F1
(5.9)
a − , a + b; a + r; − Im−1 .
2
2b
Using the fact that
f (y) =
(β)
lim
b−ma Γm (a + b)
b→∞
(β)
= 1,
(5.10)
Γm (b)
−(m−1)a−b− β2
βy
βy
lim 1 +
= e− 2 ,
(5.11)
b→∞
2b
β
βy
βy
β
(β)
(β)
a − , a + b; a + r; − Im−1 = 1 F1
lim 2 F1
a − ; a + r; − Im−1
b→∞
2
2b
2
2
βy(m−1)
β
βy
(β)
−
2
=e
r + ; a + r;
I(5.12)
1 F1
m−1
2
2
we have
ma−1
(β)
maβ
βy
βy
mβ
Γm (r)
(β)
− mβy
2
+ 1; a + r;
Im−1 ,
e
lim f (y) = (β)
1 F1
b→∞
2
2
2
2
Γm (a + r)
(5.13)
and this is the density of the maximum eigenvalue of Laguerre ensemble with parameter a. Putting a = nβ/2, we prove (5.4).
We now derive the scaling constant in the formula for the density of the Laguerre
ensemble, (4.5). The proportionality result is due to Dumitriu [9, Theorem 10.1.1,
p. 146].
Theorem 5.5. When t ≡ a − m−1
2 β − 1 is a nonnegative integer, the density of
the smallest eigenvalue of a Laguerre matrix is
densλmin (A) (y) =
m−1
my
Γ(β)
m ( 2 β + 1)
· e− 2 β ( y2 β)mt
(β)
Γm (a)
2
× 2 F0(β) ( m
2 β + 1, −t; − βy Im−1 ).
m
2β
·
(5.14)
Proof. Let C be Jacobi distributed. From (4.9), (4.13), and the identity (3.11) we
have
(β)
densλmin (C) (x) = mb · Dm
(b, a) · (1 − x)mb−1 xt
× 2 F1(β) (b − β2 , −t; b +
(β)
m−1
2 β + 1; (1
mb−1 mt
= mb · Dm (b, a) · (1 − x)
(β)
× 2 F1
(m
2β
+ 1, −t; b +
− x)Im−1 )
x
m−1
2 β
+ 1; − 1−x
x Im−1 ).
January 25, 2014
20
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
β
(with the Jacobian being 2b
(1 − x)2 ) we get:
m−1
−mt
Γ(β)
Γ(β)
m
m (a + b) · b
2 β+1
m
densλmin ( 2b C(I−C)−1 ) (y) = 2 β ·
·
m−1
β
Γ(β)
Γ(β)
m (a)
m b+ 2 β+1
−mb−1 y mt
−mt
× 1 + βy
1 + βy
2b
2β
2b
Substituting y =
2b
β
·
x
1−x
× 2 F1(β) ( m
2 β + 1, −t; b +
m−1
2 β
2
+ 1; b(− βy
)Im−1 ).
We use Proposition 5.1, (3.17), and (3.35) to take the limit as b → ∞ and obtain (5.14).
6. Numerical experiments
All formulas for the distributions and densities in this paper can be approximated
(within the limits of the computational power of modern computers) using the
software mhg [24].
It returns the truncation
M X
(β)
X
C (β) (X)C (β) (Y )
1 (a1 )(β)
κ · · · (ap )κ
· κ (β) κ
·
.
(β)
(β)
k! (b1 )κ · · · (bq )κ
Cκ (I)
k=0 κ`k
of (2.5) as well as the partial sums
X 1 (a1 )(β) · · · (ap )(β) C (β) (X)C (β) (Y )
κ
κ
· κ (β) κ
·
(β)
k! (b1 )(β)
Cκ (I)
κ · · · (bq )κ
κ`k
for k = 0, 1, . . . , M with the additional option to restrict the summation for κ1 ≤ t
(for, e.g., (4.2)). All expressions in this paper are trivially approximated then using
these partial sums.
We performed extensive numerical tests to verify the correctness of the formulas
in this paper and present four examples in Figure 1.
7. Open problems and future work
We finish with two open problems.
Several formulas in this paper, e.g., (4.2), (4.10), are only valid when certain
parameters are integers. It is an open problem to derive formulas that do not have
that restriction. The central problem there is to find an explicit expression (perhaps
in terms of p Fq(β) ) for the integral
Z
(β)
a
b
0 F0 (−X, Y )|X| |I + X| dµ(X).
R+
m
This is the multivariate version of the function Ψ, the confluent hypergeometric
function of the second kind.
The second open problem we pose is to find explicit formulas for the extreme
eigenvalues of the Hermite ensemble, which are analogous to the ones for the other
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
Wishart, =4/3, =diag(0.7,1,1.3), n=5
Laguerre, =4/3, m=3, n=5
1
1
0.9
0.9
0.8
0.8
0.7
0.7
0.6
0.6
0.5
0.5
0.4
0.4
min
min
0.3
min
0.2
max
0.1
max
theoretical CDF
min
empirical CDF
0.3
min
theoretical CDF
0.2
min
empirical CDF
0.1
max
max
0
0
5
10
15
20
21
0
0
25
5
10
x
15
theoretical CDF
empirical CDF
theoretical PDF
empirical PDF
theoretical CDF
empirical CDF
20
25
x
Trace of Wishart, =4, =diag(0.5,0.7,1,1.4), n=7
max
of Jacobi, =3, a=8, b=11, m=4
6
0.1
0.09
theoretical PDF
empirical PDF
0.08
5
0.07
theoretical PDF
empirical PDF
theoretical CDF
empirical CDF
4
0.06
3
0.05
0.04
2
0.03
0.02
1
0.01
0
0
10
20
30
40
50
60
0
0
0.2
0.4
0.6
0.8
1
Fig. 1. Numerical experiments comparing the theoretical predictions of the eigenvalue distributions
for various ensembles vs. the empirical results.
ensembles. This problem boils down to evaluating the integral
Z
e−
tr(X 2 )
2
dµ(X).
[x,∞)m
Bornemann [4] presents algorithms for numerical evaluation of distributions in
Random Matrix Theory that are expressible as Painlevé transcendents or Fredholm
determinants in the β = 1, 2, 4 cases. It is an open problem to devise good numerical
routines for the general β versions of some of these quantities, especially the bulk,
hard edge, and soft edge statistics.
January 25, 2014
22
13:18
WSPC/INSTRUCTION FILE
mvs-paper
DRENSKY, EDELMAN, GENOAR, KAN, and KOEV
Acknowledgments
We thank Iain Johnstone, Tom Koornwinder, Richard Stanley, and Brian Sutton
for insightful discussions in the process of preparation of this paper.
This research was supported by the National Science Foundation Grant DMS–
1016086 and by the Woodward Fund for Applied Mathematics at San José State
University. The Woodward Fund is a gift from the estate of Mrs. Marie Woodward
in memory of her son, Henry Teynham Woodward. He was an alumnus of the
Mathematics Department at San José State University and worked with research
groups at NASA Ames.
References
[1] Pierre-Antoine Absil, Alan Edelman, and Plamen Koev, On the largest principal
angle between random subspaces, Linear Algebra Appl. 414 (2006), 288–294.
[2] T. H. Baker and P. J. Forrester, The Calogero-Sutherland model and generalized
classical polynomials, Comm. Math. Phys. 188 (1997), no. 1, 175–216.
[3] Christopher Bingham, An antipodally symmetric distribution on the sphere, Ann.
Statist. 2 (1974), 1201–1225.
[4] F. Bornemann, On the numerical evaluation of distributions in random matrix theory:
a review, Markov Process. Related Fields 16 (2010), no. 4, 803–866.
[5] R. W. Butler and A. T. A. Wood, Laplace approximations for hypergeometric functions with matrix argument, Ann. Statist. 30 (2002), no. 4, 1155–1177.
[6] A. G. Constantine, Some non-central distribution problems in multivariate analysis,
Ann. Math. Statist. 34 (1963), 1270–1285.
[7] Alexander Dubbs and Alan Edelman, The Beta-MANOVA ensemble with general
covariance, arXiv:1309.4328, 2013.
[8] Alexander Dubbs, Alan Edelman, Plamen Koev, and Praveen Venkataramana, The
Beta–Wishart ensemble, J. Math. Phys. 54 (2013), 083507.
[9] Ioana Dumitriu, Eigenvalue statistics for the Beta-ensembles, Ph.D. thesis, Massachusetts Institute of Technology, 2003.
[10] Ioana Dumitriu and Alan Edelman, Matrix models for beta ensembles, J. Math. Phys.
43 (2002), no. 11, 5830–5847.
[11] Ioana Dumitriu and Plamen Koev, The distributions of the extreme eigenvalues of
beta–Jacobi random matrices, SIAM J. Matrix Anal. Appl. 1 (2008), 1–6.
[12] Alan Edelman, The random matrix technique of ghosts and shadows, Markov Processes and Related Fields 16 (2010), no. 4, 783–790.
[13] Alan Edelman and Plamen Koev, Eigenvalues distributions of Beta–Wishart matrices, Submitted to Random Matrices: Theory and Applications.
[14] Alan Edelman and Brian D. Sutton, The beta-Jacobi matrix model, the CS decomposition, and generalized singular value problems, Found. Comput. Math. 8 (2008),
no. 2, 259–285.
[15] P. J. Forrester, Exact results and universal asymptotics in the Laguerre random matrix ensemble, J. Math. Phys. 35 (1994), no. 5, 2539–2551.
[16] Peter J. Forrester, Log-gases and random matrices, London Mathematical Society
Monographs Series, vol. 34, Princeton University Press, Princeton, NJ, 2010.
, Probability densities and distributions for spiked Wishart β-ensembles,
[17]
arXiv:1101.2261, 2011.
[18] Kenneth I. Gross and Donald St. P. Richards, Total positivity, spherical series, and
January 25, 2014
13:18
WSPC/INSTRUCTION FILE
mvs-paper
COMPUTING WITH BETA ENSEMBLES
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
23
hypergeometric functions of matrix argument, J. Approx. Theory 59 (1989), no. 2,
224–246.
Rameshwar D. Gupta and Donald St. P. Richards, Hypergeometric functions of scalar
matrix argument are expressible in terms of classical hypergeometric functions, SIAM
J. Math. Anal. 16 (1985), no. 4, 852–858.
Iain M. Johnstone, On the distribution of the largest eigenvalue in principal components analysis, Ann. Statist. 29 (2001), no. 2, 295–327.
, Multivariate analysis and Jacobi ensembles: largest eigenvalue, TracyWidom limits and rates of convergence, Ann. Statist. 36 (2008), no. 6, 2638–2716.
MR 2485010 (2010f:62146)
Jyoichi Kaneko, Selberg integrals and hypergeometric functions associated with Jack
polynomials, SIAM J. Math. Anal. 24 (1993), no. 4, 1086–1110.
Rowan Killip and Irina Nenciu, Matrix Models for Circular Ensembles, Int. Math.
Res. Not. 50 (2004), 2665–2701.
Plamen Koev and Alan Edelman, The efficient evaluation of the hypergeometric function of a matrix argument, Math. Comp. 75 (2006), no. 254, 833–846.
P. R. Krishnaiah and T. C. Chang, On the exact distribution of the smallest root
of the Wishart matrix using zonal polynomials, Ann. Inst. Statist. Math. 23 (1971),
293–295.
Ross A. Lippert, A matrix model for the β-Jacobi ensemble, J. Math. Phys. 44 (2003),
4807–4816.
Ian G. Macdonald, Hypergeometric functions I, arXiv:1309.4568.
, Symmetric functions and Hall polynomials, Second ed., Oxford University
Press, New York, 1995.
R. J. Muirhead, Aspects of multivariate statistical theory, John Wiley & Sons Inc.,
New York, 1982.
A. Yu. Orlov, New solvable matrix integrals, Proceedings of 6th International Workshop on Conformal Field Theory and Integrable Models, vol. 19, 2004, pp. 276–293.
T. Ratnarajah, R. Vaillancourt, and M. Alvo, Eigenvalues and condition numbers of
complex random matrices, SIAM J. Matrix Anal. Appl. 26 (2004/05), no. 2, 441–456.
Atle Selberg, Remarks on a multiple integral, Norsk Mat. Tidsskr. 26 (1944), 71–78.
Jack W. Silverstein, The smallest eigenvalue of a large-dimensional Wishart matrix,
Ann. Probab. 13 (1985), no. 4, 1364–1368.
Richard P. Stanley, Some combinatorial properties of Jack symmetric functions, Adv.
Math. 77 (1989), no. 1, 76–115.
, Enumerative combinatorics. Vol. 2, Cambridge Studies in Advanced Mathematics, vol. 62, Cambridge University Press, Cambridge, 1999.
T. Sugiyama, On the distribution of the largest latent root and the corresponding latent
vector for principal component analysis, Ann. Math. Statist. 37 (1966), 995–1001.
Download