On the expectation of the norm of random matrices Stiene Riemer

advertisement
o
u
r
nal
o
f
J
on
Electr
i
P
c
r
o
ba
bility
Electron. J. Probab. 18 (2013), no. 29, 1–13.
ISSN: 1083-6489 DOI: 10.1214/EJP.v18-2103
On the expectation of the norm of random matrices
with non-identically distributed entries
Stiene Riemer∗
Carsten Schütt†
Abstract
Let Xi,j , i, j = 1, ..., n, be independent, not necessarily identically distributed random
variables with finite first moments. We show that the norm of the random matrix
(Xi,j )n
i,j=1 is up to a logarithmic factor of the order of
(Xi,j )n
E max (Xi,j )n
j=1 2 .
j=1 2 + E max
i=1,...,n
i=1,...,n
This extends (and improves in most cases) the previous results of Seginer and Latała.
Keywords: Random Matrix; Largest Singular Value; Orlicz Norm.
AMS MSC 2010: 60B20.
Submitted to EJP on June 26, 2012, final version accepted on February 8, 2013.
1
Introduction and Notation
We study the order of magnitude of the expectation of the largest singular value, i.e.
the norm of random matrices with independent entries
E (ai,j gi,j )ni,j=1 2→2 ,
where ai,j ∈ R, i, j = 1, . . . , n, gi,j , i, j = 1, . . . , n, are standard Gaussian random variables and k k2→2 the operator norm on `n
2 . There are two cases with a complete answer.
Chevet [2] showed for matrices satisfying ai,j = ai bj that the expectation is proportional
to
kak2 kbk∞ + kak∞ kbk2 ,
where kak2 denotes the Euclidean norm of a = (a1 , . . . , an ) and kak∞ = max1≤i≤n |ai |.
For diagonal matrices with diagonal elements d1 , . . . , dn the expectation of the norm
is of theq
order of the Orlicz norm k(d1 , . . . , dn )kM where the Orlicz function is given by
Rs
1
M (s) = π2 0 e− 2t2 dt [4]. This Orlicz norm is up to a logarithm of n equal to the norm
max1≤i≤n |di |.
∗ Christian-Albrechts Universität, Kiel, Germany. E-mail: riemer@math.uni-kiel.de
† Christian-Albrechts Universität, Kiel, Germany. E-mail: schuett@math.uni-kiel.de
Norm of random matrices
These two cases are of very different structure and seem to present essentially what
might occur concerning the structure of matrices. This leads us to conjecture that the
expectation for arbitrary matrices is up to a logarithmic factor equal to
max (ai,j )nj=1 2 + max k(ai,j )ni=1 k2 .
(1.1)
j=1,...,n
i=1,...,n
Latała [5] showed for arbitrary matrices
E (ai,j gi,j )ni,j=1 2→2 ≤ C max (ai,j )nj=1 2 + max k(ai,j )ni=1 k2 + (ai,j )ni,j=1 4 .
j=1,...,n
i=1,...,n
n,m
Seginer [12] showed for any n × m random matrix (Xi,j )i,j=1 of independent identically
distributed random variables
n,m n
m E (Xi,j )i,j=1 2→2 ≤ c E max (Xi,j )j=1 2 + E max k(Xi,j )i=1 k2 .
1≤j≤n
1≤i≤n
The behavior of the smallest singular value has been determined in [3, 7, 9, 10, 14].
Theorem 1.1. There is a constant c > 0 such that for all ai,j ∈ R, i, j = 1, ..., n and all
independent standard Gaussian random variables gi,j , i, j = 1, ..., n,
E (ai,j gi,j )ni,j=1 2→2
!! (ai,j )n i,j=1 1
n
(ai,j gi,j )nj=1 + E
≤ c ln e E
max
max
k(a
g
)
k
.
i,j
i,j
i=1 2
2
(ai,j )n i=1,...,n
j=1,...,n
i,j=1 ∞
The norm ratio appearing in the logarithm is at most n2 . In the same way we prove
Theorem 1.1 we can show the similar formula
E (ai,j gi,j )ni,j=1 2→2
!! (ai,j )n i,j=1 1
(ai,j )nj=1 + max k(ai,j )ni=1 k .
max
≤ c ln e 2
2
(ai,j )n i=1,...,n
j=1,...,n
i,j=1 ∞
The proof of this formula is essentially the same as the proof of Theorem 1.1. Through
out the proof we use the expression max (ai,j )n
j=1 2 instead of the expression E
i=1,...,n
max (ai,j gi,j )nj=1 2 .
i=1,...,n
The inequality of Theorem 1.1 is generalized to arbitrary random variables as in [5].
Theorem 1.2. Let Xi,j , i, j = 1, ..., n, be independent, mean zero random variables.
Then
2
E (Xi,j )ni,j=1 2→2 ≤ c (ln(en)) E max (Xi,j )nj=1 2 + E max k(Xi,j )ni=1 k2 .
i=1,...,n
j=1,...,n
On the other hand,
E
max (ai,j gi,j )nj=1 2
i=1,...,n
+E
max k(ai,j gi,j )ni=1 k2
j=1,...,n
(1.2)
is obviously smaller than E (ai,j gi,j )n
i,j=1 2→2 . We show that the expression (1.2) is
equivalent to the Musielak-Orlicz norm of the vector (1, . . . , 1), where the Orlicz functions are given through the coefficients ai,j , i, j = 1, . . . , n. Our formula (Theorem 3.1)
enables us to estimate from below the expectation of the operator norm in many cases
efficiently.
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 2/13
Norm of random matrices
Moreover, we do not know of any matrix where the expectation of the norm is not of
the same order as (1.2).
A convex function M : [0, ∞) → [0, ∞) with M (0) = 0 that is not identically 0 is called
an Orlicz function. Let M be an Orlicz function and x ∈ Rn then the Orlicz norm of x,
kxkM , is defined by
(
)
n
X
|xi |
t > 0
M
≤1
t
kxkM = inf
.
i=1
We say that two Orlicz functions M and N are equivalent (M ∼ N ) if there are strictly
positive constants c1 and c2 such that for all s ≥ 0
M (c1 s) ≤ N (s) ≤ M (c2 s).
If two Orlicz functions are equivalent, so are their norms: For all x ∈ Rn
c1 kxkM ≤ kxkN ≤ c2 kxkM .
In addition, let Mi , i = 1, ..., n, be Orlicz functions and let x ∈ Rn then the MusielakOrlicz norm of x, kxk(Mi )i , is defined by
)
n
X
|xi |
t > 0
Mi
≤1 .
t
(
kxk(Mi )i = inf
i=1
2
The upper estimate
In this section we are going to prove the upper estimate. We require the following
known lemma. In a more general form see e.g. ([13], Lemma 10).
l
Lemma 2.1. Let x
(l)
(l)
ε1 xπ(1) , ..., εn xπ(n)
(l)
=
1
√
n−l
z }| { z }| {
(1, ..., 1, 0, ..., 0), l = 1, ..., n, and let BT be the convex hull of
l
, where εi = ±1, i = 1, ..., n, and π denote permutations of {1, ..., n}.
Let k kT be the norm on Rn whose unit ball is BT . Then, for all x ∈ Rn
kxk2 ≤ kxkT ≤
p
ln(en)kxk2 .
∗
Proof. The left hand inequality is obvious. Let x ∈ Rn . Then x∗1√
, . . . , x√
n denotes the
decreasing rearrangement of the numbers |x1 |, . . . , |xn |. Let ak = k − k − 1 for k =
1, . . . , n. Then, for all x ∈ Rn
n
X
kxkT =
√
√
x∗k ( k − k − 1).
k=1
Since
√
k−
√
k−1≤
kxkT ≤
√1
k
n √
X
√
| k − k − 1|2
! 12
k=1
kxk2 ≤
n
X
1
k
! 21
kxk2 ≤
p
ln(en)kxk2 .
k=1
We denote
STn−1
=
x = (x1 , ..., xn ) ∈ S
n−1
∃i = 1, ..., n j = 1, ..., n|xj = ± √1 = i ,
i EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 3/13
Norm of random matrices
the set of extremal points of BT . Then by our previous lemma we have
kAk2→2 = sup kAxk2 ≤
p
x∈S n−1
ln(en) sup kAxk2 .
(2.1)
n−1
x∈ST
We use now the concentration of sums of independent gaussian random variables
Pn
i=1 gi zi in a Banach space ([6], formula (2.35)): For all t > 0
X=
t2
P{kXk ≥ EkXk + t} ≤ exp −
2σ(X)2
,
(2.2)
where
σ(X) = sup
kξk∗ =1
n
X
! 21
|ξ(zi )|2
.
(2.3)
i=1
Applying (2.2) to

EkGxk2 ≤ 
n
X
 21
a2ij x2j 
and

 21
n
X
σ(Gx) = max 
a2ij x2j  ,
i=1...,n
i,j=1
j=1
we immediately get the following lemma.
Lemma 2.2. For all i, j = 1, ..., n let ai,j ∈ R, let gi,j be independent standard gaussians,
!
G = (ai,j gi,j )ni,j=1 and let x ∈ B2n . For all β ≥ 1 and all x with max
i=1,...,n
n
P
j=1
a2i,j x2j
> 0 we
have
P kGxk2 > β E max (ai,j gi,j )nj=1 2 + E max k(aij gij )ni=1 k2
j=1,...,n
i=1,...,n



! 12 2 



n
P 2 2



β E max (ai,j gi,j )nj=1 + E max k(ai,j gi,j )ni=1 k
 


−
a
x


i,j
j
2
2


i=1,...,n
j=1,...,n

 1
i,j=1
!
.
≤ exp −


2
n
P






max
a2i,j x2j




i=1,...,n j=1




Proposition 2.1. For all i, j = 1, ..., n let ai,j ∈ R, let gi,j be independent standard
pπ
gaussian random variables and let G = (ai,j gi,j )n
i,j=1 . For all β with β ≥
2
p
P kGk2→2 > β ln(en) E max (ai,j gi,j )nj=1 2 + E max k(ai,j gi,j )ni=1 k2
i=1,...,n
≤
n
X
2
exp l ln(2n) − l
l=1
β
π
.
Furthermore, we get for β =
P kGk2→2 >
≤
√
j=1,...,n
p
3π ln(2n)
n n
3π ln(en) E max (ai,j gi,j )j=1 2 + E max k(ai,j gi,j )i=1 k2
i=1,...,n
j=1,...,n
1
.
n2
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 4/13
Norm of random matrices
n
P
Proof. We shall apply Lemma 2.2. We may assume that max
i=1,...,n
j=1
!
a2i,j x2j
> 0 for all
x ∈ B2n \ {0}. By (2.1)
kGk2→2 ≤
p
ln(en) sup kGxk2 .
n−1
x∈ST
Therefore, for β ∈ R>0 , we have
n
n max k(ai,j gi,j )i=1 k2
P kGk2→2 > β ln(en) E max (ai,j gi,j )j=1 2 + E
j=1,...,n
i=1,...,n
!
n
n
max k(ai,j gi,j )i=1 k
≤P
sup kGxk > β E max (aij gij )j=1 + E
.
p
2
n−1
x∈ST
2
i=1,...,n
2
j=1,...,n
(l)
For all l = 1, ..., n let Ml be the set of x(l) ∈ STn−1 , such that xj
j = 1, ..., n. Now we apply Lemma 2.2 and get
∈ {0, ± √1l } for all
p
n n
P kGk2→2 > β ln(en) E max (ai,j gi,j )j=1 2 + E max k(ai,j gi,j )i=1 k2
j=1,...,n
i=1,...,n
(
)
n
n
≤P
sup kGxk > β E max (aij gij )j=1 + E max k(ai,j gi,j )i=1 k
2
n−1
x∈ST
≤
n
X
X
2
i=1,...,n
2
j=1,...,n
(l) n n
P Gx > β E max (aij gij )j=1 2 + E max k(ai,j gi,j )i=1 k2
j=1,...,n
i=1,...,n
2
l=1 x(l) ∈Ml






 


n n

E
max
(a
g
)
+
E
max
k(a
g
)
k
n
ij ij j=1 2
i,j i,j i=1 2
 1 
X
X
i=1,...,n
j=1,...,n

≤
exp − · l β
1

2

2 

l=1 x(l) ∈Ml


n
P




max 
a2ij 


i=1,...,n
(l)
j∈{k|xk 6=0}
 12 2 



P



1
2


√

a


i,j
l

 
i=1 j∈{k|x(l) 6=0}

k
.
−
 12 

 

 
n
P


max 
a2i,j   


i=1,...,n

(l)
j∈{k|x 6=0}

n
P
k
We have

n
1 X
√ 
l i=1
 12
X

a2ij  ≤
(l)
j∈{k|xk 6=0}
max
(l)
j∈{k|xk 6=0}
k(aij )ni=1 k2 ≤ max k(aij )ni=1 k2
j=1,...,n
and by triangle inequality
E max k(ai,j gi,j )ni=1 k2
j=1,...,n
≥
≥
max E
j=1,...,n
max
j=1,...,n
n
X
! 12
|ai,j gi,j |2
(2.4)
i=1
n
X
! 21
2
2
|ai,j | (E|gi,j |)
i=1
EJP 18 (2013), paper 29.
r
=
2
max k(ai,j )ni=1 k2 .
π j=1,...,n
ejp.ejpecp.org
Page 5/13
Norm of random matrices
Therefore, we have for all β with β ≥
pπ
2

n
1 X
βE max k(ai,j gi,j )ni=1 k2 − √ 
j=1,...,n
l i=1
 12
X

a2i,j  ≥ 0.
(l)
j∈{k|xk 6=0}
Thus
p
P kGk2→2 > β ln(en) E max (ai,j gi,j )nj=1 2 + E max k(ai,j gi,j )ni=1 k2
j=1,...,n
i=1,...,n

2 







 








n


(a
g
)
E
max
n
i,j
i,j




j=1 2
X X
i=1,...,n
1 

≤
exp − · l β

 12   .


2 

l=1 x(l) ∈Ml
 



n
P



 max 
2   


a


i,j


i=1,...,n
(l)
j∈{k|xk 6=0}
Again, by (2.4) we have for all β with β ≥
P kGk2→2
≤
i=1,...,n
exp −l
P kGk2→2
≤
p
n
X
j=1,...,n
2
2l nl exp −l
l=1
2
exp l ln(2n) − l
We choose β =
≤
β
π
l=1
n
X
2
l=1 x(l) ∈Ml
=
2
p
> β ln(en) E max (ai,j gi,j )nj=1 2 + E max k(ai,j gi,j )ni=1 k2
n
X
X
n
X
pπ
β
π
β
π
.
β2
π .
3π ln(2n), thus such that 3 ln(2n) =
Then
p
n n
> 3π ln(2n) ln(en) E max (aij gij )j=1 2 + E max k(aij gij )i=1 k2
i=1,...,n
exp (−2l ln(2n)) =
n X
l=1
l=1
1
4n2
l
≤
1
n2
j=1,...,n
∞ X
l=1
1
4
l
≤
1
.
n2
Therefore
P kGk2→2 >
≤
√
n n
3π ln(en) E max (aij gij )j=1 2 + E max k(aij gij )i=1 k2
i=1,...,n
j=1,...,n
1
.
n2
Proposition 2.2. Let ai,j ∈ R, i, j = 1, ..., n and gi,j , i, j = 1, ..., n, be independent
standard Gaussian random variables, then
E (ai,j gi,j )ni,j=1 2→2
r
π √
≤
+ 3π ln(en)
E max (ai,j gi,j )nj=1 2 + E max k(ai,j gi,j )ni=1 k2 .
i=1,...,n
j=1,...,n
2
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 6/13
Norm of random matrices
Proof. We divide the estimate of E (ai,j gi,j )n
i,j=1 2→2
into two parts. Let M be set of
all points with
(ai,j gi,j )ni,j=1 2→2
√
≤ 3π ln(en) E max (ai,j gi,j )nj=1 2 + E max k(ai,j gi,j )ni=1 k2 .
j=1,...,n
i=1,...,n
Clearly,
E (ai,j gi,j )ni,j=1 2→2 χM
√
n n
≤ 3π ln(en) E max (ai,j gi,j )j=1 2 + E max k(ai,j gi,j )i=1 k2 .
j=1,...,n
i=1,...,n
Furthermore, by Cauchy-Schwarz inequality and Proposition 2.1 we get
p
2 21
E (ai,j gi,j )ni,j=1 2→2 χM c ≤ P (M c ) E (ai,j gi,j )ni,j=1 2→2


 12
 12
n
n
1
X
X
1
1
2
2
E (ai,j gi,j )ni,j=1 HS
≤
= 
|ai,j |2  ≤ max 
|ai,j |2  .
1≤i≤n
n
n i,j=1
j=1
As in (2.4)
max (aij )nj=1 2 + max k(aij )ni=1 k2
i=1,...,n
j=1,...,n
r
π
E max (aij gij )nj=1 2 + max k(aij gij )ni=1 k2 .
≤
j=1,...,n
i=1,...,n
2
Altogether, this yields
rπ n
n n
E (ai,j gi,j )i,j=1 2→2 χM c ≤
E max (ai,j gi,j )j=1 2 + max k(ai,j gi,j )i=1 k2 .
j=1,...,n
i=1,...,n
2
Summing up, we get
E (ai,j gi,j )ni,j=1 2→2
r
π √
n n
≤
+ 3π ln(en)
E max (ai,j gi,j )j=1 2 + E max k(ai,j gi,j )i=1 k2 .
i=1,...,n
j=1,...,n
2
Proof. (Theorem 1.1) W.l.o.g. we assume 0 ≤ ai,j ≤ 1, i, j = 1, ..., n and that there is a
coordinate that equals 1. For all i, j = 1, ..., n and k ∈ N we define
aki,j
=
1
2k
0
, if 21k < ai,j ≤
, else.
1
2k−1
k
k
n
Let G = (ai,j gi,j )n
of
i,j=1 and G = (ai,j gi,j )i,j=1 . Wedenote byφ(k) the
number
nonzero
k n
n
γ
n
entries in (ai,j )i,j=1 and we choose γ such that (ai,j )i,j=1 1 = 2 (ai,j )i,j=1 ∞ . Thus,
we get φ(k) 21k =
n
P
i,j=1
aki,j ≤ 2γ and therefore φ(k) ≤ 2k+γ . Therefore, the non-zero
entries of Gk are contained in a submatrix of size 2k+γ × 2k+γ . Taking this into account
and applying Proposition 2.2 to Gk
E Gk 2→2
r
π √
≤
+ 3π ln(e2k+γ ) E max (aki,j gi,j )nj=1 2 + max (aki,j gi,j )ni=1 2 .
i=1,...,n
j=1,...,n
2
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 7/13
Norm of random matrices
Since
r
π √
13 31
+ 3π ln(e2k+γ ) ≤
+ (1 + ln(2k+γ )) ≤ 8(k + γ)
2
10 10
we get

 12
n
X
k
γ
k
|aki,j gi,j |2  ≤ 8(k + γ)2 2 − 2 .
E G 2→2 ≤ 8(k + γ) E
i,j=1
Therefore,
∞
X k+γ
X
k
E Gk 2→2 ≤ 8
≤ 16
k
k .
4
4
k≥2γ 2
k≥2γ
k=1 2
X
Since one of the coordinates of the matrix is 1
E G1 2→2 ≥
∞
Z
r
|g|dt =
−∞
2
.
π
Therefore, there is a constant c such that
E kGk2→2
X
k
G ≤ 2E k≤2γ +2
2→2
The matrix
X
EkGk k2→2
k>2γ
X
k
G ≤ cE k≤2γ .
2→2
k
P
k≤2γ
G has at most
X
φ(k) ≤
k≤2γ
X
!3
(ai,j )n i,j=1 1
(ai,j )n 2γ+k ≤ 23γ+1 ≤ 2
(2.5)
i,j=1 ∞
k≤2γ
k
entries that are different from 0. Therefore, all nonzero entries of
k≤2γ G are contained in a square submatrix having less than (2.5) rows and columns. We may apply
Proposition 2.2 and get with a proper constant c
P


!3 
r
(ai,j )n √
π
i,j=1 1
 ×
E (kGk2→2 ) ≤c 
+ 3π ln e · 2 (ai,j )n 2
i,j=1 ∞



n n
X
X



aki,j gi,j  aki,j gi,j 
E  max max + j=1,...,n
i=1,...,n k≤2γ
k≤2γ
j=1 2
3


.
i=1 2
The lower estimate
Theorem 3.1. For all i, j = 1, ..., n let ai,j ∈ R≥0 and gi,j be independent standard
−1
gaussians. For all i = 1, ..., n and all s with 0 ≤ s < (ai,j )n
j=1 2

Ni (s) = s max ai,j exp −
j=1,...,n
s2

1

max a2i,j
j=1,...,n
−1
and for all s ≥ (ai,j )n
j=1 2
Ni (s) =
max ai,j
j=1,...,n
(ai,j )n j=1 2
 
(ai,j )n 2
j=1 2
 + 3 (ai,j )nj=1 exp −
2
2
max ai,j
e
j=1,...,n
EJP 18 (2013), paper 29.
1
s− (ai,j )n j=1 2
!
.
ejp.ejpecp.org
Page 8/13
Norm of random matrices
ej , j = 1, ..., n, are defined in the same way for the transposed matrix
Moreover, N
ej , j = 1, . . . , n, are Orlicz functions and
(ai,j )nj,i=1 . Then Ni , i = 1, . . . , n and N
c1 k(1)ni=1 k(Ni )i + (1)nj=1 (N
fj )j
max k(ai,j gi,j )ni=1 k2
≤ E max (ai,j gi,j )nj=1 2 + E
j=1,...,n
i=1,...,n
n n
≤ c2 k(1)i=1 k(Ni )i + (1)j=1 (N
,
f)
j j
where c1 and c2 are absolute constants.
The following example is an immediate consequence of Theorem 3.1.
Toeplitz matrices.
It covers
Example 3.2. Let A be a n × n-matrix such that for all i, = 1 . . . , n and k = 1, . . . , n

n
X

1

2
|ai,j |
2
=
j=1
1
2
n
X
|aj,k |
2
j=1
and
max |ai,j | = max |aj,k |.
1≤j≤n
1≤j≤n
Then
E
max
i=1,.....,n
(ai,j gi,j )n j=1 2
+E
max
j=1,.....,n
k(ai,j gi,j )n
i=1 k2


1


2
n
 X

√
2
|a1,j |  , ln n max |a1,j | .
∼ max 


1≤j≤n
 j=1

We associate to a random variable X a Orlicz function M by
Z
s
Z
|X|dPdt.
M (s) =
0
(3.1)
1
t ≤|X|
We have
Zs Z
|X|dPdt
M (s) =
0
Zs
=
1
t ≤|X|


Z∞
1
1

+ P (|X| ≥ u) du dt.
 P |X| ≥
t
t
0
(3.2)
1
t
Lemma 3.3. There are strictly positive constants c1 and c2 such that for all n ∈ N, all
independent random variables X1 , ..., Xn with finite first moments and for all x ∈ Rn
c1 kxk(Mi )i ≤ E max |xi Xi | ≤ c2 kxk(Mi )i ,
1≤i≤n
where M1 , ..., Mn are Orlicz functions that are associated to the random variables (3.1).
Lemma 3.3 is a generalization of the same result for identically distributed random
variables [4]. It can be generalized from the `∞ -norm to Orlicz norms.
We use the fact [11] that for all x > 0
√
1 2
2π
√
e− 2 x ≤
2
(π − 1)x + x + 2π
q Z
2
π
∞
1
2
e− 2 s ds ≤
x
EJP 18 (2013), paper 29.
q
2 1 − 21 x2
.
π xe
(3.3)
ejp.ejpecp.org
Page 9/13
Norm of random matrices
Proof. (Theorem 3.1) We apply Lemma 3.3 to the random variables

Xi = 
n
X
 12
|ai,j gi,j |2 
i = 1, . . . , n.
j=1
Now, it is enough to show that Mi ∼ Ni for all i = 1, . . . , n. We have two cases.
12 −1
Pn
2
2
E
a
g
. There are constants c1 , c2 > 0 such
j=1 i,j i,j
! 12
n
P
2
2
that for all u with u > 2E
ai,j gi,j
1
2
We consider first s <
j=1

u
max a2i,j
exp −c21



 12
n
X


2 
 ≤ P 
a2i,j gi,j
≥ u ≤ exp −c22

2
j=1
j=1,...,n

u2
.
max a2i,j
(3.4)
j=1,...,n
The right-hand side inequality follows from (2.2). The left-hand side inequality follows
from
n
X
2
2
a2i,j gi,j
≥ a2i,1 gi,1
.
j=1
Since
1
t
! 21
n
P
> 2E
j=1
Mi (s) =
for 0 < t < s, we can apply (3.4). Therefore,

 


 21
 12

Z∞
n
n

X
X
1 
1


2
2 
2
2 
ai,j gi,j
P
≥  + P
ai,j gi,j
≥ u du dt


t

t
j=1
j=1

Zs 

0
≤
2
a2i,j gi,j
1
t

Zs 
1

t
0


exp −
t2
c22
+
max a2i,j
j=1,...,n
Z∞
1
t



u2
 du dt.
exp −c22

max a2i,j

j=1,...,n


By (3.3)
Mi (s) ≤

Zs 





2
2
1
t
c
c
2
2
2
+
 dt.
−
exp − 2
max
a
exp
i,j
t
t max a2i,j
2c22 j=1,...,n
t2 max a2i,j 
j=1,...,n
0
Since
1
t
> 2E
! 21
n
P
j=1
2
a2i,j gi,j
≥
j=1,...,n
q 2 n π (ai,j )j=1 2 , we get
1
1
t
1
+ 2 max a2i,j ≤ + 2
t
2c2 j=1,...,n
t
2c2
r
π
1
c
(ai,j )nj=1 2 ≤ ,
n
2
2 (ai,j )j=1
t
2
where c is an absolute constant. Altogether, we get
Zs
Mi (s) ≤
0




Z∞
2
2 2
c
c
c
c
u
2
2
 dt =
 du.
exp − 2
exp −
t
t max a2i,j
u
max a2i,j
j=1,...,n
1
s
j=1,...,n
Passing to a new constant c2 and using (3.3) we get for all s with 0 ≤ s <
Z∞
Mi (s) ≤ c
1
s
1
2
12 −1
Pn
2
2
E
a
g
j=1 i,j i,j




2 2
2
1
c
u
s
c
2
2
 du ≤ ( max ai,j ) exp −
 . (3.5)
exp −
u
max a2i,j
c2 j=1,...,n
s2 max a2i,j
j=1,...,n
j=1,...,n
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 10/13
Norm of random matrices
From this and the definition of Ni we get that there is a constant c such that for all s
with 0 ≤ s <
1
2
12 −1
Pn
2
2
E
j=1 ai,j gi,j
Mi (s) ≤ Ni (cs).
Indeed, the inequality follows immediately from (3.5) provided that
c2
2
− 12
P
n
j=1
|ai,j |2
12
P
n
2
2
,
E
j=1 ai,j gi,j
If
Moreover,
1
2
P
n
j=1
|ai,j |2
− 12
P
n
j=1
12 
P
n
2
2
2
g
c
(E
a
)2  √2π
j=1 i,j i,j
 2
.
≤
12 exp −
4 max a2i,j
c2

j=1,...,n
c2 E
≤
2
a2i,j gi,j
√
j=1,...,n
√
!
!
2π
n
−1
+ 1 k(ai,j )j=1 k
.
c2
2π
≤ Ni
c2
Therefore, with a universal constant c the inequality Mi (s) ≤ Ni (cs) also holds for those
values of s. The inverse inequality is treated in the same way.

Now we consider s with s ≥
1
2
E
n
P
j=1
! 12 −1
2
a2i,j gij

and denote α = E
n
P
j=1
! 21
2
a2ij gij
.
The following holds
Mi (s) =
=
 



 12
 12

Z∞
n
n

X
X
1
1 


2
2 
2
2 
ai,j gi,j
≥  + P
≥ u du dt
P
ai,j gi,j


t

t
j=1
j=1

Zs 

0
1
t


 

 12
 12

Z∞
n
n

X
X
1



2
2 
2 2
ai,j gi,j
P
≥  + P
≥ u du dt
aij gij


t
t

j=1
j=1

1
Z2α 
1
0
1
t
+



 
 12
 12

Z∞
n
n

X
X
1



2 
2 
a2i,j gi,j
a2i,j gi,j
P 
≥  + P 
≥ u du dt


t
t

j=1
j=1

Zs 
1
1
2α
1
t
1
The first summand equals Mi ( 2α
). Therefore, by (3.5) the first summand is of the order
! 12 2 

2
E
 
a2i,j gi,j


max ai,j


j=1
j=1,...,n
 2

−c2
.
! 12 exp 
2


max
a
i,j
n
P


j=1,...,n
2
2


E
ai,j gi,j


n
P
j=1
We estimate the second summand. The second summand is less than or equal to

 12 

 21

 12

Zs
n
n
n

X
X
X
1
2
2 
2 
+ E
a2ij gij
dt ≤ 3E 
a2i,j gi,j
dt ≤ 3E 
a2i,j gi,j
s.


t

j=1
j=1
j=1
1

Zs 

1
2α
.
q
12 −1
Pn
2
2
2
≤ s ≤ 12 E
a
g
then,
by
(3.5)
and
j=1 i,j i,j
π max1≤j≤n ai,j ≤
2 max ai,j
Mi (s) ≤
s
c2
2α
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 11/13
Norm of random matrices

Therefore, with a universal constant c we have for all s with s ≥

Mi (s) ≤ (c − 1)s 
n
X
1
2
n
P
E
j=1

Mi (s) ≥
n
X

2
α
1
t≤
n
P
j=1
!1


 21
n
X
|ai,j |2  − 1 ≤ Ni (cs).
|ai,j |2  ≤ cs 
j=1
Now, we sketch a proof of a lower estimate. By (3.1), for all s with s ≥
Z
2
a2i,j gij
 12
j=1
Zs
! 12 −1
 21
2 
dPdt ≥
a2i,j gi,j
j=1
Zs
2
α
2
2
a2i,j gi,j

 12
n
X
2

dPdt.
a2i,j gij
Z
α
2≤
n
P
j=1
2
α
!1
j=1
2
2
a2i,j gi,j
By the definition of α
Mi (s) ≥
1
2
Zs

 12
n
X
2 
dt
E
a2i,j gi,j
j=1
2
α

 
 12 −1
n
n
X

1
 X 2 2   
2  
a2i,j gi,j
ai,j gi,j
E
s
−
2
= E

 


2
j=1
j=1

 21


 21
n
1 X 2 2 
ai,j gi,j
= E
s − 1.
2
j=1
The rest is done as in the case of the upper estimate.
References
[1] R. Adamczak, O. Guédon, A. Litvak, A. Pajor, and N. Tomczak-Jaegermann, Smallest singular value of random matrices with independent columns, Comptes Rendus Mathématique.
Académie des Sciences. Paris 346 (2008), 853–856. MR-2441920
ˆ F , Application
[2] S. Chevet, Séries des variables aléatoires gaussiennes à valeurs dans E ⊗
aux produits d’espaces de Wiener abstraits, In Séminaires sur la Géometrie des Espaces de
Banach (1977-1978), École Polytechnique, 1978. MR-0520217
[3] A. Edelman, Eigenvalues and condition numbers of random matrices, SIAM Journal on Matrix Analysis and Applications 9 (1988), 543-560. MR-0964668
[4] Y. Gordon, A. Litvak, C. Schütt and E. Werner, Orlicz Norms of Sequences of Random Variables, Annals of Probability, 2002, Vol. 30, No. 4, 1833 - 1853. MR-1944007
[5] R. Latała, Some estimates of norms of random matrices, Proceedings of the American Mathematical Society 133 (2005), 1273–1282. MR-2111932
[6] M. Ledoux, The Concentration of Measure Phenomenon, American Mathematical Society,
2001. MR-1849347
[7] A. Litvak, A. Pajor, M. Rudelson and N. Tomczak-Jaegermann, Smallest singular value of
random matrices and geometry of random polytopes, Advances in Mathematics 195 (2005),
491-523. MR-2146352
[8] G. Pisier, The Volume of Convex Bodies and Banach Space Geometry, Cambridge University
Press, 1989. MR-1036275
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 12/13
Norm of random matrices
[9] M. Rudelson and R. Vershynin, Smallest singular value of a random rectangular matrix,
Communications on Pure and Applied Mathematics 62 (2009), 1707–1739. MR-2569075
[10] M. Rudelson and R. Vershynin, The least singular value of a random square matrix is
O(n−1/2 ), Comptes Rendus Mathématique. Académie des Sciences. Paris 346 (2008), 893–
896. MR-2441928
[11] M.B. Ruskai and E. Werner, Study of a class of regularizations of 1/|x| using Gaussian integrals, SIAM J: Math. Anal. 32 (2000), 435–463. MR-1781466
[12] Y. Seginer, The expected norm of random matrices, Combinat. Probab. Comput. 9 (2000),
149–166. MR-1762786
[13] C. Schütt, On the Banach-Mazur distance of finite-dimensional symmetric Banach spaces
and the hypergeometric distribution, Studia Mathematica 72 (1982), 109–129. MR-0665413
[14] S. Szarek, Condition numbers of random matrices, Journal of Complexity 7 (1991), 131-149.
MR-1108773
EJP 18 (2013), paper 29.
ejp.ejpecp.org
Page 13/13
Download