ETNA

advertisement
ETNA
Electronic Transactions on Numerical Analysis.
Volume 33, pp. 207-220, 2009.
Copyright  2009, Kent State University.
ISSN 1068-9613.
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES
AND BLOCK ARNOLDI METHODS∗
L. ELBOUYAHYAOUI†, A. MESSAOUDI‡, AND H. SADOK§
Abstract. The solution of linear systems of equations with several right-hand sides is considered. Approximate
solutions are conveniently computed by block GMRES methods. We describe and study three variants of block
GMRES. These methods are based on three implementations of the block Arnoldi method, which differ in their
choice of inner product.
Key words. block method, GMRES method, Arnoldi method, matrix polynomial, multiple right-hand sides,
block Krylov subspace, Schur complement, characteristic polynomial.
AMS subject classifications. 65F10.
1. Introduction. Many problems in science and engineering require the solution of
large linear systems of equations with multiple right-hand sides
AX = B,
A ∈ Cn×n ,
B ∈ Cn×s ,
X ∈ Cn×s ,
1 ≤ s ≪ n.
(1.1)
Instead of applying a standard iterative method to the solution of each one of the linear systems of equations
Ax(i) = b(i)
for
i = 1, . . . , s
(1.2)
independently, it is often more efficient to apply a block method to (1.1). The first block
method, a block conjugate gradient method, was introduced by O’Leary [14] for the solution
of a linear system of equations with multiple right-hand sides (1.1) and a symmetric positive
definite matrix. For systems (1.1) with a nonsymmetric matrix, a block version of GMRES
was introduced in [29] and studied in [26, 27]. This method is based on a block version of
the standard Arnoldi process [1]; see, for example, [17, 28].
The purpose of the present paper is to compare three variants of GMRES for multiple
right-hand sides, including the block GMRES method considered in [26, 27, 29]. These
schemes are based on block Arnoldi-type methods and differ in the choice of inner product.
We provide a unified description of the methods discussed, and derive new expressions and
bounds for the residual errors.
The paper is organized as follows. Section 2 defines block GMRES iterates with the aid
of Schur complements, and presents a connection with matrix-valued polynomials. We use
these polynomials to derive some new relations in Section 3. A few examples that illustrate
the theory are provided in Section 4.
We conclude this section by introducing notation used in the remainder of this paper. We
first recall the definition of the Schur complement. [24].
D EFINITION 1.1. Let M be a matrix partitioned into four blocks
C D
M=
,
E F
∗ Received April 27, 2009. Accepted August 22, 2009. Published online on December 16, 2009. Recommended
by Lothar Reichel.
† Département de Mathématiques, Faculté des Sciences et Techniques, Université de Mohammadia, Mohammedia, Morocco. lakhdar.elbouyahyaoui@gmail.com.
‡ Département
de
Mathématiques,
Ecole
Normale
Supérieure,
Rabat,
Morocco.
abderrahim.messaoudi@gmail.com.
§ L.M.P.A, Université du Littoral, 50 rue F. Buisson BP699, F-62228 Calais Cedex, France.
sadok@lmpa.univ-littoral.fr.
207
ETNA
Kent State University
http://etna.math.kent.edu
208
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
where the submatrix F is assumed to be square and nonsingular. The Schur complement of
F in M , denoted by (M/F ), is defined by
(M/F ) = C − DF −1 E.
Throughout this paper, I and Is denote identity matrices and ek their kth column. For two
matrices Y and Z in Cn×s , we define the inner product hY, ZiF = trace(Y H Z), (where Y H
denotes the conjugate transpose of Y ). The associated norm is the Frobenius norm k · kF .
The 2-norm of a matrix X ∈ Cn×s is denoted by kXk2. The Kronecker product of the
matrices C = [ci,j ] and D is given by C ⊗ D = [ci,j D]. If X is an n × s matrix, x = vec(X)
is the ns vector obtained by stacking the s columns of the matrix X.
Finally, the roots of a matrix-valued polynomial P, which is a square matrix whose entries
are ordinary polynomials, are defined to be the roots of the ordinary polynomial det(P(t)).
2. Block minimal residual-type methods. The nonsingular linear system with multiple
right-hand sides (1.1) can be solved by Krylov subspace methods in two distinct ways. The
first approach is to apply classical GMRES [16] for linear systems of equations with singlevector right-hand sides to the s linear systems separately. The second approach is to treat all
the right-hand sides simultaneously.
Before investigating the two approaches, we need some notation. Let the initial approxi(1)
(2)
(s)
(1)
(s)
mation of the solution of (1.1) be X0 = [x0 , x0 , . . . , x0 ], and let R0 = [r0 , . . . , r0 ] =
(i)
(i)
B − AX0 be the corresponding residual, with r0 = b(i) − A x0 and B = [b(1) , . . . , b(s) ].
In what follows, we let Kk denote the block Krylov matrix
Kk = [R0 , AR0 , . . . , Ak−1 R0 ],
and Ki,k the Krylov matrix
(i)
(i)
(i)
Ki,k = [r0 , A r0 . . . , Ak−1 r0 ]
for i = 1, . . . , s. We also introduce the matrix
Wk = AKk .
2.1. Standard GMRES applied to systems with multiple right-hand sides. In this
section we apply the standard GMRES method to each one of the s linear systems of equations
(i)
(1.2). Define the ith classical Krylov subspace Kk (A, r0 ) by
(i)
(i)
(i)
(i)
Kk (A, r0 ) = span{r0 , A r0 , . . . Ak−1 r0 } ⊂ Cn .
(2.1)
(i)
It is well-known that the kth approximation xk,S of GMRES applied to the ith linear system
(1.2) satisfies
(i)
(i)
(i)
(i)
(i)
xk,S − x0 ∈ Kk (A, r0 ) and (Aj r0 )H rk,S = 0 for j = 1, . . . k,
(i)
(i)
(i)
(2.2)
where rk,S = b(i) − A xk,S . It follows that the residual vector rk,S can be written as a linear
(i)
combination of the vectors Aj r0 , j = 0, 1, . . . , k, i.e.,
(i)
(i)
(i)
rk,S = pk,S (A)r0 ,
ETNA
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
209
where
det
(i)
pk,S (t)
=
"
1
H H (i)
Ki,k
A r0
t
...
H H
Ki,k
A AKi,k
tk
#!
H AH AK )
det(Ki,k
i,k
.
The residual error for each one of the s linear systems satisfies
(i)
krk,S k22 =
H
eT1 (Ki,k+1
1
,
Ki,k+1 )−1 e1
i = 1, . . . , s;
(2.3)
see [21, 22]. Therefore, the Frobenius norm of the residual
(1)
(2)
(1)
(1)
(s)
(s)
Rk,S = [rk,S , . . . , rk,S ] = [pk,S (A)r0 , . . . , pk,S (A)r0 ]
(2.4)
can be written as
kRk,S k2F =
s
X
i=1
1
.
H
eT1 (Ki,k+1
Ki,k+1 )−1 e1
(2.5)
Similarly to the situation for standard GMRES, the residual Rk,S can be expressed in terms
of a polynomial in A. We deduce from (2.4) that
vec(Rk,S ) = PG
k,S (A) vec(R0 ),
(1)
(s)
where PG
k,S (t) = diag(pk,S (t), . . . , pk,S (t)).
(2.6)
2.2. The global GMRES method. Instead of using standard GMRES to solve each
linear system (1.2) separately, we may apply GMRES to a block diagonal matrix. The s linear
systems (1.2) can be rewritten in a compact form as (A ⊗ Is )x = vec(B), with x = vec(X).
This gives the following linear system with a single right-hand side

 (1) 

b
A



. 
.
..
(2.7)
 x =  ..  .

A
b(s)
Application of standard GMRES to (2.7) yields the global GMRES method, which also can
be defined as follows. Let
k−1
KG
U } ⊂ Cn×s
k (A, U ) = span{U, A U, . . . , A
denote the matrix Krylov subspace spanned by the matrices U, AU, . . . , Ak−1 U , where U is
an n × s matrix. Note that Z ∈ KG
k (A, U ) implies that
Z=
k
X
αj Aj−1 U ,
j=1
αj ∈ C,
j = 1, . . . , k.
At step k, the global GMRES method constructs the approximation Xk,G , which satisfies
the relations
j
Xk,G − X0 ∈ KG
k (A, R0 ) and hA R0 , Rk,G iF = 0,
j = 1, . . . , k.
The residual Rk,G = B − AXk,G satisfies the minimization property
kRk,G kF =
min
kR0 − AZkF .
Z ∈ KG
k (A, R0 )
(2.8)
ETNA
Kent State University
http://etna.math.kent.edu
210
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
The problem (2.8) is solved by applying the global Arnoldi process [8].
Global GMRES is a generalization of the global MR method proposed by Saad for approximating the inverse of a matrix [18, p. 300]. The global method also is effective, compared to block Krylov subspace methods, when applied to the solution of large and sparse
Lyapunov and Sylvester matrix equations with right-hand sides of low rank; see [9, 10, 23].
Applications of the global Arnoldi method in control theory, model reduction, and quadratic
matrix equations are given in [4–6, 30].
It is convenient to introduce the matrix product ⋄. Let Y = [Y1 , Y2 , . . . , Yp ] and
Z = [Z1 , Z2 , . . . , Zl ] be matrices of dimension n × ps and n × ls, respectively, where Yi and
Zj (i = 1, . . . , p; j = 1, . . . , l) are n × s matrices. Then ⋄ is defined by


hY1 , Z1 iF hY1 , Z2 iF . . . hY1 , Zl iF
hY2 , Z1 iF hY2 , Z2 iF . . . hY2 , Zl iF 


YH ⋄Z =
 ⊂ Cp×l .
..
..
..
..


.
.
.
.
hYp , Z1 iF
hYp , Z2 iF
. . . hYp , Zl iF
T HEOREM 2.1 ([3]). Let the matrix (AKk )H ⋄ (AKk ) be nonsingular. Then
−1
Rk,G = R0 − AKk (AKk )H ⋄ (AKk )
((AKk )H ⋄ R0 )
= PG
k,G (A)R0 ,
where
det
PG
k,G (t) =
1
t
...
(AKk )H ⋄ R0
(AKk )H ⋄ (AKk )
det((AKk )H ⋄ (AKk ))
tk
.
(2.9)
Moreover,
kRk,G k2F =
eT1
1
1
.
= T Ps
H
−1
⋄ Kk+1 ) e1
e1 ( i=1 Ki,k+1 Ki,k+1 )−1 e1
i
h
(1)
(s)
G
or equivalently
= PG
k,G (A)r0 , . . . , Pk,G (A)r0
H
(Kk+1
We also can write Rk,G
G
vec(Rk,G ) = diag PG
k,G (A), . . . , Pk,G (A) vec(R0 ).
The matrix-valued polynomial involved in the preceding two studied methods are both diagonal. In the next section, we consider a general matrix-valued polynomial.
2.3. The block GMRES method. Another approach to solving (1.1) is to consider all
the s right-hand side vectors b(i) , i = 1, . . . , s, as an n × s matrix. This leads to the block
GMRES method (BGMRES). This method determines at step k an approximate solution
Xk,B of (1.1) from the requirements
Xk,B − X0 ∈ KB
k (A, R0 ),
Rk,B = B − AXk,B ⊥ KB
k (A, AR0 ),
and
(2.10)
where
k−1
KB
U },
k (A, U ) = block span{U, AU, . . . , A
and “block span” is defined by
(
KB
k (A, U )
=
X∈C
n×s
|X =
k−1
X
i=0
i
A U Ω i ; Ωi ∈ C
s×s
for i = 0, . . . , k − 1
)
⊂ Cn×s .
ETNA
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
211
Alternatively, BGMRES can be defined by considering the approximate solution of the
ith system (1.2), which is determined by
(i)
(i)
(i)
xk,B − x0 ∈ Kk (A, R0 ) and (Aj R0 )H rk,B = 0,
j = 1, . . . , k; i = 1, . . . , s,
(2.11)
where
Kk (A, R0 ) = Range([R0 , AR0 , . . . , Ak−1 R0 ]) ⊂ Cn .
(2.12)
Note that the Krylov subspace Kk (A, R0 ) is a sum of s classical Krylov subspaces
Kk (A, R0 ) =
s
X
(i)
Kk (A, r0 ).
i=1
Each column of the residual matrix Rk,B is obtained by projecting orthogonally the corresponding column of R0 onto the block Krylov subspace A Kk (A, R0 ). Therefore, BGMRES
is a minimization method
kRk,B kF =
min
Z∈KB
k (A,R0 )
kR0 − AZkF .
The following result will be used in the sequel.
T HEOREM 2.2. ([2]) Let the matrix Wk = AKk be of full rank. Then
−1 H
Wk R0
Rk,B = R0 − AKk WkH Wk
R0
Wk
H
/
W
W
=
k .
k
WkH R0 WkH Wk
Following Vital [29], we introduce the operator
PG
k,B (A) ◦ R0 =
k
X
Ai R0 Ωi ,
i=0
−1
where Ω0 = Is , [Ω1 , . . . , Ωk ] = − WkH Wk
WkH R0 , and PG
k,B is the matrix-valued
polynomial defined by



k
Is
tIs
...
tk Is
X
 /WkH Wk  .
ti Ω i =  
(2.13)
PG
k,B (t) =
H
H
i=1
Wk R0
Wk Wk
Then the residual Rk,B can be expressed as
Rk,B = PG
k,B (A) ◦ R0 .
Theorem 2.2 helps us compare the residuals of standard GMRES applied to (1.2) and of
BGMRES applied to (1.1). The relation (2.3) is the key to developing convergence results for
GMRES [21, 22]. We have the following expression for the norm of the residuals determined
by BGMRES.
T HEOREM 2.3. Assume that the matrix Wk is of full rank. Then
(i)
krk,B k2 =
" H
(i)
(i)
T
0
e1 r0 r(i)
WkH r0
1
(i) H
Wk
r0
WkH Wk
#−1
for i = 1, . . . , s.
e1
(2.14)
ETNA
Kent State University
http://etna.math.kent.edu
212
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
Proof. From the first expression in Theorem 2.2, we deduce that
(i)
(i)
rk,B = r0 − Wk WkH Wk
Consequently,
(i) H (i)
rk,B
(i)
−1
(i)
krk,B k2 = r0
−1
(i)
WkH r0 = (I − Wk WkH Wk
(i)
(i)
= (r0 )H r0 − (r0 )H Wk WkH Wk
and so we obtain
(i)
krk,B k2
=
"
(i)
WkH )r0 .
−1
(2.15)
(i)
WkH r0 ,
!
#
(i)
(i)
(i)
(r0 )H (r0 ) (r0 )H Wk
H
/ Wk Wk .
(i)
WkH r0
WkH Wk
(2.16)
(i)
Hence, krk,B k2 is the Schur complement of WkH Wk in the matrix
"
(i)
(i)
(r0 )H r0
(i)
WkH r0
#
(i)
(r0 )H Wk
,
WkH Wk
which can be factored into a product of a block upper and a block lower triangular matrix
(UL factorization)
#
# "
−1 " (i) 2
(i)
(i)
(i)
(i)
krk,B k
0
(r0 )H r0
(r0 )H Wk
1 (r0 )H Wk WkH Wk
.
=
(i)
(i)
0
I
WkH r0
WkH Wk
WkH Wk
WkH r0
This factorization yields
eT1
" H
(i)
(i)
r0 r0
(i)
WkH r0
(i) H
Wk
r0
H
Wk Wk
#−1
e1 =
1
(i)
krk,B k2
,
(2.17)
which proves the theorem.
The above theorem allows us to improve the well-known result
min
Z∈KB
k (A,R0 )
kR0 − AZkψ ≤ max
min
i=1...s z ∈K (A,r (i) )
i
k
0
(i)
kr0 − zi k,
which was stated in [26, 27] with kZkψ = max (kzi k), and was shown by Vital in her
i=1,...,s
thesis [29]. It shows that the residual obtained by BGMRES is bounded by the maximum of
the norm of the residuals obtained by applying standard GMRES to each one of the s systems
(1.2).
T HEOREM 2.4. Let the matrix Wk be of full rank. Then
(i)
(i)
krk,B k ≤ krk,S k
for
i = 1, . . . , s
and
kRk,B kF ≤ kRk,S kF ≤ kRk,G kF .
ETNA
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
213
Proof. It suffices to show the first part of the theorem for i = 1. Let us first remark that
there exists a permutation matrix P , such that Wk = A[K1,k , . . . , Ks,k ] P . Therefore, we
(1)
can rewrite the expression of kRk,B k2 as
1
,
eT1 Fk−1 e1
(1)
krk,B k2 =
where

(1) H (1)
r0
r
 0
(AK1,k )H r0(1)
Fk = 
..


.
H (1)
(AKs,k ) r0

(1) H
AKs,k
...
r0

H
. . . (AK1,k ) AKs,k 
.
..


...
.
(1) H
AK1,k
r0
H
(AK1,k ) AK1,k
..
.
H
(AKs,k ) AK1,k
H
...
(AKs,k ) AKs,k
H
By noticing that the (k + 1) × (k + 1) principal submatrix of Fk is K1,k+1
K1,k+1 , using
Theorem 6.2 of [31, p. 177] and (2.3), we deduce that
1
(1)
krk,B k2
H
= eT1 Fk−1 e1 ≥ eT1 (K1,k+1
K1,k+1 )
−1
e1 =
1
(1)
krk,S k2
.
To prove the last inequality, we apply (2.5) and Theorem 7.2 of [15], and obtain
kRk,S kF =
s
X
eT
i=1 1
H
(Ki,k+1
1
1
= kRk,G k2F ,
≤ T Ps
H
−1
Ki,k+1 ) e1
e1 ( i=1 Ki,k+1 Ki,k+1 )−1 e1
which completes the proof.
We now examine the zeros of the matrix-valued polynomial PG
k,B .
T HEOREM 2.5. Let the matrix Wk be of full rank. Then
det(PG
k,B (t)) =
ks
(k)
Y
(α − t)
i
(k)
i=1
αi
,
(k)
where the αi , for i = 1, ..., ks, are the generalized eigenvalues of the matrix pair
{WkH Wk , WkH Kk }.
Proof. Let α be a root of det(PG
k,B ). Then from Theorem 2.2, we deduce that

det 
Is
WkH R0
αIs
...
WkH Wk
αk Is

 = 0.
(2.18)
Let us denote the ith block column of this determinant by Ci . Then by replacing the block
column Ci by Ci − αCi−1 for i = 2, . . . , k, we obtain
det(WkH Wk − αWkH Kk ) = 0,
(2.19)
which shows that α is a generalized eigenvalue of the matrix pair {WkH Wk , WkH Kk }. The
proof is completed by noticing that det(PG
k,B (0)) = 1.
ETNA
Kent State University
http://etna.math.kent.edu
214
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
3. Block Arnoldi-type algorithms for multiple starting vectors. This section provides the framework for block Arnoldi-type algorithms. These algorithms are used for determining multiple or clustered eigenvalues. They also are applied in implementations of block
and global GMRES methods.
We give a unified presentation of Arnoldi-type algorithms, which include the standard
Arnoldi algorithm applied to each column of the starting block vector, the global Arnoldi
method, and the block Arnoldi method. Let U be an n × s matrix. The Arnoldi-type algorithms construct a basis {V1• , . . . , Vk• } of a subspace of KB
k (A, U ). The basis satisfies an
orthogonality property and Hk• = (Vk• )H AVk• is upper block Hessenberg.
We examine three possibly choices of orthogonality. Let Φ• : Cn×s × Cn×s → Cs×s be
defined for • ∈ {B, S, G} by
 B
 Φ (X, Y ) = X H Y,
ΦS (X, Y ) = the diagonal of the matrix X H Y,
 G
Φ (X, Y ) = trace(X H Y )Is = hX, Y iF Is ,
for all X ∈ Cn×s and for all Y Cn×s .
If ΦB (X, Y ) = X H Y = 0, then the block-vectors X, Y are said to be block-orthogonal;
see Gutknecht [11]. Moreover, X is said to be block-normalized if X H X = Is . Of course,
the vector space of block vectors is a finite-dimensional inner product space with inner product hX, Y iF = trace(X H Y ). If hX, Y iF = 0, then X and Y are said to be F-orthogonal. If
ΦS (X, Y ) = diag(X H Y ) = 0, then we say that X and Y are diagonally orthogonal.
Using the map Φ• , we will show how the matrices Vk• and Hk• are computed.
Block Arnoldi-type algorithms
1. Let U be an n × s matrix.
•
2. Compute V1• ∈ Cn×s by determining the factorization of U : U = V1• H1,0
,
•
s×s
•
•
•
•
•
•
H1,0 ∈ C , such that H1,0 = Φ (V1 , U ) and Φ (V1 , V1 ) = Is .
3. for i = 1, . . . , k do
• Compute W = AVi• .
• for j = 1, . . . , i do
•
= Φ• (Vj• , W )
(a) Hj,i
•
(b) W = W − Vj• Hj,i
• End
•
•
•
• Compute Hi+1,i
by determining the decomposition of W : W = Vi+1
Hi+1,i
,
•
•
•
•
•
•
such that Hi+1,i = Φ (Vi+1 , W ) and Φ (Vi+1 , Vi+1 ) = Is .
4. End
We now consider two particular choices.
3.1. The block Arnoldi algorithm. For Φ• (X, Y ) = ΦB (X, Y ) = X H Y , the preceding algorithm reduces to block Arnoldi algorithm [11, 18–20, 25–27, 29], which builds an
orthonormal basis {V1B , . . . , VkB } such that the block matrix VkB = [V1B , . . . , VkB ] satisfies
H
(VkB ) VkB = Iks . It is well known that
B
B
AVkB = VkB HkB + Vk+1
Hk+1,k
EkT ,
(3.1)
where EkT = [0s , . . . , 0s , Is ] ∈ Rs×ms . Multiplying equation (3.1) by Ek , we deduce that
H
B
B
Vk+1
Hk+1,k
= AVkB − VkB VkB AVkB .
ETNA
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
215
H
B
We also have AV1B H1,0 = AU and V2B H2,1
= AV1B − V1B V1B AV1B , which imply that
H
B
V2B H2,1
H1,0 = AU − V1B V1B AU . Thus, by induction, we deduce that
H
B
B
B
B
Vk+1
Hk+1,k
Hk,k−1
· · · H1,0
= Ak U − VkB VkB Ak U.
Furthermore, if Kk = VkB RB
k is the QR decomposition of the full-rank matrix Kk , then
Vk Vk B = Kk (KkH Kk )−1 KkH . Hence, we have
B
B
B
B
Vk+1
Hk+1,k
Hk,k−1
· · · H1,0
= Ak U − Kk (KkH Kk )−1 KkH Ak U.
(3.2)
B
A
Consider the representation Vk+1
= PA
k,B (A) ◦ U . Since it is not easy to express Pk,B in
terms of Krylov matrices, we consider a monic matrix-valued polynomial, which, apart from
eA
a multiplicative matrix, is the polynomial PA
k,B . Thus, let Pk,B denote the matrix-valued
polynomial
B
e A (t) = PA (t)H B
P
k,B
k,B
k+1,k · · · H1,0 ,
and let {Zk } be the block vectors defined by Z1 = U and
B
B
B
Zk+1 = Vk+1
Hk+1,k
· · · H1,0
.
Then
A
(A) ◦ U
Zk+1 = I − Kk (KkH Kk )−1 KkH Ak U = Pek,B
e A can be expressed as
The matrix-valued polynomial P
k,B

tk Is
eA (t) = 
P
k,B
KkH Ak U
Is
...
tk−1 Is
KkH Kk
k ≥ 1.
for


 /KkH Kk  .
Applying the determinant function to this Schur complement, we obtain

 k
t Is
Is
...
tk−1 Is

det 
H
H k
K
A
U
K
K
k
k
k
e A (t)) =
.
det(P
k,B
det(KkH Kk )
eA .
The following result examines the zeros of P
k,B
T HEOREM 3.1. Let the matrix Kk be of full rank. Then
e A (t)) =
det(P
k,B
(k)
ks
Y
(k)
(t − θi ),
i=1
where θi , i = 1, ..., ks, are the eigenvalues of the matrix (KkH Kk )−1 (KkH AKk ).
e A (t)). It follows from (3.5) that
Proof. Let θ be a root of det(P
k,B

Is
det 
...
KkH Kk
θk−1 Is
θk Is
KkH Ak R0

 = 0.
(3.3)
(3.4)
(3.5)
ETNA
Kent State University
http://etna.math.kent.edu
216
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
Let Ci denote the ith block column of this determinant. Then replacing Ci by Ci − θCi−1
for i = 2, . . . , k, we obtain
det(KkH AKk − θKkH Kk ) = 0.
Since the matrix KkH Kk is nonsingular, θ is an eigenvalue of (KkH Kk )−1 KkH AKk . This result,
for the special case s = 1, is shown in [21].
Using the QR decomposition of the full-rank matrix Kk = VkB RB
k , we deduce that
−1 B B
Hk Rk .
(KkH Kk )−1 KkH AKk = RB
k
B
Consequently the roots of PA
k,B (t) are the eigenvalues of Hk .
3.2. The global Arnoldi algorithm. We have
Φ• (X, Y ) = ΦG (X, Y ) = hX, Y iF Is .
Hence, the global Arnoldi process builds an F-orthonormal basis {V1G , . . . , VkG } of KB
k (A, U ),
such that the matrix VkG = [V1G , . . . , VkG ] satisfies
G
G
AVkG = VkG HkG + Vk+1
Hk+1,k
EkT ,
where HkG = HkG ⊗ Is and the matrix HkG is a k × k Hessenberg matrix whose nonzero
entries (hG
i,j ) are defined by the following algorithm.
Global Arnoldi algorithm
1. Let U be an n × s matrix.
2. Compute V1G ∈ Cn×s by V1G = U/kU kF ,
3. for i = 1, . . . , k do
• Compute W = AViG .
• for j = 1, . . . , i do
G
(a) hG
j,i = hVj , W iF
G
(b) W = W − hG
j,i Vj
• End
G
• Compute hG
i+1,i = kW kF and set Vi+1 = W/hi+1,i .
4. End
G
G
A
It is easy to see that Hk+1,k
= hG
k+1,k Is and that Vk+1 = Pk,G (A)U . Moreover, if we
set
G
A
e A (t) = hG
P
k,G
k+1,k · · · h2,1 kU kF Pk,G (t)
eA ,
and use the explicit form of P
k,G

eA (t) =
P
k,G
det 
tk
KkH
1
k
...
tk−1
(KkH
⋄ (A U )
⋄ Kk )
det(KkH ⋄ Kk )
we can characterize the roots.
T HEOREM 3.2. Let the matrix (KkH ⋄ Kk ) be nonsingular. Then
e A (t)) =
P
k,G
s
Y
i=1
(k)
(t − θei ),


,
(3.6)
ETNA
Kent State University
http://etna.math.kent.edu
217
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
(k)
where θei , for i = 1, ..., s, are the eigenvalues of the matrix (KkH ⋄ Kk )−1 (KkH ⋄ (AKk )).
(k)
The eigenvalues θei can be called the F-Ritz values, since they also are the eigenvalues
H
G
of the Hessenberg matrix HkG = (VkG ⋄ (AV k )).
When we apply the global or the block Arnoldi processes with s = 1 and with the ith
columns of U , we obtain the standard Arnoldi process. Hence, the standard Arnoldi vectors
(i)
(i)
A
obtained with the ith columns of U can be written as vk = pk,S (A)U ei . Let Vk,S
be the
(1)
(s)
vector whose columns are vk , . . . , vk . We have
(1)
(s)
A
Vk,S
= [pk,S (A)U e1 , . . . , pk,S (A)U es ].
Consequently,
A
vec(Vk,S
) = PA
k,S (A) vec(U ),
(1)
(3.7)
(s)
where PA
k,S (t) = diag(pk,S (t), . . . , pk,S (t)).
4. Examples. We illustrate the theory developed in this paper with two examples. The
first one involves a diagonalizable matrix; the matrix of the second example is defective. In
these examples, we set X0 = 0 and U = B.
E XAMPLE 4.1. Consider the matrix and right-hand sides




1 1
−1 0 −1 1
 0 0
 0 2 0 −1



A=
 0 0 1 −1 and B =  1 1 .
−1 2
0 0 0 −2
Results obtained by the block Arnoldi and BGMRES√ methods are reported in Table 4.1.
2
A
Moreover, we have for block Arnoldi, det(PA
1,B (t)) = 18 (2t + 1)(t + 2) and det(P2,B (t)) =
√
2 2
9 (t
− 1)(t2 − 4). Hence, the eigenvalues of the matrix A are the roots of PA
2,B . We also
1
remark that the roots of PA
and
−2.
are
−
1,B
2
k
1)
2 2
+ 23 t −
3 (t
√
2 t
3 ( 3 + 1)
2)
3
3
1
2
3
3
PA
k,B (t)
2
3 (t√+
− 62
√
√
"√
"√
#
1
3 (t − 2)
t + 34
#
− 10
t
+
4)
3
t − 59 t − 38
1 2
3 (t
2
PG
k,B (t)
3
26 t
+1
1
− 13
t
(− 34 t2 − 16 t + 1
1
− 41 t2 − 12
t
5
26 t
7
13 t
+1
− 21 t2 + 13 t
− 21 t2 − 16 t + 1
TABLE 4.1
Polynomials obtained by the block Arnoldi and BGMRES methods.
On the other hand, the upper block Hessenberg matrix H2B determined by block Arnoldi
algorithm is
√
√


5 2
1
2
−1
2
18
√
√9
 2
2 
−3
5
 2
2
18
18√ 
H2B =  3√
,
11
 2 2 − 12
− 5182 
18
0
√
2
2
√
−5 182
17
9
ETNA
Kent State University
http://etna.math.kent.edu
218
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
A
Pk,G
(t)
1
√ (3t + 4)
23
k
1
√ 1
(69t2
15042
2
3
1
3
3037830 (981t
3
4
q
G
Pk,G
(t)
4
( 13 t + 1)
+ 1482t2 − 2066t − 2182)
327
2
9290 (t
88
1467 t
2
( −21
113 t +
+ 64t − 91)
3
( −2182
11809 t −
1 2
4 (t
− 1)(t2 − 4)
4589 2
11809 t
+
+ 1)
716
1687 t
+ 1)
− 1)(t2 − 4)
TABLE 4.2
Polynomials obtained by the global Arnoldi and global GMRES methods.
k
√
5
5
1
"√
6
2
6 (3t
2
3
"√
30
3
120 (15t
A
Pk,S
(t)
t+1
0
0
2t + 3
+ 2t − 5)
0
#
0
√
870
2
870 (15t + 22t − 7)
#
0
174
3
2
870 (87t + 108t − 323t − 352)
#
"√
30
0
√
(t2 − 1)(t2 − 4) 8
174
0
10
+ 15t2 − 28t − 8)
0
4
√
TABLE 4.3
Polynomials obtained by the standard Arnoldi method.
G
Pk,S
(t)
0
6t + 1
3
0
7t + 1
k
1
1
2
3
44
2
0
( −75
161 t − 161 t + 1)
178
2
0
( −3
71 t + 497 t + 1)
1 2
3
( −1
22 t − 2 t −
0
2
11 t
+ 1)
1 2
4 (t
4
0
443 2
3
( −352
t
−
1567
1567 t +
1358
1567 t
+ 1)
− 1)(t2 − 4)I2
TABLE 4.4
Polynomials obtained using the Standard GMRES method.
with the characteristic polynomial
PH2 (t) = det(tI4 − H2B ) = (t2 − 1)(t2 − 4).
The polynomials determined by global Arnoldi and global GMRES are displayed in Table 4.2. The standard Arnoldi methods yields the polynomials of Table 4.3, and the polynomial determined by standard GMRES are shown in Table 4.4.
ETNA
Kent State University
http://etna.math.kent.edu
ALGEBRAIC PROPERTIES OF THE BLOCK GMRES AND BLOCK ARNOLDI METHODS
k
A
Pk,B
(t)
(4t
−
9)
4
1
1
√
10
5
2
t−1
√
5 10
3
5
4t − 7
− 41
3
5t − 2
t
4 −1
3
4 (t − 3)
5t−7
4
219
G
Pk,B
(t)
7
t
− 18 t + 1
18
13
29
− 54
t
− 54
t+1
−3t
t−1 5t − 8
8
−t
7t − 8
TABLE 4.5
Polynomials obtained by the block Arnoldi and BGMRES methods.
E XAMPLE 4.2. Define the defective matrix

1 2
0 1
A=
0 0
0 0
and let

1
1
B=U =
0
0
1
0
1
0

0
1

0
2

2
0
.
1
1
Table 4.5 shows the results obtained by the block algorithms. The upper block Hessenberg
matrix determined by the block Arnoldi algorithm is given by
√ 
√

45
5
−3√ 10 4 10
1 
25
5 10
0 
.
 −15
√
√
H2B =

−2
−4 
20 5 10 −3√ 10
0
4 10
16
32
It has the characteristic polynomial
PH2 (t) = (t − 1)3 (t − 2).
We also have
√
2 19
A
P3,G
(t) = √ (t − 1)2 (t − 2),
67
1
G
P3,G (t) = − (t − 1)2 (t − 2),
2
"√
38
A
2
P3,S (t) = (t − 1) (t − 2)
0
0
√
66
9
#
,
1
G
P3,S
(t) = − (t − 1)2 (t − 2).
2
We remark that for all iterations except for the last one, the roots of the BGMRES polynomials are, in general, different from those of the corresponding Arnoldi polynomials. Moreover, apart from a multiplicative scalar, the determinant of the Arnoldi polynomial is the characteristic polynomial of the Hessenberg matrix obtained from the Arnoldi-type algorithms.
ETNA
Kent State University
http://etna.math.kent.edu
220
L. ELBOUYAHYAOUI, A. MESSAOUDI, AND H. SADOK
REFERENCES
[1] W. E. A RNOLDI , The principle of minimized iterations in the solution of the matrix eigenvalue problem,
Quart. Appl. Math., 9 (1951), pp. 17–29.
[2] R. B OUYOULI , K. J BILOU , A. M ESSAOUDI , AND H. S ADOK, On block minimal residual methods, Appl.
Math. Lett., 20 (2007), pp. 284–289.
[3] R. B OUYOULI , K. J BILOU , R. S ADAKA , AND H. S ADOK, Convergence properties of some block Krylov
subspace methods for multiple linear systems, J. Comput. Appl. Math., 196 (2006), pp. 498–511.
[4] C. C. C HU , M. H. L AI , AND W.S. F ENG, MIMO interconnects order reductions by using the multiple point
adaptive-order rational global Arnoldi algorithm, IECE Trans. Elect., E89-C (2006), pp 792–808.
[5] C.C. C HU , M.H. L AI , AND W.S. F ENG, The multiple point global Lanczos method for multiple-inputs
multiple-outputs interconnect order reduction, IECE Trans. Elect., E89-A (2006), pp. 2706–2716.
[6] M. H EYOUNI AND K. J BILOU, Matrix Krylov subspace methods for large scale model reduction problems,
Appl. Math. Comput., 181 (2006), pp. 1215–1228.
[7] K. J BILOU AND A. M ESSAOUDI , Matrix recursive interpolation algorithm for block linear systems. Direct
methods, Linear Algebra Appl. 294 (1999), pp. 137–154 .
[8] K. J BILOU , A. M ESSAOUDI , AND H.S ADOK, Global FOM and GMRES algorithms for matrix equations,
Appl. Numer. Math., 31 (1999), pp. 49–43.
[9] K. J BILOU AND A.J. R IQUET, Projection methods for large Lyapunov matrix equations, Linear Algebra
Appl., 31 (2006), pp. 49–43.
[10] Y. Q. L IN, Implicitly restarted global FOM and GMRES for nonsymmetric matrix equations and Sylvester
equations, Appl. Math. Comput., 167 (2005), pp. 1004–1025.
[11] M. H. G UTKNECHT, Block Krylov space methods for linear systems with multiple right-hand sides: an
introduction, in Modern Mathematical Models, Methods and Algorithms for Real World Systems, A. H.
Siddiqi, I. S. Duff, and O. Christensen, eds., Anamaya Publishers, New Delhi, 2005, pp. 420–447.
[12] A. M ESSAOUDI , Recursive interpolation algorithm: a formalism for linear equations I. Direct methods, J.
Comput. Appl. Math., 76 (1996), pp. 13–30
[13]
, Recursive interpolation algorithm: a formalism for linear equations II. Iterative methods, J. Comput.
Appl. Math., 76 (1996), pp. 31–53.
[14] D. O’L EARY, The block conjugate gradient algorithm and related methods, Linear Algebra Appl., 29 (1980),
pp. 293–322.
[15] C. R. R AO, Statistical proofs of some matrix inequalities, Linear Algebra Appl., 321 (2000), pp. 307–320.
[16] Y. S AAD AND M. S CHULTZ, GMRES: A Generalized minimal residual algorithm for solving nonsymmetric
linear systems, SIAM J. Sci. Statist. Comput., 7 (1986), pp. 856–869.
[17] Y. S AAD, Numerical Methods for Large Eigenvalue Problems, Halstead, New York, 1992.
, Iterative Methods for Sparse Linear Systems, PWS, Boston, 1996.
[18]
[19] M. S ADKANE, Block Arnoldi and Davidson methods for unsymmetric large eigenvalue problems, Numer.
Math., 64 (1993), pp. 687–706.
[20] M. S ADKANE AND M. ROBB É, Exact and inexact breakdowns in the block GMRES method, Linear Algebra
Appl., 419 (2006), pp. 265–285.
[21] H. S ADOK, Méthodes de projections pour les systèmes linéaires et non linéaires. Habilitation thesis, University of Lille 1, Lille, France, 1994.
[22]
, Analysis of the convergence of the minimal and the orthogonal residual methods, Numer. Algorithms,
40 (2005), pp. 101–115.
[23] D. K. S ALKUYEH AND F. T OUTOUNIAN, New approaches for solving large Sylvester equations, Appl. Math.
Comput., 173 (2006), pp. 9–18
[24] I. S CHUR, Potenzreihen im Innern des Einheitskreises, J. Reine Angew. Math., 147 (1917), pp. 205–232.
[25] V. S IMONCINI , Ritz and Pseudo-Ritz values using matrix polynomials, Linear Algebra Appl., 241-243 (1996),
pp. 787–801.
[26] V. S IMONCINI AND E. G ALLOPOULOS , Convergence properties of block GMRES and matrix polynomials,
Linear Algebra Appl., 247 (1996), pp. 97–119.
, An iterative method for nonsymmetric systems with multiple right-hand sides, SIAM J. Sci. Comput.,
[27]
16 (1995), pp. 917–933.
[28] G. W. S TEWART, Matrix Algorithms II: Eigensystems, SIAM, Philadelphia, 2001.
[29] B. V ITAL, Etude de quelques méthodes de résolution de problèmes linéaires de grande taille sur multiprocesseur, Ph.D. thesis, Université de Rennes, Rennes, France, 1990.
[30] N. WAGNER, A Krylov based method for quadratic matrix equations, in Angewandte und Experimentelle
Mechanik – Ein Querschnitt, K. Willner and M. Hanss, eds., Der Andere Verlag, Tönning, 2006, pp. 283–
303 .
[31] F. Z HANG, Matrix Theory, Springer, New York, 1999.
Download