UNIVERSITY OF CALIFORNIA, IRVINE On the Mathematics of Slow Light DISSERTATION

UNIVERSITY OF CALIFORNIA,
IRVINE
On the Mathematics of Slow Light
DISSERTATION
submitted in partial satisfaction of the requirements
for the degree of
DOCTOR OF PHILOSOPHY
in Mathematics
by
Aaron Thomas Welters
Dissertation Committee:
Professor Aleksandr Figotin, Chair
Professor Svetlana Jitomirskaya
Professor Abel Klein
2011
c 2011 Society for Industrial and Applied Mathematics. Reprinted with
Chapter 2 ⃝
permission. All rights reserved.
c 2011 Aaron Thomas Welters.
All other materials ⃝
DEDICATION
To Jayme
ii
TABLE OF CONTENTS
Page
ACKNOWLEDGMENTS
v
CURRICULUM VITAE
vi
ABSTRACT OF THE DISSERTATION
ix
1 Introduction
1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1.2 Electrodynamics of Lossless One-Dimensional Photonic Crystals
1.2.1 Time-Harmonic Maxwell’s Equations . . . . . . . . . . .
1.2.2 Lossless 1-D Photonic Crystals . . . . . . . . . . . . . .
1.2.3 Maxwell’s Equations as Canonical Equations . . . . . . .
1.2.4 Definition of Slow Light . . . . . . . . . . . . . . . . . .
1.3 Main Results . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1.4 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
1
1
3
3
4
4
6
7
10
2 Perturbation Analysis of Degenerate Eigenvalues from a Jordan block1
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2 The Generic Condition and its Consequences . . . . . . . . . . . . . . . . . .
2.3 Explicit Recursive Formulas for Calculating the Perturbed Spectrum . . . .
2.4 Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.5 Auxiliary Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
12
12
22
23
30
41
3 Spectral Perturbation Theory for Holomorphic Matrix Functions
3.1 On Holomorphic Matrix-Valued Functions of One Variable . . . . . .
3.1.1 Local Spectral Theory of Holomorphic Matrix Functions . . .
3.2 On Holomorphic Matrix-Valued Functions of Two Variables . . . . .
3.3 On the Perturbation Theory for Holomorphic Matrix Functions . . .
3.3.1 Eigenvalue Perturbations . . . . . . . . . . . . . . . . . . . . .
3.3.2 Eigenvector Perturbations . . . . . . . . . . . . . . . . . . . .
3.3.3 Analytic Eigenvalues and Eigenvectors . . . . . . . . . . . . .
46
46
50
63
68
68
73
79
1
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
c 2011 Society for Industrial and Applied
The contents of this chapter also appear in [61]. Copyright ⃝
Mathematics. Reprinted with permission. All rights reserved.
iii
4 Canonical Equations: A Model for Studying Slow Light
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.1.1 Model Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.2 Canonical ODEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.2.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.2.2 Energy Flux, Energy Density, and their Averages . . . . . . . . . .
4.2.3 On Points of Definite Type for Canonical ODEs . . . . . . . . . . .
4.2.4 Perturbation Theory for Canonical ODEs . . . . . . . . . . . . . . .
4.3 Canonical DAEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.3.1 The Correspondence between Canonical DAEs and Canonical ODEs
4.4 Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.5 Auxiliary Material . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Bibliography
.
.
.
.
.
.
.
.
.
.
.
93
93
94
98
99
102
103
105
107
108
109
169
196
iv
ACKNOWLEDGMENTS
I would like to begin by thanking my advisor, Professor Aleksandr Figotin. His insight,
guidance, and support have been invaluable. The efforts he has put in helping me develop as
a mathematician and a researcher went above and beyond the usual duties of an advisor. The
depth and breadth of his knowledge in mathematics and electromagnetism and his passion
for research is a continual source of inspiration for me. He has profoundly influenced my life
and my research and for all of this I am eternally grateful.
To the members of my thesis and advancement committees, Professor Svetlana Jitomirskaya,
Professor Abel Klein, Professor Ozdal Boyraz, and Dr. Ilya Vitebskiy, I would like to thank
them for their time and thoughtful input.
Thank you also to Adam Larios, Jeff Matayoshi, and Mike Rael, for many useful and enjoyable conversations.
A special thanks to Donna McConnell, Siran Kousherian, Tricia Le, Radmila Milosavljevic,
and Casey Sakasegawa, and the other staff at the UCI math department, for handling the
massive amount of paperwork and administrative details that allow me the time to concentrate on research. Their hard work behind the scenes helps make the UCI Mathematics
Department the excellent place that it is.
I would also like to thank my family. My parents, Tom and Kathy Welters, and my brothers,
Matt and Andy Welters, for being great role models on how to work hard and achieve
while still balancing ones life with respect to family, fun, and career and for instilled in
me the importance and joys of family. Thank you so much for all the love, support, and
encouragement over the years. I would like to thank my son, Ashton Welters, for making my
life happier than I thought possible. I would especially like to thank my wife and best friend,
Jayme Welters. I could not have made it without her encouragement, support, commitment,
friendship, love, and laughter. Thank you all that you have done for me over the years, it is
appreciated more than I can possibly say in words.
The work on the explicit recursive formulas in the spectral perturbation analysis of a Jordan
block was first published in SIAM Journal of Matrix Analysis and Applications, Volume 32,
no. 1, (2011) 1–22 and permission to use the copyrighted material in this thesis has been
granted.
I am thankful to the anonymous referees for helpful suggestions and insightful comments on
parts of this research. I am thankful for the University of California, Irvine, where this work
was completed.
This work was supported in part by the Air Force Office of Scientific Research (AFOSR)
under the grant FA9550-08-1-0103.
v
CURRICULUM VITAE
Aaron Thomas Welters
PARTICULARS
EDUCATION
University of California Irvine
Ph. D. in Mathematics
Advisor: Dr. Aleksandr Figotin
Irvine, CA
June 2011
St. Cloud State University
B. A. in Mathematics
Magna Cum Laude
Advisor: Dr. J. -P. Jeff Chen
St. Cloud, MN
May 2004
RESEARCH INTERESTS
My research interests are in mathematical physics, material science, electromagnetics, wave
propagation in periodic media (e.g., photonic crystals), spectral and scattering theory. I specialize in the spectral theory of periodic differential operators, nonlinear eigenvalue problems,
perturbation theory for non-self-adjoint matrices and operators, linear differential-algebraic
equations (DAEs), block operator matrices, boundary value problems, and meromorphic
Fredholm-valued operators. I apply the mathematical methods from these areas to study
spectral and scattering problems of electromagnetic waves in complex and periodic structures.
ACADEMIC HONORS
∙ Von Neumann Award for Outstanding Performance as a Graduate Student, UC Irvine, 2010
∙ Mathematics Department Scholarship, SCSU, 2003
PUBLICATIONS
2. A. Welters, Perturbation Analysis of Slow Waves for Periodic Differential-Algebraic of Definite Type. (in preparation)
1. A. Welters, On Explicit Recursive Formulas in the Spectral Perturbation Analysis of a Jordan
Block, SIAM J. Matrix Anal. Appl., 32.1 (2011), pp. 1–22.
vi
TALKS
9. UCLA PDE Seminar, University of California, UCLA, CA, February 2011.
8. 2011 AMS/MAA Joint Mathematics Meetings, New Orleans, LA, January 2011.
7. MGSC (Mathematics Graduate Student Colloquium), University of California, Irvine, CA,
November 2010.
6. UCI Mathematical Physics Seminar, University of California, Irvine, CA, November 2010.
5. MAA Southern California-Nevada Section Fall 2010 Meeting, University of California, Irvine,
CA, October 2010.
4. Arizona School of Analysis with Applications, University of Arizona, Tucson, AZ, March
2010.
3. 2010 AMS/MAA Joint Mathematics Meetings, San Francisco, CA, January 2010.
2. UCI Mathematical Physics Seminar, University of California, Irvine, CA, November 2009.
1. SCSU Mathematics Colloquium, St. Cloud State University, St. Cloud, MN, November 2009.
RESEARCH EXPERIENCE
∙ Graduate Student Researcher (GSR), UCI, September, 2007 - April, 2011. The GSR
was supported by the AFOSR grant FA9550-08-1-0103 entitled: High-Q Photonic-Crystal
Cavities for Light Amplification.
TEACHING EXPERIENCE
TEACHING ASSISTANT
∙ Math2D: Multivariable Calculus. University of California, Irvine, Summer 2006
∙ Math184: History of Mathematics. University of California, Irvine, Spring 2006
∙ Math146: Fourier Analysis. University of California, Irvine, Spring 2006
∙ Math2B: Single Variable Calculus. University of California, Irvine, Fall 2005
∙ Math2B: Single Variable Calculus. University of California, Irvine, Summer 2005
∙ Math2A: Single Variable Calculus. University of California, Irvine, Summer 2005
∙ Math2J: Infinite Series and Linear Algebra. University of California, Irvine, Spring
2005
∙ Math2B: Single Variable Calculus. University of California, Irvine, Winter 2005
∙ Math2A: Single Variable Calculus. University of California, Irvine, Fall 2004
vii
INSTITUTIONAL SERVICE
∙ Co-Organizer. “Math Graduate Student Colloquium,” University of California, Irvine;
2010-2011
∙ Graduate Recruitment Speaker. “Mathematics Graduate Recruitment Day Event,”
University of California, Irvine; April 2010.
COMPUTATIONAL SKILLS
Programming: C/C++, MATLAB, and Maple.
Markup Languages: LaTeX and Beamer.
Operating Systems: Linux/Unix based systems and Windows.
viii
ABSTRACT OF THE DISSERTATION
On the Mathematics of Slow Light
By
Aaron Thomas Welters
Doctor of Philosophy in Mathematics
University of California, Irvine, 2011
Professor Aleksandr Figotin, Chair
This thesis develops a mathematical framework based on spectral perturbation theory for
the analysis of slow light and the slow wave regime for lossless one-dimensional photonic
crystals.
Electrodynamics of lossless one-dimensional photonic crystals incorporating general bianisotropic and dispersive materials are considered. The time-harmonic Maxwell’s equations
governing the electromagnetic wave propagation through such periodic structures is shown
to reduce to a canonical system of period differential-algebraic equations (DAEs) depending
holomorphically on frequency which we call Maxwell’s DAEs. In this context, a definition
of slow light is given.
We give a detailed perturbation analysis for degenerate eigenvalues of non-self-adjoint matrices. A generic condition is considered and its consequences are studied. We prove the generic
condition implies the degenerate eigenvalue of the unperturbed matrix under consideration
has a single Jordan block in its Jordan normal form corresponding to that eigenvalue. We
find explicit recursive formulas to calculate the perturbation expansions of the splitting eigenvalues and their eigenvectors. The coefficients up to the second order for these expansions
are given.
ix
An exposition on the spectral theory and spectral perturbation theory for holomorphic matrix
functions is given. This serves as a mathematical background for the tools and theory used
in this thesis.
We give a model for studying slow light based on canonical equations which are any canonical
system of periodic differential (canonical ODEs) or differential-algebraic (canonical DAEs)
equations depending holomorphically on a spectral parameter referred to as the frequency.
For canonical ODEs, we prove formulas connecting the dispersion relation, energy flux, and
energy density in our model to the monodromy matrix of the canonical ODEs. We prove one
of the main results of this thesis relating the existence of (non-Bloch) Floquet solutions of the
canonical ODEs and the occurrence of degenerate eigenvalues with nondiagonalizable Jordan
normal form of the monodromy matrix to the band structure of the dispersion relation near
spectral stationary points. For canonical DAEs, of which it is shown Maxwell’s DAEs are
an example, a theorem is given which shows that the model for canonical DAEs, including
the Floquet, spectral, and perturbation theory, is reduced to the model for canonical ODEs.
x
Chapter 1
Introduction
1.1
Motivation
Technology has progressed to the point at which we can now fabricate materials with dimensions on the nanoscale. To give you a perspective on how small this is consider that
one nanometer (nm) is equal to one billionth of a meter (m), the width of a human hair is
about 105 nm, and the wavelength of visible light lies in the approximate range of 400–700
nm. As a result, there has emerged a new class of materials (such as metamaterials [11, 54]),
which allow the possibility of novel electromagnetic properties and effects. For example,
bianisotropic crystals can be used to observe optical singularities [6], photonic crystals [27]
can be used to slow down light [16], negative index materials can be used to create a perfect
lens [50], and metametarials derived from transformation optics can be used to induce electromagnetic invisibility [21, 22, 40, 51]. The novel electromagnetic properties and effects that
metamaterials such as photonic crystals have, is a fascinating and productive area of research
for not only physicists but for mathematicians as well (see for example [32]). In particular,
research on slow light has received considerable attention recently. With its numerous appli-
1
cations such as solar cell efficiency enhancement [10] or antenna miniaturization [49, 59, 62],
there is a growing need for applied and theoretical studies on slow light phenomena.
Although there are several methods for slowing down light including electromagnetically induced transparency (EIT) and photonic crystals, it is the latter that is becoming of particular
interest to mathematicians. A major reason for this is the mathematics of photonic crystals
has been developed significantly over the last 20 years and this gives us a rigorous mathematical setting in which to begin studying problems involving slow light. Moreover, many
of the results on slow wave phenomenon in photonic crystals are expected to have analogies with wave propagation in general linear dispersive media. Thus in order to facilitate
further applied and theoretical studies on slow light, to better understand slow light from a
mathematical perspective, and to explore analogies with slow waves in periodic media, this
thesis considers a general nondissipative but dispersive model for wave propagation using
an important class of periodic differential and differential-algebraic equations (DAEs) called
canonical equations [31]. As we will show in the next section, this model is general enough
to include electromagnetic waves governed by the time-harmonic Maxwell’s equations for
lossless one-dimensional photonic crystals whose constituent layers can be any combination
of isotropic, anisotropic, or bianisotropic materials with or without material dispersion (i.e.,
frequency-dependent response of materials). This makes our work particularly significant
in the study of slow light since metamaterials are widening the range of potential photonic
crystals that can be fabricated and so a model like ours that has the ability to analysis slow
light phenomena for a broad range of photonic crystals is in need.
2
1.2
Electrodynamics of Lossless One-Dimensional Photonic Crystals
1.2.1
Time-Harmonic Maxwell’s Equations
Electromagnetic waves will be governed by the time-harmonic Maxwell’s equations (𝑒−𝑖𝜔𝑡
convention, 𝜔 ∕= 0) without sources in Gaussian units and Cartesian coordinates (with
respect to the standard orthonormal basis vectors e1 , e2 , e3 of ℝ3 ). These equations may be
written in the 2 × 2 block operator matrix form (see [5])
⎡
⎤⎡
⎡
⎤
⎤
∇× ⎥ ⎢ E(r) ⎥
𝜔 ⎢ D(r) ⎥
⎦⎣
⎦ = −𝑖 ⎣
⎦
𝑐 B(r)
−∇× 0
H(r)
0
⎢
⎣
(1.1)
where 𝑐 is the speed of light in a vacuum, r := (𝑥1 , 𝑥2 , 𝑥3 ) are the spatial variables and
the electric field E, magnetic field H, electric induction D, and magnetic induction B take
values in ℂ3 . Here ∇× denotes the curl operator on these fields and it is given by the matrix
operator
⎡
− ∂𝑥∂ 3
∂
∂𝑥2
∂
∂𝑥3
0
− ∂𝑥∂ 1
− ∂𝑥∂ 2
∂
∂𝑥1
0
0
⎢
⎢
∇× := ⎢
⎢
⎣
⎤
⎥
⎥
⎥.
⎥
⎦
(1.2)
The linear constitutive relations in 2 × 2 block matrix form are
⎡
⎤
⎡
⎤
⎡
⎤
⎢ D(r) ⎥
⎢ E(r) ⎥
⎢ 𝜀(𝑥3 , 𝜔) 𝜉(𝑥3 , 𝜔) ⎥
⎣
⎦ = C(𝑥3 , 𝜔) ⎣
⎦ , C(𝑥3 , 𝜔) = ⎣
⎦,
B(r)
H(r)
𝜁(𝑥3 , 𝜔) 𝜇(𝑥3 , 𝜔)
(1.3)
where 𝜀, 𝜇, 𝜉, 𝜁 are 3 × 3 matrix-valued functions representing the electric permittivity, magnetic permeability, and magnetoelectric coupling tensors, respectively. Here we have tacitly
3
assumed materials which are plane parallel layers normal to the 𝑥3 -axis with each layer
consisting of a homogeneous material whose frequency dependency is implicitly indicated.
1.2.2
Lossless 1-D Photonic Crystals
We are considering lossless one-dimensional photonic crystals with dispersive materials and
hence there exists a 𝑑 > 0 and an open connected set Ω ⊆ ℂ, the frequency domain, with
Ωℝ := Ω ∩ ℝ ∕= ∅ such that
(i) C(𝑥3 , 𝜔) = C(𝑥3 + 𝑑, 𝜔), for every 𝜔 ∈ Ω and a.e. 𝑥3 .
(ii) C(𝑥3 , 𝜔)∗ = C(𝑥3 , 𝜔), for every 𝜔 ∈ Ωℝ and a.e. 𝑥3 .
(iii) C(⋅, 𝜔) ∈ 𝑀6 (𝐿2 (𝕋)), C ∈ 𝒪(Ω, 𝑀6 (𝐿2 (𝕋)))1 .
1.2.3
Maxwell’s Equations as Canonical Equations
As in [16], we seek field solutions of the form
⎡
⎡
⎤
⎤
E(𝑥3 ) ⎥
⎢ E(r) ⎥
𝑖k ⋅r ⎢
⎦,
⎣
⎦=𝑒 ⊥ ⊥⎣
H(r)
H(𝑥3 )
(1.4)
where r⊥ = (𝑥1 , 𝑥2 , 0), k⊥ = (𝑘1 , 𝑘2 , 0), and 𝑘1 , 𝑘2 ∈ ℝ. Hence solutions to the time-harmonic
Maxwell’s equations in (1.1) with the field representations (1.4) are solutions to the canonical
equations, which we call Maxwell’s DAEs,
G 𝑦 ′ (𝑥3 ) = 𝑉 (𝑥3 , 𝜔)𝑦(𝑥3 ),
1
See section 4.5 for notation.
4
(1.5)
where 𝑦(𝑥3 ) is a 6 × 1 column vector, G is a 6 × 6 singular matrix, and 𝑉 , the Hamiltonian,
is a 6 × 6 matrix-valued function having the block representations
⎡
⎤
⎡
⎡
⎤
⎤
k⊥ × ⎥
e3 × ⎥
𝜔
⎢
⎢ E(𝑥3 ) ⎥
⎢
𝑦(𝑥3 ) = ⎣
⎦.
⎦, G = 𝑖⎣
⎦ , 𝑉 (𝑥3 , 𝜔) = C(𝑥3 , 𝜔) + ⎣
𝑐
−k⊥ ×
H(𝑥3 )
−e3 ×
(1.6)
Here e3 × and k⊥ × are the matrices
⎡
⎤
⎡
⎤
0 𝑘2 ⎥
⎢ 0 −1 0 ⎥
⎢ 0
⎢
⎥
⎢
⎥
⎥ , k⊥ × := ⎢ 0
⎥
e3 × := ⎢
1
0
0
0
−𝑘
1 ⎥.
⎢
⎥
⎢
⎣
⎦
⎣
⎦
0
0 0
−𝑘2 𝑘1 0
(1.7)
In particular, they are a canonical system of differential-algebraic equations (DAEs) with
periodic coefficients that depend holomorphically on the frequency 𝜔 where the leading
matrix coefficient G ∈ 𝑀6 (ℂ) and the matrix-valued function 𝑉 : ℝ × Ω → 𝑀6 (ℂ) have the
properties:
(i) det(G ) = 0, G ∗ = −G
(ii) 𝑉 (𝑥3 , 𝜔)∗ = 𝑉 (𝑥3 , 𝜔), for each 𝜔 ∈ Ωℝ and a.e. 𝑥3 ∈ ℝ
(iii) 𝑉 (𝑥3 + 𝑑, 𝜔) = 𝑉 (𝑥3 , 𝜔), for each 𝜔 ∈ Ω and a.e. 𝑥3 ∈ ℝ
(iv) 𝑉 ∈ 𝒪(Ω, 𝑀6 (𝐿2 (𝕋))) as a function of frequency.
In Chapter 4 we will consider these types of equations further.
5
1.2.4
Definition of Slow Light
In this context, Bloch solutions of Maxwell’s DAEs, i.e. solutions of the canonical equations
in (1.5) satisfying
𝑦(𝑥3 + 𝑑) = 𝑒𝑖𝑘𝑑 𝑦(𝑥3 ),
(1.8)
for some 𝑘 ∈ ℂ, give rise to the (axial) dispersion relation
𝜔 = 𝜔(𝑘).
(1.9)
For points (𝑘0 , 𝜔 0 ) ∈ ℝ2 on the graph of this dispersion relation, i.e., the Bloch variety ℬ, a
Bloch solution 𝑦 with wavenumber-frequency pair (𝑘0 , 𝜔 0 ) is said to be propagating with its
group velocity
𝑑𝜔 .
𝑑𝑘 (𝑘0 ,𝜔0 )
(1.10)
Under certain reservations [7, 8, 27, 63], its group velocity equals its energy velocity, i.e., the
ratio of its averaged energy flux to its energy density,
𝑑𝜔 =
𝑑𝑘 (𝑘0 ,𝜔0 )
∫
1 𝑑
⟨𝑖G 𝑦(𝑥3 ), 𝑦(𝑥3 )⟩𝑑𝑥3
,
∫ 𝑑 0
1 𝑑
⟨𝑉𝜔 (𝑥3 , 𝜔 0 )𝑦(𝑥3 ), 𝑦(𝑥3 )⟩𝑑𝑥3
𝑑 0
(1.11)
with its energy flux and its energy density are (up to multiplication by a constant) the
functions of the space variable 𝑥3 given by
𝑆 = ⟨𝑖G 𝑦, 𝑦⟩,
(1.12)
𝑈 = ⟨𝑉𝜔 (⋅, 𝜔 0 )𝑦, 𝑦⟩,
(1.13)
6
respectively, where 𝑉𝜔 denotes the derivative of the Hamiltonian 𝑉 with respect to frequency
in the 𝑀𝑁 (𝐿2 (𝕋)) norm and ⟨ , ⟩ is the standard inner product for ℂ6 .
We now give an intuitive definition of slow light for lossless one-dimensional photonic crystals.
In Chapter 4 we give a more precise definition using canonical equations. We will treat the
(axial) dispersion relation 𝜔 = 𝜔(𝑘) as a multi-valued function of the wavenumber 𝑘. We
will use the notation
𝑑𝜔
∣
𝑑𝑘 (𝑘0 ,𝜔0 )
= 0 to mean (𝑘0 , 𝜔 0 ) is a stationary point on the graph of the
dispersion relation 𝜔 = 𝜔(𝑘).
Definition 1 (Slow Light) If (𝑘0 , 𝜔 0 ) ∈ ℝ2 is a stationary point on the graph of the dispersion relation 𝜔 = 𝜔(𝑘), i.e.,
𝑑𝜔 = 0,
𝑑𝑘 (𝑘0 ,𝜔0 )
(1.14)
then any propagating Bloch solution of Maxwell’s DAEs (1.5) with wavenumber-frequency
pair (𝑘, 𝜔) satisfying 0 < ∣∣(𝑘, 𝜔) − (𝑘0 , 𝜔 0 )∣∣ ≪ 1 with its group velocity
𝑑𝜔
∣
𝑑𝑘 (𝑘,𝜔)
satisfying
∣
∣ ≪ 𝑐 is called a slow wave or slow light.
∣ 𝑑𝜔
𝑑𝑘 (𝑘,𝜔)
If (𝑘0 , 𝜔 0 ) ∈ ℝ2 is a stationary point on the graph of the dispersion relation 𝜔 = 𝜔(𝑘) then
an open ball 𝐵((𝑘0 , 𝜔 0 ), 𝑟) in ℂ2 with 0 < 𝑟 ≪ 1 is called the slow wave regime.
1.3
Main Results
The objective of this dissertation is to begin developing a mathematical framework based
on spectral perturbation theory for the analysis of slow light and the slow wave regime for
lossless one-dimensional photonic crystals.
In the following we describe the main contributions of this thesis which are contained in
7
Chapters 2 and 4.
Perturbation Analysis of Degenerate Eigenvalues from a Jordan
block
A fundamental problem in the perturbation theory for non-self-adjoint matrices with a degenerate spectrum is the determination of the perturbed eigenvalues and eigenvectors. Formulas
for the higher order terms of these perturbation expansions are often needed in problems
which require an accurate asymptotic analysis.
For example, my advisor A. Figotin and his colleague, I. Vitebskiy, considered scattering
problems involving slow light in one-dimensional semi-infinite photonic crystals [3, 12–16].
They found that only in the case of the frozen mode regime could incident light enter a
photonic crystal with little reflection and be converted into a slow wave. This frozen mode
regime was found to correspond to a stationary inflection point of the dispersion relation
and a 3 × 3 Jordan block in the Jordan normal form of the unit cell transfer matrix – the
monodromy matrix of the reduced Maxwell’s equations given in [16, §5] or [16, p. 332, (180)]
which are canonical ODEs although not in canonical form. In this setting, the eigenpairs of
the monodromy matrix corresponded to Bloch waves and their Floquet multipliers. Thus in
order for them to rigorously prove the physical results and provide a better understanding of
the very essence of the frozen mode regime, they needed an asymptotic analytic description
of the frozen mode regime which required a sophisticated mathematical framework based on
the spectral perturbation theory of a Jordan block. Unfortunately, at the time when [16] was
written such a theory did not exist and hence this was a big motivating factor for Chapter
22 of this thesis.
2
c 2011 Society for Industrial and Applied
The contents of this chapter also appear in [61]. Copyright ⃝
Mathematics. Reprinted with permission. All rights reserved.
8
In Chapter 2 of this thesis we to develop the spectral perturbation theory of a Jordan block
and address the following:
1. Determine a generic condition that allows a constructive spectral perturbation analysis
for non-self-adjoint matrices with degenerate eigenvalues.
2. What connection is there between that condition and the local Jordan normal form of
the matrix perturbation corresponding to a perturbed eigenvalue?
3. Does there exist explicit recursive formulas to determine the perturbed eigenvalues and
eigenvectors for non-selfadjoint perturbations of matrices with degenerate eigenvalues?
The statement of my main results regarding those three issues are contained in Theorem 1
and Theorem 2 of this thesis which are Theorem 2.1 and Theorem 3.1 from [61]. In particular, I have developed a constructive perturbation theory for non-self-adjoint matrices with
degenerate eigenvalues and found explicit recursive formulas to calculate the perturbation
expansions of the splitting eigenvalues and their eigenvectors, under a generic condition.
Canonical Equations: A Model for Studying Slow Light
We established above that the study of slow light, i.e., slow waves for lossless one-dimensional
photonic crystals is reduced to the study of Maxwell’s DAEs (1.5) near stationary points
of the dispersion relation, i.e., in the slow wave regime. As Maxwell’s DAEs are canonical
equations, it will be beneficial to use canonical equations to study slow wave propagation. By
analogy with that physical model, we get a general mathematical model for wave propagation
in periodic structures in which we can study the dispersion relation, band structure, spectral
stationary points, slow waves, and the slow wave regime. This is exactly what we do in
Chapter 4. The contents of this chapter is part of the paper [60] currently in preparation.
9
One of the major contribution in Chapter 4 of this thesis is Theorem 48. Combined with
Theorem 35, Theorem 47, and Theorem 50, it is our answer to the question:
1. How are the analytic properties of the dispersion relation and the degree of flatness of
spectral bands near stationary points related to (non-Bloch) Floquet solutions and the
occurrence of degenerate eigenvalues and a nondiagonalizable Jordan normal form for
the monodromy matrix of the canonical equations.
To answer this question and to proof Theorem 48 we need some deep results in the spectral
perturbation theory for holomorphic matrix functions. We give an exposition of this theory
in Chapter 3.
1.4
Overview
This thesis is organized in the following manner.
Chapter 2 concerns the perturbation analysis of non-self-adjoint matrices with degenerate
eigenvalues. A generic condition is considered and its consequences are studied. It is shown
that the generic condition implies the degenerate eigenvalue of the unperturbed matrix under
consideration has a single Jordan block in its Jordan normal form corresponding to the
degenerate eigenvalue. Explicit recursive formulas are given to calculate the perturbation
expansions of the splitting eigenvalues and their eigenvectors. The coefficients up to the
second order for these expansions are conveniently listed for quick reference.
Chapter 3 is an exposition on the spectral theory and spectral perturbation theory for holomorphic matrix functions. Its content is the required background material needed in Chapter
4 to prove the main result for that chapter, namely, Theorem 48.
Chapter 4 formulates and studies a model for slow wave propagation in period structures
10
using canonical equations. Both canonical ODEs and canonical DAEs are considered. Electromagnetic wave propagation in lossless one dimensional photonic crystals was discussed in
section 1.2 as a example for the need to include DAEs into the model. Motivation for the
definitions in the model was given by a consideration of their relevance in the example.
Next, in section 4.2, canonical ODEs are considered. Results are given on energy flux and
energy density. Then in section 4.2.3 an important theorem, Theorem 47, is given on points
of definite type for canonical ODEs. In particular, the result justifies the use of the term
definite type and states that to each point of definite type of the canonical ODEs there exists
a neighborhood of such points. This is followed up immediately in the next section, section
4.2.4, with the main result of the chapter, Theorem 48, pertaining to the perturbation theory
for canonical ODEs near points of definite type. As a corollary we give a result that connects
the generic condition of Chapter 2 to points of definite type and the Jordan normal form of
the monodromy matrix of the canonical ODEs.
Next, in section 4.3, canonical DAEs are considered. A theorem is stated that tells us the
theory of canonical DAEs is reduced to the study of canonical ODEs including the Floquet,
spectral, and perturbation theory.
Finally, section 4.4 gives the proofs of all the statements in this chapter. The use of the
spectral perturbation theory of holomorphic matrix functions is use to prove the main result
of the chapter, Theorem 48.
All of the work done here was completed under the guidance and supervision of Professor
Aleksandr Figotin.
11
Chapter 2
Perturbation Analysis of Degenerate
Eigenvalues from a Jordan block1
2.1
Introduction
Consider an analytic square matrix 𝐴 (𝜀) and its unperturbed matrix 𝐴 (0) with a degenerate
eigenvalue 𝜆0 . A fundamental problem in the analytic perturbation theory of non-selfadjoint
matrices is the determination of the perturbed eigenvalues near 𝜆0 along with their corresponding eigenvectors of the matrix 𝐴 (𝜀) near 𝜀 = 0. More specifically, let 𝐴 (𝜀) be a
matrix-valued function having a range in 𝑀𝑛 (ℂ), the set of 𝑛 × 𝑛 matrices with complex
entries, such that its matrix elements are analytic functions of 𝜀 in a neighborhood of the
origin. Let 𝜆0 be an eigenvalue of the matrix 𝐴 (0) with algebraic multiplicity 𝑚 ≥ 1. Then
in this situation, it is well known [4, §6.1.7], [28, §II.1.8] that for sufficiently small 𝜀 all the
perturbed eigenvalues near 𝜆0 , called the 𝜆0 -group, and their corresponding eigenvectors may
be represented as a collection of convergent Puiseux series, i.e., convergent Taylor series in
1
c 2011 Society for Industrial and Applied
The contents of this chapter also appear in [61]. Copyright ⃝
Mathematics. Reprinted with permission. All rights reserved.
12
a fractional power of 𝜀. What is not well known, however, is how we compute these Puiseux
series when 𝐴 (𝜀) is a non-selfadjoint analytic perturbation and 𝜆0 is a defective eigenvalue
of 𝐴 (0). There are sources on the subject like [4, §7.4], [41], [47], [57, §32], and [58] but it
was found that there lacked explicit formulas, recursive or otherwise, to compute the series
coefficients beyond the first order terms. Thus the fundamental problem that this paper
addresses is to find explicit recursive formulas to determine the Puiseux series coefficients
for the 𝜆0 -group and their eigenvectors.
This problem is of applied and theoretic importance, for example, in studying the spectral
properties of dispersive media such as photonic crystals. In particular, this is especially true
in the study of slow light [16, 49, 62], where the characteristic equation, det (𝜆𝐼 − 𝐴 (𝜀)) = 0,
represents implicitly the dispersion relation for Bloch waves in the periodic crystal. In this
setting 𝜀 represents a small change in frequency, 𝐴(𝜀) is the Transfer matrix of a unit cell, and
its eigenpairs, (𝜆(𝜀), 𝑥(𝜀)), correspond to the Bloch waves. From a practical and theoretical
point of view, condition (2.1) on the dispersion relation or its equivalent formulation in
Theorem 1.(i ) of this paper regarding the group velocity for this setting, arises naturally in
the study of slow light where the Jordan normal form of the unperturbed Transfer matrix,
𝐴(0), and the perturbation expansions of the eigenpairs of the Transfer matrix play a central
role in the analysis of slow light waves.
Main Results
In this paper under the generic condition,
∂
det (𝜆𝐼 − 𝐴 (𝜀)) (𝜀,𝜆)=(0,𝜆0 ) ∕= 0,
∂𝜀
(2.1)
we show that 𝜆0 is a non-derogatory eigenvalue of 𝐴(0) and the fundamental problem mentioned above can be solved. In particular, we prove Theorem 1 and Theorem 2 which together
13
state that when condition (2.1) is true then the Jordan normal form of 𝐴 (0) corresponding
to the eigenvalue 𝜆0 consists of a single 𝑚 × 𝑚 Jordan block, the 𝜆0 -group and their corresponding eigenvectors can each be represented by a single convergent Puiseux series whose
branches are given by
𝜆ℎ (𝜀) = 𝜆0 +
∞
∑
)𝑘
(
1
𝛼𝑘 𝜁 ℎ 𝜀 𝑚
𝑘=1
𝑥ℎ (𝜀) = 𝛽 0 +
∞
∑
)𝑘
(
1
𝛽 𝑘 𝜁 ℎ𝜀 𝑚
𝑘=1
1
2𝜋
∞
for ℎ = 0, . . . , 𝑚 − 1 and any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 , {𝛼𝑘 }∞
𝑘=1 ⊆ ℂ, {𝛽 𝑘 }𝑘=0 ⊆
ℂ𝑛 , 𝛼1 ∕= 0, and 𝛽 0 is an eigenvector of 𝐴 (0) corresponding to the eigenvalue 𝜆0 . More
importantly though, Theorem 2 gives explicit recursive formulas that allows us to determine
∞
the Puiseux series coefficients, {𝛼𝑘 }∞
𝑘=1 and {𝛽 𝑘 }𝑘=0 , from just the derivatives of 𝐴 (𝜀) at
𝜀 = 0. Using these recursive formulas, we compute the leading Puiseux series coefficients up
to the second order and list them in Corollary 4.
The key to all of our results is the study of the characteristic equation for the analytic matrix
𝐴 (𝜀) under the generic condition (2.1). By an application of the implicit function theorem,
we are able to derive the functional relation between the eigenvalues and the perturbation
parameter. This leads to the implication that the Jordan normal form of the unperturbed
matrix 𝐴 (0) corresponding to the eigenvalue 𝜆0 is a single 𝑚 × 𝑚 Jordan block. From this,
we are able to use the method of undetermined coefficients along with a careful combinatorial
analysis to get explicit recursive formulas for determining the Puiseux series coefficients.
We want to take a moment here to show how the results of this paper can be used to
determine the Puiseux series coefficients up to the second order for the case in which the
non-derogatory eigenvalue 𝜆0 has algebraic multiplicity 𝑚 ≥ 2. We start by putting 𝐴 (0)
14
into the Jordan normal form [35, §6.5: The Jordan Theorem]
⎤
⎡
⎢ 𝐽𝑚 (𝜆0 )
𝑈 −1 𝐴 (0) 𝑈 = ⎣
𝑊0
⎥
⎦,
(2.2)
where (see notations at end of §1) 𝐽𝑚 (𝜆0 ) is an 𝑚 × 𝑚 Jordan block corresponding to the
eigenvalue 𝜆0 and 𝑊0 is the Jordan normal form for the rest of the spectrum. Next, define
the vectors 𝑢1 ,. . . , 𝑢𝑚 , as the first 𝑚 columns of the matrix 𝑈 ,
𝑢𝑖 := 𝑈 𝑒𝑖 , 1 ≤ 𝑖 ≤ 𝑚
(2.3)
(These vectors have the properties that 𝑢1 is an eigenvector of 𝐴 (0) corresponding to the
eigenvalue 𝜆0 , they form a Jordan chain with generator 𝑢𝑚 , and are a basis for the algebraic eigenspace of 𝐴 (0) corresponding to the eigenvalue 𝜆0 ). We then partition the matrix
𝑈 −1 𝐴′ (0)𝑈 conformally to the blocks 𝐽𝑚 (𝜆0 ) and 𝑊0 of the matrix 𝑈 −1 𝐴 (0) 𝑈 as such
⎡
⎤
∗
⎢ ∗
⎢
..
..
⎢
.
.
⎢
⎢
⎢
⎢ ∗
∗
⎢
⎢
⎢ 𝑎𝑚−1,1 ∗
⎢
𝑈 −1 𝐴′ (0)𝑈 = ⎢
⎢ 𝑎
⎢ 𝑚,1 𝑎𝑚,2
⎢
⎢
⎢ ∗
∗
⎢
⎢
..
..
⎢
.
.
⎢
⎣
∗
∗
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗ ⎥
.. . . .. .. . . .. ⎥
. . .
. . ⎥
.
⎥
⎥
⎥
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗ ⎥
⎥
⎥
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗ ⎥
⎥
⎥.
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗ ⎥
⎥
⎥
⎥
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗ ⎥
⎥
.. . . .. .. . . .. ⎥
. . ⎥
. . .
.
⎥
⎦
∗ ⋅⋅⋅ ∗ ∗ ⋅⋅⋅ ∗
(2.4)
Now, by Theorem 1 and Theorem 2, it follows that
∂
∂𝜀
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 )
𝑎𝑚,1 = − ( ∂ 𝑚
) .
∂𝜆𝑚
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
𝑚!
15
(2.5)
And hence the generic condition is true if and only if 𝑎𝑚,1 ∕= 0. This gives us an alternative
method to determine whether the generic condition (2.1) is true or not.
Lets now assume that 𝑎𝑚,1 ∕= 0 and hence that the generic condition is true. Define 𝑓 (𝜀, 𝜆) :=
det (𝜆𝐼 − 𝐴 (𝜀)). Then by Theorem 2 and Corollary 4 there is exactly one convergent Puiseux
series for the perturbed eigenvalues near 𝜆0 and one for their corresponding eigenvectors
whose branches are given by
∞
)
(
)2 ∑
)𝑘
(
(
1
1
1
𝜆ℎ (𝜀) = 𝜆0 + 𝛼1 𝜁 ℎ 𝜀 𝑚 + 𝛼2 𝜁 ℎ 𝜀 𝑚 +
𝛼𝑘 𝜁 ℎ 𝜀 𝑚
(2.6)
𝑘=3
)
1
(
)2
1
(
𝑥ℎ (𝜀) = 𝑥0 + 𝛽 1 𝜁 ℎ 𝜀 𝑚 + 𝛽 2 𝜁 ℎ 𝜀 𝑚
+
∞
∑
(
)𝑘
1
𝛽 𝑘 𝜁 ℎ𝜀 𝑚
(2.7)
𝑘=3
2𝜋
1
for ℎ = 0, . . . , 𝑚 − 1 and any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 . Furthermore, the series
coefficients up to second order may be given by
(
1/𝑚
𝛼1 = 𝑎𝑚,1 =
𝛼2 =
∂𝑓
(0, 𝜆0 )
− 1 ∂𝜀
𝑚
∂ 𝑓
(0, 𝜆0 )
𝑚! ∂𝜆𝑚
−
𝑎𝑚−1,1 + 𝑎𝑚,2
=
𝑚𝛼𝑚−2
1
(
)1/𝑚
∕= 0,
(2.8)
∂ 𝑚+1 𝑓
1
𝛼𝑚+1
1
(𝑚+1)! ∂𝜆𝑚+1
(
𝑚−1
𝑚𝛼1
(0, 𝜆0 ) +
∂𝑚𝑓
𝑚
1
𝑚! ∂𝜆
∂2𝑓
𝛼1 ∂𝜆∂𝜀
)
(0, 𝜆0 )
⎧

⎨ −Λ𝐴′ (0)𝑢1 + 𝛼2 𝑢2 , if 𝑚 = 2
𝛽 0 = 𝑢1 , 𝛽 1 = 𝛼1 𝑢2 , 𝛽 2 =

⎩
𝛼 𝑢 + 𝛼2 𝑢 , if 𝑚 > 2
2 2
(0, 𝜆0 )
)
,
(2.9)
(2.10)
1 3
for any choice of the 𝑚th root of 𝑎𝑚,1 and where Λ is given in (2.15).
The explicit recursive formulas for computing higher order terms, 𝛼𝑘 , 𝛽 𝑘 , are given by (2.24)
and (2.25) in Theorem 2. The steps which should be used to determine these higher order
terms are discussed in Remark 1 and an example showing how to calculating 𝛼3 , 𝛽 3 using
these steps, when 𝑚 ≥ 3, is provided.
16
Example
The following example may help to give a better idea of these results. Consider
⎡
− 21
1
2
⎤
⎡
⎤
⎥
⎢ 2 0 −1 ⎥
⎢
⎥
⎥
⎥.
1 ⎥ + 𝜀⎢
1
0
−
2
0
−1
⎢
⎥
2
2 ⎥
⎦
⎣
⎦
−1 1
1
1 0
0
⎢
⎢
𝐴 (𝜀) := ⎢
⎢
⎣
1
(2.11)
Here 𝜆0 = 0 is a non-derogatory eigenvalue of 𝐴 (0) of algebraic multiplicity 𝑚 = 2. We put
𝐴 (0) into the Jordan normal form
⎡
⎤
⎡
⎤
⎡
0
⎢ 0 1 0 ⎥
⎢1 1 1⎥
⎢ 1 −1
⎢
⎥
⎢
⎥
⎢
−1
⎥
⎢
⎥
⎢
𝑈 −1 𝐴 (0) 𝑈 = ⎢
1
1
⎢ 0 0 0 ⎥ , 𝑈 = ⎢0 1 1⎥ , 𝑈 = ⎢ −1
⎣
⎦
⎣
⎦
⎣
1 1 0
1
0 −1
0 0 1/2
⎤
⎥
⎥
⎥,
⎥
⎦
so that 𝑊0 = 1/2. We next define the vectors 𝑢1 , 𝑢2 , as the first two columns of the matrix
𝑈,
⎡ ⎤
⎡ ⎤
⎢1⎥
⎢1⎥
⎢ ⎥
⎢ ⎥
⎢ ⎥
⎥
𝑢1 := ⎢
⎢0⎥ , 𝑢2 := ⎢1⎥ .
⎣ ⎦
⎣ ⎦
1
1
Next we partition the matrix 𝑈 −1 𝐴′ (0)𝑈 conformally to the blocks 𝐽𝑚 (𝜆0 ) and 𝑊0 of the
matrix 𝑈 −1 𝐴 (0) 𝑈 as such
⎡
⎤⎡
0
⎢ 1 −1
⎢
𝑈 −1 𝐴′ (0)𝑈 = ⎢
1
1
⎢ −1
⎣
1
0 −1
⎥⎢
⎥⎢
⎥⎢
⎥⎢
⎦⎣
⎤ ⎡
⎤⎡
2 0 −1 ⎥ ⎢1 1 1⎥ ⎢ 0 ∗ ∗
⎥⎢
⎥ ⎢
⎢0 1 1⎥ = ⎢ 1 1 ∗
2 0 −1 ⎥
⎥⎢
⎥ ⎢
⎦ ⎣
⎦⎣
1 1 0
1 0
0
∗ ∗ ∗
17
⎤
⎥
⎥
⎥.
⎥
⎦
Here 𝑎2,1 = 1, 𝑎1,1 = 0, and 𝑎2,2 = 1. Then
∂
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 )
) ,
1 = 𝑎2,1 = − ∂𝜀( ∂ 2
∂𝜆2
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
2!
implying that the generic condition (2.1) is true. Define 𝑓 (𝜀, 𝜆) := det (𝜆𝐼 − 𝐴 (𝜀)) =
𝜆3 − 2𝜆2 𝜀 − 12 𝜆2 + 𝜆𝜀2 − 12 𝜆𝜀 + 𝜀2 + 12 𝜀. Then there is exactly one convergent Puiseux series for
the perturbed eigenvalues near 𝜆0 = 0 and one for their corresponding eigenvectors whose
branches are given by
(
ℎ
𝜆ℎ (𝜀) = 𝜆0 + 𝛼1 (−1) 𝜀
1
2
)
(
ℎ
+ 𝛼2 (−1) 𝜀
1
2
)2
+
(
(
)
)
1
1 2
𝑥ℎ (𝜀) = 𝛽 0 + 𝛽 1 (−1)ℎ 𝜀 2 + 𝛽 2 (−1)ℎ 𝜀 2 +
∞
∑
𝑘=3
∞
∑
(
ℎ
𝛼𝑘 (−1) 𝜀
1
2
)𝑘
(
)
1 𝑘
𝛽 𝑘 (−1)ℎ 𝜀 2
𝑘=3
1
for ℎ = 0, 1 and any fixed branch of 𝜀 2 . Furthermore, the series coefficients up to second
order may be given by
v(
)
u
∂𝑓
u
√
(0,
𝜆
)
√
0
𝛼1 = 1 = 1 = 𝑎2,1 = ⎷ − 1 ∂𝜀
∕= 0,
∂2𝑓
(0,
𝜆
)
2
0
2! ∂𝜆
(
)
∂2𝑓
3 1 ∂3𝑓
−
𝛼
(0,
𝜆
)
+
𝛼
(0,
𝜆
)
0
1 ∂𝜆∂𝜀
0
1 3! ∂𝜆3
1
𝑎1,1 + 𝑎2,2
( 2
)
𝛼2 = =
=
,
2
2
𝛼1 2!1 ∂∂𝜆𝑓2 (0, 𝜆0 )
⎡ ⎤
⎡ ⎤
⎢1⎥
⎢1⎥
⎢ ⎥
⎢ ⎥
′
⎥
⎢ ⎥
𝛽0 = ⎢
⎢0⎥ , 𝛽 1 = ⎢1⎥ , 𝛽 2 = −Λ𝐴 (0)𝑢1 + 𝛼2 𝑢2
⎣ ⎦
⎣ ⎦
1
1
by choosing the positive square root of 𝑎2,1 = 1 and where Λ is given in (2.15). Here
⎡
⎤
∗
⎢ 𝐽𝑚 (0)
Λ=𝑈⎣
(𝑊0 − 𝜆0 𝐼𝑛−𝑚 )−1
⎥ −1
⎦𝑈
18
⎡
⎤⎡
0
⎢1 1 1⎥ ⎢ 0 0
⎢
⎥⎢
⎥⎢
=⎢
0
⎢0 1 1⎥ ⎢ 1 0
⎣
⎦⎣
1 1 0
0 0 (1/2)−1
⎤⎡
⎤
0
⎥ ⎢ 1 −1
⎥⎢
⎥ ⎢ −1
1
1
⎥⎢
⎦⎣
1
0 −1
⎥ ⎢ 3 −1 −2 ⎥
⎥ ⎢
⎥
⎥ = ⎢ 3 −1 −2 ⎥
⎥ ⎢
⎥
⎦ ⎣
⎦
1 −1
0
⎡
⎤
𝛽 2 = −Λ𝐴′ (0)𝑢1 + 𝛼2 𝑢2
⎡ ⎤
⎡ ⎤
⎡
⎤⎡
⎤⎡ ⎤
⎢1⎥
⎢1⎥
⎢ 3 −1 −2 ⎥ ⎢ 2 0 −1 ⎥ ⎢1⎥
⎥⎢
⎥⎢ ⎥ 1 ⎢ ⎥ 1 ⎢ ⎥
⎢
⎢ ⎥
⎢ ⎥
⎥⎢
⎥⎢ ⎥
= −⎢
⎢ 3 −1 −2 ⎥ ⎢ 2 0 −1 ⎥ ⎢0⎥ + 2 ⎢1⎥ = 2 ⎢1⎥ .
⎣ ⎦
⎣ ⎦
⎣
⎦⎣
⎦⎣ ⎦
1
1
1 −1
0
1 0
0
1
Now compare this to the actual perturbed eigenvalues of our example (2.11) near 𝜆0 = 0
and their corresponding eigenvectors
1
1 1
1
𝜆ℎ (𝜀) = 𝜀 + (−1)ℎ 𝜀 2 (𝜀 + 4) 2
2
2
∞
(
(
)
)2 ∑
)𝑘
1(
ℎ 1
ℎ 1
ℎ 1
2
2
2
= (−1) 𝜀 +
(−1) 𝜀
𝛼𝑘 (−1) 𝜀
+
2
𝑘=3
⎡ ⎤ ⎡ ⎤
⎢1⎥ ⎢1⎥
⎢ ⎥ ⎢ ⎥
⎥ ⎢ ⎥
𝑥ℎ (𝜀) = ⎢
⎢0⎥ + ⎢1⎥ 𝜆ℎ (𝜀)
⎣ ⎦ ⎣ ⎦
1
1
⎡ ⎤
⎡ ⎤ ⎡ ⎤
1⎥
⎢1⎥ ⎢1⎥ (
∞
(
) 1⎢
)2 ∑
)
⎢
⎢ ⎥ ⎢ ⎥
⎥(
1
1 𝑘
ℎ
ℎ 1
⎢
⎢
⎥
⎢
⎥
⎥
𝛽 𝑘 (−1)ℎ 𝜀 2
= ⎢0⎥ + ⎢1⎥ (−1) 𝜀 2 + ⎢1⎥ (−1) 𝜀 2 +
2⎣ ⎦
⎣ ⎦ ⎣ ⎦
𝑘=3
1
1
1
1
for ℎ = 0, 1 and any fixed branch of 𝜀 2 . We see that indeed our formulas for the Puiseux
series coefficients are correct up to the second order.
19
Comparison to Known Results
There is a fairly large amount of literature on eigenpair perturbation expansions for analytic perturbations of non-selfadjoint matrices with degenerate eigenvalues (e.g. [1, 2, 4,
9, 23, 26, 28, 33, 34, 37, 41, 43, 46–48, 53, 55–58]). However, most of the literature (e.g.
[1, 2, 23, 33, 34, 37, 41, 43, 46–48, 55, 56]) contains results only on the first order expansions
of the Puiseux series or considers higher order terms only in the case of simple or semisimple
eigenvalues. For those works that do address higher order terms for defective eigenvalues
(e.g. [4, 9, 26, 28, 53, 57, 58]), it was found that there did not exist explicit recursive formulas
for all the Puiseux coefficients when the matrix perturbations were non-linear. One of the
purposes and achievements of this paper are the explicit recursive formulas (2.23)–(2.25)
in Theorem 2 which give all the higher order terms in the important case of degenerate
eigenvalues which are non-derogatory, that is, the case in which a degenerate eigenvalue of
the unperturbed matrix has a single Jordan block for its corresponding Jordan structure.
Our theorem generalizes and extends the results of [4, pp. 315–317, (4.96) & (4.97)], [57, pp.
415–418], and [58, pp. 17–20] to non-linear analytic matrix perturbations and makes explicit
the recursive formulas for calculating the perturbed eigenpair Puiseux expansions. Furthermore, in Proposition 8 we give an explicit recursive formula for calculating the polynomials
{𝑟𝑙 }𝑙∈ℕ . These polynomials must be calculated in order to determine the higher order terms
in the eigenpair Puiseux series expansions (see (2.25) in Theorem 2 and Remark 1). These
polynomials appear in [4, p. 315, (4.95)], [57, p. 414, (32.24)], and [58, p. 19, (34)] under different notation (compare with Proposition 2.41.ii) but no method is given to calculate them.
As such, Proposition 8 is an important contribution in the explicit recursive calculation of
the higher order terms in the eigenpair Puiseux series expansions.
Another purpose of this paper is to give, in the case of degenerate non-derogatory eigenvalues, an easily accessible and quickly referenced list of first and second order terms for the
Puiseux series expansions of the perturbed eigenpairs. When the generic condition (2.1) is
20
satisfied, Corollary 4 gives this list. Now for first order terms there are quite a few papers on
formulas for determining them, see for example [48] which gives a good survey of first order
perturbation theory. But for second order terms, it was difficult to find any results in the
literature similar to and as explicit as Corollary 4 for the case of degenerate non-derogatory
eigenvalues with arbitrary algebraic multiplicity and non-linear analytic perturbations. Results comparable to ours can be found in [4, p. 316], [57, pp. 415–418], [58, pp. 17-20], and
[53, pp. 37–38, 50–54, 125–128], although it should be noted that in [57, p. 417] the formula
for the second order term of the perturbed eigenvalues contains a misprint.
Overview
Section 2 deals with the generic condition (2.1). We give conditions that are equivalent to the
generic condition in Theorem 1. In §3 we give the main results of this paper in Theorem 2,
on the determination of the Puiseux series with the explicit recursive formulas for calculating
the series coefficients. As a corollary we give the exact leading order terms, up to the second
order, for the Puiseux series coefficients. Section 4 contains the proofs of the results in §2
and §3.
Notation
Let 𝑀𝑛 (ℂ) be the set of all 𝑛 × 𝑛 matrices with complex entries and ℂ𝑛 the set of all 𝑛 × 1
column vectors with complex entries. For 𝑎 ∈ ℂ, 𝐴 ∈ 𝑀𝑛 (ℂ), and 𝑥 = [𝑎𝑖,1 ]𝑛𝑖=1 ∈ ℂ𝑛 we
denote by 𝑎∗ , 𝐴∗ , and 𝑥∗ , the complex conjugate of 𝑎, the conjugate transpose of 𝐴, and the
1 × 𝑛 row vector 𝑥∗ := [ 𝑎∗1,1 ⋅ ⋅ ⋅
𝑛
∗
𝑎∗𝑛,1 ]. For 𝑥, 𝑦 ∈ ℂ we let ⟨𝑥, 𝑦⟩ := 𝑥 𝑦 be the standard
inner product. The matrix 𝐼 ∈ 𝑀𝑛 (ℂ) is the identity matrix and its 𝑗th column is 𝑒𝑗 ∈ ℂ𝑛 .
The matrix 𝐼𝑛−𝑚 is the (𝑛 − 𝑚) × (𝑛 − 𝑚) identity matrix. Define an 𝑚 × 𝑚 Jordan block
21
with eigenvalue 𝜆 to be
⎡
⎤
𝜆 1
⎢
⎥
⎢
⎥
⎢
⎥
.
.
⎢
⎥
⎢
⎥
⎢
𝐽𝑚 (𝜆) := ⎢
. . ⎥
⎥.
⎢
⎥
⎢
⎥
. 1⎥
⎢
⎣
⎦
𝜆
When the matrix 𝐴 (𝜀) ∈ 𝑀𝑛 (ℂ) is analytic at 𝜀 = 0 we define 𝐴′ (0) :=
1 𝑑𝑘 𝐴
(0).
𝑘! 𝑑𝜀𝑘
2.2
𝑑𝐴
(0)
𝑑𝜀
and 𝐴𝑘 :=
2𝜋
Let 𝜁 := 𝑒𝑖 𝑚 .
The Generic Condition and its Consequences
The following theorem, which is proved in §4, gives conditions which are equivalent to the
generic one (2.1).
Theorem 1 Let 𝐴 (𝜀) be a matrix-valued function having a range in 𝑀𝑛 (ℂ) such that its
matrix elements are analytic functions of 𝜀 in a neighborhood of the origin. Let 𝜆0 be an
eigenvalue of the unperturbed matrix 𝐴 (0) and denote by 𝑚 its algebraic multiplicity. Then
the following statements are equivalent:
( i) The characteristic polynomial det (𝜆𝐼 − 𝐴 (𝜀)) has a simple zero with respect to 𝜀 at
𝜆 = 𝜆0 and 𝜀 = 0, i.e.,
∂
det (𝜆𝐼 − 𝐴 (𝜀)) (𝜀,𝜆)=(0,𝜆0 ) ∕= 0.
∂𝜀
( ii) The characteristic equation, det(𝜆𝐼 − 𝐴 (𝜀)) = 0, has a unique solution, 𝜀 (𝜆), in a
neighborhood of 𝜆 = 𝜆0 with 𝜀 (𝜆0 ) = 0. This solution is an analytic function with a
22
zero of order 𝑚 at 𝜆 = 𝜆0 , i.e.,
𝑑𝑚−1 𝜀 (𝜆) 𝑑𝑚 𝜀 (𝜆) 𝑑0 𝜀 (𝜆) =
⋅
⋅
⋅
=
=
0,
∕= 0.
𝑑𝜆𝑚 𝜆=𝜆0
𝑑𝜆0 𝜆=𝜆0
𝑑𝜆𝑚−1 𝜆=𝜆0
( iii) There exists a convergent Puiseux series whose branches are given by
ℎ
𝜆ℎ (𝜀) = 𝜆0 + 𝛼1 𝜁 𝜀
1
𝑚
+
∞
∑
(
ℎ
𝛼𝑘 𝜁 𝜀
1
𝑚
)𝑘
,
𝑘=2
1
2𝜋
for ℎ = 0, . . . , 𝑚−1 and any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 , such that the values of
the branches give all the solutions of the characteristic equation, for sufficiently small
𝜀 and 𝜆 sufficiently near 𝜆0 . Furthermore, the first order term is nonzero, i.e.,
𝛼1 ∕= 0.
( iv) The Jordan normal form of 𝐴 (0) corresponding to the eigenvalue 𝜆0 consists of a single
𝑚 × 𝑚 Jordan block and there exists an eigenvector 𝑢0 of 𝐴 (0) corresponding to the
eigenvalue 𝜆0 and an eigenvector 𝑣0 of 𝐴 (0)∗ corresponding to the eigenvalue 𝜆∗0 such
that
⟨𝑣0 , 𝐴′ (0)𝑢0 ⟩ =
∕ 0.
2.3
Explicit Recursive Formulas for Calculating the
Perturbed Spectrum
This section contains the main results of this paper presented below in Theorem 2. To begin
we give some preliminaries that are needed to set up the theorem. Suppose that 𝐴 (𝜀) is
a matrix-valued function having a range in 𝑀𝑛 (ℂ) with matrix elements that are analytic
23
functions of 𝜀 in a neighborhood of the origin and 𝜆0 is an eigenvalue of the unperturbed
matrix 𝐴 (0) with algebraic multiplicity 𝑚. Assume that the generic condition
∂
det (𝜆𝐼 − 𝐴 (𝜀)) (𝜀,𝜆)=(0,𝜆0 ) ∕= 0,
∂𝜀
is true.
Now, by these assumptions, we may appeal to Theorem 1.(iv ) and conclude that the Jordan
canonical form of 𝐴(0) has only one 𝑚 × 𝑚 Jordan block associated with 𝜆0 . Hence there
exists a invertible matrix 𝑈 ∈ ℂ𝑛×𝑛 such that
⎡
⎢ 𝐽𝑚 (𝜆0 )
𝑈 −1 𝐴 (0) 𝑈 = ⎣
⎤
𝑊0
⎥
⎦,
(2.12)
where 𝑊0 is a (𝑛 − 𝑚) × (𝑛 − 𝑚) matrix such that 𝜆0 is not one of its eigenvalues [35, §6.5:
The Jordan Theorem].
We define the vectors 𝑢1 , . . . , 𝑢𝑚 , 𝑣1 , . . . , 𝑣𝑚 ∈ ℂ𝑛 as the first 𝑚 columns of the matrix 𝑈
∗
and (𝑈 −1 ) , respectively, i.e.,
𝑢𝑖 := 𝑈 𝑒𝑖 , 1 ≤ 𝑖 ≤ 𝑚,
(
)∗
𝑣𝑖 := 𝑈 −1 𝑒𝑖 , 1 ≤ 𝑖 ≤ 𝑚.
(2.13)
(2.14)
And define the matrix Λ ∈ 𝑀𝑛 (ℂ) by
⎡
⎤
∗
⎢ 𝐽𝑚 (0)
Λ := 𝑈 ⎣
(𝑊0 − 𝜆0 𝐼𝑛−𝑚 )−1
⎥ −1
⎦𝑈 ,
(2.15)
where (𝑊0 − 𝜆0 𝐼𝑛−𝑚 )−1 exists since 𝜆0 is not an eigenvalue of 𝑊0 (for the important properties
of the matrix Λ see Section 2.5).
24
Next, we introduce the polynomials 𝑝𝑗,𝑖 = 𝑝𝑗,𝑖 (𝛼1 , . . . , 𝛼𝑗−𝑖+1 ) in 𝛼1 ,. . . , 𝛼𝑗−𝑖+1 , for 𝑗 ≥ 𝑖 ≥
0, as the expressions
⎫




⎬
𝑝0,0 := 1, 𝑝𝑗,0 := 0, for 𝑗 > 0,
𝑖
∑
∏
𝑝𝑗,𝑖 (𝛼1 , . . . , 𝛼𝑗−𝑖+1 ) :=
𝛼𝑠𝜚 , for 𝑗 ≥ 𝑖 > 0 



𝑠1 +⋅⋅⋅+𝑠𝑖 =𝑗 𝜚=1
⎭
(2.16)
1≤𝑠𝜚 ≤𝑗−𝑖+1
and the polynomials 𝑟𝑙 = 𝑟𝑙 (𝛼1 , . . . , 𝛼𝑙 ) in 𝛼1 ,. . . , 𝛼𝑙 , for 𝑙 ≥ 1, as the expressions
∑
𝑟1 := 0, 𝑟𝑙 (𝛼1 , . . . , 𝛼𝑙 ) :=
𝑚
∏
𝛼𝑠𝜚 , for 𝑙 > 1
(2.17)
𝑠1 +⋅⋅⋅+𝑠𝑚 =𝑚+𝑙 𝜚=1
1≤𝑠𝜚 ≤𝑙
(see Section 2.5 for more details on these polynomials including recursive formulas for their
calculation).
With these preliminaries we can now state the main results of this paper. Proofs of these
results are contained in the next section.
Theorem 2 Let 𝐴 (𝜀) be a matrix-valued function having a range in 𝑀𝑛 (ℂ) such that its
matrix elements are analytic functions of 𝜀 in a neighborhood of the origin. Let 𝜆0 be an
eigenvalue of the unperturbed matrix 𝐴 (0) and denote by 𝑚 its algebraic multiplicity. Suppose
that the generic condition
∂
det (𝜆𝐼 − 𝐴 (𝜀)) (𝜀,𝜆)=(0,𝜆0 ) ∕= 0,
∂𝜀
(2.18)
is true. Then there is exactly one convergent Puiseux series for the 𝜆0 -group and one for
their corresponding eigenvectors whose branches are given by
𝜆ℎ (𝜀) = 𝜆0 +
∞
∑
𝑘=1
25
(
)𝑘
1
𝛼𝑘 𝜁 ℎ 𝜀 𝑚
(2.19)
𝑥ℎ (𝜀) = 𝛽 0 +
∞
∑
)𝑘
(
1
𝛽 𝑘 𝜁 ℎ𝜀 𝑚
(2.20)
𝑘=1
1
2𝜋
for ℎ = 0, . . . , 𝑚 − 1 and any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 with
∂
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 )
∂𝜀
𝛼𝑚
=
⟨𝑣
,
𝐴
𝑢
⟩
=
−
) ∕= 0
( ∂𝑚
𝑚
1 1
1
∂𝜆𝑚
(Here 𝐴1 denotes
𝑑𝐴
(0)
𝑑𝜀
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
𝑚!
and the vectors 𝑢1 and 𝑣𝑚 are defined in (2.13) and (2.14)). Fur-
thermore, we can choose
𝛼1 = ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩1/𝑚 ,
(2.21)
for any fixed 𝑚th root of ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩ and the eigenvectors to satisfy the normalization conditions
⟨𝑣1 , 𝑥ℎ (𝜀)⟩ = 1, ℎ = 0, ..., 𝑚 − 1.
(2.22)
Consequently, under these conditions 𝛼1 , 𝛼2 ,. . . and 𝛽 0 , 𝛽 1 , . . . are uniquely determined
and are given by the recursive formulas
⎞1/𝑚
⎛
∂
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 ) ⎟
⎜
𝛼1 = ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩1/𝑚 = ⎝− ∂𝜀( ∂ 𝑚
) ⎠
∂𝜆𝑚
−𝑟𝑠−1 +
𝑖=0
𝛼𝑠 =
𝛽𝑠 =
min{𝑠,𝑚}−1
𝑠−1
∑
∑
⟨
𝑝𝑗,𝑖
(2.23)
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
𝑚!
𝑣𝑚−𝑖 ,
⌋
⌊ 𝑚+𝑠−1−𝑗
𝑚
∑
𝑗=𝑖
⟩
𝐴𝑘 𝛽 𝑚+𝑠−1−𝑗−𝑘𝑚
𝑘=1
𝑚𝛼𝑚−1
1
⎧



⎨
𝑠
∑
𝑝𝑠,𝑖 𝑢𝑖+1 , if 0 ≤ 𝑠 ≤ 𝑚 − 1
𝑖=0
𝑗 ⌊∑
𝑚−1
𝑠−𝑚
𝑚 ⌋
∑
∑ ∑



𝑝
𝑢
−
𝑝𝑗,𝑘 Λ𝑘+1 𝐴𝑙 𝛽 𝑠−𝑗−𝑙𝑚 , if 𝑠 ≥ 𝑚
𝑠,𝑖 𝑖+1
⎩
𝑠−𝑗
𝑖=0
(2.24)
𝑗=0 𝑘=0
(2.25)
𝑙=1
where 𝑢𝑖 and 𝑣𝑖 are the vectors defined in (2.13) and (2.14), 𝑝𝑗,𝑖 and 𝑟𝑙 are the polynomials
defined in (2.16) and (2.17), ⌊⌋ denotes the floor function, 𝐴𝑘 denotes the matrix
and Λ is the matrix defined in (2.15).
26
1 𝑑𝑘 𝐴
(0),
𝑘! 𝑑𝜀𝑘
Corollary 3 The calculation of the 𝑘th order terms, 𝛼𝑘 and 𝛽 𝑘 , requires only the matrices
𝐴0 , . . . , 𝐴⌊ 𝑚+𝑘−1 ⌋ .
𝑚
Corollary 4 The coefficients of those Puiseux series up to second order are given by
)1/𝑚
∂𝑓
(0,
𝜆
)
0
− 1 ∂𝜀
= ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩1/𝑚 ,
∂𝑚𝑓
(0, 𝜆0 )
∂𝜆𝑚
⎧ 𝑚!
(
)
∂2𝑓
∂ 𝑚+1 𝑓
𝑚+1
1
1 ∂2𝑓

(0,𝜆
)+𝛼
(0,𝜆
)
(0,𝜆
)+
−
𝛼
0
1
0
0
𝑚+1
2
1

(𝑚+1)!
∂𝜆∂𝜀
2
(
𝛼1 =
∂𝜆
𝛼2 =

⎨
1
𝑚𝛼𝑚−1
( 𝑚!
1
(
1
− 𝛼𝑚+1
1
(𝑚+1)!
∂𝜀
∂𝑚𝑓
(0,𝜆0 )
∂𝜆𝑚
)
, if 𝑚 = 1
)
∂ 𝑚+1 𝑓
∂2𝑓

(0,𝜆0 )+𝛼1 ∂𝜆∂𝜀
(0,𝜆0 )

∂𝜆𝑚+1

⎩
, if 𝑚 > 1
𝑚−1 1 ∂ 𝑚 𝑓
𝑚𝛼1
( 𝑚! ∂𝜆𝑚 (0,𝜆0 ))
⎧

⎨ ⟨𝑣1 , (𝐴2 − 𝐴1 Λ𝐴1 ) 𝑢1 ⟩, if 𝑚 = 1
=
,

𝑚 ,𝐴1 𝑢2 )
⎩ (𝑣𝑚−1 ,𝐴1 𝑢1 )+(𝑣
, if 𝑚 > 1
𝑚𝛼𝑚−2
1
𝛽 0 = 𝑢1 ,
⎧

⎨ −Λ𝐴1 𝑢1 , if 𝑚 = 1
𝛽1 =
,

⎩ 𝛼1 𝑢2 , if 𝑚 > 1
⎧
(
)

2
2

−Λ𝐴
+
(Λ𝐴
)
−
𝛼
Λ
𝐴
𝑢1 , if 𝑚 = 1

2
1
1
1

⎨
𝛽2 =
,
−Λ𝐴1 𝑢1 + 𝛼2 𝑢2 , if 𝑚 = 2




⎩
𝛼2 𝑢2 + 𝛼21 𝑢3 , if 𝑚 > 2
where 𝑓 (𝜀, 𝜆) := det (𝜆𝐼 − 𝐴 (𝜀)).
Remark 1 Suppose we want to calculate the terms 𝛼𝑘+1 , 𝛽 𝑘+1 , where 𝑘 ≥ 2, using the
explicit recursive formulas given in the theorem. We may assume we already known or have
calculated
𝑘−1
𝐴0 , . . . , 𝐴⌊ 𝑚+𝑘 ⌋ , {𝑟𝑗 }𝑗=1
, {𝛼𝑗 }𝑘𝑗=1 , {𝛽 𝑗 }𝑘𝑗=0 , {{𝑝𝑗,𝑖 }𝑘𝑗=𝑖 }𝑘𝑖=0 .
(2.26)
𝑚
We need these to calculate 𝛼𝑘+1 , 𝛽 𝑘+1 and the steps to do this are indicated by the following
27
arrow diagram:
(2.45)
(2.24)
(2.44)
(2.25)
(2.26) → 𝑟𝑘 → 𝛼𝑘+1 → {𝑝𝑘+1,𝑖 }𝑘+1
𝑖=0 → 𝛽 𝑘+1 .
(2.27)
After we have followed these steps we not only will have calculated 𝛼𝑘+1 , 𝛽 𝑘+1 but we will
now know
𝑘+1
𝑘+1 𝑘+1
𝐴0 , . . . , 𝐴⌊ 𝑚+𝑘+1 ⌋ , {𝑟𝑗 }𝑘𝑗=1 , {𝛼𝑗 }𝑘+1
𝑗=1 , {𝛽 𝑗 }𝑗=0 , {{𝑝𝑗,𝑖 }𝑗=𝑖 }𝑖=0
(2.28)
𝑚
as well. But these are the terms in (2.26) for 𝑘 + 1 and so we may repeat the steps indicated
above to calculate 𝛼𝑘+2 , 𝛽 𝑘+2 .
It is in this way we see how all the higher order terms can be calculated using the results of
this paper.
Example
In order to illustrate these steps we give the following example which recursively calculates
the third order terms for 𝑚 ≥ 3.
The goal is to determine 𝛼3 , 𝛽 3 . To do this we follow the steps indicated in the above remark
with 𝑘 = 2. The first step is to collect the terms in (2.26). Assuming 𝐴0 , 𝐴1 are known then
by (2.16), (2.17), Corollary 4, and Proposition 7 we have
𝐴0 , 𝐴1 , 𝑟1 = 0, 𝛼1 = ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩1/𝑚 , 𝛼2 =
⟨𝑣𝑚−1 ,𝐴1 𝑢1 ⟩+⟨𝑣𝑚 ,𝐴1 𝑢2 ⟩
,
𝑚𝛼𝑚−2
1
𝛽 0 = 𝑢1 , 𝛽 1 = 𝛼1 𝑢2 , 𝛽 2 = 𝛼2 𝑢2 + 𝛼21 𝑢3 ,
𝑝0,0 = 1, 𝑝1,0 = 0, 𝑝1,1 = 𝛼1 , 𝑝2,0 = 0, 𝑝2,1 = 𝛼2 , 𝑝2,2 = 𝛼21 .
The next step is to determine 𝑟2 using the recursive formula for the 𝑟𝑙 ’s given in (2.45). We
28
find that
1
1
∑
1 ∑
𝑚
[(3 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼3−𝑗 𝛼𝑗+1
[(3 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼3−𝑗 𝑟𝑗 + 𝛼𝑚−2
1
2𝛼1 𝑗=1
2
𝑗=1
𝑚(𝑚 − 1) 𝑚−2 2
=
𝛼1 𝛼2 .
2
𝑟2 =
Now, since 𝑟2 is determined, we can use the recursive formula in (2.24) for the 𝛼𝑠 ’s to
calculate 𝛼3 . In doing so we find that
−𝑟2 +
𝛼3 =
min{3,𝑚}−1
2
∑
∑
𝑖=0
⟨
𝑝𝑗,𝑖
𝑣𝑚−𝑖 ,
𝑗=𝑖
⌊ 𝑚+2−𝑗
⌋
𝑚
∑
⟩
𝐴𝑘 𝛽 𝑚+2−𝑗−𝑘𝑚
𝑘=1
𝑚𝛼𝑚−1
1
−𝑟2 + 𝑝2,1 ⟨𝑣𝑚−1 , 𝐴1 𝛽 0 ⟩ + 𝑝0,0 ⟨𝑣𝑚 , 𝐴1 𝛽 2 ⟩
=
+
𝑚𝛼𝑚−1
1
𝑝2,2 ⟨𝑣𝑚−2 , 𝐴1 𝛽 0 ⟩ + 𝑝1,1 ⟨𝑣𝑚−1 , 𝐴1 𝛽 1 ⟩
𝑚𝛼𝑚−1
1
− 𝑚(𝑚−1)
𝛼𝑚−2
𝛼22 + 𝛼2 ⟨𝑣𝑚−1 , 𝐴1 𝑢1 ⟩ + ⟨𝑣𝑚 , 𝐴1 (𝛼2 𝑢2 + 𝛼21 𝑢3 )⟩
1
2
=
+
𝑚𝛼𝑚−1
1
𝛼21 ⟨𝑣𝑚−2 , 𝐴1 𝑢1 ⟩ + 𝛼1 ⟨𝑣𝑚−1 , 𝐴1 𝛼1 𝑢2 ⟩
𝑚𝛼𝑚−1
1
(
)
3−𝑚
⟨𝑣𝑚−2 , 𝐴1 𝑢1 ⟩ + ⟨𝑣𝑚−1 , 𝐴1 𝑢2 ⟩ + ⟨𝑣𝑚 , 𝐴1 𝑢3 ⟩
2
=
𝛼−1
.
1 𝛼2 +
2
𝑚𝛼𝑚−3
1
Next, since 𝛼3 is determined, we can use (2.44) to calculate {𝑝3,𝑖 }3𝑖=0 . In this case though it
suffices to use Proposition 7 and in doing so we find that
𝑝3,0 = 0, 𝑝3,1 = 𝛼3 , 𝑝3,2 = 2𝛼1 𝛼2 , 𝑝3,3 = 𝛼31 .
Finally, we can compute 𝛽 3 using the recursive formula in (2.25) for the 𝛽 𝑠 ’s. In doing so we
29
find that
𝛽3 =
⎧



⎨
3
∑
𝑝3,𝑖 𝑢𝑖+1 , if 𝑚 > 3
𝑖=0
3−𝑗
𝑗 ⌊∑
𝑚−1
3−𝑚
𝑚 ⌋

∑
∑ ∑


𝑝3,𝑖 𝑢𝑖+1 −
𝑝𝑗,𝑘 Λ𝑘+1 𝐴𝑙 𝛽 3−𝑗−𝑙𝑚 , if 𝑚 = 3
⎩
𝑖=0
𝑗=0 𝑘=0 𝑙=1
⎧

⎨ 𝑝3,1 𝑢2 + 𝑝3,2 𝑢3 + 𝑝3,3 𝑢4 , if 𝑚 > 3
=
2
∑

⎩
𝑝3,𝑖 𝑢𝑖+1 − Λ𝐴1 𝛽 0 , if 𝑚 = 3
𝑖=0
⎧

⎨ 𝛼3 𝑢2 + 2𝛼1 𝛼2 𝑢3 + 𝛼3 𝑢4 , if 𝑚 > 3
1
=

⎩ 𝛼3 𝑢2 + 2𝛼1 𝛼2 𝑢3 − Λ𝐴1 𝑢1 , if 𝑚 = 3.
This completes the calculation of the third order terms, 𝛼3 , 𝛽 3 , when 𝑚 ≥ 3.
2.4
Proofs
This section contains the proofs of the results of this paper. We begin by proving Theorem
1 of §2 on conditions equivalent to the generic condition. We next follow this up with the
proof of the main result of this paper Theorem 2. We finish by proving the Corollaries 3 and
4.
Proof of Theorem 1
To prove this theorem we will prove the chain of statements (i)⇒(ii)⇒(iii)⇒(iv)⇒(i).
We begin by proving (i)⇒(ii). Define 𝑓 (𝜀, 𝜆) := det (𝜆𝐼 − 𝐴 (𝜀)) and suppose (i) is true.
Then 𝑓 is an analytic function of (𝜀, 𝜆) near (0, 𝜆0 ) since the matrix elements of 𝐴 (𝜀) are
analytic functions of 𝜀 in a neighborhood of the origin and the determinant of a matrix
is a polynomial in its matrix elements. Also we have 𝑓 (0, 𝜆0 ) = 0 and
30
∂𝑓
∂𝜀
(0, 𝜆0 ) ∕= 0.
Hence by the holomorphic implicit function theorem [30, §1.4 Theorem 1.4.11] there exists
a unique solution, 𝜀 (𝜆), in a neighborhood of 𝜆 = 𝜆0 with 𝜀 (𝜆0 ) = 0 to the equation
𝑓 (𝜀, 𝜆) = 0, which is analytic at 𝜆 = 𝜆0 . We now show that 𝜀 (𝜆) has a zero there of order
𝑚 at 𝜆 = 𝜆0 . First, the properties of 𝜀(𝜆) imply there exists 𝜀𝑞 ∕= 0 and 𝑞 ∈ ℕ such that
(
)
𝜀 (𝜆) = 𝜀𝑞 (𝜆 − 𝜆0 )𝑞 + 𝑂 (𝜆 − 𝜆0 )𝑞+1 , for ∣𝜆 − 𝜆0 ∣ << 1. Next, by hypothesis 𝜆0 is an
eigenvalue of 𝐴 (0) of algebraic multiplicity 𝑚 hence ∂ 𝑖 𝑓 ∖∂𝜆𝑖 (0, 𝜆0 ) = 0 for 0 ≤ 𝑖 ≤ 𝑚 − 1
but ∂ 𝑚 𝑓 ∖∂𝜆𝑚 (0, 𝜆0 ) ∕= 0. Combining this with the fact that 𝑓 (0, 𝜆0 ) = 0 and
∂𝑓
∂𝜀
(0, 𝜆0 ) ∕= 0
we have
∑
𝑓 (𝜀, 𝜆) = 𝑎10 𝜀 + 𝑎0𝑚 (𝜆 − 𝜆0 )𝑚 +
𝑎𝑖𝑗 𝜀𝑖 (𝜆 − 𝜆0 )𝑗
(2.29)
𝑖+𝑗≥2, 𝑖,𝑗∈ℕ
(𝑖,𝑗)∕∈{(0,𝑗):𝑗≤𝑚}
for ∣𝜀∣ + ∣𝜆 − 𝜆0 ∣ << 1, where 𝑎10 =
∂𝑓
∂𝜀
(0, 𝜆0 ) ∕= 0 and 𝑎0𝑚 =
1 ∂𝑚𝑓
𝑚! ∂𝜆𝑚
(0, 𝜆0 ) ∕= 0. Then
using the expansions of 𝑓 (𝜀, 𝜆) and 𝜀 (𝜆) together with the identity 𝑓 (𝜀 (𝜆) , 𝜆) = 0 for
∣𝜆 − 𝜆0 ∣ << 1, we find that 𝑞 = 𝑚 and
𝜀𝑚 = −
1 ∂ 𝑚 det(𝜆𝐼−𝐴(𝜀)) 𝑚!
∂𝜆𝑚
𝑎0𝑚
(𝜆,𝜀)=(𝜆0 ,0)
=− ∂
.
𝑎10
det
(𝜆𝐼
−
𝐴
(𝜀))
∂𝜀
(𝜆,𝜀)=(𝜆0 ,0)
(2.30)
Therefore we conclude that 𝜀 (𝜆) has a zero of order 𝑚 at 𝜆 = 𝜆0 , which proves (ii).
Next, we prove (ii)⇒(iii). Suppose (ii) is true. The first part of proving (iii) involves inverting
𝜀 (𝜆) near 𝜀 = 0 and 𝜆 = 𝜆0 . To do this we expand 𝜀 (𝜆) in a power series about 𝜆 = 𝜆0 and
find that 𝜀 (𝜆) = 𝑔(𝜆)𝑚 where
(
𝑔(𝜆) = (𝜆 − 𝜆0 ) 𝜀𝑚 +
∞
∑
)1/𝑚
𝜀𝑘 (𝜆 − 𝜆0 )𝑘−𝑚
𝑘=𝑚+1
and we are taking any fixed branch of the 𝑚th root that is analytic at 𝜀𝑚 . Notice that,
for 𝜆 in a small enough neighborhood of 𝜆0 , 𝑔 is an analytic function, 𝑔 (𝜆0 ) = 0, and
𝑑𝑔
(𝜆0 )
𝑑𝜆
1/𝑚
= 𝜀𝑚
∕= 0. This implies, by the inverse function theorem for analytic functions,
that for 𝜆 in a small enough neighborhood of 𝜆0 the analytic function 𝑔 (𝜆) has an analytic
31
inverse 𝑔 −1 (𝜀) in a neighborhood of 𝜀 = 0 with 𝑔 −1 (0) = 𝜆0 . Define a multivalued function
( 1)
1
𝜆 (𝜀), for sufficiently small 𝜀, by 𝜆 (𝜀) := 𝑔 −1 𝜀 𝑚 where by 𝜀 𝑚 we mean all branches of
the 𝑚th root of 𝜀. We know that 𝑔 −1 is analytic at 𝜀 = 0 so that for sufficiently small 𝜀 the
[ 𝑑𝑔
]−1
−1
(𝜆0 )
∕= 0 we have
multivalued function 𝜆 (𝜀) is a Puiseux series. And since 𝑑𝑔𝑑𝜀 (0) = 𝑑𝜆
an expansion
∞
( 1)
( 1 )𝑘
∑
1
𝜆 (𝜀) = 𝑔 −1 𝜀 𝑚 = 𝜆0 + 𝛼1 𝜀 𝑚 +
𝛼𝑘 𝜀 𝑚 .
𝑘=2
Now suppose for fixed 𝜆 sufficiently near 𝜆0 and for sufficiently small 𝜀 that det (𝜆𝐼 − 𝐴 (𝜀)) =
0. We want to show this implies 𝜆 = 𝜆 (𝜀) for one of the branches of the 𝑚th root. We know
by hypothesis we must have 𝜀 = 𝜀 (𝜆). But as we know this implies that 𝜀 = 𝜀 (𝜆) = 𝑔(𝜆)𝑚
hence for some branch of the 𝑚th root, 𝑏𝑚 (⋅ ), we have 𝑏𝑚 (𝜀) = 𝑏𝑚 (𝑔(𝜆)𝑚 ) = 𝑔(𝜆). But
𝜆 is near enough to 𝜆0 and 𝜀 is sufficiently small that we may apply the 𝑔 −1 to both sides
yielding 𝜆 = 𝑔 −1 (𝑔(𝜆)) = 𝑔 −1 (𝑏𝑚 (𝜀)) = 𝜆 (𝜀), as desired. Furthermore, all the 𝑚 branches
𝜆ℎ (𝜀), ℎ = 0, . . . , 𝑚 − 1 of 𝜆 (𝜀) are given by taking all branches of the 𝑚th root of 𝜀 so that
ℎ
𝜆ℎ (𝜀) = 𝜆0 + 𝛼1 𝜁 𝜀
1
𝑚
+
∞
∑
(
)𝑘
1
𝛼𝑘 𝜁 ℎ 𝜀 𝑚
𝑘=2
1
2𝜋
for any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 and
[
]−1
𝑑𝑔
𝑑𝑔 −1
−1/𝑚
(0) =
(𝜆0 )
= 𝜀𝑚
∕= 0,
𝛼1 =
𝑑𝜀
𝑑𝜆
(2.31)
which proves (iii).
Next, we prove (iii)⇒(iv). Suppose (iii) is true. Define the function 𝑦 (𝜀) := 𝜆0 (𝜀𝑚 ).
Then 𝑦 is analytic at 𝜀 = 0 and
𝑑𝑦
𝑑𝜀
(0) = 𝜆1 ∕= 0. Also we have for 𝜀 sufficiently small
det (𝑦 (𝜀) 𝐼 − 𝐴 (𝜀𝑚 )) = 0. Consider the inverse of 𝑦 (𝜀), 𝑦 −1 (𝜆). It has the property
(
(
(
(
𝑚 ))
𝑚 ))
that 0 = det 𝑦 (𝑦 −1 (𝜆)) 𝐼 − 𝐴 [𝑦 −1 (𝜆)]
= det 𝜆𝐼 − 𝐴 [𝑦 −1 (𝜆)]
with 𝑦 −1 (𝜆0 ) = 0,
𝑑𝑦 −1
𝑑𝜆
𝑚
−1
(𝜆0 ) = 𝛼−1
(𝜆)] . Then 𝑔 has a zero of order 𝑚 at 𝜆0 and
1 . Define 𝑔 (𝜆) := [𝑦
det (𝜆𝐼 − 𝐴 (𝑔 (𝜆))) = 0 for 𝜆 in a neighborhood of 𝜆0 .
32
Now we consider the analytic matrix 𝐴 (𝑔 (𝜆)) − 𝜆𝐼 in a neighborhood of 𝜆 = 𝜆0 with the
constant eigenvalue 0. Because 0 is an analytic eigenvalue of it then there exists an analytic
eigenvector, 𝑥 (𝜆), of 𝐴 (𝑔 (𝜆)) − 𝜆𝐼 corresponding to the eigenvalue 0 in a neighborhood of
𝜆0 such that 𝑥 (𝜆0 ) ∕= 0. Hence for 𝜆 near 𝜆0 we have
0 = (𝐴 (𝑔 (𝜆)) − 𝜆𝐼) 𝑥 (𝜆)
(
))
)
( (
(𝜆 − 𝜆0 )𝑚 + 𝑂 (𝜆 − 𝜆0 )𝑚+1 − (𝜆 − 𝜆0 ) 𝐼 − 𝜆0 𝐼 𝑥 (𝜆)
= 𝐴 𝛼−𝑚
1
)
(
𝑑𝑥
(𝜆0 ) − 𝑥 (𝜆0 ) (𝜆 − 𝜆0 ) + ⋅ ⋅ ⋅
= (𝐴 (0) − 𝜆0 𝐼) 𝑥 (𝜆0 ) + (𝐴 (0) − 𝜆0 𝐼)
𝑑𝜆
(
)
𝑑𝑚−1 𝑥
𝑑𝑚−2 𝑥
+ (𝐴 (0) − 𝜆0 𝐼) 𝑚−1 (𝜆0 ) − 𝑚−2 (𝜆0 ) (𝜆 − 𝜆0 )𝑚−1
𝑑𝜆
𝑑𝜆
(
)
𝑚
𝑚−1
𝑑 𝑥
𝑑
𝑥
−𝑚 ′
+ (𝐴 (0) − 𝜆0 𝐼) 𝑚 (𝜆0 ) − 𝑚−1 (𝜆0 ) + 𝛼1 𝐴 (0)𝑥 (𝜆0 ) (𝜆 − 𝜆0 )𝑚
𝑑𝜆
𝑑𝜆
(
)
+ 𝑂 (𝜆 − 𝜆0 )𝑚+1 .
This implies that
𝑑𝑗−1 𝑥
𝑑𝑗 𝑥
(𝐴 (0) − 𝜆0 𝐼) 𝑥 (𝜆0 ) = 0, (𝐴 (0) − 𝜆0 𝐼) 𝑗 (𝜆0 ) = 𝑗−1 (𝜆0 ) , for 𝑗 = 1, . . . , 𝑚 − 1,
𝑑𝜆
𝑑𝜆
𝑑𝑚 𝑥
𝑑𝑚−1 𝑥
′
(𝐴 (0) − 𝜆0 𝐼) 𝑚 (𝜆0 ) = 𝑚−1 (𝜆0 ) − 𝛼−𝑚
(2.32)
1 𝐴 (0)𝑥 (𝜆0 ) .
𝑑𝜆
𝑑𝜆
The first 𝑚 equations imply that 𝑥 (𝜆0 ),
𝑚 generated by
𝑑𝑚−1 𝑥
𝑑𝜆𝑚−1
𝑑𝑥
𝑑𝜆
(𝜆0 ), . . .,
𝑑𝑚−1 𝑥
𝑑𝜆𝑚−1
(𝜆0 ) is a Jordan chain of length
(𝜆0 ). Since the algebraic multiplicity of 𝜆0 for 𝐴 (0) is 𝑚 this implies
that the there is a single 𝑚 × 𝑚 Jordan block corresponding to the eigenvalue 𝜆0 where we
𝑚−1
𝑑𝑥
𝑑
𝑥
can take 𝑥 (𝜆0 ) , 𝑑𝜆
(𝜆0 ) , ..., 𝑑𝜆
𝑚−1 (𝜆0 ) as a Jordan basis. It follows from basic properties of
Jordan chains that there exists an eigenvector 𝑣 of 𝐴(0)∗ corresponding to the eigenvalue 𝜆∗0
𝑚−1
𝑑
𝑥
such that ⟨𝑣, 𝑑𝜆
𝑚−1 (𝜆0 )⟩ = 1. Hence
⟨
0=
⟩
𝑑𝑚 𝑥
(2.32)
′
(𝐴(0) − 𝜆0 𝐼) 𝑣, 𝑚 (𝜆0 )
= 1 − 𝛼−𝑚
1 ⟨𝑣, 𝐴 (0)𝑥 (𝜆0 )⟩
𝑑𝜆
∗
33
implying that ⟨𝑣, 𝑑𝐴
(0) 𝑥 (𝜆0 )⟩ = 𝛼𝑚
1 ∕= 0. Therefore we have shown that the Jordan normal
𝑑𝜀
form of 𝐴 (0) corresponding to the eigenvalue 𝜆0 consists of a single 𝑚 × 𝑚 Jordan block and
there exists an eigenvector 𝑢 of 𝐴 (0) corresponding to the eigenvalue 𝜆0 and an eigenvector
∕ 0. This proves (iv).
𝑣 of 𝐴 (0)∗ corresponding to the eigenvalue 𝜆∗0 such that ⟨𝑣, 𝐴′ (0)𝑢⟩ =
Finally, we show (iv)⇒(i). Suppose (iv) is true. We begin by noting that since
det (𝜆0 𝐼 − 𝐴 (𝜀)) = (−1)𝑛 det ((𝐴 (0) − 𝜆0 𝐼) + 𝐴′ (0)𝜀) + 𝑜 (𝜀)
it suffices to show that
𝑆𝑛−1 :=
𝑑
det ((𝐴 (0) − 𝜆0 𝐼) + 𝐴′ (0)𝜀) 𝜀=0 ∕= 0.
𝑑𝜀
(2.33)
We will use the result from [24, Theorem 2.16] to prove (2.33). Let 𝐴 (0) − 𝜆0 𝐼 = 𝑌 Σ𝑋 ∗ be
a singular-value decomposition of the matrix 𝐴 (0) − 𝜆0 𝐼 where 𝑋, 𝑌 are unitary matrices
and Σ = diag(𝜎 1 , . . . , 𝜎 𝑛−1 , 𝜎 𝑛 ) with 𝜎 1 ≥ . . . ≥ 𝜎 𝑛−1 ≥ 𝜎 𝑛 ≥ 0 (see [35, §5.7, Theorem
2]). Now since the Jordan normal form of 𝐴 (0) corresponding to the eigenvalue 𝜆0 consists
of a single Jordan block this implies that rank of 𝐴 (0) − 𝜆0 𝐼 is 𝑛 − 1. This implies that
𝜎 1 ≥ . . . ≥ 𝜎 𝑛−1 > 𝜎 𝑛 = 0, 𝑢 = 𝑋𝑒𝑛 is an eigenvalue of 𝐴 (0) corresponding to the eigenvalue
𝜆0 , 𝑣 = 𝑌 𝑒𝑛 is an eigenvalue of 𝐴 (0) corresponding to the eigenvalue 𝜆∗0 , and there exist
nonzero constants 𝑐1 , 𝑐2 such that 𝑢 = 𝑐1 𝑢0 and 𝑣 = 𝑐2 𝑣0 .
Now using the result of [24, Theorem 2.16] for (2.33) we find that
∗
𝑆𝑛−1 = det (𝑌 𝑋 )
∑
(
)
∗ ′
𝜎 𝑖1 ⋅ ⋅ ⋅ 𝜎 𝑖𝑛−1 det (𝑌 𝐴 (0)𝑋)𝑖1 ...𝑖𝑛−1 ,
1≤𝑖1 <⋅⋅⋅<𝑖𝑛−1 ≤𝑛
where (𝑌 ∗ 𝐴′ (0)𝑋)𝑖1 ...𝑖𝑛−1 is the matrix obtained from 𝑌 ∗ 𝐴′ (0)𝑋 by removing rows and
34
columns 𝑖1 . . . 𝑖𝑛−1 . But since 𝜎 𝑛 = 0 and
(𝑌 ∗ 𝐴′ (0)𝑋)1...(𝑛−1) = 𝑒∗𝑛 𝑌 ∗ 𝐴′ (0)𝑋𝑒𝑛 = (𝑣, 𝐴′ (0)𝑢) = 𝑐∗2 𝑐1 (𝑣0 , 𝐴′ (0)𝑢0 ) ∕= 0
then 𝑆𝑛−1 = det (𝑌 𝑋 ∗ )
𝑛−1
∏
𝜎 𝑗 𝑐∗2 𝑐1 (𝑣0 , 𝐴′ (0)𝑢0 ) ∕= 0. This completes the proof.
𝑗=1
Proof of Theorem 2
We begin by noting that our hypotheses imply that statements (ii), (iii), and (iv) of Theorem
1 are true. In particular, statement (iii) implies that there is exactly one convergent Puiseux
series for the 𝜆0 -group whose branches are given by
ℎ
𝜆ℎ (𝜀) = 𝜆0 + 𝛼1 𝜁 𝜀
1
𝑚
+
∞
∑
(
)𝑘
1
𝛼𝑘 𝜁 ℎ 𝜀 𝑚 ,
𝑘=2
1
2𝜋
for ℎ = 0, . . . , 𝑚 − 1 and any fixed branch of 𝜀 𝑚 , where 𝜁 = 𝑒 𝑚 𝑖 and 𝛼1 ∕= 0. Then by well
known results [4, §6.1.7, Theorem 2], [28, §II.1.8] there exists a convergent Puiseux series for
the corresponding eigenvectors whose branches are given by
𝑥ℎ (𝜀) = 𝛽 0 +
∞
∑
(
)𝑘
1
𝛽 𝑘 𝜁 ℎ𝜀 𝑚 ,
𝑘=1
for ℎ = 0, . . . , 𝑚−1, where 𝛽 0 is an eigenvector of 𝐴0 = 𝐴 (0) corresponding to the eigenvalue
𝜆0 . Now if we examine the proof of (ii)⇒(iii) in Theorem 1 we see by equation (2.31) that
−1
𝛼𝑚
1 = 𝜀𝑚 , where 𝜀𝑚 is given in equation (2.30) in the proof of (i)⇒(iii) for Theorem 1. Thus
we can conclude that
∂
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 )
∂𝜀
𝛼𝑚
=
−
( ∂𝑚
) ∕= 0.
1
∂𝜆𝑚
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
𝑚!
35
(2.34)
Choose any 𝑚th root of ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩ and denote it by ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩1/𝑚 . By (2.34) we can just
reindexing the Puiseux series (2.19) and (2.20) and assume that
⎞1/𝑚
⎛
∂
det (𝜆𝐼 − 𝐴 (𝜀)) ∣(𝜀,𝜆)=(0,𝜆0 ) ⎟
⎜
𝛼1 = ⎝− ∂𝜀( ∂ 𝑚
) ⎠
∂𝜆𝑚
.
det(𝜆𝐼−𝐴(𝜀))∣(𝜀,𝜆)=(0,𝜆0 )
𝑚!
Next, we wish to prove that we can choose the perturbed eigenvectors (2.20) to satisfy the
normalization conditions (2.22). But this follows by Theorem 1 (iv) and the fact 𝛽 0 is an
eigenvector of 𝐴(0) corresponding to the eigenvalue 𝜆0 since then ⟨𝑣1 , 𝛽 0 ⟩ =
∕ 0 and so we may
take
𝑥ℎ (𝜀)
,
(𝑣1 ,𝑥ℎ (𝜀))
for ℎ = 0, . . . , 𝑚 − 1, to be the perturbed eigenvectors in (2.20) that satisfy
the normalization conditions (2.22).
∞
Now we are ready to begin showing that {𝛼𝑠 }∞
𝑠=1 , {𝛽 𝑠 }𝑠=0 are given by the recursive formulas
(2.23)-(2.25). The first key step is proving the following:
(𝐴0 − 𝜆0 𝐼) 𝛽 𝑠 = −
𝑠 (
∑
𝐴 𝑘 − 𝛼𝑘 𝐼 𝛽 𝑠−𝑘 , for 𝑠 ≥ 1,
(2.35)
𝛽 0 = 𝑢1 , 𝛽 𝑠 = Λ (𝐴0 − 𝜆0 𝐼) 𝛽 𝑠 , for 𝑠 ≥ 1,
(2.36)
𝑘=1
where we define 𝐴 𝑘 := 0, if
𝑚
𝑘
𝑚
)
𝑚
∕∈ ℕ.
The first equality holds since in a neighborhood of the origin
0 = (𝐴 (𝜀) − 𝜆0 (𝜀) 𝐼) 𝑥0 (𝜀) =
∞
∑
𝑠=0
(
𝑠 (
∑
𝑘=0
)
)
𝑠
𝐴 𝑘 − 𝛼𝑘 𝐼 𝛽 𝑠−𝑘 𝜀 𝑚 .
𝑚
The second equality will be proven once we show 𝛽 0 = 𝑢1 and 𝛽 𝑠 ∈ 𝑆 := span{𝑈 𝑒𝑖 ∣2 ≤
𝑖 ≤ 𝑛}, for 𝑠 ≥ 1, where 𝑈 is the matrix from (2.12). This will prove (2.36) because
Λ (𝐴0 − 𝜆0 𝐼) acts as the identity on 𝑆 by Proposition 6.i. But these follow from the facts that
36
𝑆 = {𝑥 ∈ ℂ𝑛 ∣⟨𝑣1 , 𝑥⟩ = 0} and the normalization conditions (2.22) imply that ⟨𝑣1 , 𝛽 0 ⟩ = 1
and ⟨𝑣1 , 𝛽 𝑠 ⟩ = 0, for 𝑠 ≥ 1.
The next key step in this proof is the following lemma:
Lemma 5 For all 𝑠 ≥ 0 the following identity holds
(𝐴0 − 𝜆0 𝐼) 𝛽 𝑠 =
⎧



⎨
𝑠
∑
𝑝𝑠,𝑖 𝑢𝑖 , for 0 ≤ 𝑠 ≤ 𝑚 − 1
𝑖=0
𝑗 ⌊∑
𝑚
𝑠−𝑚
𝑚 ⌋
∑
∑ ∑



𝑝𝑠,𝑖 𝑢𝑖 −
𝑝𝑗,𝑘 Λ𝑘 𝐴𝑙 𝛽 𝑠−𝑗−𝑙𝑚 , for 𝑠 ≥ 𝑚
⎩
𝑠−𝑗
𝑖=0
𝑗=0 𝑘=0
(2.37)
𝑙=1
where we define 𝑢0 := 0.
Proof.
The proof is by induction on 𝑠. The statement is true for 𝑠 = 0 since 𝑝0,0 𝑢0 =
0 = (𝐴0 − 𝜆0 𝐼) 𝛽 0 . Now suppose it was true for all 𝑟 with 0 ≤ 𝑟 ≤ 𝑠 for some nonnegative
integer 𝑠. We will show the statement is true for 𝑠 + 1 as well.
∑
Suppose 𝑠 + 1 ≤ 𝑚 − 1 then (𝐴0 − 𝜆0 𝐼) 𝛽 𝑟 = 𝑟𝑖=0 𝑝𝑟,𝑖 𝑢𝑖 for 0 ≤ 𝑟 ≤ 𝑠 and we must show
∑
that (𝐴0 − 𝜆0 𝐼) 𝛽 𝑠+1 = 𝑠+1
𝑖=0 𝑝𝑠+1,𝑖 𝑢𝑖 . Well, for 1 ≤ 𝑟 ≤ 𝑠,
(2.36)
𝛽 𝑟 = Λ (𝐴0 − 𝜆0 𝐼) 𝛽 𝑟 =
𝑟
∑
(2.40)
𝑝𝑟,𝑖 Λ𝑢𝑖 =
𝑖=0
𝑟
∑
𝑝𝑟,𝑖 𝑢𝑖+1 .
(2.38)
𝑖=1
Hence the statement is true if 𝑠 + 1 ≤ 𝑚 − 1 since
)
𝑠+1
∑(
∑ 𝑠+1−𝑘
∑
(2.35)
(2.38) 𝑠+1
(𝐴0 − 𝜆0 𝐼) 𝛽 𝑠+1 = −
𝐴 𝑘 − 𝛼𝑘 𝐼 𝛽 𝑠+1−𝑘 =
𝛼𝑘 𝑝𝑠+1−𝑘,𝑖 𝑢𝑖+1
𝑚
𝑘=1
𝑘=1 𝑖=0
(𝑠+1−𝑖
)
𝑠
∑
∑
(2.46) ∑
(2.43) 𝑠+1
=
𝛼𝑘 𝑝𝑠+1−𝑘,𝑖 𝑢𝑖+1 =
𝑝𝑠+1,𝑖 𝑢𝑖 .
𝑖=0
𝑘=1
𝑖=0
Now suppose that 𝑠 + 1 ≥ 𝑚. The proof is similar to what we just proved. By the induction
37
(2.36)
hypothesis (2.37) is true for 1 ≤ 𝑟 ≤ 𝑠 and 𝛽 𝑟 = Λ (𝐴0 − 𝜆0 𝐼) 𝛽 𝑟 thus
(2.40)
𝛽𝑟 =
⎧



⎨
𝑟
∑
𝑝𝑟,𝑖 𝑢𝑖+1 , for 0 ≤ 𝑟 ≤ 𝑚 − 1
𝑖=0
(2.39)
𝑗 ⌊∑
𝑚−1
𝑟−𝑚
𝑚 ⌋
∑
∑ ∑



𝑝
𝑢
−
𝑝𝑗,𝑘 Λ𝑘+1 𝐴𝑙 𝛽 𝑟−𝑗−𝑙𝑚 , for 𝑟 ≥ 𝑚.
𝑟,𝑖
𝑖+1
⎩
𝑟−𝑗
𝑗=0 𝑘=0
𝑖=0
𝑙=1
Hence we have
(2.35)
(𝐴0 − 𝜆0 𝐼) 𝛽 𝑠+1 = −
𝑠+1
∑(
)
𝐴 𝑘 − 𝛼𝑘 𝐼 𝛽 𝑠+1−𝑘
𝑚
𝑘=1
(2.39)
= −
⌊ 𝑠+1
𝑚 ⌋
∑
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑠+1−𝑚
∑ 𝑚−1
∑
𝑙=1
−
𝑘=1
𝑗
𝑠+1−𝑚
∑ 𝑠+1−𝑘−𝑚
∑ ∑
𝑗=0
𝑘=1
𝑠+1
∑
𝑠+1−𝑘
∑
𝑘>𝑠+1−𝑚
𝑖=0
+
(2.46)
= −
⌊ 𝑠+1
𝑚 ⌋
∑
⌋
⌊ 𝑠+1−𝑘−𝑗
𝑚
∑
𝑖=0
𝛼𝑘 𝑝𝑠+1−𝑘,𝑖 𝑢𝑖+1
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑗=0
⌊∑⌋
⌊ 𝑠+1−𝑘−𝑗
⌋
𝑚
∑
𝑖=0
= −
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑗
𝑠+1−𝑚
∑ 𝑠+1−𝑘−𝑚
∑ ∑
𝑘=1
Now let 𝑎𝑘,𝑗,𝑖 :=
⌊ 𝑠+1−𝑘−𝑗
⌋
𝑚
∑
)
𝛼𝑘 𝑝𝑠+1−𝑘,𝑖 𝑢𝑖+1
𝑘=1
𝛼𝑘 𝑝𝑗,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑘−𝑗−𝑙𝑚
𝑚
∑
𝑝𝑠+1,𝑖 𝑢𝑖
𝑖=0
𝑙=1
−
(𝑠+1−𝑖
∑
𝑙=1
𝑠+1
𝑚
(2.43)
𝑚−1
∑
𝑖=0
𝑗
𝑠+1−𝑚
∑ 𝑠+1−𝑘−𝑚
∑ ∑
𝑘=1
𝛼𝑘 𝑝𝑗,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑘−𝑗−𝑙𝑚
𝑙=1
𝑙=1
−
𝛼𝑘 𝑝𝑠+1−𝑘,𝑖 𝑢𝑖+1
𝑖=0
𝑗=0
⌋
⌊ 𝑠+1−𝑘−𝑗
𝑚
∑
𝑖=0
𝛼𝑘 𝑝𝑗,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑘−𝑗−𝑙𝑚 .
𝑙=1
𝛼𝑘 𝑝𝑗,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑘−𝑗−𝑙𝑚 . Then using the sum identity
𝑙=1
𝑗
𝑠+1−𝑚
∑ 𝑠+1−𝑘−𝑚
∑ ∑
𝑘=1
𝑗=0
𝑖=0
∑ 𝑠−𝑚
∑ 𝑠+1−𝑗−𝑚
∑
(2.48) 𝑠−𝑚
=
𝑖=0 𝑗=𝑖
𝑗
∑ 𝑠+1−𝑗−𝑚
∑ ∑
(2.46) 𝑠−𝑚
𝑎𝑘,𝑗,𝑖 =
𝑘=1
𝑗=0
𝑘=1
𝑎𝑘,𝑗,𝑖 =
𝑖=0
∑ 𝑠+1−𝑚
∑ 𝑞−𝑖
∑
(2.49) 𝑠−𝑚
𝑎𝑘,𝑗,𝑖 =
𝑖=0 𝑞=𝑖+1 𝑘=1
38
𝑗 𝑠+1−𝑗−𝑚
𝑠−𝑚
∑ ∑
∑
𝑗=0 𝑖=0
𝑎𝑘,𝑞−𝑘,𝑖
𝑘=1
𝑎𝑘,𝑗,𝑖
we can concluded that
(𝐴0 − 𝜆0 𝐼) 𝛽 𝑠+1 = −
⌊ 𝑠+1
𝑚 ⌋
∑
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑠−𝑚
∑ 𝑠+1−𝑚
∑ 𝑞−𝑖
∑
⌊ 𝑠+1−𝑞
⌋
𝑚
∑
𝑖=0 𝑞=𝑖+1 𝑘=1
𝑙=1
⌊∑⌋
𝑠+1
𝑚
=−
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝛼𝑘 𝑝𝑞−𝑘,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑞−𝑙𝑚
𝑚
∑
𝑠−𝑚
∑ 𝑠+1−𝑚
∑ ⌊
𝑖=0 𝑞=𝑖+1
⌊∑⌋
(
∑
∑ ⌋ 𝑞−𝑖
𝑠+1−𝑞
𝑚
= −
𝑘=1
𝑙=1
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑠−𝑚
∑ 𝑠+1−𝑚
∑ ⌊
𝑖=0 𝑞=𝑖+1
⌊∑⌋
∑ ⌋
𝑠+1−𝑞
𝑚
= −
𝑙=1
𝐴𝑙 𝛽 𝑠+1−𝑙𝑚 +
𝑠+1−𝑚
∑⌊
∑ 𝑞−1
𝑞=1
𝑚
(2.16) ∑
=
𝑖=0
𝑚
∑
𝑝𝑠+1,𝑖 𝑢𝑖
𝑖=0
𝑙=1
−
𝑝𝑠+1,𝑖 𝑢𝑖
𝑝𝑞,𝑖+1 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑞−𝑙𝑚
𝑠+1
𝑚
(2.47)
𝑚
∑
𝑖=0
𝑙=1
−
)
𝛼𝑘 𝑝𝑞−𝑘,𝑖 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑞−𝑙𝑚
𝑠+1
𝑚
(2.43)
𝑝𝑠+1,𝑖 𝑢𝑖
𝑖=0
𝑙=1
−
𝑝𝑠+1,𝑖 𝑢𝑖
𝑖=0
𝑙=1
−
𝑚
∑
∑ ⌋
𝑠+1−𝑞
𝑚
𝑝𝑞,𝑖+1 Λ𝑖+1 𝐴𝑙 𝛽 𝑠+1−𝑞−𝑙𝑚
𝑙=1
𝑝𝑠+1,𝑖 𝑢𝑖 −
𝑖=0
𝑗
𝑠+1−𝑚
∑ ∑
𝑗=0
𝑘=0
⌊ 𝑠+1−𝑗
𝑚 ⌋
∑
𝑝𝑗,𝑘 Λ𝑘 𝐴𝑙 𝛽 𝑠+1−𝑗−𝑙𝑚 .
𝑙=1
But this is the statement we needed to prove for 𝑠 + 1 ≥ 𝑚. Therefore by induction the
statement (2.37) is true for all 𝑠 ≥ 0 and the lemma is proved.
The lemma above is the key to prove the recursive formulas for 𝛼𝑠 and 𝛽 𝑠 as given by
(2.23)-(2.25). First we prove that 𝛽 𝑠 is given by (2.25). For 𝑠 = 0 we have already shown
𝛽 0 = 𝑢1 = 𝑝0,0 𝑢1 . So suppose 𝑠 ≥ 1. Then by (2.36) and (2.37) we find that
(2.40)
𝛽𝑠 =
⎧



⎨
𝑠
∑
𝑝𝑠,𝑖 𝑢𝑖+1 , if 0 ≤ 𝑠 ≤ 𝑚 − 1
𝑖=0
𝑗 ⌊∑
𝑚−1
𝑠−𝑚
𝑚 ⌋
∑
∑ ∑



𝑝𝑠,𝑖 𝑢𝑖+1 −
𝑝𝑗,𝑘 Λ𝑘+1 𝐴𝑙 𝛽 𝑠−𝑗−𝑙𝑚 , if 𝑠 ≥ 𝑚.
⎩
𝑠−𝑗
𝑖=0
𝑗=0 𝑘=0
𝑙=1
This proves that 𝛽 𝑠 is given by (2.25).
39
Next we will prove that 𝛼𝑠 is given by (2.23) and (2.24). We start with 𝑠 = 1 and prove 𝛼1
is given by (2.23). First, (𝐴0 − 𝜆0 𝐼)∗ 𝑣𝑚 = 0 and ⟨𝑣𝑚 , 𝑢𝑖 ⟩ = 𝛿 𝑚,𝑖 hence
⟨
⟩
𝑚
∑
(2.42) 𝑚
0 = ⟨𝑣𝑚 , (𝐴0 − 𝜆0 𝐼)𝛽 𝑚 ⟩ =
𝑣𝑚 , 𝑝𝑚,𝑖 𝑢𝑖 − 𝐴1 𝑢1
= 𝛼1 − ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩
(2.37)
𝑖=0
so that 𝛼𝑚
1 = ⟨𝑣𝑚 , 𝐴1 𝑢1 ⟩. This and identity (2.34) imply that formula (2.23) is true.
Finally, suppose that 𝑠 ≥ 2. Then (𝐴0 − 𝜆0 𝐼)∗ 𝑣𝑚 = 0 and ⟨𝑣𝑚 , 𝑢𝑖 ⟩ = 𝛿 𝑚,𝑖 implies
0 = ⟨𝑣𝑚 , (𝐴0 − 𝜆0 𝐼)𝛽 𝑚+𝑠−1 ⟩
⟨
⟩
𝑚+𝑠−1−𝑗
⌋
𝑗 ⌊
𝑚
𝑠−1
𝑚
∑
∑∑
∑
(2.37)
=
𝑣𝑚 , 𝑝𝑚+𝑠−1,𝑖 𝑢𝑖 −
𝑝𝑗,𝑘 Λ𝑘 𝐴𝑙 𝛽 𝑚+𝑠−1−𝑗−𝑙𝑚
𝑖=0
(2.48)
= 𝑝𝑚+𝑠−1,𝑚 −
𝑗=0 𝑘=0
𝑠−1
∑ 𝑠−1
∑
𝑙=1
⟨
𝑝𝑗,𝑘
(Λ∗ )𝑘 𝑣𝑚 ,
𝑘=0 𝑗=𝑘
(2.41)
= 𝑟𝑠−1 + 𝑚𝛼𝑚−1
𝛼𝑠 −
1
⌊ 𝑚+𝑠−1−𝑗
⌋
𝑚
∑
⟩
𝐴𝑙 𝛽 𝑚+𝑠−1−𝑗−𝑙𝑚
𝑙=1
𝑠−1
∑ 𝑠−1
∑
⟨
𝑝𝑗,𝑖
(Λ∗ )𝑖 𝑣𝑚 ,
⌊ 𝑚+𝑠−1−𝑗
⌋
𝑚
∑
𝑖=0 𝑗=𝑖
⟩
𝐴𝑘 𝛽 𝑚+𝑠−1−𝑗−𝑘𝑚
𝑘=1
Therefore with this equality, the fact 𝛼1 ∕= 0, and Proposition 6.iii, we can solve for 𝛼𝑠 and
we will find that it is given by (2.24). This completes the proof.
Proof of Corollaries 3 and 4
Both corollaries follow almost trivially now. To prove Corollary 3, we just examine the recursive formulas (2.23)-(2.25) in Theorem 2 to see that 𝛼𝑘 , 𝛽 𝑘 requires only 𝐴0 , . . . , 𝐴⌊ 𝑚+𝑘−1 ⌋ .
𝑚
To prove Corollary 4, we use Proposition 7 to show that
𝑝0,0 = 1, 𝑝1,0 = 𝑝2,0 = 0, 𝑝1,1 = 𝛼1 , 𝑝2,1 = 𝛼2 , 𝑝2,2 = 𝛼21
40
and then from this and (2.23)-(2.25) we get the desired result for 𝛼1 , 𝛼2 , 𝛽 0 , 𝛽 1 , 𝛽 2 in terms
of 𝐴0 , 𝐴1 , 𝐴2 . The last part to prove is the formula for 𝛼2 in terms of 𝑓 (𝜀, 𝜆) and its partial
derivatives. But the formula follows from the series representation of 𝑓 (𝜀, 𝜆) in (2.29) and
𝜆0 (𝜀) in (2.19) since, for 𝜀 in a neighborhood of the origin,
0 = 𝑓 (𝜀, 𝜆0 (𝜀))
= (𝑎10 + 𝑎0𝑚 𝑝𝑚.𝑚 ) 𝜀
⎧

⎨
(𝑎01 𝑝2,1 + 𝑎02 𝑝2,2 + 𝑎11 𝑝1,1 + 𝑎20 𝑝00 ) 𝜀2 , for 𝑚 = 1
+

⎩ (𝑎0𝑚 𝑝𝑚+1,𝑚 + 𝑎0𝑚+1 𝑝𝑚+1,𝑚+1 + 𝑎11 𝑝1,1 ) 𝜀 𝑚+1
𝑚 , for 𝑚 > 1
( 𝑚+2 )
+𝑂 𝜀 𝑚
which together with Proposition 7 implies the formula for 𝛼2 .
2.5
Auxiliary Results
Properties of the Matrix Λ
The fundamental properties of the matrix Λ defined in (2.15) which are needed in this paper
are given in the following proposition:
Proposition 6
(i) We have Λ (𝐴0 − 𝜆0 𝐼) 𝑈 𝑒1 = 0, Λ (𝐴0 − 𝜆0 𝐼) 𝑈 𝑒𝑖 = 𝑈 𝑒𝑖 , for 2 ≤ 𝑖 ≤
𝑛,
(ii) For 1 ≤ 𝑖 ≤ 𝑚 − 1 we have
Λ𝑢𝑚 = 0, Λ𝑢𝑖 = 𝑢𝑖+1
(iii) Λ∗ 𝑣1 = 0, and Λ∗ 𝑣𝑖 = 𝑣𝑖−1 , for 2 ≤ 𝑖 ≤ 𝑚.
41
(2.40)
Proof. i. Using the fact 𝐽𝑚 (0)∗ 𝐽𝑚 (0) = diag[0, 𝐼𝑚−1 ], (2.12), and (2.15) we find by block
multiplication that 𝑈 −1 Λ (𝐴0 − 𝜆0 𝐼) 𝑈 = diag[0, 𝐼𝑛−1 ]. This implies the result.
ii. & iii. The results follow from the definition of 𝑢𝑖 , 𝑣𝑖 in (2.13), (2.14) and the fact
⎡
⎤
)∗ ⎢ 𝐽𝑚 (0)
(
⎥
(𝑈 −1 Λ𝑈 )∗ = 𝑈 ∗ Λ∗ 𝑈 −1 = ⎣
⎦.
[
−1 ]∗
(𝑊0 − 𝜆0 𝐼𝑛−𝑚 )
Properties of the Polynomials in the Recursive Formulas
This appendix contains two propositions. The first proposition gives fundamental identities
that help to characterize the polynomials {𝑝𝑗,𝑖 }∞
𝑗=𝑖 and {𝑟𝑙 }𝑙∈ℕ in (2.16) and (2.17). The
second proposition gives explicit recursive formulas to calculate these polynomials.
We may assume
∑∞
𝑗=1
𝛼𝑗 𝑧 𝑗 is a convergent Taylor series and 𝛼1 ∕= 0.
Proposition 7 The polynomials {𝑝𝑗,𝑖 }∞
𝑗=𝑖 and {𝑟𝑙 }𝑙∈ℕ have the following properties:
(i)
∞
∑
𝑗=𝑖
(
𝑝𝑗,𝑖 𝑧 𝑗 =
∞
∑
)𝑖
𝛼𝑗 𝑧 𝑗
, for 𝑗 ≥ 𝑖 ≥ 0.
𝑗=1
(ii) For 𝑙 ≥ 1 we have
𝑟𝑙 = 𝑝𝑚+𝑙,𝑚 − 𝑚𝛼𝑚−1
𝛼𝑙+1 .
1
(2.41)
𝑝𝑗,𝑗 = 𝛼𝑗1 .
(2.42)
(iii) 𝑝𝑗,1 = 𝛼𝑗 , for 𝑗 ≥ 1.
(iv) For 𝑗 ≥ 0 we have
42
(v) 𝑝𝑗+1,𝑗 = 𝑗𝛼𝑗−1
1 𝛼2 , for 𝑗 > 0.
(vi) For 𝑗 ≥ 𝑖 > 0 we have
𝑗−𝑖+1
∑
𝛼𝑞 𝑝𝑗−𝑞,𝑖−1 = 𝑝𝑗,𝑖 .
(2.43)
𝑞=1
(
Proof. i. For 𝑖 ≥ 0,
∞
∑
)𝑖
𝛼𝑗 𝑧
𝑗
∞
∑
=
⋅⋅⋅
𝑠1 =1
𝑗=1
∞
∑
(
𝑠𝑖 =1
)
𝑖
∏
𝛼𝑠𝜚 𝑧 𝑠1 +⋅⋅⋅+𝑠𝑖 =
𝜚=1
∞
∑
𝑝𝑗,𝑖 𝑧 𝑗 .
𝑗=𝑖
ii. Let 𝑙 ≥ 1. Then by definition of 𝑝𝑚+𝑙,𝑚 we have
𝑚
∏
∑
𝑝𝑚+𝑙,𝑚 =
𝛼 𝑠𝜚 +
𝜚=1
𝑠1 +⋅⋅⋅+𝑠𝑚 =𝑚+𝑙
1≤𝑠𝜚 ≤𝑙+1
∃𝜚∈{1,...,𝑚} such that 𝑠𝜚 =𝑙+1
𝑚
∏
∑
𝛼 𝑠𝜚
𝜚=1
𝑠1 +⋅⋅⋅+𝑠𝑚 =𝑚+𝑙
1≤𝑠𝜚 ≤𝑙+1
∃/𝜚∈{1,...,𝑚} such that 𝑠𝜚 =𝑙+1
= 𝑚𝛼𝑚−1
𝛼𝑙+1 + 𝑟𝑙 .
1
iii. For 𝑗 ≥ 1 we have 𝑝𝑗,1 =
∑
𝑠1 =𝑗
1≤𝑠𝜚 ≤𝑗
iv. For 𝑗 ≥ 0, 𝑝0,0 = 1 and 𝑝𝑗,𝑗 =
v. For 𝑗 > 0, 𝑝𝑗+1,𝑗 =
∑
∏1
𝜚=1
𝛼 𝑠𝜚 = 𝛼 𝑗 .
∑
𝑠1 +⋅⋅⋅+𝑠𝑗 =𝑗
1≤𝑠𝜚 ≤1
𝑠1 +⋅⋅⋅+𝑠𝑗 =𝑗+1
1≤𝑠𝜚 ≤2
∏𝑗
𝜚=1
∏𝑗
𝛼 𝑠𝜚 =
𝜚=1
∑𝑗
𝛼 𝑠𝜚 =
𝜚=1
∏𝑗
𝜚=1
𝛼1 = 𝛼𝑗1 .
𝑗−1
𝛼𝑗−1
1 𝛼2 = 𝑗𝛼1 𝛼2 .
vi. It follows by
∞
∑
𝑗=𝑖
𝑗 (i)
𝑝𝑗,𝑖 𝑧 =
∞
∑
𝑗=𝑖−1
𝑝𝑗,𝑖−1 𝑧 𝑗
∞
∑
𝑝𝑗,1 𝑧 𝑗 =
𝑗=1
∞
∑
𝑗=𝑖
(𝑗−𝑖+1
∑
)
𝑝𝑗−𝑞,𝑖−1 𝑝𝑞,1
𝑧𝑗 .
𝑞=1
This next proposition gives explicit recursive formulas to calculate the polynomials {𝑝𝑗,𝑖 }∞
𝑗=𝑖
and {𝑟𝑙 }𝑙∈ℕ .
Proposition 8 For each 𝑖 ≥ 0, the sequence of polynomials, {𝑝𝑗,𝑖 }∞
𝑗=𝑖 , is given by the recur43
sive formula
𝑝𝑖,𝑖 =
𝛼𝑖1 ,
𝑝𝑗,𝑖
𝑗−1
∑
1
=
[(𝑗 + 1 − 𝑘)𝑖 − 𝑘]𝛼𝑗+1−𝑘 𝑝𝑘,𝑖 , for 𝑗 > 𝑖.
(𝑗 − 𝑖)𝛼1 𝑘=𝑖
(2.44)
Furthermore, the polynomials {𝑟𝑙 }𝑙∈ℕ are given by the recursive formula:
𝑙−1
1 ∑
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝑟𝑗 +
𝑟1 = 0, 𝑟𝑙 =
𝑙𝛼1 𝑗=1
(2.45)
𝑙−1
𝑚 𝑚−2 ∑
𝛼
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝛼𝑗+1 , for 𝑙 > 1.
𝑙 1 𝑗=1
Proof.
We begin by showing (2.44) is true. For 𝑖 = 0, (2.44) follows from the definition
of the 𝑝𝑗,0 . If 𝑖 > 0 then by (2.42) and [20, (1.1) & (3.2)] it follows that (2.44) is true. Lets
now prove (2.45). From (2.41) and (2.44), we have
𝑟𝑙 = 𝑝𝑚+𝑙,𝑚 − 𝑚𝛼𝑚−1
𝛼𝑙+1
1
𝑚+𝑙−1
1 ∑
=
[(𝑚 + 𝑙 + 1 − 𝑘)𝑚 − 𝑘]𝛼𝑚+𝑙+1−𝑘 𝑝𝑘,𝑚 − 𝑚𝛼𝑚−1
𝛼𝑙+1
1
𝑙𝛼1 𝑘=𝑚
𝑚+𝑙−1
1 ∑
[(𝑚 + 𝑙 + 1 − 𝑘)𝑚 − 𝑘]𝛼𝑚+𝑙+1−𝑘 𝑝𝑘,𝑚
=
𝑙𝛼1 𝑘=𝑚+1
(2.42)
𝑙−1
1 ∑
=
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝑝𝑚+𝑗,𝑚
𝑙𝛼1 𝑗=1
𝑙−1
(
)
1 ∑
=
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝑟𝑗 + 𝑚𝛼𝑚−1
𝛼𝑗+1
1
𝑙𝛼1 𝑗=1
(2.41)
𝑙−1
1 ∑
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝑟𝑗 +
=
𝑙𝛼1 𝑗=1
𝑙−1
𝑚 𝑚−2 ∑
𝛼
[(𝑙 + 1 − 𝑗)𝑚 − (𝑚 + 𝑗)]𝛼𝑙+1−𝑗 𝛼𝑗+1 ,
𝑙 1 𝑗=1
for 𝑙 > 1. This completes the proof.
44
Sum Identities
These double sum identities are used in the proof of Theorem 3.1
𝑑 ∑
𝑑−𝑥
∑
𝑎𝑥,𝑦 =
𝑥=𝑐 𝑦=0
𝑑−1 ∑
𝑑
∑
𝑑−𝑦
𝑑−𝑐 ∑
∑
𝑎𝑥,𝑦 ,
(2.46)
𝑎𝑥,𝑦 ,
(2.47)
𝑎𝑥,𝑦 ,
(2.48)
𝑎𝑥,𝑞−𝑥 .
(2.49)
𝑦=0 𝑥=𝑐
𝑎𝑥,𝑦 =
𝑦−1
𝑑 ∑
∑
𝑥=0 𝑦=𝑥+1
𝑦=1 𝑥=0
𝑑 ∑
𝑥
∑
𝑑 ∑
𝑑
∑
𝑎𝑥,𝑦 =
𝑥=0 𝑦=0
𝑑−𝑦
𝑑−1 ∑
∑
𝑦=𝑐 𝑥=1
𝑎𝑥,𝑦 =
𝑦=0 𝑥=𝑦
𝑞−𝑐
𝑑
∑
∑
𝑞=𝑐+1 𝑥=1
45
Chapter 3
Spectral Perturbation Theory for
Holomorphic Matrix Functions
In this section we consider holomorphic matrix-valued functions of one and two variables.
A holomorphic matrix-valued function of one variable will be referred to as a holomorphic
matrix function. Our goal is to give some basic facts on the spectral theory and spectral
perturbation theory for holomorphic matrix functions that will be needed in this thesis. The
exposition we give here is drawn from many sources, in particular, [44], [29, Appendix A], and
[19] for the spectral theory and [4, Appendix A], [23], and [34] for the spectral perturbation
theory.
3.1
On Holomorphic Matrix-Valued Functions of One
Variable
We introduce the following notation and convention for this section:
46
1. 𝒪(𝜆0 ) :=
∪
𝒪(𝐵(𝜆0 , 𝑟), ℂ) – the set of scalar functions holomorphic at 𝜆0 ∈ ℂ.
𝑟>0
2. 𝒪𝑚×𝑛 (𝜆0 ) :=
∪
𝒪(𝐵(𝜆0 , 𝑟), 𝑀𝑚,𝑛 (ℂ)) – the set of 𝑚 × 𝑛 matrix-valued functions
𝑟>0
holomorphic at 𝜆0 ∈ ℂ.
3. As convention we identify 𝒪1×1 (𝜆0 ) with 𝒪(𝜆0 ).
4. If 𝑓1 , 𝑓2 ∈ 𝒪𝑚×𝑛 (𝜆0 ) then we use the notation 𝑓1 = 𝑓2 to mean there exists an 𝑟 > 0
such that 𝑓1 , 𝑓2 ∈ 𝒪(𝐵(𝜆0 , 𝑟), 𝑀𝑚,𝑛 (ℂ)) and 𝑓1 (𝜆) = 𝑓2 (𝜆) for ∣𝜆 − 𝜆0 ∣ < 𝑟.
5. As is known, if 𝑓 ∈ 𝒪𝑚×𝑛 (𝜆0 ) then 𝑓 (𝜆) is analytic at 𝜆 = 𝜆0 , i.e., 𝑓 (𝜆) is infinitely
differentiable at 𝜆 = 𝜆0 and, denoting its 𝑗th derivative at 𝜆 = 𝜆0 by 𝑓 (𝑗) (𝜆0 ), there
exists an 𝑟 > 0 such that
∞
∑
1 (𝑗)
𝑓 (𝜆0 )(𝜆 − 𝜆0 )𝑗 , ∣𝜆 − 𝜆0 ∣ < 𝑟,
𝑓 (𝜆) =
𝑗!
𝑗=0
where the power series on the right converges absolutely in the 𝑀𝑚,𝑛 (ℂ) norm to 𝑓 (𝜆)
for ∣𝜆 − 𝜆0 ∣ < 𝑟.
We begin with the definition of the spectrum of a holomorphic matrix function and the
important notion of spectral equivalence for holomorphic matrix functions.
Definition 2 If 𝐹 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)), where 𝑈 is an open connected set in ℂ, then the spectrum of 𝐹 , denoted by 𝜎(𝐹 ), is the subset of 𝑈 such that the matrix 𝐹 (𝜆) is not invertible,
i.e.,
𝜎(𝐹 ) := {𝜆 ∈ 𝑈 ∣ det 𝐹 (𝜆) = 0}.
(3.1)
A point 𝜆0 ∈ 𝑈 is called an eigenvalue of 𝐹 provided 𝜆0 ∈ 𝜎(𝐹 ).
Definition 3 Two holomorphic matrix functions 𝐹, 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 ) are called equivalent
47
at 𝜆0 provided there exists 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0 such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are
invertible matrices and
𝐹 (𝜆) = 𝑁 (𝜆)𝐺(𝜆)𝑀 (𝜆),
(3.2)
for every 𝜆 ∈ 𝐵(𝜆0 , 𝑟).
For the local spectral theory of holomorphic matrix functions the follows theorem plays a
central role:
Theorem 9 (Local Smith Form) Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) with 𝜆0 ∈ 𝜎(𝐹 ). Then 𝐹 is equivalent at 𝜆0 to a unique diagonal matrix 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 ) with
𝐺(𝜆) = diag{0, . . . , 0, (𝜆 − 𝜆0 )𝑚𝜇+1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1},
(3.3)
where 0 ≤ 𝜇 ≤ 𝑔 ≤ 𝑛 and the number of zeros down the diagonal is 𝜇, when 𝜇 = 𝑔 there are
no (𝜆 − 𝜆0 ) terms appearing in the diagonal, when 𝜇 < 𝑔 we have 𝑚𝜇+1 ≥ . . . ≥ 𝑚𝑔 ≥ 1 with
{𝑚𝜇+𝑗 }𝑔𝑗=1 ⊆ ℕ, and when 𝑔 = 𝑛 there are no ones appearing in the diagonal.
Proof. This existence is proven in [29, pp. 414–415, Theorem A.6.1] and uniqueness follows
from [19, p. 106, Theorem 4.3.1].
Definition 4 The matrix 𝐺 in (3.3) is called the local Smith form of 𝐹 corresponding to
the eigenvalue 𝜆0 .
Now its obvious from the theorem that
𝑔 = dim(ker(𝐹 (𝜆0 ))),
(3.4)
𝜇 = dim(ker(𝐹 (𝜆))), 0 < ∣𝜆 − 𝜆0 ∣ ≪ 1,
(3.5)
48
and, since det(𝐹 ) ∈ 𝒪(𝜆0 ),
𝜇 = 0 ⇔ det(𝐹 ) ∕= 0.
(3.6)
But an interpretation of the numbers {𝑚𝜇+𝑗 }𝑔𝑗=1 is not as straight forward. Thus we now
give an important example showing how to interpret this theorem in a special case to the
structure of the Jordan normal form of a matrix corresponding to an eigenvalue. Moreover,
the following example will help to motivate the discussion in the next section.
Example 1. Let 𝐴 ∈ 𝑀𝑛 (ℂ) and define 𝐹 (𝜆) := 𝐴 − 𝜆𝐼𝑛 for 𝜆 ∈ ℂ. Let 𝜆0 be an eigenvalue
of 𝐴. Then the hypotheses of Theorem 9 are satisfied and so 𝐹 is equivalent at 𝜆0 to a
diagonal matrix
𝐺(𝜆) = diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}.
(3.7)
where 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1 and {𝑚𝑗 }𝑔𝑗=1 ⊆ ℕ. The numbers 𝑔, 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 :=
𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 are called the geometric multiplicity, partial multiplicities, and algebraic
multiplicity, respectively, of the eigenvalue 𝜆0 of the matrix 𝐴. It follows from [18, §2.2] and
[18, p. 657, §A.3, Theorem A.3.4] that the Jordan normal form of the matrix 𝐴 corresponding
to the eigenvalue 𝜆0 consists of 𝑔 Jordan blocks of order 𝑚1 , . . . , 𝑚𝑔 , respectively, with
eigenvalue 𝜆0 , that is, there exists matrices 𝑆 ∈ 𝑀𝑛 (ℂ) and 𝑊0 ∈ 𝑀𝑛−𝑚 (ℂ) such that 𝑆
is an invertible matrix, 𝜆0 is not an eigenvalue of the matrix 𝑊0 , and 𝑆 −1 𝐴𝑆 is the block
diagonal matrix
𝑆 −1 𝐴𝑆 = diag{𝐽𝑚1 (𝜆0 ), . . . , 𝐽𝑚𝑔 (𝜆0 ), 𝑊0 },
49
(3.8)
where 𝐽𝑚𝑗 (𝜆0 ) is the 𝑚𝑗 × 𝑚𝑗 Jordan block given by
⎡
𝜆 1
⎢ 0
⎢
⎢
. .
⎢
⎢
𝐽𝑚𝑗 (𝜆0 ) = ⎢
. .
⎢
⎢
⎢
.
⎢
⎣
⎤
⎥
⎥
⎥
⎥
⎥
⎥ , for 𝑗 = 1, . . . , 𝑔,
⎥
⎥
⎥
1⎥
⎦
𝜆0
(3.9)
where all unmarked elements are zeros.
In the next section we use Theorem 9 and this example to analogously define for an eigenvalue of a holomorphic matrix function, its geometric multiplicity, partial multiplicities, and
algebraic multiplicity. We also introduce the notion of a Jordan chain for a holomorphic
matrix function and again use this example to compare it with the standard definition of a
Jordan chain of a matrix (see [35, §6.3] for the standard definitions in the matrix case).
3.1.1
Local Spectral Theory of Holomorphic Matrix Functions
For holomorphic matrix functions, a local spectral theory can be given based on the local
Smith form. We now give a brief exposition of this theory.
Definition 5 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ). Then ker(𝐹 (𝜆0 )) is called the eigenspace of 𝐹 at 𝜆0 .
If 𝜑 ∈ ker(𝐹 (𝜆0 )) and 𝜑 ∕= 0 then 𝜑 is called an eigenvector of 𝐹 corresponding to the
eigenvalue 𝜆0 .
Definition 6 If 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝜆0 is an eigenvalue 𝐹 then we say 𝜆0 is an eigenvalue
of 𝐹 of finite (infinite) algebraic multiplicity if det(𝐹 ) ∕= 0 (det(𝐹 ) = 0).
Definition 7 Let 𝐹 be any holomorphic matrix function such that 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝜆0 an
50
eigenvalue of 𝐹 of finite algebraic multiplicity. By Theorem 9 and (3.6) there exists unique
numbers 𝑔, 𝑚1 , . . . , 𝑚𝑔 ∈ ℕ satisfying 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 such that the local Smith form of 𝐹
corresponding to the eigenvalue 𝜆0 is
𝐺(𝜆) = diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}.
(3.10)
The positive integers 𝑔, 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 := 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 are called the geometric
multiplicity, partial multiplicities, and algebraic multiplicity, respectively, of the
eigenvalue 𝜆0 of 𝐹 . We say 𝜆0 is a semisimple eigenvalue of 𝐹 if 𝑚1 = ⋅ ⋅ ⋅ = 𝑚𝑔 = 1.
The following proposition gives a simple but important characterization of the geometric
and algebraic multiplicity.
Proposition 10 Suppose 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝜆0 is an eigenvalue of 𝐹 of finite algebraic
multiplicity. Let 𝑔 and 𝑚 be the geometric multiplicity and algebraic multiplicity of the
eigenvalue 𝜆0 of 𝐹 . Then 𝑔 is the dimension of ker(𝐹 (𝜆0 )) and 𝑚 is the order of the zero of
the function det(𝐹 (𝜆)) at 𝜆 = 𝜆0 .
Proof.
Let 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 be the partial multiplicities of the eigenvalue 𝜆0 of 𝐹 so
that 𝑚 = 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 . By Theorem 9 we know that 𝐹 is equivalent at 𝜆0 to its local
Smith form. Thus there exists 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0 such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are
invertible matrices and
𝐹 (𝜆) = 𝑁 (𝜆) diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}𝑀 (𝜆),
for every 𝜆 ∈ 𝐵(𝜆0 , 𝑟). It follows from this representation that the first 𝑔 columns of the
matrix 𝑀 (𝜆0 )−1 are a basis for ker(𝐹 (𝜆0 )) so that 𝑔 = dim(ker(𝐹 (𝜆0 ))). Moreover, since
51
det(𝑁 ), det(𝑀 ) ∈ 𝒪(𝜆0 ) with det(𝑁 (𝜆0 )), det(𝑀 (𝜆0 )) ∕= 0, we have
det(𝐹 (𝜆)) = det(𝑁 (𝜆) diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}𝑀 (𝜆))
= (𝜆 − 𝜆0 )𝑚 det(𝑁 (𝜆) det(𝑀 (𝜆))
= (𝜆 − 𝜆0 )𝑚 det(𝑁 (𝜆0 )) det(𝑀 (𝜆0 )) + 𝑂((𝜆 − 𝜆0 )𝑚+1 ), as 𝜆 → 𝜆0 .
Hence 𝑚 is the order of the zero of the function det(𝐹 (𝜆)) at 𝜆 = 𝜆0 . This completes the
proof.
Definition 8 Let 𝐹 be any holomorphic matrix function such that 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝜆0 an
𝑛
eigenvalue of 𝐹 . A sequence {𝜑𝑗 }𝑙−1
𝑗=0 ⊆ ℂ , where 𝑙 ∈ ℕ, is called a Jordan chain of 𝐹 of
length 𝑙 corresponding to the eigenvalue 𝜆0 (associated with the eigenvector 𝜑0 ) if
𝜑0 ∕= 0 and
𝑗
∑
1 (ℎ)
𝐹 (𝜆0 )𝜑𝑗−ℎ = 0,
ℎ!
ℎ=0
𝑗 = 0, . . . , 𝑙 − 1.
(3.11)
The maximal length of a Jordan chain of 𝐹 corresponding to the eigenvalue 𝜆0 and associated
with 𝜑0 is called the multiplicity of the eigenvector 𝜑0 and is denoted by 𝑚(𝜑0 ), i.e.,
𝑚(𝜑0 ) := sup{𝑙 ∈ ℕ : {𝜑𝑗 }𝑙−1
𝑗=0 satisfies (3.11)}.
(3.12)
Example 2. Let 𝐴 ∈ 𝑀𝑛 (ℂ) and define 𝐹 (𝜆) := 𝐴 − 𝜆𝐼𝑛 for 𝜆 ∈ ℂ. Then by Definition
5, 𝜆0 is an eigenvalue of 𝐴 if and only if 𝜆0 is an eigenvalue of 𝐹 . We will now show that
𝑛
a sequence of vectors {𝜑𝑗 }𝑙−1
𝑗=0 ⊆ ℂ is a Jordan chain of 𝐴 of length 𝑙 corresponding to the
eigenvalue 𝜆0 (see [35, p. 230, §6.3] for definition) if and only if {𝜑𝑗 }𝑙−1
𝑗=0 is a Jordan chain of
𝐹 of length 𝑙 corresponding to the eigenvalue 𝜆0 .
52
Suppose 𝜆0 is an eigenvalue of 𝐴 and {𝜑𝑗 }𝑙−1
𝑗=0 is a Jordan chain of 𝐴 of length 𝑙 corresponding
to the eigenvalue 𝜆0 , i.e.,
𝜑0 ∕= 0, (𝐴 − 𝜆0 𝐼𝑛 )𝜑𝑗 = 𝜑𝑗−1 ,
𝑗 = 0, . . . , 𝑙 − 1
where 𝜑−1 := 0. Then 𝜑0 ∕= 0 and
𝑗
∑
1 (ℎ)
𝐹 (𝜆0 )𝜑𝑗−ℎ = 𝐹 (0) (𝜆0 )𝜑𝑗 + 𝐹 (1) (𝜆0 )𝜑𝑗−1 = (𝐴 − 𝜆0 𝐼𝑛 )𝜑𝑗 − 𝜑𝑗−1 = 0
ℎ!
ℎ=0
for 𝑗 = 0, . . . , 𝑙 − 1. This implies by Definition 8 that the sequence {𝜑𝑗 }𝑙−1
𝑗=0 is a Jordan chain
of 𝐹 of length 𝑙 corresponding to the eigenvalue 𝜆0 .
Conversely, if {𝜑𝑗 }𝑙−1
𝑗=0 is a Jordan chain of 𝐹 of length 𝑙 corresponding to the eigenvalue 𝜆0
then by Definition 8 this sequence satisfies
𝑗
∑
1 (ℎ)
𝜑0 ∕= 0 and 0 =
𝐹 (𝜆0 )𝜑𝑗−ℎ = (𝐴 − 𝜆0 𝐼𝑛 )𝜑𝑗 − 𝜑𝑗−1 ,
ℎ!
ℎ=0
𝑗 = 0, . . . , 𝑙 − 1,
𝑙−1
where 𝜑−1 := 0. This implies {𝜑𝑗 }𝑗=0
⊆ ℂ𝑛 is a Jordan chain of 𝐴 of length 𝑙 corresponding
to the eigenvalue 𝜆0 .
Definition 9 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ). A vector-valued function 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) is called a generating function of order 𝑙 for 𝐹 at the eigenvalue 𝜆0 , where 𝑙 ∈ ℕ, if 𝜑(𝜆0 ) ∕= 0 and the
function 𝐹 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) has a zero of order ≥ 𝑙 at 𝜆0 , i.e.,
𝜑(𝜆0 ) ∕= 0 𝑎𝑛𝑑 𝐹 (𝜆)𝜑(𝜆) = 𝑂((𝜆 − 𝜆0 )𝑙 ), as 𝜆 → 𝜆0
(3.13)
Proposition 11 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ). Then 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) is a generating function of order 𝑙
53
for 𝐹 at the eigenvalue 𝜆0 if and only if 𝜆0 is an eigenvalue 𝐹 and the sequence
𝜑𝑗 :=
1 (𝑗)
𝜑 (𝜆0 ),
𝑗!
𝑗 = 0, . . . , 𝑙 − 1
(3.14)
is a Jordan chain of F of length 𝑙 corresponding to the eigenvalue 𝜆0 .
Proof. This statement and its proof can be found in [44, p. 57, §II.11, Lemma 11.3]. The
result follows from the fact if 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) then 𝐹 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) and
hence
𝐹 (𝜆)𝜑(𝜆) =
∞
∞
∑
∑
1 (𝑗)
1 (𝑗)
𝐹 (𝜆0 )(𝜆 − 𝜆0 )𝑗
𝜑 (𝜆0 )(𝜆 − 𝜆0 )𝑗
𝑗!
𝑗!
𝑗=0
𝑗=0
𝑗
∞ ∑
∑
1
1 (ℎ)
𝐹 (𝜆0 )
𝜑(𝑗−ℎ) (𝜆0 )
=
ℎ!
(𝑗
−
ℎ)!
𝑗=0 ℎ=0
𝑗
𝑙−1 ∑
∑
1 (ℎ)
1
=
𝐹 (𝜆0 )
𝜑(𝑗−ℎ) (𝜆0 ) + 𝑂((𝜆 − 𝜆0 )𝑙 ), as 𝜆 → 𝜆0 .
ℎ!
(𝑗 − ℎ)!
𝑗=0 ℎ=0
Definition 10 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) such that 𝜆0 an eigenvalue of 𝐹 of finite algebraic mul𝑚 −1
𝑗
tiplicity. We say {𝜑𝑖,𝑗 }𝑖=0
, 𝑗 = 1, . . . , 𝑔 is a canonical system of Jordan chains of 𝐹
corresponding to the eigenvalue 𝜆0 if the following three conditions are satisfied:
1. The numbers 𝑔, 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 := 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 are the geometric multiplicity,
partial multiplicities, and algebraic multiplicity, respectively, of the eigenvalue 𝜆0 of 𝐹 .
𝑚 −1
𝑗
2. The sequence {𝜑𝑖,𝑗 }𝑖=0
is a Jordan chain of 𝐹 of length 𝑚𝑗 , for 𝑗 = 1, . . . , 𝑔 and
𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1.
3. The vectors 𝜑0,𝑗 , 𝑗 = 1, . . . , 𝑔 form a basis for the eigenspace ker(𝐹 (𝜆0 )).
54
Example 3. Let 𝐴 ∈ 𝑀𝑛 (ℂ) and define 𝐹 (𝜆) := 𝐴 − 𝜆𝐼𝑛 for 𝜆 ∈ ℂ. Then it can be shown
𝑚 −1
𝑗
that {𝜑𝑖,𝑗 }𝑖=0
, 𝑗 = 1, . . . , 𝑔 is a canonical system of Jordan chains of 𝐹 corresponding to
∪
𝑚𝑗 −1
the eigenvalue 𝜆0 if and only if the set of vectors 𝑔𝑗=1 {𝜑𝑖,𝑗 }𝑖=0
form a Jordan basis for the
generalized eigenspace of the matrix 𝐴 corresponding to the eigenvalue 𝜆0 (see [35, p. 232,
𝑚 −1
𝑗
§6.4] for the definition) where {𝜑𝑖,𝑗 }𝑖=0
is a Jordan chain of 𝐴 of length 𝑚𝑗 corresponding
to the eigenvalue 𝜆0 and 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1.
Lemma 12 The relation ∼ defined by 𝐹 ∼ 𝐺 if 𝐹, 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 ) are equivalent at 𝜆0 , is
an equivalence relation on 𝒪𝑛×𝑛 (𝜆0 ).
Proof.
First, its obvious the relation ∼ is reflexive since the identity matrix 𝐼𝑛 as a
constant function belongs in 𝒪𝑛×𝑛 (𝜆0 ). Now lets show the relation is symmetric. Suppose
𝐹, 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝐹 ∼ 𝐺. This implies there exists 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0
such that 𝐹 (𝜆) = 𝑁 (𝜆)𝐺(𝜆)𝑀 (𝜆) for ∣𝜆 − 𝜆0 ∣ < 𝑟 and 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are invertible matrices.
It follows from [45, p. 7, §1.2, Proposition 1.2.5] that there exists a possibly smaller 𝑟 > 0
such that 𝑁 −1 (𝜆) := 𝑁 (𝜆)−1 and 𝑀 −1 (𝜆) := 𝑀 (𝜆)−1 exist for ∣𝜆 − 𝜆0 ∣ < 𝑟 and 𝑁 −1 , 𝑀 −1 ∈
𝒪(𝐵(𝜆0 , 𝑟), 𝑀𝑛 (ℂ)). But then 𝑁 −1 , 𝑀 −1 ∈ 𝒪𝑛×𝑛 (𝜆0 ), 𝑁 −1 (𝜆), 𝑀 −1 (𝜆0 ) are invertible, and
𝐺(𝜆) = 𝑁 −1 (𝜆)𝐹 (𝜆)𝑀 −1 (𝜆) for ∣𝜆 − 𝜆0 ∣ < 𝑟. This proves 𝐺 ∼ 𝐹 and so ∼ is symmetric. We
now prove ∼ is transitive. Indeed, let 𝐹 ∼ 𝐺 and 𝐺 ∼ 𝐻. Then there exists 𝑁1 , 𝑁2 , 𝑀1 , 𝑀2 ∈
𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0 such that 𝐹 (𝜆) = 𝑁1 (𝜆)𝐺(𝜆)𝑀1 (𝜆) = 𝑁1 (𝜆)𝑁2 (𝜆)𝐻(𝜆)𝑀1 (𝜆)𝑀2 (𝜆)
for ∣𝜆 − 𝜆0 ∣ < 𝑟 but the products 𝑁1 𝑁2 and 𝑀1 𝑀2 are in 𝒪𝑛×𝑛 (𝜆0 ) and are invertible at 𝜆0 .
Thus 𝐹 ∼ 𝐻 and so ∼ is transitive. This proves the relation ∼ is an equivalence relation on
𝒪𝑛×𝑛 (𝜆0 ).
Proposition 13 Let 𝐹, 𝐺 be any holomorphic matrix function such that 𝐹, 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 )
are equivalent at 𝜆0 . Then
1. 𝜆0 is an eigenvalue of 𝐹 of finite algebraic multiplicity if and only if 𝜆0 is an eigenvalue
55
of 𝐺 of finite algebraic multiplicity.
2. Let 𝜆0 be an eigenvalue of 𝐹 of finite algebraic multiplicity. Then the local Smith
forms of 𝐹 and 𝐺 corresponding to the eigenvalue 𝜆0 are equal and, in particular, the
geometric, partial, and algebraic multiplicities of the eigenvalue 𝜆0 of 𝐹 and 𝐺 are the
same.
3. Let 𝜆0 be an eigenvalue of 𝐹 of finite algebraic multiplicity. Let 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) be
such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are invertible matrices and
𝐹 (𝜆) = 𝑁 (𝜆)𝐺(𝜆)𝑀 (𝜆)
(3.15)
for every ∣𝜆 − 𝜆0 ∣ < 𝑟. If {𝜑𝑗 }𝑙𝑗=0 is a Jordan chain of 𝐹 of length 𝑙 corresponding to
the eigenvalue 𝜆0 then {˜
𝜑𝑗 }𝑙𝑗=0 is a Jordan chain of 𝐺 of length 𝑙 corresponding to the
eigenvalue 𝜆0 where
𝑗
𝜑
˜ 𝑗 :=
Proof.
1 ∑ 1 (ℎ)
𝑀 (𝜆0 )𝜑𝑗−ℎ ,
𝑗! ℎ=0 ℎ!
𝑗 = 0, . . . , 𝑙 − 1.
(3.16)
Let 𝐹, 𝐺 be any holomorphic matrix function such that 𝐹, 𝐺 ∈ 𝒪𝑛×𝑛 (𝜆0 ) are
equivalent at 𝜆0 . It follows from Lemma 12 and Theorem 9 that the local Smith forms of
𝐹 and 𝐺 are equal. This implies that 𝜆0 is an eigenvalue of 𝐹 finite algebraic multiplicity
if and only if 𝜆0 is an eigenvalue of 𝐺 finite algebraic multiplicity. It also implies that if 𝜆0
is an eigenvalue of 𝐹 finite algebraic multiplicity then the geometric, partial, and algebraic
multiplicities of the eigenvalue 𝜆0 of 𝐹 and 𝐺 are the same.
Suppose that 𝜆0 is an eigenvalue of 𝐹 finite algebraic multiplicity. Let 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) be
such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are invertible matrices and
𝐹 (𝜆) = 𝑁 (𝜆)𝐺(𝜆)𝑀 (𝜆)
56
(3.17)
for every ∣𝜆 − 𝜆0 ∣ < 𝑟. Suppose {𝜑𝑗 }𝑙𝑗=0 is a Jordan chain of 𝐹 of length 𝑙 corresponding to
the eigenvalue 𝜆0 . Define the vector-valued function 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) by
𝜑(𝜆) :=
𝑙−1
∑
𝜑𝑗 (𝜆 − 𝜆0 )𝑗 ,
𝜆 ∈ ℂ.
𝑗=0
It follows from Proposition 11 that 𝜑 is a generating function of order 𝑙 for 𝐹 at the eigenvalue
𝜆0 . This implies
𝜑(𝜆0 ) ∕= 0,
𝐹 (𝜆)𝜑(𝜆) = 𝑂((𝜆 − 𝜆0 )𝑙 ), as 𝜆 → 𝜆0 .
But since 𝑁 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝑁 (𝜆0 ) is invertible this implies there exists a possibly smaller
𝑟 > 0 such that 𝑁 −1 (𝜆) := 𝑁 (𝜆)−1 exists for ∣𝜆 − 𝜆0 ∣ < 𝑟 and 𝑁 −1 ∈ 𝒪(𝐵(𝜆0 , 𝑟), 𝑀𝑛 (ℂ)).
This implies
𝐺(𝜆)𝑀 (𝜆)𝜑(𝜆) = 𝑁 −1 (𝜆)𝐹 (𝜆)𝜑(𝜆) = 𝑂((𝜆 − 𝜆0 )𝑙 ), as 𝜆 → 𝜆0 .
But since 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and 𝑀 (𝜆0 ) is invertible then 𝑀 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) and
𝑀 (𝜆0 )𝜑(𝜆0 ) ∕= 0,
𝐺(𝜆)𝑀 (𝜆)𝜑(𝜆) = 𝑂((𝜆 − 𝜆0 )𝑙 ), as 𝜆 → 𝜆0 .
This implies by Definition 9 that 𝑀 𝜑 is a generating function of order 𝑙 for 𝐹 at the eigenvalue
𝜆0 and so by Proposition 11, {˜
𝜑𝑗 }𝑙𝑗=0 is a Jordan chain of 𝐺 of length 𝑙 corresponding to the
eigenvalue 𝜆0 where
𝑗
1 ∑ 1 (ℎ)
1 𝑑𝑗
(𝑀 (𝜆)𝜑(𝜆)) 𝜆=𝜆0 =
𝑀 (𝜆0 )𝜑𝑗−ℎ ,
𝜑
˜ 𝑗 :=
𝑗! 𝑑𝜆𝑗
𝑗! ℎ=0 ℎ!
𝑗 = 0, . . . , 𝑙 − 1.
This completes the proof.
Theorem 14 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) such that 𝜆0 an eigenvalue of 𝐹 of finite algebraic mul57
tiplicity. Then there exists a canonical system of Jordan chains of 𝐹 corresponding to the
eigenvalue 𝜆0 .
Proof.
Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) such that 𝜆0 an eigenvalue of 𝐹 of finite algebraic multiplicity.
Denote by 𝑔, 𝑚1 , . . . , 𝑚𝑔 , 𝑚 the geometric, partial, and algebraic multiplicities, respectively,
of the eigenvalue 𝜆0 of 𝐹 where 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1 and 𝑚 = 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 . By Theorem
9 and Lemma 12 there exists 𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0 such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are
invertible matrices and
𝐺(𝜆) = 𝑁 (𝜆)𝐹 (𝜆)𝑀 (𝜆),
for every 𝜆 ∈ 𝐵(𝜆0 , 𝑟) where
𝐺(𝜆) = diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}.
Let 𝑒1 , . . . , 𝑒𝑛 denote the standard orthonormal basis vectors in ℂ𝑛 . Define vector-valued
functions 𝜑𝑗 ∈ 𝒪𝑛×1 (𝜆0 ), 𝑗 = 1, . . . , 𝑔 by
𝜑𝑗 (𝜆) = 𝑒𝑗 ,
𝜆 ∈ ℂ,
𝑗 = 1, . . . , 𝑔.
(3.18)
Then since
𝜑𝑗 (𝜆0 ) = 𝑒𝑗 ∕= 0,
𝐺(𝜆)𝜑𝑖 (𝜆) = (𝜆 − 𝜆0 )𝑚𝑗 𝑒𝑗 = 𝑂((𝜆 − 𝜆0 )𝑚𝑖 ), as 𝜆 → 𝜆0 ,
(3.19)
each 𝜑𝑗 is a generating function of order 𝑚𝑗 for 𝐺 at the eigenvalue 𝜆0 and so by Proposition
𝑚 −1
𝑗
11 the sequence {𝜑𝑖,𝑗 }𝑖=0
defined by
𝜑𝑖,𝑗 :=
1 (𝑖)
𝜑 (𝜆0 ) = 𝛿 0,𝑖 𝑒𝑗 ,
𝑖! 𝑗
58
𝑖 = 0, . . . , 𝑚𝑗 − 1,
(3.20)
where 𝛿 0,𝑖 is the Kronecker delta symbol, is a Jordan chain of 𝐺 of length 𝑚𝑗 corresponding to
𝑚 −1
𝑗
the eigenvalue 𝜆0 , for 𝑗 = 1, . . . , 𝑔. It follows that {𝜑𝑖,𝑗 }𝑖=0
, 𝑗 = 1, . . . , 𝑔 is canonical system
of Jordan chains for 𝐺 corresponding to the eigenvalue 𝜆0 , that is, we have constructed
a canonical system of Jordan chains for the local Smith form of 𝐹 corresponding to the
𝑚 −1
𝑗
eigenvalue 𝜆0 . By Proposition 13 the sequence {˜
𝜑𝑖,𝑗 }𝑖=0
defined by
𝑖
𝜑
˜ 𝑖,𝑗
1 ∑ 1 (ℎ)
𝑀 (𝜆0 )𝜑𝑖−ℎ,𝑗 ,
:=
𝑖! ℎ=0 ℎ!
𝑖 = 0, . . . , 𝑚𝑗 − 1
(3.21)
is a Jordan chain of 𝐹 of length 𝑚𝑗 corresponding to the eigenvalue 𝜆0 , for 𝑗 = 1, . . . , 𝑔. Moreover, since 𝑒1 , . . . , 𝑒𝑔 is a basis for ker 𝐺(𝜆0 ), 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are invertible, and 𝑁 (𝜆0 )𝐺(𝜆0 ) =
𝐹 (𝜆0 )𝑀 (𝜆0 ) then 𝜑
˜ 0,𝑗 = 𝑀 (𝜆0 )𝑒𝑗 , 𝑗 = 1, . . . , 𝑔 is a basis for ker 𝐹 (𝜆0 ). Thus we have shown
the numbers 𝑔, 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 = 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 are the geometric multiplicity, partial
multiplicities, and algebraic multiplicity, respectively, of the eigenvalue 𝜆0 of 𝐹 , the sequence
𝑚 −1
𝑗
{𝜑𝑖,𝑗 }𝑖=0
is a Jordan chain of 𝐹 of length 𝑚𝑗 , for 𝑗 = 1, . . . , 𝑔 and 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1, and
the vectors 𝜑0,𝑗 , 𝑗 = 1, . . . , 𝑔 form a basis for the eigenspace ker(𝐹 (𝜆0 )). Therefore, by Defi𝑚 −1
𝑗
nition 10, {𝜑𝑖,𝑗 }𝑖=0
, 𝑗 = 1, . . . , 𝑔 is a canonical system of Jordan chains of 𝐹 corresponding
to the eigenvalue 𝜆0 . This completes the proof.
It will be important in this thesis to know when an eigenvalue of a holomorphic matrix
function is both of finite algebraic multiplicity and semisimple. The next proposition gives
us a simple sufficient condition for this to be true.
Proposition 15 If 𝜆0 is an eigenvalue of 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and the equation
𝐹 (1) (𝜆0 )𝜑0 = −𝐹 (𝜆0 )𝜑1
(3.22)
has no solution for 𝜑0 ∈ ker(𝐹 (𝜆0 ))/{0} and 𝜑1 ∈ ℂ𝑛 then 𝜆0 is an eigenvalue of 𝐹 of finite
algebraic multiplicity and a semisimple eigenvalue of 𝐹 .
59
Proof.
Let the hypotheses of this proposition be true for some 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ). Then the
hypothesis that (3.22) has no solution for 𝜑0 ∈ ker(𝐹 (𝜆0 ))/{0} and 𝜑1 ∈ ℂ𝑛 is equivalent
by Definition 8 to there are no Jordan chains of 𝐹 of length 𝑙 ≥ 2 corresponding to the
eigenvalue 𝜆0 . Thus if 𝜆0 was an eigenvalue of 𝐹 but not of finite algebraic multiplicity then
by Theorem 9 we could find a vector-valued function 𝜑 ∈ 𝒪𝑛×1 (𝜆0 ) and an 𝑟 > 0 such that
𝐹 (𝜆)𝜑(𝜆) = 0 for ∣𝜆 − 𝜆0 ∣ < 𝑟. By Proposition 11 we could find a Jordan chain of order 𝑙
for any 𝑙 ∈ ℕ and in particular, for 𝑙 ≥ 2, a contradiction. Hence 𝜆0 must be an eigenvalue
of 𝐹 of finite algebraic multiplicity. If 𝜆0 was an eigenvalue of 𝐹 but was not a semisimple
eigenvalue of 𝐹 then by Theorem 14 there would exist a Jordan chain of 𝐹 of length ≥ 2,
a contradiction. Therefore, 𝜆0 is an eigenvalue of 𝐹 of finite algebraic multiplicity and a
semisimple eigenvalue of 𝐹 . This completes the proof.
Another important result needed in this thesis is given by the next proposition.
Proposition 16 Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and let 𝑓 ∈ 𝒪𝑛×𝑛 (𝑧0 ) be such that 𝑓 (𝑧0 ) = 𝜆0 . Define
𝐺(𝑧) := 𝐹 (𝑓 (𝑧)) for all 𝑧 such that the composition is well-defined. Then 𝐺 ∈ 𝒪𝑛×𝑛 (𝑧0 )
and 𝜆0 is an eigenvalue of 𝐹 if and only if 𝑧0 is an eigenvalue of 𝐺. Moreover, if 𝜆0 is an
eigenvalue of 𝐹 and 𝑓 ′ (𝑧0 ) ∕= 0 then
1. 𝜆0 is an eigenvalue of 𝐹 of finite algebraic multiplicity if and only if 𝑧0 is an eigenvalue
of 𝐺 of finite algebraic multiplicity.
2. Let 𝜆0 be an eigenvalue of 𝐹 of finite algebraic multiplicity. Then the geometric, partial,
and algebraic multiplicities of the eigenvalue 𝜆0 of 𝐹 and of the eigenvalue 𝑧0 of 𝐺 are
in one-to-one correspondence.
Proof.
Let 𝐹 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and let 𝑓 ∈ 𝒪𝑛×𝑛 (𝑧0 ) such that 𝑓 (𝑧0 ) = 𝜆0 . It follows there
exists an 𝜖 > 0 such that 𝐹 ∈ 𝒪(𝐵(𝜆0 , 𝜖), 𝑀𝑛 (ℂ)). Thus there exists a 𝛿 > 0 such that
𝑓 ∈ 𝒪(𝐵(𝜆0 , 𝛿), ℂ)) and if 𝑧 ∈ 𝐵(𝜆0 , 𝛿) then 𝑓 (𝑧) ∈ 𝐵(𝜆0 , 𝜖). It then follows 𝐺(𝑧) = 𝐹 (𝑓 (𝑧))
60
is well-defined on 𝐵(𝜆0 , 𝛿) and 𝐺′ (𝑧) = 𝑓 ′ (𝑧)𝐹 ′ (𝑓 (𝑧)) for every 𝑧 ∈ 𝐵(𝜆0 , 𝛿). Thus 𝐺 ∈
𝒪(𝐵(𝑧0 , 𝛿), 𝑀𝑛 (ℂ)) and so 𝐺 ∈ 𝒪𝑛×𝑛 (𝑧0 ). It follows from the fact det 𝐹 (𝜆0 ) = det 𝐺(𝑧0 )
that 𝜆0 is an eigenvalue of 𝐹 if and only if 𝑧0 is an eigenvalue of 𝐺.
Suppose 𝜆0 is an eigenvalue of 𝐹 . Then by the local Smith form, Theorem 9, there exists
𝑁, 𝑀 ∈ 𝒪𝑛×𝑛 (𝜆0 ) and an 𝑟 > 0 with 𝑟 ≤ 𝜖 such that 𝑁 (𝜆0 ), 𝑀 (𝜆0 ) are invertible matrices
and
𝐹 (𝜆) = 𝑁 (𝜆) diag{0𝜇 , (𝜆 − 𝜆0 )𝑚𝜇+1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 𝐼𝑛−𝑔 }𝑀 (𝜆),
for every 𝜆 ∈ 𝐵(𝜆0 , 𝑟), where 0 ≤ 𝜇 ≤ 𝑔 ≤ 𝑛, 0𝜇 is the 𝜇 × 𝜇 zero matrix, 𝐼𝑛−𝑔 is the
(𝑛−𝑔)×(𝑛−𝑔) identity matrix, the number of zeros down the diagonal is 𝜇, when 𝜇 = 𝑔 there
are no (𝜆 − 𝜆0 ) terms appearing in the diagonal, when 𝜇 < 𝑔 we have 𝑚𝜇+1 ≥ . . . ≥ 𝑚𝑔 ≥ 1
with {𝑚𝜇+𝑗 }𝑔𝑗=1 ⊆ ℕ, and when 𝑔 = 𝑛 there are no ones appearing in the diagonal.
Again we can find a 𝛿 1 ≤ 𝛿 such that 𝑓 (𝑧) ∈ 𝐵(𝜆0 , 𝑟) for 𝑧 ∈ 𝐵(𝑧0 , 𝛿 1 ). But from what we
˜ (𝑧) := 𝑁 (𝑓 (𝑧)) and 𝑀
˜ (𝑧) := 𝑀 (𝑓 (𝑧)), for all 𝑧 such that the
just proved this implies for 𝑁
˜, 𝑀
˜ ∈ 𝒪𝑛×𝑛 (𝑧0 ) with 𝑁
˜ (𝑧0 ) = 𝑁 (𝜆0 ), 𝑀
˜ (𝑧0 ) = 𝑀 (𝜆0 )
composition is well-defined, that 𝑁
invertible matrices such that
𝐺(𝑧) = 𝐹 (𝑓 (𝑧))
˜ (𝑧) diag{0𝜇 , (𝑓 (𝑧) − 𝜆0 )𝑚𝜇+1 , . . . , (𝑓 (𝑧) − 𝜆0 )𝑚𝑔 , 𝐼𝑛−𝑔 }𝑀
˜ (𝑧)
=𝑁
for every 𝑧 ∈ 𝐵(𝑧0 , 𝛿 1 ).
Now since 𝑓 ′ (𝑧0 ) ∕= 0 then for the function 𝑓0 (𝑧) :=
𝑓 (𝑧)
𝑧−𝑧0
for 𝑧 ∈ 𝐵(𝑧0 , 𝛿 1 ) we have
𝒪(𝐵(𝑧0 , 𝛿 1 ), ℂ) with 𝑓0 (𝑧0 ) = 𝑓 ′ (𝑧0 ) ∕= 0. Thus there exists a 𝛿 2 > 0 with 𝛿 2 ≤ 𝛿 1 such
that 𝑓0−1 (𝑧) :=
𝑧−𝑧0
𝑓 (𝑧)
with 𝑧 ∈ 𝐵(𝑧0 , 𝛿 2 ) is well-defined and 𝑓0−1 ∈ 𝒪(𝐵(𝑧0 , 𝛿 2 ), ℂ).
61
ˆ (𝑧) for 𝑧 ∈ 𝐵(𝑧0 , 𝛿 2 ) by
Define the matrix 𝑀
˜ (𝑧).
ˆ (𝑧) = diag{𝐼𝜇 , 𝑓0 (𝑧)𝑚𝜇+1 , . . . , 𝑓0 (𝑧)𝑚𝑔 , 𝐼𝑛−𝑔 }𝑀
𝑀
ˆ ∈ 𝒪(𝐵(𝑧0 , 𝛿 2 ), 𝑀𝑛 (ℂ)) with 𝑀
ˆ (𝑧0 ) an invertible matrix and 𝑀
ˆ (𝑧)
It follows from this that 𝑀
is invertible for every 𝑧 ∈ 𝐵(𝑧0 , 𝛿 2 ). Moreover, since we have (𝑓 (𝑧) − 𝜆0 )𝑗 = (𝑧 − 𝑧0 )𝑗 𝑓0 (𝑧)𝑗 ,
it follows that
𝐺(𝑧) = 𝐹 (𝑓 (𝑧))
˜ (𝑧) diag{0𝜇 , (𝑓 (𝑧) − 𝜆0 )𝑚𝜇+1 , . . . , (𝑓 (𝑧) − 𝜆0 )𝑚𝑔 , 𝐼𝑛−𝑔 }𝑀
˜ (𝑧)
=𝑁
˜ (𝑧) diag{0𝜇 , (𝑧 − 𝑧0 )𝑚𝜇+1 , . . . , (𝑧 − 𝑧0 )𝑚𝑔 , 𝐼𝑛−𝑔 }𝑀
ˆ (𝑧)
=𝑁
for every 𝑧 ∈ 𝐵(𝑧0 , 𝛿 2 ).
Therefore by the uniqueness portion of Theorem 9 it follows that the local Smith form of 𝐹
corresponding to the eigenvalue 𝜆0 is
diag{0𝜇 , (𝜆 − 𝜆0 )𝑚𝜇+1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 𝐼𝑛−𝑔 }
and the local Smith form of 𝐺 corresponding to the eigenvalue 𝑧0 is
diag{0𝜇 , (𝑧 − 𝑧0 )𝑚𝜇+1 , . . . , (𝑧 − 𝑧0 )𝑚𝑔 , 𝐼𝑛−𝑔 }.
The proof of the proposition now follows from the uniqueness of those local Smith forms.
The following example arises in our thesis in Floquet theory when studying the Jordan normal
form of the monodromy matrix for canonical systems of periodic differential equations with
period 𝑑.
62
Example 4. Let 𝐴 ∈ 𝑀𝑛 (ℂ) and define 𝐹 (𝜆) := 𝐴 − 𝜆𝐼𝑛 for 𝜆 ∈ 𝑀𝑛 (ℂ). Define 𝐺(𝑘) :=
𝐹 (𝑒𝑖𝑘𝑑 ) = 𝐴−𝑒𝑖𝑘𝑑 𝐼𝑛 , for 𝑘 ∈ ℂ. Let 𝜆0 be an eigenvalue of the matrix 𝐴. Then we know from
example 1 that if we let 𝑔, 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 , 𝑚 := 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 denote the geometric, partial,
and algebraic multiplicities, respectively, of the eigenvalue 𝜆0 of 𝐹 then the Jordan normal
form of the matrix 𝐴 corresponding to the eigenvalue 𝜆0 has exactly 𝑔 Jordan blocks of order
𝑚1 , . . . , 𝑚𝑔 , respectively, with eigenvalue 𝜆0 . Now let 𝑘0 ∈ ℂ such that 𝑒𝑖𝑘0 𝑑 = 𝜆0 . Then,
since
𝑑 𝑖𝑘𝑑
𝑒
𝑑𝑘
= 𝑖𝑑𝑒𝑖𝑘𝑑 ∕= 0 for any 𝑘, it follows from Proposition 16 that 𝑘0 is an eigenvalue of 𝐺
with finite algebraic multiplicity such that its geometric, partial, and algebraic multiplicities
are 𝑔, 𝑚1 , . . . , 𝑚𝑔 and 𝑚, respectively. By the same proposition the converse is also true.
We conclude by noting that it can be helpful sometimes to study the spectrum of the matrix
𝐴 by studying the spectrum of the holomorphic matrix function 𝐺(𝑘) := 𝐴 − 𝑒𝑖𝑘𝑑 𝐼𝑛 . This
perspective is found useful when considering perturbation problems such as when 𝐴 := 𝐴(𝜔)
depends on a perturbation parameter 𝜔.
3.2
On Holomorphic Matrix-Valued Functions of Two
Variables
We introduce the following notation and convention for this section:
(i) 𝒪(𝑈, 𝐸) := {𝑓 : 𝑈 → 𝐸) ∣ 𝑓 is holomorphic on 𝑈 } where 𝑈 is an open connected set
in ℂ2 , (𝐸, ∣∣ ⋅ ∣∣) is a Banach space, and a function 𝑓 : 𝑈 → 𝐸 is said to be holomorphic
on 𝑈 provided the partial derivatives of 𝑓 exist at each point in 𝑈 , i.e., the limits
∂𝑓 := 𝑓𝜆 (𝜆0 , 𝑧0 ) := lim (𝜆 − 𝜆0 )−1 (𝑓 (𝜆, 𝑧0 ) − 𝑓 (𝜆0 , 𝑧0 ))
𝜆→𝜆0
∂𝜆 (𝜆,𝑧)=(𝜆0 ,𝑧0 )
∂𝑓 := 𝑓𝑧 (𝜆0 , 𝑧0 ) := lim (𝑧 − 𝑧0 )−1 (𝑓 (𝜆0 , 𝑧) − 𝑓 (𝜆0 , 𝑧0 ))
𝑧→𝑧0
∂𝑧 (𝜆,𝑧)=(𝜆0 ,𝑧0 )
63
exist for every (𝜆0 , 𝑧0 ) ∈ 𝑈 .
(ii) 𝒪((𝜆0 , 𝑧0 )) :=
∪
𝒪(𝐵((𝜆0 , 𝑧0 ), 𝑟), ℂ) – the set of scalar functions holomorphic at
𝑟>0
(𝜆0 , 𝑧0 ) ∈ ℂ2 .
(iii) 𝒪𝑚×𝑛 ((𝜆0 , 𝑧0 )) :=
∪
𝒪(𝐵((𝜆0 , 𝑧0 ), 𝑟), 𝑀𝑚,𝑛 (ℂ)) – the set of 𝑚 × 𝑛 matrix-valued func-
𝑟>0
tions holomorphic at (𝜆0 , 𝑧0 ) ∈ ℂ2 .
(iv) As convention we identify 𝒪1×1 ((𝜆0 , 𝑧0 )) with 𝒪((𝜆0 , 𝑧0 )).
(v) If 𝑓1 , 𝑓2 ∈ 𝒪𝑚×𝑛 ((𝜆0 , 𝑧0 )) then we use the notation 𝑓1 = 𝑓2 to mean there exists
an 𝑟 > 0 such that 𝑓1 , 𝑓2 ∈ 𝒪(𝐵((𝜆0 , 𝑧0 ), 𝑟), 𝑀𝑚,𝑛 (ℂ)) and 𝑓1 (𝜆, 𝑧) = 𝑓2 (𝜆, 𝑧) for
∣∣(𝜆, 𝑧) − (𝜆0 , 𝑧0 )∣∣ℂ < 𝑟.
The following is an important characterization of holomorphic functions as analytic functions:
Lemma 17 If 𝑓 ∈ 𝒪𝑚×𝑛 ((𝜆0 , 𝑧0 )) then 𝑓 (𝜆, 𝑧) is analytic at (𝜆, 𝑧) = (𝜆0 , 𝑧0 ), that is,
𝑗1 +𝑗2
the partial and mixed partial derivatives 𝑓 (𝑗1 ,𝑗2 ) (𝜆0 , 𝑧0 ) := ∂∂𝑗1 𝜆∂ 𝑗2𝑓𝑧 (𝜆,𝑧)=(𝜆0 ,𝑧0 ) exist for every 𝑗1 , 𝑗2 ∈ ℕ ∪ {0} and there exists an 𝑟 > 0 such that
𝑓 (𝜆, 𝑧) =
∞
∑
𝑗1 ,𝑗2
1
𝑓 (𝑗1 ,𝑗2 ) (𝜆0 , 𝑧0 )(𝜆 − 𝜆0 )𝑗1 (𝑧 − 𝑧0 )𝑗2 , ∣𝜆 − 𝜆0 ∣ < 𝑟, ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑗
!𝑗
!
=0 1 2
where the power series on the right converges absolutely in the 𝑀𝑚,𝑛 (ℂ) norm to 𝑓 (𝜆, 𝑧) for
(𝜆, 𝑧) ∈ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟).
Proof. The space 𝑀𝑚,𝑛 (ℂ) is finite dimensional and since all norms on a finite dimensional
space are equivalent, we need only show that the statement is true for the norm given by
∣∣[𝑎𝑖,𝑗 ]𝑚,𝑛
[𝑎𝑖,𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∣∣∞ := max ∣𝑎𝑖,𝑗 ∣,
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (ℂ).
𝑖,𝑗
64
Thus it follows from this that 𝒪𝑚×𝑛 ((𝜆0 , 𝑧0 )) = 𝑀𝑚,𝑛 (𝒪((𝜆0 , 𝑧0 ))) and the statement will be
true if we can show that its true for every 𝑓 ∈ 𝒪((𝜆0 , 𝑧0 )). But this statement for functions
in 𝒪((𝜆0 , 𝑧0 )) is a well known fact from the theory of functions of several complex variables
[17, p. 39, Theorem 3.2] and follows from a deep theorem of Hartogs (see [17, p. 21, Theorem
1.6]).
In the spectral perturbation theory of holomorphic matrix functions, the following theorem
plays a central role:
Theorem 18 (Weierstrass Preparation Theorem) If 𝑓 ∈ 𝒪((𝜆0 , 𝑧0 )), 𝑓 (𝜆0 , 𝑧0 ) = 0,
and 𝑓 ∕= 0, then there exists an 𝑟 > 0 such that
𝑓 (𝜆, 𝑧) = (𝑧 − 𝑧0 )𝜇 𝑓0 (𝜆, 𝑧)𝑓1 (𝜆, 𝑧),
𝑓1 (𝜆, 𝑧) ∕= 0, and
𝑓0 (𝜆, 𝑧) = (𝜆 − 𝜆0 )𝑚 + 𝑎1 (𝑧)(𝜆 − 𝜆0 )𝑚−1 + ⋅ ⋅ ⋅ + 𝑎𝑚−1 (𝑧)(𝜆 − 𝜆0 ) + 𝑎𝑚 (𝑧)
(3.23)
(3.24)
for every (𝜆, 𝑧) ∈ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟). Here 𝜇, 𝑚 ∈ ℕ ∪ {0}, 𝑎𝑗 ∈ 𝒪(𝐵(𝑧0 , 𝑟), ℂ) with
𝑎𝑗 (𝑧0 ) = 0 for 𝑗 = 1, . . . , 𝑚 − 1, and 𝑓1 ∈ 𝒪(𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟), ℂ).
Proof. The proof of this statement follows from [17, p. 55, §I.4, Theorem 4.1].
It follows from this that to study the zero set of the function 𝑓 ∈ 𝒪((𝜆0 , 𝑧0 )) near (𝜆0 , 𝑧0 )
we need to study the zero set of the monic polynomial 𝑓0 in 𝜆 (if 𝑚 ≥ 1) given by (3.24),
whose coefficients 𝑎1 , . . . , 𝑎𝑚 are in 𝒪(𝑧0 ) and satisfy 𝑎1 (𝑧0 ) = ⋅ ⋅ ⋅ = 𝑎𝑚 (𝑧0 ) = 0.
Hence we define the set
𝒪0 ((𝜆0 , 𝑧0 )) := {𝑓 ∈ 𝒪((𝜆0 , 𝑧0 )) ∣ 𝑓 has the form in (3.24)}.
65
(3.25)
Let 𝑓 ∈ 𝒪0 ((𝜆0 , 𝑧0 )). Then 𝑓 (𝜆, 𝑧0 ) = (𝜆 − 𝜆0 )𝑚 for some 𝑚 ∈ ℕ ∪ {0}. We call deg 𝑓 := 𝑚
the degree of 𝑓 .
A function 𝑓 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) with 𝑓 ∕= 1 is said to be irreducible if whenever 𝑓 = 𝑝1 𝑝2 with
𝑝1 , 𝑝2 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) then 𝑝1 = 1 or 𝑝2 = 1.
Theorem 19 Every 𝑓 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) with 𝑓 ∕= 1, factors uniquely into a finite product of
irreducibles, i.e., there exists unique irreducibles 𝑝1 , . . . , 𝑝𝑙 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) (not necessarily
distinct) such that
𝑓=
𝑙
∏
𝑝𝑗 .
(3.26)
𝑗=1
Furthermore, deg 𝑓 = deg 𝑝1 + ⋅ ⋅ ⋅ + deg 𝑝𝑙 . Moreover, there exists an 𝑟 > 0 such that for
each 𝑗 = 1, . . . , 𝑙 and for each fixed 𝑧 satisfying 0 < ∣𝑧 − 𝑧0 ∣ < 𝑟, the polynomial 𝑝𝑗 (𝜆, 𝑧) has
only simple roots.
Proof.
The uniqueness of a factorization into irreducibles follows from [4, p. 397, §A.2.6,
Theorem 1] and our choice of notation in A.2.(v). From the definition of the degree of an
element from 𝒪0 ((𝜆0 , 𝑧0 )) we have
(𝜆 − 𝜆0 )
deg 𝑓
= 𝑓 (𝜆, 𝑧0 ) =
𝑙
∏
𝑙
∏
𝑝𝑗 (𝜆, 𝑧0 ) =
(𝜆 − 𝜆0 )deg 𝑝𝑗 = (𝜆 − 𝜆0 )deg 𝑝1 +⋅⋅⋅+deg 𝑝𝑙
𝑗=1
𝑗=1
implying deg 𝑓 = deg 𝑝1 + ⋅ ⋅ ⋅ + deg 𝑝𝑙 . The latter part of this theorem follows from [4, p.
399, §A.3.2, Theorem 2].
Definition 11 Let 𝑞 ∈ ℕ and 𝑟 > 0. If 𝑣 ∈ 𝒪(𝐵(0,
√
𝑞
𝑟), 𝑀𝑚,𝑛 (ℂ)), where
√
𝑞
𝑟 denotes the
real 𝑞th root of 𝑟, and is given by the convergent power series
𝑣(𝜀) =
∞
∑
𝑐𝑗 𝜀𝑗 , ∣𝜀∣ <
𝑗=0
66
√
𝑞
𝑟
(3.27)
then the multi-valued function
1
𝑢(𝑧) := 𝑣((𝑧 − 𝑧0 ) 𝑞 ) =
∞
∑
𝑗
𝑐𝑗 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
(3.28)
𝑗=0
is called a convergent Puiseux series expanded about 𝑧0 with limit point 𝑐0 , domain
𝐵(𝑧0 , 𝑟), and period 𝑞.
Fix any branch of the 𝑞th root function and denote its evaluation at (𝑧 − 𝑧0 ) by (𝑧 − 𝑧0 )1/𝑞
1
(e.g., using the principal branch of the 𝑞th root (𝑧 − 𝑧0 )1/𝑞 = ∣𝑧 − 𝑧0 ∣1/𝑞 𝑒 𝑞 𝑖arg(𝑧−𝑧0 ) where
−𝜋 < arg(𝑧 −𝑧0 ) ≤ 𝜋). Then the single-valued functions 𝑢0 , . . . , 𝑢𝑞−1 given by the convergent
series
ℎ
𝑢ℎ (𝑧) := 𝑣(𝜁 (𝑧 − 𝑧0 )
1/𝑞
)=
∞
∑
𝑐𝑗 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑗 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞 − 1, (3.29)
𝑗=0
where 𝜁 is any primitive 𝑞th root of unity (e.g., 𝜁 = 𝑒𝑖2𝜋/𝑞 ), are called the branches of the
Puiseux series 𝑢. The values of the Puiseux series 𝑢 at a point 𝑧 ∈ 𝐵(𝑧0 , 𝑟) are the values
of its branches at this point, namely, 𝑢0 (𝑧), . . . , 𝑢𝑞−1 (𝑧). The graph of the Puiseux series 𝑢
∪
is the union of the graphs of its branches, namely, 𝑞−1
ℎ=0 {(𝑧, 𝑢ℎ (𝑧)) : ∣𝑧 − 𝑧0 ∣ < 𝑟}.
Theorem 20 Let 𝑓 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) with 𝑓 ∕= 1. Let 𝑓 =
𝑙
∏
𝑝𝑗 be a factorization of 𝑓 into
𝑗=1
unique irreducibles 𝑝1 , . . . , 𝑝𝑙 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) (not necessarily distinct) with deg 𝑓 = deg 𝑝1 +
⋅ ⋅ ⋅ + deg 𝑝𝑙 and 𝑟 > 0 be such that for each 𝑗 = 1, . . . , 𝑙 and for each fixed 𝑧 satisfying
0 < ∣𝑧 − 𝑧0 ∣ < 𝑟, the polynomial 𝑝𝑗 (𝜆, 𝑧) has only simple roots. Then there exists 𝑙 convergent
Puiseux series 𝜆1 (𝑧), . . . , 𝜆𝑙 (𝑧) expanded about 𝑧0 with limit point 𝜆0 , domain 𝐵(𝑧0 , 𝑟), and
periods deg 𝑝1 , . . . , deg 𝑝𝑙 , respectively, such that for every (𝜆, 𝑧) ∈ 𝐵(𝑧0 , 𝑟) × ℂ,
deg 𝑝𝑗 −1
𝑝𝑗 (𝜆, 𝑧) =
∏
(𝜆 − 𝜆𝑗,ℎ (𝑧)),
ℎ=0
67
𝑗 = 1, . . . , 𝑙
(3.30)
and
𝑝𝑗 −1
𝑙 deg∏
∏
𝑓 (𝜆, 𝑧) =
(𝜆 − 𝜆𝑗,ℎ (𝑧)),
𝑗=1
(3.31)
ℎ=0
where 𝜆𝑗,ℎ (𝑧), ℎ = 0, . . . , deg 𝑝𝑗 − 1 denotes all the branches of the Puiseux series 𝜆𝑗 (𝑧), for
𝑗 = 1, . . . , 𝑙.
Proof. This statement follows from [4, p. 406, §A.5.4, Theorem 3].
3.3
On the Perturbation Theory for Holomorphic Matrix Functions
In this section, we consider a holomorphic matrix-valued function of two variables 𝐿 ∈
𝒪𝑛×𝑛 ((𝜆0 , 𝑧0 )) which satisfies det(𝐿(𝜆0 , 𝑧0 )) = 0. For such a function, we can consider
𝐿(𝜆, 𝑧) as a family, indexed by a perturbation parameter 𝑧, of holomorphic matrix functions
in the variable 𝜆. We will give basic facts regarding the dependency on 𝑧 of the eigenvalues
𝜆(𝑧) and corresponding eigenvectors 𝜑(𝑧) of the holomorphic matrix function 𝐿(𝜆, 𝑧) when
(𝜆, 𝑧) is near (𝜆0 , 𝑧0 ). In order to avoid confusion we denote by 𝐿(⋅, 𝑧), the holomorphic
matrix function 𝐿(𝜆, 𝑧) in the variable 𝜆 for a fix 𝑧.
3.3.1
Eigenvalue Perturbations
We begin by defining the spectrum for a holomorphic matrix-valued function of two variables
in a manner similar to the one variable case.
Definition 12 If 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) then the spectrum of 𝐿, denoted by 𝜎(𝐿), is the subset
68
of 𝑈 such that the matrix 𝐿(𝜆, 𝑧) is not invertible, i.e.,
𝜎(𝐿) := {(𝜆, 𝑧) ∈ 𝑈 ∣ det 𝐿(𝜆, 𝑧) = 0}.
(3.32)
Notice that (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿) if and only if 𝜆0 ∈ 𝜎(𝐿(⋅, 𝑧0 )), that is, 𝜆0 is an eigenvalue of the
holomorphic matrix function 𝐿(⋅, 𝑧0 ) ∈ 𝒪(𝜆0 ). Thus the spectral theory developed in the
previous sections for holomorphic matrix functions applies.
Conceptually, we can consider the eigenvalues 𝜆(𝑧) of 𝐿(⋅, 𝑧) as a multi-valued function of 𝑧
implicitly defined by the equation det 𝐿(𝜆, 𝑧) = 0. Thus it is useful to introduce the notation
𝜎(𝐿)−1 for its graph, i.e.,
𝜎(𝐿)−1 := {(𝑧, 𝜆) ∣ (𝜆, 𝑧) ∈ 𝜎(𝐿)}.
(3.33)
With this in mind, the next theorem can be interpreted as a description of the local analytic
properties of this multi-valued function.
Theorem 21 Suppose 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) such that (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿). If 𝜆0 an eigenvalue
of 𝐿(⋅, 𝑧0 ) with algebraic multiplicity 𝑚 then for any 𝜀 > 0 there exists an 𝛿 > 0 such
that 𝜎(𝐿)−1 ∩ 𝐵(𝑧0 , 𝛿) × 𝐵(𝜆0 , 𝜀) is the union of the graphs of 𝑙 convergent Puiseux series
𝜆1 (𝑧), . . . , 𝜆𝑙 (𝑧) expanded about 𝑧0 with limit point 𝜆0 , domain 𝐵(𝑧0 , 𝛿), and periods 𝑞1 , . . . , 𝑞𝑙 ,
respectively, which satisfy 𝑚 = 𝑞1 + ⋅ ⋅ ⋅ + 𝑞𝑙 .
Proof.
First, 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) implies 𝑓 := det(𝐿) ∈ 𝒪(𝑈, ℂ) since the determinant of
a matrix is just a polynomial in its entries. Hence 𝜎(𝐿) is the zero set of the holomorphic
function 𝑓 : 𝑈 → ℂ and since (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿) then 𝑓 (𝜆0 , 𝑧0 ) = 0. Suppose 𝜆0 an eigenvalue of 𝐿(⋅, 𝑧0 ) with algebraic multiplicity 𝑚. Then by Proposition 10 the order of the
zero of 𝑓 (𝜆, 𝑧0 ) = det(𝐿(𝜆, 𝑧0 )) at 𝜆 = 𝜆0 is 𝑚 and so 𝑓 (⋅, 𝑧0 ) ∕= 0. By the Weierstrass
69
Preparation Theorem, Theorem 18 above, there exists 𝑓0 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) with degree 𝑚 and
𝑓1 ∈ 𝒪((𝜆0 , 𝑧0 )) with 𝑓1 (𝜆0 , 𝑧0 ) ∕= 0 such that 𝑓 = 𝑓0 𝑓1 in 𝒪((𝜆0 , 𝑧0 )). By Theorem 19,
there exists unique irreducibles 𝑝1 , . . . , 𝑝𝑙 ∈ 𝒪0 ((𝜆0 , 𝑧0 )) (not necessarily distinct) such that
in 𝒪((𝜆0 , 𝑧0 )) we have
𝑓 = 𝑓1 𝑓0 = 𝑓1
𝑙
∏
𝑝𝑗 .
𝑗=1
Moreover, by the same theorem, deg 𝑓0 = deg 𝑝1 + ⋅ ⋅ ⋅ + deg 𝑝𝑙 and there exists an 𝑟 > 0 such
that for each 𝑗 = 1, . . . , 𝑙 and for each fixed 𝑧 satisfying 0 < ∣𝑧 − 𝑧0 ∣ < 𝑟, the polynomial
𝑝𝑗 (𝜆, 𝑧) has only simple roots. By taking a smaller 𝑟, we may also assume without loss of
generality that 𝑓, 𝑓1 , 𝑝1 , . . . , 𝑝𝑙 ∈ 𝒪(𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟), ℂ) and
𝑓 (𝜆, 𝑧) = 𝑓1 (𝜆, 𝑧)
𝑙
∏
𝑝𝑗 (𝜆, 𝑧),
𝑓1 (𝜆, 𝑧) ∕= 0,
𝑗=1
for every (𝜆, 𝑧) ∈ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟). We now define the numbers 𝑞𝑗 := deg 𝑝𝑗 , 𝑗 = 1, . . . , 𝑙.
Thus 𝑚 = deg 𝑓0 = deg 𝑝1 + ⋅ ⋅ ⋅ + deg 𝑝𝑙 = 𝑞1 + ⋅ ⋅ ⋅ + 𝑞𝑙 . By Theorem 20 there exists 𝑙
convergent Puiseux series 𝜆1 (𝑧), . . . , 𝜆𝑙 (𝑧) expanded about 𝑧0 with limit point 𝜆0 , domain
𝐵(𝑧0 , 𝑟), and periods 𝑞1 , . . . , 𝑞𝑙 , respectively, such that
𝑞𝑗 −1
𝑝𝑗 (𝜆, 𝑧) =
∏
(𝜆 − 𝜆𝑗,ℎ (𝑧)),
𝑗 = 1, . . . , 𝑙
ℎ=0
where 𝜆𝑗,ℎ (𝑧), ℎ = 0, . . . , 𝑞𝑗 − 1 denotes all the branches of the Puiseux series 𝜆𝑗 (𝑧), for
𝑗 = 1, . . . , 𝑙. This implies
𝑗 −1
𝑙 𝑞∏
∏
det(𝐿(𝜆, 𝑧)) = 𝑓1 (𝜆, 𝑧)
(𝜆 − 𝜆𝑗,ℎ (𝑧)),
𝑗=1 ℎ=0
for every (𝜆, 𝑧) ∈ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟).
70
𝑓1 (𝜆, 𝑧) ∕= 0,
Let 𝜀 > 0 be given. Then since
lim 𝜆𝑗,ℎ (𝑧) = 𝜆𝑗,ℎ (𝑧0 ) = 𝜆0 ,
𝑧→𝑧0
ℎ = 0, . . . , 𝑞𝑗 − 1,
𝑗 = 1, . . . , 𝑙,
ℎ = 0, . . . , 𝑞𝑗 − 1,
𝑗 = 1, . . . , 𝑙,
there exists 𝛿 > 0 with 𝛿 ≤ 𝑟 such that
∣𝜆𝑗,ℎ (𝑧) − 𝜆0 ∣ < min{𝜀, 𝑟},
if ∣𝑧 − 𝑧0 ∣ < 𝛿. Then, since 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝛿) ⊆ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟), we have
𝜎(𝐿) ∩ 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝛿) = 𝜎(𝐿) ∩ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟) ∩ 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝛿)
𝑗 −1
𝑙 𝑞∪
∪
=
{(𝜆𝑗,ℎ (𝑧), 𝑧) : (𝜆𝑗,ℎ (𝑧), 𝑧) ∈ 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟)} ∩ 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝛿)
𝑗=1 ℎ=0
𝑗 −1
𝑙 𝑞∪
∪
{(𝜆𝑗,ℎ (𝑧), 𝑧) ∣ 𝑧 ∈ 𝐵(𝑧0 , 𝛿) }
=
𝑗=1 ℎ=0
It follows from this that if we just take the convergent Puiseux series 𝜆1 (𝑧), . . . , 𝜆𝑙 (𝑧) to have
the domain 𝐵(𝑧0 , 𝛿) instead of 𝐵(𝑧0 , 𝑟) then we have 𝑙 convergent Puiseux series expanded
about 𝑧0 with limit point 𝜆0 , domain 𝐵(𝑧0 , 𝛿), and periods 𝑞1 , . . . , 𝑞𝑙 , respectively, which
satisfy 𝑚 = deg 𝑞1 + ⋅ ⋅ ⋅ + deg 𝑞𝑙 and such that the union of their graphs is
𝑗 −1
𝑙 𝑞∪
∪
{(𝑧, 𝜆𝑗,ℎ (𝑧)) ∣ 𝑧 ∈ 𝐵(𝑧0 , 𝛿) } = {(𝑧, 𝜆) ∣ (𝜆, 𝑧) ∈ 𝜎(𝐿) ∩ 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝛿)}
𝑗=1 ℎ=0
= 𝜎(𝐿)−1 ∩ 𝐵(𝑧0 , 𝛿) × 𝐵(𝜆0 , 𝜀).
This completes the proof.
Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and assume that (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿) and 𝜆0 an eigenvalue of 𝐿(⋅, 𝑧0 ) with
finite algebraic multiplicity. Denote by 𝑔, 𝑚1 ≥ . . . ≥ 𝑚𝑔 , and 𝑚 = 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 the
geometric multiplicity, partial multiplicities, and algebraic multiplicity, respectively of the
71
eigenvalue 𝜆0 of 𝐿(⋅, 𝑧0 ). Then it follows Theorem 21 that given an 𝜀 > 0 there exists a 𝛿 > 0
such that, for ∣𝑧 −𝑧0 ∣ < 𝛿, the spectrum of 𝐿(𝜆, 𝑧) in ∣𝜆−𝜆0 ∣ < 𝜀, that is, 𝜎(𝐿(⋅, 𝑧))∩𝐵(𝜆0 , 𝜀),
consists of 𝑚 eigenvalues and they are the values of 𝑙 convergent Puiseux series of the form
𝜆𝑗 (𝑧) = 𝜆0 +
∞
∑
𝑠
𝑐𝑗,𝑠 (𝑧 − 𝑧0 ) 𝑞𝑗 , ∣𝑧 − 𝑧0 ∣ < 𝛿,
𝑗 = 1, . . . , 𝑙
(3.34)
𝑠=1
where there periods satisfy 𝑚 = 𝑞1 + . . . + 𝑞𝑙 . Specifically, letting 𝜆𝑗,ℎ (𝑧), ℎ = 0, . . . , 𝑞𝑗 − 1
denote all the branches of the Puiseux series 𝜆𝑗 (𝑧), for 𝑗 = 1, . . . , 𝑙 then
𝑙
∪
𝜎(𝐿(⋅, 𝑧)) ∩ 𝐵(𝜆0 , 𝜀) =
{𝜆𝑗,ℎ (𝑧) : ℎ = 0, . . . , 𝑞𝑗 − 1}
(3.35)
𝑗=1
for each 𝑧 ∈ 𝐵(𝑧0 , 𝛿).
In general the only connection between the numbers 𝑙 and 𝑞1 , . . . , 𝑞𝑙 with the geometric mul∑𝑔
∑𝑙
tiplicity 𝑔 and partial multiplicities 𝑚1 , . . . , 𝑚𝑔 is the equality 𝑚 =
𝑗=1 𝑚𝑗 .
𝑗=1 𝑞𝑗 =
Because of this two important problems arise in the spectral perturbation theory of holomorphic matrix functions:
1. Find general conditions involving just the partial derivatives of 𝐿 at (𝜆0 , 𝑧0 ) up to first
order, namely, 𝐿(𝜆0 , 𝑧0 ), 𝐿𝜆 (𝜆0 , 𝑧0 ), and 𝐿𝑧 (𝜆0 , 𝑧0 ) such that we may determine the
values 𝑙, 𝑞1 , . . . , 𝑞𝑙 and, in particular, those conditions which imply 𝑙 = 𝑔 and 𝑞𝑗 = 𝑚𝑗 ,
for 𝑗 = 1, . . . , 𝑔.
2. Find formulas for the first order coefficients of those Puiseux series, i.e., formulas for
coefficient 𝑐𝑗,1 , 𝑗 = 1, . . . , 𝑙 in (3.34).
Several papers have addressed these problems (see [23, 34, 37–39]) and significant progress
was made towards their resolution.
72
3.3.2
Eigenvector Perturbations
We will now proceed to give a useful description for the dependency on 𝑧 of the perturbed
eigenvectors 𝜑(𝑧) of the holomorphic matrix function 𝐿(𝜆, 𝑧) corresponding to perturbed
eigenvalue 𝜆(𝑧) when (𝜆, 𝑧) is near a point (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿).
Definition 13 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)). We say 𝜆(𝑧) is a eigenvalue Puiseux series of
𝐿(⋅, 𝑧) provided it is a convergent Puiseux series whose values at each point 𝑧 in its domain
are eigenvalues of 𝐿(⋅, 𝑧). If
𝜆(𝑧) = 𝜆0 +
∞
∑
𝑠
𝑐𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝛿
(3.36)
𝑠=1
is a eigenvalue Puiseux series of 𝐿(⋅, 𝑧) then we say 𝜑(𝑧) is an eigenvector Puiseux series
of 𝐿(⋅, 𝑧) corresponding to the eigenvalue Puiseux series 𝜆(𝑧) provided it is a convergent
Puiseux series with
𝜑(𝑧) = 𝛽 0 +
∞
∑
𝑠
𝛽 𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟
(3.37)
𝑠=1
𝑛
where 0 < 𝑟 ≤ 𝛿, {𝛽 𝑠 }∞
𝑠=0 ⊆ ℂ , 𝛽 0 ∕= 0, and such that if we fix any branch of the 𝑞th root
function and denote its evaluation at (𝑧 − 𝑧0 ) by (𝑧 − 𝑧0 )1/𝑞 then their branches
𝜆ℎ (𝑧) = 𝜆0 +
𝜑ℎ (𝑧) = 𝛽 0 +
∞
∑
𝑠=1
∞
∑
𝑐𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞,
(3.38)
𝛽 𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞,
(3.39)
𝑠=1
where 𝜁 is any primitive 𝑞th root of unity, satisfy
𝐿(𝜆ℎ (𝑧), 𝑧)𝜑ℎ (𝑧) = 0,
𝜑ℎ (𝑧) ∕= 0,
73
ℎ = 0, . . . , 𝑞 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟.
(3.40)
We call the pair (𝜆(𝑧), 𝜑(𝑧)) an eigenpair Puiseux series of 𝐿(⋅, 𝑧).
The following lemma tells us that to every eigenvalue Puiseux series there exists a corresponding eigenvector Puiseux series.
Lemma 22 Suppose 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)). If 𝜆(𝑧) is a eigenvalue Puiseux series of 𝐿(⋅, 𝑧) then
there exists an eigenvector Puiseux series 𝜑(𝑧) of 𝐿(⋅, 𝑧) corresponding to the eigenvector
Puiseux series 𝜆(𝑧).
Proof.
Assume 𝜆(𝑧) be an eigenvalue Puiseux series of 𝐿(⋅, 𝑧). This implies for some
√
𝑞 ∈ ℕ, 𝛿 > 0, 𝑣 ∈ 𝒪(𝐵(0, 𝑞 𝛿), ℂ), and some (𝜆0 , 𝑧0 ) ∈ 𝑈 , 𝜆(𝑧) is a convergent Puiseux series
with
1
𝑞
𝜆(𝑧) = 𝑣((𝑧 − 𝑧0 ) ) = 𝜆0 +
∞
∑
𝑠
𝑐𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝛿
(3.41)
𝑠=1
and taking any fixed branch of the 𝑞th root and denoting its evaluation at 𝑧−𝑧0 by (𝑧−𝑧0 )1/𝑞 ,
its branches
ℎ
1/𝑞
𝜆ℎ (𝑧) = 𝑣(𝜁 (𝑧 − 𝑧0 )
) = 𝜆0 +
∞
∑
𝑐𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞, (3.42)
𝑠=1
where 𝜁 is any primitive 𝑞th root of unity, satisfy (𝜆ℎ (𝑧), 𝑧) ∈ 𝑈 for ∣𝑧 − 𝑧0 ∣ < 𝛿 and
det 𝐿(𝜆ℎ (𝑧), 𝑧) = 0, ∣𝑧 − 𝑧0 ∣ < 𝛿,
ℎ = 0, . . . , 𝑞 − 1,
(3.43)
since the values 𝜆ℎ (𝑧), ℎ = 0, . . . , 𝑞−1 for ∣𝑧−𝑧0 ∣ < 𝛿 are eigenvalues of 𝐿(⋅, 𝑧) by assumption.
√
We introduce a new variable 𝜀 ∈ 𝐵(0, 𝑞 𝛿). Then 𝑧(𝜀) := 𝜀𝑞 + 𝑧0 ∈ 𝐵(𝑧0 , 𝛿) is an analytic
√
function. Moreover, for each 𝜀 ∈ 𝐵(0, 𝑞 𝛿) there exists an ℎ ∈ {0, . . . , 𝑞 − 1} such that 𝜀 =
𝜁 ℎ (𝑧(𝜀)−𝑧0 )1/𝑞 . This implies (𝑣(𝜀), 𝑧(𝜀)) = (𝜆ℎ (𝑧(𝜀)), 𝑧(𝜀)) ∈ 𝑈 and hence det 𝐿(𝑣(𝜀), 𝑧(𝜀)) =
74
det 𝐿(𝜆ℎ (𝑧), 𝑧) = 0. Thus we have shown for every 𝜀 ∈ 𝐵(0,
√
𝑞
𝑟), the analytic functions 𝑣(𝜀)
and 𝑧(𝜀) satisfy (𝑣(𝜀), 𝑧(𝜀)) ∈ 𝑈 and so have ranges in the domain of the analytic function
𝐿 and satisfy
det 𝐿(𝑣(𝜀), 𝑧(𝜀)) = 0.
(3.44)
√
𝑞
𝛿) → 𝑀𝑛 (ℂ) defined by 𝐹 (𝜀) := 𝐿(𝑣(𝜀), 𝑧(𝜀)) is an
√
√
analytic function as well, i.e., 𝐹 ∈ 𝒪(𝐵(0, 𝑞 𝛿), 𝑀𝑛 (ℂ)), and hence det 𝐹 ∈ 𝒪(𝐵(0, 𝑞 𝛿), ℂ)
But this implies the function 𝐹 : 𝐵(0,
with
det 𝐹 = 0.
(3.45)
But this implies that 0 is an eigenvalue of the holomorphic matrix function 𝐹 of infinite
algebraic multiplicity. Consider the local Smith form 𝐺 of 𝐹 corresponding to the eigenvalue
0 of infinite algebraic multiplicity. By Theorem 9 it must be of the form
′
′
𝐺(𝜀) = diag{0, . . . , 0, 𝜀𝑚𝜇+1 , . . . , 𝜀𝑚𝑔 , 1, . . . , 1}
(3.46)
where 1 ≤ 𝜇 ≤ 𝑔 = dim ker 𝐹 (0) = dim ker 𝐿(𝜆0 , 𝑧0 ) ≤ 𝑛, the number of zeros down the
diagonal is 1 ≤ 𝜇 = dim ker 𝐹 (𝜀) = dim ker 𝐿(𝑣(𝜀), 𝑧(𝜀)) for 0 < 𝜀 ≪ 1, when 𝜇 = 𝑔
there are no 𝜀 terms down the diagonal, when 𝜇 < 𝑔 we have 𝑚′𝜇+1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚′𝑔 ≥ 1 with
{𝑚′𝜇+𝑗 }𝑔𝑗=1 ⊆ ℕ, and when 𝑔 = 𝑛 there are no ones appearing in the diagonal. Hence by
√
√
Theorem 9 and Lemma 12 there exists an 𝑟 > 0 with 0 < 𝑞 𝑟 ≤ 𝑞 𝛿 and matrix functions
√
𝑀, 𝑁 ∈ 𝒪(𝐵(0, 𝑞 𝑟), 𝑀𝑛 (ℂ)) such that 𝑀 (𝜀), 𝑁 (𝜀) are invertible matrices for every 𝜀 ∈
√
𝐵(0, 𝑞 𝑟) and satisfy
′
′
diag{0, . . . , 0, 𝜀𝑚𝜇+1 , . . . , 𝜀𝑚𝑔 , 1, . . . , 1} = 𝑁 (𝜀)𝐹 (𝜀)𝑀 (𝜀)
= 𝑁 (𝜀)𝐿(𝑣(𝜀), 𝑧(𝜀))𝑀 (𝜀), ∣𝜀∣ <
75
√
𝑞
𝑟.
(3.47)
Denote the standard orthonormal basis vectors in ℂ𝑛 by 𝑒1 , . . . , 𝑒𝑛 . Then
𝜙𝑗 (𝜀) := 𝑀 (𝜀)𝑒𝑗 ,
𝑗 = 1, . . . , 𝜇
(3.48)
√
are the first 𝜇 columns of 𝑀 (𝜀) and so are a basis for ker 𝐿(𝑣(𝜀), 𝑧(𝜀)) for every 𝜀 ∈ 𝐵(0, 𝑞 𝑟).
√
√
Hence for 𝑗 = 1, . . . , 𝜇 we have 𝜙𝑗 ∈ 𝒪(𝐵(0, 𝑞 𝑟), ℂ𝑛 ) and 𝜙𝑗 (𝜀) ∕= 0 for every 𝜀 ∈ 𝐵(0, 𝑞 𝑟).
In particular, 𝜙1 ∈ 𝒪(𝐵(0,
√
𝑞
𝑟), ℂ𝑛 ) and satisfies
𝜑1 (𝜀) ∕= 0, ∣𝜀∣ <
𝐿(𝑣(𝜀), 𝑧(𝜀))𝜑1 (𝜀) = 0,
√
𝑞
𝑟.
(3.49)
√
𝑞
𝑟), ℂ𝑛 ) it follows that it has a power series expansion about the
√
point 𝑧0 which converges for ∣𝜀∣ < 𝑞 𝑟. Thus we have
Now since 𝜙1 ∈ 𝒪(𝐵(0,
𝜙1 (𝑧) = 𝛽 0 +
∞
∑
𝛽 𝑠 𝜀𝑠 , ∣𝜀∣ <
√
𝑞
𝑟.
(3.50)
𝑠=1
𝑛
𝑛
where {𝛽 𝑠 }∞
𝑠=0 ⊆ ℂ , 𝛽 0 ∕= 0, and the series converges absolutely in the ℂ norm for every
√
𝜀 ∈ 𝐵(0, 𝑞 𝑟). It follows that a convergent Puiseux series 𝜑(𝑧) is given by
1
𝑞
𝜑(𝑧) = 𝜙1 ((𝑧 − 𝑧0 ) ) = 𝛽 0 +
∞
∑
𝑠
𝛽 𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟.
(3.51)
𝑠=1
We will now show that 𝜑(𝑧) is an eigenvector Puiseux series of 𝐿(⋅, 𝑧) corresponding to
the eigenvalue Puiseux series 𝜆(𝑧). Fix any branch of the 𝑞th root function denoting its
evaluation at (𝑧 − 𝑧0 ) by (𝑧 − 𝑧0 )1/𝑞 and let 𝜁 be any primitive 𝑞th root of unity. Now if
√
√
∣𝑧 − 𝑧0 ∣ < 𝑟 then 𝜀ℎ (𝑧) := 𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 ∈ 𝐵(0, 𝑞 𝑟) ⊆ 𝐵(0, 𝑞 𝛿) for ℎ = 0, . . . , 𝑞 and so it
follows that the branches of 𝜆(𝑧) and 𝜑(𝑧) are given by the convergent series
𝜆ℎ (𝑧) = 𝜆0 +
∞
∑
𝑐𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
𝑠=1
76
ℎ = 0, . . . , 𝑞,
(3.52)
𝜑ℎ (𝑧) = 𝛽 0 +
∞
∑
𝛽 𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞.
(3.53)
𝑠=1
Moreover, if ∣𝑧 − 𝑧0 ∣ < 𝑟 then we have 𝜀ℎ (𝑧) = 𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 ∈ 𝐵(0,
√
𝑞
𝑟) ⊆ 𝐵(0,
√
𝑞
𝛿) implies
𝑧 = 𝑧(𝜀ℎ (𝑧)), 𝜆ℎ (𝑧) = 𝑣(𝜀ℎ (𝑧)), (𝑣(𝜀ℎ (𝑧)), 𝑧(𝜀ℎ (𝑧))) = (𝜆ℎ (𝑧), 𝑧) ∈ 𝑈 , 𝜑ℎ (𝑧) = 𝜙1 (𝜀ℎ (𝑧)) ∕= 0,
and
𝐿(𝜆ℎ (𝑧), 𝑧)𝜑ℎ (𝑧) = 𝐿(𝑣(𝜀ℎ (𝑧)), 𝑧(𝜀ℎ (𝑧)))𝜙1 (𝜀ℎ (𝑧)) = 0
(3.54)
for ℎ = 0, . . . , 𝑞 − 1. Therefore 𝜑(𝑧) is an eigenvector Puiseux series of 𝐿(⋅, 𝑧) corresponding
to the eigenvalue Puiseux series 𝜆(𝑧). This completes the proof.
Theorem 23 Suppose 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) such that (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿). If 𝜆0 an eigenvalue of
𝐿(⋅, 𝑧0 ) with algebraic multiplicity 𝑚 then for any 𝜀 > 0 there exists a 𝑟 > 0 and 𝑙 eigenvalue
Puiseux series (𝜆1 (𝑧), 𝜑1 (𝑧)), . . . , (𝜆𝑙 (𝑧), 𝜑𝑙 (𝑧)) expanded about 𝑧0 with domain 𝐵(𝑧0 , 𝑟) and
periods 𝑞1 , . . . , 𝑞𝑙 , respectively, satisfying 𝑚 = 𝑞1 +⋅ ⋅ ⋅+𝑞𝑙 such that if we fix any branch of the
𝑞th root function, for 𝑞 ∈ {𝑞1 , . . . , 𝑞𝑙 }, denoting its evaluation at (𝑧 − 𝑧0 ) by (𝑧 − 𝑧0 )1/𝑞 and
let 𝜁 𝑞 be any primitive 𝑞th root of unity then their branches given by the convergent series
𝜆𝑗,ℎ (𝑧) = 𝜆0 +
∞
∑
𝑐𝑗,𝑠 (𝜁 ℎ𝑞𝑗 (𝑧 − 𝑧0 )1/𝑞𝑗 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟, ℎ = 0, . . . , 𝑞𝑗 − 1,
(3.55)
𝛽 𝑗,ℎ (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞𝑗 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟,
(3.56)
𝑠=1
𝜑𝑗,ℎ (𝑧) = 𝛽 𝑗,0 +
∞
∑
ℎ = 0, . . . , 𝑞𝑗 − 1.
𝑠=1
satisfy
det 𝐿(𝜆𝑗,ℎ (𝑧), 𝑧) = 0,
ℎ = 0, . . . , 𝑞𝑗 − 1,
𝜎(𝐿) ∩ 𝐵(𝜆0 , 𝜀) × 𝐵(𝑧0 , 𝑟) =
𝑗 −1
𝑙 𝑞∪
∪
𝑗 = 1, . . . , 𝑙,
∣𝑧 − 𝑧0 ∣ < 𝑟
{(𝜆𝑗,ℎ (𝑧), 𝑧) : ∣𝑧 − 𝑧0 ∣ < 𝑟}
(3.57)
(3.58)
𝑗=1 ℎ=1
𝜎(𝐿(⋅, 𝑧)) ∩ 𝐵(𝜆0 , 𝜀) =
𝑗 −1
𝑙 𝑞∪
∪
{𝜆𝑗,ℎ (𝑧)}, ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑗=1 ℎ=1
77
(3.59)
𝐿(𝜆ℎ (𝑧), 𝑧)𝜑ℎ (𝑧) = 0,
𝜑ℎ (𝑧) ∕= 0,
ℎ = 0, . . . , 𝑞𝑗 − 1,
𝑗 = 1, . . . , 𝑙,
∣𝑧 − 𝑧0 ∣ < 𝑟. (3.60)
Proof. This theorem follows directly from Theorem 21 and Lemma 22.
Now in general the zero-order eigenvectors {𝛽 𝑗,0 }𝑔𝑗=1 ⊆ ker 𝐿(𝜆0 , 𝑧0 ) need not be a basis for
ker 𝐿(𝜆0 , 𝑧0 ), the eigenspace of 𝐿(⋅, 𝑧0 ) corresponding to the eigenvalue 𝜆0 . Because of this
two important problems arise in the spectral perturbation theory of holomorphic matrix
functions:
1. Find general conditions involving just the partial derivatives of 𝐿 at (𝜆0 , 𝑧0 ) up to first
order, namely, 𝐿(𝜆0 , 𝑧0 ), 𝐿𝜆 (𝜆0 , 𝑧0 ), and 𝐿𝑧 (𝜆0 , 𝑧0 ) such that we may determine whether
or not we can choose eigenpair Puiseux series (𝜆1 (𝑧), 𝜑1 (𝑧)), . . . , (𝜆𝑙 (𝑧), 𝜑𝑙 (𝑧)) of 𝐿(⋅, 𝑧)
as given in the above theorem so that their zero-order eigenvectors {𝜑𝑗 (𝑧0 )}𝑔𝑗=1 span
the whole eigenspace ker 𝐿(𝜆0 , 𝑧0 ).
2. Find formulas to determine the zero-order eigenvectors {𝛽 𝑗,0 }𝑔𝑗=1 .
The literature which addresses these problems is scarce and this author is only aware of
[23, 34] which addresses these problems for general perturbations of holomorphic matrix
functions. Even in the special case 𝐿(𝜆, 𝑧) = 𝐴(𝑧) − 𝜆𝐼, few results exist with the exception
of [4, p. 270, Theorem 4; pp. 310-311, Theorem 8, Corollary 2, & Corollary 3], [41, Theorem
2], [58, p. 64, Chap. 2, §7.10, Theorem 6], and [47, Theorem 2.2]. In particular, the second
problem remains open even in this case, when total splitting at the first-order approximation
(see [4, p. 311] for definition) does not occur, and seems a difficult problem to resolve.
78
3.3.3
Analytic Eigenvalues and Eigenvectors
In this section we give a survey of the main results from [23, 34] that we will need in our
thesis.
Preliminaries. We first begin with some preliminaries.
Definition 14 Let 𝐹 ∈ 𝒪(𝑈, 𝑀𝑚,𝑛 (ℂ)) where 𝑈 ⊆ ℂ is an open connected set. Define
𝑈 ∗ := {𝜆 ∣ 𝜆 ∈ 𝑈 }. The matrix function 𝐹 ∗ : 𝑈 ∗ → 𝑀𝑚,𝑛 (ℂ) defined by 𝐹 ∗ (𝜆) = 𝐹 (𝜆)∗ is
called the adjoint of the holomorphic matrix function 𝐹 .
Lemma 24 If 𝐹 ∈ 𝒪(𝑈, 𝑀𝑚,𝑛 (ℂ)) where 𝑈 ⊆ ℂ is an open connected set then 𝑈 ∗ is an
open connected set and 𝐹 ∗ ∈ 𝒪(𝑈 ∗ , 𝑀𝑚,𝑛 (ℂ)).
Proof.
If 𝐹 ∈ 𝒪(𝑈, 𝑀𝑚,𝑛 (ℂ)) then 𝐹 = [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 where 𝑎𝑖𝑗 ∈ 𝒪(𝑈, ℂ). Now obviously
𝑈 ∗ is an open connected set and from complex analysis we know 𝑎∗𝑖𝑗 (𝜆) := 𝑎𝑖𝑗 (𝜆) belongs to
∗
𝒪(𝑈 ∗ , ℂ). But this implies 𝐹 ∗ = [𝑎∗𝑗𝑖 ]𝑛,𝑚
𝑖=1,𝑗=1 ∈ 𝒪(𝑈 , 𝑀𝑚,𝑛 (ℂ)). This completes the proof.
Proposition 25 Let 𝐹 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)). Then
1. 𝜆0 ∈ ℂ is an eigenvalue of 𝐹 if and only if 𝜆0 is an eigenvalue of 𝐹 ∗ .
2. 𝜆0 is an eigenvalue of 𝐹 of finite algebraic multiplicity if and only if 𝜆0 is an eigenvalue
of 𝐹 ∗ of finite algebraic multiplicity.
3. If 𝜆0 is an eigenvalue of 𝐹 of finite algebraic multiplicity then the geometric, partial,
and algebraic multiplicities of 𝜆0 coincide with the geometric, partial, and algebraic
multiplicities, respectively, of the eigenvalue 𝜆0 of 𝐹 ∗ .
79
Proof. The first and second properties follows from the equality det 𝐹 (𝜆0 ) = det 𝐹 (𝜆0 )∗ =
det 𝐹 ∗ (𝜆0 ) for any 𝜆0 ∈ 𝑈 . To prove the third part let 𝜆0 be an eigenvalue of 𝐹 of finite
algebraic multiplicity and let 𝑔, 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 , 𝑚 := 𝑚1 + . . . + 𝑚𝑔 be its geometric, partial,
and algebraic multiplicities, respectively. Then by Theorem 9 there exists 𝑁, 𝑀 ∈ 𝒪(𝜆0 )
and an 𝑟 > 0 such that 𝑀 (𝜆0 ), 𝑁 (𝜆0 ) are invertible and for every 𝜆 ∈ 𝐵(𝜆0 , 𝑟) we have
𝐹 (𝜆) = 𝑁 (𝜆) diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}𝑀 (𝜆).
where if 𝑔 = 𝑛 there are no ones appearing in the diagonal. But this implies for any
𝜆 ∈ 𝐵(𝜆0 , 𝑟) we have
𝐹 ∗ (𝜆) = 𝑀 ∗ (𝜆) diag{(𝜆 − 𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1}𝑁 ∗ (𝜆).
where if 𝑔 = 𝑛 there are no ones appearing in the diagonal. And since 𝑀 ∗ , 𝑁 ∗ ∈ 𝒪(𝜆0 ) with
𝑀 ∗ (𝜆0 )) = 𝑀 (𝜆0 )∗ , 𝑁 ∗ (𝜆0 )) = 𝑁 (𝜆0 )∗ are invertible, this implies that 𝐺∗ (𝜆) = diag{(𝜆 −
𝜆0 )𝑚1 , . . . , (𝜆 − 𝜆0 )𝑚𝑔 , 1, . . . , 1} is the local Smith form of 𝐹 ∗ corresponding to the eigenvalue
𝜆0 . Therefore, 𝑔, 𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 , 𝑚 are the geometric, partial, and algebraic multiplicities,
respectively, of the eigenvalue 𝜆0 of 𝐹 ∗ . This completes the proof.
Definition 15 Let 𝑈 ⊆ ℂ2 . Define 𝑈 ∗ := {(𝜉, 𝜂) ∣ (𝜉, 𝜂) ∈ 𝑈 }. If 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ))
where 𝑈 is an open connected set then the matrix function 𝐿∗ : 𝑈 ∗ → 𝑀𝑛 (ℂ) defined by
𝐿∗ (𝜉, 𝜂) = 𝐿(𝜉, 𝜂)∗ is called the adjoint of the holomorphic matrix function 𝐿.
Lemma 26 If 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) where 𝑈 ⊆ ℂ2 is an open connected set then 𝑈 ∗ is an open
connected set and 𝐿∗ ∈ 𝒪(𝑈 ∗ , 𝑀𝑛 (ℂ)).
Proof.
Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) where 𝑈 ⊆ ℂ2 is an open connected set. Obviously 𝑈 ∗ is
an open connected set. The adjoint of 𝐿 is 𝐿∗ (𝜉, 𝜂) := 𝐿(𝜉, 𝜈). Now for any (𝜉 0 , 𝜂 0 ) ∈ 𝑈 ∗
80
there exist (𝜆0 , 𝑧0 ) ∈ 𝑈 such that (𝜉 0 , 𝜂 0 ) = (𝜆0 , 𝑧0 ). We define 𝐹1 (𝜉) := 𝐿(𝜉, 𝑧0 ) = 𝐿∗ (𝜉, 𝜂 0 )
and 𝐹2 (𝜂) := 𝐿(𝜆0 , 𝜂) = 𝐿∗ (𝜉 0 , 𝜂). It follows from the one variable case that 𝐹1 ∈ 𝒪(𝜉 0 ) and
𝐹2 ∈ 𝒪(𝜂 0 ). This implies the partial derivatives of 𝐿∗ (𝜉, 𝜂) exist at (𝜉 0 , 𝜂 0 ). But since this
is true for every (𝜉 0 , 𝜂 0 ) ∈ 𝑈 ∗ then 𝐿∗ ∈ 𝒪(𝑈 ∗ , ℂ). This completes the proof.
Lemma 27 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)). Then 𝜆(𝑧) ∈ 𝜎(𝐿(⋅, 𝑧)) if and only if 𝜆(𝜂) ∈ 𝜎(𝐿∗ (⋅, 𝜂)),
where 𝜂 = 𝑧. Moreover,
𝜎(𝐿∗ ) = 𝜎(𝐿)∗ .
Proof.
(3.61)
If 𝜆(𝑧) ∈ 𝜎(𝐿(⋅, 𝑧)) then 0 = det 𝐿(𝜆(𝑧), 𝑧) = det 𝐿(𝜆(𝑧), 𝑧)∗ = det 𝐿∗ (𝜆(𝑧), 𝑧) =
det 𝐿∗ (𝜆(𝜂), 𝜂) where 𝜂 = 𝑧. Hence 𝜆(𝜂) ∈ 𝜎(𝐿∗ (⋅, 𝜂)). Conversely, if 𝜆(𝜂) ∈ 𝜎(𝐿∗ (⋅, 𝜂))
then letting 𝜂 = 𝑧 we have 0 = det 𝐿∗ (𝜆(𝜂), 𝜂) = det 𝐿∗ (𝜆(𝑧), 𝑧) = det 𝐿(𝜆(𝑧), 𝑧)∗ =
det 𝐿(𝜆(𝑧), 𝑧). Hence 𝜆(𝑧) ∈ 𝜎(𝐿(⋅, 𝑧)). The latter part of this statement follows from
the identity det 𝐿∗ (𝜉, 𝜂) = det 𝐿(𝜉, 𝜂)∗ = det 𝐿(𝜉, 𝜂) for any (𝜉, 𝜂) ∈ 𝑈 ∗ . This completes the
proof.
Definition 16 Suppose that 𝜆(𝑧) is a convergent Puiseux series given by
𝜆(𝑧) =
∞
∑
𝑠
𝑐𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟
(3.62)
𝑠=0
where {𝑐𝑠 }∞
𝑠=0 ⊆ ℂ. Then a convergent Puiseux series is given by
𝜆∗ (𝜂) :=
∞
∑
𝑠
𝑐𝑠 (𝜂 − 𝑧0 ) 𝑞 , ∣𝜂 − 𝑧0 ∣ < 𝑟.
(3.63)
𝑠=0
We call the Puiseux series 𝜆∗ (𝜂) the adjoint of the Puiseux series 𝜆(𝑧).
Lemma 28 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)). Then 𝜆(𝑧) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑧) if
81
and only if 𝜆∗ (𝜂) is an eigenvalue Puiseux series of 𝐿∗ (⋅, 𝜂), where 𝜂 = 𝑧. Moreover, if 𝜆ℎ (𝑧),
ℎ = 0, . . . , 𝑞 − 1 are all the branches of 𝜆(𝑧) and 𝜆∗ℎ (𝜂), ℎ = 0, . . . , 𝑞 − 1 are all the branches
of 𝜆∗ (𝜂) then for 𝜂 = 𝑧 we have
𝑞−1
𝑞−1
∪
{𝜆∗ℎ (𝜂)}
ℎ=0
𝑞−1
∪
=
∪
{𝜆ℎ (𝜂)},
(3.64)
ℎ=0
𝑞−1
{𝜆ℎ (𝑧)} =
ℎ=0
∪
{𝜆∗ℎ (𝑧)}.
(3.65)
ℎ=0
Proof. Its obvious that
𝜆(𝑧) = 𝜆0 +
∞
∑
𝑠
𝑐𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟
(3.66)
𝑠=1
is a convergent Puiseux series if and only if its adjoint
∗
𝜆 (𝜂) := 𝜆0 +
∞
∑
𝑠
𝑐𝑠 (𝜂 − 𝑧0 ) 𝑞 , ∣𝜂 − 𝑧0 ∣ < 𝑟
(3.67)
𝑠=1
is a convergent Puiseux series since (𝜆∗ )∗ (𝑧) = 𝜆(𝑧). Denote the principal branch of the 𝑞th
√
1
√
root by 𝑞 𝑧 − 𝑧0 := 𝑞 ∣𝑧 − 𝑧0 ∣𝑒 𝑞 𝑖arg(𝑧−𝑧0 ) for 0 ≤ arg(𝑧 − 𝑧0 ) ≤ 2𝜋 and let 𝜁 := 𝑒𝑖2𝜋/𝑞 . Denote
√
1
another branch of the 𝑞th root by (𝑧 −𝑧0 )1/𝑞 := 𝑞 ∣𝑧 − 𝑧0 ∣𝑒 𝑞 𝑖arg(𝑧−𝑧0 ) for −𝜋 ≤ arg(𝑧 −𝑧0 ) < 𝜋
and notice 𝜁 = 𝑒−𝑖2𝜋/𝑞 is a primitive 𝑞th root of unity. A quick calculation shows
√
𝑞
𝑧 − 𝑧0 = (𝑧 − 𝑧0 )1/𝑞 ,
√
(𝑧 − 𝑧0 )1/𝑞 = 𝑞 𝑧 − 𝑧0 ,
for every 𝑧 ∈ ℂ.
82
(3.68)
(3.69)
Now all the branches of 𝜆(𝑧) and 𝜆∗ (𝜂) are given by the convergent series
𝜆ℎ (𝑧) = 𝜆0 +
∞
∑
√
𝑐𝑠 (𝜁 ℎ 𝑞 𝑧 − 𝑧0 )𝑠 ,
ℎ = 0, . . . , 𝑞 − 1 ∣𝑧 − 𝑧0 ∣ < 𝑟,
𝑠=1
𝜆∗ℎ (𝜂) = 𝜆0 +
∞
∑
ℎ
𝑐𝑠 (𝜁 (𝜂 − 𝑧0 )1/𝑞 )𝑠 ,
ℎ = 0, . . . , 𝑞 − 1, ∣𝜂 − 𝑧0 ∣ < 𝑟.
𝑠=1
Making the substituitions 𝜂 = 𝑧 and 𝑧 = 𝜂 as input to those branches and then using (3.68)
and (3.69), we prove the identities
𝜆ℎ (𝑧) = 𝜆∗ℎ (𝑧),
ℎ = 0, . . . , 𝑞 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟,
(3.70)
𝜆∗ℎ (𝜂) = 𝜆ℎ (𝜂),
ℎ = 0, . . . , 𝑞 − 1, ∣𝜂 − 𝑧0 ∣ < 𝑟.
(3.71)
This proves
𝑞−1
𝑞−1
𝑞−1
∪
∪
∪
{𝜆∗ℎ (𝜂)}
ℎ=0
=
{𝜆ℎ (𝜂)},
ℎ=0
𝑞−1
{𝜆ℎ (𝑧)} =
ℎ=0
∪
{𝜆∗ℎ (𝑧)}
ℎ=0
which is just the set of all values of the Puiseux series 𝜆∗ (𝜂) and 𝜆(𝑧) at 𝜂 and 𝑧, respectively,
where 𝜂 = 𝑧. And hence, since for any convergent Puiseux series with say period 𝑞, the set
of all values at a point in its domain is independent of our choice of the branch of the 𝑞th
root function and primitive 𝑞th root of unity that we use, this proves the latter part of the
statement of this lemma. It now follows this and Lemma 27 that the values of the branches
of 𝜆(𝑧) are eigenvalues of 𝐿(⋅, 𝑧) if and only if the branches of its adjoint 𝜆∗ (𝜂) are eigenvalues
of 𝐿∗ (⋅, 𝜂), where 𝜂 = 𝑧. And hence 𝜆(𝑧) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑧) if and only
if 𝜆∗ (𝜂) is an eigenvalue Puiseux series of 𝐿∗ (⋅, 𝜂), where 𝜂 = 𝑧. This completes the proof.
On Results from the Spectral Perturbation Theory. We are now ready to begin
giving some of the major results from [23, 34] that we will need in our thesis. The following
definition comes from [34].
83
Definition 17 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and (𝜆0 , 𝑧0 ) ∈ 𝑈 . Let (𝜆(𝑧), 𝜑(𝑧)) be an eigenpair
Puiseux series of 𝐿(⋅, 𝑧) such that 𝜆(𝑧) and 𝜑(𝑧) are expanded about 𝑧0 with limit point 𝜆0
and 𝛽 0 , respectively. The vector 𝛽 0 is called a generating eigenvector of 𝐿(⋅, 𝑧) (at the
point (𝜆0 , 𝑧0 ) and associated with 𝜆(𝑧)).
Remark. It should be noted that in the case 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) with (𝜆0 , 𝑧0 ) ∈ 𝑈 and 𝜆0 is
an eigenvalue of 𝐿(⋅, 𝑧0 ) of finite algebraic multiplicity, that by Theorem 23 there exists an
eigenpair Puiseux series (𝜆(𝑧), 𝜑(𝑧)) of 𝐿(⋅, 𝑧) such that 𝜆(𝑧) and 𝜑(𝑧) are expanded about
𝑧0 with limit point 𝜆0 and 𝛽 0 , respectively, for some 𝛽 0 . And hence a generating eigenvector
𝛽 0 of 𝐿(⋅, 𝑧) at the point (𝜆0 , 𝑧0 ) and associated with 𝜆(𝑧) exists. And, in particular, we
have 𝛽 0 ∕= 0 and 𝛽 0 ∈ ker(𝐿(𝜆0 , 𝑧0 )). Moreover, by Lemma 28 and Lemma 22 it follows
there exists an eigenpair Puiseux series (𝜆∗ (𝜂), 𝛾(𝜂)) of 𝐿∗ (⋅, 𝜂) such that 𝜆∗ (𝜂) and 𝛾(𝜂)
are expanded about 𝑧0 with limit point 𝜆0 and 𝛾 0 , respectively, for some 𝛾 0 . Thus 𝛾 0 is
a generating eigenvector of 𝐿∗ (⋅, 𝜂) at the point (𝜆0 , 𝑧0 ) and associated with 𝜆∗ (𝜂), where
𝜆∗ (𝜂) is the adjoint of the Puiseux series 𝜆(𝑧) and 𝐿∗ is the adjoint of 𝐿. And, in particular,
𝛾 0 ∕= 0 and 𝛾 0 ∈ ker(𝐿∗ (𝜆0 , 𝑧0 )) = ker 𝐿(𝜆0 , 𝑧0 )∗ .
The next proposition is equivalent to Lemma 7 from [34] and its proof comes from [23].
As we will shall soon see this proposition is actually one of the key results in the spectral
perturbation theory of holomorphic matrix functions.
Proposition 29 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and (𝜆0 , 𝑧0 ) ∈ 𝑈 . Let 𝜆1 (𝑧), 𝜆2 (𝑧) be any two eigenvalue Puiseux series of 𝐿 expanded about 𝑧0 with common limit point 𝜆0 . Suppose that for
some branch 𝜆1,ℎ1 (𝑧) of 𝜆1 (𝑧) and some branch 𝜆2,ℎ2 (𝑧) of 𝜆2 (𝑧) there existed an 𝑟 > 0 such
that 𝜆1,ℎ1 (𝑧) ∕= 𝜆2,ℎ2 (𝑧) for ∣𝑧 − 𝑧0 ∣ < 𝑟. Then for any generating eigenvector 𝛽 0 of 𝐿(⋅, 𝑧) at
the point (𝜆0 , 𝑧0 ) and associated with 𝜆1 (𝑧) and for any generating eigenvector 𝛾 0 of 𝐿∗ (⋅, 𝜂)
84
at the point (𝜆0 , 𝑧0 ) and associated with 𝜆∗2 (𝜂) we have
⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 , 𝛾 0 ⟩ℂ = 0.
(3.72)
Proof. By assumption there exists eigenvalue Puiseux series
𝜆1 (𝑧) =
∞
∑
𝑠
𝑐1,𝑠 (𝑧 − 𝑧0 ) 𝑞1 , ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑠=0
𝜆2 (𝑧) =
∞
∑
𝑠
𝑐2,𝑠 (𝑧 − 𝑧0 ) 𝑞2 , ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑠=0
of 𝐿(⋅, 𝑧) such that for some fixed branch of the 𝑞1 th and 𝑞2 th root functions say 𝑓1 and 𝑓2
with its evaluation at (𝑧 − 𝑧0 ) given by 𝑓1 (𝑧 − 𝑧0 ) and 𝑓2 (𝑧 − 𝑧0 ), respectively, and some fixed
primitive 𝑞1 th and 𝑞2 th root of unity say 𝜁 1 and 𝜁 2 , respectively, we have all their branches
given by the convergent series
𝜆1,ℎ (𝑧) = 𝜆0 +
𝜆2,ℎ (𝑧) = 𝜆0 +
∞
∑
𝑠=1
∞
∑
𝑐1,𝑠 (𝜁 ℎ1 𝑓1 (𝑧 − 𝑧0 ))𝑠 ,
ℎ = 0, . . . , 𝑞1 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑐2,𝑠 (𝜁 ℎ2 𝑓2 (𝑧 − 𝑧0 ))𝑠 ,
ℎ = 0, . . . , 𝑞2 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟
𝑠=1
and for some ℎ1 ∈ {0, . . . , 𝑞1 − 1} and ℎ2 ∈ {0, . . . , 𝑞2 − 1},
𝜆1,ℎ1 (𝑧) ∕= 𝜆2,ℎ2 (𝑧), ∣𝑧 − 𝑧0 ∣ < 𝑟.
Now let 𝛽 0 be a generating eigenvector of 𝐿(⋅, 𝑧) at the point (𝜆0 , 𝑧0 ) and associated with
𝜆1 (𝑧) and 𝛾 0 be a generating eigenvector of 𝐿∗ (⋅, 𝜂) at the point (𝜆0 , 𝑧0 ) and associated
with 𝜆∗2 (𝜂). By definition of generating eigenvector, there exists an eigenpair Puiseux series
(𝜆(𝑧), 𝜑(𝑧)) of 𝐿(⋅, 𝑧) such that 𝜆(𝑧) and 𝜑(𝑧) are expanded about 𝑧0 with limit point 𝜆0 and
𝛽 0 , respectively, and there exists an eigenpair Puiseux series (𝜆∗2 (𝜂), 𝛾(𝜂)) of 𝐿∗ (⋅, 𝜂) such
85
that 𝜆(𝑧) and 𝜑(𝑧) are expanded about 𝑧0 with limit point 𝜆0 and 𝛾 0 , respectively.
Now because of the definition of eigenvector Puiseux series and the definition of adjoint of
a Puiseux series and because none of the properties just mentioned change by going to a
smaller domain for 𝑧, i.e., taking 𝑟 smaller, we can assume without loss of generality that all
the branches of 𝜑(𝑧), 𝜆∗2 (𝜂), 𝛾(𝜂) are given by the convergent series
∞
∑
𝜑ℎ (𝑧) = 𝛽 0 +
𝛽 1,𝑠 (𝜁 ℎ1 𝑓1 (𝑧 − 𝑧0 ))𝑠 ,
𝑠=1
∞
∑
𝜆∗2,ℎ (𝜂) = 𝜆0 +
𝛾 ℎ (𝜂) = 𝜆0 +
𝑐2,𝑠 (𝜁 ℎ2 𝑓2 (𝜂 − 𝑧0 ))𝑠 ,
𝑠=1
∞
∑
𝜒𝑠 (𝜁 ℎ2 𝑓2 (𝜂 − 𝑧0 ))𝑠 ,
ℎ = 0, . . . , 𝑞1 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞2 − 1, ∣𝜂 − 𝑧0 ∣ < 𝑟,
ℎ = 0, . . . , 𝑞2 − 1, ∣𝜂 − 𝑧0 ∣ < 𝑟,
𝑠=1
respectively, and satisfy
ℎ = 0, . . . , 𝑞1 − 1, ∣𝑧 − 𝑧0 ∣ < 𝑟,
𝐿(𝜆1,ℎ (𝑧), 𝑧)𝜑ℎ (𝑧) = 0,
𝐿∗ (𝜆∗2,ℎ (𝜂), 𝜂)𝛾 ℎ (𝜂) = 0,
ℎ = 0, . . . , 𝑞2 − 1, ∣𝜂 − 𝑧0 ∣ < 𝑟.
Now it follows from Lemma 28 that
𝑞2 −1
∪
ℎ=0
𝑞2 −1
{𝜆∗2,ℎ (𝜂)}
=
∪
{𝜆2,ℎ (𝜂)}, ∣𝜂 − 𝑧0 ∣ < 𝑟.
ℎ=0
Thus for every 𝜂 ∈ 𝐵(𝑧0 , 𝑟) there exists ℎ(𝜂) ∈ {0, . . . , 𝑞2 − 1} such that
𝜆∗2,ℎ(𝜂) (𝜂) = 𝜆2,ℎ2 (𝜂).
Hence if ∣𝑧 − 𝑧0 ∣ < 𝑟 then for 𝜂 = 𝑧 we have
0 = 𝐿∗ (𝜆∗2,ℎ(𝜂) (𝜂), 𝜂)𝛾 ℎ(𝜂) (𝜂) = 𝐿(𝜆2,ℎ2 (𝜂), 𝜂)∗ 𝛾 ℎ(𝜂) (𝜂) = 𝐿(𝜆2,ℎ2 (𝑧), 𝑧)∗ 𝛾 ℎ(𝑧) (𝑧)
86
and thus
⟨𝛾 ℎ(𝑧) (𝑧), 𝐿(𝜆2,ℎ2 (𝑧), 𝑧)𝜑ℎ1 (𝑧)⟩ℂ = ⟨𝐿(𝜆2,ℎ2 (𝑧), 𝑧)∗ 𝛾 ℎ(𝑧) (𝑧), 𝜑ℎ1 (𝑧)⟩ℂ = 0.
Now it follows from Lemma 17, since 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and (𝜆0 , 𝑧0 ) ∈ 𝑈 , that 𝐿 is analytic at (𝜆, 𝑧) = (𝜆0 , 𝑧0 ), i.e., the partial and mixed partial derivatives 𝐿(𝑗1 ,𝑗2 ) (𝜆0 , 𝑧0 ) :=
∂ 𝑗1 +𝑗2 𝐿 exist for every 𝑗1 , 𝑗2 ∈ ℕ ∪ {0} and there exists an 𝑟0 > 0 such that
∂ 𝑗1 𝜆∂ 𝑗2 𝑧 (𝜆,𝑧)=(𝜆0 ,𝑧0 )
𝐿(𝜆, 𝑧) =
∞
∑
𝑗1 ,𝑗2
1
𝐿(𝑗1 ,𝑗2 ) (𝜆0 , 𝑧0 )(𝜆 − 𝜆0 )𝑗1 (𝑧 − 𝑧0 )𝑗2 , ∣𝜆 − 𝜆0 ∣ < 𝑟0 , ∣𝑧 − 𝑧0 ∣ < 𝑟0
𝑗
!𝑗
!
=0 1 2
where the power series on the right converges absolutely in the 𝑀𝑛 (ℂ) norm to 𝐿(𝜆, 𝑧) for
(𝜆, 𝑧) ∈ 𝐵(𝜆0 , 𝑟0 ) × 𝐵(𝑧0 , 𝑟0 ). We may assume without loss of generality that 𝑟0 ≤ 𝑟.
Thus from the facts above and the facts
𝑥𝑗 − 𝑦 𝑗 = (𝑥 − 𝑦)
𝑗−1
∑
𝑥𝑙 𝑦 𝑗−1−𝑙 ,
𝑥, 𝑦 ∈ ℂ,
𝑗∈ℕ
𝑙=0
lim𝑧→𝑧0 𝜆1,ℎ1 (𝑧) = lim 𝜆1,ℎ2 (𝑧) = 𝜆0
𝑧→𝑧0
lim𝑧→𝑧0 𝛾 ℎ(𝑧) (𝑧) = lim 𝛾 ℎ (𝑧) = 𝛾 0 ,
𝑧→𝑧0
ℎ = 0, . . . , 𝑞2 − 1
it follows that
(𝜆1,ℎ1 (𝑧) − 𝜆0 )𝑗 − (𝜆2,ℎ2 (𝑧) − 𝜆0 )𝑗
= 𝑜(1), as 𝑧 → 𝑧0 , if 𝑗 ≥ 2
𝜆1,ℎ1 (𝑧) − 𝜆2,ℎ2 (𝑧)
and so
𝐿(𝜆1,ℎ1 (𝑧), 𝑧) − 𝐿(𝜆2,ℎ2 (𝑧), 𝑧)
= 𝐿(1,0) (𝜆0 , 𝑧0 ) + 𝑜(1), as 𝑧 → 𝑧0
𝜆1,ℎ1 (𝑧) − 𝜆2,ℎ2 (𝑧)
87
which implies
⟨𝛾 ℎ(𝑧) (𝑧), 𝐿(𝜆1,ℎ1 (𝑧), 𝑧)𝜑ℎ1 (𝑧)⟩ℂ − ⟨𝛾 ℎ(𝑧) (𝑧), 𝐿(𝜆2,ℎ2 (𝑧), 𝑧)𝜑ℎ1 (𝑧)⟩ℂ
0=
𝜆1,ℎ1 (𝑧) − 𝜆2,ℎ2 (𝑧)
⟩
⟨
𝐿(𝜆1,ℎ1 (𝑧), 𝑧) − 𝐿(𝜆2,ℎ2 (𝑧), 𝑧)
𝜑ℎ1 (𝑧)
= 𝛾 ℎ(𝑧) (𝑧),
𝜆1,ℎ1 (𝑧) − 𝜆2,ℎ2 (𝑧)
ℂ
= ⟨𝛾 0 , 𝐿(1,0) (𝜆0 , 𝑧0 )𝛽 0 ⟩ℂ + 𝑜(1), as 𝑧 → 𝑧0 .
Hence we must have
0 = ⟨𝛾 0 , 𝐿(1,0) (𝜆0 , 𝑧0 )𝛽 0 ⟩ℂ = ⟨𝛾 0 , 𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 ⟩ℂ = ⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 , 𝛾 0 ⟩ℂ ,
where by our notation 𝐿𝜆 (𝜆0 , 𝑧0 ) := 𝐿(1,0) (𝜆0 , 𝑧0 ). Therefore ⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 , 𝛾 0 ⟩ℂ = 0. This
completes the proof.
Lemma 30 Let 𝜆(𝑧) be a convergent Puiseux series expanded about 𝑧0 with domain 𝐵(𝑧0 , 𝑟0 )
and period 𝑞. Suppose 𝜆(𝑧) is not a single-valued analytic function of 𝑧 in 𝐵(𝑧0 , 𝑟0 ). Then
if we let 𝜆1 (𝑧), . . . , 𝜆𝑞−1 (𝑧) be all the branches of 𝜆(𝑧) with respect to a fix branch of the 𝑞th
root function and a fix primitive 𝑞th root of unity, there exists ℎ1 , ℎ2 ∈ {0, . . . , 𝑞 − 1} and an
𝑟 > 0 with 𝑟 ≤ 𝑟0 such that
𝜆ℎ1 (𝑧) ∕= 𝜆ℎ2 (𝑧), 0 < ∣𝑧 − 𝑧0 ∣ < 𝑟.
Proof.
(3.73)
Suppose 𝜆(𝑧) is a convergent Puiseux series expanded about 𝑧0 with domain
𝐵(𝑧0 , 𝑟0 ) and period 𝑞, where
𝜆(𝑧) =
∞
∑
𝑠
𝑐𝑠 (𝑧 − 𝑧0 ) 𝑞 , ∣𝑧 − 𝑧0 ∣ < 𝑟0 ,
𝑠=0
and suppose that 𝜆(𝑧) is not a single-valued analytic function of 𝑧 in 𝐵(𝑧0 , 𝑟0 ), i.e., there
exists 𝑠 ∈ ℕ such that
𝑠
𝑞
∕∈ ℕ and 𝑐𝑠 ∕= 0. We will denote by 𝑠0 the smallest such 𝑠 with that
88
property. Note that for the Puiseux series 𝜆(𝑧) with the series representation give above we
must have the period 𝑞 ≥ 2.
Now fix a branch of the 𝑞th root and denote its evaluation at 𝑧 − 𝑧0 by (𝑧 − 𝑧0 )1/𝑞 and fix
some primitive 𝑞th root of unity denoting it by 𝜁. Then the branches of 𝜆(𝑧) are given by
the convergent series
𝜆ℎ (𝑧) =
∞
∑
𝑐𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 , ∣𝑧 − 𝑧0 ∣ < 𝑟0 ,
ℎ = 0, . . . , 𝑞 − 1.
𝑠=0
By the definition of 𝑠0 there exists an 𝑙 ∈ ℕ ∪ {0} such that
𝜆ℎ (𝑧) =
𝑙
∑
𝑗
ℎ
1/𝑞 𝑠0
𝑐𝑗𝑞 (𝑧 − 𝑧0 ) + 𝑐𝑠0 (𝜁 (𝑧 − 𝑧0 )
) +
𝑠0 −1
𝑞
∞
∑
= 𝑙, 𝑐𝑠0 ∕= 0, and
𝑐𝑠 (𝜁 ℎ (𝑧 − 𝑧0 )1/𝑞 )𝑠 ,
𝑠=𝑠0 +1
𝑗=0
for ∣𝑧 − 𝑧0 ∣ < 𝑟0 and ℎ = 0, . . . , 𝑞 − 1. It follows from this that for any ℎ1 , ℎ2 ∈ {0, . . . , 𝑞 − 1}
we have
𝑠0
∣𝜆ℎ1 (𝑧) − 𝜆ℎ2 (𝑧)∣ = ∣𝑐𝑠0 ∣∣𝜁 ℎ1 𝑠0 − 𝜁 ℎ2 𝑠0 ∣∣𝑧 − 𝑧0 ∣ 𝑞 + 𝑂(∣𝑧 − 𝑧0 ∣
𝑠0 +1
𝑞
), as 𝑧 → 𝑧0 .
The leading order term here can only be zero if ∣𝜁 ℎ1 𝑠0 − 𝜁 ℎ2 𝑠0 ∣ = 0 which would imply
𝜁 (ℎ1 −ℎ2 )𝑠0 = 1 and since 𝜁 is a primitive 𝑞th root of unity this implies
letting ℎ1 := 1, ℎ2 := 0 ∈ {0, . . . , 𝑞 − 1} (since 𝑞 ≥ 2) then, because
(ℎ1 −ℎ2 )𝑠0
𝑞
(ℎ1 −ℎ2 )𝑠0
𝑞
=
∈ ℤ. Thus
𝑠0
𝑞
∕∈ ℤ, we
have ∣𝜁 ℎ1 𝑠0 − 𝜁 ℎ2 𝑠0 ∣ =
∕ 0 and so
∣𝜆ℎ1 (𝑧) − 𝜆ℎ2 (𝑧)∣
𝑠0
∣𝑐𝑠0 ∣∣𝜁 ℎ1 𝑠0 − 𝜁 ℎ2 𝑠0 ∣∣𝑧 − 𝑧0 ∣ 𝑞
= 1 + 𝑜(1),
𝑧 ∕= 𝑧0 , as 𝑧 → 𝑧0 .
Therefore, since lim 𝜆ℎ (𝑧) = 𝑐0 for ℎ = 0, . . . , 𝑞 − 1, it follows there exists 𝑟 > 0 with 𝑟 ≤ 𝑟0
𝑧→𝑧0
such that
𝜆ℎ1 (𝑧) ∕= 𝜆ℎ2 (𝑧), 0 < ∣𝑧 − 𝑧0 ∣ < 𝑟.
89
This completes the proof.
The following theorem is Theorem 9 from [34].
Theorem 31 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and suppose (𝜆0 , 𝑧0 ) ∈ 𝑈 . Let 𝜆(𝑧) be an eigenvalue
Puiseux series of 𝐿(⋅, 𝑧) expanded about 𝑧0 with limit point 𝜆0 . Assume that for every generating eigenvector 𝛽 of 𝐿(⋅, 𝑧) at the point (𝜆0 , 𝑧0 ) and associated with 𝜆(𝑧) there exists a
generating eigenvector 𝛾 of 𝐿∗ (⋅, 𝜂) at the point (𝜆0 , 𝑧0 ) and associated with 𝜆∗ (𝜂) such that
⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽, 𝛾⟩ℂ ∕= 0,
(3.74)
where 𝜆∗ (𝜂) denotes the adjoint of 𝜆(𝑧) and 𝐿∗ denotes the adjoint of 𝐿. Then 𝜆(𝑧) is a
single-valued analytic function of 𝑧 and there exists an eigenvector Puiseux series 𝜑(𝑧) of
𝐿(⋅, 𝑧) corresponding to 𝜆(𝑧) which is also a single-valued analytic function of 𝑧.
Proof.
Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and suppose (𝜆0 , 𝑧0 ) ∈ 𝑈 . Let 𝜆(𝑧) be an eigenvalue
Puiseux series of 𝐿(⋅, 𝑧) expanded about 𝑧0 , domain 𝐵(𝑧0 , 𝑟0 ), with limit point 𝜆0 , and
period 𝑞. Suppose the hypotheses of this theorem are true for 𝜆(𝑧). But suppose that 𝜆(𝑧)
was not a single-valued analytic function of 𝑧 ∈ 𝐵(𝑧0 , 𝑟0 ). Then by Lemma 30, if we let
𝜆1 (𝑧), . . . , 𝜆𝑞−1 (𝑧) be all the branches of 𝜆(𝑧) with respect to a fix branch of the 𝑞th root
function and a fix primitive 𝑞th root of unity, there exists ℎ1 , ℎ2 ∈ {0, . . . , 𝑞 − 1} and an
𝑟 > 0 with 𝑟 ≤ 𝑟0 such that
𝜆ℎ1 (𝑧) ∕= 𝜆ℎ2 (𝑧), 0 < ∣𝑧 − 𝑧0 ∣ < 𝑟.
(3.75)
Now by Lemma 22 there exists an eigenvector Puiseux series 𝜑(𝑧) of 𝐿(⋅, 𝑧) corresponding
to the eigenvalue Puiseux series 𝜆(𝑧). Let 𝛽 0 denote the limit point of 𝜑(𝑧) when expanded
about 𝑧0 . By definition 𝛽 0 is a generating eigenvector of 𝐿(⋅, 𝑧) at the point (𝜆0 , 𝑧0 ) and
90
associated with 𝜆(𝑧) and so by our hypotheses their exists a eigenvalue Puiseux series 𝛾(𝜂),
expanded about 𝑧0 with limit point 𝛾 0 , of 𝐿∗ (⋅, 𝜂) corresponding to the eigenvalue Puiseux
series 𝜆∗ (𝜂) such that
⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 , 𝛾 0 ⟩ℂ ∕= 0.
(3.76)
But according to Proposition 29 with 𝜆1 (𝑧) := 𝜆2 (𝑧) = 𝜆(𝑧) we must have
⟨𝐿𝜆 (𝜆0 , 𝑧0 )𝛽 0 , 𝛾 0 ⟩ℂ = 0.
(3.77)
This is a contradiction. Therefore, 𝜆(𝑧) is single-valued analytic function of 𝑧 ∈ 𝐵(𝑧0 , 𝑟0 )
and so by Lemma 22 there exists an eigenvector Puiseux series 𝜑(𝑧) of 𝐿(⋅, 𝑧) corresponding
to 𝜆(𝑧) which is also a single-valued analytic function of 𝑧. This completes the proof.
The following proposition comes from [23, Lemma 4.1].
Proposition 32 Let 𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and suppose (𝜆0 , 𝑧0 ) ∈ 𝜎(𝐿). Assume (𝜆(𝑧), 𝜑(𝑧)) is
an eigenpair Puiseux series of 𝐿(⋅, 𝑧) such that both 𝜆(𝑧) and 𝜑(𝑧) are analytic functions at
𝑧0 with 𝜆(𝑧0 ) = 𝜆0 . Denote the order of the zero of the function 𝜆(𝑧) − 𝜆0 at 𝑧0 by 𝑞. Then
𝜑(𝑧) is a generating function of order 𝑞 for 𝐿(𝜆0 , ⋅) at the eigenvalue 𝑧0 .
Proof.
By hypotheses there exists an 𝑟 > 0 such that 𝐵(𝜆0 , 𝑟) × 𝐵(𝑧0 , 𝑟) ⊆ 𝑈 , 𝜆(⋅) ∈
𝒪(𝐵(𝑧0 , 𝑟), ℂ), and 𝜑(⋅) ∈ 𝒪(𝐵(𝑧0 , 𝑟), ℂ𝑛 ) such that 𝜑(𝑧0 ) ∕= 0 and 𝐿(𝜆(𝑧), 𝑧)𝜑(𝑧) = 0 for
every 𝑧 ∈ 𝐵(𝑧0 , 𝑟). Thus it follows from this by using the power series expansion of 𝐿 at
(𝜆0 , 𝑧0 ) from Lemma 17, that
𝜑(𝑧0 ) ∕= 0,
𝐿(𝜆0 , 𝑧)𝜑(𝑧) = 𝐿(𝜆0 , 𝑧)𝜑(𝑧) − 𝐿(𝜆(𝑧), 𝑧)𝜑(𝑧)
= (𝐿(𝜆0 , 𝑧) − 𝐿(𝜆(𝑧), 𝑧))𝜑(𝑧)
= −(𝜆(𝑧) − 𝜆0 )𝐿𝜔 (𝜆0 , 𝑧0 )𝜑(𝑧) + 𝑜(𝜆(𝑧) − 𝜆0 )
91
= −𝜆(𝑞) (𝑧0 )(𝑧 − 𝑧0 )𝑞 𝐿𝜔 (𝜆0 , 𝑧0 )𝜑(𝑧0 ) + 𝑜((𝑧 − 𝑧0 )𝑞 )
= 𝑂((𝑧 − 𝑧0 )𝑞 ), as 𝑧 → 𝑧0 .
By definition of generating function this implies 𝜑(𝑧) is a generating function of order 𝑞 for
𝐿(𝜆0 , ⋅) at the eigenvalue 𝑧0 .
92
Chapter 4
Canonical Equations: A Model for
Studying Slow Light
4.1
Introduction
In this chapter we consider a general nondissipative but dispersive model for wave propagation using an important class of periodic differential and differential-algebraic equations
(DAEs) called canonical equations [31]. This model is general enough to include electromagnetic waves governed by the time-harmonic Maxwell’s equations for lossless one-dimensional
photonic crystals whose constituent layers can be any combination of isotropic, anisotropic,
or bianisotropic materials with or without material dispersion (i.e., frequency-dependent response of materials). This makes our work particularly significant in the study of slow light
since metamaterials are widening the range of potential photonic crystals that can be fabricated and so a model like ours that has the ability to analysis slow light phenomena for a
broad range of photonic crystals is in need.
93
4.1.1
Model Formulation
In this chapter1 , we will denoted a wave by 𝑦 and its evaluation at a position 𝑥 by 𝑦(𝑥),
where 𝑥 ∈ ℝ, 𝑦(𝑥) ∈ ℂ𝑁 , and 𝑦 ∈ (𝐿2𝑙𝑜𝑐 (ℝ))𝑁 . The frequency domain Ω is an open connected
set in ℂ and the real frequency domain Ωℝ := Ω ∩ ℝ is nonempty.
Wave propagation will be governed by canonical systems of differential equations with periodic coefficients that depend holomorphically on the frequency parameter 𝜔, i.e., differential
equations of the form
G 𝑦 ′ (𝑥) = 𝑉 (𝑥, 𝜔)𝑦(𝑥),
(4.1)
where the leading matrix coefficient G ∈ 𝑀𝑁 (ℂ) and the matrix-valued function 𝑉 : ℝ×Ω →
𝑀𝑁 (ℂ), the Hamiltonian, have the properties:
(i) G ∗ = −G
(ii) 𝑉 (𝑥, 𝜔)∗ = 𝑉 (𝑥, 𝜔), for each 𝜔 ∈ Ωℝ and almost every 𝑥 ∈ ℝ
(iii) 𝑉 (𝑥 + 𝑑, 𝜔) = 𝑉 (𝑥, 𝜔), for each 𝜔 ∈ Ω and almost every 𝑥 ∈ ℝ
(iv) 𝑉 ∈ 𝒪(Ω, 𝑀𝑁 (𝐿𝑝 (𝕋))) as a function of frequency where
⎧

⎨ 1 if det(G ) ∕= 0,
𝑝=

⎩ 2 if det(G ) = 0.
In the case det(G ) = 0, the domain of equation (4.1) and definition of its leading term are
1,1
(ℝ))𝑁 }, G 𝑦 ′ := G (𝑃⊥ 𝑦)′ ,
𝒟 := {𝑦 ∈ (𝐿2𝑙𝑜𝑐 (ℝ))𝑁 : 𝑃⊥ 𝑦 ∈ (𝑊𝑙𝑜𝑐
1
See section 4.5 for notation and convention.
94
(4.2)
where 𝑃 , 𝑃⊥ ∈ 𝑀𝑁 (ℂ) denote the projections onto the kernel and range of G , respectively.
1,1
In the case det(G ) ∕= 0, we take the standard domain 𝒟 := (𝑊𝑙𝑜𝑐
(ℝ))𝑁 and definition of 𝑦 ′
as the weak derivative of the entries of 𝑦.
The equations of the form (4.1) with det(G ) = 0 are differential-algebraic equations (DAEs)
and are not just ODEs. Their solutions must satisfy an algebraic equation as well, namely,
𝑃 𝑉 (𝑥, 𝜔)𝑦(𝑥) = 0 for a.e. 𝑥 ∈ ℝ. Thus, in order to guarantee the existence of a nontrivial
solution for each frequency and a solution space which depends holomorphically on frequency,
we impose an additional hypothesis for equations of the form (4.1) whenever det(G ) = 0,
namely, (G + 𝑃 𝑉 (𝑥, 𝜔)𝑃 )−1 exists for each 𝜔 ∈ Ω and a.e. 𝑥 ∈ ℝ and as a function of
frequency
(G + 𝑃 𝑉 𝑃 )−1 ∈ 𝒪(Ω, 𝑀𝑁 (𝐿∞ (0, 𝑑))).
(4.3)
We will refer to this as the index-1 hypothesis.
Differential-algebraic equations of the form given by (4.1), for real frequencies, are known as
canonical equations [31] and, in the case det(G ) ∕= 0, they are known as canonical differential
equations, linear canonical systems, linear Hamiltonian systems, or Hamiltonian equations.
We distinguish between the two classes of canonical equations of the form (4.1) by referring
to those with det(G ) = 0 as canonical DAEs and those with det(G ) ∕= 0 as canonical ODEs.
We introduce the following definitions into our model:
Definition 18 We say 𝑦 is a solution of equation (4.1) provided 𝑦 ∈ 𝒟 and equation (4.1)
is satisfied for some 𝜔 ∈ Ω and for a.e. 𝑥 in ℝ. In which case we call 𝜔 its frequency.
Definition 19 We say 𝑦 is a Floquet solution provided for some 𝜔 ∈ Ω, it is a solution
95
of equation (4.1) with 𝜔 as its frequency and has the Floquet representation
𝑦(𝑥) = 𝑒𝑖𝑘𝑥
𝑚
∑
𝑥𝑗 𝑢𝑗 (𝑥),
(4.4)
𝑗=0
where 𝑚 ∈ ℕ ∪ {0} and 𝑢𝑗 ∈ 𝒟 ∩ (𝐿𝑝 (𝕋))𝑁 , for 𝑗 = 1, . . . , 𝑚. We call 𝑘, (𝑘, 𝜔), and 𝜆 := 𝑒𝑖𝑘𝑑
its wavenumber, wavenumber-frequency pair, and Floquet multiplier, respectively.
If 𝑢𝑚 ∕= 0 then we call 𝑚 its order. We call 𝑦 a Bloch solution if it is a Floquet solution
with 𝑚 = 0. If, in addition, 𝑦 is a Bloch solution with real frequency then we call it a
propagating wave if it has a real wavenumber and an evanescent wave otherwise.
Definition 20 We define the Bloch variety to be the set
ℬ := {(𝑘, 𝜔) ∈ ℂ × Ω ∣ (𝑘, 𝜔) is the wavenumber-frequency pair of some
(4.5)
nontrivial Bloch solution of equation (4.1)}
and the real Bloch variety to be the set
ℬℝ := ℬ ∩ ℝ2 .
(4.6)
Definition 21 We define the dispersion relation denoted by 𝜔 = 𝜔(𝑘) as the multivalued function whose graph is the Bloch variety. We say a point (𝑘0 , 𝜔 0 ) ∈ ℬℝ is amicable
if for every 𝜖 > 0 there exists an 𝛿 > 0 such that ℬ ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) is the union
of the graphs of a finite number of real analytic functions, say 𝜔 1 (𝑘), . . . , 𝜔 𝑔 (𝑘), with the
property 𝜔 1 (𝑘0 ) = ⋅ ⋅ ⋅ = 𝜔 𝑔 (𝑘0 ) = 𝜔 0 . We call these functions the band functions and
their derivatives
𝑔
𝑑𝜔 1
, . . . , 𝑑𝜔
𝑑𝑘
𝑑𝑘
the group velocities. The intersection of the graph of a band
function with ℝ2 is called a band.
Definition 22 For any solution 𝑦 of equation (4.1) with real frequency 𝜔 0 we define its
96
energy flux and energy density to be the functions of position given by
⟨𝑖G 𝑦, 𝑦⟩ℂ
(4.7)
⟨𝑉𝜔 (⋅, 𝜔 0 )𝑦, 𝑦⟩ℂ
(4.8)
respectively, where 𝑉𝜔 denotes the derivative of the Hamiltonian 𝑉 with respect to frequency in
the 𝑀𝑁 (𝐿𝑝 (𝕋)) norm. We define its average energy flux and average energy density
to be the averages
∫
1 𝑑
⟨𝑖G 𝑦(𝑥), 𝑦(𝑥)⟩ℂ 𝑑𝑥
𝑑 0
∫
1 𝑑
⟨𝑉𝜔 (𝑥, 𝜔 0 )𝑦(𝑥), 𝑦(𝑥)⟩ℂ 𝑑𝑥
𝑑 0
(4.9)
(4.10)
respectively. We also define its energy velocity as the ratio of its average energy flux to
its average energy density provided the latter is nonzero, i.e.,
∫
1 𝑑
⟨𝑖G 𝑦(𝑥), 𝑦(𝑥)⟩ℂ 𝑑𝑥
𝑑 0
.
∫
𝑑
1
⟨𝑉
(𝑥,
𝜔
)𝑦(𝑥),
𝑦(𝑥)⟩
𝑑𝑥
𝜔
0
ℂ
𝑑 0
(4.11)
Remark: The integrals (4.9) and (4.10) are well-defined because Corollary 40, Corollary 43,
and Theorem 50 show the functions (4.7) and (4.8) are in 𝐿1loc (ℝ).
Definition 23 We say (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the canonical equations
in (4.1) and say these equations are of definite type at this point provided the average energy
density of any nontrivial Bloch solution 𝑦 with (𝑘0 , 𝜔 0 ) as its wavenumber-frequency pair is
nonzero, i.e.,
1
𝑑
∫
𝑑
⟨𝑉𝜔 (𝑥, 𝜔 0 )𝑦(𝑥), 𝑦(𝑥)⟩ℂ 𝑑𝑥 ∕= 0.
0
97
(4.12)
The canonical equations in (4.1) along with the definitions just given constitute the model
of wave propagation that we will analyze in this thesis. Moreover, based on our studies
on electromagnetic wave propagation in one-dimensional photonic crystals, it is found by
physical considerations that a point of definite type is the right assumption in which to
begin a perturbation analysis from stationary points of the dispersion relation.
We are particularly interested in the perturbation analysis of canonical equations in the slow
wave regime in which points in the Bloch variety (𝑘, 𝜔) belong to a small neighborhood of
a point (𝑘0 , 𝜔 0 ) in the real Bloch variety where (𝑘0 , 𝜔 0 ) is both a point of definite type for
the canonical equations (such a point is amicable (see Theorem 48 and Theorem 50)) and a
stationary point of the dispersion relation, i.e., at (𝑘0 , 𝜔 0 ) one of the bands has a stationary
point. For this model and with this regime in mind, we give the following definition for a
slow wave:
Definition 24 Let (𝑘0 , 𝜔 0 ) be any point which is amicable and 0 < 𝑟 ≪ 1. Then any
propagating Bloch wave 𝑦 with wavenumber-frequency pair (𝑘, 𝜔) belonging to a band with
a stationary point at (𝑘0 , 𝜔 0 ) and satisfying 0 < ∣∣(𝑘, 𝜔) − (𝑘0 , 𝜔 0 )∣∣ℂ < 𝑟 is called a slow
wave.
The main purpose of this chapter is to begin developing a mathematical framework for this
model, including the Floquet, spectral, and perturbation theory, and to use this framework
to analyze in detail the slow wave regime.
4.2
Canonical ODEs
In this section we consider the canonical ODEs
J 𝜓 ′ (𝑥) = 𝐴(𝑥, 𝜔)𝜓(𝑥),
98
1,1
𝜓 ∈ (𝑊loc
(ℝ))𝑛
(4.13)
where, as discussed in section 4.1.1, the leading matrix coefficient J ∈ 𝑀𝑛 (ℂ) and the
Hamiltonian 𝐴 : ℝ × Ω → 𝑀𝑛 (ℂ) have the properties:
(i) det(J ) ∕= 0, J ∗ = −J ,
(ii) 𝐴(𝑥, 𝜔)∗ = 𝐴(𝑥, 𝜔), for each 𝜔 ∈ Ωℝ and a.e. 𝑥 ∈ ℝ,
(iii) 𝐴(𝑥 + 𝑑, 𝜔) = 𝐴(𝑥, 𝜔), for each 𝜔 ∈ Ω and a.e. 𝑥 ∈ ℝ,
(iv) 𝐴 ∈ 𝒪(Ω, 𝑀𝑁 (𝐿1 (𝕋))) as a function of frequency.
Denote by 𝐴𝜔 the derivative of the Hamiltonian 𝐴 with respect to frequency in the 𝑀𝑛 (𝐿1 (𝕋))
norm.
4.2.1
Preliminaries
We begin with some preliminary results on the solutions of the canonical ODEs in (4.13)
and the Floquet theory relating to the matricant and monodromy matrix of these canonical ODEs. The dependency of the matricant and monodromy matrix on frequency is also
discussed. We will hold off on proving the results of this subsection until section 4.4. The
statements in this subsection are known, but we state them here in order to setup the main
results of this chapter. We also will give proofs of these statements for completeness in
section 4.4.
Proposition 33 For each 𝜔 ∈ Ω, let Ψ(⋅, 𝜔) denote the fundamental matrix solution of
the canonical ODEs in (4.13) satisfying Ψ(0, 𝜔) = 𝐼𝑛 , i.e., the unique function Ψ(⋅, 𝜔) ∈
1,1
𝑀𝑛 (𝑊loc
(ℝ)) satisfying a.e. the matrix differential equation with initial condition
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
99
(4.14)
Then the matrix-valued function Ψ : ℝ × Ω → 𝑀𝑛 (ℂ) has the following properties:
(i) 𝜓 is a solution of the canonical ODEs in (4.13) at the frequency 𝜔 ∈ Ω if and only if
𝜓 = Ψ(⋅, 𝜔)𝛾 for some 𝛾 ∈ ℂ𝑛 . Moreover, 𝛾 is unique.
(ii) For every (𝑥, 𝜔) ∈ ℝ × Ω,
∫
Ψ(𝑥, 𝜔) = 𝐼𝑛 +
𝑥
J −1 𝐴(𝑡, 𝜔)Ψ(𝑡, 𝜔)𝑑𝑡.
(4.15)
0
(iii) For every 𝑥 ∈ ℝ, Ψ(𝑥, ⋅) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)). We will denote the partial derivative of the
function Ψ with respect to frequency in the 𝑀𝑛 (ℂ) norm by Ψ𝜔 .
1,1
(iv) For every (𝑥, 𝜔) ∈ ℝ × Ω, Ψ𝜔 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) and
∫
Ψ𝜔 (𝑥, 𝜔) =
𝑥
J −1 𝐴𝜔 (𝑡, 𝜔)Ψ(𝑡, 𝜔) + J −1 𝐴(𝑡, 𝜔)Ψ𝜔 (𝑡, 𝜔)𝑑𝑡
(4.16)
0
1,1
(v) For every (𝑥, 𝜔) ∈ ℝ × Ω, Ψ−1 (𝑥, 𝜔) := Ψ(𝑥, 𝜔)−1 exists and Ψ−1 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
(vi) For every (𝑥, 𝜔) ∈ ℝ × Ω,
Ψ(𝑥 + 𝑑, 𝜔) = Ψ(𝑥, 𝜔)Ψ(𝑑, 𝜔).
(4.17)
Now for each fixed 𝜔 ∈ Ω, Ψ(⋅, 𝜔) and Ψ(𝑑, 𝜔) are the matricant and monodromy matrix,
respectively, of the canonical ODEs in (4.13). Thus the previous proposition implies the
following:
Theorem 34 (Floquet-Lyapunov) Let 𝜔 ∈ Ω. Then there exists a matrix 𝐾(𝜔) ∈ 𝑀𝑛 (ℂ)
1,1
and a function 𝐹 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) such that for every 𝑥 ∈ ℝ,
Ψ(𝑥, 𝜔) = 𝐹 (𝑥, 𝜔)𝑒𝑖𝑥𝐾(𝜔) ,
100
𝐹 (𝑥 + 𝑑, 𝜔) = 𝐹 (𝑥, 𝜔), 𝐹 (0, 𝜔) = 𝐼𝑛 , 𝐹 −1 (𝑥, 𝜔) := 𝐹 (𝑥, 𝜔)−1 exists, and 𝐹 −1 (⋅, 𝜔) ∈
1,1
𝑀𝑛 (𝑊loc
(ℝ)).
Theorem 35 If 𝜓 is a Floquet solution of the canonical ODEs in (4.13) with wavenumberfrequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚 then 𝜓 = Ψ(⋅, 𝜔)𝛾 where
𝛾 is a generalized eigenvector of Ψ(𝑑, 𝜔) of order 𝑚 + 1 corresponding to the eigenvalue 𝜆.
Conversely, if 𝛾 is a generalized eigenvector of order 𝑚 + 1 of the monodromy matrix Ψ(𝑑, 𝜔)
corresponding to the eigenvalue 𝜆 then for any 𝑘 ∈ ℂ such that 𝜆 = 𝑒𝑖𝑘𝑑 , 𝜓 = Ψ(⋅, 𝜔)𝛾 is
a Floquet solution of the canonical ODEs in (4.13) with wavenumber-frequency pair (𝑘, 𝜔),
Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚.
Corollary 36 Let 𝜔 ∈ Ω. Let {𝛾 𝑗 }𝑛𝑗=1 be a Jordan basis for the matrix Ψ(𝑑, 𝜔). For
𝑗 = 1, . . . , 𝑛, let 𝑙𝑗 , 𝜆𝑗 , and 𝑘𝑗 be such that 𝛾 𝑗 is a generalized eigenvalue of Ψ(𝑑, 𝜔) of
order 𝑙𝑗 corresponding to the eigenvalue 𝜆𝑗 = 𝑒𝑖𝑘𝑗 𝑑 . Define {𝜓 𝑗 }𝑛𝑗=1 by 𝜓 𝑗 := Ψ(⋅, 𝜔)𝛾 𝑗 for
𝑗 = 1, . . . , 𝑛. Then the following statements are true:
(i) The set of solutions of the canonical ODEs in (4.13) at the frequency 𝜔 is a vector
space over ℂ and {𝜓 𝑗 }𝑛𝑗=1 is a basis for this space.
(ii) For 𝑗 = 1, . . . , 𝑛, 𝜓 𝑗 is a Floquet solution of the canonical ODEs in (4.13) with
wavenumber-frequency pair (𝑘𝑗 , 𝜔), Floquet multiplier 𝜆𝑗 = 𝑒𝑖𝑘𝑗 𝑑 , and order 𝑙𝑗 − 1.
Corollary 37 Let ℬ denote the Bloch variety of the canonical ODEs in (4.13). Then, for
any 𝑙 ∈ ℤ,
ℬ = (2𝜋𝑙/𝑑, 0) + ℬ.
(4.18)
Corollary 38 Define the function 𝐷 : ℂ × Ω → ℂ by
(
)
𝐷(𝑘, 𝜔) := det 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔) .
101
(4.19)
Then 𝐷 is a nonconstant holomorphic function and its zero set is the Bloch variety of the
canonical ODEs in (4.13), i.e.,
ℬ = {(𝑘, 𝜔) ∈ ℂ × Ω ∣ 𝐷(𝑘, 𝜔) = 0}.
4.2.2
(4.20)
Energy Flux, Energy Density, and their Averages
The following results describe how the energy flux, energy density, their averages, and points
of definite type of the canonical ODEs in (4.13) are related to the leading matrix coefficient,
the matricant, and the monodromy matrix.
We start by considering the energy flux and its average. The following proposition is the key
result, as we will see in this chapter, in the study of energy flux for canonical equations.
1,1
Proposition 39 For each 𝜔 ∈ Ωℝ we have Ψ(⋅, 𝜔)∗ 𝑖J Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) and
Ψ(⋅, 𝜔)∗ 𝑖J Ψ(⋅, 𝜔) = 𝑖J .
(4.21)
This proposition is important because it tells us the energy flux of any solution of canonical
ODEs with real frequency is a conserved quantity. A more precise statement of this is found
in the next corollary.
Corollary 40 For each 𝜔 ∈ Ωℝ and every 𝛾 1 , 𝛾 2 ∈ ℂ𝑛 , if 𝜓 1 = Ψ(⋅, 𝜔)𝛾 1 and 𝜓 2 = Ψ(⋅, 𝜔)𝛾 2
1,1
then ⟨𝑖J 𝜓 1 , 𝜓 2 ⟩ℂ ∈ 𝑊loc
(ℝ) and
1
𝑑
∫
⟨𝑖J 𝜓 1 , 𝜓 2 ⟩ℂ = ⟨𝑖J 𝛾 1 , 𝛾 2 ⟩ℂ ,
(4.22)
⟨𝑖J 𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥 = ⟨𝑖J 𝛾 1 , 𝛾 2 ⟩ℂ .
(4.23)
𝑑
0
102
We will now consider the energy density and its average.
Lemma 41 For each 𝜔 ∈ Ωℝ we have 𝐴𝜔 (⋅, 𝜔) ∈ 𝑀𝑛 (𝐿1 (𝕋)) and
𝐴𝜔 (⋅, 𝜔)∗ = 𝐴𝜔 (⋅, 𝜔).
(4.24)
Theorem 42 For each 𝜔 ∈ Ωℝ we have
1,1
(ℝ)),
Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(4.25)
Ψ(⋅, 𝜔)∗ 𝐴𝜔 (⋅, 𝜔)Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝐿1loc (ℝ)),
(4.26)
Ψ(⋅, 𝜔)∗ 𝐴𝜔 (⋅, 𝜔)Ψ(⋅, 𝜔) = (Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔))′ ,
∫
1 𝑑
1
Ψ(𝑡, 𝜔)∗ 𝐴𝜔 (𝑡, 𝜔)Ψ(𝑡, 𝜔)𝑑𝑡 = Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔).
𝑑 0
𝑑
(4.27)
(4.28)
Corollary 43 For each 𝜔 ∈ Ωℝ and every 𝛾 1 , 𝛾 2 ∈ ℂ𝑛 , if 𝜓 1 = Ψ(⋅, 𝜔)𝛾 1 and 𝜓 2 = Ψ(⋅, 𝜔)𝛾 2
1,1
then ⟨Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ ∈ 𝑊loc
(ℝ), ⟨𝐴𝜔 (⋅, 𝜔)𝜓 1 , 𝜓 2 ⟩ℂ ∈ 𝐿1loc (ℝ) and
⟨𝐴𝜔 (⋅, 𝜔)𝜓 1 , 𝜓 2 ⟩ℂ = ⟨Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ ′ ,
1
𝑑
4.2.3
∫
0
𝑑
1
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥 = ⟨Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ .
𝑑
(4.29)
(4.30)
On Points of Definite Type for Canonical ODEs
Let (𝑘, 𝜔) ∈ ℬℝ and on the eigenspace of the monodromy matrix Ψ(𝑑, 𝜔) corresponding to
the eigenvalue 𝑒𝑖𝑘𝑑 we define the sesquilinear form 𝑞(𝑘,𝜔) by
𝑞(𝑘,𝜔) (𝛾 1 , 𝛾 2 ) :=
1
𝑑
∫𝑑
0
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥 (= 𝑑1 ⟨Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ ),
(4.31)
𝛾 1 , 𝛾 2 ∈ ker(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)),
𝜓 1 = Ψ(⋅, 𝜔)𝛾 1 ,
𝜓 2 = Ψ(⋅, 𝜔)𝛾 2 .
It follows from Corollary 43 that 𝑞(𝑘,𝜔) is well-defined.
103
Lemma 44 For each (𝑘, 𝜔) ∈ ℬℝ , 𝑞(𝑘,𝜔) is a Hermitian form, that is, 𝑞(𝑘,𝜔) satisfies
𝑞(𝑘,𝜔) (𝛾 1 , 𝛾 2 ) = 𝑞(𝑘,𝜔) (𝛾 2 , 𝛾 1 )
(4.32)
for every 𝛾 1 , 𝛾 2 ∈ ker(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)). In particular, 𝑞(𝑘,𝜔) (𝛾, 𝛾) ∈ ℝ for every 𝛾 ∈
ker(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)).
Denote the sign function by sgn(𝑥) :=
𝑥
∣𝑥∣
∈ {−1, 1}, for 𝑥 ∈ ℝ/{0}.
Proposition 45 The canonical ODEs in (4.13) are of definite type at a point (𝑘0 , 𝜔 0 ) ∈ ℬℝ
if and only if 𝑞(𝑘0 ,𝜔0 ) is a definite sesquilinear form which is bounded, i.e., the following
properties hold:
(i) For every nonzero 𝛾 ∈ ker(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)) we have
𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾) ∈ ℝ/{0}
(4.33)
and its sign sgn(𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾)) =: sgn(𝑞(𝑘0 ,𝜔0 ) ) is independent of the choice of 𝛾.
(ii) The sesquilinear form
⟨𝛾 1 , 𝛾 2 ⟩(𝑘0 ,𝜔0 ) := sgn(𝑞(𝑘0 ,𝜔0 ) )𝑞(𝑘0 ,𝜔0 ) (𝛾 1 , 𝛾 2 ),
(4.34)
𝛾 1 , 𝛾 2 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 ))
is inner product on ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
(iii) There exists constants 𝐶1 , 𝐶2 > 0 such that
𝐶1 ∣∣𝛾∣∣2ℂ ≤ ∣𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾)∣ ≤ 𝐶2 ∣∣𝛾∣∣2ℂ , for all 𝛾 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
104
(4.35)
Lemma 46 The family of sesquilinear forms {𝑞(𝑘,𝜔) }(𝑘,𝜔)∈ℬℝ depends continuously on the
indexing parameter (𝑘, 𝜔) in the following sense:
If {(𝑘𝑗 , 𝜔 𝑗 )}𝑗∈ℕ ⊆ ℬℝ , {𝛾 𝑗 }𝑗∈ℕ ⊆ ℂ𝑛 such that 𝛾 𝑗 ∈ ker(𝑒𝑖𝑘𝑗 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 𝑗 )) with ∣∣𝛾 𝑗 ∣∣ℂ = 1
for all 𝑗 ∈ ℕ and
∣∣⋅∣∣ℂ
∣∣⋅∣∣ℂ
(𝑘𝑗 , 𝜔 𝑗 ) −−→ (𝑘0 , 𝜔 0 ) and 𝛾 𝑗 −−→ 𝛾 0 , as 𝑗 → ∞
then (𝑘0 , 𝜔 0 ) ∈ ℬℝ , 𝛾 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) with ∣∣𝛾 0 ∣∣ℂ = 1 and
∣⋅∣
𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛾 𝑗 , 𝛾 𝑗 ) −
→ 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 ) as 𝑗 → ∞.
(4.36)
Theorem 47 If (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the canonical ODEs in (4.13)
then there exists an 𝑟 > 0 such that every (𝑘, 𝜔) ∈ 𝐵((𝑘0 , 𝜔 0 ), 𝑟) ∩ ℬℝ is a point of definite
type and
sgn(𝑞(𝑘,𝜔) ) = sgn(𝑞(𝑘0 ,𝜔0 ) ).
4.2.4
(4.37)
Perturbation Theory for Canonical ODEs
This section contains the main results of this chapter on the perturbation theory for canonical
ODEs.
Theorem 48 Suppose the canonical ODEs in (4.13) are of definite type at (𝑘0 , 𝜔 0 ) ∈ ℬℝ .
Let 𝑔 be the number of Jordan blocks (geometric multiplicity) in the Jordan form of the
monodromy matrix Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 and 𝑚1 ≥ ⋅ ⋅ ⋅ ≥
𝑚𝑔 ≥ 1 the dimensions of each of those Jordan blocks (partial multiplicities). Define 𝑚 :=
𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 (algebraic multiplicity). Let 𝜖 > 0 be given. Then there exists an 𝛿 > 0 such
105
that
(i) The order of the zero of 𝐷(𝑘0 , 𝜔) at 𝜔 = 𝜔 0 is 𝑔 and the order of the zero of 𝐷(𝑘, 𝜔 0 )
at 𝑘 = 𝑘0 is 𝑚.
(ii) The set ℬ ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) is the union of the graphs of 𝑔 nonconstant real analytic
functions 𝜔 1 (𝑘), . . . , 𝜔 𝑔 (𝑘) given by the convergent power series
𝜔 𝑗 (𝑘) = 𝜔 0 + 𝜈 𝑗,𝑚𝑗 (𝑘 − 𝑘0 )𝑚𝑗 +
∞
∑
𝜈 𝑗,𝑙 (𝑘 − 𝑘0 )𝑙 , ∣𝑘 − 𝑘0 ∣ < 𝛿
(4.38)
𝑙=𝑚𝑗 +1
where
𝜈 𝑗,𝑚𝑗 ∕= 0,
(4.39)
for 𝑗 = 1, . . . , 𝑔. Moreover, there exists analytic functions 𝜑1 (𝑘), . . . , 𝜑𝑔 (𝑘) belonging
to 𝒪(𝐵(𝑘0 , 𝛿), ℂ𝑛 ) such that
𝜑𝑗 (𝑘) ∕= 0, Ψ(𝑑, 𝜔 𝑗 (𝑘))𝜑𝑗 (𝑘) = 𝑒𝑖𝑘𝑑 𝜑𝑗 (𝑘), ∣𝑘 − 𝑘0 ∣ < 𝛿,
𝑗 = 1, . . . , 𝑔,
(4.40)
the vectors 𝜑1 (𝑘), . . . , 𝜑𝑔 (𝑘) are linearly independent for ∣𝑘−𝑘0 ∣ < 𝛿, and, in particular,
the vectors 𝜑1 (𝑘0 ), . . . , 𝜑𝑔 (𝑘0 ) form a basis for ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
Corollary 49 The conditions
𝐷(𝑘0 , 𝜔 0 ) = 0,
∂𝐷
(𝑘0 , 𝜔 0 ) ∕= 0, (𝑘0 , 𝜔 0 ) ∈ ℝ × Ωℝ
∂𝜔
(4.41)
are satisfied if and only if (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the canonical ODEs in
(4.13) and the Jordan normal form of the monodromy matrix Ψ(𝑑, 𝜔 0 ) corresponding to the
eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 consists of a single Jordan block, i.e., dim ker(𝜆0 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) = 1.
106
Remark. In this case, by denoting 𝜆0 := 𝑒𝑖𝑘0 𝑑 , the conditions of the previous corollary are
det(𝜆0 𝐼𝑛 − Ψ(𝑑, 𝜔)) = 𝐷(𝑘0 , 𝜔) = 0 and
∂
∂𝜔
det(𝜆𝐼𝑛 − Ψ(𝑑, 𝜔))∣(𝜔0 ,𝜆0 ) =
∂𝐷
(𝑘0 , 𝜔 0 )
∂𝜔
∕= 0,
but this is just the generic condition (2.1) in chapter 2 (see also [61, See (1.1)]). This
condition was throughly investigated in chapter 2 (see also [61, See (1.1)]) for general matrix
perturbations. And, in particular, in Theorem 2 of chapter 2 (see also [61, Theorem 3.1])
explicit recursive formulas are given for the unique eigenvalue Puiseux series 𝜆(𝜔) which
satisfies 𝜆(𝜔 0 ) = 𝜆0 and for its associated eigenvector Puiseux series of the monodromy
matrix Ψ(𝑑, 𝜔) near the frequency 𝜔 = 𝜔 0 . Moreover, their Puiseux series coefficients up to
the second order are conveniently listed in Corollary 4 of chapter 2 (see also [61, Corollary
3.3]).
4.3
Canonical DAEs
In this section we consider the canonical DAEs
G 𝜓 ′ (𝑥) = 𝑉 (𝑥, 𝜔)𝑦(𝑥),
𝑦∈𝒟
(4.42)
where, as discussed in §1, the leading matrix coefficient G ∈ 𝑀𝑛 (ℂ) and the Hamiltonian
𝑉 : ℝ × Ω → 𝑀𝑁 (ℂ) have the properties:
(i) det(G ) = 0, G ∗ = −G ,
(ii) 𝑉 (𝑥, 𝜔)∗ = 𝑉 (𝑥, 𝜔), for each 𝜔 ∈ Ωℝ and a.e. 𝑥 ∈ ℝ,
(iii) 𝑉 (𝑥 + 𝑑, 𝜔) = 𝑉 (𝑥, 𝜔), for each 𝜔 ∈ Ω and a.e. 𝑥 ∈ ℝ,
(iv) 𝑉 ∈ 𝒪(Ω, 𝑀𝑁 (𝐿2 (𝕋))) as a function of frequency.
107
Denote by 𝑉𝜔 the derivative of the Hamiltonian 𝑉 with respect to frequency in the 𝑀𝑛 (𝐿2 (𝕋))
norm.
The domain 𝒟 of the canonical DAEs in (4.42) and definition of its leading term are
1,1
𝒟 := {𝑦 ∈ (𝐿2𝑙𝑜𝑐 (ℝ))𝑁 : 𝑃⊥ 𝑦 ∈ (𝑊𝑙𝑜𝑐
(ℝ))𝑁 }, G 𝑦 ′ := G (𝑃⊥ 𝑦)′ ,
(4.43)
where 𝑃 , 𝑃⊥ ∈ 𝑀𝑁 (ℂ) denote the projections onto the kernel and range of G , respectively.
We shall assume that the canonical DAEs in (4.42) satisfy the index-1 hypothesis, namely,
(G + 𝑃 𝑉 𝑃 )−1 ∈ 𝒪(Ω, 𝑀𝑁 (𝐿∞ (0, 𝑑))).
(4.44)
The goal of this section is to describe the theory of canonical DAEs including the Floquet
and spectral perturbation theory in terms of the theory developed in the previous section
for canonical ODEs. We will only present the statement of our main result, the proof and
consequences will appear in [60].
4.3.1
The Correspondence between Canonical DAEs and Canonical ODEs
The following represents our main result in the study of canonical DAEs:
Theorem 50 Let 𝑛 := dim ran(G ). Then there exists an 𝑛 × 𝑛 system of canonical ODEs
J 𝜓 ′ (𝑥) = 𝐴(𝑥, 𝜔)𝜓(𝑥),
1,1
𝜓 ∈ (𝑊loc
(ℝ))𝑛
(4.45)
with J ∈ 𝑀𝑛 (ℂ), 𝐴 ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝕋))), and a function 𝑄 ∈ 𝒪(Ω, 𝑀𝑁 ×𝑛 (𝐿2 (𝕋))) such that
1,1
the multiplication map 𝑄(⋅, 𝜔) : (𝑊loc
(ℝ))𝑛 → 𝒟 is injective and the following properties
108
hold:
(i) The solutions of the canonical DAEs in (4.42) with frequency 𝜔 are given by 𝑦 =
𝑄(⋅, 𝜔)𝜓 where 𝜓 is a solution of the canonical ODEs in (4.45) with frequency 𝜔.
(ii) If 𝑦 = 𝑄(⋅, 𝜔)𝜓 then 𝑦 is Floquet solution of the canonical DAEs with wavenumberfrequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚 if and only if 𝜓 is a
Floquet solution of the canonical ODEs with wavenumber-frequency pair (𝑘, 𝜔), Floquet
multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚.
(iii) If 𝑦 = 𝑄(⋅, 𝜔 0 )𝜓 is a solution of the canonical DAEs with frequency 𝜔 0 ∈ Ωℝ then its
energy flux and energy density are in 𝐿1loc (ℝ) and
⟨𝑖G 𝑦, 𝑦⟩ℂ = ⟨𝑖J 𝜓, 𝜓⟩ℂ , ⟨𝑉𝜔 (⋅, 𝜔 0 )𝑦, 𝑦⟩ℂ = ⟨𝐴𝜔 (⋅, 𝜔 0 )𝜓, 𝜓⟩ℂ .
(4.46)
To summarize this theorem, the study of canonical DAEs (with index-1 hypothesis), including the Floquet, spectral, and perturbation theory, is reduced to the study of canonical
ODEs and the theory already developed in this chapter. In [60] we will elaborate on this
correspondence in more detail and discuss the spectral perturbation theory as well.
4.4
Proofs
Proofs for Section 4.2.1
We begin by introducing a differential operator associated to the canonical ODEs in (4.13).
We consider the frequency dependent operator
𝑇 (𝜔)𝜓 := 𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓,
109
1,1
𝜓 ∈ (𝑊loc
(ℝ))𝑛 ,
𝜔 ∈ Ω.
(4.47)
1,1
Lemma 51 For each 𝜔 ∈ Ω, 𝑇 (𝜔) is a linear operator with 𝑇 (𝜔) : (𝑊loc
(ℝ))𝑛 → (𝐿1loc (ℝ))𝑛 .
Proof.
1,1
(ℝ))𝑛 → (𝐿1loc (ℝ))𝑛 given by
We begin by proving the operator 𝑇 (𝜔) : (𝑊loc
1,1
1,1
(ℝ))𝑛 we
(ℝ))𝑛 . Then by definition of the space (𝑊loc
(4.47) is well-defined. Let 𝜓 ∈ (𝑊loc
have 𝜓 ′ ∈ (𝐿1loc (ℝ))𝑛 . We now show J −1 𝐴(⋅, 𝜔)𝜓 ∈ (𝐿1loc (ℝ))𝑛 . To do this let (𝑎, 𝑏) ⊆
ℝ be any bounded interval. It follows from Lemma 88 that 𝐴(⋅, 𝜔)∣(𝑎,𝑏) ∈ 𝑀𝑛 (𝐿1 (𝑎, 𝑏))
and from Lemma 69 that 𝜓∣(𝑎,𝑏) ∈ (𝑊 1,1 (𝑎, 𝑏))𝑛 . It then follows from Lemma 83 that
𝐴(⋅, 𝜔)∣(𝑎,𝑏) 𝜓∣(𝑎,𝑏) ∈ (𝐿1 (𝑎, 𝑏))𝑛 and hence J −1 𝐴(⋅, 𝜔)∣(𝑎,𝑏) 𝜓∣(𝑎,𝑏) ∈ (𝐿1 (𝑎, 𝑏))𝑛 . From this
it follows that (J −1 𝐴(⋅, 𝜔)𝜓)∣(𝑎,𝑏) = J −1 𝐴(⋅, 𝜔)∣(𝑎,𝑏) 𝜓∣(𝑎,𝑏) ∈ (𝐿1 (𝑎, 𝑏))𝑛 for every bounded
interval (𝑎, 𝑏) ⊆ ℝ, this implies J −1 𝐴(⋅, 𝜔)𝜓 ∈ (𝐿1loc (ℝ))𝑛 . And hence 𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓 ∈
1,1
(𝐿1loc (ℝ))𝑛 . This proves that the operator 𝑇 (𝜔) : (𝑊loc
(ℝ))𝑛 → (𝐿1loc (ℝ))𝑛 given by (4.47) is
well-defined. Therefore, since its obvious the operator 𝑇 (𝜔) is linear, the proof is complete.
Lemma 52 𝜓 is a solution of the canonical ODEs in (4.13) with frequency 𝜔 ∈ Ω if and
only if 𝜓 ∈ ker(𝑇 (𝜔)).
Proof.
Let 𝜓 is a solution of the canonical ODEs in (4.13) with frequency 𝜔 ∈ Ω. This
1,1
means by Definition 18 that 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and for a.e. 𝑥 ∈ ℝ 𝜓 satisfies the equation
J 𝜓 ′ (𝑥) = 𝐴(𝑥, 𝜔)𝜓(𝑥).
Hence for a.e. 𝑥 ∈ ℝ satisfies the equation
𝜓 ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)𝜓(𝑥).
110
But this implies, since as we showed in the proof of Lemma 4.47 that J −1 𝐴(⋅, 𝜔)𝜓 ∈
(𝐿1loc (ℝ))𝑛 , that 𝑇 (𝜔)𝜓 = 𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓 = 0 in (𝐿1loc (ℝ))𝑛 . This proves 𝜓 ∈ ker(𝑇 (𝜔)).
1,1
Conversely, if 𝜓 ∈ ker(𝑇 (𝜔)) then 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and 0 = 𝑇 (𝜔)𝜓 = 𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓 in
(𝐿1loc (ℝ))𝑛 . This implies that for a.e. 𝑥 ∈ ℝ, 𝜓 satisfies the equation
𝜓 ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)𝜓(𝑥),
or equivalently, the equation
J 𝜓 ′ (𝑥) = 𝐴(𝑥, 𝜔)𝜓(𝑥).
Therefore by Definition 18 this means 𝜓 is a solution of the canonical ODEs in (4.13) with
frequency 𝜔 ∈ Ω. This completes the proof.
Let (𝑎, 𝑏) ⊆ ℝ be a bounded interval. Then we can consider (𝑊 1,1 (𝑎, 𝑏))𝑛 and (𝐿1 (𝑎, 𝑏))𝑛 as
1,1
a subspace of (𝑊loc
(ℝ))𝑛 and (𝐿1loc (ℝ))𝑛 , respectively, by identifying them with the image
of those latter spaces under the surjective map 𝑓 7→ 𝑓 ∣(𝑎,𝑏) . Then we can restrict the domain
of 𝑇 (𝜔) to the subspace (𝑊 1,1 (𝑎, 𝑏))𝑛 and it becomes a differential operator on (𝑊 1,1 (𝑎, 𝑏))𝑛
with range in (𝐿1 (𝑎, 𝑏))𝑛 . It is with this perspective in mind that we introduce and consider
the following frequency dependent operator
𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 := 𝜓 ′ − J −1 𝐴(⋅, 𝜔)∣(𝑎,𝑏) 𝜓,
𝜓 ∈ (𝑊 1,1 (𝑎, 𝑏))𝑛 ,
𝜔 ∈ Ω.
(4.48)
Proposition 53 For each 𝜔 ∈ Ω we have
𝑇 ∣(𝑎,𝑏) (𝜔) : (𝑊 1,1 (𝑎, 𝑏))𝑛 → (𝐿1 (𝑎, 𝑏))𝑛
111
(4.49)
𝑇 ∣(𝑎,𝑏) (𝜔) ∈ L ((𝑊 1,1 (𝑎, 𝑏))𝑛 , (𝐿1 (𝑎, 𝑏))𝑛 )
(4.50)
𝑇 ∣(𝑎,𝑏) ∈ 𝒪(Ω, L ((𝑊 1,1 (𝑎, 𝑏))𝑛 , (𝐿1 (𝑎, 𝑏))𝑛 )).
(4.51)
(𝑇 (𝜔)𝜓)∣(𝑎,𝑏) = 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) ,
(4.52)
Moreover,
1,1
for every 𝜔 ∈ Ω, every 𝜓 ∈ (𝑊loc
(ℝ))𝑛 , and every bounded interval (𝑎, 𝑏) ⊆ ℝ.
Proof.
To be begin we note that by hypothesis 𝐴 ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝕋))). Let (𝑎, 𝑏) ⊆ ℝ)
be any bounded interval. It follows from Lemma 88 that for each 𝜔 ∈ Ω, 𝐴∣(𝑎,𝑏) (⋅, 𝜔) :=
𝐴(⋅, 𝜔)∣(𝑎,𝑏) ∈ 𝑀𝑛 (𝐿1 (𝑎, 𝑏)). It then follows from Lemma 93 that 𝐴∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝑎, 𝑏))).
It then follows from Lemma 89 with 𝑝 = ∞, 𝑞 = 𝑠 = 1 that we have 𝐴˜ := J −1 𝐴∣(𝑎,𝑏) ∈
𝒪(Ω, 𝑀𝑛 (𝐿1 (𝑎, 𝑏))). But according to our definition we have
˜ 𝜔)𝜓,
𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 𝜓 ′ − 𝐴(⋅,
𝜓 ∈ (𝑊 1,1 (𝑎, 𝑏))𝑛 ,
𝜔 ∈ Ω.
with 𝐴˜ ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝑎, 𝑏))). It now follows from the fact ()′ ∈ L ((𝑊 1,1 (𝑎, 𝑏))𝑛 , (𝐿1 (𝑎, 𝑏))𝑛 )
and Lemma 83 that 𝑇 ∣(𝑎,𝑏) (𝜔) : (𝑊 1,1 (𝑎, 𝑏))𝑛 → (𝐿1 (𝑎, 𝑏))𝑛 is a well-defined linear operator
and 𝑇 ∣(𝑎,𝑏) (𝜔) ∈ L ((𝑊 1,1 (𝑎, 𝑏))𝑛 , (𝐿1 (𝑎, 𝑏))𝑛 ), for each 𝜔 ∈ Ω. It follows now from both
Lemma 79 and Lemma 90 that 𝑇 ∣(𝑎,𝑏) ∈ 𝒪(Ω, L ((𝑊 1,1 (𝑎, 𝑏))𝑛 , (𝐿1 (𝑎, 𝑏))𝑛 )).
1,1
Now let 𝜔 ∈ Ω, 𝜓 ∈ (𝑊loc
(ℝ))𝑛 , and (𝑎, 𝑏) ⊆ ℝ) be a bounded interval. Then by Lemma
69 we have 𝜓∣(𝑎,𝑏) ∈ (𝑊 1,1 (𝑎, 𝑏))𝑛 and so 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) ∈ (𝐿1 (𝑎, 𝑏))𝑛 . But (𝑇 (𝜔)𝜓)∣(𝑎,𝑏) ∈
(𝐿1 (𝑎, 𝑏))𝑛 since 𝑇 (𝜔)𝜓 ∈ (𝐿1loc (ℝ))𝑛 and by Lemma 69 we have 𝜓∣′(𝑎,𝑏) = 𝜓 ′ ∣(𝑎,𝑏) . Thus we
conclude that
(𝑇 (𝜔)𝜓)∣(𝑎,𝑏) = (𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓)∣(𝑎,𝑏) = 𝜓∣′(𝑎,𝑏) − J −1 𝐴(⋅, 𝜔)∣(𝑎,𝑏) 𝜓∣(𝑎,𝑏)
112
= 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) .
And hence we have shown that (𝑇 (𝜔)𝜓)∣(𝑎,𝑏) = 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) for every 𝜔 ∈ Ω, every
1,1
𝜓 ∈ (𝑊loc
(ℝ))𝑛 , and every bounded interval (𝑎, 𝑏) ⊆ ℝ. This completes the proof.
Corollary 54 Let 𝜔 ∈ Ω. Then the following statements are equivalent:
(i) 𝜓 is a solution of the canonical ODEs in (4.13) with frequency 𝜔.
(ii) 𝜓 ∈ ker(𝑇 (𝜔)).
1,1
(iii) 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and 𝜓∣(𝑎,𝑏) ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)) every bounded interval (𝑎, 𝑏) ⊆ ℝ.
Proof. We have already proven in Lemma 52 that statement (i) is equivalent to statement
(ii). We now complete the proof of this corollary by proving statement (ii) is equivalent to
1,1
statement (iii). Suppose 𝜓 ∈ ker(𝑇 (𝜔)) for some 𝜔 ∈ Ω. Then 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and for any
bounded interval (𝑎, 𝑏) ⊆ ℝ by Proposition 53 we have 0 = (𝑇 (𝜔)𝜓)∣(𝑎,𝑏) = 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) im1,1
plying 𝜓∣(𝑎,𝑏) ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)). Conversely, if 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and 𝜓∣(𝑎,𝑏) ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)) for
every bounded interval (𝑎, 𝑏) ⊆ ℝ then by Proposition 53 this implies 0 = 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) =
(𝑇 (𝜔)𝜓)∣(𝑎,𝑏) for every bounded interval (𝑎, 𝑏) ⊆ ℝ. But since 𝑇 (𝜔)𝜓 ∈ (𝐿1loc (ℝ))𝑛 this
implies 𝑇 (𝜔)𝜓 = 0 in (𝐿1loc (ℝ))𝑛 and therefore 𝜓 ∈ ker(𝑇 (𝜔)). This completes the proof.
In the following proposition, we collect together some facts from [45, §II.2.5] that will be
need in this chapter.
Proposition 55 For every bounded interval (𝑎, 𝑏) ⊆ ℝ with 0 ∈ (𝑎, 𝑏) and for each 𝜔 ∈ Ω,
there exists a unique function Ψ∣(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) satisfying for a.e. 𝑥 ∈ (𝑎, 𝑏) the
113
matrix differential equation with initial condition
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
(4.53)
Moreover, the following statements are true:
(i) If 𝜓 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)) then there exists a unique 𝛾 ∈ ℂ𝑛 such that 𝜓 = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾.
Conversely, if 𝛾 ∈ ℂ𝑛 and we define 𝜓 := Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾 then 𝜓 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)).
(ii) For every (𝑥, 𝜔) ∈ [𝑎, 𝑏] × Ω,
∫
Ψ∣(𝑎,𝑏) (𝑥, 𝜔) = 𝐼𝑛 +
𝑥
J −1 𝐴(𝑡, 𝜔)Ψ∣(𝑎,𝑏) (𝑡, 𝜔)𝑑𝑡.
(4.54)
0
(iii) As a function of frequency, Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))).
−1
(iv) For every (𝑥, 𝜔) ∈ [𝑎, 𝑏] × Ω, Ψ∣−1
exists and as function of
(𝑎,𝑏) (𝑥, 𝜔) := Ψ∣(𝑎,𝑏) (𝑥, 𝜔)
1,1
position Ψ∣−1
(𝑎, 𝑏)).
(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊
Proof.
Let us fix a bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0. We proceed by proving the
existence and uniqueness portions of this proposition first and then we will prove statements
(i)–(iv).
We begin by reminding the reader that, by the proof of Proposition 53, we have
˜ 𝜔)𝜓,
𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 𝜓 ′ − 𝐴(⋅,
𝜓 ∈ (𝑊 1,1 (𝑎, 𝑏))𝑛 ,
𝜔 ∈ Ω.
˜ 𝜔) := J −1 𝐴(⋅, 𝜔)∣(𝑎,𝑏) for 𝜔 ∈ Ω. Following [45, p.
with 𝐴˜ ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝑎, 𝑏))) where 𝐴(⋅,
69, §II.2.5, Definition 2.5.2], we introduce two new definitions.
Definition 25 Let 𝜔 0 ∈ Ω. A matrix 𝑌0 ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) is called a fundamental matrix
of 𝑇 ∣(𝑎,𝑏) (𝜔 0 )𝜓 = 0 if for each 𝜓 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔 0 )) there exists a 𝛾 ∈ ℂ𝑛 such that 𝜓 = 𝑌0 𝛾.
114
A matrix function 𝑌 : Ω → 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) is called a fundamental matrix function of
𝑇 ∣(𝑎,𝑏) 𝜓 = 0 if 𝑌 (𝜔) is a fundamental matrix of 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 for each 𝜔 ∈ Ω.
We now prove the existence portion of this proposition. By [45, p. 69, §II.2.5, Theorem 2.5.3]
there exists a fundamental matrix function 𝑌 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))) of 𝑇 ∣(𝑎,𝑏) 𝜓 = 0 with
the property that for all 𝜔 ∈ Ω, 𝑌 (𝑎, 𝜔) = 𝐼𝑛 and 𝑌 (⋅, 𝜔) is invertible in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)).
Hence for each 𝜔 ∈ Ω, there exists 𝑌 −1 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) such that 𝑌 −1 (⋅, 𝜔)𝑌 (⋅, 𝜔) =
𝑌 (⋅, 𝜔)𝑌 −1 (⋅, 𝜔) = 𝐼𝑛 . But the functions 𝑌 (⋅, 𝜔), 𝑌 −1 (⋅, 𝜔) : [𝑎, 𝑏] → 𝑀𝑛 (ℂ) are continuous
implying 𝑌 (𝑥, 𝜔)𝑌 −1 (𝑥, 𝜔) = 𝐼𝑛 for all 𝑥 ∈ [𝑎, 𝑏]. Thus 𝑌 (𝑥, 𝜔) is invertible for all 𝑥 ∈ [𝑎, 𝑏]
and 𝑌 −1 (𝑥, 𝜔) = 𝑌 (𝑥, 𝜔)−1 .
Now for each 𝜔 ∈ Ω, we define Ψ∣(𝑎,𝑏) (⋅, 𝜔) := 𝑌 (⋅, 𝜔)𝑌 −1 (0, 𝜔). It follows by [45, p. 71,
§II.2.5, Proposition 2.5.4] that Ψ∣(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) is a fundamental matrix of
𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 0 for each 𝜔 ∈ Ω. It then follows from [45, p. 72, §II.2.5, Corollary 2.5.5] that
Ψ∣(𝑎,𝑏) (⋅, 𝜔) is a solution of the equation
˜ 𝜔)Ψ = 0
Ψ′ − 𝐴(⋅,
in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) for each 𝜔 ∈ Ω. But this means for each 𝜔 ∈ Ω, Ψ∣(𝑎,𝑏) (⋅, 𝜔) satisfies for
a.e. 𝑥 ∈ (𝑎, 𝑏) the matrix differential equation with initial condition
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
(4.55)
This proves existence.
We now prove the uniqueness portion of this proposition. Let 𝜔 ∈ Ω. Suppose Φ ∈
𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) satisfies for a.e. 𝑥 ∈ (𝑎, 𝑏) the matrix differential equation with initial condi-
115
tion
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
(4.56)
It follows that Φ is a solution of the equation
˜ 𝜔)Ψ = 0
Ψ′ − 𝐴(⋅,
in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) with 0 ∈ [𝑎, 𝑏] and Φ(0) = 𝐼𝑛 . This implies by [45, p. 73, §II.2.5, Proposition 2.5.9] that Φ is a fundamental matrix of 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 0. But from the existence
portion of this proof, 𝑌 (⋅, 𝜔) is also a fundamental matrix of 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 0 and so by [45,
p. 71, §II.2.5, Proposition 2.5.4] there exists an invertible matrix 𝐶 ∈ 𝑀𝑛 (ℂ) such that Φ =
𝑌 (⋅, 𝜔)𝐶. Thus since 𝐼𝑛 = Φ(0) = 𝑌 (0, 𝜔)𝐶 this implies Φ = 𝑌 (⋅, 𝜔)𝑌 −1 (0, 𝜔) = Ψ∣(𝑎,𝑏) (⋅, 𝜔).
This proves uniqueness.
Now we will prove statements (i)–(iv) of this proposition. We begin with statement (i). First,
it follows from what we just proved that for each 𝜔 ∈ Ω, Ψ∣(𝑎,𝑏) (⋅, 𝜔) = 𝑌 (⋅, 𝜔)𝑌 −1 (0, 𝜔) ∈
𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) is a fundamental matrix of 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓 = 0. By Definition 25 this means
if 𝜓 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)) then there exists 𝛾 ∈ (ℂ)𝑛 such that 𝜓 = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾. But as
discussed 𝑌 (⋅, 𝜔) is invertible with inverse 𝑌 −1 (⋅, 𝜔)−1 ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)). Uniqueness of the
representation 𝜓 = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾 follows since 𝛾 = 𝑌 (0, 𝜔)𝑌 −1 (⋅, 𝜔)𝜓 in (𝑊 1,1 (𝑎, 𝑏))𝑛 .
Let 𝜔 ∈ Ω, let 𝛾 ∈ (ℂ)𝑛 , and define 𝜓 := Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾. Now because 𝜓 = 𝑌 (⋅, 𝜔)(𝑌 −1 (0, 𝜔)𝛾)
and 𝑌 is a fundamental matrix function of 𝑇 ∣(𝑎,𝑏) 𝜓 = 0, it follows by Definition 25 that
𝜓 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)). This completes the proof of statement (i).
116
The proof of statement (ii) is straightforward. Let (𝑥, 𝜔) ∈ [𝑎, 𝑏] × Ω. Then Ψ := Ψ∣(𝑎,𝑏) (⋅, 𝜔)
satisfies (4.53) and so upon integrating and applying Lemma 70 we arrive at
𝑥
∫
∫
′
𝑥
Ψ (𝑡, 𝜔)𝑑𝑡 = 𝐼𝑛 +
Ψ∣(𝑎,𝑏) (𝑥, 𝜔) = 𝐼𝑛 +
∫0 𝑥
= 𝐼𝑛 +
J −1 𝐴(𝑡, 𝜔)Ψ(𝑡, 𝜔)𝑑𝑡
0
J −1 𝐴(𝑡, 𝜔)Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝑑𝑡.
0
This proves statement (ii).
We complete the proof of this proposition now by proving statements (iii) and (iv). We have
already shown that there exists 𝑌 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))) such that for each 𝜔 ∈ Ω, 𝑌 (⋅, 𝜔)
is invertible in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) and, denoting this inverse in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) by 𝑌 −1 (⋅, 𝜔), we
have for every 𝑥 ∈ [𝑎, 𝑏], 𝑌 (𝑥, 𝜔) is invertible with 𝑌 −1 (𝑥, 𝜔) = 𝑌 (𝑥, 𝜔)−1 and Ψ∣(𝑎,𝑏) (⋅, 𝜔) =
𝑌 (⋅, 𝜔)𝑌 −1 (0, 𝜔). This proves statement (iv) since for every (𝑥, 𝜔) ∈ [𝑎, 𝑏] × Ω, we have
−1
Ψ∣(𝑎,𝑏) (𝑥, 𝜔) is invertible with inverse Ψ∣−1
= 𝑌 (0, 𝜔)𝑌 −1 (𝑥, 𝜔) and
(𝑎,𝑏) (𝑥, 𝜔) := Ψ∣(𝑎,𝑏) (𝑥, 𝜔)
−1
Ψ∣−1
(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) by Lemma 72.
(𝑎,𝑏) (⋅, 𝜔) = 𝑌 (0, 𝜔)𝑌
We will now show Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))). Since 𝑌 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))) and
𝑌 (⋅, 𝜔) is invertible in 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) with the inverse 𝑌 −1 (⋅, 𝜔) for every 𝜔 ∈ Ω, it follows
from [45, p. 66, §II.2.3, Proposition 2.3.3] and [45, p. 7, §I.1.2, Proposition 1.2.5] that
𝑌 −1 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))). It follows from this and Lemma 91 that if we let 𝛿 0 denote the
evaluation map at 𝑥 = 0 then 𝑌 −1 (0, ⋅) = 𝛿 0 𝑌 −1 ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)). But the map 𝜄 : 𝑀𝑛 (ℂ) →
𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) given by (𝜄𝐵)(⋅) := 𝐵 is a continuous linear map and so by Lemma 78 we
have 𝑌 −1 (0, ⋅) = 𝜄𝛿 0 𝑌 −1 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))). Therefore by Lemma 78 and Lemma 72
we have Ψ∣(𝑎,𝑏) = 𝑌 −1 𝜄𝛿 0 𝑌 −1 ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))). This proves statement (iii) and hence
the proof of the proposition is complete.
From the functions Ψ∣(𝑎,𝑏) (⋅, 𝜔) we define a matrix-valued function Ψ : ℝ × Ω → 𝑀𝑛 (ℂ) by
Ψ(𝑥, 𝜔) := Ψ∣(𝑎,𝑏) (𝑥, 𝜔)
117
(4.57)
for each (𝑥, 𝜔) ∈ ℝ × Ω and for any interval (𝑎, 𝑏) containing 0 and 𝑥.
Lemma 56 The matrix-valued function Ψ : ℝ × Ω → 𝑀𝑛 (ℂ) is well-defined. Moreover, for
1,1
(ℝ)) and for any 𝑎, 𝑏 ∈ ℝ with 𝑎 < 0 < 𝑏 we have
each 𝜔 ∈ Ω, Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
Ψ(⋅, 𝜔)∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (⋅, 𝜔).
Proof.
(4.58)
First, fix 𝜔 ∈ Ω and let (𝑢, 𝑣) be an interval such that 0 ∈ (𝑢, 𝑣) ⊆ (𝑎, 𝑏). We will
begin by showing
Ψ∣(𝑎,𝑏) (⋅, 𝜔)∣(𝑢,𝑣) = Ψ∣(𝑢,𝑣) (⋅, 𝜔).
(4.59)
We define 𝑌 := Ψ∣(𝑎,𝑏) (⋅, 𝜔)∣(𝑢,𝑣) so that 𝑌 (0) = 𝐼𝑛 . Then by [45, §II.2.2, Proposition 2.2.1]
we have 𝑌 ∈ 𝑀𝑛 (𝑊 1,1 (𝑢, 𝑣)) and 𝑌 ′ = Ψ∣(𝑎,𝑏) (⋅, 𝜔)′ ∣(𝑢,𝑣) . This implies by the definition of
Ψ∣(𝑎,𝑏) (⋅, 𝜔) in Proposition 55 that 𝑌 is a function in 𝑀𝑛 (𝑊 1,1 (𝑢, 𝑣)) satisfying the for a.e.
𝑥 ∈ (𝑢, 𝑣) the matrix differential equation with initial condition
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
By the uniqueness portion of Proposition 55 we conclude that 𝑌 := Ψ∣(𝑢,𝑣) (⋅, 𝜔). Thus we
have shown that Ψ∣(𝑎,𝑏) (⋅, 𝜔)∣(𝑢,𝑣) = Ψ∣(𝑢,𝑣) (⋅, 𝜔).
We now prove the function Ψ defined by (4.57) is well-defined. Let (𝑥, 𝜔) ∈ ℝ × Ω. Let
(𝑎1 , 𝑏1 ) and (𝑎2 , 𝑏2 ) be any intervals containing 0 and 𝑥. Then there exists 𝑢, 𝑣 such that
(𝑢, 𝑣) = (𝑎1 , 𝑏1 ) ∩ (𝑎2 , 𝑏2 ). Now since 0 and 𝑥 are in the interval (𝑢, 𝑣) then Ψ∣(𝑢,𝑣) (⋅, 𝜔)
and Ψ∣(𝑢,𝑣) (𝑥, 𝜔) are well-defined by Proposition 55. And thus since 0 ∈ (𝑢, 𝑣) ⊆ (𝑎𝑗 , 𝑏𝑗 ) for
𝑗 = 1, 2, we get from (4.59) that
Ψ∣(𝑎1 ,𝑏1 ) (𝑥, 𝜔) = Ψ∣(𝑢,𝑣) (𝑥, 𝜔) = Ψ∣(𝑎2 ,𝑏2 ) (𝑥, 𝜔).
118
This proves the function Ψ is well-defined.
1,1
(ℝ)) for any 𝜔 ∈ Ω its enough to show, by Lemma 71,
Next, to show that Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
that Ψ(⋅, 𝜔)∣(𝑎,𝑏) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) for any interval (𝑎, 𝑏) containing 0. But this follows from
the definition (4.57) since for any 𝑥 ∈ (𝑎, 𝑏)
Ψ(𝑥, 𝜔) = Ψ∣(𝑎,𝑏) (𝑥, 𝜔)
which implies
Ψ(⋅, 𝜔)∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)).
This completes the proof of the lemma.
We are in a position now to prove Proposition 33.
Proof. [Proposition 33] To begin this proof we need to show that the matrix-valued
function Ψ : ℝ × Ω → 𝑀𝑛 (ℂ), as defined by (4.57), for a fixed 𝜔 ∈ Ω satisfies a.e. 𝑥 ∈ ℝ the
matrix differential equation with initial condition
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 ,
1,1
and Ψ(⋅, 𝜔) is the only function in 𝑀𝑛 (𝑊loc
(ℝ)) that does so. After which we will prove that
it has the properties described in Proposition 33.
This first part follows now from the previous lemma since for any 𝜔 ∈ Ω and interval (𝑎, 𝑏)
containing 0 we have Ψ(𝑥, 𝜔)∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (𝑥, 𝜔) for all 𝑥 ∈ (𝑎, 𝑏) and so by Proposition
55 the function Ψ(⋅, 𝜔) satisfies a.e. 𝑥 ∈ (𝑎, 𝑏) the matrix differential equation with initial
119
condition
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
Hence this implies Ψ(⋅, 𝜔) satisfies a.e. 𝑥 ∈ ℝ the matrix differential equation with initial
condition
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
1,1
Lets now prove uniqueness. Fix 𝜔 ∈ Ω. Suppose Ψ1 ∈ 𝑀𝑛 (𝑊loc
(ℝ)) satisfied a.e. 𝑥 ∈ ℝ the
matrix differential equation with initial condition
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
Then for any interval (𝑎, 𝑏) containing 0, Ψ1 ∣(𝑎,𝑏) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) and satisfies a.e. 𝑥 ∈ (𝑎, 𝑏)
the matrix differential equation with initial condition
Ψ′ (𝑥) = J −1 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ(0) = 𝐼𝑛 .
and so by the uniqueness part of Proposition 55 this implies Ψ1 (⋅)∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (⋅, 𝜔). By
definition (4.57) this implies Ψ1 (⋅)∣(𝑎,𝑏) = Ψ(⋅, 𝜔)∣(𝑎,𝑏) . Since this is true for any interval (𝑎, 𝑏)
containing 0 we must have Ψ1 (⋅) = Ψ(⋅, 𝜔). This proves uniqueness.
Now we will prove Propositions 33.(i)–(vi). We start with property (i). Suppose that 𝜓 is
a solution of the canonical ODEs in (4.13) at the frequency 𝜔 ∈ Ω. Then we must prove
1,1
there exists a unique 𝛾 ∈ ℂ𝑛 such that, as elements of (𝑊loc
(ℝ))𝑛 , 𝜓 = Ψ(⋅, 𝜔)𝛾. We begin
by proving the uniqueness statement. Suppose that there did exist 𝛾 ∈ ℂ𝑛 such that 𝜓 =
1,1
1,1
Ψ(⋅, 𝜔)𝛾. It follows from this, since 𝜓 ∈ (𝑊loc
(ℝ))𝑛 and Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) by Lemma
120
70 they have unique locally absolutely continuous representative functions 𝜓(⋅) : ℝ → ℂ𝑛
and Ψ(⋅, 𝜔) : ℝ → 𝑀𝑛 (ℂ) which are continuous, that 𝜓(𝑥) = Ψ(𝑥, 𝜔)𝛾 for all 𝑥 ∈ ℝ. In
particular, since Ψ(0, 𝜔) = 𝐼𝑛 this means 𝜓(0) = Ψ(0, 𝜔)𝛾 = 𝛾. This proves uniqueness of 𝛾.
Lets now prove existence. By Lemma 52 we have 𝜓 ∈ ker(𝑇 (𝜔)). By Corollary 54 we have
𝜓∣(𝑎,𝑏) ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)) for any bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0 and so by Proposition
55.((i)) there exists a 𝛾∣(𝑎,𝑏) ∈ ℂ𝑛 such that 𝜓∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾∣(𝑎,𝑏) . But 0 ∈ (𝑎, 𝑏) and so
by (4.57) we have 𝛾 := 𝜓(0) = 𝜓(0)∣(𝑎,𝑏) = 𝜓∣(𝑎,𝑏) (0) = Ψ∣(𝑎,𝑏) (0, 𝜔)𝛾∣(𝑎,𝑏) = Ψ(0, 𝜔)𝛾∣(𝑎,𝑏) =
𝛾∣(𝑎,𝑏) . Thus from this and Lemma 56 we have 𝜓∣(𝑎,𝑏) = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾 = Ψ(⋅, 𝜔)∣(𝑎,𝑏) 𝛾 =
Ψ(⋅, 𝜔)𝛾∣(𝑎,𝑏) . But since this is true for every bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0 this
1,1
implies as elements of (𝑊loc
(ℝ))𝑛 we have 𝜓 = Ψ(⋅, 𝜔)𝛾. This proves existence.
1,1
Conversely, let 𝛾 ∈ (ℝ)𝑛 and define 𝜓 := Ψ(⋅, 𝜔)𝛾. By Lemma 73 we have 𝜓 ∈ (𝑊loc
(ℝ))𝑛 .
Then for any bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0 we have by Lemma 56 and Proposition
55.(i) that 𝜓∣(𝑎,𝑏) = Ψ(⋅, 𝜔)∣(𝑎,𝑏) 𝛾 = Ψ∣(𝑎,𝑏) (⋅, 𝜔)𝛾 ∈ ker(𝑇 ∣(𝑎,𝑏) (𝜔)). It follows from this and
Proposition 53 that we have (𝑇 (𝜔)𝜓)∣(𝑎,𝑏) = 𝑇 ∣(𝑎,𝑏) (𝜔)𝜓∣(𝑎,𝑏) = 0 in (𝐿1 (𝑎, 𝑏))𝑛 . But this is
true for every bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0 and so it follows this and Lemma 4.47
that 𝑇 (𝜔)𝜓 = 0 in (𝐿1loc (ℝ))𝑛 which means 𝜓 ∈ ker(𝑇 (𝜔)). Therefore by Corollary 54 this
implies 𝜓 is a solution of the canonical ODEs in (4.13) with frequency 𝜔. This completes
the proof of property (i).
Next we will prove property (ii). Let (𝑥, 𝜔) ∈ ℝ × Ω. Take any interval (𝑎, 𝑏) containing 0
and 𝑥. Then by Proposition 55 and the previous lemma we have
∫
𝑥
Ψ(𝑥, 𝜔) = Ψ∣(𝑎,𝑏) (𝑥, 𝜔) = 𝐼𝑛 +
J −1 𝐴(𝑡, 𝜔)Ψ∣(𝑎,𝑏) (𝑡, 𝜔)𝑑𝑡
0
∫ 𝑥
= 𝐼𝑛 +
J −1 𝐴(𝑡, 𝜔)Ψ(𝑡, 𝜔)𝑑𝑡.
0
This proves property (ii).
121
Next we will prove property (iii). That is, we must show for each fixed 𝑥 ∈ ℝ, that Ψ(𝑥, ⋅) ∈
𝒪(Ω, 𝑀𝑛 (ℂ)). To do this we fix 𝑥 ∈ ℝ and choose any bounded interval (𝑎, 𝑏) ⊆ ℝ with
0, 𝑥 ∈ (𝑎, 𝑏). Then by (4.57) we have Ψ(𝑥, ⋅) = Ψ∣(𝑎,𝑏) (𝑥, ⋅). Let 𝛿 𝑥 : 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑛 (ℂ)
be the evaluation map at 𝑥 defined by 𝛿 𝑥 𝐵 := 𝐵(𝑥) – the value of the unique absolutely
continuous representative of 𝐵 ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) at 𝑥. Then by Proposition 55.(iii) we
have Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))) and so by Lemma 91 we have 𝛿 𝑥 Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)).
And thus we have Ψ(𝑥, ⋅) = Ψ∣(𝑎,𝑏) (𝑥, ⋅) = 𝛿 𝑥 Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)) as desired. This proves
property (iii).
Next we prove property (iv). To start, since Ψ(𝑥, ⋅) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)) for each 𝑥 ∈ ℝ, we can
define Ψ𝜔 : ℝ × Ω → 𝑀𝑛 (ℂ) to be the partial derivative with respect to frequency in the
𝑀𝑛 (ℂ) norm of the function Ψ : ℝ × Ω → 𝑀𝑛 (ℂ), i.e.,
Ψ𝜔 (𝑥, 𝜔 0 ) := lim (𝜔 − 𝜔 0 )−1 (Ψ(𝑥, 𝜔) − Ψ(𝑥, 𝜔 0 )), ∀(𝑥, 𝜔 0 ) ∈ ℝ × Ω.
𝜔→𝜔 0
1,1
Our goal is to prove that for any (𝑥, 𝜔 0 ) ∈ ℝ × Ω, Ψ𝜔 (⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) and
∫
Ψ𝜔 (𝑥, 𝜔 0 ) =
𝑥
J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 ) + J −1 𝐴(𝑡, 𝜔 0 )Ψ𝜔 (𝑡, 𝜔 0 )𝑑𝑡.
0
To do this we begin by choosing any bounded interval (𝑎, 𝑏) ⊆ ℝ with 0 ∈ (𝑎, 𝑏). Next, let
ℐ : 𝑀𝑛 (𝐿1 (𝑎, 𝑏)) → 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)) be the integral map defined by
∫
(ℐ𝐵)(𝑥) :=
𝑥
𝐵(𝑡)𝑑𝑡,
𝐵 ∈ 𝑀𝑛 (𝐿1 (𝑎, 𝑏)),
𝑥 ∈ (𝑎, 𝑏).
0
Now by hypothesis we have 𝐴 ∈ 𝒪(Ω, 𝐿1 (𝕋)) and so by Lemma 93 we have 𝐴∣(𝑎,𝑏) ∈
𝒪(Ω, 𝐿1 (𝑎, 𝑏)) and (𝐴∣(𝑎,𝑏) )𝜔 = 𝐴𝜔 ∣(𝑎,𝑏) . It then follows from this and Lemma 89 with
𝑝 = ∞, 𝑞 = 𝑠 = 1 that J −1 𝐴∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝐿1 (𝑎, 𝑏)) and (J −1 𝐴∣(𝑎,𝑏) )𝜔 = J −1 𝐴𝜔 ∣(𝑎,𝑏) .
From this and the fact by Proposition 55.(iii) we have Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))), it follows by Lemma 90 with 𝑝 = 1 that J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑛 (𝐿1 (𝑎, 𝑏))) whose derivative
122
is (J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) )𝜔 = J −1 𝐴𝜔 ∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) + J −1 𝐴∣(𝑎,𝑏) (Ψ∣(𝑎,𝑏) )𝜔 . It follows from this and
Lemma 92 that we must have ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ) ∈ 𝒪(Ω, 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏))) and
(ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ))𝜔 = ℐ(J −1 𝐴𝜔 ∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) + J −1 𝐴∣(𝑎,𝑏) (Ψ∣(𝑎,𝑏) )𝜔 ).
This and Lemma 91 imply that for any 𝑥 ∈ (𝑎, 𝑏), 𝛿 𝑥 ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ))
and
(𝛿 𝑥 ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ))𝜔 = 𝛿 𝑥 ℐ(J −1 𝐴𝜔 ∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) + J −1 𝐴∣(𝑎,𝑏) (Ψ∣(𝑎,𝑏) )𝜔 ).
(4.60)
Now from Proposition 55.(ii) it follows that we have the identity
Ψ(𝑥, ⋅) = Ψ∣(𝑎,𝑏) (𝑥, ⋅) = 𝛿 𝑥 Ψ∣(𝑎,𝑏) = 𝐼𝑛 + 𝛿 𝑥 ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ), ∀𝑥 ∈ (𝑎, 𝑏).
From this and (4.60) it follows that
Ψ𝜔 (𝑥, ⋅) = 𝛿 𝑥 (Ψ∣(𝑎,𝑏) )𝜔 = (𝐼𝑛 + 𝛿 𝑥 ℐ(J −1 𝐴∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) ))𝜔
= 𝛿 𝑥 ℐ(J −1 𝐴𝜔 ∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) + J −1 𝐴∣(𝑎,𝑏) (Ψ∣(𝑎,𝑏) )𝜔 ), ∀𝑥 ∈ (𝑎, 𝑏).
(4.61)
But this implies for any (𝑥, 𝜔 0 ) ∈ (𝑎, 𝑏) × Ω we have Ψ𝜔 (𝑥, 𝜔 0 ) = 𝛿 𝑥 (Ψ∣(𝑎,𝑏) )𝜔 (𝜔 0 ) =
(Ψ∣(𝑎,𝑏) )𝜔 (𝑥, 𝜔 0 ). And hence this shows that for each 𝜔 0 ∈ Ω,
Ψ𝜔 (⋅, 𝜔 0 )∣(𝑎,𝑏) = (Ψ∣(𝑎,𝑏) )𝜔 (⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝑊 1,1 (𝑎, 𝑏)).
(4.62)
But since this is true for any bounded interval (𝑎, 𝑏) ⊆ ℝ with 0 ∈ (𝑎, 𝑏), then by Lemma
1,1
71 for any 𝜔 0 ∈ Ω we have Ψ𝜔 (⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝑊loc
(ℝ)). Moreover, it follows that for any
(𝑥, 𝜔 0 ) ∈ ℝ × Ω we can choose a bounded interval (𝑎, 𝑏) ⊆ ℝ with 0, 𝑥 ∈ (𝑎, 𝑏) so that by
123
(4.57), (4.61), and (4.62) we have
Ψ𝜔 (𝑥, 𝜔 0 ) = 𝛿 𝑥 ℐ(J −1 𝐴𝜔 ∣(𝑎,𝑏) Ψ∣(𝑎,𝑏) + J −1 𝐴∣(𝑎,𝑏) (Ψ∣(𝑎,𝑏) )𝜔 )(𝜔 0 )
∫ 𝑥
J −1 𝐴𝜔 ∣(𝑎,𝑏) (𝑡, 𝜔 0 )Ψ∣(𝑎,𝑏) (𝑡, 𝜔 0 ) + J −1 𝐴∣(𝑎,𝑏) (𝑡, 𝜔 0 )(Ψ∣(𝑎,𝑏) )𝜔 (𝑡, 𝜔 0 )𝑑𝑡
=
∫0 𝑥
=
J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 ) + J −1 𝐴(𝑡, 𝜔 0 )Ψ𝜔 (𝑡, 𝜔 0 )𝑑𝑡.
0
This completes the proof of Proposition 33.(iv).
Next we prove property (v) of Proposition 33. The goal is to prove for every (𝑥, 𝜔) ∈ ℝ × Ω
the matrix Ψ(𝑥, 𝜔) is invertible and, denoting Ψ−1 (𝑥, 𝜔) := Ψ(𝑥, 𝜔)−1 , that Ψ−1 (⋅, 𝜔) ∈
1,1
𝑀𝑛 (𝑊loc
(ℝ)). Well, by Proposition 55.(iv) for any bounded interval (𝑎, 𝑏) ⊆ ℝ containing
0 and any (𝑥, 𝜔) ∈ (𝑎, 𝑏) × Ω we have Ψ∣(𝑎,𝑏) (𝑥, 𝜔) is invertible and denoting Ψ∣−1
(𝑎,𝑏) (𝑥, 𝜔) :=
1,1
Ψ∣(𝑎,𝑏) (𝑥, 𝜔)−1 we have Ψ∣−1
(𝑎, 𝑏)). But by (4.57) and since Ψ(𝑥, 𝜔) =
(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊
Ψ∣(𝑎,𝑏) (𝑥, 𝜔) this implies Ψ−1 (𝑥, 𝜔) = Ψ∣−1
(𝑎,𝑏) (𝑥, 𝜔) for any (𝑥, 𝜔) ∈ (𝑎, 𝑏) × Ω. This proves for
every (𝑥, 𝜔) ∈ ℝ × Ω the matrix Ψ(𝑥, 𝜔) is invertible and for any 𝜔 ∈ Ω, Ψ−1 (⋅, 𝜔)∣(𝑎,𝑏) =
1,1
Ψ∣−1
(𝑎, 𝑏)) for any bounded interval (𝑎, 𝑏) ⊆ ℝ containing 0. And therefore
(𝑎,𝑏) (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊
1,1
by Lemma 71 this proves Ψ−1 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)) for any 𝜔 ∈ Ω. This completes the proof
of property (v) of Proposition 33.
Finally, we will complete the proof of Proposition 33 by proving property (vi). To do this
we must show that for every (𝑥, 𝜔) ∈ ℝ × Ω we have Ψ(𝑥 + 𝑑, 𝜔) = Ψ(𝑥, 𝜔)Ψ(𝑑, 𝜔). Fix
1,1
𝜔 ∈ Ω and define Φ := Ψ(⋅ + 𝑑, 𝜔)Ψ(𝑑, 𝜔)−1 . We now show that Φ ∈ 𝑀𝑛 (𝑊loc
(ℝ)). First,
1,1
1,1
since Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)), it follows from Lemma 74 that Ψ(⋅ + 𝑑, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
1,1
(ℝ)). It now
Second, since Ψ(𝑑, 𝜔)−1 is just a constant matrix then Ψ(𝑑, 𝜔)−1 ∈ 𝑀𝑛 (𝑊loc
1,1
follows from Lemma 73 that the product Φ = Ψ(⋅ + 𝑑, 𝜔)Ψ(𝑑, 𝜔)−1 ∈ 𝑀𝑛 (𝑊loc
(ℝ)). We will
1,1
now show that Φ ∈ 𝑀𝑛 (𝑊loc
(ℝ)) satisfies a.e. the matrix differential equation with initial
124
condition
1,1
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ ∈ 𝑀𝑛 (𝑊loc
(ℝ)), Ψ(0) = 𝐼𝑛 .
1,1
1,1
(ℝ)) denote the translation operator as
(ℝ)) → 𝑀𝑛 (𝑊loc
To do this we let ℒ𝑑 : 𝑀𝑛 (𝑊loc
defined in Lemma 74. Then Φ = ℒ𝑑 (Ψ(⋅, 𝜔)Ψ(𝑑, 𝜔)−1 ). Then by Lemma 74 and Lemma 73
we have
J Φ′ = J (ℒ𝑑 (Ψ(⋅, 𝜔)Ψ(𝑑, 𝜔)−1 ))′ = J ℒ𝑑 (Ψ(⋅, 𝜔)Ψ(𝑑, 𝜔)−1 )′
= J ℒ𝑑 (Ψ(⋅, 𝜔)′ Ψ(𝑑, 𝜔)−1 ) = ℒ𝑑 (J Ψ(⋅, 𝜔)′ )Ψ(𝑑, 𝜔)−1 .
But Proposition 33 implies J Ψ(⋅, 𝜔)′ = 𝐴(⋅, 𝜔)Ψ(⋅, 𝜔) and since by hypothesis 𝐴(⋅ + 𝑑, 𝜔) =
𝐴(⋅, 𝜔) this implies ℒ𝑑 (𝐴(⋅, 𝜔)Ψ(⋅, 𝜔)) = 𝐴(⋅ + 𝑑, 𝜔)Ψ(⋅ + 𝑑, 𝜔) = 𝐴(⋅, 𝜔)ℒ𝑑 (Ψ(⋅, 𝜔)). Hence
these facts imply
J Φ′ = ℒ𝑑 (J Ψ(⋅, 𝜔)′ )Ψ(𝑑, 𝜔)−1 = ℒ𝑑 (𝐴(⋅, 𝜔)Ψ(⋅, 𝜔))Ψ(𝑑, 𝜔)−1
= 𝐴(⋅, 𝜔)ℒ𝑑 (Ψ(⋅, 𝜔))Ψ(𝑑, 𝜔)−1 = 𝐴(⋅, 𝜔)Φ.
But this is an equality of elements in 𝑀𝑛 (𝐿1loc (ℝ)) and so together with the fact that Φ(0) =
1,1
Ψ(0 + 𝑑, 𝜔)Ψ(𝑑, 𝜔)−1 = 𝐼𝑛 this implies that Φ ∈ 𝑀𝑛 (𝑊loc
(ℝ)) satisfies a.e. the matrix
differential equation with initial condition
1,1
J Ψ′ (𝑥) = 𝐴(𝑥, 𝜔)Ψ(𝑥), Ψ ∈ 𝑀𝑛 (𝑊loc
(ℝ)), Ψ(0) = 𝐼𝑛 .
Now though the uniqueness portion of Proposition 33 implies that we must have Ψ(⋅, 𝜔) = Φ.
Thus Ψ(𝑥 + 𝑑, 𝜔) = Ψ(𝑥, 𝜔)Ψ(𝑑, 𝜔) for a.e. 𝑥 ∈ ℝ and so, since Proposition 33.(ii) implies
the function Ψ(𝑥, 𝜔) is continuous as a function of 𝑥 ∈ ℝ, we must have Ψ(𝑥 + 𝑑, 𝜔) =
Ψ(𝑥, 𝜔)Ψ(𝑑, 𝜔) for all 𝑥 ∈ ℝ. This proves Proposition 33.(vi) and therefore completes the
125
proof of Proposition 33.
We will use Proposition 33 now to prove the Floquet-Lyapunov Theorem.
Proof. [Theorem 34] We fix 𝜔 ∈ Ω. To simplify notation we will denote the matricant
by Φ := Ψ(⋅, 𝜔) and we will agree to treat, in the rest of this proof, all objects as depending
implicitly on the frequency 𝜔. Our goal here is to show that there exists a matrix 𝐾 ∈ 𝑀𝑛 (ℂ)
1,1
and a function 𝐹 ∈ 𝑀𝑛 (𝑊loc
(ℝ)) such that for every 𝑥 ∈ ℝ,
Φ(𝑥) = 𝐹 (𝑥)𝑒𝑖𝑥𝐾 ,
1,1
where 𝐹 (𝑥 + 𝑑) = 𝐹 (𝑥), 𝐹 (0) = 𝐼𝑛 , 𝐹 −1 (𝑥) := 𝐹 (𝑥)−1 exist, and 𝐹 −1 ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
To begin, let 𝜆1 , . . . , 𝜆𝑠 denote the distinct eigenvalues of the monodromy matrix Φ(𝑑). Let
log(𝑧) be any branch of the logarithm which is analytic at each 𝜆𝑗 , for 𝑗 = 1, . . . , 𝑠. Using
the Riesz-Dunford functional calculus as described for matrices in [35, pp. 304–334, §9], it
follows that the matrix 𝐾 :=
1
𝑖𝑑
log(Φ(𝑑)) satisfies
𝑒𝑖𝑑𝐾 = Φ(𝑑),
where 𝑒𝑧 is the exponential function. From the functional calculus it follows that the matrix
1,1
function 𝐵(𝑥) := 𝑒𝑖𝑥𝐾 , for 𝑥 ∈ ℝ has the properties 𝐵 ∈ 𝑀𝑛 (𝑊loc
(ℝ)), 𝐵(𝑥) is invertible
1,1
with inverse 𝐵 −1 (𝑥) := 𝑒−𝑖𝑥𝐾 for 𝑥 ∈ ℝ, and 𝐵 −1 ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
We define the function 𝐹 by 𝐹 (𝑥) := Φ(𝑥)𝑒−𝑖𝑥𝐾 , 𝑥 ∈ ℝ. It then follows from Lemma
1,1
73 that 𝐹 ∈ 𝑀𝑛 (𝑊loc
(ℝ)). We also find that for every 𝑥 ∈ ℝ, 𝐹 (𝑥) is invertible with
inverse 𝐹 −1 (𝑥) := 𝑒𝑖𝑥𝐾 Φ(𝑥)−1 . It follows from Proposition 33.(v) and Lemma 73 that 𝐹 −1 ∈
1,1
𝑀𝑛 (𝑊loc
(ℝ)). Furthermore, since Φ(0) = 𝐼𝑛 = 𝑒−𝑖0𝐾 then 𝐹 (0) = 𝐼𝑛 . Moreover, since
𝑒𝑖𝑑𝐾 = Φ(𝑑), 𝑒−𝑖𝑑𝐾 = Φ(𝑑)−1 , and, by Proposition 33.(vi), Φ(𝑥 + 𝑑) = Φ(𝑥)Φ(𝑑) for every
126
𝑥 ∈ ℝ, we have
𝐹 (𝑥 + 𝑑) = Φ(𝑥 + 𝑑)𝑒−𝑖(𝑥+𝑑)𝐾 = Φ(𝑥)Φ(𝑑)𝑒−𝑖𝑑𝐾 𝑒−𝑖𝑥𝐾 = Φ(𝑥)Φ(𝑑)Φ(𝑑)−1 𝑒−𝑖𝑥𝐾
= Φ(𝑥)𝑒−𝑖𝑥𝐾 = 𝐹 (𝑥), ∀𝑥 ∈ ℝ.
This completes the proof of Theorem 34.
Using the Floquet-Lyapunov Theorem and its proof we will now establish the validity of
Theorem 35.
Proof. [Theorem 35] We first prove the statement: if 𝜓 is a nontrivial Floquet solution
of the canonical ODEs in (4.13) with wavenumber-frequency pair (𝑘, 𝜔), Floquet multiplier
𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚 ∈ ℕ ∪ {0} then 𝜓 = Ψ(⋅, 𝜔)𝛾 where 𝛾 is a generalized eigenvector of
the monodromy matrix Ψ(𝑑, 𝜔) of order 𝑚 + 1 corresponding to the eigenvalue 𝜆. We prove
this by induction on the order 𝑚 of the Floquet solution. Before we begin we will prove the
following lemma:
1,1
𝑛
1
𝑛
Lemma 57 If {𝑢𝑗 }𝑚
𝑗=0 ⊆ (𝑊loc (ℝ)) ∩ (𝐿 (𝕋)) and for a.e. 𝑥 ∈ ℝ,
𝑚
∑
𝑥𝑗 𝑢𝑗 (𝑥) = 0
𝑗=0
then 𝑢1 = ⋅ ⋅ ⋅ = 𝑢𝑚 = 0.
Proof. We prove the statement by induction. The statement is obviously true for 𝑚 = 0.
Suppose the statement is true for 𝑚 ∈∈ ℕ ∪ {0}. Lets show its true for 𝑚 + 1. Suppose
1,1
𝑛
1
𝑛
{𝑢𝑗 }𝑚+1
𝑗=0 ⊆ (𝑊loc (ℝ)) ∩ (𝐿 (𝕋)) and satisfies for a.e. 𝑥 ∈ ℝ,
𝑚+1
∑
𝑥𝑗 𝑢𝑗 (𝑥) = 0
𝑗=0
127
𝑛
By Lemma 70 we can assume {𝑢𝑗 }𝑚+1
𝑗=0 ⊆ (𝐴𝐶loc (ℝ)) and satisfy 𝑢𝑗 (𝑥 + 𝑑) = 𝑢𝑗 (𝑥) for all
𝑥 ∈ ℝ and for 𝑗 = 0, . . . , 𝑚 + 1. Hence letting 𝑢𝑗,𝑖 denote the 𝑖th row entry of 𝑢𝑗 we have
𝑢𝑗,𝑖 ∈ 𝐴𝐶loc (ℝ) is a 𝑑-periodic function and hence is bounded on ℝ, for 𝑗 = 0, . . . , 𝑚 + 1,
𝑖 = 0, . . . , 𝑛. This implies
𝑥𝑗 𝑢𝑗,𝑖 (𝑥)
=0
𝑥→∞ 𝑥𝑚+1
lim
for 𝑗 = 0, . . . , 𝑚, 𝑖 = 0, . . . , 𝑛 and there exists 𝑥𝑖 ∈ [0, 𝑑) such that
𝑢𝑚+1,𝑖 (𝑥𝑖 ) = sup ∣𝑢𝑚+1,𝑖 (𝑥)∣
𝑥∈ℝ
for 𝑖 = 0, . . . , 𝑛. Now it follows from our hypothesis that
𝑚+1
∑
𝑥𝑗 𝑢𝑗,𝑖 (𝑥) = 0, ∀𝑥 ∈ ℝ,
𝑖 = 1, . . . , 𝑛.
𝑗=0
Thus we conclude for 𝑙 ∈ ℕ,
sup ∣𝑢𝑚+1,𝑖 (𝑥)∣ = 𝑢𝑚+1,𝑖 (𝑥𝑖 ) = lim 𝑢𝑚+1,𝑖 (𝑥𝑖 + 𝑙𝑑)
𝑙→∞
𝑥∈ℝ
= lim −
𝑙→∞
𝑚
∑
𝑗=0
(𝑥𝑖 + 𝑙𝑑)𝑗 𝑢𝑗,𝑖 (𝑥𝑖 + 𝑙𝑑)
=0
(𝑥𝑖 + 𝑙𝑑)𝑚+1
for 𝑖 = 0, . . . , 𝑛. This implies 𝑢𝑚+1 = 0. But now the sequence {𝑢𝑗 }𝑚
𝑗=0 satisfies the hypotheses of this lemma and so by the induction hypotheses we conclude that 𝑢𝑚 = 𝑢𝑚 − 1 = ⋅ ⋅ ⋅ =
𝑢0 = 0. Therefore by induction the statement is true for all 𝑚 ∈ ℕ ∪ {0}. This completes
the proof.
We now begin the proof of the above statement. Let us show the base case 𝑚 = 0 is
true. Suppose 𝜓 is a Floquet solution of the canonical ODEs in (4.13) with wavenumberfrequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚 = 0. Then by Definition
128
1,1
(ℝ))𝑛 ∩ (𝐿1 (𝕋))𝑛 with 𝑢0 ∕= 0. By
19, 𝜓(𝑥) = 𝑒𝑖𝑘𝑥 𝑢0 (𝑥) for a.e. 𝑥 ∈ ℝ, for some 𝑢0 ∈ (𝑊loc
Lemma 57 we conclude 𝜓 ∕= 0 and so by Proposition 33.(i), 𝜓 = Ψ(⋅, 𝜔)𝛾 for some nonzero
𝛾 ∈ ℂ𝑛 . These facts and Proposition 33.(vi) imply for a.e. 𝑥 ∈ ℝ, Ψ(𝑥, 𝜔)Ψ(𝑑, 𝜔)𝛾 =
𝜓(𝑥 + 𝑑) = 𝑒𝑖𝑘𝑥 𝑒𝑖𝑘𝑑 𝑢0 (𝑥 + 𝑑) = 𝜆𝑒𝑖𝑘𝑑 𝑢0 (𝑥) = 𝜆𝜓(𝑥) = Ψ(𝑥, 𝜔)𝜆𝛾. Hence Proposition 33.(v)
implies Ψ(𝑑, 𝜔)𝛾 = 𝜆𝛾. But this implies, since 𝛾 ∕= 0, that 𝛾 is a generalized eigenvector of
Ψ(𝑑, 𝜔) of order 1 corresponding to the eigenvalue 𝜆. This proves the base case.
Suppose now the statement is true for 𝑚 ∈ ℕ ∪ {0}. We will now show its true for 𝑚 + 1.
Let 𝜓 be a Floquet solution of the canonical ODEs in (4.13) with wavenumber-frequency
pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚 + 1. Then by Definition 19, 𝜓(𝑥) =
∑
1,1
𝑚+1
𝑗
𝑛
1
𝑛
𝑒𝑖𝑘𝑥 𝑚+1
𝑗=0 𝑥 𝑢𝑗 (𝑥) for a.e. 𝑥 ∈ ℝ, for some {𝑢𝑗 }𝑗=0 ⊆ (𝑊loc (ℝ)) ∩ (𝐿 (𝕋)) with 𝑢𝑚+1 ∕= 0.
By Lemma 57 we conclude 𝜓 ∕= 0 and so by Proposition 33.(i), 𝜓 = Ψ(⋅, 𝜔)𝛾 for some nonzero
𝛾 ∈ ℂ𝑛 . We define
𝛾˜ := (Ψ(𝑑, 𝜔) − 𝜆𝐼𝑛 )𝛾,
˜ := Ψ(⋅, 𝜔)˜
𝜓
𝛾.
˜ is a Floquet solution of the canonical ODEs in (4.13) with wavenumberWe will now prove 𝜓
frequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚. First, it follows by Proposi˜ is a solution of the canonical ODEs in (4.13) at the frequency 𝜔 ∈ Ω. Next,
tion 33.(i) that 𝜓
∑ ()
by Proposition 33.(vi), the Floquet representation of 𝜓, and the fact (𝑥+𝑑)𝑗 = 𝑗𝑙=0 𝑗𝑙 𝑑𝑗−𝑙 𝑥𝑙 ,
it follows for a.e. 𝑥 ∈ ℝ that
˜
𝜓(𝑥)
= Ψ(𝑥, 𝜔)(Ψ(𝑑, 𝜔) − 𝑒𝑖𝑘𝑑 𝐼𝑛 )𝛾 = 𝜓(𝑥 + 𝑑) − 𝑒𝑖𝑘𝑑 𝜓(𝑥)
=𝑒
𝑖(𝑘+𝑑)𝑥
𝑚+1
∑
𝑗
(𝑥 + 𝑑) 𝑢𝑗 (𝑥 + 𝑑) − 𝑒
𝑗=0
𝑖𝑘𝑑 𝑖𝑘𝑥
𝑒
𝑚+1
∑
𝑥𝑗 𝑢𝑗 (𝑥)
𝑗=0
𝑗−1 ( )
𝑚+1
∑∑
(
) 𝑖𝑘𝑑
𝑗 𝑗−𝑙 𝑙 𝑖𝑘𝑑
𝑗
𝑗
𝑖𝑘𝑥
=𝑒
(𝑥 + 𝑑) − 𝑥 𝑒 𝑢𝑗 (𝑥) = 𝑒
𝑑 𝑥 𝑒 𝑢𝑗 (𝑥)
𝑙
𝑗=0
𝑗=1 𝑙=0
( 𝑚 ( )
)
𝑚
𝑚
∑
∑ 𝑗
∑
𝑖𝑘𝑥
𝑙
𝑗+1−𝑙 𝑖𝑘𝑑
𝑖𝑘𝑥
=𝑒
𝑥
𝑑
𝑒 𝑢𝑗+1 (𝑥) = 𝑒
𝑥𝑙 𝑢˜𝑗 (𝑥),
𝑙
𝑙=0
𝑗=𝑙
𝑙=0
𝑖𝑘𝑥
𝑚+1
∑
129
where 𝑢˜𝑗 =
∑𝑚 (𝑗 ) 𝑗+1−𝑙 𝑖𝑘𝑑
1,1
(ℝ))𝑛 ∩ (𝐿1 (𝕋))𝑛 for 𝑗 = 0, . . . , 𝑚 and 𝑢˜𝑚 =
𝑒 𝑢𝑗+1 ∈ (𝑊loc
𝑗=𝑙 𝑙 𝑑
˜ is a Floquet solution of the canonical ODEs in (4.13) with
𝑑𝑒𝑖𝑘𝑑 𝑢𝑚+1 ∕= 0. But this proves 𝜓
wavenumber-frequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚. Hence by the
induction hypothesis and the uniqueness portion of Proposition 33.(i) we conclude that 𝛾˜ is
a generalized eigenvector of Ψ(𝑑, 𝜔) of order 𝑚 + 1 corresponding to the eigenvalue 𝜆. But
since 𝛾˜ = (Ψ(𝑑, 𝜔) − 𝜆𝐼𝑛 )𝛾 this implies 𝛾 is a generalized eigenvector of Ψ(𝑑, 𝜔) of order
𝑚 + 2 corresponding to the eigenvalue 𝜆. This proves the statement for 𝑚 + 1. Therefore by
induction the statement is true for all 𝑚 ∈ ℕ ∪ {0}.
To complete the proof of this theorem we must show that if 𝛾 is a generalized eigenvector
of Ψ(𝑑, 𝜔) of order 𝑚 + 1 corresponding to the eigenvalue 𝜆 then for any 𝑘 ∈ ℂ such
that 𝜆 = 𝑒𝑖𝑘𝑑 , 𝜓 = Ψ(⋅, 𝜔)𝛾 is a Floquet solution of the canonical ODEs in (4.13) with
wavenumber-frequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚.
Let 𝛾 be a generalized eigenvector of Ψ(𝑑, 𝜔) of order 𝑚 + 1 corresponding to the eigenvalue
𝜆 where 𝑚 ∈ ℕ ∪ {0}. Define 𝜓 := Ψ(⋅, 𝜔)𝛾. By Proposition 33.(i), 𝜓 is a nontrivial solution
of the canonical ODEs in (4.13) at the frequency 𝜔 ∈ Ω. We will now show that 𝜓 is a
Floquet solution with the desired properties.
To simplify notation we will denote the matricant by Φ := Ψ(⋅, 𝜔). Then by the Floquet1,1
Lyapunov Theorem there exists a matrix 𝐾 ∈ 𝑀𝑛 (ℂ) and a function 𝐹 ∈ 𝑀𝑛 (𝑊loc
(ℝ)) such
that for every 𝑥 ∈ ℝ,
Φ(𝑥) = 𝐹 (𝑥)𝑒𝑖𝑥𝐾 ,
1,1
where 𝐹 (𝑥 + 𝑑) = 𝐹 (𝑥), 𝐹 (0) = 𝐼𝑛 , 𝐹 −1 (𝑥) := 𝐹 (𝑥)−1 exist, and 𝐹 −1 ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
Moreover, by the proof of the Floquet-Lyapunov Theorem we can assume that
𝐾=
1
log(Φ(𝑑))
𝑖𝑑
130
where, letting 𝜆1 , . . . , 𝜆𝑠 denote the distinct eigenvalues of the monodromy matrix Φ(𝑑),
log(𝑧) is a branch of the logarithm which is analytic at each 𝜆𝑙 , for 𝑙 = 1, . . . , 𝑠. Fix an
𝑥 ∈ ℝ. We define the function
𝑥
𝑓 (𝑧) := 𝑒 𝑑 𝑙𝑜𝑔(𝑧)
with the same domain as the function log(𝑧). It follows that 𝑓 is analytic at eigenvalues of
the monodromy matrix Φ(𝑑) as well. By the Riesz-Dunford functional calculus as described
1
for matrices in [35, pp. 304–334, §9] we can define the matrix 𝑓 (Φ(𝑑)). But 𝑓 (𝑧) = 𝑒𝑖𝑥( 𝑖𝑑 𝑙𝑜𝑔(𝑧))
and so it follows by [35, pp. 324, §9.7, Theorem 2] that
𝑓 (Φ(𝑑)) = 𝑒𝑖𝑥𝐾 .
For each 𝑙 = 1, . . . , 𝑠, let 𝑃𝑙 ∈ 𝑀𝑛 (ℂ) denote the projection onto the generalized eigenspace of
the monodromy matrix Φ(𝑑) corresponding to the eigenvalue 𝜆𝑙 . Without loss of generality
we may assume 𝜆1 = 𝜆. Let 𝑚𝑙 denote the index of the eigenvalue 𝜆𝑙 for 𝑙 = 1, . . . , 𝑠. Let
𝑓𝑙,𝑗 denote the value of the 𝑗th derivative of 𝑓 (𝑧) at the eigenvalue 𝜆𝑙 for 𝑙 = 1, . . . , 𝑠 and
𝑗 = 0, . . . , 𝑚𝑘 − 1. By spectral resolution of 𝑓 (Φ(𝑑)) (see [35, p. 314, §9.5, Theorem 1], [35,
p. 319, §9.5, Theorem 3], and [35, p. 321, §9.6, Theorem 1]) we have
𝑓 (Φ(𝑑)) =
𝑠 𝑚
𝑙 −1
∑
∑
𝑓𝑙,𝑗
𝑙=1 𝑗=0
𝑗!
(Φ(𝑑) − 𝜆𝑙 𝐼𝑛 )𝑗 𝑃𝑙
Now since by hypothesis 𝛾 is a generalized eigenvector of Φ(𝑑) of order 𝑚 + 1 corresponding
to the eigenvalue 𝜆1 this implies
𝑃1 𝛾 = 𝛾,
𝑃𝑙 𝛾 = 0, for 𝑙 ∕= 1,
(Φ(𝑑) − 𝜆1 𝐼𝑛 )𝑚 𝛾 ∕= 0, (Φ(𝑑) − 𝜆1 𝐼𝑛 )𝑗 𝛾 = 0, for 𝑗 > 𝑚.
131
Thus we have
𝑓 (Φ(𝑑))𝛾 =
𝑚
∑
𝑓1,𝑗
𝑗=0
𝑗!
(Φ(𝑑) − 𝜆1 𝐼𝑛 )𝑗 𝛾,
𝑓1,𝑗 =
𝑑𝑗 𝑓 .
𝑑𝑧 𝑗 𝑧=𝜆1
But we have
)
(𝑥
) 𝑥
(𝑥) (𝑥
𝑑𝑗 𝑓 − 1 ⋅⋅⋅
− (𝑗 − 1) 𝑒( 𝑑 −(𝑗−1)) log(𝜆1 ) ,
=
𝑗
𝑑𝑧 𝑧=𝜆1
𝑑
𝑑
𝑑
𝑗 = 1, . . . , 𝑚.
We define the polynomials in the variable 𝑥 by
𝑝0 (𝑥) := 1,
𝑝𝑗 (𝑥) :=
(𝑥) (𝑥
𝑑
)
(𝑥
)
− 1 ⋅⋅⋅
− (𝑗 − 1) 𝑒−(𝑗−1) log(𝜆1 ) .
𝑑
𝑑
Choose any 𝑘 ∈ ℂ such that
𝜆1 = 𝑒𝑖𝑘𝑑 .
Then, since 𝜆1 = 𝑒log(𝜆1 ) which follows from the fact log(𝑧) is a branch of the logarithm
analytic at 𝜆1 , there exist 𝑞 ∈ ℤ such that
𝑖𝑘𝑑 = log(𝜆1 ) + 𝑖2𝜋𝑞.
From these facts it follows that
𝑒
𝑖𝑥𝐾
𝛾 = 𝑓 (Φ(𝑑))𝛾 =
𝑚
∑
𝑓1,𝑗
𝑗=0
= 𝑒𝑖𝑘𝑥
𝑚
∑
𝑝𝑗 (𝑥)
𝑗=0
𝑗!
𝑗!
𝑗
(Φ(𝑑) − 𝜆1 𝐼𝑛 ) 𝛾 =
𝑚
∑
𝑗=0
𝑒−𝑖𝑥2𝜋𝑞/𝑑 (Φ(𝑑) − 𝜆1 𝐼𝑛 )𝑗 𝛾.
132
𝑥
𝑒 𝑑 log(𝜆1 )
𝑝𝑗 (𝑥)
(Φ(𝑑) − 𝜆1 𝐼𝑛 )𝑗 𝛾
𝑗!
And from this we conclude
𝜓(𝑥) = Φ(𝑥)𝛾 = 𝐹 (𝑥)𝑒𝑖𝑥𝐾 𝛾 = 𝑒𝑖𝑘𝑥
𝑚
∑
𝑝𝑗 (𝑥)
𝑗=0
𝑗!
𝑒−𝑖𝑥2𝜋𝑞/𝑑 𝐹 (𝑥)(Φ(𝑑) − 𝜆𝐼𝑛 )𝑗 𝛾,
(4.63)
1,1
for every 𝑥 ∈ ℝ. But 𝑣𝑗 (⋅) := 𝑗!1 𝑒−𝑖(⋅)2𝜋𝑞/𝑑 𝐹 (⋅)(Φ(𝑑) − 𝜆𝐼𝑛 )𝑗 𝛾 ∈ (𝑊loc
(ℝ))𝑛 ∩ (𝐿1 (𝕋))𝑛 and the
polynomial 𝑝𝑗 (⋅) has degree 𝑗, for 𝑗 = 0, . . . , 𝑚. Moreover, 𝑣𝑚 (0) =
1
(Φ(𝑑)
𝑚!
− 𝜆𝐼𝑛 )𝑚 𝛾 ∕= 0
and 𝜆 = 𝑒𝑖𝑘𝑑 . From these facts and the representation (4.63) we conclude that for any 𝑘 ∈ ℂ
such that 𝜆 = 𝑒𝑖𝑘𝑑 , 𝜓 is Floquet solution of the canonical ODEs in (4.13) with wavenumberfrequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 𝑚. This completes the proof of
Proposition 35.
Proof. [Corollary 36] Fix an 𝜔 ∈ Ω. Let {𝛾 𝑗 }𝑛𝑗=1 be a Jordan basis for the matrix Ψ(𝑑, 𝜔).
Then for 𝑗 = 1, . . . , 𝑛, we can define 𝑙𝑗 , 𝜆𝑗 , and 𝑘𝑗 to be such that 𝛾 𝑗 is a generalized
eigenvalue of Ψ(𝑑, 𝜔) of order 𝑙𝑗 corresponding to the eigenvalue 𝜆𝑗 = 𝑒𝑖𝑘𝑗 𝑑 . We define
{𝜓 𝑗 }𝑛𝑗=1 by 𝜓 𝑗 := Ψ(⋅, 𝜔)𝛾 𝑗 for 𝑗 = 1, . . . , 𝑛. Then since {𝛾 𝑗 }𝑛𝑗=1 is a basis for (ℂ)𝑛 it follows
by 33.(i) and 33.(v) that the set of solutions of the canonical ODEs in (4.13) at the frequency
𝜔 is a vector space over ℂ and {𝜓 𝑗 }𝑛𝑗=1 is a basis for that space. Moreover, by Theorem 35,
𝜓 𝑗 is a Floquet solution of the canonical ODEs in (4.13)with wavenumber-frequency pair
(𝑘𝑗 , 𝜔), Floquet multiplier 𝜆𝑗 = 𝑒𝑖𝑘𝑗 𝑑 , and order 𝑙𝑗 − 1, for 𝑗 = 1, . . . , 𝑛. This completes the
proof.
Proof. [Corollary 37 & Corollary 38] Let ℬ denote the Bloch variety of the canonical
ODEs in (4.13), i.e.,
ℬ := {(𝑘, 𝜔) ∈ ℂ × Ω ∣ (𝑘, 𝜔) is the wavenumber-frequency pair of some nontrivial
Bloch solution of the canonical ODEs in (4.13)}.
133
Define the function 𝐷 : ℂ × Ω → ℂ by
(
)
𝐷(𝑘, 𝜔) := det 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)
and the set 𝒞 ⊆ ℂ × Ω by
𝒞 := {(𝑘, 𝜔) ∈ ℂ × Ω ∣ 𝐷(𝑘, 𝜔) = 0}.
Now Corollary 37 follows from Corollary 38 since 𝐷(𝑘 + 2𝜋/𝑑, 𝜔) = 𝐷(𝑘, 𝜔). Hence to prove
these corollaries we need only show the function 𝐷 is a nonconstant holomorphic function
and ℬ = 𝒞.
We will now prove ℬ = 𝒞. Let (𝑘, 𝜔) ∈ ℬ. Then there exists a 𝜓 ∕= 0 which is a Bloch
solution of the canonical ODEs in (4.13) with wavenumber-frequency pair (𝑘, 𝜔). But this
implies by Definition 19 that 𝜓 is a Floquet solution of the canonical ODEs in (4.13) with
wavenumber-frequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 1. By Theorem 35
this implies 𝜓 = Ψ(⋅, 𝜔)𝛾 where 𝛾 is an eigenvector of the monodromy matrix Ψ(⋅, 𝜔) with
(
)
𝜆 = 𝑒𝑖𝑘𝑑 as its corresponding eigenvalue. Thus we must have det 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔) = 0 and
hence (𝑘, 𝜔) ∈ 𝒞. This proves ℬ ⊆ 𝒞.
(
)
To prove the reverse inclusion, let (𝑘, 𝜔) ∈ 𝒞. Then det 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔) = 0 and this
implies 𝜆 = 𝑒𝑖𝑘𝑑 is an eigenvalue of the monodromy matrix Ψ(𝑑, 𝜔). Let 𝛾 be an eigenvector of
the monodromy matrix Ψ(𝑑, 𝜔) corresponding to the eigenvalue 𝜆 = 𝑒𝑖𝑘𝑑 . Then Theorem 35
implies 𝜓 := Ψ(⋅, 𝜔)𝛾 is a Floquet solution of the canonical ODEs in (4.13) with wavenumberfrequency pair (𝑘, 𝜔), Floquet multiplier 𝜆 = 𝑒𝑖𝑘𝑑 , and order 1. This and Lemma 57 imply 𝜓
is a nontrivial Bloch solution of the canonical ODEs in (4.13) with wavenumber-frequency
pair (𝑘, 𝜔) and hence (𝑘, 𝜔) ∈ ℬ. This proves 𝒞 ⊆ ℬ. Therefore ℬ = 𝒞.
134
We will now prove the function 𝐷 : ℂ × Ω → ℂ is a nonconstant function. First, 𝑓 (𝜆, 𝜔) :=
det (𝜆𝐼𝑛 − Ψ(𝑑, 𝜔)) is a monic polynomial of degree 𝑛 in the variable 𝜆 for each 𝜔 ∈ ℂ. Thus
𝑓 : ℂ × Ω → ℂ is a nonconstant function. Now 𝐷(𝑘, 𝜔) = 𝑓 (𝑒𝑖𝑘𝑑 , 𝜔) and this implies 𝐷 must
be a nonconstant function for otherwise we could fix an 𝜔 0 ∈ Ω and then there would exist
𝑐 ∈ ℂ such that 𝑓 (𝑒𝑖𝑘𝑑 , 𝜔 0 ) − 𝑐 = 0 for every 𝑘 ∈ ℂ implying, since 𝑓 (𝜆, 𝜔 0 ) − 𝑐 is a monic
polynomial of degree 𝑛 in the variable 𝜆, the function 𝑔(𝑘) := 𝑒𝑖𝑘𝑑 maps into a finite set of
values in ℂ which is a contradiction. Thus we have proven the function 𝐷 is a nonconstant
function.
We complete the proof of these corollaries now by proving the function 𝐷 : ℂ × Ω → ℂ
is a holomorphic function. By Lemma 80 we have 𝒪(Ω, 𝑀𝑛 (ℂ)) = 𝑀𝑛 (𝒪(Ω, ℂ)). Thus
by Proposition 33.(iii) we have Ψ(𝑑, ⋅) ∈ 𝑀𝑛 (𝒪(Ω, ℂ)), i.e., the entries of the monodromy
matrix function Ψ(𝑑, ⋅) are holomorphic functions of frequency. This implies the entries of
the matrix 𝑀 (𝑘, 𝜔) := 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔) are holomorphic functions of the variable (𝑘, 𝜔) in
the domain ℂ × Ω. And since the function 𝐹 (𝐵) := det(𝐵) is a polynomial in the entries of
the matrix 𝐵 ∈ 𝑀𝑛 (ℂ) this implies 𝐷(𝑘, 𝜔) = 𝐹 (𝑀 (𝑘, 𝜔)) is a holomorphic functions of the
variable (𝑘, 𝜔) in the domain ℂ × Ω. This completes the proof.
Proofs for Section 4.2.2
We begin with proving a key result in the study of the energy flux for canonical ODEs.
Proof. [Proposition 39] Fix an 𝜔 ∈ Ωℝ . Then by Proposition 33 we have Ψ(⋅, 𝜔) ∈
1,1
1,1
𝑀𝑛 (𝑊loc
(ℝ)), by Lemma 75 we have Ψ(⋅, 𝜔)∗ ∈ 𝑀𝑛 (𝑊loc
(ℝ)), and hence by Lemma 73 we
1,1
(ℝ)). For simplification of notation let Φ := Ψ(⋅, 𝜔).
have Ψ(⋅, 𝜔)∗ 𝑖J Ψ(⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
1,1
To complete the proof we must show Φ∗ 𝑖J Φ = 𝑖J in 𝑀𝑛 (𝑊loc
(ℝ)). First, it follows from
Proposition 33 that Φ′ = J −1 𝐴(⋅, 𝜔)Φ in 𝑀𝑛 (𝐿1loc (ℝ)). Second, by our assumptions we
135
have (J −1 )∗ = −J −1 and 𝐴(⋅, 𝜔)∗ = 𝐴(⋅, 𝜔) in 𝑀𝑛 (𝐿1loc (ℝ)). Thus by these facts and by
Lemmas 75 and 73 we have
(Φ∗ 𝑖J Φ)′ = (Φ′ )∗ 𝑖J Φ + Φ∗ 𝑖J Φ′ = (J −1 𝐴(⋅, 𝜔)Φ)∗ 𝑖J Φ + Φ∗ 𝑖𝐴(⋅, 𝜔)Φ
= −𝑖Φ∗ 𝐴(⋅, 𝜔)Φ + 𝑖Φ∗ 𝐴(⋅, 𝜔)Φ = 0.
It follows from this that there exists a constant matrix 𝐶 ∈ 𝑀𝑛 (ℂ) such that
Φ∗ 𝑖J Φ = 𝐶.
But since Φ(0) = Ψ(0, 𝜔) = 𝐼𝑛 this implies 𝐶 = 𝑖J and hence
Φ∗ 𝑖J Φ = 𝑖J .
This completes the proof.
Proof.
[Lemma 41] Let 𝜔 0 ∈ Ωℝ . By Lemma 76 and the hypotheses the Hamilto-
nian satisfies 𝐴(⋅, 𝜔)∗ = 𝐴(⋅, 𝜔) as an element of 𝑀𝑛 (𝐿1 (𝕋))) for every 𝜔 ∈ Ωℝ and 𝐴 ∈
𝒪(Ω, 𝑀𝑛 (𝐿1 (𝕋))) as a function of frequency. By definition since 𝐴𝜔 is derivative of the 𝐴
with respect to frequency in the Banach space 𝑀𝑛 (𝐿1 (𝕋)) we have 𝐴𝜔 (⋅, 𝜔) ∈ 𝑀𝑛 (𝐿1 (𝕋)) for
every 𝜔 ∈ Ω and
lim ∥(𝜔 − 𝜔 0 )−1 (𝐴(⋅, 𝜔) − 𝐴(⋅, 𝜔 0 )) − 𝐴𝜔 (⋅, 𝜔 0 )∥𝐿1 (𝕋) = 0.
𝜔→𝜔 0
It then follows from these facts and Lemma 76 that
0 = lim ∥(𝜔 − 𝜔 0 )−1 (𝐴(⋅, 𝜔) − 𝐴(⋅, 𝜔 0 )) − 𝐴𝜔 (⋅, 𝜔 0 )∥𝐿1 (𝕋)
𝜔→𝜔 0
𝜔∈Ωℝ
(
)∗
= lim ∥ (𝜔 − 𝜔 0 )−1 (𝐴(⋅, 𝜔) − 𝐴(⋅, 𝜔 0 )) − 𝐴𝜔 (⋅, 𝜔 0 ) ∥𝐿1 (𝕋)
𝜔→𝜔 0
𝜔∈Ωℝ
136
= lim ∥(𝜔 − 𝜔 0 )−1 (𝐴(⋅, 𝜔) − 𝐴(⋅, 𝜔 0 )) − 𝐴𝜔 (⋅, 𝜔 0 )∗ ∥𝐿1 (𝕋) .
𝜔→𝜔 0
𝜔∈Ωℝ
From which it follows that 𝐴𝜔 (⋅, 𝜔 0 )∗ = 𝐴𝜔 (⋅, 𝜔 0 ) in 𝑀𝑛 (𝐿1 (𝕋))). This completes the proof.
Proof. [Theorem 42] Let 𝜔 0 ∈ Ωℝ . Then Proposition 33 implies Ψ(⋅, 𝜔 0 ), Ψ𝜔 (⋅, 𝜔 0 ) ∈
1,1
1,1
(ℝ)). Thus by Lemma 73 we
(ℝ)) and by Lemma 75 have Ψ(⋅, 𝜔 0 )∗ ∈ 𝑀𝑛 (𝑊loc
𝑀𝑛 (𝑊loc
1,1
have Ψ(⋅, 𝜔 0 )∗ J Ψ𝜔 (⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝑊loc
(ℝ)).
Thus by notation 4.5.(xxix) we must have
(Ψ(⋅, 𝜔 0 )∗ J Ψ𝜔 (⋅, 𝜔 0 ))′ ∈ 𝑀𝑛 (𝐿1loc (ℝ)). Now since 𝐴𝜔 (⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝐿1 (𝕋))) ⊆ 𝑀𝑛 (𝐿1loc (ℝ))
then it follows from Lemma 84 that we also have Ψ(⋅, 𝜔 0 )∗ 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 ) ∈ 𝑀𝑛 (𝐿1loc (ℝ)).
Our next goal is to show that as elements of 𝑀𝑛 (𝐿1loc (ℝ)),
Ψ(⋅, 𝜔 0 )∗ 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 ) = (Ψ(⋅, 𝜔 0 )∗ J Ψ𝜔 (⋅, 𝜔 0 ))′ .
We begin by defining two operators that will play a key role in our proof. The first is the
integral map defined by
∫
𝑥
(ℐ𝐵)(𝑥) :=
𝐵(𝑡)𝑑𝑡,
𝐵 ∈ (𝐿1loc (ℝ))𝑛 ,
𝑥 ∈ ℝ.
0
1,1
By Lemma 87, the map ℐ : (𝐿1loc (ℝ))𝑛 → (𝑊loc
(ℝ))𝑛 is well-defined and
(ℐ𝑓 )′ = 𝑓,
1,1
for every 𝑓 ∈ (𝑊loc
(ℝ))𝑛 . The second map is defined for each 𝜔 ∈ Ω by
∫
(𝑆(𝜔)𝑓 )(𝑥) := Ψ(𝑥, 𝜔)
𝑥
Ψ(𝑡, 𝜔)−1 𝑓 (𝑡)𝑑𝑡,
0
137
𝑓 ∈ (𝐿1loc (ℝ))𝑛 ,
𝑥 ∈ ℝ.
1,1
(ℝ))𝑛 is a right inverse of
Lemma 58 For each 𝜔 ∈ Ω, the map 𝑆(𝜔) : (𝐿1loc (ℝ))𝑛 → (𝑊loc
1,1
the map 𝑇 (𝜔) : (𝑊loc
(ℝ))𝑛 → (𝐿1loc (ℝ))𝑛 defined in (4.47), i.e., for every 𝑓 ∈ (𝐿1loc (ℝ))𝑛 we
have
𝑇 (𝜔)𝑆(𝜔)𝑓 = 𝑓.
Proof.
1,1
(ℝ))𝑛 is wellLet 𝜔 ∈ Ω. Lets first prove the map 𝑆(𝜔) : (𝐿1loc (ℝ))𝑛 → (𝑊loc
defined. By Proposition 33.(v) it follows that Ψ−1 (𝑥, 𝜔) := Ψ(𝑥, 𝜔)−1 exists for every
1,1
𝑥 ∈ ℝ and Ψ−1 (⋅, 𝜔) ∈ 𝑀𝑛 (𝑊loc
(ℝ)). Let 𝑓 ∈ (𝐿1loc (ℝ))𝑛 . Then it follows by Lemma
84 that Ψ−1 (⋅, 𝜔)𝑓 ∈ (𝐿1loc (ℝ))𝑛 and applying the integral map to this yields ℐ(Ψ−1 (⋅, 𝜔)𝑓 ) ∈
1,1
(𝑊loc
(ℝ))𝑛 . By Lemma 73 and the definition of 𝑆(𝜔) we find that
1,1
𝑆(𝜔)𝑓 = Ψ(⋅, 𝜔)ℐ(Ψ−1 (⋅, 𝜔)𝑓 ) ∈ (𝑊loc
(ℝ))𝑛 .
Thus the map is well-defined.
1,1
Next, recall that the map 𝑇 (𝜔) : (𝑊loc
(ℝ))𝑛 → (𝐿1loc (ℝ))𝑛 in (4.47) is
𝑇 (𝜔)𝜓 := 𝜓 ′ − J −1 𝐴(⋅, 𝜔)𝜓,
1,1
𝜓 ∈ (𝑊loc
(ℝ))𝑛 .
We now show 𝑆(𝜔) is a right inverse of 𝑇 (𝜔). Let 𝑓 ∈ (𝐿1loc (ℝ))𝑛 then by Lemma 73 we have
(𝑆(𝜔)𝑓 )′ = (Ψ(⋅, 𝜔)ℐ(Ψ−1 (⋅, 𝜔)𝑓 ))′ = Ψ(⋅, 𝜔)′ ℐ(Ψ−1 (⋅, 𝜔)𝑓 ) + Ψ(⋅, 𝜔)(ℐ(Ψ−1 (⋅, 𝜔)𝑓 ))′
= J −1 𝐴(⋅, 𝜔)Ψ(⋅, 𝜔)ℐ(Ψ−1 (⋅, 𝜔)𝑓 ) + Ψ(⋅, 𝜔)Ψ−1 (⋅, 𝜔)𝑓 = J −1 𝐴(⋅, 𝜔)𝑆(𝜔)𝑓 + 𝑓.
Hence it follows that
𝑇 (𝜔)𝑆(𝜔)𝑓 = (𝑆(𝜔)𝑓 )′ − J −1 𝐴(⋅, 𝜔)𝑆(𝜔)𝑓 = 𝑓.
138
Thus 𝑆(𝜔) is a right inverse of 𝑇 (𝜔). This completes the proof.
Let 𝛾 ∈ ℂ𝑛 . Define 𝜙 := Ψ𝜔 (⋅, 𝜔 0 )𝛾 and 𝑓 := J −1 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 )𝛾. We will now show
that 𝜙 = 𝑆(𝜔 0 )𝑓 . It follows from Lemma 84 that 𝑓 ∈ (𝐿1loc (ℝ))𝑛 . It follows by Proposition
1,1
33.(iv) that 𝜙 ∈ (𝑊loc
(ℝ))𝑛 and satisfies
(
)
𝜙 = Ψ𝜔 (⋅, 𝜔 0 )𝛾 = ℐ J −1 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 )𝛾 + J −1 𝐴(⋅, 𝜔 0 )Ψ𝜔 (⋅, 𝜔 0 )𝛾
(
)
= ℐ 𝑓 + J −1 𝐴(⋅, 𝜔 0 )𝜙
since J −1 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 )𝛾 ∈ 𝑀𝑛 (𝐿1loc (ℝ)) by Lemma 84. Hence by Lemma 87 it follows
that
𝑇 (𝜔 0 )𝜙 = 𝜙′ − J −1 𝐴(⋅, 𝜔 0 )𝜙 = 𝑓.
1,1
But by Lemma 58 we also have 𝑇 (𝜔 0 )𝑆(𝜔 0 )𝑓 = 𝑓 with 𝑆(𝜔 0 )𝑓 ∈ (𝑊loc
(ℝ))𝑛 . Hence
1,1
𝜓 := 𝜙 − 𝑆(𝜔 0 )𝑓 ∈ (𝑊loc
(ℝ))𝑛 and satisfies
𝑇 (𝜔 0 )𝜓 = 0,
that is, 𝜓 ∈ ker(𝑇 (𝜔 0 )). Thus by Corollary 54 and Proposition 33.(i) there exists a unique
𝛾˜ ∈ ℂ𝑛 such that 𝜓 = Ψ(⋅, 𝜔 0 )˜
𝛾 . In fact, 𝛾˜ = 𝜓(0) = 𝜙(0)−(𝑆(𝜔 0 )𝑓 )(0). But (𝑆(𝜔)𝑓 )(0) = 0
and 𝜙(0) = Ψ𝜔 (0, 𝜔 0 )𝛾 = 0. Thus we have 𝜓 = 0 and so 𝜙 = 𝑆(𝜔 0 )𝑓 as desired.
Now it follows from what we have just shown that for every 𝛾 ∈ ℂ𝑛 ,
Ψ𝜔 (⋅, 𝜔 0 )𝛾 = 𝑆(𝜔 0 )(J −1 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 )𝛾)
139
from which it follows by the definition of 𝑆(𝜔 0 ) and Lemma 87 that
∫
𝑥
Ψ(𝑡, 𝜔 0 )−1 J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝛾𝑑𝑡
Ψ𝜔 (𝑥, 𝜔 0 )𝛾 = Ψ(𝑥, 𝜔 0 )
∫0 𝑥
= Ψ(𝑥, 𝜔 0 )
Ψ(𝑡, 𝜔 0 )−1 J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡𝛾,
∀𝑥 ∈ ℝ.
0
But since this is true for every 𝛾 ∈ ℂ𝑛 this implies
∫
Ψ𝜔 (𝑥, 𝜔 0 ) = Ψ(𝑥, 𝜔 0 )
𝑥
Ψ(𝑡, 𝜔 0 )−1 J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡, ∀𝑥 ∈ ℝ.
0
Now if we multiply both sides by Ψ(𝑥, 𝜔)∗ J and apply Proposition 39 we get
∫
𝑥
Ψ(𝑡, 𝜔 0 )−1 J −1 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡
Ψ(𝑥, 𝜔 0 ) J Ψ𝜔 (𝑥, 𝜔 0 ) = Ψ(𝑥, 𝜔) J Ψ(𝑥, 𝜔 0 )
0
∫ 𝑥
∫ 𝑥
−1
−1
=J
Ψ(𝑡, 𝜔 0 ) J 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡 =
Ψ(𝑡, 𝜔 0 )∗ 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡
∗
∗
0
0
for every 𝑥 ∈ ℝ. It follows from this and Lemma 87 that
(Ψ(⋅, 𝜔 0 )∗ J Ψ𝜔 (⋅, 𝜔 0 ))′ = Ψ(⋅, 𝜔 0 )∗ 𝐴𝜔 (⋅, 𝜔 0 )Ψ(⋅, 𝜔 0 )
and
1
𝑑
∫
0
𝑑
1
Ψ(𝑡, 𝜔 0 )∗ 𝐴𝜔 (𝑡, 𝜔 0 )Ψ(𝑡, 𝜔 0 )𝑑𝑡 = Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔 0 ).
𝑑
This completes the proof.
Proof. [Corollary 43] Let 𝜔 ∈ Ωℝ and 𝛾 1 , 𝛾 2 ∈ ℂ𝑛 . Then by Theorem 42 we have
1,1
Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 ∈ (𝑊loc
(ℝ))𝑛
Ψ(⋅, 𝜔)∗ 𝐴𝜔 (⋅, 𝜔)Ψ(⋅, 𝜔)𝛾 1 ∈ (𝐿1loc (ℝ))𝑛
(Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 )′ = Ψ(⋅, 𝜔)∗ 𝐴𝜔 (⋅, 𝜔)Ψ(⋅, 𝜔)𝛾 1
140
1,1
(ℝ),
Hence it follows from this and Lemma 75 that ⟨Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ ∈ 𝑊loc
⟨𝐴𝜔 (⋅, 𝜔)𝜓 1 , 𝜓 2 ⟩ℂ ∈ 𝐿1loc (ℝ), and ⟨Ψ(⋅, 𝜔)∗ J Ψ𝜔 (⋅, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ ′ = ⟨𝐴𝜔 (⋅, 𝜔)𝜓 1 , 𝜓 2 ⟩ℂ . And by
Theorem 42 we conclude that
1
𝑑
∫
0
𝑑
1
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥 = ⟨Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)𝛾 1 , 𝛾 2 ⟩ℂ .
𝑑
This completes the proof.
Proofs for Section 4.2.3
We begin by proving the sesquilinear form defined in (4.31) is a Hermitian form.
Proof. [Lemma 44] Let (𝑘, 𝜔) ∈ ℬℝ . By Lemma 41, for any 𝛾 1 , 𝛾 2 ∈ ker (𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)),
if we let 𝜓 1 := Ψ(⋅, 𝜔)𝛾 1 , 𝜓 2 := Ψ(⋅, 𝜔)𝛾 2 then we have
∫
∫
1 𝑑
1 𝑑
𝑞(𝑘,𝜔) (𝛾 1 , 𝛾 2 ) =
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥 =
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥), 𝜓 2 (𝑥)⟩ℂ 𝑑𝑥
𝑑 0
𝑑 0
∫
∫
1 𝑑
1 𝑑
=
⟨𝜓 (𝑥), 𝐴𝜔 (𝑥, 𝜔)𝜓 1 (𝑥)⟩ℂ 𝑑𝑥 =
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 2 (𝑥), 𝜓 1 (𝑥)⟩ℂ 𝑑𝑥 = 𝑞(𝑘,𝜔) (𝛾 2 , 𝛾 1 ),
𝑑 0 2
𝑑 0
implying 𝑞(𝑘,𝜔) (𝛾 1 , 𝛾 1 ) ∈ ℝ. This completes the proof.
Proof. [Proposition 45] By Definition 23, the canonical ODEs in (4.13) are of definite
type at a point (𝑘0 , 𝜔 0 ) ∈ ℬℝ if and only if
1
𝑑
∫
𝑑
⟨𝐴𝜔 (𝑥, 𝜔 0 )𝜓(𝑥), 𝜓(𝑥)⟩ℂ 𝑑𝑥 ∕= 0
0
for any nontrivial Bloch solution 𝜓 of the canonical ODEs in (4.13) with wavenumberfrequency (𝑘0 , 𝜔 0 ). But by Theorem 35, 𝜓 is a nontrivial Bloch solution 𝜓 of the canonical
ODEs in (4.13) with wavenumber-frequency (𝑘0 , 𝜔 0 ) if and only if 𝜓 = Ψ(⋅, 𝜔 0 )𝛾 where 𝛾
141
is an eigenvector of the monodromy matrix Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 .
And hence the canonical ODEs in (4.13) are of definite type at a point (𝑘0 , 𝜔 0 ) ∈ ℬℝ if and
only if ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) ∕= {0} and
𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾) ∕= 0, for every nonzero 𝛾 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
(4.64)
Thus to complete the proof we need only show that (4.64) implies 𝑞(𝑘0 ,𝜔0 ) , as defined in (4.31),
is a definite sesquilinear form which is bounded.
Suppose that (4.64) is true. Then it follows from Lemma 44 that 𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾) ∈ ℝ/{0}
for every nonzero 𝛾 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )). Suppose there existed nonzero 𝛾 − , 𝛾 + ∈
ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) such that 𝑞(𝑘0 ,𝜔0 ) (𝛾 − 𝛾 − ) < 0 < 𝑞(𝑘0 ,𝜔0 ) (𝛾 + , 𝛾 + ). Then we can define a
continuous real-valued function 𝑓 : [0, 1] → ℝ by
𝑓 (𝑥) := 𝑞(𝑘0 ,𝜔0 ) (𝑥𝛾 + + (1 − 𝑥)𝛾 − , 𝑥𝛾 + + (1 − 𝑥)𝛾 − ).
Now 𝑓 (0) < 0 < 𝑓 (1) and so by the intermediate value theorem that there exists 𝑥0 ∈
(0, 1) such that 𝑓 (𝑥0 ) = 0 implying by hypothesis that 𝑥0 𝛾 + + (1 − 𝑥0 )𝛾 − = 0. But this
implies 0 < 𝑥20 𝑞(𝑘0 ,𝜔0 ) (𝛾 + , 𝛾 + ) = (1 − 𝑥0 )2 𝑞(𝑘0 ,𝜔0 ) (𝛾 − , 𝛾 − ) < 0, a contradiction. Thus we
conclude that sgn(𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾)) =: sgn(𝑞(𝑘0 ,𝜔0 ) ) is independent of the choice of nonzero 𝛾 in
ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
We define a sesquilinear form by
⟨𝛾 1 , 𝛾 2 ⟩(𝑘0 ,𝜔0 ) := sgn(𝑞(𝑘0 ,𝜔0 ) )𝑞(𝑘0 ,𝜔0 ) (𝛾 1 , 𝛾 2 ),
𝛾 1 , 𝛾 2 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
It follows from what we have just shown that ⟨ , ⟩(𝑘0 ,𝜔0 ) is an inner product on the finite-
142
dimensional vector space ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) and so
1
1
2
∣∣𝛾∣∣(𝑘0 ,𝜔0 ) := ⟨𝛾, 𝛾⟩(𝑘
= ∣𝑞(𝑘0 ,𝜔0 ) (𝛾, 𝛾)∣ 2 ,
0 ,𝜔 0 )
𝛾 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )),
defines a norm on this space as does the Euclidean norm ∣∣ ⋅ ∣∣ℂ . But since any two norms
on a finite dimensional vector space are equivalent this implies there exists 𝐶1 , 𝐶2 > 0 such
that
𝐶1 ∣∣𝛾∣∣2ℂ ≤ ∣∣𝛾∣∣2(𝑘0 ,𝜔0 ) ≤ 𝐶2 ∣∣𝛾∣∣2ℂ
for every 𝛾 ∈ ker (𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )). This completes the proof.
Proof. [Lemma 46] Let {(𝑘𝑗 , 𝜔 𝑗 )}𝑗∈ℕ be a sequence in ℬℝ and {𝛾 𝑗 }𝑗∈ℕ a sequence in ℂ𝑛
such that 𝛾 𝑗 ∈ ker(𝑒𝑖𝑘𝑗 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 𝑗 )), ∣∣𝛾 𝑗 ∣∣ℂ = 1 for all 𝑗 ∈ ℕ and
∣∣⋅∣∣ℂ
∣∣⋅∣∣ℂ
(𝑘𝑗 , 𝜔 𝑗 ) −−→ (𝑘0 , 𝜔 0 ) and 𝛾 𝑗 −−→ 𝛾 0 , as 𝑗 → ∞,
for some 𝑘0 , 𝜔 0 ∈ ℂ and 𝛾 0 ∈ ℂ𝑛 . We will now show that ∣∣𝛾 0 ∣∣ℂ = 1, 𝛾 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 −
Ψ(𝑑, 𝜔 0 )), and (𝑘0 , 𝜔 0 ) ∈ ℬℝ .
∣∣⋅∣∣ℂ
First, since 𝛾 𝑗 −−→ 𝛾 0 as 𝑗 → ∞, then
1 = lim ∣∣𝛾 𝑗 ∣∣ℂ = ∣∣𝛾 0 ∣∣ℂ .
𝑗→∞
Second, if we define a matrix-valued function by 𝑀 (𝑘, 𝜔) := 𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔), (𝑘, 𝜔) ∈ ℂ × Ω
then this function is continuous since
∣∣𝑀 (𝑘, 𝜔) − 𝑀 (𝑘0 , 𝜔 0 )∣∣ℂ ≤ ∣∣𝑀 (𝑘, 𝜔) − 𝑀 (𝑘0 , 𝜔)∣∣ℂ + ∣∣𝑀 (𝑘0 , 𝜔) − 𝑀 (𝑘0 , 𝜔 0 )∣∣ℂ
1
= ∣𝑒𝑖𝑘𝑑 − 𝑒𝑖𝑘0 𝑑 ∣𝑛 2 + ∣∣Ψ(𝑑, 𝜔) − Ψ(𝑑, 𝜔 0 )∣∣ℂ
143
and by Proposition 33.(iii) we have Ψ(𝑑, ⋅) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ) implying ∣∣Ψ(𝑑, 𝜔) − Ψ(𝑑, 𝜔 0 )∣∣ℂ →
0 as 𝜔 → 𝜔 0 . Thus we conclude from this that
∣∣𝑀 (𝑘0 , 𝜔 0 )𝛾 0 ∣∣ℂ = lim ∣∣𝑀 (𝑘𝑗 , 𝜔 𝑗 )𝛾 𝑗 − 𝑀 (𝑘0 , 𝜔 0 )𝛾 0 ∣∣ℂ
𝑗→∞
≤ lim ∣∣𝑀 (𝑘𝑗 , 𝜔 𝑗 )𝛾 𝑗 − 𝑀 (𝑘0 , 𝜔 0 )𝛾 𝑗 ∣∣ℂ + lim ∣∣𝑀 (𝑘0 , 𝜔 0 )𝛾 𝑗 − 𝑀 (𝑘0 , 𝜔 0 )𝛾 0 ∣∣ℂ
𝑗→∞
𝑗→∞
≤ lim ∣∣𝑀 (𝑘𝑗 , 𝜔 𝑗 ) − 𝑀 (𝑘0 , 𝜔 0 )∣∣ℂ ∣∣𝛾 𝑗 ∣∣ℂ + lim ∣∣𝑀 (𝑘0 , 𝜔 0 )∣∣ℂ ∣∣𝛾 𝑗 − 𝛾 0 ∣∣ℂ
𝑗→∞
𝑗→∞
=0
and hence 𝛾 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )). It now follows from this and Corollary 38 that
(𝑘0 , 𝜔 0 ) ∈ ℬℝ .
Now we complete the proof by showing
∣⋅∣
𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛾 𝑗 , 𝛾 𝑗 ) −
→ 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 ), as 𝑗 → ∞.
(4.65)
First, it follows from Corollary 43 that for any (𝑘, 𝜔) ∈ ℬℝ and for any 𝛽 1 , 𝛽 2 ∈ ker(𝑒𝑖𝑘𝑑 𝐼𝑛 −
Ψ(𝑑, 𝜔)) we have
1
𝑞(𝑘,𝜔) (𝛽 1 , 𝛽 2 ) = ⟨Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)𝛽 1 , 𝛽 2 ⟩ℂ .
𝑑
Thus we extend the definition of 𝑞(𝑘,𝜔) to all of ℂ𝑛 × ℂ𝑛 in the obvious way by
1
𝑞(𝑘,𝜔) (𝛽 1 , 𝛽 2 ) := ⟨Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)𝛽 1 , 𝛽 2 ⟩ℂ ,
𝑑
𝛽 1 , 𝛽 2 ∈ ℂ𝑛 .
It follows from this definition that 𝑞(𝑘,𝜔) : ℂ𝑛 × ℂ𝑛 → ℂ is a continuous sesquilinear form
since
∣𝑞(𝑘,𝜔) (𝛽 1 , 𝛽 2 )∣ ≤ 𝑑−1 ∣∣Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)∣∣ℂ ∣∣𝛽 1 ∣∣ℂ ∣∣𝛽 2 ∣∣ℂ , ∀𝛽 1 , 𝛽 2 ∈ ℂ𝑛 .
144
Hence for any (𝑘, 𝜔) ∈ ℬℝ and any 𝛽 ∈ ℂ𝑛 we have the inequality
∣𝑞(𝑘,𝜔) (𝛽, 𝛽) − 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 )∣ ≤ ∣𝑞(𝑘,𝜔) (𝛽, 𝛽) − 𝑞(𝑘,𝜔) (𝛽, 𝛾 0 )∣
+ ∣𝑞(𝑘,𝜔) (𝛽, 𝛾 0 ) − 𝑞(𝑘0 ,𝜔0 ) (𝛽, 𝛾 0 )∣
+ ∣𝑞(𝑘0 ,𝜔0 ) (𝛽, 𝛾 0 ) − 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 )∣
and the inequalities
∣𝑞(𝑘,𝜔) (𝛽, 𝛽) − 𝑞(𝑘,𝜔) (𝛽, 𝛾 0 )∣ = ∣𝑞(𝑘,𝜔) (𝛽, 𝛽 − 𝛾 0 )∣
≤ 𝑑−1 ∣∣Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)∣∣ℂ ∣∣𝛽∣∣ℂ ∣∣𝛽 − 𝛾 0 ∣∣ℂ ,
∣𝑞(𝑘0 ,𝜔0 ) (𝛽, 𝛾 0 ) − 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 )∣ = ∣𝑞(𝑘0 ,𝜔0 ) (𝛽 − 𝛾 0 , 𝛾 0 )∣
≤ 𝑑−1 ∣∣Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)∣∣ℂ ∣∣𝛽 − 𝛾 0 ∣∣ℂ ∣∣𝛾 0 ∣∣ℂ ,
∣𝑞(𝑘,𝜔) (𝛽, 𝛾 0 ) − 𝑞(𝑘0 ,𝜔0 ) (𝛽, 𝛾 0 )∣ = 𝑑−1 ∣∣Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔)
− Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 )∣∣ℂ ∣∣𝛽∣∣ℂ ∣∣𝛾 0 ∣∣ℂ
And thus it follows from these inequalities that (4.65) is true if
∣∣⋅∣∣ℂ
∣⋅∣
Ψ(𝑑, 𝜔)∗ J Ψ𝜔 (𝑑, 𝜔) −−→ Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 ), as 𝜔 −
→ 𝜔0.
But by Lemma 77, multiplication of matrices in 𝑀𝑛 (ℂ) is a continuous bilinear map and so
in order to prove the latter statement we need only show
∣∣⋅∣∣ℂ
∣∣⋅∣∣ℂ
∣⋅∣
Ψ(𝑑, 𝜔)∗ −−→ Ψ(𝑑, 𝜔 0 )∗ and Ψ𝜔 (𝑑, 𝜔) −−→ Ψ𝜔 (𝑑, 𝜔 0 ), as 𝜔 −
→ 𝜔0.
145
Convergence of the first limit follows from the fact that the adjoint map ()∗ : 𝑀𝑛 (ℂ) →
𝑀𝑛 (ℂ) is a norm preserving map, i.e., ∣∣𝐵 ∗ ∣∣ℂ = ∣∣𝐵∣∣ℂ for every 𝐵 ∈ 𝑀𝑛 (ℂ), since ∣∣Ψ(𝑑, 𝜔)∗ −
Ψ(𝑑, 𝜔 0 )∗ ∣∣ℂ = ∣∣(Ψ(𝑑, 𝜔)−Ψ(𝑑, 𝜔 0 ))∗ ∣∣ℂ = ∣∣Ψ(𝑑, 𝜔)−Ψ(𝑑, 𝜔 0 )∣∣ℂ → 0 as 𝜔 → 𝜔 0 . Now for the
other limit, by Lemma 81 since Ψ(𝑑, ⋅) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)) and, by Proposition 33.(iii), Ψ𝜔 (𝑑, ⋅)
is its derivative in the ∣∣ ⋅ ∣∣ℂ norm then Ψ𝜔 (𝑑, ⋅) ∈ 𝒪(Ω, 𝑀𝑛 (ℂ)). But this implies Ψ𝜔 (𝑑, ⋅) is
∣∣⋅∣∣ℂ
a continuous matrix-valued function on Ω in the ∣∣ ⋅ ∣∣ℂ norm and so Ψ𝜔 (𝑑, 𝜔) −−→ Ψ𝜔 (𝑑, 𝜔 0 )
∣⋅∣
as 𝜔 −
→ 𝜔 0 . Therefore (4.65) is true and this completes the proof.
Proof. [Theorem 47] Suppose (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the canonical
ODEs in (4.13). We will first show that there exists an 𝑟 > 0 such that every (𝑘, 𝜔) ∈
𝐵((𝑘0 , 𝜔 0 ), 𝑟) ∩ ℬℝ is a point of definite type for the canonical ODEs in (4.13). We will prove
this by contradiction. Suppose the statement was not true. Then we can find a sequence of
1,1
∞
𝑛
points {(𝑘𝑗 , 𝜔 𝑗 )}∞
𝑗=1 ⊆ ℬℝ and a sequence {𝜓 𝑗 }𝑗=1 ⊆ (𝑊loc (ℝ)) such that 𝜓 𝑗 is a nontrivial
Bloch solution of the canonical ODEs in (4.13) with wavenumber-frequency pair (𝑘𝑗 , 𝜔 𝑗 ) and
1
𝑑
∫
𝑑
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 𝑗 (𝑥), 𝜓 𝑗 (𝑥)⟩ℂ 𝑑𝑥 = 0,
0
for every 𝑗 ∈ ℕ. By Theorem 35 there exists a nonzero 𝛾 𝑗 ∈ ker(𝑒𝑖𝑘𝑗 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 𝑗 )) such that
𝜓 𝑗 = Ψ(𝑑, 𝜔 𝑗 )𝛾 𝑗 , for every 𝑗 ∈ ℕ. Let 𝛽 𝑗 :=
𝛾𝑗
,
∣∣𝛾 𝑗 ∣∣ℂ
for 𝑗 ∈ ℕ. Then {𝛽 𝑗 }∞
𝑗=1 is a sequence
of vectors with unit norm in ℂ𝑛 and hence since {𝛽 ∈ ℂ𝑛 : ∣∣𝛽∣∣ℂ = 1} is a compact set
∣∣⋅∣∣ℂ
this implies there exists a convergent subsequence {𝛽 𝑗𝑙 }∞
𝑙=1 such that 𝛽 𝑗𝑙 −−→ 𝛽 0 as 𝑙 → ∞
for some 𝛽 0 ∈ ℂ𝑛 with ∣∣𝛽 0 ∣∣ℂ = 1. Thus since {(𝑘𝑗𝑙 , 𝜔 𝑗𝑙 )}𝑙∈ℕ ⊆ ℬℝ , {𝛽 𝑗𝑙 }𝑗∈ℕ ⊆ ℂ𝑛 such that
𝛽 𝑗𝑙 ∈ ker(𝑒𝑖𝑘𝑗𝑙 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 𝑗𝑙 )) with ∣∣𝛽 𝑗𝑙 ∣∣ℂ = 1 for all 𝑙 ∈ ℕ and
∣∣⋅∣∣ℂ
∣∣⋅∣∣ℂ
(𝑘𝑗𝑙 , 𝜔 𝑗𝑙 ) −−→ (𝑘0 , 𝜔 0 ) and 𝛽 𝑗𝑙 −−→ 𝛽 0 , as 𝑙 → ∞
146
then by Lemma 46, 𝛽 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) with ∣∣𝛽 0 ∣∣ℂ = 1 and
∣⋅∣
→ 𝑞(𝑘0 ,𝜔0 ) (𝛽 0 , 𝛽 0 ) as 𝑙 → ∞.
𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛽 𝑗𝑙 , 𝛽 𝑗𝑙 ) −
𝑙
𝑙
But for every 𝑙 ∈ ℕ we have
𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛽 𝑗𝑙 , 𝛽 𝑗𝑙 ) = ∣∣𝛾 𝑗𝑙 ∣∣−2
ℂ 𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛾 𝑗𝑙 , 𝛾 𝑗𝑙 )
𝑙
𝑙
𝑙
𝑙
∫ 𝑑
1
= ∣∣𝛾 𝑗𝑙 ∣∣−2
⟨𝐴𝜔 (𝑥, 𝜔)𝜓 𝑗𝑙 (𝑥), 𝜓 𝑗𝑙 (𝑥)⟩ℂ 𝑑𝑥 = 0
ℂ
𝑑 0
implying
𝑞(𝑘0 ,𝜔0 ) (𝛽 0 , 𝛽 0 ) = 0, ∣∣𝛽 0 ∣∣ℂ = 1,
𝛽 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )).
By Proposition 45 this contradicts that (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the
canonical ODEs in (4.13). Thus we conclude that there exists an 𝑟 > 0 such that every
(𝑘, 𝜔) ∈ 𝐵((𝑘0 , 𝜔 0 ), 𝑟) ∩ ℬℝ is a point of definite type for the canonical ODEs in (4.13).
Now let {(𝑘𝑗 , 𝜔 𝑗 )}∞
𝑗=1 ⊆ ℬℝ be any sequence converging to (𝑘0 , 𝜔 0 ) in the Euclidean norm
∣∣ ⋅ ∣∣ℂ . Then since the set {𝛽 ∈ ℂ𝑛 : ∣∣𝛽∣∣ℂ = 1} is compact we can find a sequence {𝛾 𝑗 }∞
𝑗=1
such that 𝛾 𝑗 ∈ ker(𝑒𝑖𝑘𝑗 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 𝑗 )) with ∣∣𝛾 𝑗 ∣∣ℂ = 1 for all 𝑗 ∈ ℕ and
∣∣⋅∣∣ℂ
𝛾 𝑗 −−→ 𝛾 0 , as 𝑗 → ∞
for some 𝛾 0 ∈ {𝛽 ∈ ℂ𝑛 : ∣∣𝛽∣∣ℂ = 1}. From what we just proved there exists a 𝐽 ∈ ℕ such
that if 𝑗 ≥ 𝐽 then (𝑘𝑗 , 𝜔 𝑗 ) is a point of definite type for the canonical ODEs in (4.13). Thus
by Proposition 45 and Lemma 46 it follows that 𝛾 0 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) with ∣∣𝛾 0 ∣∣ℂ = 1
and
sgn(𝑞(𝑘𝑗 ,𝜔𝑗 ) ) =
𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛾 𝑗 , 𝛾 𝑗 )
∣𝑞(𝑘𝑗 ,𝜔𝑗 ) (𝛾 𝑗 , 𝛾 𝑗 )∣
∣⋅∣
−
→
𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 )
∣𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 )∣
147
= sgn(𝑞(𝑘0 ,𝜔0 ) ), as 𝑗 → ∞.
From which we conclude that for 𝑗 sufficiently large we must have sgn(𝑞(𝑘𝑗 ,𝜔𝑗 ) ) = sgn(𝑞(𝑘0 ,𝜔0 ) ).
Therefore, since this is true for an arbitrary sequence in ℬℝ , we conclude that there exists an
𝑟 > 0 such that every (𝑘, 𝜔) ∈ 𝐵((𝑘0 , 𝜔 0 ), 𝑟) ∩ ℬℝ is a point of definite type for the canonical
ODEs in (4.13) and sgn(𝑞(𝑘,𝜔) ) = sgn(𝑞(𝑘0 ,𝜔0 ) ). This completes the proof.
Proofs for Section 4.2.4
We now begin proving our main results on the perturbation theory for canonical ODEs. Our
proof relies heavily on the spectral perturbation theory for holomorphic matrix functions
and the reader may wish to review chapter 3 before preceding.
Proof. [Theorem 48] Let (𝑘0 , 𝜔 0 ) ∈ ℬℝ be a point of definite type for the canonical ODEs
in (4.13). Let 𝑔 be the number of Jordan blocks (geometric multiplicity) in the Jordan
form of the monodromy matrix Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 and
𝑚1 ≥ ⋅ ⋅ ⋅ ≥ 𝑚𝑔 ≥ 1 the dimensions of each of those Jordan blocks (partial multiplicities).
Define 𝑚 := 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 (algebraic multiplicity). Let 𝜖 > 0 be given.
Choose an 𝑟0 > 0 such that 0 < 𝑟0 ≤ 𝜖 and 𝐵(𝜔 0 , 𝑟0 ) ∈ Ω. Define the function 𝐿 : 𝑈 →
𝑀𝑛 (ℂ) by
𝐿(𝜔, 𝑘) := Ψ(𝑑, 𝜔) − 𝑒𝑖𝑘𝑑 𝐼𝑛 .
(4.66)
where 𝑈 := 𝐵(𝜔 0 , 𝑟0 ) × 𝐵(𝑘0 , 𝑟0 ). By definition of a holomorphic matrix-valued function of
two variables, see 3.2.(i), and the fact Ψ(𝑑, ⋅) ∈ 𝒪(𝐵(𝜔 0 , 𝑟0 ), 𝑀𝑛 (ℂ)), it follows that
𝐿 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)).
(4.67)
We will now use the results from section 3.2 and give a spectral perturbation analysis of the
148
holomorphic matrix function 𝐿(⋅, 𝑘) near the point (𝜔 0 , 𝑘0 ) ∈ 𝜎(𝐿) = ℬ −1 .
First, since 𝐵(𝜔 0 , 𝑟0 )∗ = 𝐵(𝜔 0 , 𝑟0 ), the adjoint of the holomorphic function Ψ(𝑑, 𝜔), see
Definition 14, is
Ψ∗ (𝑑, 𝜔) = Ψ(𝑑, 𝜔)∗ ,
𝜔 ∈ 𝐵(𝜔 0 , 𝑟0 ),
(4.68)
and by Lemma 24 we have
Ψ∗ (𝑑, ⋅) ∈ 𝒪(𝐵(𝜔 0 , 𝑟0 ), 𝑀𝑛 (ℂ)).
(4.69)
Next, since 𝑈 ∗ = 𝑈 , the adjoint 𝐿∗ of the holomorphic function of two variables 𝐿, see
Definition 15, is
𝐿∗ (𝜔, 𝑘) = 𝐿(𝜔, 𝑘)∗ , (𝜔, 𝑘) ∈ 𝑈,
(4.70)
𝐿∗ ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)).
(4.71)
and by Lemma 26 we have
We now prove series of lemmas which relate the spectrum of 𝐿 to the spectrum of its adjoint
𝐿∗ .
Lemma 59 For every (𝜔, 𝑘) ∈ 𝑈 we have
𝐿∗ (𝜔, 𝑘)J = −𝑒−𝑖𝑘𝑑 Ψ∗ (𝑑, 𝜔)J 𝐿(𝜔, 𝑘).
149
(4.72)
Proof. By Proposition 39 we have
Ψ∗ (𝑑, 𝜔)J Ψ(𝑑, 𝜔) = Ψ(𝑑, 𝜔)∗ J Ψ(𝑑, 𝜔) = J
for every 𝐵(𝜔 0 , 𝑟0 ) ∩ ℝ ⊆ Ωℝ . Thus if (𝜔, 𝑘) ∈ 𝑈 ∩ ℝ2 then
− 𝑒−𝑖𝑘𝑑 Ψ∗ (𝑑, 𝜔)J 𝐿(𝜔, 𝑘) = −𝑒−𝑖𝑘𝑑 Ψ(𝑑, 𝜔)∗ J (Ψ(𝑑, 𝜔) − 𝑒𝑖𝑘𝑑 𝐼𝑛 )
= Ψ(𝑑, 𝜔)∗ J − 𝑒−𝑖𝑘𝑑 J = (Ψ(𝑑, 𝜔)∗ − 𝑒−𝑖𝑘𝑑 )J
= 𝐿∗ (𝜔, 𝑘)J .
But for the matrix-valued function 𝑓 (𝜔, 𝑘) := −𝑒−𝑖𝑘𝑑 Ψ∗ (𝑑, 𝜔)J 𝐿(𝜔, 𝑘)−𝐿∗ (𝜔, 𝑘)J we have
𝑓 ∈ 𝒪(𝑈, 𝑀𝑛 (ℂ)) and its identically equal to zero on 𝑈 ∩ℝ2 where 𝑈 = 𝐵(𝜔 0 , 𝑟0 )×𝐵(𝑘0 , 𝑟0 ).
By Lemma 17, 𝑓 is an analytic function of two variables identically equal to zero on 𝑈 ∩ ℝ2
hence using the power series expansion of 𝑓 at the point (𝜔 0 , 𝑘0 ) and the fact 𝑓 = 0 on
𝑈 ∩ ℝ2 we can conclude 𝑓 = 0 in an open connected set of ℂ2 in 𝑈 which contains (𝜔 0 , 𝑘0 ).
Thus, since 𝑈 is an open connected set on which 𝑓 is analytic, by analytic continuation we
conclude that 𝑓 = 0 on 𝑈 . This completes the proof of the lemma.
The following two lemmas are the key to the proof of this theorem.
Lemma 60 If 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit
point 𝜔 0 then 𝜔(𝑘) and 𝜔 ∗ (𝑘), the adjoint of the Puiseux series 𝜔(𝑘), both are eigenvalue
Puiseux series of 𝐿∗ (⋅, 𝑘) and 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit point 𝜔 0 .
Proof.
By Lemma 59, the fact J and Ψ(𝑑, 𝜔) are invertible matrices for every 𝜔 ∈ Ω,
implies the holomorphic matrix function 𝐿(⋅, 𝑘) and its adjoint 𝐿∗ (⋅, 𝑘) are equivalent, i.e,
𝐿(⋅, 𝑘) ∼ 𝐿∗ (⋅, 𝑘), at 𝜔 in the sense of holomorphic matrix functions from chapter 3 for every
(𝜔, 𝑘) ∈ 𝑈 . In particular, 𝐿(⋅, 𝑘) and its adjoint 𝐿∗ (⋅, 𝑘) have the same eigenvalues for every
150
𝑘 ∈ 𝐵(𝑘0 , 𝑟0 ), i.e., for every 𝑘 ∈ 𝐵(𝑘0 , 𝑟0 )
𝜎(𝐿∗ (⋅, 𝑘)) = 𝜎(𝐿(⋅, 𝑘)).
This implies if 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit
point 𝜔 0 then by Lemma 28 its adjoint 𝜔 ∗ (𝑘) is an eigenvalue Puiseux series of 𝐿∗ (⋅, 𝑘)
expanded about 𝑘0 = 𝑘0 with limit point 𝜔 0 = 𝜔 0 . But since 𝜎(𝐿∗ (⋅, 𝑘)) = 𝜎(𝐿(⋅, 𝑘)) then
all the values of the branches of the Puiseux series 𝜔 ∗ (𝑘) and 𝜔(𝑘) are eigenvalues of both
𝐿∗ (⋅, 𝑘) and 𝐿(⋅, 𝑘) which, by the definition of eigenvalue Puiseux series, means 𝜔 ∗ (𝑘) and
𝜔(𝑘) are also eigenvalue Puiseux series of 𝐿∗ (⋅, 𝑘) expanded about 𝑘0 with limit point 𝜔 0 .
This completes the proof.
Lemma 61 If 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit
point 𝜔 0 and 𝛽 is generating eigenvector of 𝐿(⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and associated with
𝜔(𝑘) then J 𝛽 is a generating eigenvector of 𝐿∗ (⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and associated with
𝜔(𝑘). Moreover,
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛽, J 𝛽⟩ℂ ∕= 0.
Proof.
(4.73)
Let 𝜔(𝑘) be an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit
point 𝜔 0 and 𝛽 is generating eigenvector of 𝐿(⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and associated
with 𝜔(𝑘). By definition of generating eigenvector, there exists a eigenpair Puiseux series
(𝜔(𝑘), 𝛾(𝑘)) of 𝐿(⋅, 𝑘) expanded about 𝑘0 such that 𝜔(𝑘) has limit point 𝜔 0 and 𝛾(𝑘) has
limit point 𝛽. By Lemma 59 and the fact det(J ) ∕= 0 it follows that (𝜔(𝑘), J 𝛾(𝑘)) is
an eigenpair Puiseux series of 𝐿∗ (⋅, 𝑘) expanded about 𝑘0 . Hence J 𝛾(𝑘0 ) = J 𝛽 0 is the
limit point of the eigenvector Puiseux series J 𝛾(𝑘). By definition of generating eigenvector
this means J 𝛽 is a generating eigenvector of 𝐿∗ (⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and associated
with 𝜔(𝑘). Moreover, since 𝛽 is a generating eigenvector of 𝐿(⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) this
151
means 𝛽 ∕= 0 and 𝛽 ∈ ker 𝐿(𝜔 0 , 𝑘0 ) = ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )). This means by Lemma 45 and
Corollary 43 that
1
1
0 ∕= 𝑞( 𝑘0 ,𝜔0 ) (𝛽, 𝛽) = ⟨Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 )𝛽, 𝛽⟩ℂ = ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛽, Ψ(𝑑, 𝜔 0 )𝛽⟩ℂ
𝑑
𝑑
𝑖𝑘0 𝑑
1
𝑒
= ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛽, 𝑒𝑖𝑘0 𝑑 𝛽⟩ℂ = −
⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛽, J 𝛽⟩ℂ .
𝑑
𝑑
Therefore,
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 , J 𝛽 0 ⟩ℂ = ⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛽, J 𝛽⟩ℂ ∕= 0.
This completes the proof of the lemma.
Proposition 62 If 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with
limit point 𝜔 0 then 𝜔(𝑘) is a single-valued real analytic nonconstant function of 𝑘.
Proof. Let 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit point
𝜔 0 . Then by Lemma 60 we know that 𝜔 ∗ (𝑘) is also an eigenvalue Puiseux series of 𝐿(⋅, 𝑘)
expanded about 𝑘0 with limit point 𝜔 0 . Let 𝜔 1 (𝑘) := 𝜔(𝑘) and 𝜔 2 (𝑘) := 𝜔 ∗ (𝑘). By definition
of the adjoint of a convergent Puiseux series we have 𝜔 ∗2 (𝑘) = (𝜔 ∗ )∗ (𝑘) = 𝜔(𝑘) = 𝜔 1 (𝑘).
Suppose that for some branch 𝜔 1,ℎ1 (𝑘) of 𝜔 1 (𝑘) and some branch 𝜔 2,ℎ2 (𝑘) of 𝜔 2 (𝑘) there
existed an 𝑟 > 0 such that 𝜔 1,ℎ1 (𝑘) ∕= 𝜔 2,ℎ2 (𝑘) for ∣𝑘 − 𝑘0 ∣ < 𝑟. Then this implies by
Proposition 29 that for every generating eigenvector 𝛽 of 𝐿(⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and
associated with 𝜔 1 (𝑘) and for any generating eigenvector 𝛾 of 𝐿∗ (⋅, 𝑘) at the point (𝜔 0 , 𝑘0 )(=
(𝜔 0 , 𝑘0 )) and associated with 𝜔 ∗2 (𝑘)(= 𝜔 1 (𝑘)) we have
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛽, 𝛾⟩ℂ = 0.
But by Lemma 61 we know that J 𝛽 is a generating eigenvector of 𝐿∗ (⋅, 𝑘) at the point
152
(𝜔 0 , 𝑘0 ) = (𝜔 0 , 𝑘0 ) and associated with 𝜔 ∗2 (𝑘) = 𝜔 1 (𝑘) and
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛽, J 𝛽⟩ℂ ∕= 0.
Thus we have a contradiction. This contradiction implies that there must exist an 𝑟 > 0
such that if 𝜔 1,ℎ1 (𝑘) is any branch of 𝜔 1 (𝑘) and if 𝜔 2,ℎ2 (𝑘) is any branch of 𝜔 2 (𝑘) then
𝜔 1,ℎ1 (𝑘) = 𝜔 2,ℎ2 (𝑘) for ∣𝑘 − 𝑘0 ∣ < 𝑟. Thus if we fix the branch 𝜔 2,ℎ2 (𝑘) then all branches of
𝜔 1 (𝑘) are equal to 𝜔 2,ℎ2 (𝑘) for ∣𝑘 −𝑘0 ∣ < 𝑟. By Lemma 30 this implies 𝜔 1 (𝑘) is a single-valued
analytic function of 𝑘 for 𝑘 ∈ 𝐵(𝑘0 , 𝑟). For the exact same reasoning this implies 𝜔 2 (𝑘) is
a single-valued analytic function of 𝑘 and we have 𝜔 1 (𝑘) = 𝜔 2 (𝑘) for all 𝑘 ∈ 𝐵(𝑘0 , 𝑟). But
since 𝜔(𝑘) = 𝜔 1 (𝑘) = 𝜔 2 (𝑘) = 𝜔 ∗ (𝑘) where 𝜔 ∗ (𝑘) is the adjoint of the analytic function 𝜔(𝑘),
this implies by definition of adjoint of a Puiseux series, that 𝜔(𝑘) = 𝜔 ∗ (𝑘) = 𝜔(𝑘) for all
𝑘 ∈ 𝐵(𝑘0 , 𝑟). And so 𝜔(𝑘) is a single-valued analytic function in 𝐵(𝑘0 , 𝑟) and is real-valued
for real 𝑘, i.e., is a real analytic function of 𝑘 in 𝐵(𝑘0 , 𝑟).
To complete the proof we need only show that 𝜔(𝑘) is a nonconstant function of 𝑘, i.e.,
𝜔(𝑘) = 𝜔(𝑘0 ) for every 𝑘 in 𝐵(𝑘0 , 𝑟) is a contradiction. By hypothesis we have 𝜔(𝑘0 ) = 𝜔 0 and
since 𝜔(𝑘) is an eigenvalue Puiseux series of 𝐿(⋅, 𝑘) this means 0 = det 𝐿(𝜔(𝑘), 𝑘) for every
𝑘 in 𝐵(𝑘0 , 𝑟). If 𝜔(𝑘) was a constant function then this would mean 0 = det 𝐿(𝜔(𝑘), 𝑘) =
det 𝐿(𝜔 0 , 𝑘) = det(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) for every 𝑘 in 𝐵(𝑘0 , 𝑟). But det(𝜆𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) is a
monic polynomial in 𝜆 and so has a nonzero but finite number of zeros where as 𝑓 (𝑘) := 𝑒𝑖𝑘𝑑
being a nonconstant analytic function is an open map and hence 𝑓 (𝐵(𝑘0 , 𝑟)) is a nonempty
open set in ℂ and hence contains an uncountable number of elements. But then because
0 = det(𝑓 (𝑘)𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) for every 𝑘 in 𝐵(𝑘0 , 𝑟) this implies 𝑓 (𝐵(𝑘0 , 𝑟)) has only a finite
number of elements since they are all roots of det(𝜆𝐼𝑛 −Ψ(𝑑, 𝜔 0 )). A contradiction. Therefore
𝜔(𝑘) is a nonconstant function. This completes the proof.
Lemma 63 Let 𝜔 1 (𝑘) and 𝜔 2 (𝑘) are two different eigenvalue Puiseux series of 𝐿(⋅, 𝑘) ex-
153
panded about 𝑘0 with limit point 𝜔 0 . If 𝛽 1 , 𝛽 2 are generating eigenvectors of 𝐿(⋅, 𝑘) at the
point (𝜔 0 , 𝑘0 ) corresponding to 𝜔 1 (𝑘) and 𝜔 2 (𝑘), respectively, then
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛽 1 , J 𝛽 2 ⟩ℂ = 0.
(4.74)
Proof. Let 𝜔 1 (𝑘) and 𝜔 2 (𝑘) are two different eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded
about 𝑘0 with limit point 𝜔 0 . Let 𝛽 1 , 𝛽 2 are generating eigenvectors of 𝐿(⋅, 𝑘) at the point
(𝜔 0 , 𝑘0 ) corresponding to 𝜔 1 (𝑘) and 𝜔 2 (𝑘), respectively. Then by Proposition 62 𝜔 1 (𝑘) and
𝜔 2 (𝑘) are single-valued real analytic functions and hence there adjoints satisfy 𝜔 ∗𝑗 (𝑘) =
𝜔 𝑗 (𝑘) = 𝜔 𝑗 (𝑘) for 𝑗 = 1, 2. By Lemma 61 we know J 𝛽 2 is a generating eigenvector of 𝐿∗ (⋅, 𝑘)
at the point (𝜔 0 , 𝑘0 ) = (𝜔 0 , 𝑘0 ) and associated with 𝜔 2 (𝑘). This implies by Proposition 29,
since 𝜔 1 (𝑘) and 𝜔 ∗2 (𝑘) = 𝜔 2 (𝑘) are different analytic functions, that
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛽 1 , J 𝛽 2 ⟩ℂ = 0.
(4.75)
This completes the proof.
Proposition 64 The holomorphic matrix function 𝐿(⋅, 𝑘0 ) has 𝜔 0 an eigenvalue of finite
algebraic multiplicity. Furthermore, it is a semisimple eigenvalue of 𝐿(⋅, 𝑘0 ). Moreover, its
geometric multiplicity as an eigenvalue of 𝐿(⋅, 𝑘0 ) is 𝑔 and this is the order of the analytic
function det 𝐿(𝜔, 𝑘0 ) = 𝐷(𝑘0 , 𝜔) at 𝜔 = 𝜔 0 .
Proof.
First, since (𝑘0 , 𝜔 0 ) ∈ ℬℝ then 0 = 𝐷(𝑘0 , 𝜔 0 ) = det(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) =
(−1)𝑛 det(𝐿(𝜔 0 , 𝑘0 )) so that 𝜔 0 ∈ 𝜎(𝐿(⋅, 𝑘0 )) and hence by definition 𝜔 0 an eigenvalue of
the holomorphic matrix function 𝐿(⋅, 𝑘0 ). To show that it is an eigenvalue of finite algebraic
multiplicity and a semisimple eigenvalue, it suffices to show, by Proposition 15, that
𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 = −𝐿(𝜔 0 , 𝑘0 )𝛾 1
154
has no solution for 𝛾 0 ∈ ker(𝐿(𝜔 0 , 𝑘0 ))/{0} and 𝛾 1 ∈ ℂ𝑛 .
Suppose there existed vectors 𝛾 0 ∈ ker(𝐿(𝜔 0 , 𝑘0 ))/{0}, 𝛾 1 ∈ ℂ𝑛 such that 𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 =
−𝐿(𝜔 0 , 𝑘0 )𝛾 1 . It follows from this, the fact ker(𝐿(𝜔 0 , 𝑘0 )) = ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )), Proposition 45 along with the fact (𝑘0 , 𝜔 0 ) is a point of definite type, that we have
1
0 ∕= 𝑞(𝑘0 ,𝜔0 ) (𝛾 0 , 𝛾 0 ) = ⟨Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 )𝛾 0 , 𝛾 0 ⟩ℂ
𝑑
1
1
= ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛾 0 , Ψ(𝑑, 𝜔 0 )𝛾 0 ⟩ℂ = ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛾 0 , 𝑒𝑖𝑘0 𝑑 𝛾 0 ⟩ℂ
𝑑
𝑑
𝑒𝑖𝑘0 𝑑
𝑒𝑖𝑘0 𝑑
=−
⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛾 0 , J 𝛾 0 ⟩ℂ = −
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 , J 𝛾 0 ⟩ℂ .
𝑑
𝑑
But 𝐿(𝜔 0 , 𝑘0 )𝛾 0 = 0 and so by Lemma 59 this implies 0 = 𝐿∗ (𝜔 0 , 𝑘0 )J 𝛾 0 = 𝐿(𝜔 0 , 𝑘0 )∗ J 𝛾 0
which implies
⟨𝐿(𝜔 0 , 𝑘0 )𝛾 1 , J 𝛾 0 ⟩ℂ = ⟨𝛾 1 , 𝐿(𝜔 0 , 𝑘0 )J 𝛾 0 ⟩ℂ = 0
which implies
0 ∕= −
𝑒𝑖𝑘0 𝑑
𝑒𝑖𝑘0 𝑑
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 , J 𝛾 0 ⟩ℂ = −
⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝛾 0 + 𝐿(𝜔 0 , 𝑘0 )𝛾 1 , J 𝛾 0 ⟩ℂ = 0.
𝑑
𝑑
This is a contradiction. Thus 𝜔 0 is an eigenvalue of 𝐿(⋅, 𝑘0 ) of finite algebraic multiplicity
and a semisimple eigenvalue. And so its geometric multiplicity equals its algebraic multiplicity and hence by Proposition 10, its geometric multiplicity is dim ker(𝐿(𝜔 0 , 𝑘0 )) =
dim ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) = 𝑔 and its algebraic multiplicity again 𝑔 is equal to the order of
the analytic function det(𝐿(𝜔, 𝑘0 )) at 𝜔 = 𝜔 0 . And since det(𝐿(𝜔, 𝑘0 )) = (−1)𝑛 det(𝑒𝑖𝑘0 𝑑 𝐼𝑛 −
Ψ(𝑑, 𝜔 0 )) = 𝐷(𝑘0 , 𝜔) for 𝜔 ∈ 𝐵(𝜔 0 , 𝑟0 ) this implies 𝑔 is the order of the zero of the analytic
function 𝐷(𝑘0 , 𝜔) at 𝜔 = 𝜔 0 . This completes the proof of the proposition.
Proposition 65 The holomorphic matrix function 𝐿(𝜔 0 , ⋅) has 𝑘0 an eigenvalue of finite
algebraic multiplicity. Moreover, 𝑘0 as an eigenvalue of 𝐿(𝜔 0 , ⋅) has 𝑔, 𝑚1 ≥ . . . ≥ 𝑚𝑔 , 𝑚
155
as its geometric, partial, and algebraic multiplicities, respectively. In particular, the order of
the zero of the analytic function det 𝐿(𝜔 0 , 𝑘) = 𝐷(𝑘, 𝜔 0 ) at 𝑘 = 𝑘0 is 𝑚.
Proof. We have 𝐿(𝜔 0 , 𝑘) = Ψ(𝑑, 𝜔 0 )−𝑒𝑖𝑘𝑑 𝐼𝑛 . If we define the holomorphic matrix functions
𝐹 (𝜆) := Ψ(𝑑, 𝜔 0 ) − 𝜆𝐼𝑛 and 𝐺(𝑘) := Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘𝑑 𝐼𝑛 then it follows that 𝐺(𝑘) = 𝐹 (𝑒𝑖𝑘𝑑 ). It
thus follows that 𝑘0 is an eigenvalue of 𝐺, 𝜆0 := 𝑒𝑖𝑘0 𝑑 is an eigenvalue of 𝐹 , and there is a oneto-one correspondence with their geometric, partial, and algebraic multiplicities. But as we
know by Example 1 in chapter 3 and by definition of the numbers 𝑔, 𝑚1 ≥ . . . ≥ 𝑚𝑔 , and 𝑚,
it follows that the geometric, partial, and algebraic multiplicities of the eigenvalue 𝜆0 of 𝐹 are
just, in the Jordan normal form of the matrix Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 , 𝑔
the number of Jordan blocks, their orders 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 = 𝑚1 + ⋅ ⋅ ⋅ + 𝑚𝑔 , respectively.
Thus 𝑔, 𝑚1 , . . . , 𝑚𝑔 , and 𝑚 are the geometric, partial, and algebraic multiplicities of the
eigenvalue 𝑘0 of 𝐿(𝜔 0 , 𝑘). By Proposition 10, since 𝑚 being the algebraic multiplicity of the
eigenvalue 𝑘0 of 𝐿(𝜔 0 , ⋅), we know that 𝑚 is the order of the zero of the function det 𝐿(𝜔 0 , 𝑘)
at 𝑘 = 𝑘0 . And since det 𝐿(𝜔 0 , 𝑘) = 𝐷(𝑘, 𝜔 0 ), the proof is complete.
Theorem 66 There exists a 𝛿 > 0 such that 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) ⊆ 𝑈 and ℬ ∩ 𝐵(𝑘0 , 𝛿) ×
𝐵(𝜔 0 , 𝜖) is the union of the graphs of 𝑔 nonconstant real analytic functions 𝜔 1 (𝑘), . . . , 𝜔 𝑔 (𝑘)
given by convergent power series
𝑚𝑗
𝜔 𝑗 (𝑘) = 𝜔 0 + 𝜈 𝑗,𝑚𝑗 (𝑘 − 𝑘0 )
+
∞
∑
𝜈 𝑗,𝑙 (𝑘 − 𝑘0 )𝑙 , ∣𝑘 − 𝑘0 ∣ < 𝛿
(4.76)
𝑙=𝑚𝑗 +1
where
𝜈 𝑗,𝑚𝑗 ∕= 0,
(4.77)
for 𝑗 = 1, . . . , 𝑔. Moreover, there exists analytic functions 𝜑1 (𝑘), . . . , 𝜑𝑔 (𝑘) belonging to
156
𝒪(𝐵(𝑘0 , 𝛿), ℂ𝑛 ) such that
𝜑𝑗 (𝑘) ∕= 0,
𝐿(𝜔 𝑗 (𝑘), 𝑘)𝜑𝑗 (𝑘) = 0, ∣𝑘 − 𝑘0 ∣ < 𝛿,
𝑗 = 1, . . . , 𝑔
(4.78)
and the vectors 𝜑1 (𝑘), . . . , 𝜑𝑔 (𝑘) are linearly independent for ∣𝑘 − 𝑘0 ∣ < 𝛿.
Proof. The spectrum 𝜎(𝐿) of the holomorphic matrix-valued function 𝐿 is given by
𝜎(𝐿) = {(𝜔, 𝑘) ∈ 𝑈 ∣ det 𝐿(𝜔, 𝑘) = 0}.
It follows from the definition of 𝐿 and the set 𝑈 = 𝐵(𝜔 0 , 𝑟0 ) × 𝐵(𝑘0 , 𝑟0 ) ⊆ Ω × ℂ that the
inverse set 𝜎(𝐿)−1 is a subset of the Bloch variety ℬ since
𝜎(𝐿)−1 = {(𝑘, 𝜔) ∈ 𝐵(𝑘0 , 𝑟0 ) × 𝐵(𝜔 0 , 𝑟0 ) ∣ det 𝐿(𝜔, 𝑘) = 0}
= {(𝑘, 𝜔) ∈ 𝐵(𝑘0 , 𝑟0 ) × 𝐵(𝜔 0 , 𝑟0 ) ∣ det 𝐷(𝑘, 𝜔) = 0}
= ℬ ∩ 𝐵(𝑘0 , 𝑟0 ) × 𝐵(𝜔 0 , 𝑟0 ).
Now by Proposition 64 the algebraic multiplicity of the eigenvalue 𝜔 0 of 𝐿(⋅, 𝑘0 ) is 𝑔 and so
by Theorem 21, there exists an 𝛿 > 0, which we may assume without loss of generality 𝛿 ≤ 𝑟0 ,
such that 𝜎(𝐿)−1 ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) is the union of the graphs of 𝑙 convergent Puiseux
series 𝜔 1 (𝑘), . . . , 𝜔 𝑙 (𝑘) expanded about 𝑘0 with limit point 𝜔 0 , domain 𝐵(𝑘0 , 𝛿), and periods
𝑞1 , . . . , 𝑞𝑙 , respectively, which satisfy 𝑔 = 𝑞1 +⋅ ⋅ ⋅+𝑞𝑙 . This implies each 𝜔 𝑗 (𝑘) is an eigenvalue
Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 with limit point 𝜔 0 and so by Proposition 62,
𝜔 𝑗 (𝑘) is a single-valued nonconstant real analytic function of 𝑘 in the domain 𝐵(𝑘0 , 𝛿), for
𝑗 = 1, . . . , 𝑙. Then, since 𝛿 ≤ 𝑟0 ≤ 𝜖 and 𝜎(𝐿)−1 = ℬ ∩ 𝐵(𝑘0 , 𝑟0 ) × 𝐵(𝜔 0 , 𝑟0 ), we have
𝜎(𝐿)−1 ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) = ℬ ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) and it is the union of the graphs of 𝑔
convergent Puiseux series 𝜔 1 (𝑘), . . . , 𝜔 𝑙 (𝑘) single-valued nonconstant real analytic function
of 𝑘 in the domain 𝐵(𝑘0 , 𝛿) with 𝜔 𝑗 (𝑘0 ) = 𝜔 0 , for 𝑗 = 1, . . . , 𝑙.
157
Let 𝜔 1 (𝑘), . . . , 𝜔 𝑔ˆ(𝑘) denote all those analytic functions which are not identically equal on
𝐵(𝑘0 , 𝛿).
This implies now that they have a convergent series expansion
𝜔 𝑗 (𝑘) = 𝜔 0 + 𝜈 𝑗,𝑞𝑗 (𝑘 − 𝑘0 )𝑞𝑗 +
∞
∑
𝜈 𝑗,𝑙 (𝑘 − 𝑘0 )𝑙 , ∣𝑘 − 𝑘0 ∣ < 𝛿
𝑙=𝑞𝑗 +1
where
𝜈 𝑗,𝑞𝑗 ∕= 0,
for 𝑗 = 1, . . . , 𝑔ˆ. Without loss of generality we may assume 𝑞1 ≥ ⋅ ⋅ ⋅ ≥ 𝑞𝑔ˆ.
Now by the Weierstrass Preparation Theorem, see Theorem 18 in the chapter 3, since det 𝐿 ∈
𝒪((𝜔 0 , 𝑘0 )), det 𝐿(𝜔 0 , 𝑘0 ) = 0, det 𝐿(⋅, 𝑘0 ) ∕= 0 which follows since 𝐿(⋅, 𝑘0 ) has 𝜔 0 as a
semisimple eigenvalue, then
det 𝐿 = 𝑓0 𝑓1 ,
𝑓0 ∈ 𝒪0 ((𝜔 0 , 𝑘0 )),
𝑓1 ∈ 𝒪((𝜔 0 , 𝑘0 ))
where 𝑓1 (𝜔, 𝑘) ∕= 0 for (𝜔, 𝑘) ∈ 𝐵(𝜔 0 , 𝑟) × 𝐵(𝑘0 , 𝑟) with 0 < 𝑟 ≪ 1 and deg 𝑓0 = 𝑔, since 𝑔 is
the order of the zero of det 𝐿(𝜔, 𝑘0 ) = 𝑓0 (𝜔, 𝑘0 )𝑓1 (𝜔, 𝑘0 ) because by Proposition 64 we know
𝜔 0 is a semisimple eigenvalue of 𝐿(⋅, 𝑘0 ), which means its geometric multiplicity 𝑔 equals
its algebraic multiplicity equals the order of the zero of det 𝐿(𝜔, 𝑘0 ) at 𝜔 = 𝜔 0 which thus
equals the order of the zero of 𝑓0 (𝜔, 𝑘0 ) at 𝜔 = 𝜔 0 which is by definition the deg 𝑓0 . Now by
Theorem 19 since 𝑓0 ∕= 1 because deg 𝑓0 = 𝑔 > 1 there exists unique and distinct irreducibles
𝑝1 , . . . , 𝑝𝑠 ∈ 𝒪0 ((𝜔 0 , 𝑘0 )) such that
𝑓0 =
𝑠
∏
𝑗=1
158
𝑚(𝑗)
𝑝𝑗
where 𝑚(𝑗) is the unique number representing the multiplicity of the irreducible factor 𝑝𝑗 ,
∑𝑠
for 𝑗 = 1, . . . , 𝑠 and 𝑔 = deg 𝑓0 =
ˆ𝑗 (𝜔, 𝑘) := 𝜔 − 𝜔 𝑗 (𝑘) =
𝑗=1 𝑚(𝑗) deg 𝑝𝑗 . But for 𝑝
(𝜔 − 𝜔 0 ) − (𝜔 𝑗 (𝑘) − 𝜔 0 ) we have 𝑝ˆ𝑗 ∈ 𝒪0 ((𝜔 0 , 𝑘0 )) and it irreducible, for 𝑗 = 1, . . . , 𝑔ˆ. Now
though as we have shown 𝜎(𝐿)−1 ∩ 𝐵(𝑘0 , 𝛿) × 𝐵(𝜔 0 , 𝜖) is the graph of the 𝜔 1 (𝑘), . . . , 𝜔 𝑔ˆ(𝑘)
and hence since det 𝐿 = 𝑓0 𝑓1 then the zeros of 𝑓0 are exactly the zeros of 𝑝ˆ𝑗 ∈ 𝒪0 ((𝜔 0 , 𝑘0 ))
implying by Theorem 20 and the uniqueness of the factorization into irreducibles we must
have after possible reordering of the indices, 𝑔ˆ = 𝑠, 𝑝𝑗 = 𝑝ˆ𝑗 for 𝑗 = 1, . . . , 𝑔ˆ, and
𝑔ˆ
∏
𝑓0 =
𝑚(𝑗)
𝑝ˆ𝑗
.
𝑗=1
This of course implies, since deg 𝑝ˆ𝑗 = 1 for each 𝑗, that
𝑔 = deg 𝑓0 =
𝑔ˆ
∑
𝑚(𝑗)
deg 𝑝ˆ𝑗
=
𝑗=1
𝑔ˆ
∑
𝑚(𝑗) deg 𝑝ˆ𝑗 =
𝑗=1
𝑔ˆ
∑
𝑚(𝑗).
𝑗=1
Now it follows from what we have just shown that
det 𝐿(𝜔 0 , 𝑘) = 𝑓0 (𝜔 0 , 𝑘)𝑓1 (𝜔 0 , 𝑘) = 𝑓1 (𝜔 0 , 𝑘)
𝑔ˆ
∏
𝑚(𝑗)
𝑝ˆ𝑗
(𝜔 0 , 𝑘)
𝑗=1
𝑔ˆ
𝑔ˆ
∏
∏
𝑚(𝑗)
= 𝑓1 (𝜔 0 , 𝑘) (𝜔 0 − 𝜔 𝑗 (𝑘))
= 𝑓1 (𝜔 0 , 𝑘) (−𝜈 𝑗,𝑞𝑗 (𝑘 − 𝑘0 )𝑞𝑗 + 𝑜((𝑘 − 𝑘0 )𝑞𝑗 ))𝑚(𝑗)
𝑗=1
(
=
𝑓1 (𝜔 0 , 𝑘)
𝑔ˆ
𝑗=1
)
∏
(−𝜈 𝑗,𝑞𝑗 )𝑚(𝑗)
∑𝑔ˆ
𝑗=1 𝑞𝑗 𝑚(𝑗)
(𝑘 − 𝑘0 )
(
)
∑𝑔ˆ
+ 𝑜 (𝑘 − 𝑘0 ) 𝑗=1 𝑞𝑗 𝑚(𝑗)
𝑗=1
𝑔ˆ
∏
𝑓1 (𝜔 0 , 𝑘0 ) (−𝜈 𝑗,𝑞𝑗 )𝑚(𝑗)
(
=
)
∑𝑔ˆ
(𝑘 − 𝑘0 )
𝑗=1 𝑞𝑗 𝑚(𝑗)
(
∑𝑔ˆ
+ 𝑜 (𝑘 − 𝑘0 )
𝑗=1 𝑞𝑗 𝑚(𝑗)
)
𝑗=1
𝑔ˆ
∏
as 𝑘 → 𝑘0 . The leading term 𝑓1 (𝜔 0 , 𝑘0 ) (−𝜈 𝑗,𝑞𝑗 )𝑚(𝑗) ∕= 0 implying the order of the zero
𝑗=1
∑𝑔ˆ
of det 𝐿(𝜔 0 , 𝑘) at 𝑘 = 𝑘0 is 𝑗=1 𝑞𝑗 𝑚(𝑗), but by Proposition 65 the order of the zero of
159
det 𝐿(𝜔 0 , 𝑘) at 𝑘 = 𝑘0 is 𝑚. This implies
𝑚=
𝑔ˆ
∑
𝑞𝑗 𝑚(𝑗).
𝑗=1
Lemma 67 For 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1, 𝜔 𝑗 (𝑘) is a semisimple eigenvalue of 𝐿(⋅, 𝑘) and
𝑚(𝑗) = dim ker 𝐿(𝜔 𝑗 (𝑘), 𝑘),
𝑗 = 1, . . . , 𝑔ˆ.
(4.79)
Proof. For 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1 we have 𝜔 𝑗1 (𝑘) ∕= 𝜔 𝑗2 (𝑘) for 𝑗1 ∕= 𝑗2 with 1 ≤ 𝑗1 , 𝑗2 ≤ 𝑔ˆ, since
by assumption the functions 𝜔 1 (𝑘), . . . , 𝜔 𝑔ˆ(𝑘) where not identically equal on 𝐵(𝑘0 , 𝛿). And
since
𝑔ˆ
∏
det 𝐿(𝜔, 𝑘) = 𝑓0 (𝜔, 𝑘)𝑓1 (𝜔, 𝑘) = 𝑓1 (𝜔, 𝑘) (𝜔 − 𝜔 𝑗 (𝑘))𝑚(𝑗) ,
𝑓1 (𝜔, 𝑘) ∕= 0
𝑗=1
for all (𝜔, 𝑘) ∈ 𝐵(𝜔 0 , 𝑟) × 𝐵(𝑘0 , 𝑟) for 0 < 𝑟 ≪ 1, these fact together with 𝜔 𝑗 (𝑘0 ) = 𝜔 0
imply that if 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1 then the order of the zero of the function det 𝐿(⋅, 𝑘) at
𝜔 = 𝜔 𝑗 (𝑘) is 𝑚(𝑗) which by 10 this is the algebraic multiplicity of the eigenvalue 𝜔 𝑗 (𝑘) of
the holomorphic matrix function 𝐿(⋅, 𝑘), for 𝑗 = 1, . . . , 𝑔ˆ. Now it follows by the local Smith
form (see Theorem 9) of the holomorphic matrix function 𝐿(𝜔 𝑗 (𝑘), 𝑘) at the eigenvalue 𝑘0
that
𝜇(𝑗) := dim ker 𝐿(𝜔 𝑗 (𝑘), 𝑘),
𝑗 = 1, . . . , 𝑔ˆ, 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1,
are constant. Our goal is to now show that 𝜇(𝑗) = 𝑚(𝑗), for 𝑗 = 1, . . . , 𝑔ˆ. Notice first that by
definition of 𝜇(𝑗), it is the geometric multiplicity of the eigenvalue of 𝜔 𝑗 (𝑘) of the holomorphic
matrix function 𝐿(⋅, 𝑘) and since 𝑚(𝑗) is the algebraic multiplicity of the eigenvalue 𝜔 𝑗 (𝑘) of
160
the holomorphic matrix function 𝐿(⋅, 𝑘) for 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1. But it follows from Theorem
47 that if ∣𝑘 − 𝑘0 ∣ ≪ 1 and 𝑘 is real then (𝑘, 𝜔 𝑗 (𝑘)) is point of definite type for the canonical
ODEs in (4.13) and hence by Proposition 64 we know 𝜔 𝑗 (𝑘) is a semisimple eigenvalue of
the holomorphic matrix function 𝐿(⋅, 𝑘) which by definition of semisimple eigenvalue means
its geometric multiplicity 𝜇(𝑗) equals its algebraic multiplicity 𝑚(𝑗), i.e., 𝜇(𝑗) ≤ 𝑚(𝑗), for
𝑗 = 1, . . . , 𝑔ˆ. And therefore this implies for all 𝑘 satisfying 0 < ∣𝑘 − 𝑘0 ∣ ≪ 1 we have
𝑚(𝑗) = 𝜇(𝑗) = dim ker 𝐿(𝜔 𝑗 (𝑘), 𝑘),
𝑗 = 1, . . . , 𝑔ˆ,
which means 𝜔 𝑗 (𝑘) is a semisimple eigenvalue of 𝐿(⋅, 𝑘). This completes the proof.
Now it follows this lemma and the local Smith form, Theorem 9, of the holomorphic matrix
function 𝐿(𝜔 𝑗 (𝑘), 𝑘) in the variable 𝑘 with eigenvalue 𝑘0 , that there exists an 𝑟 > 0 and there
exists function 𝑁𝑗 , 𝑀𝑗 ∈ 𝒪(𝐵(𝑘0 , 𝑟), 𝑀𝑛 (ℂ)) with 𝑁𝑗 (𝑘), 𝑀𝑗 (𝑘) invertible matrices and
𝑁𝑗 (𝑘)𝐿(𝜔 𝑗 (𝑘), 𝑘)𝑀𝑗 (𝑘) = diag{0, . . . , 0, (𝑘 − 𝑘0 )𝑚𝑗,𝑚(𝑗)+1 , . . . , (𝑘 − 𝑘0 )𝑚𝑗,𝑛 }
(4.80)
for every 𝑘 ∈ 𝐵(𝑘0 , 𝑟), where the number of zeros down the diagonal is given by 𝑚(𝑗) and
{𝑚𝑗,𝑖 }𝑛𝑖=𝑚(𝑗)+1 is a decreasing sequence of nonnegative integers, for 𝑗 = 1, . . . , 𝑔ˆ.
Let 𝑒1 , . . . , 𝑒𝑛 denote the standard orthonormal vectors from ℂ𝑛 . We now define the vectorvalued functions
𝜑𝑗,𝑖 (𝑘) := 𝑀𝑗 (𝑘)𝑒𝑖 ,
𝑖 = 1, . . . , 𝑚(𝑗),
𝑗 = 1, . . . , 𝑔ˆ, ∣𝑘 − 𝑘0 ∣ < 𝑟.
It follows, since 𝑀𝑗 ∈ 𝒪(𝐵(𝑘0 , 𝑟), 𝑀𝑛 (ℂ)), that 𝜑𝑗,𝑖 ∈ 𝒪(𝐵(𝑘0 , 𝑟), ℂ𝑛 ) for all 𝑖, 𝑗 and
𝐿(𝜔 𝑗 (𝑘), 𝑘)𝜑𝑗,𝑖 (𝑘) = 0,
𝑖 = 1, . . . , 𝑚(𝑗),
161
𝑗 = 1, . . . , 𝑔ˆ, ∣𝑘 − 𝑘0 ∣ < 𝑟.
Lemma 68 For every 𝑘 with ∣𝑘 − 𝑘0 ∣ ≪ 1, the vectors 𝜑𝑗,𝑖 (𝑘), 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ
𝑚(𝑖)
are linear independent, {𝜑𝑗,𝑖 (𝑘)}𝑖=1 is a basis for ker 𝐿(𝜔 𝑗 (𝑘), 𝑘), for 𝑗 = 1, . . . , 𝑔ˆ, and
∪𝑔ˆ
𝑚(𝑖)
𝑗=1 {𝜑𝑗,𝑖 (𝑘0 )}𝑖=1 is a basis for ker 𝐿(𝜔 0 , 𝑘0 ).
Proof. For the first part of the theorem it suffices to prove that their limit points 𝜑𝑗,𝑖 (𝑘0 ),
𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ are linearly independent. Suppose they were not. Then there
∑ˆ ∑𝑚(𝑗)
exists constants 𝑐𝑗,𝑖 , 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ with 𝑔𝑗=1
𝑖=1 ∣𝑐𝑗,𝑖 ∣ > 0 such that
0=
𝑔ˆ 𝑚(𝑗)
∑
∑
𝑐𝑗,𝑖 𝜑𝑗,𝑖 (𝑘0 ).
𝑗=1 𝑖=1
But if 𝑗1 , 𝑗2 ∈ {1, . . . , 𝑔ˆ, 𝑖1 ∈ {1, . . . , 𝑚(𝑗1 ), and 𝑖2 ∈ {1, . . . , 𝑚(𝑗2 )} then (𝜔 𝑗1 (𝑘), 𝜑𝑖1 ,𝑗1 (𝑘))
and (𝜔 𝑗2 (𝑘), 𝜑𝑖2 ,𝑗2 (𝑘)) are eigenpair Puiseux series of 𝐿(⋅, 𝑘) expanded about 𝑘0 . By the
definition of generating eigenvector, 𝜑𝑗1 ,𝑖1 (𝑘0 ) and 𝜑𝑗2 ,𝑖2 (𝑘0 ) are generating eigenvectors of
𝐿(⋅, 𝑘) at the point (𝜔 0 , 𝑘0 ) and associated with 𝜔 1 (𝑘) and 𝜔 2 (𝑘), respectively. Now if 𝑗1 ∕= 𝑗2
then 𝜔 𝑗1 (𝑘) and 𝜔 𝑗2 (𝑘) are two different eigenvalue Puiseux series of 𝐿(⋅, 𝑘) expanded about
𝑘0 with the same limit point 𝜔 0 . This implies by Lemma 63 that
0 = ⟨𝐿𝜔 (𝜔 0 , 𝑘0 )𝜑𝑗1 ,𝑖1 (𝑘0 ), J 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩ℂ = −⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩ℂ
= −𝑒𝑖𝑘0 𝑑 𝑒−𝑖𝑘0 𝑑 ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩ℂ
= −𝑒−𝑖𝑘0 𝑑 ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝜑𝑗1 ,𝑖1 (𝑘0 ), Ψ(𝑑, 𝜔 0 )𝜑𝑗2 ,𝑖2 (𝑘0 )⟩ℂ
= −𝑒−𝑖𝑘0 𝑑 ⟨Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 )𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩ℂ
= −𝑑𝑒−𝑖𝑘0 𝑑 𝑞(𝑘0 ,𝜔0 ) (𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 ))
= −𝑑𝑒−𝑖𝑘0 𝑑 sgn(𝑞(𝑘0 ,𝑑) )⟨𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩(𝑘0 ,𝜔0 ) .
162
Thus we have the orthogonality relations
⟨𝜑𝑗1 ,𝑖1 (𝑘0 ), 𝜑𝑗2 ,𝑖2 (𝑘0 )⟩(𝑘0 ,𝜔0 ) = 0,
Now if we let 𝜓 𝑗 :=
∑𝑚(𝑗)
𝑖=1
𝑐𝑗,𝑖 𝜑𝑗,𝑖 (𝑘0 ) then it follows from these orthogonality relations that
⟨𝜓 𝑗1 , 𝜓 𝑗2 ⟩(𝑘0 ,𝜔0 ) = 0,
And hence since 0 =
∑𝑔ˆ
𝑗=1
𝑗1 ∕= 𝑗2 .
∑𝑚(𝑗)
0=
𝑖=1
𝑐𝑗,𝑖 𝜑𝑗,𝑖 (𝑘0 ) =
⟨ 𝑔ˆ
∑
𝑗=1
∑𝑔ˆ
𝑗=1
𝑗1 ∕= 𝑗2 .
𝜓 𝑗 it follows that
⟩
= ⟨𝜓 𝑗1 , 𝜓 𝑗1 ⟩(𝑘0 ,𝜔0 ) .
𝜓 𝑗 , 𝜓 𝑗1
(𝑘0 ,𝜔 0 )
But by Proposition 45 we know that ⟨ , ⟩(𝑘0 ,𝜔0 ) is an inner product on ker 𝐿(𝜔 0 , 𝑘0 ) and
so we conclude 𝜓 𝑗 = 0 for every 𝑗. But 𝑀𝑗 (𝑘0 ) is an invertible matrix and hence 0 =
∑𝑚(𝑗)
∑
𝑀𝑗 (𝑘0 )−1 𝜓 𝑗 = 𝑚(𝑗)
𝑖=1 ∣𝑐𝑗,𝑖 ∣ = 0 for every 𝑗. Thus we conclude
𝑖=1 𝑐𝑗,𝑖 𝑒𝑖 which implies that
∑𝑔ˆ ∑𝑚(𝑗)
𝑖=1 ∣𝑐𝑗,𝑖 ∣ = 0 a contradiction. Therefore the limit points 𝜑𝑗,𝑖 (𝑘0 ), 𝑖 = 1, . . . , 𝑚(𝑗),
𝑗=1
𝑗 = 1, . . . , 𝑔ˆ are linearly independent. This completes the proof.
Without loss of generality we may take 0 < 𝑟 ≤ 𝛿 such that the consequences of Lemma 68 is
true for the functions 𝜑𝑗,𝑖 (𝑘)), 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ. Define the functions 𝜔 𝑗,𝑖 (𝑘) :=
∑ˆ ∑𝑚(𝑗)
𝜔 𝑗 (𝑘), 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ on 𝐵(𝑘0 , 𝑟). It follows that there is 𝑔𝑗=1
𝑖=1 = 𝑔 of
them, they are nonconstant real analytic functions with the union of there graphs being ℬ ∩
𝐵(𝑘0 , 𝑟)×𝐵(𝜔 0 , 𝜀). The order of the zero of 𝜔 𝑗 (𝑘)−𝜔 0 at 𝑘 = 𝑘0 is 𝑞𝑗,𝑖 := 𝑞𝑗 , 𝑖 = 1, . . . , 𝑚(𝑗),
∑ˆ ∑𝑚(𝑗)
∑𝑔ˆ
𝑗 = 1, . . . , 𝑔ˆ such that 𝑔𝑗=1
𝑖=1 𝑞𝑖,𝑗 =
𝑗=1 𝑚(𝑗)𝑞𝑖,𝑗 = 𝑚. Moreover, the functions 𝜑𝑗,𝑖 (𝑘),
𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ are analytic functions belonging to 𝒪(𝐵(𝑘0 , 𝑟), ℂ𝑛 ) such that
for each 𝑘 ∈ 𝐵(𝑘0 , 𝑟) the vectors 𝜑𝑗,𝑖 (𝑘), 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ are linear independent,
∪ˆ
𝑚(𝑖)
𝑚(𝑖)
{𝜑𝑗,𝑖 (𝑘)}𝑖=1 are a basis for ker 𝐿(𝜔 𝑗,𝑖 (𝑘), 𝑘), for 𝑗 = 1, . . . , 𝑔ˆ, and 𝑔𝑗=1
{𝜑𝑗,𝑖 (𝑘0 )}𝑖=1 is a
basis for ker 𝐿(𝜔 0 , 𝑘0 ). Thus the proof this theorem will be complete if we can show that 𝑞𝑗,𝑖 ,
163
𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ are the dimensions of the Jordan blocks in the Jordan normal
form of Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 .
It follows from what we just stated that (𝜔 𝑖,𝑗 (𝑘), 𝜑𝑗,𝑖 (𝑘)), 𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ are
eigenpair Puiseux series of 𝐿(⋅, 𝑘). And the order of the zero of 𝜔 𝑗,𝑖 (𝑘) − 𝜔 0 at 𝑘 = 𝑘0 is 𝑞𝑗,𝑖 ,
𝑖 = 1, . . . , 𝑚(𝑗), 𝑗 = 1, . . . , 𝑔ˆ. We denote any one of them by (𝜔(𝑘), 𝜑(𝑘)). Then both 𝜔(𝑘)
and 𝜑(𝑘) are analytic at 𝑘 = 𝑘0 with 𝜔(𝑘0 ) = 𝜔 0 . Let 𝑞 be the order of the zero of 𝜔 𝑗 (𝑘) − 𝜔 0
at 𝑘 = 𝑘0 . It follows from Proposition 32 that 𝜑(𝑘) is a generating function of order 𝑞 for
𝐿(𝜔 0 , ⋅) at the eigenvalue 𝑘0 . It follows by Proposition 11 that
𝜑𝑙 :=
1 (𝑙)
𝜑 (𝜆0 ),
𝑗!
𝑙 = 0, . . . , 𝑞 − 1
is a Jordan chain of 𝐿(𝜔 0 , ⋅) of length 𝑞 corresponding to the eigenvalue 𝑘0 . By definition of
a Jordan chain for a holomorphic function we have
𝜑0 ∕= 0 and
𝑙
∑
1 (0,ℎ)
𝐿
(𝜔 0 , 𝑘0 )𝜑𝑙−ℎ = 0,
ℎ!
ℎ=0
𝑙 = 0, . . . , 𝑞 − 1.
But 𝐿(0,0) (𝜔 0 , 𝑘0 ) = (Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘0 𝑑 𝐼𝑛 ) and 𝐿(0,ℎ) (𝜔 0 , 𝑘0 ) = −(𝑖𝑑)ℎ 𝑒𝑖𝑘0 𝑑 𝐼𝑛 for ℎ ≥ 0. Hence
for 𝑙 = 0, . . . , 𝑞 − 1 we have
0 = (Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘0 𝑑 𝐼𝑛 )𝜑0 ,
𝜑0 ∕= 0,
𝑙
𝑙
∑
∑
1 (0,ℎ)
𝑖𝑘0 𝑑
0=
𝐿
(𝜔 0 , 𝑘0 )𝜑𝑙−ℎ = (Ψ(𝑑, 𝜔 0 ) − 𝑒 𝐼𝑛 )𝜑𝑙 +
−(𝑖𝑑)ℎ 𝑒𝑖𝑘0 𝑑 𝜑𝑙−ℎ ,
ℎ!
ℎ=0
ℎ=1
Thus it follow that
(Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘0 𝑑 𝐼𝑛 )𝜑0 = 0,
𝜑0 ∕= 0,
(Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘0 𝑑 𝐼𝑛 )𝑙 𝜑𝑙 = (𝑖𝑑𝑒𝑖𝑘0 𝑑 )𝑙 𝜑0 , 1 ≤ 𝑙 ≤ 𝑞 − 1
164
𝑙≥1
This implies that 𝛾 𝑞−1 :=
𝜑𝑞−1
(𝑖𝑑𝑒𝑖𝑘0 𝑑 )𝑞−1
is a generalized eigenvector of Ψ(𝑑, 𝜔) of order 𝑞 corre-
sponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 . Thus if we define
𝛾 𝑙 := (Ψ(𝑑, 𝜔 0 ) − 𝑒𝑖𝑘0 𝑑 𝐼𝑛 )𝑞−1−𝑙 𝛾 𝑞−1 ,
𝑙 = 0, . . . , 𝑞 − 1
𝑞−1
is a Jordan chain of Ψ(𝑑, 𝜔 0 ) of length 𝑞 corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑
then {𝛾 𝑙 }𝑙=0
with 𝛾 0 = 𝜑0 the eigenvector.
𝑞
−1
𝑗,𝑖
Thus to each pair (𝜔 𝑖,𝑗 (𝑘), 𝜑𝑗,𝑖 (𝑘)) we get a Jordan chain, {𝛾 𝑗,𝑖,𝑙 }𝑙=0
, of Ψ(𝑑, 𝜔 0 ) of length
𝑞𝑗,𝑖 − 1 corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 with 𝛾 𝑗,𝑖,0 = 𝜑𝑗,𝑖 (𝑘0 ) the eigenvector. Thus since
∪ˆ
𝑚(𝑖)
{𝜑𝑗,𝑖 (𝑘0 )}𝑖=1
the elements in a Jordan chain are always linearly independent and, since 𝑔𝑗=1
is a basis for ker 𝐿(𝜔 0 , 𝑘0 ), the eigenvectors of these chains 𝛾 𝑗,𝑖,0 = 𝜑𝑗,𝑖 (𝑘0 ), 𝑖 = 1, . . . , 𝑚(𝑗),
𝑗 = 1, . . . , 𝑔ˆ are linearly independent and this implies the collection of Jordan chains
∪𝑔ˆ
𝑚(𝑖)
𝑗=1 {𝜑𝑗,𝑖 (𝑘0 )}𝑖=1 are linearly independent. But the dimension of the generalized eigenspace
for Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 is equal to its algebraic multiplicity which is
∪ˆ
𝑚(𝑖)
{𝜑𝑗,𝑖 (𝑘0 )}𝑖=1 are linearly independent and there
𝑚. And the collection of Jordan chains 𝑔𝑗=1
∑ˆ ∑𝑚(𝑗)
are a total of 𝑔𝑗=1
𝑖=1 𝑞𝑖,𝑗 = 𝑚 of them and they belong to the generalized eigenspace
for Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 . Therefore the collection of Jordan chains
∪𝑔ˆ
𝑚(𝑖)
𝑗=1 {𝜑𝑗,𝑖 (𝑘0 )}𝑖=1 are a basis for the generalized eigenspace for Ψ(𝑑, 𝜔 0 ) corresponding to
the eigenvalue 𝑒𝑖𝑘0 𝑑 , and in particular, they are a Jordan basis for that space for matrix
Ψ(𝑑, 𝜔 0 ). This implies in the Jordan normal form of Ψ(𝑑, 𝜔 0 ) corresponding to the eigen∑ˆ ∑𝑚(𝑗)
value 𝑒𝑖𝑘0 𝑑 it has 𝑔𝑗=1
𝑖=1 = 𝑔 Jordan blocks whose orders are exactly 𝑞𝑖,𝑗 , 𝑖 = 1, . . . , 𝑚(𝑗),
𝑗 = 1, . . . , 𝑔ˆ, but by hypothesis these orders where 𝑚1 , . . . , 𝑚𝑔 implying by the uniqueness
of the orders of the Jordan blocks these lists have the same numbers in them. This therefore
completes the proof.
The proof of Theorem 48.(i) now follows from Proposition 64 and Proposition 65. The
proof of Theorem 48.(ii) follows from Theorem 66. This completes the proof of Theorem 48.
165
Proof. [Corollary 49] Suppose the conditions
𝐷(𝑘0 , 𝜔 0 ) = 0,
∂𝐷
(𝑘0 , 𝜔 0 ) ∕= 0, (𝑘0 , 𝜔 0 ) ∈ ℝ × Ωℝ
∂𝜔
are satisfied where 𝐷(𝑘, 𝜔) = det(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔)), defined in Corollary 38. By Corollary
38 the first condition implies (𝑘0 , 𝜔 0 ) ∈ ℬℝ . Choose 𝑟 > 0 such that 𝐵(𝜔 0 , 𝑟) ⊆ Ω. Let
𝜆0 := 𝑒𝑖𝑘0 𝑑 and 𝑀 (𝜀) := Ψ(𝑑, 𝜀 + 𝜔 0 ) for 𝜀 ∈ 𝐵(0, 𝑟). By Proposition 33.(iii) we have
Ψ(𝑑, ⋅) ∈ 𝒪(𝐵(𝜔 0 , 𝑟), 𝑀𝑛 (ℂ)) and hence 𝑀 (⋅) ∈ 𝒪(𝐵(0, 𝑟), 𝑀𝑛 (ℂ)) which implies 𝑀 (𝜀) is
an analytic matrix function of the perturbation parameter 𝜀. Furthermore, 𝜆0 is an eigenvalue
of the unperturbed matrix 𝑀 (0) = Ψ(𝑑, 𝜔 0 ) since 0 = 𝐷(𝜔 0 , 𝑘0 ) = det(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) =
det(𝜆0 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )). Moreover, we have
∂
∂
det(𝜆𝐼𝑛 − 𝑀 (𝜀))∣(0,𝜆0 ) =
det(𝜆𝐼𝑛 − Ψ(𝑑, 𝜀 + 𝜔 0 ))∣(0,𝜆0 )
∂𝜀
∂𝜀
∂
∂
=
det(𝜆𝐼𝑛 − Ψ(𝑑, 𝜔))∣(𝜔0 ,𝜆0 ) =
det(𝑒𝑖𝑘𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔))∣(𝜔0 ,𝑘0 )
∂𝜔
∂𝜔
∂𝐷
=
(𝑘0 , 𝜔 0 ) ∕= 0
∂𝜔
Thus the generic condition, defined in [61, see (1.1)], is satisfied for the matrix function
𝑀 (𝜀) for the unperturbed matrix 𝑀 (0) with the eigenvalue 𝜆0 . By [61, Theorem 2.1.(iv)]
the Jordan normal form of 𝑀 (0) = Ψ(𝑑, 𝜔 0 )) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑
consists of a single 𝑚 × 𝑚 Jordan block and there exists an eigenvector 𝛾 0 of Ψ(𝑑, 𝜔 0 ) and
an eigenvector 𝛽 0 of Ψ(𝑑, 𝜔 0 )∗ corresponding to the eigenvalue 𝜆0 = 𝑒−𝑖𝑘0 𝑑 such that
⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛾 0 , 𝛽 0 ⟩ℂ ∕= 0.
But since the Jordan normal form of Ψ(𝑑, 𝜔 0 )) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑
consists of a single 𝑚 × 𝑚 Jordan block, this implies that dim ker(𝜆0 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) = 1.
166
But because Ψ(𝑑, 𝜔 0 )∗ is a adjoint of Ψ(𝑑, 𝜔 0 ) this implies the Jordan normal form corresponding to the eigenvalue 𝜆0 = 𝑒−𝑖𝑘0 𝑑 also consists of a single 𝑚 × 𝑚 Jordan block so that
dim ker(𝜆0 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) = 1 as well. And therefore, since any eigenvector 𝛾 of Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 must be a scalar multiple of 𝛾 0 and similarly any eigenvector
𝛽 of Ψ(𝑑, 𝜔 0 )∗ corresponding to the eigenvalue 𝜆0 must be a scalar multiple of 𝛽 0 , we must
have
⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛾, 𝛽⟩ℂ ∕= 0
for any eigenvector 𝛾 of Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 and for any
eigenvector 𝛽 of Ψ(𝑑, 𝜔 0 )∗ corresponding to the eigenvalue 𝜆0 = 𝑒−𝑖𝑘0 𝑑 . That is the key
statement.
Indeed, Proposition 39 implies
Ψ(𝑑, 𝜔 0 )∗ J Ψ(𝑑, 𝜔 0 ) = J
and by Proposition 33.(v) we know that Ψ(𝑑, 𝜔 0 ) is an invertible matrix and hence for any
eigenvector 𝛾 of Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 we have
Ψ(𝑑, 𝜔 0 )∗ J 𝛾 = J Ψ(𝑑, 𝜔 0 )−1 𝛾 = 𝑒−𝑖𝑘0 𝑑 J 𝛾.
Then since J is invertible this implies J 𝛾 is an eigenvector of Ψ(𝑑, 𝜔 0 )∗ corresponding to
the eigenvalue 𝑒−𝑖𝑘0 𝑑 = 𝜆0 . But this means
0 ∕= ⟨Ψ𝜔 (𝑑, 𝜔 0 )𝛾, J 𝛾⟩ℂ = −𝑒𝑖𝑘0 𝑑 𝑒−𝑖𝑘0 𝑑 ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛾, 𝛾⟩ℂ
= −𝑒−𝑖𝑘0 𝑑 ⟨J Ψ𝜔 (𝑑, 𝜔 0 )𝛾, Ψ(𝑑, 𝜔 0 )𝛾⟩ℂ
= −𝑒−𝑖𝑘0 𝑑 ⟨Ψ(𝑑, 𝜔 0 )∗ J Ψ𝜔 (𝑑, 𝜔 0 )𝛾, 𝛾⟩ℂ
167
−𝑖𝑘0 𝑑 1
= −𝑑𝑒
𝑑
∫
𝑑
⟨𝐴𝜔 (𝑥, 𝜔)𝜓(𝑥), 𝜓(𝑥)⟩ℂ 𝑑𝑥,
𝜓 := Ψ(⋅, 𝜔 0 )𝛾,
0
where the last equality follows from Corollary 43.
Thus we have shown for any 𝛾 ∈
ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) with 𝛾 ∕= 0 we have
1
𝑑
∫
𝑑
⟨𝐴𝜔 (𝑥, 𝜔)𝜓(𝑥), 𝜓(𝑥)⟩ℂ 𝑑𝑥 ∕= 0,
𝜓 := Ψ(⋅, 𝜔 0 )𝛾.
0
Therefore if 𝜓 is any nontrivial Bloch solution of the canonical ODEs in 4.13 with (𝑘0 , 𝜔 0 )
as its wavenumber-frequency pair, the by Theorem 35 there exists an eigenvector 𝛾 of the
monodromy matrix Ψ(𝑑, 𝜔 0 ) corresponding to the eigenvalue 𝑒𝑖𝑘0 𝑑 such that 𝜓 = Ψ(⋅, 𝜔 0 )𝛾,
i.e., 𝛾 ∈ ker(𝑒𝑖𝑘0 𝑑 𝐼𝑛 − Ψ(𝑑, 𝜔 0 )) with 𝛾 ∕= 0, and thus from what we just proved we must have
1
𝑑
∫
𝑑
⟨𝐴𝜔 (𝑥, 𝜔)𝜓(𝑥), 𝜓(𝑥)⟩ℂ 𝑑𝑥 ∕= 0.
0
This proves (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the canonical ODEs in 4.13. Moreover, from what we have shown the Jordan normal form of Ψ(𝑑, 𝜔 0 )) corresponding to the
eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 consists of a single 𝑚 × 𝑚 Jordan block.
We will now prove the converse. Suppose that (𝑘0 , 𝜔 0 ) ∈ ℬℝ is a point of definite type for the
canonical ODEs in (4.13) such that the Jordan normal form of Ψ(𝑑, 𝜔 0 ) corresponding to the
eigenvalue 𝜆0 = 𝑒𝑖𝑘0 𝑑 consists of a single 𝑚 × 𝑚 Jordan block, for some 𝑚 ∈ ℕ. By Theorem
48.(i), since 𝑔 = 1 in this case, we know that the order of the zero of the function 𝐷(𝑘0 , 𝜔)
at 𝜔 = 𝜔 0 is 1. But this is equivalent to 𝐷(𝑘0 , 𝜔 0 ) = 0 and
converse is true. This completes the proof.
168
∂𝐷
(𝑘0 , 𝜔 0 )
∂𝜔
∕= 0. Therefore the
4.5
Auxiliary Material
In this section we give some notation, convention, and background material frequently used
throughout this chapter.
Notation and Convention
(i) 𝑀𝑚,𝑛 (𝐸) := the set of all 𝑚 × 𝑛 matrices with entries in a space 𝐸.
(ii) 𝑀𝑚 (𝐸) := 𝑀𝑚,𝑚 (𝐸).
(iii) 𝐸 𝑚 := 𝑀𝑚,1 (𝐸).
(iv) As convention we identify 𝐸 with 𝐸 1 and the set of 𝑚-tuples whose entries are in 𝐸
with 𝑀1,𝑚 (𝐸).
(v) 𝐴𝑇 :=the transpose of the matrix 𝐴 ∈ 𝑀𝑚,𝑛 (𝐸).
(vi) (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) ) := the Banach space with norm
(
∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸) :=
𝑚 ∑
𝑛
∑
) 21
∣∣𝑎𝑖𝑗 ∣∣2
, for any 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐸)
(4.81)
𝑖=1 𝑖=1
where (𝐸, ∣∣ ⋅ ∣∣) is a Banach space.
(vii) L (𝐸, 𝐹 ) := the Banach space of all continuous linear operators from a Banach space
(𝐸, ∣∣ ⋅ ∣∣𝐸 ) to a Banach space (𝐹, ∣∣ ⋅ ∣∣𝐹 ) with the operator norm
∣∣𝑇 𝑒∣∣𝐹
𝑒∈𝐸, ∣∣𝑒∣∣𝐸
∣∣𝑇 ∣∣L (𝐸,𝐹 ) := sup
𝑒∕=0
(viii) 𝐵(𝑎, 𝑟) := {𝑒 ∈ 𝐸 : ∣∣𝑒 − 𝑎∣∣ < 𝑟}, where (𝐸, ∣∣ ⋅ ∣∣) is a Banach space, 𝑎 ∈ 𝐸, and
𝑟 > 0.
169
(ix) 𝒪(Ω, 𝐸) := {𝑓 : Ω → 𝐸 ∣ 𝑓 is holomorphic on Ω} where (𝐸, ∣∣ ⋅ ∣∣) is a Banach space, Ω
an open connected set in ℂ, and 𝑓 : Ω → 𝐸 is said to be holomorphic on Ω provided
𝑓 is differentiable at each point in Ω, i.e., the limit
𝑓𝜔 (𝜔 0 ) := lim (𝜔 − 𝜔 0 )−1 (𝑓 (𝜔) − 𝑓 (𝜔 0 ))
𝜔→𝜔 0
exists for every 𝜔 0 ∈ Ω. The function 𝑓𝜔 : Ω → 𝐸 is called the derivative of 𝑓 .
(x) (ℂ, ∣ ⋅ ∣):= complex numbers with the standard norm ∣𝑎 + 𝑖𝑏∣ :=
√
𝑎2 + 𝑏2 and complex
conjugation 𝑎 + 𝑖𝑏 := 𝑎 − 𝑖𝑏, for 𝑎, 𝑏 ∈ ℝ.
(xi) ∣∣ ⋅ ∣∣ℂ := ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (ℂ)
(xii) 𝐼𝑚 := the identity matrix in 𝑀𝑚 (ℂ).
(xiii) 𝐴 := the matrix formed from the matrix 𝐴 ∈ 𝑀𝑚,𝑛 (ℂ) by replacing each entry by its
complex conjugate.
𝑇
(xiv) 𝐴∗ := 𝐴 , the adjoint of the matrix 𝐴 ∈ 𝑀𝑚,𝑛 (ℂ).
(xv) ⟨𝑢, 𝑣⟩ℂ := 𝑢∗ 𝑣, the standard inner product on ℂ𝑚 .
(xvi) (𝐿𝑝 (𝑎, 𝑏), ∣∣ ⋅ ∣∣𝑝 ) := the Banach space of measurable functions 𝑓 on (𝑎, 𝑏) (modulo
functions which vanish almost everywhere (a.e.)) such that
(∫
∣∣𝑓 ∣∣𝑝 :=
𝑏
) 𝑝1
∣𝑓 (𝑥)∣𝑝 𝑑𝑥
< ∞ (1 ≤ 𝑝 < ∞),
𝑎
∣∣𝑓 ∣∣∞ := ess sup𝑥∈(𝑎,𝑏) ∣𝑓 (𝑥)∣ < ∞ (𝑝 = ∞).
(xvii) ∣∣ ⋅ ∣∣𝑝 := ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐿𝑝 (𝑎,𝑏))
(xviii) 𝐿𝑝loc (ℝ) := the space of all measurable functions 𝑓 on ℝ (modulo functions which vanish
a.e.) such that 𝑓 ∣(𝑎,𝑏) ∈ 𝐿𝑝 (𝑎, 𝑏) for all 𝑎 < 𝑏.
170
(xix) 𝐿𝑝 (𝕋) := {𝑓 ∈ 𝐿𝑝loc (ℝ) : 𝑓 (𝑥 + 𝑑) = 𝑓 (𝑥) for a.e. 𝑥 ∈ ℝ}, the Banach space with norm
∣∣𝑓 ∣∣𝐿𝑝 (𝕋)
( ∫ 𝑑
) 𝑝1
1
:=
∣𝑓 (𝑥)∣𝑝 𝑑𝑥
< ∞ (1 ≤ 𝑝 < ∞),
𝑑 0
∣∣𝑓 ∣∣𝐿∞ (𝕋) := ess sup𝑥∈(0,𝑑) ∣𝑓 (𝑥)∣ < ∞ (𝑝 = ∞).
(xx) ∣∣ ⋅ ∣∣𝐿𝑝 (𝕋) := ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))
(xxi) As convention we identify an element 𝐴 of 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) or 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) with any
one of its representative functions 𝐴(⋅) (and visa versa). And hence any identities,
inequalities, etc. throughout this chapter are to be understood as being true almost
everywhere unless otherwise stated.
(xxii) As convention we identify an element 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))) with the function
𝐴 : (𝑎, 𝑏) × Ω → 𝑀𝑚,𝑛 (ℂ) defined by 𝐴(⋅, 𝜔) := 𝐴(𝜔). Similarly, a function 𝐴 :
(𝑎, 𝑏) × Ω → 𝑀𝑚,𝑛 (ℂ) is said to be in 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))) as a function of frequency
and we write 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))) provided 𝐴(𝜔) := 𝐴(⋅, 𝜔) ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) for
each 𝜔 ∈ Ω and 𝐴(⋅) ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))).
(xxiii) As convention we identify an element 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))) with the function 𝐴 :
ℝ × Ω → 𝑀𝑚,𝑛 (ℂ) defined by 𝐴(⋅, 𝜔) := 𝐴(𝜔). Similarly, a function 𝐴 : ℝ × Ω →
𝑀𝑚,𝑛 (ℂ) is said to be in 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))) as a function of frequency and we write
𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))) provided 𝐴(𝜔) := 𝐴(⋅, 𝜔) ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))) for each 𝜔 ∈ Ω and
𝐴(⋅) ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)))).
(xxiv)
∫
𝑈
𝐴(𝑥)𝑑𝑥 :=
[∫
𝑈
𝑎𝑖𝑗 (𝑥)𝑑𝑥
]𝑚,𝑛
𝑖=1,𝑗=1
, for any measurable set 𝑈 contained in [𝑎, 𝑏] and any
𝑝
𝑝
𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐿 (𝑎, 𝑏)) ∪ 𝑀𝑚,𝑛 (𝐿loc (ℝ)).
(xxv) As convention we set
∫𝑥
𝑐
𝐴(𝑡)𝑑𝑡 :=
∫
[𝑐,𝑥]
𝐴(𝑡)𝑑𝑡 if 𝑥 ≤ 𝑐 and
∫𝑥
𝑐
𝐴(𝑡)𝑑𝑡 := −
𝑐 < 𝑎, whenever 𝑐, 𝑥 ∈ [𝑎, 𝑏] and 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) ∪ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)).
171
∫𝑐
𝑥
𝐴(𝑡)𝑑𝑡 if
(xxvi) 𝑊 1,1 (𝑎, 𝑏) := {𝑓 ∈ 𝐿1 (𝑎, 𝑏) ∣ 𝑓 ′ ∈ 𝐿1 (𝑎, 𝑏)}, the standard Sobolev space with norm
∣∣𝑓 ∣∣1,1 := ∣∣𝑓 ∣∣1 + ∣∣𝑓 ′ ∣∣1
where 𝑓 ′ denotes the weak derivative of 𝑓 .
(xxvii) ∣∣ ⋅ ∣∣1,1 := ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝑊 1,1 (𝑎,𝑏))
1,1
(ℝ) := {𝑓 ∈ 𝐿1loc (ℝ) ∣ 𝑓 ′ ∈ 𝐿1loc (ℝ)}.
(xxviii) 𝑊loc
[ ]𝑚,𝑛
1,1
1,1
(𝑎, 𝑏)) or 𝑀𝑚,𝑛 (𝑊loc
(ℝ)).
(xxix) 𝐴′ := 𝑎′𝑖𝑗 𝑖=1,𝑗=1 , for any 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 in 𝑀𝑚,𝑛 (𝑊
1,1
(xxx) As convention we always identify an element 𝐴 of 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) or 𝑀𝑚,𝑛 (𝑊loc
(ℝ))
with its unique absolutely continuous or unique locally absolutely continuous representative, respectively, (see Lemma 70 below) so that, according to this identification,
∫𝑥
𝐴(𝑥) = 𝐴(𝑢) + 𝑢 𝐴′ (𝑡)𝑑𝑡 for all 𝑢, 𝑥 ∈ [𝑎, 𝑏], if 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)), and for all
1,1
𝑢, 𝑥 ∈ ℝ, if 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)).
(xxxi) As convention we identify an element 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) with the function
𝐴 : [𝑎, 𝑏] × Ω → 𝑀𝑚,𝑛 (ℂ) defined by 𝐴(𝑥, 𝜔) := 𝐴(𝜔)(𝑥) – the value of the function
𝐴 : Ω → 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) at 𝜔 ∈ Ω and the value of the unique absolutely continuous
representative of 𝐴(𝜔) at 𝑥 ∈ [𝑎, 𝑏]. Similarly, a function 𝐴 : 𝐼 × Ω → 𝑀𝑚,𝑛 (ℂ),
where 𝐼 := (𝑎, 𝑏) or [𝑎, 𝑏], is said to be in 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) and we write 𝐴 ∈
𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) provided 𝐴(𝜔) := 𝐴(⋅, 𝜔) ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) for each 𝜔 ∈ Ω
and 𝐴(⋅) ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))).
Background Material
The following is background material for this chapter.
In what follows we will need notation for the derivative of a function in the classical sense. Let
172
𝑓 : 𝐼 → ℂ be a complex-valued function defined on an interval 𝐼 ⊆ ℝ. If 𝑓 is differentiable
(in the classical sense) everywhere on 𝐼 we denote its derivative by 𝑑𝑓 /𝑑𝑥, otherwise we
define the function 𝑑𝑓 /𝑑𝑥 : 𝐼 → ℂ by
⎧

⎨ lim𝑥→𝑡
𝑑𝑓 :=

𝑑𝑥 𝑥=𝑡
⎩
𝑓 (𝑥)−𝑓 (𝑡)
,
𝑥−𝑡
if the limit exists
0,
else.
1,1
Lemma 69 If 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) then 𝐴∣(𝑎,𝑏) ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) and 𝐴∣′(𝑎,𝑏) = 𝐴′ ∣(𝑎,𝑏) for
any bounded interval (𝑎, 𝑏) ⊆ ℝ.
Proof.
1,1
We first prove the statement in the case 𝑚 = 𝑛 = 1. Suppose 𝑓 ∈ 𝑊loc
(ℝ). Then
by the definition of the weak derivative ()′ since 𝑓 ′ ∈ 𝐿1loc (ℝ) we have
∫
∫
′
𝜑(𝑥)𝑓 ′ (𝑥)𝑑𝑥, ∀𝜑 ∈ 𝐶0∞ (ℝ),
𝜑 (𝑥)𝑓 (𝑥)𝑑𝑥 = −
ℝ
ℝ
where 𝐶0∞ (ℝ) is the set of all infinitely differentiable function on ℝ with compact support
and 𝜑′ = 𝑑𝜙/𝑑𝑥 for 𝜑 ∈ 𝐶0∞ (ℝ). This implies that for any (𝑎, 𝑏) open interval and any
𝜙 ∈ 𝐶0∞ (𝑎, 𝑏), the set of infinitely differentiable functions on (𝑎, 𝑏) with compact support, we
can take the zero extension of 𝜙, i.e.,
⎧

⎨ 𝜙(𝑥) if 𝑥 ∈ (𝑎, 𝑏)
𝜙𝑒 (𝑥) :=

⎩ 0
else.
so that 𝜙𝑒 ∈ 𝐶0∞ (ℝ). And hence
∫
𝑏
′
∫
𝜑′𝑒 (𝑥)𝑓 (𝑥)𝑑𝑥
𝜑 (𝑥)𝑓 (𝑥)𝑑𝑥 =
𝑎
ℝ
∫
=−
𝜑𝑒 (𝑥)𝑓 ′ (𝑥)𝑑𝑥
ℝ
∫
=−
𝑏
𝜑(𝑥)𝑓 ′ (𝑥)𝑑𝑥, ∀𝜑 ∈ 𝐶0∞ (𝑎, 𝑏).
𝑎
173
This implies by the definition of the weak derivative that 𝑓 ∣(𝑎,𝑏) ∈ 𝑊 1,1 (𝑎, 𝑏) and 𝑓 ∣′(𝑎,𝑏) =
𝑓 ′ ∣(𝑎,𝑏) . This proves the statement for 𝑚 = 𝑛 = 1. For the case 𝑚, 𝑛 ≥ 1, it follows
𝑚,𝑛
now trivially from the notation 4.5.(xxix) and the fact [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∣(𝑎,𝑏) = [𝑎𝑖𝑗 ∣(𝑎,𝑏) ]𝑖=1,𝑗=1 for
1
1
[𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐿 (𝑎, 𝑏)) ∪ 𝑀𝑚,𝑛 (𝐿loc (ℝ)). This completes the proof.
Definition 26 For any compact interval [𝑎, 𝑏] ⊆ ℝ, denote by 𝐴𝐶[𝑎, 𝑏] the space of absolutely
continuous functions on [𝑎, 𝑏], i.e.,
𝐴𝐶[𝑎, 𝑏] := {𝑓 : [𝑎, 𝑏] → ℂ ∣ 𝑓 is differentiable a.e., 𝑑𝑓 /𝑑𝑥 ∈ 𝐿1 (𝑎, 𝑏), and
∫ 𝑢 𝑑𝑓 𝑓 (𝑢) = 𝑓 (𝑣) +
𝑑𝑡, ∀𝑢, 𝑣 ∈ [𝑎, 𝑏] }.
𝑣 𝑑𝑥 𝑥=𝑡
We denote by 𝐴𝐶loc (ℝ) the space of locally absolutely continuous functions on ℝ, i.e.,
𝐴𝐶loc (ℝ) := {𝑓 : ℝ → ℂ ∣ 𝑓 is differentiable a.e., 𝑑𝑓 /𝑑𝑥 ∈ 𝐿1loc (ℝ), and
∫ 𝑢 𝑑𝑓 𝑓 (𝑢) = 𝑓 (𝑣) +
𝑑𝑡, ∀𝑢, 𝑣 ∈ ℝ }.
𝑣 𝑑𝑥 𝑥=𝑡
For 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐴𝐶[𝑎, 𝑏]) we define
𝑑𝐴
𝑑𝑥
:= [
𝑑𝑎𝑖𝑗 𝑚,𝑛
]
𝑑𝑥 𝑖=1,𝑗=1
∈ 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)). Sim-
ilarly, for 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐴𝐶loc (ℝ)) we define the matrix
𝑑𝐴
𝑑𝑥
:= [
𝑑𝑎𝑖𝑗 𝑚,𝑛
]
𝑑𝑥 𝑖=1,𝑗=1
∈
𝑀𝑚,𝑛 (𝐿1loc (ℝ)).
Lemma 70 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) if and only if 𝐴(⋅) ∈ 𝑀𝑚,𝑛 (𝐴𝐶[𝑎, 𝑏]) for some representative function 𝐴(⋅) of 𝐴. Furthermore, this representative function is unique. Moreover, for
the absolutely continuous representative 𝐴(⋅) of 𝐴 we have 𝐴′ =
∫
𝐴(𝑥) = 𝐴(𝑢) +
𝑑𝐴(⋅)
𝑑𝑥
and
𝑥
𝐴′ (𝑡)𝑑𝑡, ∀𝑢, 𝑥 ∈ [𝑎, 𝑏].
𝑢
1,1
Similarly, 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) if and only if 𝐴(⋅) ∈ 𝑀𝑚,𝑛 (𝐴𝐶loc (ℝ))) for some representative
function 𝐴(⋅) of 𝐴. Furthermore, this representative function is unique. Moreover, for the
174
locally absolutely continuous representative 𝐴(⋅) of 𝐴 we have 𝐴′ =
𝑑𝐴(⋅)
𝑑𝑥
and
𝑥
∫
𝐴′ (𝑡)𝑑𝑡, ∀𝑢, 𝑥 ∈ ℝ.
𝐴(𝑥) = 𝐴(𝑢) +
𝑢
Proof.
We first prove the statement for 𝑚 = 𝑛 = 1. By [45, p. 56, §II.2.1, Proposition
2.1.5], if 𝑓 ∈ 𝑊 1,1 (𝑎, 𝑏) then 𝑓 has a representative function 𝑓 (⋅) : [𝑎, 𝑏] → ℂ such that
∫
𝑓 (𝑥) = 𝑓 (𝑢) +
𝑥
𝑓 ′ (𝑡)𝑑𝑡, ∀𝑢, 𝑥 ∈ [𝑎, 𝑏].
𝑢
This implies 𝑓 (⋅) ∈ 𝐴𝐶[𝑎, 𝑏] with 𝑑𝑓 (⋅)/𝑑𝑥 = 𝑓 ′ . To prove uniqueness we note that if
𝑔 ∈ 𝐴𝐶[𝑎, 𝑏] is another representative function of 𝑓 then since 𝑓 (⋅) is as well we have
𝑔(𝑥) = 𝑓 (𝑥) for a.e. 𝑥 ∈ (𝑎, 𝑏). But then by the fact absolutely continuous functions on [𝑎, 𝑏]
are continuous functions on [𝑎, 𝑏] this implies 𝑔(𝑥) = 𝑓 (𝑥) for every 𝑥 ∈ [𝑎, 𝑏]. This proves
uniqueness. On the other hand, if 𝑓 (⋅) ∈ 𝐴𝐶[𝑎, 𝑏] then by [45, p. 55, §II.2.1, Proposition
2.1.3] we have 𝑓 ∈ 𝑊 1,1 (𝑎, 𝑏).
1,1
Now suppose 𝑓 ∈ 𝑊loc
(ℝ). Then by Lemma 69 we have 𝑓 ∣(𝑎,𝑏) ∈ 𝑊 1,1 (𝑎, 𝑏) for any bounded
open interval (𝑎, 𝑏) and 𝑓 ∣′(𝑎,𝑏) = 𝑓 ′ ∣(𝑎,𝑏) . Hence from what we just proved 𝑓 ∣(𝑎,𝑏) has an
absolutely continuous representative function 𝑓 ∣(𝑎,𝑏) (⋅) : [𝑎, 𝑏] → ℂ such that
∫
𝑥
𝑓 ∣(𝑎,𝑏) (𝑥) = 𝑓 ∣(𝑎,𝑏) (𝑢) +
∫𝑢 𝑥
= 𝑓 ∣(𝑎,𝑏) (𝑢) +
𝑓 ∣′(𝑎,𝑏) (𝑡)𝑑𝑡
∫
= 𝑓 ∣(𝑎,𝑏) (𝑢) +
𝑥
𝑓 ′ ∣(𝑎,𝑏) (𝑡)𝑑𝑡
𝑢
𝑓 ′ (𝑡)𝑑𝑡, ∀𝑢, 𝑥 ∈ [𝑎, 𝑏].
𝑢
Define a function 𝑔(⋅) : ℝ → ℂ by
𝑔(𝑥) := 𝑓 ∣(𝑎,𝑏) (𝑥), for any bounded interval (𝑎, 𝑏) with 𝑥 ∈ (𝑎, 𝑏).
We first show this function is well-defined. Let (𝑎, 𝑏) and (𝑢, 𝑣) be any two bounded open
175
interval such that 𝐼 := (𝑎, 𝑏) ∩ (𝑢, 𝑣) ∕= ∅ and let 𝑓 ∣(𝑎,𝑏) (⋅) ∈ 𝐴𝐶[𝑎, 𝑏] be an absolutely continuous representative function of 𝑓 ∣(𝑎,𝑏) ∈ 𝑊 1,1 (𝑎, 𝑏) and 𝑓 ∣(𝑢,𝑣) (⋅) ∈ 𝐴𝐶[𝑢, 𝑣] be an absolutely
continuous representative function of 𝑓 ∣(𝑢,𝑣) ∈ 𝑊 1,1 (𝑢, 𝑣) . Now fix any representative function 𝑓˜ : ℝ → ℂ of 𝑓 . Then 𝑓 ∣(𝑎,𝑏) (𝑥) = 𝑓˜(𝑥) for a.e. 𝑥 ∈ [𝑎, 𝑏] and 𝑓 ∣(𝑢,𝑣) (𝑥) = 𝑓˜(𝑥) for
a.e. 𝑥 ∈ [𝑢, 𝑣] implying 𝑓 ∣(𝑎,𝑏) (𝑥) = 𝑓˜(𝑥) = 𝑓 ∣(𝑢,𝑣) (𝑥) for a.e. 𝑥 ∈ 𝐼. Hence since 𝑓 ∣(𝑎,𝑏) (⋅)
and 𝑓 ∣(𝑢,𝑣) (⋅) are continuous on 𝐼 this implies 𝑓 ∣(𝑎,𝑏) (𝑥) = 𝑓 ∣(𝑢,𝑣) (𝑥) for all 𝑥 ∈ 𝐼. This
proves 𝑔(⋅) : ℝ → ℂ is a well-defined function. From the definition of 𝑔 it also proves
1,1
𝑔∣(𝑎,𝑏) = 𝑓 ∣(𝑎,𝑏) ∈ 𝑊 1,1 (𝑎, 𝑏) for any bounded open interval (𝑎, 𝑏) and hence 𝑔 = 𝑓 ∈ 𝑊loc
(ℝ).
It also implies that for any 𝑢, 𝑥 ∈ ℝ, if we take any bounded open interval (𝑎, 𝑏) containing
𝑢, 𝑥, then we have
∫
𝑔(𝑥) = 𝑓 ∣(𝑎,𝑏) (𝑥) = 𝑓 ∣(𝑎,𝑏) (𝑢) +
𝑥
′
∫
𝑓 (𝑡)𝑑𝑡 = 𝑔(𝑢) +
𝑢
𝑥
𝑓 ′ (𝑡)𝑑𝑡.
𝑢
Thus implying 𝑔(⋅) ∈ 𝐴𝐶loc (ℝ) with 𝑑𝑔(⋅)/𝑑𝑥 = 𝑓 ′ . Hence we have proven 𝑓 has a locally
absolutely continuous representative 𝑓 (⋅) := 𝑔(⋅) ∈ 𝐴𝐶loc (ℝ) with 𝑑𝑓 (⋅)/𝑑𝑥 = 𝑓 ′ and such
that
∫
𝑓 (𝑥) = 𝑓 (𝑢) +
𝑥
𝑓 ′ (𝑡)𝑑𝑡, ∀𝑢, 𝑥 ∈ ℝ.
𝑢
To prove uniqueness we note that if 𝑓1 , 𝑓2 ∈ 𝐴𝐶loc (ℝ) are two representative functions of 𝑓
then 𝑓1 and 𝑓2 are equal a.e. on ℝ and so since functions in 𝐴𝐶loc (ℝ) are continuous this
implies 𝑓1 and 𝑓2 are equal everywhere on ℝ. This proves uniqueness.
Suppose now that 𝑓 ∈ 𝐴𝐶loc (ℝ). Then we have 𝑓 ∈ 𝐿1loc (ℝ) and 𝑑𝑓 /𝑑𝑥 ∈ 𝐿1loc (ℝ). We need
only show now that 𝑓 has 𝑑𝑓 /𝑑𝑥 as its weak derivative, i.e., 𝑓 ′ = 𝑑𝑓 /𝑑𝑥. To do this we need
to show
∫
∫
𝑑𝑓 𝜑 (𝑡)𝑓 (𝑡)𝑑𝑡 = − 𝜑(𝑡) 𝑑𝑡, ∀𝜑 ∈ 𝐶0∞ (ℝ),
𝑑𝑥 𝑥=𝑡
ℝ
ℝ
′
176
where 𝐶0∞ (ℝ) is the set of all infinitely differentiable function on ℝ with compact support.
Let 𝜑 ∈ 𝐶0∞ (ℝ) and let [𝑎, 𝑏] be any nonempty compact interval such that supp(𝜙) ⊆ [𝑎, 𝑏].
Then by integration by parts we have
∫
∫
′
𝑏
∫
′
𝜑 (𝑡)𝑓 (𝑡)𝑑𝑡 = 𝑓 (𝑏)𝜑(𝑏) − 𝑓 (𝑎)𝜑(𝑎) −
𝑎
∫
∫ 𝑏
𝑑𝑓 𝑑𝑓 𝜑(𝑡) 𝑑𝑡 = − 𝜑(𝑡) 𝑑𝑡.
=−
𝑑𝑥 𝑥=𝑡
𝑑𝑥 𝑥=𝑡
ℝ
𝑎
𝜑 (𝑡)𝑓 (𝑡)𝑑𝑡 =
ℝ
𝑏
𝜑(𝑡)
𝑎
𝑑𝑓 𝑑𝑡
𝑑𝑥 𝑥=𝑡
And therefore we have shown
∫
∫
𝑑𝑓 𝜑 (𝑡)𝑓 (𝑡)𝑑𝑡 = − 𝜑(𝑡) 𝑑𝑡, ∀𝜑 ∈ 𝐶0∞ (ℝ),
𝑑𝑥 𝑥=𝑡
ℝ
ℝ
′
1,1
which proves 𝑓 ′ = 𝑑𝑓 /𝑑𝑥 ∈ 𝐿1loc (ℝ) and hence 𝑓 ∈ 𝑊loc
(ℝ).
This completes the prove in the case 𝑚 = 𝑛 = 1. The case 𝑚, 𝑛 ≥ 1 follows now trivially
from the case 𝑚 = 𝑛 = 1 by the notation 4.5.(xxiv) and 4.5.(xxix). This completes the
proof.
The following lemma, along with Lemma 69 and Lemma 70, completes our characterization
1,1
of spaces 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) and 𝑀𝑚,𝑛 (𝑊loc
(ℝ)).
Lemma 71 Let 𝐴 : ℝ → 𝑀𝑚,𝑛 (ℂ) be a function such that there exists an increasing se∪∞
quence of intervals {(𝑎𝑗 , 𝑏𝑗 )}∞
𝑗=1 with ℝ =
𝑗=1 (𝑎𝑗 , 𝑏𝑗 ) such that for every 𝑗, 𝐴∣(𝑎𝑗 ,𝑏𝑗 ) ∈
1,1
𝑀𝑚,𝑛 (𝑊 1,1 (𝑎𝑗 , 𝑏𝑗 )). Then 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)).
We first prove the statement for 𝑚 = 𝑛 = 1. Let 𝑓 : ℝ → ℂ be a function
∪∞
such that there exists an increasing sequence of intervals {(𝑎𝑗 , 𝑏𝑗 )}∞
𝑗=1 with ℝ =
𝑗=1 (𝑎𝑗 , 𝑏𝑗 )
Proof.
such that for every 𝑗, 𝑓 ∣(𝑎𝑗 ,𝑏𝑗 ) ∈ 𝑊 1,1 (𝑎𝑗 , 𝑏𝑗 ). It follows from this that on every interval
(𝑎𝑗 , 𝑏𝑗 ) the function 𝑓 on this interval is equal a.e. to a function which is measurable and
integrable on (𝑎𝑗 , 𝑏𝑗 ) and this implies the function 𝑓 restricted to (𝑎𝑗 , 𝑏𝑗 ) is a measurable and
177
integrable function. Thus since ℝ =
∪∞
and the restriction of 𝑓 : ℝ → ℂ to (𝑎𝑗 , 𝑏𝑗 )
𝑗=1 (𝑎𝑗 , 𝑏𝑗 )
is measurable and integrable for every 𝑗, this implies 𝑓 is a measurable function which is
locally integrable and so 𝑓 ∈ 𝐿1loc (ℝ).
Now 𝑓 ∣(𝑎𝑗 ,𝑏𝑗 ) ∈ 𝑊 1,1 (𝑎𝑗 , 𝑏𝑗 ) for every 𝑗 and so by Lemma 70 there exists an absolutely
continuous function 𝑔𝑗 ∈ 𝐴𝐶[𝑎𝑗 , 𝑏𝑗 ] such that 𝑓 (𝑥) = 𝑔𝑗 (𝑥) for a.e. 𝑥 ∈ (𝑎𝑗 , 𝑏𝑗 ) for each 𝑗.
But this implies for any fix 𝑗0 we must have 𝑔𝑗 (𝑥) = 𝑓 (𝑥) = 𝑔𝑗0 (𝑥) for a.e. 𝑥 ∈ (𝑎𝑗0 , 𝑏𝑗0 ) for any
fixed 𝑗 ≥ 𝑗0 . But the functions 𝑔𝑗 , 𝑔𝑗0 are continuous on [𝑎𝑗0 , 𝑏𝑗0 ] and so this implies 𝑔𝑗 (𝑥) =
𝑔𝑗0 (𝑥) for all 𝑥 ∈ [𝑎𝑗0 , 𝑏𝑗0 ] and every 𝑗 ≥ 𝑗0 . But this implies 𝑔(𝑥) := lim𝑗→∞ 𝑔𝑗 (𝑥) exists for
every 𝑥 ∈ ℝ and so defines a function 𝑔 : ℝ → ℂ. Moreover, from what we proved we have
𝑔(𝑥) = 𝑔𝑗 (𝑥) for all 𝑥 ∈ [𝑎𝑗0 , 𝑏𝑗0 ] and any 𝑗. Thus for every 𝑗 we have 𝑔∣[𝑎𝑗 ,𝑏𝑗 ] = 𝑔𝑗 ∈ 𝐴𝐶[𝑎𝑗 , 𝑏𝑗 ]
implying 𝑔 is differentiable a.e. on [𝑎𝑗 , 𝑏𝑗 ] with 𝑑𝑔/𝑑𝑥∣[𝑎𝑗 ,𝑏𝑗 ] = 𝑑𝑔𝑗 /𝑑𝑥 = 𝑔𝑗′ ∈ 𝐿1 (𝑎𝑗 , 𝑏𝑗 ). But
∪
1
since ℝ = ∞
𝑗=1 (𝑎𝑗 , 𝑏𝑗 ) this implies 𝑔 is differentiable a.e. on ℝ with 𝑑𝑔/𝑑𝑥 ∈ 𝐿loc (ℝ). Hence
if 𝜑 ∈ 𝐶0∞ (ℝ) then we can find 𝑗 sufficiently large so that supp(𝜑) ⊆ [𝑎𝑗 , 𝑏𝑗 ] and so
∫
∫
′
𝑏𝑗
∫
𝑏𝑗
𝑎𝑗
𝑏𝑗
∫
=
𝑎𝑗
∫
′
𝑏𝑗
𝜑 (𝑡)𝑔𝑗 (𝑡)𝑑𝑡 =
𝜑 (𝑡)𝑔(𝑡)𝑑𝑡 =
𝜑 (𝑡)𝑔(𝑡)𝑑𝑡 =
ℝ
′
𝜑(𝑡)𝑔𝑗′ (𝑡)𝑑𝑡
𝑎𝑗
𝑎𝑗
∫
𝑑𝑔 𝑑𝑔 𝜑(𝑡) 𝑑𝑡.
𝜑(𝑡) 𝑑𝑡 =
𝑑𝑥 𝑥=𝑡
𝑑𝑥 𝑥=𝑡
ℝ
1,1
This implies that 𝑔 ′ = 𝑑𝑔/𝑑𝑥 ∈ 𝐿1loc (ℝ) and therefore 𝑔 ∈ 𝑊loc
(ℝ). But 𝑓 = 𝑔 a.e. on ℝ and
1,1
so 𝑓 ∈ 𝑊loc
(ℝ). This proves the statement for 𝑛 = 𝑚 = 1. The case where 𝑚, 𝑛 ≥ 1 follows
trivially now from the case 𝑛 = 𝑚 = 1. This completes the proof.
Lemma 72 Let (𝑎, 𝑏) ⊆ ℝ be a bounded interval.
Then the matrix multiplication ⋅ :
𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) × 𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝑊 1,1 (𝑎, 𝑏)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈
𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)), 𝐵 ∈ 𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)), is a continuous bilinear map and
(𝐴𝐵)′ = 𝐴′ 𝐵 + 𝐴𝐵 ′ .
178
We first prove the statement for 𝑚 = 𝑛 = 𝑟 = 1. To begin, let (𝑎, 𝑏) ⊆ ℝ
Proof.
be a bounded interval. Then by [45, p. 65, §II.2.3, Proposition 2.3.1], multiplication ⋅ :
𝑊 1,1 (𝑎, 𝑏) × 𝑊 1,1 (𝑎, 𝑏) → 𝑊 1,1 (𝑎, 𝑏), where 𝑓 ⋅ 𝑔 := 𝑓 𝑔 for 𝑓, 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏), is a continuous
bilinear map and
(𝑓 𝑔)′ = 𝑓 ′ 𝑔 + 𝑓 𝑔 ′ .
This proves the statement for 𝑚 = 𝑛 = 𝑟 = 1. The statement for 𝑚, 𝑛, 𝑟 ≥ 1 now follows
from this, Lemma 77, and notation 4.5.(xxix). This completes the proof.
1,1
1,1
1,1
Lemma 73 Matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) × 𝑀𝑛,𝑟 (𝑊loc
(ℝ)) → 𝑀𝑚,𝑟 (𝑊loc
(ℝ)),
1,1
1,1
where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)), 𝐵 ∈ 𝑀𝑛,𝑟 (𝑊loc
(ℝ)), is a well-defined map and
(𝐴𝐵)′ = 𝐴′ 𝐵 + 𝐴𝐵 ′ .
Proof.
We first prove the statement for 𝑚 = 𝑛 = 𝑟 = 1. To begin, let (𝑎, 𝑏) ⊆ ℝ be any
1,1
bounded interval. It follows from Lemma 69 and Lemma 72 that if 𝑓, 𝑔 ∈ 𝑊loc
(ℝ) then we
have (𝑓 𝑔)∣(𝑎,𝑏) = 𝑓 ∣(𝑎,𝑏) 𝑔∣(𝑎,𝑏) ∈ 𝑊 1,1 (𝑎, 𝑏). Now since this is true for every bounded interval
1,1
(𝑎, 𝑏) ⊆ ℝ, this implies by Lemma 71 that 𝑓 𝑔 ∈ 𝑊loc
(ℝ). Moreover, by Lemma 69 and
Lemma 72 we have for every bounded interval (𝑎, 𝑏) ⊆ ℝ,
((𝑓 𝑔)∣(𝑎,𝑏) )′ = (𝑓 ∣(𝑎,𝑏) 𝑔∣(𝑎,𝑏) )′ = (𝑓 ∣(𝑎,𝑏) )′ 𝑔∣(𝑎,𝑏) + 𝑓 ∣(𝑎,𝑏) (𝑔∣(𝑎,𝑏) )′
= 𝑓 ′ ∣(𝑎,𝑏) 𝑔∣(𝑎,𝑏) + 𝑓 ∣(𝑎,𝑏) 𝑔 ′ ∣(𝑎,𝑏) .
It follows from this that 𝑓 ′ 𝑔 + 𝑓 𝑔 ′ ∈ 𝐿1loc (ℝ) and for any 𝜑 ∈ 𝐶0∞ (ℝ) we can find a bounded
interval (𝑎, 𝑏) ⊆ ℝ such that supp(𝜑) ⊆ (𝑎, 𝑏) implying
∫
′
∫
𝜑 (𝑥)(𝑓 𝑔)(𝑥)𝑑𝑥 =
ℝ
𝑏
𝜑′ (𝑥)(𝑓 𝑔)∣(𝑎,𝑏) (𝑥)𝑑𝑥
𝑎
179
∫
𝑏
𝜑(𝑥)(𝑓 ′ ∣(𝑎,𝑏) (𝑥)𝑔∣(𝑎,𝑏) (𝑥) + 𝑓 ∣(𝑎,𝑏) (𝑥)𝑔 ′ ∣(𝑎,𝑏) (𝑥))𝑑𝑥
=−
𝑎
∫
𝑏
𝜑(𝑥)(𝑓 ′ (𝑥)𝑔(𝑥) + 𝑓 (𝑥)𝑔 ′ (𝑥))𝑑𝑥
=−
𝑎
∫
=−
𝑏
𝜑(𝑥)(𝑓 ′ 𝑔 + 𝑓 𝑔 ′ )(𝑥)𝑑𝑥.
𝑎
And because this is true for every 𝜑 ∈ 𝐶0∞ (ℝ) this means the weak derivative of 𝑓 𝑔 is
𝑓 ′ 𝑔 + 𝑓 𝑔 ′ , i.e.,
(𝑓 𝑔)′ = 𝑓 ′ 𝑔 + 𝑓 𝑔 ′ .
This proves the statement for 𝑚 = 𝑛 = 𝑟 = 1. The statement for 𝑚, 𝑛, 𝑟 ≥ 1 now follows
trivially from this one by notation 4.5.(xxix). This completes the proof.
1,1
Lemma 74 Let 𝐸 = 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) or 𝑀𝑚,𝑛 (𝑊loc
(ℝ)), where 1 ≤ 𝑝 ≤ ∞. The translation
operator ℒ𝑑 : 𝐸 → 𝐸, where (ℒ𝑑 𝐴)(⋅) = 𝐴(⋅+𝑑) for 𝐴 ∈ 𝐸, is a well-defined map. Moreover,
1,1
for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) we have
(ℒ𝑑 𝐴)′ = ℒ𝑑 𝐴′ .
Proof. We first prove the statement for 𝑛 = 𝑚 = 1. Let 𝑓 ∈ 𝐿𝑝loc (ℝ). Then (ℒ𝑑 𝑓 )(⋅) : ℝ →
ℂ given by (ℒ𝑑 𝑓 )(𝑥) := 𝑓 (𝑥 + 𝑑), 𝑥 ∈ ℝ is a measurable function. And for any bounded
interval (𝑎, 𝑏) ⊆ ℝ, since ∣∣(ℒ𝑑 𝑓 )∣(𝑎,𝑏) ∣∣𝑝 = ∣∣𝑓 ∣(𝑎+𝑑,𝑏+𝑑) ∣∣𝑝 , we have (ℒ𝑑 𝑓 )∣(𝑎,𝑏) ∈ 𝐿𝑝 (𝑎, 𝑏). This
1,1
(ℝ) then 𝑓 ′ ∈ 𝐿1loc (ℝ) and from what we have just
proves ℒ𝑑 𝑓 ∈ 𝐿𝑝loc (ℝ). Now if 𝑓 ∈ 𝑊loc
1,1
proved, ℒ𝑑 𝑓 ′ ∈ 𝐿1loc (ℝ). We will now show ℒ𝑑 𝑓 ∈ 𝑊loc
(ℝ) with (ℒ𝑑 𝑓 )′ = ℒ𝑑 𝑓 ′ . To start, we
have (𝜑(⋅ − 𝑑))′ = 𝜑′ (⋅ − 𝑑) ∈ 𝐶0∞ (ℝ) for all 𝜑 ∈ 𝐶0∞ (ℝ) since ()′ = 𝑑/𝑑𝑥 is just the derivative
in the classical sense on smooth functions. From this and the fact the Lebesgue integral on
180
ℝ is translation invariant it follows that
∫
∫
′
𝜙′ (𝑥 − 𝑑)(ℒ𝑑 𝑓 )(𝑥 − 𝑑)𝑑𝑥
∫ℝ
∫
′
=
𝜙 (𝑥 − 𝑑)𝑓 (𝑥)𝑑𝑥 =
𝜙′ (𝑥 − 𝑑)𝑓 (𝑥)𝑑𝑥
ℝ
∫ℝ
∫
′
= (𝜙(⋅ − 𝑑)) (𝑥)𝑓 (𝑥)𝑑𝑥 =
𝜙(𝑥 − 𝑑)𝑓 ′ (𝑥)𝑑𝑥
ℝ
∫ℝ
∫
=
𝜙(𝑥)𝑓 ′ (𝑥 + 𝑑)𝑑𝑥 =
𝜙(𝑥)(ℒ𝑑 𝑓 ′ )(𝑥)𝑑𝑥, 𝜑 ∈ 𝐶0∞ (ℝ).
𝜙 (𝑥)(ℒ𝑑 𝑓 )(𝑥)𝑑𝑥 =
ℝ
ℝ
ℝ
But this implies (ℒ𝑑 𝑓 )′ = ℒ𝑑 𝑓 ′ and, since we have already shown ℒ𝑑 𝑓 ′ ∈ 𝐿1loc (ℝ), this proves
1,1
ℒ𝑑 𝑓 ∈ 𝑊loc
(ℝ). This proves the statement for 𝑚 = 𝑛 = 1. The statement for 𝑚, 𝑛, 𝑟 ≥ 1
now follows trivially from this one by notation 4.5.(xxix). This completes the proof.
Lemma 75 For 1 ≤ 𝑝 ≤ ∞ and 𝑚, 𝑛 ∈ ℕ, the adjoint extends via 𝐴∗ (⋅) := 𝐴(⋅)∗ for
𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)), to a well-defined map
∗
: 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) → 𝑀𝑛,𝑚 (𝐿𝑝loc (ℝ)). Moreover, if
1,1
1,1
𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) then 𝐴∗ ∈ 𝑀𝑛,𝑚 (𝑊loc
(ℝ)) and
(𝐴∗ )′ = (𝐴′ )∗ .
Proof.
In the case 𝑚 = 𝑛 = 1, it just follows from the facts that 𝑓 ∈ 𝐿1loc (ℝ) if and only
1,1
1,1
(ℝ) if and only if Re(𝑓 ), Im(𝑓 ) ∈ 𝑊loc
(ℝ), Re(𝑓 ′ ) =
if Re(𝑓 ), Im(𝑓 ) ∈ 𝐿1loc (ℝ) and 𝑓 ∈ 𝑊loc
′
(Re(𝑓 ))′ , Im(𝑓 ′ ) = (Im(𝑓 ))′ , and 𝑓 ′ = 𝑓 . In the case 𝑚, 𝑛 ≥ 1, if 𝐴 := {𝑎𝑖𝑗 }𝑚,𝑛
𝑖=1,𝑗=1 ∈
1
𝑀𝑚,𝑛 (𝐿1loc (ℝ)) then, by the case 𝑚 = 𝑛 = 1, we have 𝐴∗ = {𝑎∗𝑖𝑗 }𝑛,𝑚
𝑗=1,𝑖=1 ∈ 𝑀𝑛,𝑚 (𝐿loc (ℝ)) and
1,1
1,1
if 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) then 𝐴∗ ∈ 𝑀𝑛,𝑚 (𝑊loc
(ℝ)) with
{ }𝑛,𝑚
𝑛,𝑚
(𝐴∗ )′ = {𝑎𝑖𝑗 ′ }𝑗=1,𝑖=1 = 𝑎𝑖𝑗 ′ 𝑗=1,𝑖=1 = (𝐴′ )∗ .
This completes the proof.
181
Lemma 76 If 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)) and 1 ≤ 𝑝 ≤ ∞ then 𝐴∗ ∈ 𝑀𝑛,𝑚 (𝐿𝑝 (𝕋)) with ∣∣𝐴∗ ∣∣𝐿𝑝 (𝕋) =
∣∣𝐴∣∣𝐿𝑝 (𝕋) .
Proof. By Lemma 75 the map ∗ : 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) → 𝑀𝑛,𝑚 (𝐿𝑝loc (ℝ)) extending the adjoint by
𝐴∗ (⋅) = 𝐴(⋅)∗ for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) is well-defined. By Lemma 74, the translation operator
ℒ𝑑 : 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) → 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) given by (L𝑑 𝐴)(⋅) = 𝐴(⋅ + 𝑑) for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ))
is well-defined. Thus for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)), since 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)) ⊆ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)), then
𝐴∗ , ℒ𝑑 𝐴∗ , (ℒ𝑑 𝐴)∗ are all well-defined and belong to 𝑀𝑛,𝑚 (𝐿𝑝loc (ℝ)). Also, 𝐵 ∈ 𝑀𝑛,𝑚 (𝐿𝑝 (𝕋))
if and only if 𝑀𝑛,𝑚 (𝐿𝑝loc (ℝ)) and ℒ𝑑 𝐵 = 𝐵. Thus since 𝐴∗ ∈ 𝑀𝑛,𝑚 (𝐿𝑝loc (ℝ)) and
ℒ𝑑 𝐴∗ = 𝐴∗ (⋅ + 𝑑) = 𝐴(⋅ + 𝑑)∗ = (ℒ𝑑 𝐴)∗ = 𝐴∗
then 𝐴∗ ∈ 𝑀𝑛,𝑚 (𝐿𝑝 (𝕋)).
Now we prove ∣∣𝐴∗ ∣∣𝐿𝑝 (𝕋) = ∣∣𝐴∣∣𝐿𝑝 (𝕋) for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)). We start with the case 𝑚 = 𝑛 =
1. If 𝑓 ∈ 𝐿𝑝 (𝕋) then as we just proved 𝑓 ∈ 𝐿𝑝 (𝕋) and
∣∣𝑓 ∣∣𝐿𝑝 (𝕋)
) 𝑝1
( ∫ 𝑑
) 𝑝1 ( ∫ 𝑑
1
1
𝑝
𝑝
∣𝑓 (𝑥)∣ 𝑑𝑥
=
∣𝑓 (𝑥)∣ 𝑑𝑥
= ∣∣𝑓 ∣∣𝐿𝑝 (𝕋) (1 ≤ 𝑝 < ∞),
=
𝑑 0
𝑑 0
∣∣𝑓 ∣∣𝐿∞ (𝕋) = ∣∣𝑓 ∣∣∞ = ∣∣𝑓 ∣∣∞ = ∣∣𝑓 ∣∣𝐿∞ (𝕋) (𝑝 = ∞).
This proves the statement for 𝑚 = 𝑛 = 1. Consider now the case 𝑚, 𝑛 ≥ 1. If 𝐴 :=
𝑛,𝑚
𝑝
𝑝
∗
{𝑎𝑖𝑗 }𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐿 (𝕋)) then 𝐴 = {𝑎𝑖𝑗 }𝑗=1,𝑖=1 ∈ 𝑀𝑛,𝑚 (𝐿loc (ℝ)) and so
(
∣∣𝐴∗ ∣∣𝐿𝑝 (𝕋) =
𝑚
𝑛 ∑
∑
) 21
∣∣𝑎𝑖𝑗 ∣∣2𝐿𝑝 (𝕋)
(
=
𝑗=1 𝑖=1
𝑛
𝑚 ∑
∑
) 12
∣∣𝑎𝑖𝑗 ∣∣2𝐿𝑝 (𝕋)
= ∣∣𝐴∣∣𝐿𝑝 (𝕋) .
𝑖=1 𝑗=1
And hence the statement is true for 𝑚, 𝑛 ≥ 1. This completes the proof.
182
General Statement
In this section we give the statements we need for general Banach spaces. We split this
section into statements that do not involve the frequency 𝜔 and those that do.
Parameter Independent
This is the collection of statements which do not involve the frequency parameter.
Definition 27 Let 𝐸1 , 𝐸2 , and 𝐸3 be Banach spaces. A function ⋅ : 𝐸1 × 𝐸2 → 𝐸3 is called
a continuous bilinear map provided it satisfies the following three properties:
(i) For any 𝑎1 ∈ 𝐸1 , the map 𝑎2 7→ 𝑎1 ⋅ 𝑎2 is a linear map from 𝐸2 to 𝐸3 .
(ii) For any 𝑎2 ∈ 𝐸2 , the map 𝑎1 7→ 𝑎1 ⋅ 𝑎2 is a linear map from 𝐸1 to 𝐸3 .
(iii) There exists a constant 𝐶 > 0 such that ∣∣𝑎1 ⋅ 𝑎2 ∣∣𝐸3 ≤ 𝐶∣∣𝑎1 ∣∣𝐸1 ∣∣𝑎2 ∣∣𝐸2 for all 𝑎1 ∈ 𝐸1 ,
𝑎2 ∈ 𝐸2 .
Lemma 77 Let 𝐸1 , 𝐸2 , and 𝐸3 be Banach spaces and ⋅ : 𝐸1 × 𝐸2 → 𝐸3 be a bilinear map.
Define the map ⋅ : 𝑀𝑚,𝑛 (𝐸1 ) × 𝑀𝑛,𝑟 (𝐸2 ) → 𝑀𝑚,𝑟 (𝐸3 ) defined by
[
𝐴 ⋅ 𝐵 :=
𝑛
∑
]𝑚,𝑟
𝑎𝑖𝑘 ⋅ 𝑏𝑘𝑗
𝑘=1
𝑖=1,𝑗=1
𝑛,𝑟
for any 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐸1 ), 𝐵 := [𝑏𝑖𝑗 ]𝑖=1,𝑗=1 ∈ 𝑀𝑛,𝑟 (𝐸2 ). Then ⋅ : 𝑀𝑚,𝑛 (𝐸1 ) ×
𝑀𝑛,𝑟 (𝐸2 ) → 𝑀𝑚,𝑟 (𝐸3 ) is a continuous bilinear map.
Proof. The map is obviously linear in both of its arguments and so properties (i) and (ii)
are satisfied of Definition 27. Now choose any 𝐶 > 0 such that ∣∣𝑎1 ⋅ 𝑎2 ∣∣𝐸3 ≤ 𝐶∣∣𝑎1 ∣∣𝐸1 ∣∣𝑎2 ∣∣𝐸2
183
for all 𝑎1 ∈ 𝐸1 , 𝑎2 ∈ 𝐸2 . Then
⎛
∣∣𝐴 ⋅ 𝐵∣∣𝐸3 = ⎝
𝑖=1 𝑗=1
(
≤
𝑚 ∑
𝑟 ∑
𝑛
∑
𝑚 ∑
𝑟 ∑
𝑛
∑
𝑘=1
2 ⎞ 12 (
) 12
𝑚 ∑
𝑟 ∑
𝑛
∑
𝑎𝑖𝑘 ⋅ 𝑏𝑘𝑗 ⎠ ≤
∣∣𝑎𝑖𝑘 ⋅ 𝑏𝑘𝑗 ∣∣2𝐸3
𝑖=1 𝑗=1 𝑘=1
𝐸3
) 21
𝐶 2 ∣∣𝑎𝑖𝑘 ∣∣2𝐸1 ∣∣𝑏𝑘𝑗 ∣∣2𝐸3
𝑖=1 𝑗=1 𝑘=1
≤𝐶
(( 𝑚 𝑛
∑∑
𝑖=1 𝑘=1
∣∣𝑎𝑖𝑘 ∣∣2𝐸1
)( 𝑛 𝑟
∑∑
)) 21
∣∣𝑏𝑘𝑗 ∣∣2𝐸3
𝑘=1 𝑗=1
= 𝐶∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸1 ) ∣∣𝐵∣∣𝑀𝑛,𝑟 (𝐸2 ) .
This proves property (iii) of Definition 27 and thus we have proved the lemma.
Parameter Dependent
This is the collection of statements which involve the frequency parameter.
Lemma 78 Let 𝐸1 , 𝐸2 , and 𝐸3 be Banach spaces and ⋅ : 𝐸1 × 𝐸2 → 𝐸3 be a bilinear map.
If 𝑢 ∈ 𝒪(Ω, 𝐸1 ) and 𝑣 ∈ 𝒪(Ω, 𝐸2 ) then 𝑢 ⋅ 𝑣 ∈ 𝒪(Ω, 𝐸3 ) and
(𝑢 ⋅ 𝑣)𝜔 = 𝑢𝜔 ⋅ 𝑣 + 𝑢 ⋅ 𝑣𝜔 .
Proof. First we choose any 𝐶 > 0 such that ∣∣𝑎1 ⋅ 𝑎2 ∣∣𝐸3 ≤ 𝐶∣∣𝑎1 ∣∣𝐸1 ∣∣𝑎2 ∣∣𝐸2 for all 𝑎1 ∈ 𝐸1 ,
𝑎2 ∈ 𝐸2 . Now let 𝑢 ∈ 𝒪(Ω, 𝐸1 ) and 𝑣 ∈ 𝒪(Ω, 𝐸2 ). Let 𝜔 0 ∈ Ω. Then
(𝜔 − 𝜔 0 )−1 (𝑢(𝜔) ⋅ 𝑣(𝜔) − 𝑢(𝜔 0 ) ⋅ 𝑣(𝜔 0 )) − (𝑢𝜔 (𝜔 0 ) ⋅ 𝑣(𝜔 0 ) + 𝑢(𝜔 0 ) ⋅ 𝑣𝜔 (𝜔 0 ))
𝐸3
= [(𝜔 − 𝜔 0 )−1 (𝑢(𝜔) − 𝑢𝜔 (𝜔 0 ))] ⋅ 𝑣(𝜔) + 𝑢(𝜔 0 ) ⋅ [(𝜔 − 𝜔 0 )−1 (𝑣(𝜔) − 𝑣𝜔 (𝜔 0 ))]𝐸3
≤ 𝐶∣∣(𝜔 − 𝜔 0 )−1 (𝑢(𝜔) − 𝑢𝜔 (𝜔 0 ))∣∣𝐸1 ∣∣(𝜔 − 𝜔 0 )−1 (𝑣(𝜔) − 𝑣𝜔 (𝜔 0 ))∣∣𝐸2
184
Thus as 𝜔 → 𝜔 0 the RHS of that inequality goes to zero implying the LHS of that inequality
does as well which implies the function (𝑢 ⋅ 𝑣)(𝜔) := 𝑢(𝜔) ⋅ 𝑣(𝜔) is differentiable at 𝜔 0 and
(𝑢 ⋅ 𝑣)𝜔 (𝜔 0 ) = (𝑢𝜔 ⋅ 𝑣)(𝜔 0 ) + (𝑢 ⋅ 𝑣𝜔 )(𝜔 0 ).
Therefore 𝑢 ⋅ 𝑣 ∈ 𝒪(Ω, 𝐸3 ) and (𝑢 ⋅ 𝑣)𝜔 = 𝑢𝜔 ⋅ 𝑣 + 𝑢 ⋅ 𝑣𝜔 .
Lemma 79 Let 𝐸1 and 𝐸2 be Banach spaces. If ℒ ∈ 𝒪(Ω, L (𝐸1 , 𝐸2 )) and 𝑢 ∈ 𝒪(Ω, 𝐸1 )
then ℒ𝑢 ∈ 𝒪(Ω, 𝐸2 ) and
(ℒ𝑢)𝜔 = ℒ𝜔 𝑢 + ℒ𝑢𝜔 .
Proof. From the Banach spaces L (𝐸1 , 𝐸2 ) and 𝐸1 , we define the map ⋅ : L (𝐸1 , 𝐸2 )×𝐸1 →
𝐸2 by ℒ ⋅ 𝑢 := ℒ𝑢. Then it is a continuous bilinear map with ∣∣ℒ ⋅ 𝑢∣∣𝐸2 ≤ ∣∣ℒ∣∣L (𝐸1 ,𝐸2 ) ∣∣𝑢∣∣𝐸1 .
Thus it follows from Lemma 78 above that if ℒ ∈ 𝒪(Ω, L (𝐸1 , 𝐸2 )) and 𝑢 ∈ 𝒪(Ω, 𝐸1 ) then
ℒ𝑢 ∈ 𝒪(Ω, 𝐸2 ) and (ℒ𝑢)𝜔 = ℒ𝜔 𝑢 + ℒ𝑢𝜔 . This completes the proof.
Lemma 80 Let 𝐸 be a Banach space. Then 𝒪(Ω, 𝑀𝑚,𝑛 (𝐸)) = 𝑀𝑚,𝑛 (𝒪(Ω, 𝐸)) and for any
𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐸)),
𝐴𝜔 = [(𝑎𝑖𝑗 )𝜔 ]𝑚,𝑛
𝑖=1,𝑗=1 .
Proof. Let (𝐸, ∣∣⋅∣∣) be a Banach space and (𝑀𝑚,𝑛 (𝐸), ∣∣⋅∣∣𝑀𝑚,𝑛 (𝐸) ) the Banach space whose
norm is defined in (4.81). To begin, we know that any two norms on a finite dimensional
vector space over ℂ are equivalent so that there exists constants 𝐶1 , 𝐶2 > 0 such that
𝐶1 ∣∣𝐵∣∣𝑀𝑚,𝑛 (ℂ),∞ ≤ ∣∣𝐵∣∣𝑀𝑚,𝑛 (ℂ),2 ≤ 𝐶2 ∣∣𝐵∣∣𝑀𝑚,𝑛 (ℂ),∞ , ∀𝐵 ∈ 𝑀𝑚,𝑛 (ℂ),
185
where ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (ℂ),∞ , ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (ℂ),2 : 𝑀𝑚,𝑛 (ℂ) → [0, ∞) are the norms defined by
(
∣∣𝐵∣∣𝑀𝑚,𝑛 (ℂ),∞ :=
max
1≤𝑖≤𝑚,1≤𝑗≤𝑛
∣𝑏𝑖𝑗 ∣, ∣∣𝐵∣∣𝑀𝑚,𝑛 (ℂ),2 :=
𝑚 ∑
𝑛
∑
) 21
∣𝑏𝑖𝑗 ∣2
,
𝑖=1 𝑖=1
for any 𝐵 := [𝑏𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (ℂ).
Now we define a function ∣ ⋅ ∣ : 𝑀𝑚,𝑛 (𝐸) → [0, ∞) by
∣𝐴∣ := [∣∣𝑎𝑖𝑗 ∣∣]𝑚,𝑛
∀𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ,
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐸).
We define two norms ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ , ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),2 : 𝑀𝑚,𝑛 (𝐸) → [0, ∞) by
∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸),∞ := ∣𝐴∣𝑀𝑚,𝑛 (ℂ),∞ , ∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸),2 := ∣𝐴∣𝑀𝑚,𝑛 (ℂ),2
for any 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐸). But since ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),2 = ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) it follows that
𝐶1 ∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸),∞ ≤ ∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸) ≤ 𝐶2 ∣∣𝐴∣∣𝑀𝑚,𝑛 (𝐸),∞ , ∀𝐴 ∈ 𝑀𝑚,𝑛 (𝐸).
From this it follows that the identity map 𝜄 : (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) ) → (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅
∣∣𝑀𝑚,𝑛 (𝐸),∞ ) belongs to L ((𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) ), (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ )) and has a continuous inverse. These facts and Lemma 79 imply that 𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) )) =
𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ )) and (𝜄𝐴)𝜔 = 𝜄𝐴𝜔 = 𝐴𝜔 for all 𝐴 ∈ 𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅
∣∣𝑀𝑚,𝑛 (𝐸) )).
Now it follows from the definition of the norm ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ that the statement of this
lemma is true for 𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ )) and hence 𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) )) =
𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸),∞ )) = 𝑀𝑚,𝑛 (𝒪(Ω, 𝐸)) with
[(𝑎𝑖𝑗 )𝜔 ]𝑚,𝑛
𝑖=1,𝑗=1 = (𝜄𝐴)𝜔 = 𝜄𝐴𝜔 = 𝐴𝜔
186
for every 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝒪(Ω, (𝑀𝑚,𝑛 (𝐸), ∣∣ ⋅ ∣∣𝑀𝑚,𝑛 (𝐸) )). This completes the proof of the
lemma.
Lemma 81 If 𝑓 ∈ 𝒪(Ω, 𝐸) then 𝑓𝜔 ∈ 𝒪(Ω, 𝐸).
Proof.
This statement is proven in [19, pp. 21–22, Theorem 1.8.4 & Theorem 1.8.5].
Specific Statements
In the second section we give the statements we need for the specific Banach spaces used in
this chapter. We split this section into statements that do not involve the frequency 𝜔 and
those that do.
Parameter Independent
This is the collection of statements which do not involve the frequency parameter.
Lemma 82 If 𝑝, 𝑞, 𝑠 ∈ ℝ with 1 ≤ 𝑝, 𝑞, 𝑠 ≤ ∞ and
1
𝑝
+
1
𝑞
=
1
𝑠
then matrix multiplication ⋅ :
𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))×𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝐿𝑠 (𝑎, 𝑏)), where 𝐴⋅𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)),
𝐵 ∈ 𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)), is a continuous bilinear map.
Proof.
Multiplication ⋅ : 𝐿𝑝 (𝑎, 𝑏) × 𝐿𝑞 (𝑎, 𝑏) → 𝐿𝑠 (𝑎, 𝑏), where 𝑓 ⋅ 𝑔 := 𝑓 𝑔 for 𝑓 ∈ 𝐿𝑝 (𝑎, 𝑏),
𝑔 ∈ 𝐿𝑠 (𝑎, 𝑏), is a bilinear map since it follows from Holder’s inequality that
∣∣𝑓 𝑔∣∣𝑠 ≤ ∣∣𝑓 ∣∣𝑝 ∣∣𝑔∣∣𝑞 .
187
Thus by Lemma 77 it follows that matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))×𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)) →
𝑀𝑚,𝑟 (𝐿𝑠 (𝑎, 𝑏)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) and 𝐵 ∈ 𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)), is also
continuous bilinear map. This completes the proof.
Lemma 83 Let 1 ≤ 𝑝 ≤ ∞. Then the map of matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) ×
𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝐿𝑝 (𝑎, 𝑏)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) and 𝐵 ∈
𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)), is a continuous bilinear map.
Proof.
Let (𝐶[𝑎, 𝑏], ∣∣ ⋅ ∣∣∞ ) denote the Banach space of all continuous functions on [𝑎, 𝑏]
with norm ∣∣𝑔∣∣∞ := sup𝑥∈[𝑎,𝑏] ∣𝑔(𝑥)∣. By our convention, if 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏) then 𝑔 is absolutely
continuous on [𝑎, 𝑏] and hence 𝑔 ∈ 𝐶[𝑎, 𝑏]. Thus we have 𝑊 1,1 (𝑎, 𝑏) ⊆ 𝐶[𝑎, 𝑏]. Now by [45,
p. 57, §II.2.1, Proposition 2.1.7] we have for the inclusion map 𝜄 : 𝑊 1,1 (𝑎, 𝑏) → 𝐶[𝑎, 𝑏] given
by 𝜄𝑔 = 𝑔 that 𝜄 ∈ L (𝑊 1,1 (𝑎, 𝑏), 𝐶[𝑎, 𝑏]). This implies there exists a 𝐶 > 0 such that
sup ∣𝑔(𝑥)∣ = ∣∣𝑔∣∣∞ ≤ 𝐶∣∣𝑔∣∣1,1 , for every 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏).
𝑥∈[𝑎,𝑏]
It follows from this that for all 𝑓 ∈ 𝐿𝑝 (𝑎, 𝑏), 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏) we have
∣∣𝑓 𝑔∣∣𝑝 ≤ ∣∣𝑓 ∣∣𝑝 ∣∣𝑔∣∣∞ ≤ 𝐶∣∣𝑓 ∣∣𝑝 ∣∣𝑔∣∣1,1 .
This implies that multiplication ⋅ : 𝐿𝑝 (𝑎, 𝑏) × 𝑊 1,1 (𝑎, 𝑏) → 𝐿𝑝 (𝑎, 𝑏), where 𝑓 ⋅ 𝑔 := 𝑓 𝑔 for
𝑓 ∈ 𝐿𝑝 (𝑎, 𝑏), 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏), is a continuous bilinear map. Therefore it follows by Lemma 77
that the matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) × 𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝐿𝑝 (𝑎, 𝑏)), where
𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)), 𝐵 ∈ 𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)), is also a continuous bilinear map.
This completes the proof.
Lemma 84 Let 1 ≤ 𝑝 ≤ ∞. Then the map of matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) ×
1,1
𝑀𝑛,𝑟 (𝑊loc
(ℝ)) → 𝑀𝑚,𝑟 (𝐿𝑝loc (ℝ)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) and 𝐵 ∈
1,1
𝑀𝑛,𝑟 (𝑊loc
(ℝ)), is a well-defined map.
188
Proof.
1,1
(ℝ)). Then, for any bounded interval
Let 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝loc (ℝ)) and 𝐵 ∈ 𝑀𝑛,𝑟 (𝑊loc
(𝑎, 𝑏) ⊆ ℝ by Lemma 83, (𝐴𝐵)∣(𝑎,𝑏) = 𝐴∣(𝑎,𝑏) 𝐵∣(𝑎,𝑏) ∈ 𝑀𝑚,𝑟 (𝐿𝑝 (𝑎, 𝑏)). But this implies
𝐴𝐵 ∈ 𝑀𝑚,𝑟 (𝐿𝑝loc (ℝ)). This completes the proof.
Lemma 85 For each 𝑥 ∈ [𝑎, 𝑏], the evaluation map at 𝑥 defined by 𝛿 𝑥 : 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) →
𝑀𝑚,𝑛 (ℂ) where 𝛿 𝑥 𝐴 := 𝐴(𝑥) for 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)). Then
𝛿 𝑥 ∈ L (𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)), 𝑀𝑚,𝑛 (ℂ)).
(4.82)
Proof. It follows from the proof of Lemma 83 that there exists a constant 𝐶 > 0 such that
∣𝛿 𝑥 𝑔∣ ≤ 𝐶∣∣𝑔∣∣1,1 , for every 𝑔 ∈ 𝑊 1,1 (𝑎, 𝑏).
Now the evaluation map is linear so we need only prove its bounded. If 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈
𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) we have
(
∣∣𝛿 𝑥 𝐴∣∣ℂ = ∣∣𝐴(𝑥)∣∣ℂ =
𝑚 ∑
𝑛
∑
) 21
∣𝑎𝑖𝑗 (𝑥)∣2
1
1
≤ 𝑚 2 𝑛 2 max ∣𝑎𝑖𝑗 (𝑥)∣
1≤𝑖≤𝑚
1≤𝑗≤𝑛
𝑖=1 𝑗=1
1
2
1
2
1
2
≤ 𝐶𝑚 𝑛 max ∣∣𝑎𝑖𝑗 ∣∣1,1 ≤ 𝐶𝑚 𝑛
1≤𝑖≤𝑚
1≤𝑗≤𝑛
1
2
( 𝑚 𝑛
∑∑
) 12
∣∣𝑎𝑖𝑗 ∣∣21,1
1
1
= 𝐶𝑚 2 𝑛 2 ∣∣𝐴∣∣1,1 .
𝑖=1 𝑗=1
1
1
This implies that ∣∣𝛿 𝑥 ∣∣L (𝑀𝑚,𝑛 (𝑊 1,1 (𝑎,𝑏)),𝑀𝑚,𝑛 (ℂ)) ≤ 𝐶𝑚 2 𝑛 2 and hence it follows that 𝛿 𝑥 ∈
L (𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)), ℂ). This completes the proof.
Lemma 86 Let 𝑎, 𝑏 ∈ ℝ with 𝑎 < 𝑏 and let 𝑢 ∈ [𝑎, 𝑏]. We define the integral map by
∫
(ℐ𝐴)(𝑥) :=
𝑥
𝐴(𝑡)𝑑𝑡,
𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)),
𝑢
189
𝑥 ∈ (𝑎, 𝑏).
Then ℐ ∈ L (𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)), 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) and
(ℐ𝐴)′ = 𝐴,
for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)).
Proof. By [45, p. 56, §II.2.1, Proposition 2.1.5] it follows that the map ℐ : 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)) →
𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) is a well-defined map which is linear and (ℐ𝐴)′ = 𝐴. Thus we need only
prove that it is bounded to complete the proof. Now if 𝑓 ∈ 𝐿1 (𝑎, 𝑏) then it follows from the
definition of the norm ∣∣ ⋅ ∣∣1,1 on 𝑊 1,1 (𝑎, 𝑏) that
∣∣ℐ𝑓 ∣∣1,1 = ∣∣(ℐ𝑓 )′ ∣∣1 + ∣∣ℐ𝑓 ∣∣1 = ∣∣𝑓 ∣∣1 + ∣∣ℐ𝑓 ∣∣1 ≤ (𝑏 − 𝑎 + 1)∣∣𝑓 ∣∣1 .
1
Hence it follows that if 𝐴 := [𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐿 (𝑎, 𝑏)) then
(
∣∣ℐ𝐴∣∣1,1 =
𝑚 ∑
𝑛
∑
) 21
∣∣ℐ𝑎𝑖𝑗 ∣∣21,1
1
1
≤ 𝑚 2 𝑛 2 max ∣∣ℐ𝑎𝑖𝑗 ∣∣1,1
1≤𝑖≤𝑚
1≤𝑗≤𝑛
𝑖=1 𝑗=1
1
2
1
2
1
2
≤ (𝑏 − 𝑎 + 1)𝑚 𝑛 max ∣∣𝑎𝑖𝑗 ∣∣1 ≤ (𝑏 − 𝑎 + 1)𝑚 𝑛
1
2
( 𝑚 𝑛
∑∑
1≤𝑖≤𝑚
1≤𝑗≤𝑛
1
) 12
∣∣𝑎𝑖𝑗 ∣∣21
𝑖=1 𝑗=1
1
= (𝑏 − 𝑎 + 1)𝑚 2 𝑛 2 ∣∣𝐴∣∣1 .
1
1
This implies ∣∣ℐ∣∣L (𝑀𝑚,𝑛 (𝐿1 (𝑎,𝑏)),𝑀𝑚,𝑛 (𝑊 1,1 (𝑎,𝑏))) ≤ (𝑏 − 𝑎 + 1)𝑚 2 𝑛 2 and hence it follows that
ℐ ∈ L (𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)), 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))). This completes the proof.
1,1
(ℝ)) defined by
Lemma 87 The integral map ℐ : 𝑀𝑚,𝑛 (𝐿1loc (ℝ)) → 𝑀𝑚,𝑛 (𝑊loc
∫
(ℐ𝐴)(𝑥) :=
𝑥
𝐴(𝑡)𝑑𝑡,
𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1loc (ℝ)),
0
190
𝑥 ∈ ℝ,
is well-defined and satisfies
(ℐ𝐴)′ = 𝐴,
for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1loc (ℝ)).
Proof.
Let 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1loc (ℝ)). Then it follows that the function 𝐵𝐴 (⋅) : ℝ → 𝑀𝑚,𝑛 (ℂ),
defined by
∫
𝑥
𝐴(𝑡)𝑑𝑡,
𝐵𝐴 (𝑥) :=
𝑥 ∈ ℝ,
0
1,1
is well-defined. Moreover, by Lemma 86 we have that 𝐵𝐴 ∣(𝑎,𝑏) ∈ 𝑀𝑚,𝑛 (𝑊loc
(𝑎, 𝑏)) for any
bounded interval (𝑎, 𝑏) in ℝ containing 0 and 𝐵𝐴 ∣′(𝑎,𝑏) = 𝐴∣(𝑎,𝑏) . By Lemma 71 it follows
1,1
that 𝐵𝐴 ∈ 𝑀𝑚,𝑛 (𝑊loc
(ℝ)) and 𝐵𝐴 ′ ∣(𝑎,𝑏) = 𝐵𝐴 ∣′(𝑎,𝑏) = 𝐴∣(𝑎,𝑏) for any bounded interval (𝑎, 𝑏)
in ℝ containing 0. This proves 𝐵𝐴′ = 𝐴. But by the definition of the integral map we have
1,1
(ℝ)) is
ℐ𝐴 = 𝐵𝐴 for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1loc (ℝ)). Therefore ℐ : 𝑀𝑚,𝑛 (𝐿1loc (ℝ)) → 𝑀𝑚,𝑛 (𝑊loc
well-defined and (ℐ𝐴)′ = 𝐴 for any 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1loc (ℝ)). This completes the proof.
Lemma 88 Let 1 ≤ 𝑝 ≤ ∞ and (𝑎, 𝑏) ⊆ ℝ be a bounded open interval. Define the restriction
map by 𝐴 7→ 𝐴∣(𝑎,𝑏) where 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)). Then ∣(𝑎,𝑏) ∈ L (𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)), 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)).
Proof.
Fix 𝑝 ∈ [1, ∞] and a bounded open interval (𝑎, 𝑏). Pick 𝑞 ∈ ℕ such that (𝑎, 𝑏) ⊆
[−𝑞𝑑, 𝑞𝑑]. Let 𝑓 ∈ 𝐿𝑝 (𝕋). Then 𝑓 ∈ 𝐿𝑝loc (ℝ) and 𝑓 (𝑥 + 𝑑) = 𝑓 (𝑥) for a.e. 𝑥 ∈ ℝ. If 𝑝 = ∞
then
∣∣𝑓 ∣(𝑎,𝑏) ∣∣∞ ≤ ∣∣𝑓 ∣(−𝑞𝑑,𝑞𝑑) ∣∣∞ = ∣∣𝑓 ∣(0,𝑑) ∣∣∞ = ∣∣𝑓 ∣∣𝐿∞ (𝕋) .
191
If 𝑝 ∈ [0, ∞) then by periodicity we have
∫ 𝑗𝑑
∣𝑓 (𝑥)∣𝑝 𝑑𝑥 =
(𝑗−1)𝑑
∫𝑑
0
∣𝑓 (𝑥)∣𝑝 𝑑𝑥 for any 𝑗 ∈ ℤ.
Hence implying
∣∣𝑓 ∣(𝑎,𝑏) ∣∣𝑝𝑝
𝑏
∫
∣𝑓 (𝑥)∣𝑝 𝑑𝑥
−𝑞𝑑
𝑎
𝑞
𝑗𝑑
∑∫
𝑝
∣𝑓 (𝑥)∣ 𝑑𝑥 +
𝑞 ∫
∑
(𝑗−1)𝑑
𝑗=1
𝑞
=
𝑞𝑑
∣𝑓 (𝑥)∣ 𝑑𝑥 ≤
=
=
∫
𝑝
𝑝
∣𝑓 (𝑥)∣ 𝑑𝑥 +
0
𝑗=1
𝑞 ∫
∑
∣𝑓 (𝑥)∣𝑝 𝑑𝑥
−𝑗𝑑
𝑗=1
𝑑
∑∫
−(𝑗−1)𝑑
𝑑
∣𝑓 (𝑥)∣𝑝 𝑑𝑥
0
𝑗=1
𝑑
∫
= 2𝑞
0
∣𝑓 (𝑥)∣𝑝 𝑑𝑥 = 2𝑞𝑑∣∣𝑓 ∣∣𝑝𝐿𝑝 (𝕋) .
1
Thus for any 𝑝 ∈ [1, ∞] we have ∣∣𝑓 ∣(𝑎,𝑏) ∣∣𝑝 ≤ (2𝑞𝑑) 𝑝 ∣∣𝑓 ∣∣𝐿𝑝 (𝕋) . But then for any 𝐴 :=
𝑝
[𝑎𝑖𝑗 ]𝑚,𝑛
𝑖=1,𝑗=1 ∈ 𝑀𝑚,𝑛 (𝐿 (𝕋)) we have
(
∣∣𝐴∣(𝑎,𝑏) ∣∣𝑝 =
𝑚 ∑
𝑛
∑
) 12
∣∣𝑎𝑖𝑗 ∣(𝑎,𝑏) ∣∣2𝑝
1
1
≤ 𝑚 2 𝑛 2 max ∣∣𝑎𝑖𝑗 ∣(𝑎,𝑏) ∣∣𝑝
1≤𝑖≤𝑚
1≤𝑗≤𝑛
𝑖=1 𝑗=1
1
1
1
≤ (2𝑞𝑑) 𝑝 𝑚 2 𝑛 2 max ∣∣𝑎𝑖𝑗 ∣∣𝐿𝑝 (𝕋)
1≤𝑖≤𝑚
1≤𝑗≤𝑛
1
𝑝
1
2
≤ (2𝑞𝑑) 𝑚 𝑛
1
2
( 𝑚 𝑛
∑∑
) 21
∣∣𝑎𝑖𝑗 ∣∣2𝐿𝑝 (𝕋)
1
1
1
= (2𝑞𝑑) 𝑝 𝑚 2 𝑛 2 ∣∣𝐴∣∣𝐿𝑝 (𝕋) .
𝑖=1 𝑗=1
And hence we conclude the restriction map ∣(𝑎,𝑏) : 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)) → 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) is a
1
1
1
bounded linear operator with norm bounded above by (2𝑞𝑑) 𝑝 𝑚 2 𝑛 2 . This proves the statement for 𝑝 ∈ [1, ∞].
Parameter Dependent
This is the collection of statements which involve the frequency parameter.
Lemma 89 Let 𝑝, 𝑞, 𝑠 ∈ ℝ with 1 ≤ 𝑝, 𝑞, 𝑠 ≤ ∞ and
192
1
𝑝
+ 1𝑞 = 1𝑠 . If 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)))
and 𝐵 ∈ 𝒪(Ω, 𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏))) then 𝐴𝐵 ∈ 𝒪(Ω, 𝑀𝑚,𝑟 (𝐿𝑠 (𝑎, 𝑏))) and
(𝐴𝐵)𝜔 = 𝐴𝜔 𝐵 + 𝐴𝐵𝜔 .
Proof. By Lemma 82 we know that the map of matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) ×
𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝐿𝑠 (𝑎, 𝑏)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) and 𝐵 ∈
𝑀𝑛,𝑟 (𝐿𝑞 (𝑎, 𝑏)), is a continuous bilinear map. Thus by Lemma 78 we must have 𝐴𝐵 ∈
𝒪(Ω, 𝑀𝑚,𝑟 (𝐿𝑠 (𝑎, 𝑏))) and
(𝐴𝐵)𝜔 = 𝐴𝜔 𝐵 + 𝐴𝐵𝜔 .
This completes the proof.
Lemma 90 Let 1 ≤ 𝑝 ≤ ∞. If 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏))) and 𝐵 ∈ 𝒪(Ω, 𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)))
then 𝐴𝐵 ∈ 𝒪(Ω, 𝑀𝑛,𝑟 (𝐿𝑝 (𝑎, 𝑏))) and
(𝐴𝐵)𝜔 = 𝐴𝜔 𝐵 + 𝐴𝐵𝜔 .
Proof. By Lemma 83 we know that the map of matrix multiplication ⋅ : 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) ×
𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑚,𝑟 (𝐿𝑝 (𝑎, 𝑏)), where 𝐴 ⋅ 𝐵 := 𝐴𝐵 for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)), 𝐵 ∈
𝑀𝑛,𝑟 (𝑊 1,1 (𝑎, 𝑏)), is a continuous bilinear map. Thus by Lemma 78 we must have 𝐴𝐵 ∈
𝒪(Ω, 𝑀𝑚,𝑟 (𝐿𝑝 (𝑎, 𝑏))) and
(𝐴𝐵)𝜔 = 𝐴𝜔 𝐵 + 𝐴𝐵𝜔 .
This completes the proof.
Lemma 91 Let 𝑥 ∈ [𝑎, 𝑏] and 𝛿 𝑥 : 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)) → 𝑀𝑚,𝑛 (ℂ) be the evaluation map
193
at 𝑥, i.e., 𝛿 𝑥 𝐴 = 𝐴(𝑥) for all 𝐴 ∈ 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)). If 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) then
𝛿 𝑥 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (ℂ)) and
(𝛿 𝑥 𝐴)𝜔 = 𝛿 𝑥 𝐴𝜔 .
Proof.
By Lemma 85 we know 𝛿 𝑥 ∈ L (𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏)), 𝑀𝑚,𝑛 (ℂ)). Thus by Lemma 79,
if 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) then 𝛿 𝑥 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (ℂ)) and
(𝛿 𝑥 𝐴)𝜔 = 𝛿 𝑥 𝐴𝜔 .
This completes the proof.
Lemma 92 Let 𝑎, 𝑏 ∈ ℝ, 𝑎 < 𝑏, and 𝑢 ∈ [𝑎, 𝑏]. Let ℐ : 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)) → 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))
be the integral map, i.e.,
∫
𝑥
𝐴(𝑡)𝑑𝑡,
(ℐ𝐴)(𝑥) :=
𝐴 ∈ 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)),
𝑥 ∈ (𝑎, 𝑏).
𝑢
If 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏))) then ℐ𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) and
(ℐ𝐴)𝜔 = ℐ𝐴𝜔 .
Proof.
By Lemma 86 we know that ℐ ∈ L (𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏)), 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))). Thus by
Lemma 79, if 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿1 (𝑎, 𝑏))) then ℐ𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑊 1,1 (𝑎, 𝑏))) and
(ℐ𝐴)𝜔 = ℐ𝐴𝜔 .
This completes the proof.
194
Lemma 93 Let 𝑝 ∈ [1, ∞) and (𝑎, 𝑏) be any bounded open interval in ℝ. Let ∣(𝑎,𝑏) :
𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)) → 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)) be the restriction map, i.e., the domain restriction map 𝐴 7→
𝐴∣(𝑎,𝑏) for 𝐴 ∈ 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)). If 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝕋))) then 𝐴∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)))
and
(𝐴∣(𝑎,𝑏) )𝜔 = 𝐴𝜔 ∣(𝑎,𝑏) .
Proof.
By Lemma 88 we know that ∣(𝑎,𝑏) ∈ L (𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)), 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)). Thus by
Lemma 79, if 𝐴 ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝑀𝑚,𝑛 (𝐿𝑝 (𝕋)))) then we have 𝐴∣(𝑎,𝑏) ∈ 𝒪(Ω, 𝑀𝑚,𝑛 (𝐿𝑝 (𝑎, 𝑏)))
and
(𝐴∣(𝑎,𝑏) )𝜔 = 𝐴𝜔 ∣(𝑎,𝑏) .
This completes the proof.
195
Bibliography
[1] A. L. Andrew, K.-w. E. Chu, and P. Lancaster. Derivatives of eigenvalues and eigenvectors of matrix functions. SIAM J. Matrix Anal. Appl., 14(4):903–926, 1993.
[2] A. L. Andrew and R. C. E. Tan. Iterative computation of derivatives of repeated
eigenvalues and the corresponding eigenvectors. Numer. Linear Algebra Appl., 7(4):151–
167, 2000.
[3] J. Ballato, A. Ballato, A. Figotin, and I. Vitebskiy. Frozen light in periodic stacks of
anisotropic layers. Phys. Rev. E, 71(3):036612, Mar 2005.
[4] H. Baumgärtel. Analytic perturbation theory for matrices and operators, volume 15 of
Operator Theory: Advances and Applications. Birkhäuser Verlag, Basel, 1985.
[5] D. W. Berreman. Optics in stratified and anisotropic media: 4 x 4-matrix formulation.
J. Opt. Soc. Am., 62(4):502–510, 1972.
[6] M. V. Berry. The optical singularities of bianisotropic crystals. Proc. R. Soc. Lond.
Ser. A Math. Phys. Eng. Sci., 461(2059):2071–2098, 2005.
[7] L. Brillouin. Wave propagation in periodic structures. Electric filters and crystal lattices.
Dover Publications Inc., New York, N. Y., 2nd edition, 1953.
[8] L. Brillouin. Wave propagation and group velocity. Pure and Applied Physics, Vol. 8.
Academic Press, New York, 1960.
[9] K.-w. E. Chu. On multiple eigenvalues of matrices depending on several parameters.
SIAM J. Numer. Anal., 27(5):1368–1385, 1990.
[10] D. Duche, L. Escoubas, J.-J. Simon, P. Torchio, W. Vervisch, and F. Flory. Slow bloch
modes for enhancing the absorption of light in thin films for photovoltaic cells. Applied
Physics Letters, 92(19):193310 –193310–3, may 2008.
[11] M. A. Fiddy and R. Tsu. Understanding metamaterials. Waves Random Complex
Media, 20(2):202–222, 2010.
[12] A. Figotin and I. Vitebskiy. Nonreciprocal magnetic photonic crystals. Phys. Rev. E,
63(6):066609, May 2001.
196
[13] A. Figotin and I. Vitebskiy. Electromagnetic unidirectionality in magnetic photonic
crystals. Phys. Rev. B, 67(16):165210, Apr 2003.
[14] A. Figotin and I. Vitebskiy. Oblique frozen modes in periodic layered media. Phys. Rev.
E, 68(3):036609, Sep 2003.
[15] A. Figotin and I. Vitebskiy. Gigantic transmission band-edge resonance in periodic
stacks of anisotropic layers. Phys. Rev. E, 72(3):036619, Sep 2005.
[16] A. Figotin and I. Vitebskiy. Slow light in photonic crystals. Waves Random Complex
Media, 16(3):293–382, 2006.
[17] B. A. Fuks. Theory of analytic functions of several complex variables. Translated by A.
A. Brown, J. M. Danskin and E. Hewitt. American Mathematical Society, Providence,
R.I., 1963.
[18] I. Gohberg, P. Lancaster, and L. Rodman. Invariant subspaces of matrices with applications, volume 51 of Classics in Applied Mathematics. Society for Industrial and Applied
Mathematics (SIAM), Philadelphia, PA, 2006. Reprint of the 1986 original.
[19] I. Gohberg and J. Leiterer. Holomorphic operator functions of one variable and applications, volume 192 of Operator Theory: Advances and Applications. Birkhäuser Verlag,
Basel, 2009. Methods from complex analysis in several variables.
[20] H. W. Gould. Coefficient identities for powers of Taylor and Dirichlet series. Amer.
Math. Monthly, 81:3–14, 1974.
[21] A. Greenleaf, M. Lassas, and G. Uhlmann. On nonuniqueness for Calderón’s inverse
problem. Math. Res. Lett., 10(5-6):685–693, 2003.
[22] U. G. Greenleaf A., Lassas M. Anisotropic conductivities that cannot be detected by
EIT. Physiological Measurement, 24:413–419(7), 2003.
[23] R. Hryniv and P. Lancaster. On the perturbation of analytic matrix functions. Integral
Equations Operator Theory, 34(3):325–338, 1999.
[24] I. C. F. Ipsen and R. Rehman. Perturbation bounds for determinants and characteristic
polynomials. SIAM J. Matrix Anal. Appl., 30(2):762–776, 2008.
[25] J. D. Jackson. Classical electrodynamics. John Wiley & Sons Inc., New York, 2nd
edition, 1975.
[26] C.-P. Jeannerod and E. Pflügel. A reduction algorithm for matrices depending on
a parameter. In Proceedings of the 1999 International Symposium on Symbolic and
Algebraic Computation (Vancouver, BC), pages 121–128 (electronic), New York, 1999.
ACM.
[27] J. D. Joannopoulos, S. G. Johnson, J. N. Winn, and R. D. Meade. Photonic Crystals:
Molding the Flow of Light. Princeton University Press, 2nd edition, 2008.
197
[28] T. Kato. Perturbation theory for linear operators. Classics in Mathematics. SpringerVerlag, Berlin, 1995. Reprint of the 1980 edition.
[29] V. Kozlov and V. Maz′ ya. Differential equations with operator coefficients with applications to boundary value problems for partial differential equations. Springer Monographs
in Mathematics. Springer-Verlag, Berlin, 1999.
[30] S. G. Krantz. Function theory of several complex variables. The Wadsworth &
Brooks/Cole Mathematics Series. Wadsworth & Brooks/Cole Advanced Books & Software, Pacific Grove, CA, second edition, 1992.
[31] S. G. Kreı̆n and E. O. Utochkina. An implicit canonical equation in a Hilbert space.
Ukrain. Mat. Zh., 42(3):388–390, 1990.
[32] P. Kuchment. The mathematics of photonic crystals. In Mathematical modeling in
optical science, volume 22 of Frontiers Appl. Math., pages 207–272. SIAM, Philadelphia,
PA, 2001.
[33] P. Lancaster. On eigenvalues of matrices dependent on a parameter. Numer. Math.,
6:377–387, 1964.
[34] P. Lancaster, A. S. Markus, and F. Zhou. Perturbation theory for analytic matrix
functions: the semisimple case. SIAM J. Matrix Anal. Appl., 25(3):606–626 (electronic),
2003.
[35] P. Lancaster and M. Tismenetsky. The theory of matrices. Computer Science and
Applied Mathematics. Academic Press Inc., Orlando, FL, second edition, 1985.
[36] L. Landau and E. Lifshitz. Electrodynamics of Continuous Media. Pergamon Press,
Oxford, 2nd edition, 1984.
[37] H. Langer and B. Najman. Remarks on the perturbation of analytic matrix functions.
II. Integral Equations Operator Theory, 12(3):392–407, 1989.
[38] H. Langer and B. Najman. Remarks on the perturbation of analytic matrix functions.
III. Integral Equations Operator Theory, 15(5):796–806, 1992.
[39] H. Langer and B. Najman. Leading coefficients of the eigenvalues of perturbed analytic
matrix functions. Integral Equations Operator Theory, 16(4):600–604, 1993.
[40] U. Leonhardt. Optical Conformal Mapping. Science, 312(5781):1777–1780, 2006.
[41] V. B. Lidskiı̆. On the theory of perturbations of nonselfadjoint operators. Z̆. Vyčisl.
Mat. i Mat. Fiz., 6(1):52–60, 1966.
[42] I. V. Lindell, A. H. Sihvola, P. Puska, and L. H. Ruotanen. Conditions for the parameter
dyadics of lossless bianisotropic media. Microwave and Optical Technology Letters,
8(5):268–272, 1995.
198
[43] Y. Ma and A. Edelman. Nongeneric eigenvalue perturbations of Jordan blocks. Linear
Algebra Appl., 273:45–63, 1998.
[44] A. S. Markus. Introduction to the spectral theory of polynomial operator pencils, volume 71 of Translations of Mathematical Monographs. American Mathematical Society,
Providence, RI, 1988. Translated from the Russian by H. H. McFaden, Translation
edited by Ben Silver, With an appendix by M. V. Keldysh.
[45] R. Mennicken and M. Möller. Non-self-adjoint boundary eigenvalue problems, volume
192 of North-Holland Mathematics Studies. North-Holland Publishing Co., Amsterdam,
2003.
[46] C. D. Meyer and G. W. Stewart. Derivatives and perturbations of eigenvectors. SIAM
J. Numer. Anal., 25(3):679–691, 1988.
[47] J. Moro, J. V. Burke, and M. L. Overton. On the Lidskii-Vishik-Lyusternik perturbation
theory for eigenvalues of matrices with arbitrary Jordan structure. SIAM J. Matrix
Anal. Appl., 18(4):793–817, 1997.
[48] J. Moro and F. M. Dopico. First order eigenvalue perturbation theory and the Newton
diagram. In Applied mathematics and scientific computing (Dubrovnik, 2001), pages
143–175. Kluwer/Plenum, New York, 2003.
[49] G. Mumcu, K. Sertel, J. Volakis, I. Vitebskiy, and A. Figotin. RF propagation in finite
thickness unidirectional magnetic photonic crystals. Antennas and Propagation, IEEE
Transactions on, 53(12):4026 – 4034, Dec. 2005.
[50] J. B. Pendry. Negative refraction makes a perfect lens. Phys. Rev. Lett., 85(18):3966–
3969, Oct 2000.
[51] J. B. Pendry, D. Schurig, and D. R. Smith. Controlling Electromagnetic Fields. Science,
312(5781):1780–1782, 2006.
[52] A. Serdyukov, I. Semchenko, S. Tretyakov, and A. Sihvola. Electromagnetics of Bianisotropic Materials: Theory and Applications. Gordon and Breach, Amsterdam, 2001.
[53] A. P. Seyranian and A. A. Mailybaev. Multiparameter stability theory with mechanical
applications, volume 13 of Series on Stability, Vibration and Control of Systems. Series
A: Textbooks, Monographs and Treatises. World Scientific Publishing Co. Inc., River
Edge, NJ, 2003.
[54] A. Sihvola. Metamaterials in electromagnetics. Metamaterials, 1(1):2 – 11, 2007.
[55] J. G. Sun. Eigenvalues and eigenvectors of a matrix dependent on several parameters.
J. Comput. Math., 3(4):351–364, 1985.
[56] J. G. Sun. Multiple eigenvalue sensitivity analysis. Linear Algebra Appl., 137/138:183–
211, 1990.
199
[57] M. M. Vaı̆nberg and V. A. Trenogin. Theory of branching of solutions of non-linear
equations. Noordhoff International Publishing, Leyden, 1974. Translated from the
Russian by Israel Program for Scientific Translations.
[58] M. I. Višik and L. A. Ljusternik. Solution of some perturbation problems in the case
of matrices and self-adjoint or non-selfadjoint differential equations. I. Russian Math.
Surveys, 15(3):1–73, 1960.
[59] J. Volakis, G. Mumcu, K. Sertel, C.-C. Chen, M. Lee, B. Kramer, D. Psychoudakis, and
G. Kiziltas. Antenna miniaturization using magnetic-photonic and degenerate bandedge crystals. Antennas and Propagation Magazine, IEEE, 48(5):12–28, Oct. 2006.
[60] A. Welters. Perturbation analysis of slow waves for periodic differential-algebraic equations of definite type. In preparation.
[61] A. Welters. On explicit recursive formulas in the spectral perturbation analysis of a
Jordan block. SIAM J. Matrix Anal. Appl., 32(1):1–22, 2011.
[62] S. Yarga, K. Sertel, and J. Volakis. Degenerate band edge crystals for directive antennas.
Antennas and Propagation, IEEE Transactions on, 56(1):119–126, Jan. 2008.
[63] P. Yeh. Electromagnetic propagation in birefringent layered media. J. Opt. Soc. Amer.,
69(5):742–756, 1979.
200