Uploaded by iii iii

kythe pk greens functions and linear differential equations

advertisement
GREEN’S
FUNCTIONS
AND LINEAR
DIFFERENTIAL
EQUATIONS
Theory, Applications,
and Computation
CHAPMAN & HALL/CRC APPLIED MATHEMATICS
AND NONLINEAR SCIENCE SERIES
Series Editor Goong Chen
Published Titles
Advanced Differential Quadrature Methods, Zhi Zong and Yingyan Zhang
Computing with hp-ADAPTIVE FINITE ELEMENTS, Volume 1, One and Two Dimensional
Elliptic and Maxwell Problems, Leszek Demkowicz
Computing with hp-ADAPTIVE FINITE ELEMENTS, Volume 2, Frontiers: Three
Dimensional Elliptic and Maxwell Problems with Applications, Leszek Demkowicz,
Jason Kurtz, David Pardo, Maciej Paszyński, Waldemar Rachowicz, and Adam Zdunek
CRC Standard Curves and Surfaces with Mathematica®: Second Edition,
David H. von Seggern
Discovering Evolution Equations with Applications: Volume 1-Deterministic Equations,
Mark A. McKibben
Exact Solutions and Invariant Subspaces of Nonlinear Partial Differential Equations in
Mechanics and Physics, Victor A. Galaktionov and Sergey R. Svirshchevskii
Geometric Sturmian Theory of Nonlinear Parabolic Equations and Applications,
Victor A. Galaktionov
Green’s Functions and Linear Differential Equations: Theory, Applications,
and Computation, Prem K. Kythe
Introduction to Fuzzy Systems, Guanrong Chen and Trung Tat Pham
Introduction to non-Kerr Law Optical Solitons, Anjan Biswas and Swapan Konar
Introduction to Partial Differential Equations with MATLAB®, Matthew P. Coleman
Introduction to Quantum Control and Dynamics, Domenico D’Alessandro
Mathematical Methods in Physics and Engineering with Mathematica, Ferdinand F. Cap
Mathematical Theory of Quantum Computation, Goong Chen and Zijian Diao
Mathematics of Quantum Computation and Quantum Technology, Goong Chen,
Louis Kauffman, and Samuel J. Lomonaco
Mixed Boundary Value Problems, Dean G. Duffy
Multi-Resolution Methods for Modeling and Control of Dynamical Systems,
Puneet Singla and John L. Junkins
Optimal Estimation of Dynamic Systems, John L. Crassidis and John L. Junkins
Quantum Computing Devices: Principles, Designs, and Analysis, Goong Chen,
David A. Church, Berthold-Georg Englert, Carsten Henkel, Bernd Rohwedder,
Marlan O. Scully, and M. Suhail Zubairy
A Shock-Fitting Primer, Manuel D. Salas
Stochastic Partial Differential Equations, Pao-Liu Chow
CHAPMAN & HALL/CRC APPLIED MATHEMATICS
AND NONLINEAR SCIENCE SERIES
GREEN’S
FUNCTIONS
AND LINEAR
DIFFERENTIAL
EQUATIONS
Theory, Applications,
and Computation
Prem K. Kythe
University of New Orleans
Louisiana, USA
Chapman & Hall/CRC
Taylor & Francis Group
6000 Broken Sound Parkway NW, Suite 300
Boca Raton, FL 33487-2742
© 2011 by Taylor and Francis Group, LLC
Chapman & Hall/CRC is an imprint of Taylor & Francis Group, an Informa business
No claim to original U.S. Government works
Printed in the United States of America on acid-free paper
10 9 8 7 6 5 4 3 2 1
International Standard Book Number-13: 978-1-4398-4009-2 (Ebook-PDF)
This book contains information obtained from authentic and highly regarded sources. Reasonable
efforts have been made to publish reliable data and information, but the author and publisher cannot
assume responsibility for the validity of all materials or the consequences of their use. The authors and
publishers have attempted to trace the copyright holders of all material reproduced in this publication
and apologize to copyright holders if permission to publish in this form has not been obtained. If any
copyright material has not been acknowledged please write and let us know so we may rectify in any
future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced,
transmitted, or utilized in any form by any electronic, mechanical, or other means, now known or
hereafter invented, including photocopying, microfilming, and recording, or in any information storage or retrieval system, without written permission from the publishers.
For permission to photocopy or use material electronically from this work, please access www.copyright.com (http://www.copyright.com/) or contact the Copyright Clearance Center, Inc. (CCC), 222
Rosewood Drive, Danvers, MA 01923, 978-750-8400. CCC is a not-for-profit organization that provides licenses and registration for a variety of users. For organizations that have been granted a photocopy license by the CCC, a separate system of payment has been arranged.
Trademark Notice: Product or corporate names may be trademarks or registered trademarks, and are
used only for identification and explanation without intent to infringe.
Visit the Taylor & Francis Web site at
http://www.taylorandfrancis.com
and the CRC Press Web site at
http://www.crcpress.com
Contents
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Notations and Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
1. Some Basic Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1. Euclidean Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.1. Metric Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.2. Inner Product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2. Classes of Continuous Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3. Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3.1. Convergence of Sequences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3.2. Weak Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3.3. Metric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.4. Convergence of Infinite Series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3.5. Tests for Convergence of Positive Series . . . . . . . . . . . . . . . . . . . . . . . . 6
1.4. Functionals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.4.1. Examples of Linear Functionals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.5. Linear Transformations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.6. Cramer’s Rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.7. Green’s Identities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.8. Differentiation and Integration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.8.1. Leibniz’s Rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.8.2. Integration by Parts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.9. Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.9.1. Bessel’s Inequality for Fourier Series . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.9.2. Bessel’s Inequality for Square-Integrable Functions . . . . . . . . . . . . . 14
1.9.3. Schwarz’s Inequality for Infinite Sequences . . . . . . . . . . . . . . . . . . . . 15
1.10. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2. The Concept of Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1. Generalized Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.1. Heaviside Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
2.1.2. Delta Function in Curvilinear Coordinates . . . . . . . . . . . . . . . . . . . . . 27
2.2. Singular Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2.3. The Concept of Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.4. Linear Operators and Inverse Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
v
vi
CONTENTS
2.4.1. Linear Operators and Inverse Operators . . . . . . . . . . . . . . . . . . . . . . . 34
2.4.2. Adjoint Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
2.5. Fundamental Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
2.6. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
3. Sturm-Liouville Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.1. Ordinary Differential Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.1.1. Initial and Boundary Conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.1.2. General Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3.1.3. Method of Variation of Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
3.2. Initial Value Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
3.2.1. One-Sided Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
3.2.2. Wronskian Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
3.2.3. Systems of First-Order Differential Equations . . . . . . . . . . . . . . . . . . 55
3.3. Boundary Value Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
3.3.1. Sturm-Liouville Boundary Value Problems . . . . . . . . . . . . . . . . . . . . 56
3.3.2. Properties of Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
3.3.3. Green’s Function Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
3.4. Eigenvalue Problem for Sturm-Liouville Systems . . . . . . . . . . . . . . . . . . . . 64
3.4.1. Eigenpairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
3.4.2. Orthonormal Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
3.4.3. Eigenfunction Expansion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
3.4.4. Data for Eigenvalue Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
3.5. Periodic Sturm-Liouville Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
3.6. Singular Sturm-Liouville Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
3.7. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
4. Bernoulli’s Separation Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
4.1. Coordinate Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
4.2. Partial Differential Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
4.3. Bernoulli’s Separation Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
4.3.1. Laplace’s Equation in a Cube . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
4.3.2. Laplace’s Equation in a Cylinder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
4.3.3. Laplace’s Equation in a Sphere . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
4.3.4. Helmholtz’s Equation in Cartesian Coordinates . . . . . . . . . . . . . . . . . 92
4.3.5. Helmholtz’s Equation in Spherical Coordinates . . . . . . . . . . . . . . . . . 93
4.3.6. Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
4.4. Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
4.5. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
5. Integral Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
5.1. Integral Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
5.2. Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
5.2.1. Definition of Dirac Delta Function . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
5.3. Fourier Integral Theorems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
5.3.1. Properties of Fourier Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
5.3.2. Fourier Transforms of Derivatives of a Function . . . . . . . . . . . . . . . 127
CONTENTS
vii
5.3.3. Convolution Theorems for Fourier Transform . . . . . . . . . . . . . . . . . 127
5.4. Fourier Sine and Cosine Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
5.4.1. Properties of Fourier Sine and Cosine Transforms . . . . . . . . . . . . . 130
5.4.2. Convolution Theorems for Fourier Sine and Cosine Transforms . 131
5.5. Finite Fourier Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
5.5.1. Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
5.5.2. Periodic Extensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
5.5.3. Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
5.6. Multiple Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
5.7. Hankel Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
5.8. Summary: Variables of Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
5.9. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
6. Parabolic Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
6.1. 1-D Diffusion Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
6.1.1. Sturm-Liouville System for 1-D Diffusion Equation . . . . . . . . . . . 144
6.1.2. Green’s Function for 1-D Diffusion Equation . . . . . . . . . . . . . . . . . 146
6.2. 2-D Diffusion Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
6.2.1. Dirichlet Problem for the General Parabolic Equation in a Square 149
6.3. 3-D Diffusion Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
6.3.1. Electrostatic Analog . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
6.4. Schrödinger Diffusion Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
6.5. Min-Max Principle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6.6. Diffusion Equation in a Finite Medium . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6.7. Axisymmetric Diffusion Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
6.8. 1-D Heat Conduction Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158
6.9. Stefan Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160
6.10. 1-D Fractional Diffusion Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
6.10.1. 1-D Fractional Diffusion Equation in Semi-Infinite Medium . . . 165
6.11. 1-D Fractional Schrödinger Diffusion Equation . . . . . . . . . . . . . . . . . . . 166
6.12. Eigenpairs and Dirac Delta Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
6.13. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 170
7. Hyperbolic Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
7.1. 1-D Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
7.1.1. Sturm-Liouville System for 1-D Wave Equation . . . . . . . . . . . . . . . 175
7.1.2. Vibrations of a Variable String . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
7.1.3. Green’s Function for 1-D Wave Equation . . . . . . . . . . . . . . . . . . . . . 179
7.2. 2-D Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
7.3. 3-D Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
7.4. 2-D Axisymmetric Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
7.5. Vibrations of a Circular Membrane . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
7.6. 3-D Wave Equation in a Cube . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
7.7. Schrödinger Wave Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186
7.8. Hydrogen Atom . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
7.8.1. Harmonic Oscillator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
viii
CONTENTS
7.9. 1-D Fractional Nonhomogeneous Wave Equation . . . . . . . . . . . . . . . . . . . 190
7.10. Applications of the Wave Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193
7.10.1. Cauchy Problem for 2-D and 3-D Wave Equation . . . . . . . . . . . . 193
7.10.2. d’Alembert Solution of the Cauchy Problem for Wave Equation 194
7.10.3. Free Vibration of a Large Circular Membrane . . . . . . . . . . . . . . . . 196
7.10.4. Hyperbolic or Parabolic Equations in Terms of Green’s Functions 196
7.11. Laplace Transform Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198
7.12. Quasioptics and Diffraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
7.12.1. Diffraction of Monochromatic Waves . . . . . . . . . . . . . . . . . . . . . . . 201
(a) Fraunhofer Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
(b) Fresnel Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
7.13. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
8. Elliptic Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209
8.1. Green’s Function for 2-D Laplace’s Equation . . . . . . . . . . . . . . . . . . . . . . 209
8.2. 2-D Laplace’s Equation in a Rectangle . . . . . . . . . . . . . . . . . . . . . . . . . . . . 211
8.3. Green’s Function for 3-D Laplace’s Equation . . . . . . . . . . . . . . . . . . . . . . 212
8.3.1. Laplace’s Equation in a Rectangular Parallelopiped . . . . . . . . . . . . 213
8.4. Harmonic Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 217
8.5. 2-D Helmholtz’s Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 218
8.5.1. Closed-Form Green’s Function for Helmholtz’s Equation . . . . . . . 219
8.6. Green’s Function for 3-D Helmholtz’s Equation . . . . . . . . . . . . . . . . . . . . 220
8.7. 2-D Poisson’s Equation in a Circle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221
8.8. Method for Green’s Function in a Rectangle . . . . . . . . . . . . . . . . . . . . . . . 226
8.9. Poisson’s Equation in a Cube . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
8.10. Laplace’s Equation in a Sphere . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231
8.11. Poisson’s Equation and Green’s Function in a Sphere . . . . . . . . . . . . . . 235
8.12. Applications of Elliptic Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237
8.12.1. Dirichlet Problem for Laplace’s Equation . . . . . . . . . . . . . . . . . . . . 237
8.12.2. Neumann Problem for Laplace’s Equation . . . . . . . . . . . . . . . . . . . 237
8.12.3. Robin Problem for Laplace’s Equation . . . . . . . . . . . . . . . . . . . . . . 239
8.12.4. Dirichlet Problem for Helmholtz’s Equation . . . . . . . . . . . . . . . . . 239
8.12.5. Dirichlet Problem for Laplace’s Equation in the Half-Plane . . . . 240
8.12.6. Dirichlet Problem for Laplace’s Equation in a Circle . . . . . . . . . . 241
8.12.7. Dirichlet Problem for Laplace’s Equation in the Quarter Plane . 241
8.12.8. Vibration Equation for the Unit Sphere . . . . . . . . . . . . . . . . . . . . . . 243
8.13. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244
9. Spherical Harmonics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
9.1. Historical Sketch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
9.2. Laplace’s Solid Spherical Harmonics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252
9.2.1. Orthonormalization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
9.2.2. Condon-Shortley Phase Factor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
9.2.3. Spherical Harmonics Expansion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 257
9.2.4. Addition Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
9.2.5. Laplace’s Coefficients . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
CONTENTS
ix
9.3. Surface Spherical Harmonics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261
9.3.1. Poisson Integral Representation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 266
9.3.2. Representation of a Function f (θ, φ) . . . . . . . . . . . . . . . . . . . . . . . . . 268
9.3.3. Addition Theorem for Spherical Harmonics . . . . . . . . . . . . . . . . . . . 269
9.3.4. Discrete Energy Spectrum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 271
9.3.5. Further Developments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 274
9.4. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 276
10. Conformal Mapping Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
10.1. Definitions and Theorems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
10.1.1. Cauchy-Riemann Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
10.1.2. Conformal Mapping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282
10.1.3. Symmetric Points . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283
10.1.4. Cauchy’s Integral Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284
10.1.5. Mean-Value Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284
10.2. Dirichlet Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 285
10.2.1. Dirichlet Problem for a Circle in the (x, y)-Plane . . . . . . . . . . . . . 289
10.3. Neumann Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290
10.4. Green’s and Neumann’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 293
10.4.1. Laplacian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 293
10.4.2. Green’s Function for a Circle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 295
10.4.3. Green’s Function for an Ellipse . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 297
10.4.4. Green’s Function for an Infinite Strip . . . . . . . . . . . . . . . . . . . . . . . 299
10.4.5. Green’s Function for an Annulus . . . . . . . . . . . . . . . . . . . . . . . . . . . 302
10.5. Computation of Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 303
10.5.1. Interpolation Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 304
10.6. Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 309
A. Adjoint Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 314
B. List of Fundamental Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 317
B.1. Linear Ordinary Differential Operator with Constant Coefficients . . . . 317
B.2. Fundamental Solutions for the Operators . . . . . . . . . . . . . . . . . . . . . . . . . . 317
B.3. Elliptic Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 317
B.4. Helmholtz Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 318
B.5. Fundamental Solution for the Cauchy-Riemann Operator . . . . . . . . . . . 319
B.6. Fundamental Solution for the Diffusion Operator . . . . . . . . . . . . . . . . . . . 319
B.7. Schrödinger Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
B.8. Fundamental Solution for the Wave Operator . . . . . . . . . . . . . . . . . . . . . . 321
B.9. Fundamental Solution for the Fokker-Plank Operator . . . . . . . . . . . . . . . 321
B.10. Klein-Gordon Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321
C. List of Spherical Harmonics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322
C.1. Legendre’s Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322
C.2. Associated Legendre’s Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
C.3. Relations with or without Condon-Shortley Phase Factor . . . . . . . . . . . . 324
C.4. Laguerre’s Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 326
x
CONTENTS
C.5. Associated Laguerre’s Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
D. Tables of Integral Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 329
D.1. Laplace Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 329
D.2. Fourier Cosine Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 332
D.3. Fourier Sine Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333
D.4. Complex Fourier Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334
D.5. Finite Sine Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 335
D.6. Finite Cosine Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336
D.7. Zero-Order Hankel Transform Pairs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337
E. Fractional Derivatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338
F. Systems of Ordinary Differential Equations . . . . . . . . . . . . . . . . . 341
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 345
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349
Preface
Boundary value problems associated with ordinary and partial differential equations
have been an integral part of mathematics, mathematical physics, and applied sciences, and Green’s functions for these problems have become an important subject,
particularly appealing to mathematicians, physicists and applied scientists. Although
Green’s functions were first introduced by George Green in 1828, with the physical
interpretation as an ‘influence function’ and ‘potential’ in certain mechanical problems, specially the string problem, these functions have been developed and widely
used during the past 60 years, firstly as part of an exciting research area and later
as part of a curriculum in courses on partial differential equations offered by mathematics, engineering and physics departments to their senior and graduate students.
In most textbooks on ordinary and partial differential equations and boundary value
problems there is generally a single chapter on Green’s functions that provides a
selected portion of this topic with a few basic results and examples, while in others
there exists some sort of limited material scattered throughout the book. Although
the significance of these functions in solving boundary value problems is justifiably
indicated in many books, the material in such a single chapter becomes either very
difficult or just a collection of a few well known boundary value problems that are
solved to justify the presence of this topic.
There are very few books written on the subject of Green’s functions, although
some of the notables are those by Roach [1970], Stakgold [1979], and Sagan [1989].
These books were written to provided theory and examples to justify the development
of the subject, but they are, however, not suitable as textbooks. There are other highly
specialized and research-oriented books on Green’s functions that confine to certain
restricted fields, such as potentials, diffusion and waves, solid state physics, quantum
mechanics, lattice Schrödinger operators, and boundary element method. Such books
certainly do not count as standard textbooks.
Overview
Our aim is to provide a kind of textbook on Green’s Functions that satisfies the
following criteria:
xi
xii
PREFACE
1. It is simple enough for the average senior or graduate student to understand and
grasp the beauty and power of the subject.
2. It has enough mathematical rigor to withstand criticism from the experts.
3. It has a large number of examples and exercises from different areas of mathematics,
applied science and engineering.
4. It provides motivation to appeal to students and teachers alike, by providing
sufficient theoretical basis which leads to the development of Green’s function method
which is applied to solve initial and boundary value problems involving different linear
ordinary and partial differential equations.
5. It possesses a robust self-contained text, full of explanation on different problems, graphical representations where necessary, and a few appendices for certain
background material.
6. It contains about one hundred solved examples and and twice as many exercises
with hints and answers and difficult ones with adequate hints and sometimes complete
solution.
7. It describes and uses the following methods for solving initial and boundary value
problems, which are explained with clarity and detail: (i) classical method of variation
of parameters; (ii) a generalization of method of variation of parameters to construct
one-sided Green’s functions for initial value problems involving linear ordinary differential equations; (iii) a variation of the above method, called the Wronskian method,
to find one-sided Green’s functions for initial value problems involving linear ordinary
differential equations; (iv) a variation of the above method, called the Green’s function method, to construct Green’s functions for boundary value problems involving
linear ordinary differential equations, with a step-by-step procedure and some useful
shortcuts; (v) Bernoulli’s separation method for linear partial differential equations;
(vi) integral transform method to determine Green’s functions for linear parabolic,
hyperbolic, and elliptic equations; (vii) method of images for linear elliptic equations
and related Green’s functions; and (viii) conformal mapping method for determining
Green’s functions for linear elliptic equations; and (ix) an interpolation method for
numerical construction of Green’s functions for convex and starlike regions.
8. The subject material of this textbook is arranged in a manner that tries to eliminate
the two detrimental effects on students, namely, to impart the impression to mathematically inclined students that they are not left out, that they are not dwelling in
a textbook that consists of some rather dull manipulations with integrals or infinite
series; and to technically oriented students that the subject of linear ordinary and
partial differential equations is not to be treated simply by the method of variation of
parameters, or separation of variables, or integral transforms.
Outline
In most books on the subject, the approach to introduce the concept of Green’s
functions has been firstly to provide a definition of the Dirac delta function, secondly
to solve a boundary value problem involving a nonhomogeneous ordinary differential
PREFACE
xiii
equation, like the one (Example 2.9) solved in this book, and finally to define the kernel
of the solution, obtained in the form of an integral equation as Green’s function for
the boundary value problem in question. This inverse approach is also seen in the
case of linear partial differential equations.
The approach adopted in this book is more direct. The first task has been to
explain and define the Dirac delta function precisely and to unravel the mystery
surrounding this generalized function. Then the concept of Green’s function and its
relationship with the Dirac delta function has been carefully established. After these
two hurdles are overcome, the mystique of the development of the entire subject of
Green’s functions and its application in solving linear ordinary and partial differential
equations has been unfolded in a manner that makes the subject simply appealing and
useful.
The book starts in Chapter 1 with an introduction to some basic results and definitions from topics such as Euclidean space, specially the metric space and the concept
of inner product; classes of continuous functions and those that are infinitely differentiable with compact support; convergence of sequences, their weak and strong
convergence, convergence in the mean, and convergence of infinite series; linear functionals, and linear transformations; Cramer’s rule; divergence theorem and Green’s
identities; Leibniz’s rule for differentiation of integrals and for the nth derivative of
product of two functions; formulas for integration by parts; Bessel’s inequality for
Fourier series and for square-integrable functions; Schwarz’s inequality for infinite
sequences; and Parseval’s equality or completeness relation.
The concept of a Green’s function is presented in Chapter 2 by first introducing the ‘generalized function’ known as the Dirac delta function through a limiting
process for certain admissible ‘test functions’ which are infinitely differentiable with
compact support. This approach to define the Dirac delta function seems to be the
simplest as compared to more advanced definitions, available prior to 1945, that
d
arose out of Schwarz’s theory of distributions, namely, (i) δ(x) =
H(x), due to
dx
∞
Heaviside; (ii) δ(x) = limn→∞ fn (x) or δ = n=0 fn for suitable functions fn ,
due
δ(x) = 0 for x = 0, and
∞ to Fourier, Kirchhoff, Heaviside, Jordan and Pauli; (iii)
∞
δ(x)
dx
=
1,
due
to
Dirac
and
Heaviside;
and
(iv)
δ(x−a)f
(x) dx = f (a),
−∞
−∞
∞
or −∞ δ(x)f (x) dx = f (0), due to Fourier, Heaviside, and Dirac. These and other
variations of such definitions did not provide any simple mathematical basis for operating with this generalized function. In fact, the theory of distributions evolved
without making the δ-function as its starting point, and then played a significant role
in future research on partial differential equations, and especially on the theory of
Green’s functions. Basic theorem for the construction of Green’s functions and their
important properties are established in this chapter, which ends with a very simple introduction to fundamental solutions for some differential operators. These solutions,
also known as Green’s functions in the large, or ‘free-space’ Green’s functions, or
singular solutions, are very useful in the development and application of the boundary
element method.
xiv
PREFACE
A thorough review and detailed description of the construction of one-sided
Green’s functions for initial value problems, and that of Green’s functions for boundary value problems for linear ordinary differential equations are presented in Chapter 3. The method of variation of parameters is generalized to produce an effective
method and its variation, called the Wronskian method, to generate one-sided Green’s
functions for initial value problems. A step-by-step procedure for the Green’s function method to construct Green’s functions for boundary value problems in general
and Sturm-Liouville systems in particular is included in this chapter, together with
periodic and singular Sturm-Liouville systems. This chapter ends with a detailed
presentation of eigenvalue problems for Sturm-Liouville systems.
Linear partial differential equations begin in Chapter 4, which is devoted to a
complete description of Bernoulli’s separation method. This method has been very
prominent since 1735 when Daniel Bernoulli formulated the principle of coexistence
of small oscillations, which, after using Taylor’s and John Bernoulli’s theory of vibrating string, led him to believe that the general solution of this problem could be
found in the form of a trigonometric series. But no method for determining the coefficients of this series solution was available until about 25 years later when in 1812
Fourier wrote and later published his extensive memoir Théorie du mouvement de
la chaleur dans les corps solides in the French journal ‘Mémoires de l’Académie’,
1819-1822, wherein he considered the following five problems: (i) one-dimensional
flow of heat; (ii) two-dimensional flow of heat in a rectangle; (iii) three-dimensional
flow of heat in a rectangular parallelopiped; (iv) flow of heat in a sphere when the
temperature depends only on the distance from its center; and (v) flow of heat in
a right circular cylinder when the temperature depends only on the distance from
the axis. These problems also included the question of radiation. In the first three
problems the series solution, when one or more dimensions become infinite, was
shown to degenerate into what we now call ‘Fourier’s integrals’. On the question of
determination of the coefficients in the theory of trigonometric series, it was for the
first time that the real importance of this series was shown and Fourier’s name was
associated with its development. Six problems and fourteen examples are solved in
this chapter by using Bernoull’s separation method.
To make this book self contained, a detailed description of integral transforms, especially the Laplace, Fourier, Fourier sine and cosine, finite Fourier, multiple Laplace
and Fourier, and Hankel transform, is presented in Chapter 5 with properties of these
transforms, along with many examples.
Parabolic equations are dealt in Chapter 6, which describes methods to construct
Green’s functions and solve Dirichlet and other problems for the diffusion equation,
Schrödinger diffusion equation, axisymmetric diffusion equation in one-, two-, and
three-dimensions, and one-dimensional fractional diffusion and Schrödinger diffusion
equations. Hyperbolic equations are studied in Chapter 7 where detailed construction
of Green’s functions for wave equation, axisymmetric wave equation, Schödinger
wave equation in one-, two- and three-dimensions, and wave equation in a cube are
presented; one-dimensional fractional wave equation is studied; and application to the
PREFACE
xv
Cauchy problem and its d’Alembert solution is studied. Other applications, like the
free vibrations of a large circular membrane, are included. The problem of diffraction
in quasi-optics with both Fraunhoffer and Fresnel approximations for diffraction of
monochromatic waves is presented. The investigation into the mathematical aspect of
the hydrogen atom, in conjunction with Schödinger’s wave equation, makes a special
topic presented in this chapter.
Chapter 8 deals with elliptic equations, and Green’s functions for Laplace’s equation in two and three dimensions, including Laplace’s equation in a rectangular parallelopiped, are constructed. Harmonic functions are introduced, and Green’s functions
for two- and three-dimensional Helmholtz’s and Poisson’s equations are determined.
Some important applications of elliptic equations, namely, the Dirichlet, Neumann
and Robin problems, and the vibration equation for the unit sphere, are presented in
detail.
Spherical harmonics, which are the angular portion of a set of Laplace’s equation in
spherical coordinates, are discussed in Chapter 9. They are useful in many theoretical
and physical applications in physics, seismology. geodesy, spectral analysis, magnetic
fields, and quantum mechanics. Both solid and surface spherical harmonics are
analyzed in detail, including the Condon-Shortley phase factor, and discrete energy
spectrum for the hydrogen atom are studied in some detail.
The conformal mapping method to determine Green’s and Neumann’s functions
for the circle, ellipse, infinite strip, and annular region is presented in Chapter 10. The
discussion starts with some useful definitions and results from conformal mapping,
and the relationship between an analytic function and Green’s function is established, which is used to present an interpolation method for numerically constructing
Green’s functions for different types of convex and starlike regions. The computational part has been simplified to the extent that a good quality digital calculator
is sufficient to produce the required numerical values, although a Mathematica
notebook greens.nb and some projects are available in Kythe et al. [2003].
Six appendices, one each on adjoint operators, fundamental solutions, list of harmonics, tables of integral transforms used in the book, fractional derivatives, and
one-sided Green’s functions for systems of linear first-order ordinary differential
equations, are presented in Appendix A through F, which are followed by the Bibliography and the Index.
Layout of the Material
The material of this book ranges from average to challenged sections, which can be
differently suitable for students and readers depending on their background. The basic
prerequisite, however, is a thorough knowledge of differential and integral calculus,
advanced calculus, and an elementary course on linear ordinary differential equations.
This background assumes that the readers are at least at the senior undergraduate level.
The readers with these skills can easily go through the first three chapters. Readers
with advanced knowledge of elementary linear partial differential equations of the
parabolic, hyperbolic and elliptic types together with some understanding of the
xvi
PREFACE
method of separation of variables (called Bernoulli’s separation method in this book)
and integral transforms will be able to go through the first nine chapters, in part or
whole, depending on their interest and academic requirements. It is assumed that such
readers are at the level of graduate students in mathematics, applied mathematics,
or physics. Some portions of the book, if found too demanding, can be omitted,
depending on the discretion of the reader or instructor. Books dealing with the kind
of broad subject as this one cater to the needs of students and readers from different
branches of science, engineering and technology. The simplest rule to get full use of
such a book is to decide on the requirements and choose the relevant topics.
The author takes this opportunity to thank Mr. Robert B. Stern, Executive Editor,
CRC Press/Taylor & Francis Group, for encouragement to complete this book; his
editorial staff for a very efficient job and cooperation; the two reviewers who took
time to communicate some valuable suggestions; and finally my friend Michael R.
Schäferkotter for help and advice freely given whenever needed.
Shreveport, LA
Northfield, MN
Notations and Definitions
A list of the notations, definitions, and abbreviations used in this book is given below.
a , thermal conductivity
2
a0 =
= 0.529 Å, first Bohr radius
mc2
a.e. , almost everywhere
[aij ], elements of an n × n matrix A
a0 , an , bn , Fourier coefficients, n = 1, 2, . . .
arg{z} , argument of a complex number z
A\B , complement of a set B with respect to a set A
A × B , product of the sets A and B
A , closure of a set A
A = [aij ] , 1 ≤ i ≤ m, 1 ≤ j ≤ n, or A = [c1 | c2 | · · · |cj | · · · |cn ], a matrix with
elements aij , or a matrix A where cj , i ≤ j ≤ n, denotes the jth column of A
AT , transpose of a matrix A
Å , Ångström, angstrom unit is a unit of length equal to 0.1 nanometre or a × 10−10
metres; named after Anders Jonas Ångström (1814-1874), a Swedish physicist;
1 Å = 100.00 × 10−12 m or 0.1 nm
= 328.10−12 ft = 3.9370 × 10−9 in (US/Imperial units)
B(r, x0 ) , open ball of radius r centered at a point x0
B(ε, x0 ) , neighborhood of a point x0 ∈ R, or ε < x − x0 < ε, for an arbitrary small
ε>0
B(r, a) , open disk of radius r and center at a in C
B(r, a) , closed disk of radius r and center at a in C
B(1, 0) ≡ U , open unit disk in C
B1 , B2 , linear differential operators of order ≤ 1 defining boundary conditions
c , wave speed, or wave velocity
const , constant
cof (Aji ), cofactors of an n × n matrix A
[c1 |c2 | · · · cj | · · · cn ], representation of an n × n matrix A, where cj , 1 ≤ j ≤ n
denote the columns of the matrix A
xvii
xviii
NOTATIONS AND DEFINITIONS
C , capacity of a conductor; cross-section of a pipe or channel
C k (D) , class of real-valued functions continuous together with all the derivatives
up to order k inclusive, 0 ≤ k < ∞, in a domain D ∈ Rn .
∞
C (D) , class of functions infinitely differentiable in D, i.e., their continuous partial
derivatives of all orders exist
∞
C0 (D) , class of functions which are infinitely differentiable on D and vanish outside
some bounded region (index 0 indicates compact support)
C0∞ (R) , class of functions with compact support on R and infinitely differentiable
on R
C , complex plane
det [A], determinant of a square matrix A
d(x, y) , metric or distance on R
ds , line element
dS , surface element
dx = dx dy dz , volume element in R3
D , domain; also, differential operator d/dx
D̄ = D ∪ ∂D , closure of a domain D
dn
dn
Dn ≡ n or n , nth derivative with respect to x or t
dx
dt
∂p
p
Dt ≡ p , 0 < p ≤ 1 , Caputo time derivative
∂t
Dβ [f (x)] , Riemann-Liouville derivative of order β, 0 ≤ β ≤ 2
1/D , inverse operator of D
∂D , boundary of a domain D
D , class of admissible functions in C0∞ (R) which have compact support over the
intervals of the form [−a, a] and which approach the Dirac δ-function (or δdistribution) as a → 0; that is, the space of all distributions on C0∞ (Rn )
e.g. , for example
x(Latin, exempli garatia)
2
2
erf(x) = √
e−t dt , error function
π 0
∞
2
2
erfc(x) = 1 − erf(x) = √
e−t dt , complementary error function
π x
E , modulus of elasticity; kinetic energy; total energy
Ext(Γ) , interior of a simple closed contour Γ
Ejn , energy levels of the Hydrogen atom
EI , flexural rigidity of a beam
E(x, y, z) = u(x, y, z) eikz , harmonic wave function
E(r) = {Ψ(r)} , radial wave, where Ψ(r) is a complex wave
EN (r) , radial wave for N slits
∞
zm
Ep,q (z) =
, p, q > 0 , Mittag-Leffler function
m=0 Γ(pm + q)
Eq(s) , equation(s) (when followed by an equation number)
NOTATIONS AND DEFINITIONS
xix
E , electric field
Fig. (Figs.) , abbreviation for Figure (Figures), followed by numerical tag(s)
F (s) = L{f (t)} , Laplace transform of f (t)
1/2
f , norm of f , defined by f, f
{fn (x)} , a sequence of real-valued functions
fs (n) , fs (n, y) , finite Fourier sine transform
∞
nπx
2
fs (n) sin
, inverse finite Fourier sine transform
a 1
a
fc (n) , fc (n, y) , finite Fourier cosine transform
∞
nπx
fc (0) 2 +
fc (n) cos
, inverse finite Fourier sine transform
a
a 1
a
f ◦ g , composition of functions f and g such that (f ◦ g)(x) = f (g(x))
F , field
Fc {f (x)} ≡ fc (α) , Fourier cosine transform
F −1 {fc (α)} ≡ f (x) , inverse Fourier cosine transform
Fs {f (x)} ≡ fs (α) , Fourier sine transform
F −1 {fs (α)} ≡ f (x) , inverse Fourier sine transform
F{f (x)} ≡ f(α) , Fourier complex trannsform
F −1 {f(α)} ≡ f (x) , inverse Fourier complex transform
g , electric charge; total distribution charge density
1
2
gε (x) = √
e−x /(2ε) , Gaussian functions, or Gaussian distribution
2πε
t
t
F + G = 0 f (t − u)g(u) du = 0 f (u)g(t − u) du , convolution of F and G
g(x, s) , one-sided Green’s function for initial value problems
gij (x, s) , one-sided Green’s function for a system of n first-order ordinary differential
equations
G(t, t ), or G(x, x ), also G(t, s) or G(x, s) , Green’s function
G(x, x ) , Green’s function, also written as G(x − x )
G(x, x ; t, t ) ≡ G(x − x ; t − t ) , Green’s function for a space-time operator
h , Planck’s constant, h = 2π = 1.054 × 10−34 joules-sec, or = 6.625 × 10−27 erg-sec , Planck’s constant
π (1)
(1)
hn (x) =
H
(x) , spherical Bessel function of first kind and order n
2x n+1/2
H(t) or H(x) , Heaviside unit step function
H(z), or Hpq , Fox H-function H(z), introduced by Fox [1961], is defined as
(a1 , α1 ), . . . , (ap , αp )
m,n
H(z) = Hp,q z
(b1 , β1 ), . . . , (bp , βp )
m
n
Γ(1 − aj + αj s)
1
j=1 Γ(bj − βj s)
j=1
=
,
2 i π C pj=n+1 Γ(aj + αj s) qj=m+1 Γ(1 − bj + βj s)
xx
NOTATIONS AND DEFINITIONS
where 0 ≤ n ≤ p, 0 ≤ m ≤ q; αj , βj > 0, s complex, and aj , bj are complex
numbers such that no pole of Γ(bj − βj s) for j = 1, . . . , m coincides with any
pole of Γ(1 − aj + αj s) for j = 1, . . . , n; and C is a contour in the complex
bj + k
aj − 1 − k
s-plane from γ − i ∞ to γ + i ∞ such that
and
lie to the left
βj
αj
and right of C, respectively (Prudnikov et al. [1990:626]).
n
2 d
2
Hn (x) = (−1)n ex
e−x , Hermite polynomials of degree n
n
dx
(1)
(2)
H0 (kr), H0 (kr) , Hankel functions of the first and second kind, respectively, and
(1,2)
of order n; H0 (kr) = J0 (kr) ± i Y0 (kr)
20
H12
, see H(z) Fox H-function
Hn {f (x)} ≡ fˆn (σ) , Hankel transform of order n
H0 {f (x)} ≡ fˆ0 (σ) , zero-order Hankel transform
Hn−1 {fˆn (σ)} ≡ f (x) , inverse Hankel transform of order n
2 d2
2 d2
2 2
2
H=−
+
2π
ν
mt
=
−
+ 12 mω 2 t2 , Hamiltonian operator
2m dt2
2m dt2
i.e. , that is (Latin id est)
iff , if and only if
i, j, k, unit vectors along the rectangular coordinates axes x, y, and z, respectively
I , moment of inertia; current intensity
Int(Γ) , interior of a simple closed contour Γ
I(x) , intensity of a wave
I0 (z) , modified Bessel function of the first kind and of order zero
In (x) , modified
Besselfunction of the first kind and order n, defined by In (x) =
e−i nπ/2 Jn ei π/2 x
I(z, z0 ) , index or winding number of a simple closed contour with respect to a point
z0 in the complex plane
, imaginary part of a complex quantity
j , total angular momentum eigenvalue
∂x ∂x ∂x
∂u ∂v ∂w
(x, y, z)
∂(x, y, z)
∂y ∂y ∂y
J , Jacobian, defined by J
=
=
(u, v, w)
∂(u, v, w)
∂u ∂v ∂w
∂z ∂z ∂z
∂u ∂v ∂w
Jn (x) , Bessel function of first kind and order n = 0, 1, 2, . . .
k , thermal diffusivity; spring constant
K0 (z) , modified Bessel function of the third kind and of order zero
Kn (x) , modified Bessel function of the third kind and order n, defined by
π [I−n (x) − In (x)]
∞2 sin nπ
l2 = {x ∈ X : i=1 |xi |2 < ∞}, vector space
log z = ln |z| + i arg{z} , (multiple-valued) logarithm function in C
Kn (x) =
NOTATIONS AND DEFINITIONS
xxi
L1 ([a, b]), vector space of first-order integrable functions on R
L2 ([a, b]), vector (Hilbert) space of square integrable functions on R or C
Lp ([a, b]), p ≥ 1, vector space of p-order integrable functions on R
L , linear differential operator; induction coefficient of a conductor
L∗ , adjointoperator to a differential operator L
d
du
L[u] ≡
p(x)
+ q(x)u, a < x < b , Sturm-Liouville operator
dx
dx
∂
L D,
, transient operator
∂t
−1
L , inverse (integral) operator
Ln (x) , Laguerre polynomials of order n
Lm
n (x) , associated Laguerre polynomials of order n and degree m
Ln (t, φ, t1 , φ1 ) , Laplace’s coefficients, t = cos θ
L{f (t)} ≡ F (s) , Laplace transform
L−1 {F (s)} ≡ f (t) , inverse Laplace transform
n , outward normal perpendicular to the boundary of a curve or surface; azimuthal
(orbital angular) quantum number
nx , ny , nz , components of the outward normal n along x, y, and z axis, respectively
N (z, z ) , Neumann’s function
N = Z+ , set of natural numbers
m , mass; mass of a quantum particle; magnetic quantum number
p.v. , Cauchy’s principal value of an integral
∂u
p , partial derivative ux , or
; also, pressure
∂x
P (D) , ordinary differential operator, defined by a polynomial of degree n of the
form P (D) = a0 (x) D n + a1 (x) Dn−1 + · · · + an−1 (x) D + an (x), a0 (x) = 0
P (r) , probability function
P (x) , gravitational potential
Pn (x) , Legendre polynomials of degree n = 0, 1, 2, . . . ; Legendre’s coefficients, or
surface zonal harmonics
Pn (cos θ) , Laplace’s coefficient of degree n
Pnm (cos θ) or Pnm (t) , associated Legendre polynomials of order m = 0, ±1, ±2, . . .
and degree n = 0, 1, 2, . . .
∂u
q , partial derivative uy , or
∂y
Qn (x) , Legendre function of the second kind of order n
∂2u
r , partial derivative uxx , or
; also, radial axis
∂x2
y
(r, θ) , polar coordinates: x = r cos θ, y = r sin θ, r = x2 + y 2 , θ = arctan
x
(r, θ, z) , polar cylindrical coordinates: x = r cos θ, y = r sin θ, r = x2 + y 2 , θ =
y
arctan , z = z
x
xxii
NOTATIONS AND DEFINITIONS
(r, θ, φ) , spherical coordinates: x = r sin θ cos φ, y = r sin θ sin φ, z = r cos θ,
z
y
r = x2 + y 2 + z 2 , θ = arccos , φ = arctan
r
x
r n Pn (cos θ) or r −n−1 Pn (cos θ) , solid zonal harmonics
r n Ynm (t, φ) or r−n−1 Ynm (t, φ) , solid spherical harmonics of degree n
|r| , average probability distance for the hydrogen atom
R = {0 < x < a, 0 < y < b} , rectangle
Rm
n (r) , radial wave function
R , real line
Rn , Euclidean n-space; R1 ≡ R
R+ , set of positive real numbers
, real part of a complex quantity
s , variable of the Laplace transform; spin quantum number
s(x, x ) , singularity
function

1,
x
>
0,


sgn(x) = 0, x = 0,
signum function


−1, x < 0,
sin x
sinc(x) =
, sinc function
x
supp f , support of a continuous function f
∞
S=
sn , sum of an infinite sequence {sn }
n=1
Sn = s1 + s2 + · · · + sn , partial sum of an infinite sequence {sn }
S(r, x0 ) boundary (surface) of the open ball B(r, x0 ), or {x, |x − x0 | = r}
2π n/2
Sn (1) =
, surface area of the unit ball B(1, 0)in Rn
Γ(n/2)
∂ 2u
t , partial derivative uyy , or
; also, time
∂y 2
t , source point ; singularity
T , linear transformation; also, kinetic energy
u , dependent variable; displacement; temperature
uc (x) or uc (t) , complementary function for an ordinary differential equation
up (x) or uc (t) , particular integral for an ordinary differential equation
u1,s , u2,s , u3,s , first, second, and third state of radial energy of the hydrogen atom,
respectively
u∗ (x, x ) , fundamental solution, or ‘free-space’ Green’s function, or Green’s function in the large
U ≡ B(1, 0) , open unit disk
V , volume; potential energy
W (t) , Lambert W-function, or Omega function, or Product Log function. It is
the inverse function of f (w) = w ew , where w is a complex number, i.e., z =
W (z) eW (z) ; this function is multiple-valued except at z = 0. If we restrict it to
NOTATIONS AND DEFINITIONS
xxiii
real numbers and keep w real, then the function is defined only for x ≥ −1/e; it
is double-valued on the interval (−1/e, 0). If we impose an additional restriction
of w ≥ −1, then the function is single-valued, and is denoted by W0 (x), where
W0 (0) = 0 and W0 (−1/e) = −1. The other branch on [−1/e, 0] with w ≤ −1
is denoted by W−1 (x) and decreases from W−1 (−1/e) = −1 to W−1 (0− ) =
dW
W (z)
−∞. This function satisfies the differential equation
=
for
dz
z (1 + W (z))
z = −1/e.
u1 (x)
···
un (x)
u1 (x)
···
un (x)
W (x) = W (u1 , . . . , un ) =
, Wronskian
···
···
···
(n−1)
(n−1)
u1
(x) · · · un
(x)
Wj (x) or Wj (t) , determinant obtained from the Wronskian W (x) or W (t) by replacing the jth column by [0 0 · · · 1]T
∞
zn
W (z, p, q) =
, p, q > 0 , Wright function
n=0 n! Γ(pm + q)
{xn } , a sequence of real numbers
{xnk } , a subsequence of real numbers
(x, y, z) , cartesian coordinates
X , metric space
x , a point (x1 , x2 , . . . , xn ) in Rn ; a field point
x , source point; singularity
|x − y| , Euclidean distance between the points x, y ∈ Rn
π
π
yn (x) =
Yn+1/2 (x) = (−1)n
Jn−1/2 (x) , spherical Bessel function of
2x
2x
order n
Y0 (x) , Bessel function of second kind of order zero
Yn (x) , Bessel function of the second kind of order n, defined by
cos nπ Jn (x) − J−n (x)
Yn (x) =
; sometimes denoted by Nn (x)
sin nπ
Ynm (θ, φ) , spherical harmonics
z , a complex number z = x + i y
z ∗ , inverse (symmetrical) point of z with respect to the unit circle in C
|z| = x2 + y 2 , z = x + i y , modulus of a complex number z
Z , set of integers
Z+ , set of positive integers
α , variable of Fourier transform, of Fourier sine and cosine transforms
γ = 0.577215665 , Euler gamma
Γ , simple contour, or path; boundary of a domain
∞
Γ(z) = 0 tα−1 (1 − t)α dt, {z} > 0 , gamma function
δ(x), δ(x, x ), δ(x, s) , Dirac delta function; also denoted by δ(x − x )
δmn , Kronecker delta, equal to 1 if m = n and 0 if m = n
ζ = (α, β, γ) , variable of 3-D Fourier transform
xxiv
NOTATIONS AND DEFINITIONS
α , wavelength
λn , eigenvalues
1
ε
λε (x) =
, Cauchy densities, or Lorentz curves
π x2 + ε2
ν , vibration frequency
φ , phase; latitude (azimuth)
φn , eigenfunctions
φε (t), ‘cap’ function
(λn , φn ) , eigenpairs
ν , vibration frequency
ρ(x) , weight function
ρ(r) , charge density function
σ , variable of zero-order Hankel transform
ς = (α, β) , variable of 2-D Fourier transform
θ , colatitude (polar angle)
ϑ-function , Riemann theta function is defined as
ϑ(z; τ ) = exp 2π i 12 mT τ m + mT z ,
m
where z ∈ Cn is a complex vector, τ ∈ Hn is the Siegel upper half-plane,
and T denotes the transpose. For n = 1, H is the upper half-plane in C. It
is related to the Jacobi theta functions, where for our interest ϑ00 is defined by
∞ n 2
ϑ00 (z; τ ) = ϑ(z; τ ) = ϑ3 (z; q), where ϑ00 (w; q) =
w2 q n , where
n=−∞
w = ei πz is called the argument and q = ei πτ the nome. A useful defini i+∞ i πτ u2
e
cos(2uz + πu)
tion is: ϑ00 (z; τ ) = −i
du. For more details, see
sin(πu)
i−∞
Abramowitz and Stegun [1968: §16.27ff]; Akhizer [1990], Pierpont [1959], and
Dubrovin [1981].
χD (x), characteristic function of a domain D
Ψ(r) , monochromatic complex wave
ω , frequency; angular frequency of vibrations
∂B(r, a) = {z ∈ C : |z − a| = r} , circle of radius r and center a
∂D , boundary of the domain D
1-D, 2-D, 3-D, and 4-D, one-dimensional, two-dimensional, three-dimensional, and
four-dimensional (three space dimensions and one time), respectively
0 = (0, 0, . . . , 0), zero vector, null vector, or origin in Rn
n
1 =(1, 1, . . . , 1), unit vector
in R
n
n!
n
=
,
= 1 , binomial coefficients
k
k! (n − k)!
0
n!
= 1 · 2 · . . . (n − 1)n, 0! = 1! == (−1)! = 1 , factorial n
x, y , inner product of x, y ∈ R
· , norm
NOTATIONS AND DEFINITIONS
b
f, g =
f (x) g(x) dx , inner product of functions f and g
a
, line or surface integral
, double (surface) integral
, triple (volume) integral
∂
∂
∂
+j
+k
∂x
∂y
∂z
∂2
∂2
∂2
∇2 , Laplacian
+
+
∂x2
∂y 2
∂z 2
2
2
∇ + k , Helmholtz operator
∇4 , biharmonic operator, defined by ∇4 u = ∇2 (∇2 u)
∂
− k∇2 , diffusion operator; heat conduction operator
∂t
∂
, partial derivative with respect to n
∂n
∂
∂ ∂
−
+ x , Fokker-Plank operator
∂t ∂x ∂x
2
∂
c ≡ 2 − c2 ∇2 , d’Alembertian, or wave operator; also, ≡ 1
∂t
, end of an example or of a proof
! ! , attention sign
∇ , grad, defined by ∇ = i
xxv
This page intentionally left blank
1
Some Basic Results
In this chapter we discuss some basic definitions and present results which are needed
to study Green’s functions and linear ordinary and partial differential equations.
Proofs of these results can be found in standard textbooks on advanced calculus
and real analysis. The notation used in this book, although standard, is presented
prior to this chapter. Readers familiar with the topics covered in this chapter may still
like to read it; others are advised to study them thoroughly.
1.1. Euclidean Space
A real finite dimensional vector space on which an inner product is defined is called
the Euclidean space, which is denoted by Rn , n = 1, 2, 3, . . . . Let F be a field, let
“+” denote a mapping of Rn × Rn into Rn , and let “·” denote a mapping of F × Rn
into Rn . The elements x ∈ Rn are called vectors, such that x = (x1 , x2 , . . . , xn )
represents a (position) vector of a point with cartesian coordinates (x1 , x2 , . . . , xn ).
The elements of the field F are called scalars, while the operation “+” defined
on Rn is called vector addition and the mapping “·” the scalar multiplication or
multiplication of vectors by scalars. Then for each x, y ∈ Rn there is a unique
element x + y ∈ Rn , called the sum of x and y, and for each x ∈ Rn and α ∈ F
there is a unique element α · x = α x ∈ Rn , called the multiplication of x by
α. The non-empty set Rn and the field F along with the above two mappings of
vector addition and scalar multiplication constitute a vector space or a linear space
if the following axioms are satisfied: (i) x + y = y + x for every x, y ∈ Rn ; (ii)
x + (y + z) = (x + y) + z for every x, y, z ∈ Rn ; (iii) There is a unique vector
in Rn , called the zero vector or the null vector or the origin, denoted by 0 such
that 0 + x = x + 0 = x for all x, y ∈ Rn ; (iv) α(x + y) = αx + αy for every
α ∈ F and for every x, y ∈ Rn ; (v) (α + β)x = αx + βx for all α, β ∈ F and
for every x ∈ Rn ; (vi) (α β)x = α(βx) for all α, β ∈ F and for every x ∈ Rn ;
and (vii) 0 x = x + 0 = 0 for all x ∈ Rn ; (viii) 1 x = x for all every x ∈ Rn ,
where 1 = (1, 1, . . . , 1) denotes the unit vector directed from the origin 0 along the
coordinate axes.
1
2
1. SOME BASIC RESULTS
Since n denotes the dimension of the space, and since we will be mostly dealing
with n = 1, 2, 3, we will use the following notation: We denote R1 simply by R
which is the set of all points on the real axis. The 2-D space R2 is the set of all points
(x, y) in the real plane, and the 3-D space R3 represents the set of all points (x, y, z).
The point 0 represents the origin of coordinates, and the vector 1 is represented by
i, j, k in R3 , which are the unit vectors along the rectangular coordinates axes x, y,
and z, respectively.
We will denote by R+ the set of nonnegative real numbers, by Z the set of integers,
and by N the set of natural numbers. Notice that N = Z+ . Since |x − y| defines
the Euclidean distance between points x and y in Rn , where x = (x1 , . . . , xn ) and
y = (y1 , . . . , yn ), we define an open ball of radius r centered at a point x0 ∈ Rn by
{x : |x − x0 | < r}, and denote it by B(r, x0 ). The boundary (surface) of the open
2π n/2
ball B(r, x0 ) is denoted by S(r, x0 ) = {x : |x − x0 | = r}, where Sn (1) =
Γ(n/2)
is the surface area of the unit ball B(1, 0) in Rn . The ε-neighborhood of a point
x0 ∈ R is defined by B(ε, x0 ), or −ε < x − x0 < ε for an arbitrarily small ε > 0.
The complement of a set B with respect to a set A is denoted by A\B, the product
of the sets A and B by A × B, and the closure of a set A by A.
1.1.1. Metric Space. Since R is a metric space, let d be a real-valued function
such that d : R → R, where d has the following properties: (i) d(x, y) ≥ 0 for
all x, y ∈ R, and d(x, y) = 0 iff x = y; (ii) d(x, y) = d(y, x) for all x, y ∈
R (symmetry property); and (iii) d(x, y) ≤ d(x, z) + d(z, y) for all x, y, z ∈ R
(triangular inequality). The function d is called a metric (or distance) on R, and
d(x, y) = |x − y| for all x, y ∈ R; obviously, d(x, y) = |x − y| = 0 iff x = y, and
d(x, y) = |x − y| = |(x − z) + (z − y)| ≤ |x − z| + |z − y| = d(x, y) + d(z, y) for
all x, y, z ∈ R. These properties also hold on the space Rn
1.1.2. Inner Product. For every
x,
y ∈ R the inner product
x, y possesses
the following
three properties: (i) x, y > 0 for all x= 0 and
x,
y = 0 ifx = 0;
(ii) x, y = y, x for all x, y ∈ R; (iii) α x + βy, z = α x, z + β x, y for all
x,
for all α, β ∈ F . The inner product for vectors in Rn is defined by
y, z ∈ R
and
n
x, y = i=1 xi yi for any vectors x = (x1 , . . . , xn ) and y = (y1 , . . . , yn ) in Rn .
For real-valued functions f (x) and g(x) defined on an interval (a, b) ∈ R,
where
a
can be −∞ and b can be ∞, the inner product of f and g is denoted by f, g and
defined by
b
f, g =
f (x) g(x) dx.
(1.1)
a
The weighted inner product of f and g with weight function ρ > 0 on the interval
(a, b) is defined by
b
f, g ρ =
f (x)g(x)ρ(x) dx.
(1.2)
a
1.3. CONVERGENCE
3
1.2. Classes of Continuous Functions
A real- (or complex-) valued function f is said to belong to the class C k (D) if it
is continuous together with all the derivatives up to order k inclusive, 0 ≤ k < ∞,
in a domain D ∈ Rn . The functions f in the class C k (D) which admit continuous
continuations in the closure D̄ = D ∪ ∂D, where ∂D denotes the boundary of the
domain D, form the class of functions C k (D̄). The class C ∞ (D) consists of functions
f which are infinitely differentiable in D, i.e., their continuous partial derivatives
of all orders exist. These classes are linear sets; thus, every linear combination
λf + µg, where λ and µ are arbitrary real or complex numbers, also belongs to this
class. Further, a function defined on D ⊂ R is said to belong to the class C0∞ (D)
if it is infinitely differentiable on D and vanishes outside some bounded region.
The support of a continuous function f (written supp f ) is the closure of the set
{x ∈ D : f (x) = 0}. Then the class C0k (D) denotes the set of functions in C k (D)
that have compact support, where the index 0 indicates compact support. A function
is said to be of compact support if it is equal to zero outside a given bounded set in
its domain. A function f with compact support on R is said to belong to the class
C0∞ (R) if it is infinitely differentiable on R.
1.3. Convergence
Various results on convergence of sequences and infinite series in R are discussed.
1.3.1. Convergence of Sequences. A sequence {xn } in a set X ⊆ R is a
function f : Z+ → X. Thus, if {xn } is a sequence in X, then f (n) = xn for each
n ∈ Z+ . Let d (x, xn ) denote the distance (or metric) between x and xn . If {xn } is
a sequence of points in X, and if x is a point of X, then the sequence {xn } is said
to converge to x if for every ε > 0 there is an integer N such that for all n ≥ N ,
d (x, xn ) < ε, i.e., xn belong to the ε-neighborhood of the point x for all n ≥ N . In
general, N depends on ε, i.e., N = N (ε), and we write lim xn = x, or alternatively,
n
xn → x as n → ∞. If there is no x ∈ X to which the sequence converges, then we
say that the sequence {xn } diverges. Further, if the range of f (n) is bounded, then the
sequence {xn } is said to be bounded. In this definition the range of f (n) may consists
of a finite number or an infinite number of points. If the range of f consists of one
point, then we say that the sequence is a constant sequence, Obviously, all constant
sequences are convergent. Let {xn } be a sequence in X, and let n1 , n2 , . . . , nk , . . .
be a sequence of positive integers which is strictly increasing, i.e., nj > nk for all
j > k. Then the sequence {xnk } is called a subsequence of {xn }. If the subsequence
{xnk } converges, then its limit is called a subsequential limit of {xn }. Let {xn } be
a sequence in X. Then:
(i) There is at most one point x ∈ X such that lim xn = x ;
n
(ii) If {xn } is convergent, then it is bounded;
(iii) {xn } converges to a point x ∈ X iff every ball (neighborhood) about x contains
4
1. SOME BASIC RESULTS
all but a finite number of terms in {xn };
(iv) {xn } converges to a point x ∈ X iff every subsequence of {xn } converges to x;
(v) If {xn } converges to x ∈ X and if y ∈ X, then lim d (xn , y) = d(x, y);
n
(vi) If {xn } converges to x ∈ X and if the sequence {yn } of X converges to y ∈ X,
the lim d (xn , yn ) = d(x, y); and
n
(vii) If {xn } converges to x ∈ X, and if there is a y ∈ X and a c > 0 such that
d (xn , y) ≤ c for all n ∈ Z+ , then d(x, y) ≤ c.
A sequence {xn } ∈ X is said to be a Cauchy sequence if for every ε > 0
there is an integer N such that d (xn , xm ) < ε whenever n, m ≥ N . Then (i) every
convergent sequence in a metric space is a Cauchy sequence; (ii) if {xn } is a Cauchy
sequence, then {xn } is a bounded sequence; (iii) if a Cauchy sequence {xn } contains
a convergent subsequence {xm }, then the sequence {xn } is convergent.
A function f : X → Y , where X, Y ⊆ R, is continuous at a point x0 ∈ X iff for
every sequence {xn } of points in X which converges to x0 the corresponding
sequence
{f (xn )} converges to the point f (x0 ) in Y ; i.e., lim f (x0 ) = f
n→∞
lim xn
n→∞
=
f (x0 ) whenever lim xn = x0 . Let f be a mapping from X into Y . Then (i) f
n→∞
is continuous on X iff the inverse image of each open subset of {Y, dy } is open in
{X, dx }; and (ii) f is continuous on X iff the inverse image of each closed subset of
{Y, dy } is closed in {X, dx }. Moreover, let f be a mapping from X into Y , and let
g be a mapping from Y into Z. If f is continuous on X and g is continuous on Y ,
then the composite mapping f ◦ g of X into Z is continuous on X.
Let X, Y ⊆ R, and let {fn } be a sequence of functions from X into Y . If
{fn (x)} converges at each x ∈ X, then we say that {fn } is pointwise convergent,
and write: lim fn = f , where f is defined for every x ∈ X. In other words, we say
n
that the sequence {fn } is pointwise convergent to a function f if for every ε > 0
and every x ∈ X there is an integer N = N (ε, x) such that dy (fn (x), f (x)) < ε
whenever n ≥ N (ε, x). In general, N (ε, x) is not necessarily bounded. However,
if N (ε, x) is bounded for all x ∈ X, then we say that the sequence {fn } converges
to f uniformly on X. Equivalently, let M (ε) = sup N (ε, x) < ∞. Then we say
x∈X
that the sequence {fn } converges uniformly to f on X if for every ε > 0 there is an
M (ε) such that dy (fn (x), f (x)) < ε whenever n ≥ M (ε) for all x ∈ X. Further, if
the sequence {fn } converges uniformly to f on X, then f is continuous on X. Also,
if f is continuous on X and if Z is a compact subset of X, then (i) f is uniformly
continuous on Z; (ii) f is bounded on Z; and (iii) If Z = ∅, f attains its infimum and
supremum on Z; i.e., there exists x0 , x1 ∈ Z such that f (x0 ) = inf{f (x : x ∈ Z}
and f (x1 ) = sup{f (x) : x ∈ Z}.
1.3.2. Weak Convergence. A sequence {xn } of elements in X is said to
converge
weakly
to an element x ∈ X if for every x ∈ X the inner product
xn , x → x, x . If the sequence {xn } converges to x ∈ X , i.e., if xn − x → 0
1.3. CONVERGENCE
5
as n → ∞, then we call this convergence strong convergence, or the sequence {xn }
convergences in the norm, to distinguish it from weak convergence. Let {xn } be a
sequence in X which converges in the norm to x ∈ X. Then {xn } converges weakly
to x. This result states that convergence in the norm (or strong convergence) implies
weak convergence. Note that every weakly convergent sequence in Rn is convergent.
Example 1.1. For each n ∈ N, define fn on [0, 2π] by fn (x) = cos nx. Then
the sequence {fn } ⊂ Lp ([0, 2π]) for each p ≥ 1. The Riemann-Lebesgue lemma1
shows that fn → 0 weakly for all p ≥ 1.
Example 1.2. The vector space Lp ([a, b]) of p-order integrable functions,
p ≥ 1, consists of all continuous functions f and g on the interval [a, b] with metric
b
1/p
defined by d(f, g) =
|f (x) − g(x)|p dx
. For p = 1 it represents the space
a
of first-order integrable function and for p = 2 of the square-integrable functions on
1 if x ∈ D,
the interval [a, b]. Let D be a subset of R, and χD (x) =
be the
0 if x ∈ R\D,
characteristic function of D. Take fn = n χ[0,1/n] . Then {fn } ⊂ Lp ([0, 1]) for all
p ≥ 1, and fn → 0 a.e., but fn → 0 weakly.
1.3.3. Metric. Let X = Rn , and define the vector space l2 = x ∈ X :
∞
2
i=1 |xi | < ∞ . If y ∈ l2 is defined by y = (d1 , d2 , . . . , dn , . . .), then the metric
1/2
∞
2
for this space is defined by d(x, y) =
for every x, y ∈ l2 .
i=1 |ci − di |
Example 1.3. Consider the sequence {xn } defined in the vector space l2 by the
vectors x1 = (1, 0, . . . , 0, . . . ), x2 = (0, 1, 0, . . . , 0, . . . ), x3 = (0, 0, 1, . . . , 0, . . . ),
. . . . To show that this sequence converges weakly, note that if x = (c1 , c2 , . . . , cn . . .),
then every x ∈ l2 can be represented as the inner product with some fixed vector
∞
y = (d1 , d2 , . . . , dn , . . .), i.e., x, x =
ci di . For our sequence {xn } we
i=1
have
xn , x = dn , and since dn → 0 as n → ∞ for every y ∈ l2 , we find that
xn , x → 0 as n → ∞ for every x ∈ l2 . Thus, {xn } converges to 0 weakly.
However, xn → 0 strongly, because xn = 1.
1.3.4. Convergence of Infinite Series. The sum of an infinite sequence {sn },
denoted by
∞
S=
sn = s1 + s2 + . . . + sn + . . . ,
(1.3)
n=1
is called an infinite series. With each series there is associated a sequence of partial
sums: Sn = s1 + . . . + sn . If lim Sn = S is a finite number, then the series (1.3)
n→+∞
is said to converge and S is called its sum. If lim Sn does not exist, the series
n→+∞
1
This lemma states that if fn ∈ L1 ([0, 2π]), then
lim fn (x) = 0.
|n|→∞
6
1. SOME BASIC RESULTS
(1.3) is said to diverge. There are two cases in which a series diverges:
(a) lim Sn = ∞, or (b) as n increases, Sn increases and decreases (i.e., oscillates)
n→+∞
without reaching a limit. For convergence we have the following useful results:
(i) A convergent (divergent) series remains convergent (convergent) even after the
removal or alteration of any or all of its first n terms.
(ii) The sum of a convergent series is unique.
(iii)If sn converges to S,
then ksn , where k is any constant, converges to
kS; if sn diverges, so does ksn .
(iv) If sn converges, then lim sn = 0. The converse is not true; for example,
for the harmonic series 1 +
the series diverges.
(v) If lim sn = 0, then
n→∞
from the harmonic series.
1
2
n→∞
1
+ 14
3
+
+ ··· +
1
n
+ · · · we have lim sn = 0, but
n→∞
sn diverges. The converse is not true, as is obvious
1.3.5. Tests for Convergence of Positive Series. A series
sn , all of
whose terms are positive, is called a positive series. A positive series is convergent if
the sequence {Sn } of its partial sums is bounded. The following tests for convergence
of positive series are based on the fact that the sequences of their partial sums is always
nondecreasing.
(a) Integral Test. Let f (n) denote the general terms sn of a positive series
sn . If f (x) > 0 and never increases on the interval x > x0 , wherex0 is some
∞
positive integer, then the series sn converges or diverges according as x0 f (x) dx
exists or does not exist.
(b) Comparison Test for Convergence. A positive series sn is convergent if each term (perhaps, after a finite number of terms) is
less than or equal to
the corresponding term of a known convergent positive series cn .
(c) Comparison Test for Divergence. A positive series sn is divergent
if each term (perhaps, after a finite number of terms) isequal or greater than the
corresponding term of a known divergent positive series dn .
sn+1
(d) Ratio Test. A positive series
sn converges if lim
< 1, and
n→∞ sn
sn+1
sn+1
diverges if lim
> 1. If lim
= 1, the test fails to indicate either
n→∞ sn
n→∞ sn
convergence or divergence.
1.4. Functionals
Let f, φ =
f (x)φ(x) dx denote a real number associated with each x ∈ Rn
Rn
for every test function φ ∈ C0∞ (Rn ). Then f is said to be a functional on Rn . For
1.4. FUNCTIONALS
7
example, the Fourier series of f ∈ C 1 [0, π], defined by
f (x) =
∞
bn sin nx,
where bn =
n=1
2
π
π
f (x) sin nx dx,
(1.4)
0
is a functional on R, with test functions in the set {sin x, sin 2x, · · · }. Some useful
properties of test functions are:
(i) If φ1 (x) and φ2 (x) are test functions on Rn , so is their linear combination
c1 φ1 (x) + c2 φ2 (x), where c1 and c2 are real numbers;
(ii) If φ(x) ∈ C0∞ (Rn ), so do all partial derivatives of φ(x) belong to the class
C0∞ (Rn );
(iii) If φ(x) ∈ C0∞ (Rn ) and α(x) is infinitely differentiable, then the product
α(x)φ(x) belongs to the class C ∞ (Rn ); and
(iv) If φ (x1 , · · · , xm ) ∈ C0∞ (Rm ) and ψ (xm+1 , · · · , xn ) ∈ C0∞ (Rn−m ), then
φ (x1 , · · · , xm ) ψ (xm+1 , · · · , xn ) ∈ C0∞ (Rn ).
A functional f on Rn is said to be linear if f, λφ1 + µφ2 = λ f, φ1 + µ f, φ2
for all real numbers λ, µ and all φ1,2 ∈ C0∞ (Rn ). Note that f, 0 = 0, and
m
m
f,
αn φn =
αn f, φn .
n=1
n=1
A linear
functional
f on C0∞ (Rn ) is said to be continuous if the numerical se
quence f, φm → 0 as m → ∞, where {φm (x)} is a null sequence in C ∞ (Rn ),
i.e., supp φm , m = 1, 2, · · · , is contained in a sufficiently large ball, and
lim max |Dk φm (x)| = 0
m→∞ x∈Rn
for every multi-index k, |k| ≤ n.
∞
n
A continuous
linear functional f on C0 (R ) is said to be a distribution. The
number f, φ is called the value of f at φ, or the action of f on φ. The space D of all
distributions on C0∞ (Rn ) is a linear space. A locally integrable function f (x) ∈ Rn
generates an n-dimensional distribution f such that for all φ ∈ C0∞ (Rn )
f, φ =
f (x)φ(x) dx
n
R∞
∞
(1.5)
=
···
f (x1 , · · · , xn ) φ (x1 , · · · , xn ) dx1 · · · dxn .
−∞
−∞
Hence, every locally integrable function f can be regarded as a distribution.
Let f1 (x) and f2 (x) be two different continuous functions. Then eachgenerates
∞
n
a different
distribution
such
that there exists a φ in C0 (R ) for which f1 , φ =
f2 , φ , i.e., f1 − f2 , φ = 0. Two functions f
1 and f2 are said to be equal almost
|f1 − f2 | dx = 0. Hence, two
everywhere (a.e.) on a bounded domain Ω if
Ω
8
1. SOME BASIC RESULTS
locally integrable functions which are equal a.e. generate the same distribution. A
distribution of the form (1.5), where f (x) is locally integrable, is said to be regular.
All other distributions are called singular, although formula (1.5) can be used formally
for such distributions. Note that if f ∈ D, then the distribution f vanishes in any
region lying outside supp f , i.e.,
f, φ = 0, supp f ∩ supp φ = ∅.
(1.6)
1.4.1. Examples of Linear Functionals. Some examples of linear functionals
are as follows.
Example 1.4. Consider the space of functions C[a, b]. Then the mapping
b
f1 (x) = a x(s) ds, x ∈ C[a, b] is a linear functional on C[a, b]. Also, the function
defined by f2 (x) = x(s0 ), x ∈ C[a, b], s0 ∈ [a, b], is a linear functional on C[a, b],
b
and so is the function f3 (x) = a x(s) x0 (s) ds, where x0 is a fixed element of C[a, b]
and x any element in C[a, b].
Example 1.5. Let X be a vector space over a field F . Let X = F n , and denote
x ∈ X by x = {x1 , . . . , xn }. The mapping f4 (x) = x is a linear functional on X.
Also, let α = {α1 , . . . , αn } ∈ X be fixed and let x = {x1 , . . . , xn } be an arbitrary
n
element of X. Then the function f5 (x) =
α i xi is a linear functional on X.
i=1
1.5. Linear Transformations
A mapping T of a linear space X into a linear space Y , where X and Y are vector
spaces over some field F , is called a linear transformation or linear operator if
(i) T (x + y) = T (x) + t(y) for all x, y ∈ X; and
(ii) T (αx) = α T (x) for all x ∈ X and for all α ∈ F .
Often we find it convenient to write T ∈ L(X, Y ) to specify that T is a linear
transformation from a linear space X into a linear space Y . This means that L(X, Y )
denotes the set of all linear transformations from a linear space X into a linear space
Y . This definition implies
transformation from a linear space X into
n that T is a linear
n
a linear space Y iff T
α i xi =
αi T (xi ) for all xi ∈ X and for all αi ∈ F ,
i=1
i=1
i = 1, . . . , n. This result is also known as the principle of superposition. We cite
some examples of linear transformations.
Example 1.6. Let X = Y denote the space of real-valued function of the class
t
C[a, b]. Let T : X → Y be defined by T [x](t) = a x(s) ds, a ≤ t ≤ b, where
integration is in the Riemann sense. Then T is a linear transformation.
Example 1.7. Let X denote the set of function x(t) in the class C[a, b], and
let vector addition and scalar multiplication be defined respectively by (x + y)(t) =
dx(t)
x(t) + y(t) and (αx)(t) = αx(t) for all t ∈ [a, b]. Then [T ]x(t) =
is a linear
dt
1.6. CRAMER’S RULE
9
transformation.
Example 1.8. Let X denote the space of all complex-valued function f (t)
defined on the interval [0, ∞) such that f (t) is Riemann-integrable and such that
lim |f (t)| < k eαt , where k is some positive constant and α any real number. Let
t→∞
vector addition and scalar multiplication be defined as in Example 1.7. Let Y denote a
linear space of complex functions of a complex variable s = σ+iω. Then the mapping
∞
T : X → Y defined by [T ]f (s) = 0 f (t) e−st dt is a linear transformation, and is
called the Laplace transform of f (t) (see Chapter 5).
Example 1.9. Let X be the space of real-valued functions f (t) ∈ C[a, b]. Let
k(s, t) be a real-valued function defined for a ≤ s ≤ b, a ≤ t ≤ b, such that for each
b
x ∈ X the Riemann integral a k(s, t) f (t) dt exists and defines a continuous function
b
of s on [a, b]. Let T1 : X → X be defined by [T1 ]f (s) = y(s) = a k(s, t) f (t) dt.
Then T1 is a linear transformation, and this equation is called the Fredholm integral
equation of the first kind.
1.6. Cramer’s Rule
An n × n matrix A is said to be nonsingular (or invertible) iff there exists an n × n
matrix B such that AB = I, where I is the n × n identity matrix. Consider a system
of m linear algebraic equations in n unknowns:
a11 x1 + · · · + a1n xn = b1 ,
a21 x1 + · · · + a2n xn = b2 ,
..
.
(1.7)
am1 x1 + · · · + amn xn = bm ,
where the coefficients aij , 1 ≤ i ≤ m, 1 ≤ j ≤ n, and right-hand side bi , 1 ≤ i ≤ m,
are all given real (or complex) numbers. Here the index i in the coefficients aij refers
to the row and the index j the column in which the coefficient occurs. In matrix
notation the system (1.7) is represented by Ax = b, where the left side is
  n a x 


a11 · · · a1n  x1  
 j=1 1j j 

..
..
.. 
..
..

Ax =
=
.
(1.8)
.
.
.
.
 .  

 n

an1 · · · amn
xm
j=1 amj xj
The matrix A is also represented by A = [aij ], or by A = [c1 |c2 | · · · cj | · · · |cn ],
where cj , 1 ≤ j ≤ n, denotes the jth column of A.
Let A = [aij ] be an n × n matrix. If the ith row and jth column of A are deleted,
the remaining (n − 1) rows and (n − 1) columns form another matrix Mij , called
i+j
the
nminor of aij . The cofactor of [a
ijn] is defined as cij = (−1) det[M ], such that
i=1 aij cik = 0 for j = k, and
i=1 aij ckj = 0 for i = k. These relations can
10
1. SOME BASIC RESULTS
n
n
also be written as i=1 aij cik = det[A] δjk and i=1 aij cik = det[A] δik , where
δik is the Kronecker delta.
If an n×n matrix A is invertible, the associated matrix operator A, represented by
(1.8), has its precise range, and the matrix equation Ax = b has the solution x = Bb,
where the n × n matrix B whose ith column is bi has the property AB = I. The
matrix B is called the inverse of the matrix A and is denoted by A−1 . Note that
(i) A−1 A = AA−1 = I; (ii) A−1 is nonsingular (invertible), and (iii) (AB)−1 =
B −1 A−1 .
The solution of an n × n linear system Ax = b exists and is unique when A is
nonsingular; it is given by x = A−1 b. If b = 0, then the equation Ax = 0 has only
the trivial solution x = 0. When A is singular, the equation Ax = b cannot be solved
for all b, and if a solution exists, it is not unique; but in this case the equation Ax = 0
has a nontrivial solution. For a nonsingular matrix A the inverse A−1 is given by
1
A−1 =
AT , det[A] = 0,
(1.9)
det[A]
where AT denotes the transpose of A.
Creamer’s Rule. Let A be an n × n matrix with det[A] = 0, and denote the
columns of A by c1 , c2 , . . . , cn . Then the unique solution of the equation Ax = b is
xi =
det [c1 | · · · |b | · · · |cn ]
,
det[A]
i = 1, 2, . . . , n.
(1.10)
Proof. Since x = A−1 b, let the entries of A−1 be denoted by dij so that A−1 =
cofAji
[dij ]. Then by (1.9) we get dij =
, where cofAji denotes the cofactor of Aji .
det[A]
Hence, for each j = 1, 2, . . . , n, we have
xi =
n
cofAji
j=1
det[A]
bj =
det [c1 | · · · |b | · · · |cn ]
.
det[A]
1.7. Green’s Identities
Let D be a finite domain in Rn , n = 1, 2, 3, bounded by a piecewise smooth, orientable
surface (or boundary) ∂D, and let F be a scalar function and G a vector function in
the class C 3 (D). Then
Gradient theorem:
∇F dD =
n F dS,
D
∂D
∇ · G dD =
Divergence theorem:
D
n · G dS,
∂D
∇ × G dD =
Stokes’s theorem:
D
G · t dS,
∂D
1.7. GREEN’S IDENTITIES
11
where nis the outward normal to the surface ∂D, t is the tangent vector at a point
on ∂D,
denotes the surface or line integral, and dS (or ds) denotes the surface
(or line) element depending on the dimension of D. The divergence theorem in
the above form is also known as the Gauss theorem. This theorem implies that
∇ · [G ∇h − h∇g] = G∇2 h − h∇2 G = G∇2 h, since ∇2 G = 0. Thus,
∂h
∂G 2
G∇ h dξ dη =
G
−h
ds.
(1.11)
∂n
∂n
D
∂D
The Stokes’s theorem in R2 is a generalization of Green’s theorem which states that
if G = (G1 , G2 ) is a continuously differentiable vector field defined on a region
containing D ∪ ∂D ⊂ R2 such that ∂D is a Jordan contour, then
∂G2
∂G1
−
dx dy =
G1 dx+G2 dy.
(1.12)
∂x
∂y
D
∂D
The gradient and the divergence theorems are valid in R3 if the surface integral on
the left side is replaced by a volume integral and the line integral on the right is replaced
by a surface integral. Let the functions M (x, y, z), N (x, y, z), and P (x, y, z), where
(x, y, z) ∈ D, be the components of the vector G in R3 . Then, by the divergence
theorem
∂M
∂N
∂P
+
+
dx
∂x
∂y
∂z
D
(1.13)
=
[M cos(n, x) + N cos(n, y) + P cos(n, z)] dS,
∂D
where dx = dx dy dz, dS denotes the surface element, ∂D denotes the boundary
of D, and cos(n, x), cos(n, y), and cos(n, z) the direction cosines of n. If we take
∂v
∂v
∂v
M =u
,N =u
, and P = u
, then (1.13) yields
∂x
∂y
∂z
∂u ∂v
∂u ∂v ∂u ∂v
∂v
+
+
dx =
u
dS −
u∇2 v dx,
∂x ∂x ∂y ∂y
∂z ∂z
∂n
D
∂D
D
(1.14)
which is known as Green’s first identity. Moreover, if we interchange u and v in
(1.13), we get
∂u ∂v
∂u ∂v
∂u ∂v
∂u
+
+
dx =
v
dS −
v∇2 u dx.
∂x ∂x ∂y ∂y
∂z ∂z
D
∂D ∂n
D
(1.15)
If we subtract (1.14) from (1.15), we obtain Green’s second identity:
2
∂v
∂u
2
u∇ v − v∇ u dx =
u
−v
dS,
(1.16)
∂n
∂n
D
∂D
which is also known asGreen’s reciprocity theorem. This result also holds in R2 .
Note that Green’s identities are valid even if the domain D is bounded by finitely
12
1. SOME BASIC RESULTS
many closed surfaces; however, in that case the surface integrals must be evaluated
over all surfaces that make the boundary of D, and in R2 the line integrals must be
evaluated over all paths that make the boundary of D. If f and g are real and harmonic
in D ⊂ R2 and Γ = ∂D, then from (1.16)
∂g
∂f
f
−g
ds = 0.
(1.17)
∂n
∂n
Γ
Let D be a simply connected region in the complex plane with boundary Γ. Let
z0 be any point inside D, and let D be the region obtained by indenting from D a disk
of radius ε and center at z0 , where ε > 0 is small (Fig. A.1 (a)). Then ∂D consists
of the contour Γ together with the contour Γε .
r
ε
zο
.z
Γε
Γ
zο •
Dε
ε
Γε
Γ
(a)
(b)
Fig. 1.1.
If we set f = u and g = log r in (1.17), where z ∈ D and r = |z − z0 |, then, since
∂
∂
=−
on Γε , we get
∂n
∂r
∂ (log r)
∂u
u
∂u
u
− log r
ds −
− log r
ds = 0.
(1.18)
∂n
∂n
r
∂r
Γ
Γε
Since
2π
u
1
ds = lim
u(z0 + ε eiθ ) ε dθ = 2π u (z0 ) ,
ε→0 Γ r
ε→0 0
ε
ε
2π
∂u
∂u
lim
log r
ds = lim
log ε
ε dθ = 0,
ε→0 Γ
ε→0
∂r
∂ε
0
ε
lim
we let ε → 0 in (1.18) and obtain
2π u(z0 ) =
∂ (log r)
∂u
u
− log r
ds,
∂n
∂n
Γ
(1.19)
1.8. DIFFERENTIATION AND INTEGRATION
13
which is known as Green’s third identity. Note that Eq (1.19) gives the value of a
∂u
harmonic function u at an interior point in terms of the boundary values of u and
.
∂n
If the contour Γ is continuously differentiable (has no corners) and if the point z0 is
on Γ, then, instead of indenting the entire disk of radius ε, we indent from D a half
disk with center at the point z0 (Fig. 1.1(b)), and Green’s third identity becomes
∂ (log r)
∂u
π u(z0 ) = p.v.
u
− log r
ds
∂n
∂n
Γ
(1.20)
where p.v. denotes the principal value of the integral, i.e., it is the limit as r → 0 of
the integral over the contour Γ obtained by deleting that part of Γ which lies within
the circle of radius ε and center z0 .
1.8. Differentiation and Integration
Some useful formulas on Leibniz’s rule and integration by pats in Rn , n = 1, 2, are
presented.
1.8.1. Leibniz’s Rules.
(a) For differentiation of integrals:
d
dx
φ2 (x)
φ2 (x)
F (x, t) dt =
φ1 (x)
φ1 (x)
∂F
dφ2
dφ1
dt + F (φ2 , x)
− F (φ1 , x)
. (1.21)
∂x
dx
dx
(b) For the nth derivative of a product of two functions u and v:
n
n
n−1
D (uv) = u(D v) +
(Du)(D
v) +
(D2 u)(Dn−2 v) + · · · + (Dn u),
1
2
(1.22)
n
where
are the binomial coefficients.
k
n
n
1.8.2. Integration by Parts. Let f (x) and g(x) be piecewise continuous
functions in an interval (a, b) with their first derivatives continuous in (a, b). Then
using integration by parts we get
b
g
a
df
dx = −
dx
b
f
a
!b
dg
!
dx + f g ! .
dx
a
(1.23)
Further, if f (x) and g(x) be piecewise continuous functions in an interval (a, b)
with their first and second derivatives piecewise continuous in (a, b), and h(x) is a
14
1. SOME BASIC RESULTS
continuous function in (a, b), then using integration by parts we get
b
a
b 2
b
d df
d f
df dg
df !!b
g
dx =
g 2 dx = −
dx + g ! ,
dx dx
dx
dx a
a
a dx dx
b
2
g
a
2
d
d f
h(x) 2
dx2
dx
b
dx =
2
h(x)
a
d
+g
dx
h(x)
(1.24)
2
d fd g
dx
dx2 dx2
d2 f
dx2
− h(x)
d2 f dg
dx2 dx
b
.
a
(1.25)
Let a finite domain D ⊂ R2 be bounded by a smooth closed curve ∂D, and let w and
F be scalar functions continuous on D. Then the gradient and divergence theorems
lead to two useful identities in R2 , which in component form are:
∂F
∂w
w
dx dy = −
F dx dy +
nx w F ds,
∂x
D
D ∂x
∂D
(1.26)
∂F
∂w
w
dx dy = −
F dx dy +
ny w F ds.
∂y
D
D ∂y
∂D
1.9. Inequalities
Some useful inequalities are presented for ready reference.
1.9.1. Bessel’s Inequality for Fourier Series. A piecewise continuous
∞ a0 function f (x) can be represented as the Fouries series f (x) =
+
an cos nx+
2 n=1
bn sin nx , where the series on the right converges pointwise, uniformly and in the
mean to f (x) on the interval [−π, π]. The following inequality, known as Bessel’s
inequality, holds:
∞
1
a20 2
+
an + b2n ≤
2
π
n=1
π
f 2 (x) dx.
−π
(1.27)
1.9.2. Bessel’s Inequality for Square-Integrable Functions. Let a given
function f (x), defined on an interval (a, b), be represented by a uniformly continuous
∞
series of the form f (x) =
cn φn (x), where the coefficients cn are constants and φn
n=1
are square-integrable functions orthogonal with respect to a positive weight function
b
f φn ρ dx
ρ(x). Then cn = a
, which yields Bessel’s inequality:
φ2n ρ dx
I
1.10. EXERCISES
∞
b
b
φ2n ρ dx ≤
c2n
f 2 ρ dx.
a
n=1
15
(1.28)
a
If this series converges in the mean to f (x), that is, if
b
n
2
lim
f (x) −
ck φk (x) ρ(x) dx = 0,
n→∞
a
∞
then
k=1
n=1
b
c2n
φ2n ρ dx
a
b
f 2 ρ dx,
=
(1.29)
(1.30)
a
which is known as Parseval’s equality, or Parseval’s completeness relation in the
b
sense that if the limit (1.29) holds for every function f for which
f 2 ρ dx is finite,
a
then the set of functions {φ1 , φ2 , . . . } is said to be complete.
1.9.3. Schwarz’s Inequality for Infinite Sequences. Let cn and dn be two
sequences of numbers . Then
!
!
! ∞
!
!
!
c
d
n n! ≤
!
n=1
∞
c2n
"1/2 ∞
n=1
"1/2
d2n
,
(1.31)
n=1
which is known as Schwarz’s inequality for the two infinite sequences.
1.10. Exercises
1.1. Compute Sn (1) for n = 1, 2, and 3, and interpret your results in R, R2 , and R3 .
Ans. 2, 2π, and 4π/3, respectively.
1.2. Prove that the limit of a convergent sequence is unique.
Hint. Suppose that lim sn = s, and lim sn = t, where |s − t| > 2ε > 0.
n→∞
n→∞
Then show that there are contradictory properties in the ε-neighborhoods of s and
t, namely, these neighborhoods have no points in common, and each contains all
but a finite number of terms of the sequence. Hence s = t.
1.3. Prove that the harmonic series diverges.
Hint. S1 > 2, S8 > 2.5, S16 > 3, S32 > 3.5, S64 > 4, . . . .
1.4. If {sn } is a sequence of non-zero terms and if lim sn = ∞, then show that
n→+∞
lim 1/sn = 0.
n→+∞
Hint. Let ε > 0. Then, lim sn = ∞ implies that for any M > 1/ε, there
n→+∞
exists an integer m > 0 such that whenever n > m then |sn | > M > 1/ε. For
this m, |1/sn | < 1/M < ε for n > m.
16
1. SOME BASIC RESULTS
1.5. Determine whether or not each of the following sequences is bounded, whether
#
3$
nπ
convergent or divergent, and whether oscillating: (a) n +
; (b) sin
;
n
4
# n! $
# ln n $
√
3
(c)
n2 ; (d)
; (e)
.
10n
n
1.6. For a > 1 show that lim an = +∞.
n→+∞
Hint. Let M > 0 and take a = 1 + b, b > 0. Then use binomial expansion of
an = (1 + b)n to show that an > M when n > M/b.
%
1.7. Show that lim n 1/np = 1, p > 0.
n→+∞
Hint. Use np = e(p ln n)/n .
1.8. Show that the infinite arithmetic series a + (a + d) + (a + 2d) + · · · +
[a + (n − 1)d] + · · · diverges for a2 + d2 > 0.
Hint. Sn =
1
2
n [2a + (n − 1)d], and lim Sn = ∞ unless a = d = 0.
n→+∞
1.9. Show that the infinite geometric series a + ar + ar 2 + · · · + ar n−1 + · · · , a = 0,
a
converges to
if |r| < 1 and diverges if |r| ≥ 1.
1−r
a − arn
Hint. Sn =
.
1−r
1.10. Use integral test to check convergence of the following series:
1
1
1
1
(a) √ + √ + √ + √ + · · · ;
3
5
7
9
1
1
1
(b) 1 + p + p + p + · · · ;
2
3
4
1
(c) sin π + 14 sin 12 π + 19 sin 13 π + 16
sin 14 π + · · · .
1
1
; take f (x) = √
. Ans. Diverges.
2n + 1
2x + 1
1
1
(b) f (n) = sn = p ; take f (x) = p . Discuss the cases: (i) p > 1, (ii) p = 1,
n
x
and (iii) p < 1. Ans. Case (i) the series converges; cases (ii) and (iii) it diverges.
∞
1
π
1
(c) f (n) = sn = 2 sin . Consider 2 f (x) dx and show that it is equal to ,
n
n
π
and the series converges.
Hint. (a) f (n) = sn = √
1.11. Use the comparison test to check convergence of the following series:
1 1
1
1
1
(a) + +
+
+ ··· + 2
+···;
2 5 10 17
n +1
22 + 1 32 + 1 42 + 1
(b) 1 + 3
+
+
+···;
2 + 1 33 + 1 43 + 1
1.10. EXERCISES
17
1
1
1
+ + + ···.
2! 3! 4!
1
1
n2 + 1
1
Hint. (a) sn = + < 2 ; Ans. Converges; (b) sn = 3
≥ ; Ans.
n 1
n
n +1
n
1
1
1
diverges. (c) sn = , and use n! ≥ 2n−1 , so
≤ n−1 , and compare with the
n!
n!
2
convergent geometric series. Ans. Converges.
(c) 1 +
1.12. Use the ratio test to check convergence of the following series:
1
2
3
4
1
2!
3!
4!
+
+ 3 + 4 + · · · ; (b) + 2 + 3 + 4 · · · .
3 32
3
3
3 3
3
3
22 + 1 32 + 1 42 + 1
(c) 1 + 3
+
+
+ ···.
2 + 1 33 + 1 43 + 1
n
n+1
n!
Hint. (a) sn = n , sn+1 = n+1 ; Ans. Converges; (b) sn = n , sn+1 =
3
3
3
(n + 1)!
n2 + 1
(n + 1)2 + 1
.
Ans.
Diverges;
(c)
s
=
,
s
=
; Ans. Test
n
n+1
3n+1
n3 + 1
(n + 1)3 + 1
fails.
(a)
2
The Concept of Green’s Functions
Green’s function for a differential equation is its solution when the forcing term is
the Dirac delta function due to a unit point source (or sink) in a given domain. This
solution provides us with a method to generate solutions of linear ordinary and partial
differential equations subject to different kinds of boundary conditions and internal
sources. This method is important in a variety of physical problems and applications,
and has a special role in the study of certain problems of mathematical physics.
2.1. Generalized Functions
The construction of Green’s functions for different types of boundary value problems
is an important subject in itself since these functions are used in a wide range of
applications. Before developing the subject of Green’s functions we must have a
clear understanding of the mathematical behavior of generalized functions. To this
end we will first study certain elementary aspects of generalized functions, followed
by the concept of Green’s function.
A point source is represented by the Dirac delta function, which belongs to a class
of functions known as generalized functions (or distributions, as they are often
called, although care must be taken to distinguish them from ordinary distributions).
The existence of distributions in the generalized sense is based on relations which are
not functions in the traditional sense. The physical examples we want to consider are
related to the determination of the intensity of a concentrated force F acting at time
t = 0 or at the origin x = 0 of a rectangular cartesian coordinate system.
Example 2.1. Consider the vibrating system of a spring, and suppose that an
external force F (t) is applied to the interval −t0 ≤ t ≤ t0 , such that
0,
t0 < |t|,
F (t) =
(2.1)
f (t), t0 ≥ |t|,
18
2.1. GENERALIZED FUNCTIONS
19
where the measure m(t) of the strength of this force at the point t = 0 can be
∞
t0
represented as m(t) =
F (t) dt =
f (t) dt. Now, suppose that f (t) is
−∞
−t0
increased while t0 is decreased in such a way that m(t) remains constant. Then we
have what is known as an impulsive force, and for m(t) = 1 (unit force) we have
t0
lim
f (t) dt = 1. There are many functions f (t) which satisfy the requirement
t0 →0
−t0
(2.1), and one such function is f (t) =
F (t) =
1
. Using this function, let
2t0

0
if t0 < |t|,
1

2t0
if t0 ≥ |t|.
(2.2)
This function F (t) is an ordinary function which belongs to the class C0∞ (I), where
I = [−t0 ≤ t ≤ t0 ]. However, F (t) → δ(t) as t0 → 0; that is, in the limit as
t0 → 0 the function F (t) converges to the Dirac delta-function (henceforth written
δ-function)1 which is not a function in the ordinary sense; it is called a generalized
function. This limit is taken in the weal sense, which means that for any continuous
function φ(t) the weak limit of F (t) is
t0 →0
t0
lim
t0
F (t)φ(t) dt =
−t0
δ(t)φ(t) dt = φ(0).
−t0
Physically, the δ-function signifies a force of strength unity, a sudden impulse, at time
t = 0 which is called the source point or singularity of the δ-function. If the source
point is translated to t , then the function (2.2) becomes

0
F (t) =

1
2t0
if t0 < |t − t |,
if t0 ≥ |t − t |,
such that the weak limit of F (t) as t0 → 0 is δ(t−t ). Note that if the time variable t in
Eq (2.2) is replaced by the space variable x, then lim F (x) = δ (x − x ), which will
x0 →0
signify a sudden impulse of unit strength at the source point x = x . The δ-function
in either case is also written as δ (t − t ) or δ (x − x ).
Example 2.2. Consider the 2-D case of a concentrated force F of unit strength
applied at the origin x = 0, x = (x, y), positioned at the boundary of a region D
of the (x, y)-plane and directed along the z-axis. To measure F let this force be
1 This function was introduced by P. A. M. Dirac in 1926-27. The development of the
analysis in this and next two sections has been reiterated by various authors, but it can be
traced back to examples and ideas found in Dirac [1926-27; 1947].
20
2. THE CONCEPT OF GREEN’S FUNCTIONS
uniformly distributed over a circular area of radius ε and center at the origin. Then
the mean value of the force F over the region D is given by the function

 1
if x2 + y 2 < ε,
π ε2
Fε (x, y) =
(2.3)

0 if x2 + y 2 > ε.
Note that D Fε dx dy = 1. Let the pointwise limit value of the function Fε (x, y)
as ε → 0 be denoted by δ(x, y). Then
δ(x, y) =
+∞ if x2 + y 2 = 0,
0 if x2 + y 2 = 0.
(2.4)
Also,
since the measure of intensity of the force F is equal to 1, we must have
δ(x,
y) dx dy = 1. But the function δ(x, y) defined by (2.4) does not reproduce
D
the value of the force. However, this situation is avoided if instead of the pointwise
limit of Fε we take the weak limit such that for any continuous function φ(x, y) the
weak limit of Fε is
lim
Fε (x, y)φ(x, y) dx dy ≡
δ(x, y)φ(x, y) dx dy = φ(0, 0).
ε→0
D
D
In fact, the weak limit of Fε (x, y) is a functional which assigns the value φ(0, 0) to a
function φ(x, y). Thus, we say that the Dirac δ-function is a weak
limit of the sequence
{Fε }, i.e., lim {Fε (x, y)} = δ(x, y), and we have established D δ(x, y) dx dy =
ε→0
1.
This example can be extended to higher dimensions. If the origin is translated
(shifted)
to a point (x , y ), then the δ-function is represented by δ(x, y; x , y ), and
δ(x, y; x , y ) dx dy = 1. The origin or the point x = (x , y ) is called the
D
source point or singularity of the Dirac δ-function.
There are other descriptions of this function, but they all lead to the same properties
in R, which are
t +ε
0 if t = t ,
δ(t−t ) =
and
δ(t − t ) dt = 1,
(2.5)
+∞ if t → t ,
t −ε
where ε > 0 is an arbitrarily small real number. This property implies that
t +ε
t −ε
f (t) δ(t − t ) dt = f (t ),
(2.6)
which is known as the translation or shifting property of the δ-function. Similar
results hold in higher dimensions.
We will denote by D the class of those functions in C0∞ (R) which have compact
support over the intervals of the form [−a, a] and which approach the Dirac δ-function
2.1. GENERALIZED FUNCTIONS
21
(or δ-distribution) as a → 0. The members of the class D are also called test
functions. To legitimize the δ-function, we will consider some of these test functions
defined in the ordinary sense on an interval I = [−a, a], where a > 0. Note that
the choice of a test function is made from the class D of functions which satisfy the
following two conditions: (i) It must belong to the class C0∞ (R), and (ii) it must
approach the δ-function as a → 0.2
Example 2.3. Let us modify the definition of the function (2.2) by writing
0 for |t| > ε,
fε (t) =
(2.7)
1
for |t| < ε.
2ε
As ε → 0, the rectangles
become narrower and taller (see Fig. 2.1), but their areas
remain constant, since fε (t) dt = 1 for any ε > 0. This means that the δ-function
can be defined as the limit δ(t) = lim fε (t). Obviously, this limit is zero everywhere
ε→0
except at the point t = 0. This has led to the commonly applied mathematical
definition of the Dirac δ-function as
∞ for t = 0,
δ(t) =
(2.8)
0 for |t| > 0,
subject to the additional normalization condition that the area under this function is
equal to 1.
fε(t )
0 ε
t
Fig. 2.1. Delta Function as a Limit of Rectangular Functions.
Properties of the δ-function are:
(i) δ(x) = 0 for x = 0,
∞
(ii)
δ(x) dx = 1
(iii)
−∞
∞
f (x)δ(x) dx = f (0).
−∞
2 The
class C0∞ (R) is defined in $ 1.2.
(2.9)
22
2. THE CONCEPT OF GREEN’S FUNCTIONS
These properties of the Dirac delta-function δ(x) show that it is not a function: It is
zero for every x with the exception of x = 0 where it is infinite and its integral over
the entire real axis is equal to 1. This function is not analytic, but can be obtained as
a limiting case of either analytic continuous or piecewise continuous functions with
finite support, i.e., belonging to the class D of functions in C0∞ (R). The δ-function
is not a function in the strict sense; it has been legitimized with the above defining
properties. However, this function is very useful only as part of an integrand and never
as an end result. By shifting the singularity (source point) from x = 0 to x = x = 0,
this function is written as δ(x − x ). Then the third result in (2.9) becomes
∞
−∞
f (x)δ(x − x ) dx = f (x ).
(2.10)
which is known as the translation or shifting property of the δ-function. In the 3-D
case and using spherical coordinates we have for the singularity (or source) at the
origin
0
∞ π 2π
0
δ(r)r 2 sin θ dr dθ dφ =
∞
∞
∞
δ(x)δ(y)δ(z) dx dy dz = 1,
−∞ −∞ −∞
0
(2.11)
whereas for the singularity at r = r1 , Eq (2.11) becomes
∞
0
π
0
0
2π
δ(r2 − r1 )r22 sin θ2 dr2 dθ2 dφ2 = 1.
(2.12)
Note that δ(r1 − r2 ) = δ(r2 − r1 ).
The definition (2.8) is consistent with the classical definition of a function and
automatically satisfies the second property in (2.9). An important consequence of
this property is that if f (t) is any continuous function, then
∞
ε
δ(t)f (t) dt = lim
ε→0
−∞
δ(t)f (t) dt = f (0).
−ε
To prove, note that by definition (2.7)
∞
ε
δ(t)f (t) dt = lim
−∞
ε→0
−ε
1
1
f (t) dt = lim 2εf (t ) = f (0),
ε→0
2ε
2ε
where t is a point at which f (t) takes its average value in (−ε, ε), such that t ∈
(−ε, ε), and therefore, t → 0 as ε → 0.
The dimension of the δ-function, which is important to physicists and engineers
but mostly ignored by mathematicians, can be determined from the fact that this
function is one of those few self-similar functions whose argument can be a space
variable or a time variable, and therefore, depending on the dimension of its argument,
2.1. GENERALIZED FUNCTIONS
23
the δ-function has a nonzero dimension. For example, the dimension of the δ-function
of time δ(t) is equal to the inverse time, i.e., the dimension of frequency because, by
∞
definition, −∞ δ(t) dt = 1, which is dimensionless. In other words, the dimension
of δ(t) is equal to the dimension of the inverse function 1/t.
The Dirac δ-function is used in electronics to represent the unit impulse and
sampled signals. It also represents a unit mass ideally concentrated at the origin; if
concentrated elsewhere, e.g., at the integer n, it is denoted by δn , and the δ-function
translated at n is defined by δn (t) = δ(t − n). Although the above definition of the
δ-function is heuristically relevant, it is not mathematically correct. In the Riemann
sense the integral of the δ-function defined by (2.10) is not well-defined, or in the
Lebesgue sense it equals zero. However, for each ε > 0 the integral
∞
Tε [φ] =
fε (t)φ(t) dt
(2.13)
−∞
exists for any fixed continuous function φ, and it converges to the value φ(0) as
ε → 0+ ; that is, lim Tε [φ] = T [φ] = φ(0). This definition can provide us with
ε→0+
a valid definition of δ-function as a limit of the integrals of type (2.13). In fact, the
integral (2.13) defines a linear functional on the test function φ(t), generated by the
function fε (t) which is known as the kernel of the functional Tε . The notion of a
functional is more general than that of a function because a functional depends on a
variable which is a function itself but its values are real numbers. The linear functional
Tε [φ] on the test function φ is generated by the integral (2.13) with the kernel fε (t),
where the test function φ belongs to the class D of smooth functions in C0∞ (R) with
compact support. The functional that assigns to each test function its value at 0 will
correspond to the Dirac delta distribution. The choice of test functions can be any
function in the class D.
To show that the value of a continuous function f at any point t is determined by
the value of the functional (2.13) on all test functions in the class D, we will discuss
by an example.
Example 2.4. Consider the function

 C exp − 1
,
1 − t2
ω(t) =

0, for |t| ≥ 1,
for |t| < 1,
(2.14)
∞
where the constant C is determined from the normalization condition −∞ ω(t) dt
= 1, which gives C ≈ 2.25. The function ω(t) ∈ C ∞ ; it vanishes outside the interval
[−1, 1] which is bounded, and it is differentiable everywhere including the endpoints
−1 and 1 such that at −1 (or at 1) all the right (or left) derivatives are zero and the
left (or right) derivatives are identically zero. The function ω(t), after rescaling for
each ε > 0, gives a new function
ωε (t) =
1
ω
ε
t
ε
,
(2.15)
24
2. THE CONCEPT OF GREEN’S FUNCTIONS
which has compact support since it vanishes outside the interval [−ε, ε], and it is
infinitely
differentiable which can be checked by applying the chain rule. Also,
∞
ω
(t)
dt = 1. Applying the generalized mean value theorem for integrals and
ε
−∞
assuming that the function f (t) is continuous, we find the value of the functional as
Tε [f ] =
∞
−∞
f (t) ωε (t) dt → f (0) as ε → 0.
Thus, the value of f can be recovered by evaluating the functionals Tε at f .
We have shown that any linear functional T [φ] which is continuous on the set
D of functions in C ∞ is a distribution. A functional T is said to be linear on D if
T [α φ+βψ] = α T [φ]+β T [ψ] for any test functions φ and ψ in D, where α and β are
any real (complex) numbers. A functional T on D is said to be continuous if for any
sequence of functions φk (t) ∈ D which converge to a test function φ(t), the numbers
T [φk ] which are the values of the functional T on φk converge to the number T [φ].
The convergence in this definition is understood in the following sense: The supports
of all φk are all contained in a fixed bounded set in their domain, and as k → ∞, the
functions φk , including all their derivatives φm
k , m = 1, 2, . . . , converge uniformly
(m)
to the corresponding limit test function φ and its derivatives, i.e., φk → φ(m) for
each m = 0, 1, 2, . . . .
It is interesting to note that the support of the function δ(t) is the single point
t = 0. For this reason the δ-function has the physical significance of the source
density for a unit source concentrated at a single point known as the source point.
Besides the functions defined in (2.2), (2.7) and (2.15), there are several other
test functions that belong to the class D. They are given below, and their graphs are
presented in Fig. 2.2:
2
2
2
eε /(|t| −ε ) , |t| < ε,
(a)
φε (t) =
0, |t| > ε,
such that φε (0) = e−1 and |φε (t)| ≤ e−1 (known as the ‘cap’ function);
(b)
φj (t) =
1
ωε (t),
εj
j = 1, 2, · · · , |t| < ε,
where ωε (t) is defined in (2.15);
(c)
φj (t) =
j
1
,
π 1 + j 2 t2
j = 1, 2, · · · , |t| < ε;
(d)
2 2
j
φj (t) = √ e−j t ,
π
j = 1, 2, · · · , |t| < ε;
(e)
φj (t) =
1 sin2 jt
,
jπ t2
j = 1, 2, · · · , |t| < ε.
2.1. GENERALIZED FUNCTIONS
25
φ1 ( t )
φ (t )
ε
ε = 1/4
ε = 1/4
(a)
(b)
ε = 1/2
ε = 1/2
ε=1
ε=1
t
0
−1
φ j( t )
φ j( t )
j=4
(c)
j=4
(d)
j=2
j=2
j=1
−3
0
1
0
3
j=1
t
−3
0
t
3
φ j( t )
j=4
(e)
j=2
j=1
t
−3
0
3
Fig. 2.2. Graphs of Test Functions (a)–(e).
Moreover, the functions
2
1
√ e−x /4ε ,
2 πε
1
x
sin ,
π
ε
1
ε
,
π x2 + ε2
ε
x
sin2 ,
πx2
ε
with their domains of definition restricted to a compact set in R, say |x| < ε, are other
examples of test functions belonging to the class D. All test functions in the class D
approach δ(t) in the limit as ε → 0.
26
2. THE CONCEPT OF GREEN’S FUNCTIONS
The 2-D examples the test functions are:
(i) The ‘cap’ function
ωε (x, y) =
2
2
2
Cε e−ε /(ε −|x| ) ,
0,
if |x|2 = x2 + y 2 ≤ ε,
if x2 + y 2 > ε,
where the constant Cε is determined from the condition that
ωε (x, y) dx dy = 1.
D
2
2
1
√ e−|x| /(2ε ) ,
ε 2π
(ii) The function ωε (x, y) =
|x|2 = x2 + y 2 , where
ωε (x, y) dx dy = 1.
D
To summarize, some important properties of the δ-function are as follows:
∞
1.
f (t) δ(t − t ) dt = f (t ) (Fundamental property);
−∞
∞
2.
−∞
f (t ) δ(t − t ) dt = f (t )
∞
−∞
δ(t − t ) dt = f (t );
3. f (t) δ(t − t ) = f (t ) δ(t − t );
4. t δ(t) = 0;
5. δ(t − t ) = δ(t − t);
∞
1 for t > 0,
6.
δ(s) ds =
= H(t);
0 for t < 0
−∞
d
H(t) = δ(t),
dt
where H(t) is the Heaviside unit step function (Fig. 2.3). Note that the last property
is not true in the classical sense since H(t) is not differentiable at t = 0. In the above
results the point t = t is the source point (or singularity) for the δ-function.
7.
The δ-function can be similarly defined for the space variable x, and all the results
given above hold for δ(x), where x ∈ Rn , or for δ(x − x ) where x is the source
point for the δ-function. As in the 1-D case, the δ-function in R2 is formally written
as δ(x) = δ(x) δ(y), and δ(x; t) = δ(x) δ(y) δ(t); similarly, in R3 it is formally
written as δ(x) = δ(x) δ(y) δ(z), and δ(x; t) = δ(x) δ(y) δ(z) δ(t).
2.1.1. Heaviside Function. A very useful application of the property (iii)
in
(2.9)
provides us with an alternative definition of the function δ(t): If f, g =
f (t)g(t) dt = f (0) for continuous or piecewise continuous functions on R, then
the function g(t) must be the δ-function. We will use this result in the following
example to discuss the Heaviside function H(t).
2.1. GENERALIZED FUNCTIONS
27
H(t )
1
t
0
Fig. 2.3. Heaviside Function H(t).
Example 2.5. An alternative definition of the δ-function is based on the relationship between the Heaviside function H(t), defined in the above Property 6, and
the function δ(t). This can be established
∞ as follows: The Heaviside function H(t)
defines the distribution f, H =
f (t) dt, and so, in the generalized sense, its
0
∞
derivative is defined by f, H = H, −f = −
f (t) dt = f (0); thus,
0
H (t) = δ(t).
(2.16)
H(t + ε) − H(t − ε)
, by property (iii) in (2.9) we have
ε→0
ε
∞
∞
1
f, H = lim
f (t) dt −
f (t) dt
ε→0 ε
−ε
0
1 0
= lim
f (t) dt = f (0), since f is continuous at t = 0,
ε→0 ε −ε
Alternately, since H = lim
which implies (2.16). Since H(t) is not differentiable at t = 0 in the classical sense,
the distributional definition (2.16) for H (t) when formally integrated gives (2.16).
The function H(t − t ) is similarly defined by
1, t > t ,
H(t − t ) =
(2.17)
0, t < t .
The Heaviside function is generally used to restrict a time-dependent solution for
initial value problem to the interval t > t .
2.1.2. Delta Function in Curvilinear Coordinates. The δ-function in
the polar coordinates (r, θ) in R2 is given by
δ(x) δ(y) =
δ(r)
,
2πr
r 2 = x2 + y 2 ,
(2.18)
and in the spherical coordinates (r, θ, φ) in R3 by
δ(x) δ(y) δ(z) =
δ(r)
,
4πr2
r 2 = x2 + y 2 + z 2 .
(2.19)
28
2. THE CONCEPT OF GREEN’S FUNCTIONS
For definition of these coordinate systems, see §4.1. In general, for the two-dimensional
case, if the cartesian coordinates x = (x, y) are transformed into general coordinates
u = (u, v), then
(x , y )
δ (x − x ) = δ (x − x ) δ (y − y ) = δ (u − u ) J
(u , v )
−1
(2.20)
,
(x , y ) = 0, where x = (x , y ) and u = (u , v ).
(u , v )
Similarly, for the three-dimensional case, where x = (x, y, z), x = (x , y , z ),
u = (u, v, w), and u = (u , v , w ), the transformation is given by
provided the Jacobian J
δ (x − x ) = δ (x − x ) δ (y − y ) δ (z − z )
−1
(x , y , z )
= δ (u − u ) J
,
(u , v , w )
(2.21)
(x , y , z ) = 0. We will ignore the proofs and also the singular case
(u , v , w )
when the Jacobian J is zero; they can be found in Stakgold [1979]. Some useful
results in the polar and the spherical coordinates are:
provided J
δ (r − r ) δ (θ − θ )
in R2 ,
r
δ (r − r ) δ (θ − θ ) δ (φ − φ )
δ (x − x ) δ (y − y ) δ (z − z ) =
r 2 sin φ
δ (x − x ) δ (y − y ) =
in R3 .
The δ-function in all dimensions has the following properties:
f (x) δ (x − x ) dx =
D
and
D
δ (k (x − x )) dx =
D
f (x )
0
if x ∈ D,
if x ∈ D,
1
δ (x − x ) dx,
k
(2.22)
(2.23)
where the integral is single, double, or triple, depending on the domain D.
Note that δ (x − x ) is also written as δ (x, x ), and has the units [L−1 ]. Also, we
often write δ (x, x ) for δ (x − x ), δ (x, t) for δ (x) δ(t), and δ (x) for δ (x, 0).
! ! For the sake of simplicity we will henceforth use s in stead of t or x when
we discuss Green’s functions in R in this and the following chapter.
2.2. SINGULAR DISTRIBUTIONS
29
2.2. Singular Distributions
We will define singular distributions as limits of regular functions. As we have
seen, the Dirac delta distribution
can be represented as a limit of a sequence of
integral functionals Tk [φ] =
∞
−∞
fk (x) φ(x) dx with respect to the kernels that are
regular functions. This leads to the representation of δ(x) in the limit such that
Tk [φ] → δ(φ) as k → ∞ for each φ ∈ D. The convergence in this situation is always
the weak convergence (see §1.3.2 and §2.1). We will consider the cases of three
regular functions, namely the Gaussian functions, Cauchy densities, and complexvalued oscillating functions, in the following examples.
Example 2.6. Consider the family of Gaussian functions
gε (x) = √
2
1
e−x /(2ε) ,
2πε
which are normalized by the condition that
∞
−∞
(2.24)
ε > 0,
gε (x) dx = 1. Let fk (x) =
g1/k (x), k = 1, 2, . . . be a weakly approximating sequence. Then ε → 0 as k → ∞,
and the graphs of the approximating Gaussian functions gk (x) show higher and higher
peaks and move more and more close toward x = 0, but the area under each one of
them remains constant (see Fig. 2.4.).
f k ( x)
x
Fig. 2.4. Graphs of the First Three Gaussian Functions gk (x).
Example 2.7. Consider the Cauchy densities (or the Lorentz curves, as the
physicists call them)
λε (x) =
1
ε
,
π x2 + ε 2
(2.25)
30
2. THE CONCEPT OF GREEN’S FUNCTIONS
x
Fig. 2.5. Graphs of {fε (x)}.
and define the kernels fk (x) = λ1/k (x). These kernels, which look almost like the
Gaussian functions (Fig. 2.4) but differ from them significantly, weakly converge
to the δ-function. In fact, both the Gaussian functions and the Cauchy densities are
infinitely differentiable and integrable functions for all values of x ∈ R, since the
dx
integral
= arctan x has finite limit at x = −∞ and x = ∞, but at x = 0
2
x +1
gε (0) = √
1
,
2πε
and
λε (0) =
1
,
πε
and thus both go to +∞ as ε → 0. But these two functions differ from each other
as follows: The Gaussian function decays exponentially to 0 as x → ±∞, whereas
ε
the Cauchy densities, being asymptotic to
, decay more slowly to 0 as x → ±∞
πx2
than the Gaussian, and the area under them are less concentrated around x = 0 than
in the case of the Gaussian functions. In particular, for f (x) = x2
∞
∞
x2 −x2 /(2ε)
1
ε
√
Tf [gε ] =
e
dx = ε, Tf [λε ] =
x2 2
dx = ∞,
π
x
+
ε2
2πε
−∞
−∞
which shows that the functional with the Gaussian function are well defined while
those with the Cauchy densities are not.
Example 2.8. Consider the complex-valued functions
i −ix2 /(2ε)
fε (x) =
e
, ε > 0.
2πε
(2.26)
These functions are found in quantum mechanics and quasi-optics.3 The graph of
1
{fε (x)} is given in Fig. 2.5. Note that |fε (x)| = √
= const, and diverges to
2πε
In quasi-optics {fε (x)} turn out to be Green’s functions of a monochromatic wave
in the Fresnel approximation (see §7.12.1(b)).
3
2.3. THE CONCEPT OF GREEN’S FUNCTIONS
31
∞ as ε → 0. However, these functions converge weakly to the δ-function as ε → 0.
It follows from the fact that these functions oscillate at higher and higher speed as ε
becomes smaller and smaller.
2.3. The Concept of Green’s Functions
Green’s function G(x, x ) for a linear ordinary or partial differential operator L
satisfies the following system: L[G(x, x )] = δ(x − x ) in a domain D, and subject
to the boundary conditions BG(x, x ) = 0 on ∂D, where B is a linear ordinary or
partial differential operator whose order is less than that of L. If the domain D is the
entire space, the solution to this system is known as the fundamental or (singularity)
solution for the operator L (see §2.5).
To understand the concept of Green’s functions4 we will take some examples and
firstly discuss Green’s functions for ordinary differential equations.
Example 2.9. Consider the boundary value problem of the forced, transverse
vibrations of a taut string of length l, defined by
L[u] =
d2 u
+ k 2 u = −f (t),
dt2
0 < t < l,
(2.27)
where L denotes the linear differential operator, and subject to the boundary conditions
u(0) = 0 = u(l), which imply that the endpoints of the string are kept fixed. The
general solution of Eq (2.27) is
u(t) = A(t) cos kt + B(t) sin kt,
where the functions A(t) and B(t) are to be determined by the method of variation
of parameters. Thus, we solve the equations
A (t) cos kt + B (t) sin kt = 0,
−kA (t) sin kt + kB (t) cos kt = −f (t)
to find A (t) and B (t), which are given by
A (t) =
f (t) sin kt
,
k
and B (t) = −
f (t) cos kt
.
k
Integrating these functions with respect to t to obtain A(t) and B(t), we find the
formal solution of Eq (2.27) as
cos kt t
sin kt t
u(t) =
f (s) sin ks ds −
f (s) cos ks ds,
(2.28)
k
k
a
b
4 George Green (1793–1841) was an English miller and self-taught mathematician who
also made contributions to the theory of electricity and magnetism, including the term
‘potential’.
32
2. THE CONCEPT OF GREEN’S FUNCTIONS
where the constants a and b are determined from the two boundary conditions. Applying the boundary condition u(0) = 0 to Eq (2.28) we find that a must be chosen
0
such that
f (s) sin ks ds = 0. Since f (s) is arbitrary, we must choose a = 0.
a
Similarly, on applying the boundary condition u(l) = 0 to Eq (2.28), we must have
cos kl l
sin kl l
u(l) =
f (s) sin ks ds −
f (s) cos ks ds = 0,
k
k
0
b
or
sin kl 0
1 l
f (s) cos ks ds =
f (s) sin k(s − l) ds.
(2.29)
k
k 0
b
Combining (2.28) and (2.29), the solution of this boundary value problem is given by
l
1 t
sin kt
u(t) =
f (s) sin k(s − t) ds −
f (s) sin k(s − l) ds
k 0
k sin kl 0
t
l
sin ks sin k(l − t)
sin kt sin k(l − s)
=
f (s)
ds +
f (s)
ds (2.30)
k
sin
kl
k sin kl
0
t
l
=
f (s) G(t, s) ds,
0
where

sin ks sin k(l − t)


, for 0 ≤ s ≤ t,
k sin kl
G(t, s) =
(2.31)

 sin kt sin k(l − s) , for t ≤ s ≤ l.
k sin kl
This function G(t, s) is a two-point function of position known as Green’s function
for the boundary value problem considered here.5 This function exists provided
sin kl = 0. Notice that Green’s function defined by (2.31) is independent of the
forcing term f (t), but it depends only on the associated homogeneous differential
equation and the boundary conditions. This implies that in all such problems with
different function f (t) the knowledge of Green’s function allows us to construct the
solution u(t) of the boundary value problem (2.27).
Example 2.10. Consider the following boundary value problem:
L[u] = u − k 2 u = f (x),
a < x < b;
u(a) = 0 = u(b).
(2.32)
The solution for this problem, using the method of variation of parameters, is
x
1
cosh k(b − x − s + a) − cosh k(b − x + s − a)
u(x) =
f (s) ds
2k a
sinh k(b − a)
b
1
cosh k(b − x − s + a) − cosh k(b + x − s − a)
+
f (s) ds.
2k x
sinh k(b − a)
5
This function was called the ‘influence function’ by George Green in 1828 when he
solved the string problem.
2.3. THE CONCEPT OF GREEN’S FUNCTIONS
33
This can be simplified as
u(x) = −
1
k
sinh k(b − x) sinh k(s − a)
f (s) ds
sinh k(b − a)
a
b
sinh k(x − a) sinh k(b − s)
+
f (s) ds .
sinh k(b − a)
x
x
Now, if we define
G(x, s) =

sinh k(x − a) sinh k(b − s)


−
k sinh k(b − a)
if a ≤ x < s,

sinh k(b − x) sinh k(s − a)

−
k sinh k(b − a)
if s < x ≤ b,
(2.33)
we can express the solution (2.32) as
u(x) =
b
G(x, s)f (s) ds.
(2.34)
a
The function G(x, s) is known as Green’s function for the problem (2.32).
In these examples we notice that if the function f (x) is taken with a minus sign,
as in (2.27), Green’s function defined by (2.31) does not have a minus sign; this
situation has, however, reversed in Example 2.10. Mathematically it does not matter
whether f (x) is taken with a minus or a plus sign. Now, from these two examples we
can derive some conclusions which should help us solve the problem of constructing
Green’s functions for more general initial and boundary value problems. We notice
the following properties of Green’s function G(x, s):
1. This function satisfies the given differential equation, i.e., G ± k 2 G = 0 in each
intervals 0 ≤ s < t, t < s ≤ l (in Example 2.9), or a ≤ x < s, s < x ≤ b (in
Example 2.10);
2. In Examples 2.9 and 2.10, the function G is continuous at the point s = x since
sin kx sin k(l − x)
= lim G(x, s),
k sin kl
s→x+
sinh k(x − a) sinh k(b − x)
lim G(x, s) =
= lim+ G(x, s).
k sinh k(b − a)
s→x−
s→x
lim G(x, s) =
s→x−
3. The derivative dG/ds is discontinuous at s = x, since in Example 2.9
cos kx sin k(l − x)
,
sin kl
sin kx cos k(l − x)
G (x, x+ ) = lims→x+ G (x, s) = −
.
sin kl
Thus, G (x, x+ ) − G (x, x− ) = −1. Similar results hold for Example 2.10;
G (x, x− ) = lims→x− G (x, s) =
34
2. THE CONCEPT OF GREEN’S FUNCTIONS
4. Since G(x, 0) = G(x, l) = 0, the function G satisfies the boundary conditions;
5. The function G is symmetric in its arguments, i.e., G(x, s) = G(s, x).
2.4. Linear Operators and Inverse Operators
Definition of linear operators, their inverse and adjoint operators, and their role in determining Green’s functions for linear differential equations with prescribed boundary
conditions are discussed.
2.4.1. Linear Operators and Inverse Operators. The linear differential
operator L was introduced in Examples 2.9 and 2.10. We will now introduce the
inverse operator L−1 and express the solution of boundary value problems as
b
−1
u(x) = L [f ] (x) =
G(x, s)f (s) ds,
(2.35)
or symbolically as L−1 [ ](·) =
a
b
G(x, s) (·) ds. We will now show that G(x, s)
a
is the solution of the boundary value problem
u − k 2 u = δ(x − s),
a < x < b,
u(a) = u(b) = 0.
(2.36)
Note that in some textbooks, Green’s function for boundary value problems is defined
as the solution of the equation u −k 2 u = −δ(x−s), a < x < b; u(a) = u(b) =
0, which assumes an impulsive force of strength −1 at the source point (or singularity)
x = s. In Eq (2.36) we are assuming this force to be of strength +1. Before we solve
this differential equation, we will prove the following result:
Theorem 2.1. The solution of the differential equation
P (x) u + Q(x) u + R(x) u = δ(x − s),
a < x < b,
(2.37)
subject to the boundary conditions u(a) = u(b) = 0, satisfies the condition
1
u (x+ ) − u (x− ) =
, provided P (x) = 0.
P (x)
Proof. Define F (x) = e
F u +
Q/P dx
. Then Eq (2.37) can be written as
Q
R
F
F u + F u = δ(x − s),
P
P
P
where we have dropped the arguments in the functions F, P, Q and R, or as
d
R
F
(F u ) + F u = δ(x − s),
dx
P
P
2.4. LINEAR OPERATORS AND INVERSE OPERATORS
35
Integrating both sides of this equation from x − ε to x + ε, we get
x+ε
x−ε
d
(F u ) dx +
dx
x+ε
R
F u dx =
P
x−ε
x+ε
x−ε
F (x)
F (x)
δ(x − s) dx =
.
P (x)
P (x)
Now, if we take the limit as ε → 0, the second term in the above equation vanishes
and we have
F (x)
lim F (x + ε) u (x + ε) − F (x − ε) u (x − ε) =
,
ε→0
P (x)
or
F (x)
F (x) u (x+ ) − u (x− ) =
,
P (x)
which yields the required result.
Example 2.11. Consider the problem (2.36) whose solution is given by
u(x) =
A sinh k(x − a)
B sinh k(b − x)
if x < s,
if x > s,
(2.38)
where the solution (2.38) satisfies the boundary conditions and the constants A and
B satisfy the condition of continuity for u(x) and the jump condition for u (x) (see
Theorem 2.1). Thus,
A sinh k(s − a) = B sinh k(b − s),
−Bk cosh k(b − s) − Ak cosh k(s − a) = 1.
Solving this set of equations for A and B and substituting them into Eq (2.38) and
simplifying, we get
G(x, s) =

sinh k(x − a) sinh k(b − s)


−
k sinh k(b − a)
if a ≤ x < s,

sinh k(s − a) sinh k(b − x)

−
k sinh k(b − a)
if s < x ≤ b.
Notice that this Green’s function is the same as (2.33) and it satisfies the same five
conditions given at the end of the previous section.
2.4.2. Adjoint Operators. If L is a linear differential operator with dependent
variable u and independent variable x, then an operator L∗ which satisfies the relation
b
v L[u] dx =
a
a
b
b
u L∗ [v] dx + M (u, v, u , v , x) ,
a
(2.39)
36
2. THE CONCEPT OF GREEN’S FUNCTIONS
where M (u, v, u , v , x) represents the boundary terms obtained after integration by
parts, is called the adjoint operator of L. An operator L is said to be self-adjoint if
L = L∗ . For more on adjoint operators, see Appendix A.
The basic result in the theory of Green’s functions is as follows:
Theorem 2.2. If G (x, x ) is Green’s function for the linear operator L
and G∗ (x, x ) is Green’s function for its adjoint operator L∗ , then G (x, x ) =
G∗ (x , x).
Before we prove this theorem, we will discuss two boundary value problems
involving second-order differential operators. This will not only clarify the concept
of Green’s functions but also show the significance of this theorem.
Example 2.12. (Problem I): Let L be a second order ordinary differential
operator. Consider the boundary value problem:
L[u] = f (x) in (a, b), such that B1 u
x=a
= 0, B2 u
x=b
= 0,
where L[u] = A(x)u + B(x)u + C(x)u, and B1 and B2 are linear differential
operators of order 1 defining the boundary conditions. Then integration by parts
yields
b
b v L[u] dx = A(x)vu − A(x)v u + B(x)vu +
a
a
b
uL∗ [v] dx,
a
where L∗ [v] = A(x)v + [2A (x) − B(x)] v + [A (x) − B (x) + C(x)] v. If we
require v to satisfy the boundary conditions B1∗ v = 0 and B2∗ v = 0, where
x=a
x=b
b
the operators B1∗ and B2∗ are chosen such that A(x)vu − A(x)v +B(x)vu = 0,
a
then the pair of the boundary conditions B1∗ v
= 0 and B2∗ v = 0 become the
x=a
x=b
adjoint boundary conditions for the given pair of the boundary conditions B1 u
=
x=a
0 and B2 u
= 0. Next, consider the following boundary value problem:
x=b
L[u]
b) with the homogeneous boundary conditions B1 ux=a = 0,
= f (x) in (a,
B2 ux=b = 0. Let G∗ (x, s) be the solution of the problem:
L∗ [v] = δ (x − s)
in (a, b), such that B1∗ v x=a
= 0, B2∗ v
x=b
= 0.
Then if we replace v by G∗ (x, s) in Eq (2.39), it reduces to
b
a
G∗ (x, s) L[u] dx =
b
a
u L∗ [G∗ ] (x, s) dx =
b
u δ (x − s) dx = u (s) .
a
2.4. LINEAR OPERATORS AND INVERSE OPERATORS
But since
b
G∗ (x, s) L[u] dx =
a
to
b
37
G∗ (x, s) f (x) dx, the above equation reduces
a
b
u (s) =
G∗ (x, s) f (x) dx.
a
Moreover, since s is an arbitrary point, we interchange x and s in the above result
and get
b
u(x) =
G∗ (s, x) f (s) ds.
(2.40)
a
In this solution Green’s function for the adjoint operator L∗ turns out to be such that
G∗ (s, x) = G(x, s), where s is used for x .
Example 2.13. (Problem II): Consider a second-order partial differential operator L, and let u(x) and v(x) be two C 2 -functions. Then, integrating by parts, we
get
v(x) L[u](x) dx =
D
M (u, v) dS +
∂D
u(x)L∗ [v](x) dx,
(2.41)
D
where M (u, v) is a differential operator of the first order, ∂D is the boundary of the
region D, L∗ is the adjoint operator of L, and dx = dx dy dz is the volume element.
Consider the following boundary value problem:
L[u] = f (x)
in D, such that Bu = 0 on ∂D,
where B is a linear partial differential operator of the first order. To find the solution
of this problem, we use Eq (2.41) and the solution of the following problem:
L∗ [v] = δ(x, y) in D, such that B ∗ v = 0 on ∂D,
(2.42)
where B ∗ is a linear partial differential operator of the first order such that M (u, v) =
0 on ∂D. The boundary condition B ∗ v = 0 on ∂D is known as the adjoint boundary
condition for Bu = 0 on ∂D. The solution v (x, x ) of the problem (2.42) is denoted
by G∗ (x, x ) and is known as Green’s function for the problem (2.42). Using (2.41),
we get
G∗ (x, x ) L[u](x) dx =
D
M (u, G∗ ) dS +
∂D
u(x) L∗ [G∗ ] (x, x ) dx,
D
which gives
D
G∗ (x, x ) f (x) dx =
D
u(x) δ (x − x ) dx = u (x ) .
(2.43)
38
2. THE CONCEPT OF GREEN’S FUNCTIONS
Since x is arbitrary, we interchange x and x and obtain
u (x) =
G∗ (x , x) f (x ) dx .
D
Then using Theorem 2.2, we obtain
u(x) =
G(x, x ) f (x ) dx .
D
It turns out that even when the boundary condition on u is nonhomogeneous, we can
find the solution from Eq (2.43). In the above solution we needed Green’s function
G∗ (x, x ) for the adjoint operator L∗ defined in the problem (2.42).
Proof of Theorem 2.2. In this proof we will show that G∗ (x, x ) =
G (x , x), where G∗ (x, x ) is solution of the equation L∗ [G∗ ] (x, x ) = δ (x − x )
subject to homogeneous boundary conditions. Clearly, G (x, z1 ) satisfies the differential equation L[G] (x, z1 ) = δ (x − z1 ) in D, and the boundary condition
G (xs , z1 ) = 0 on the boundary ∂D, where xs is any point on ∂D, while G∗ (x, z2 )
satisfies the differential equation L∗ [G∗ ] (x, z2 ) = δ (x − z2 ) in D and the boundary condition G∗ (xs , z2 ) = 0 on ∂D. Multiplying the first differential equation by
G∗ (x, z2 ) and integrating over the whole region D, we get
G∗ (x, z2 ) L[G] (x, z1 ) dx =
G∗ (x, z2 ) δ (x − z1 ) dx = G∗ (z1 , z2 ) .
D
D
G∗ (x, z2 ) L[G] (x, z1 ) dx by parts, we get
If we integrate
D
G∗ (x, z2 ) L[G] (x, z1 ) dx =
G (x, z1 ) L∗ [G∗ ] (x, z2 ) dx,
D
D
G (x, z1 ) δ (x − z2 ) dx = G (z2 , z1 ) .
=
D
The boundary terms vanish due to the boundary conditions. Hence,
G (z2 , z1 ) = G∗ (z1 , z2 ) .
(2.44)
Replacing z2 by x and z1 by x , the theorem is proved.
Corollary 2.2. (Reciprocity Theorem.) Green’s function is symmetric for
the self-adjoint operators with respect to x and x , i.e., G(x, x ) = G(x , x).
Proof. The proof follows from Green’s theorem (1.12). Substitute G1 =
G(x, x ), G2 = G(x, x ) in the identity by dividing the interval of integration into
three subintervals x0 ≤ x ≤ x , x ≤ x ≤ x , x ≤ x ≤ x1 , treating each
separately. The proof is completed by taking into account both discontinuities at the
points x = x and x = x .
2.4. LINEAR OPERATORS AND INVERSE OPERATORS
39
Courant and Hilbert [1968:355] note that “The symmetry of Green’s function
expresses a reciprocity frequently occurring in physics: If the force 1, applied at the
point x , produces the result G(x, x ) at the point x, then the same force 1 acting at
x produces the same result at x .”
Theorem 2.3. If f (x) is a continuous or piecewise continuous function
of x, then the function
x1
u(x) =
G(x, s)f (s) ds
(2.45)
x0
is a solution of the equation L[u] = −f (x), which satisfies the prescribed
boundary conditions.
Proof. From (2.45) we have
x1
u (x) =
G (x, s)f (s) ds,
x0
x1
u (x) =
G (x, s)f (s) ds +
x0
x1
G (x, s)f (s) ds
x0
+ G (x, x − 0)f (x) − G (x.x + 0)f (x)
x1
=
x
x0 1
=
G (x, s)f (s) ds + [G (x, x + 0) − G (x, x − 0)] f (x)
G (x, s)f (s) ds −
x0
f (x)
.
p(x)
x
Thus, pu + p u − qu = x01 (pg + p G − qG) f (s) ds − f (x). This completes
the proof since L[G] = 0. To prove the converse, we integrate by parts setting u = G
over two intervals of integration x0 ≤ x ≤ s and s ≤ x ≤ x1 . Thus, the general
solution is
x1
x1
u(x) =
G(x, s)f (s) ds + pG u .
x0
x0
An application of Theorem 2.2 will allow us to express the solution of the partial
differential equation L[u](x) = δ(x, x ) in terms of Green’s function G (x, x ).
Therefore, it is not required to use the adjoint operator L∗ to find a Green’s function.
Example 2.14. Consider the boundary value problem u + 2u + u = f (x),
u(0) = 0 = u(l). Its solution is given by
s
s
x s
u(x) = −
(l − x) f (s) e−(x−s) ds −
(l − s) f (s) e−(x−s) ds, (2.46)
l 0
0 l
which leads to Green’s function
 x
 − (l − s) e−(x−s)
l
G(x, s) =
 − s (l − x) e−(x−s)
l
for 0 ≤ x < s,
for s < x ≤ l.
(2.47)
40
2. THE CONCEPT OF GREEN’S FUNCTIONS
Note that the same Green’s function is obtained if we directly solve the problem
u + 2u + u = δ(x − s), u(0) = 0 = u(l). In this case the operator D2 + 2D + 1 is
not self-adjoint, and the adjoint operator is given by D2 − 2D + 1. Green’s function
G∗ (x, s) for this adjoint operator satisfies the boundary value problem u −2u +u =
δ(x − s), u(0) = 0 = u(l). But the solution G∗ (x, s) for this problem is such that
G∗ (s, x) = G(x, s). The function G∗ (s, x) is given by (2.47). Green’s function in
this case is not symmetric. Also note that once Green’s function (2.47) is known,
the solution u of the problem, given by (2.46), can be easily determined. This is
known as the Green’s function method of finding the solution to boundary value
problems.
The Green’s function is the response at the point x from an impulse at the point
s. It is the influence function which depends on the geometry of the domain and the
boundary conditions.
We will show that for every equation and boundary condition the Green’s function
gives a kind of ‘inverse’. We have the following result:
Theorem 2.4. The Green’s function G(x, s) for a linear differential equation L[u] = f is the solution when f is a δ-function at s: L[G] = δ(x − s).
For bounded f the solution for L[u] = f is a superposition of the free-space
Green’s functions (or fundamental solutions) such that
u(x) =
b
G(x, s)f (s) ds.
(2.48)
a
Proof. By applying L on both sides of (2.48) we find that the right side is
L[G] = δ(x − s). Thus, (2.48) becomes an identity, and hence L[u] = f .
The equation L[G] = δ(x − s) shows that G is the inverse of L. This becomes
clear if we compare this situation with the finite-dimensional case where L is a matrix.
In the language of algebra, when L multiplies a column of the matrix L−1 , it produces
a column of the identity matrix I. Notice that a column of I is like a δ-function, where
the nonzero part is concentrated at a single point.
Thus, expressing f (x) as an integral
of the δ-function is like writing a vector f as j fj δj , where δj is the jth column of
the identity matrix. This means that formula (2.6) corresponds to the identity f = If .
Next, note that in the continuous case there is a ‘column’ for every point s. In the
identity matrix this column is δ concentrated at a single point. In L−1 the column is
G which is response to δ. Then (2.48) is precisely equivalent to u = L−1 f , where
we have an integral operator instead of a sum over all columns. Thus, the response
to a distributed source f is the sum or integral of the response to the point source δ,
which is the kind of inverse operation provided by the Green’s function.
The existence of Green’s function is equivalent to that of a unique solution of the
homogeneous boundary value problem for the equation L[u] = f (x). Therefore, the
following alternative exists: Under given homogeneous boundary conditions, either
the equation L[u] = f (x) has a uniquely determined solution u(x) for every given
x, or the homogeneous equation L[u] = 0 has a nontrivial solution.
2.5. FUNDAMENTAL SOLUTIONS
41
It is obvious from the above examples that Green’s function G(x, s) does not
depend on the input function f (t). Let T is the linear (integral) operator defined by
x
T : f −→
G(x, s) f (s) ds,
x
or T [f ](x) =
x0
G(x, s)f (s) ds,
x0
where the operator T is equivalent to the inverse linear operator L−1 introduced in
(2.35), and G is a continuous function on the square I × I = {(x, s) | x, s ∈ I}.
This formulation leads to inverse problems which are typically ill posed. Examples
of ill-posed inverse problems include integral equations of the first kind, tomography,
and inverse scattering. Since these topics are beyond the scope of this book, interested
readers are advised to consult the following literature: Groetsch [1984], Porter and
Sterling [1993], Kythe and Puri [2002], Kythe and Schäferkotter [2005] on integral
equations of the first kind where the last two books also include useful Mathematica
codes.
In Chapter 3 we will study a generalization of the variation of parameters method,
which leads to the Green’s function method for initial and boundary value problems,
while in Chapters 4 and 5 we study Bernoulli’s separation method and integral transforms which will be useful in developing different methods for determining Green’s
functions for partial differential operators, including the Sturm-Liouville systems,
parabolic, hyperbolic and elliptic equations, and fractional partial differential equations. This will be followed by the conformal mapping method and computational
construction of Green’s functions in different types of convex and starlike 2-D regions.
2.5. Fundamental Solutions
Consider the boundary value problem
L[u](x) = f (x) in D ⊂ Rn ,
(2.49a)
B[u] = 0 on
(2.49b)
∂D = S,
where f ∈ C(D), and (2.49b) represents linear initial and boundary conditions. This
problem has a unique solution in terms of Green’s function G(x, x ):
u(x) =
G(x, x ) f (x )dx ,
(2.50)
D
where dx denotes integration with respect to the variable x (source point or singularity). Green’s function G(x, x ) is singular at the fixed point x ∈ D, and the
singular part of G(x, x ), denoted by u∗ (x, x ), is known as the fundamental solution (or singular solution, or ‘free-space’ Green’s function, or Green’s function in the
large) for the operator L such that
L[u∗ ](x)(x, x ) = δ(x, x ).
(2.51)
42
2. THE CONCEPT OF GREEN’S FUNCTIONS
Note that Eq (2.51) is simply Eq (2.49a), where u(x) is replaced by u∗ (x, x ) and f (x)
by δ(x, x ). In particular, for example, the fundamental solution for the Laplacian
operator ∇2 is unique up to a harmonic function. Some important properties of the
fundamental solution u∗ for the operator L are:
(i) u∗ (x, x ) is defined everywhere except at x = x where it is singular.
(ii) u∗ (x, x ) is unique up to a harmonic function of x, i.e., any two fundamental
solutions for L with the same pole x differ by a solution of the homogeneous
equation Lu∗ = 0. In the absence of boundary conditions the homogeneous
equation will have many solutions, and u∗ will not be unique.
(iii) u∗ (x, x ) = u∗ (x , x).
(iv) u∗ (x, x ) depends only on the distance r = |x − x | for all n ≥ 1.
(v) ∇2x u∗ = 0 for all x = x and n ≥ 1.
∂u∗
(vi)
ds = 1, where S denotes the surface of the sphere |x − x | = r for any
S ∂n
∂
r > 0, n is the outward normal to S, and
= n · ∇x .
∂n
Physically, the function u∗ in (2.51) is the response to a concentrated unit source
located at x = x . An important physical application of the fundamental solution
u∗ (x) is that it enables us to solve the nonhomogeneous equation
L[u](x) = f (x),
(2.52)
where f ∈ C0∞ (Rn ). In fact, let f (x) represent the source term in the form of a total
point sources f (x )δ(x − x ), i.e.,
f (x) =
Rn
f (x )δ(x − x ) dx .
Then, in view of (2.51), there exists an influence function f (x )u∗ (x − x ) at each
point. Note that if L does not have constant coefficients, then
u(x) =
Rn
u∗ (x, x )f (x ) dx
(2.53)
is the solution of (2.52).
Example 2.15. (Potential flow) The potential
flow ina one-dimensional region
d
du
is defined by the Sturm-Liouville equation
a(x)
= f (x), 0 ≤ x ≤ l,
dx
dx
where u denotes the potential. This equation is found in problems of transverse
deflection of a cable, axial deformation of a bar, heat transfer along a fin in heat
exchangers, flow though pipes, laminar incompressible flow through a channel under
constant pressure gradient, linear flow through porous media, and electrostatics. Let
2.5. FUNDAMENTAL SOLUTIONS
43
a function u∗ (x, s), which is sufficiently continuous and differentiable as often as
needed, be a solution of
d
du∗
a
= δ(x, s).
dx
dx
For a = const (for a homogeneous isotropic medium), the fundamental solution for
this equation is given by
u∗ (x, s) =
1
(l − r) ,
2a
r = |x − s| .
Example 2.16. (Bending of an elastic beam) This problem is defined by
d2 d2 u b
= f (x),
dx2 dx2
0 ≤ x ≤ l,
where b = EI is the flexural rigidity of the beam (E being the modulus of elasticity
and I the moment of inertia). Let u∗ (x, s) be the fundamental solution of
b
d 4 u∗
= δ(x, s).
dx4
The fundamental solution for this equation, subject to the boundary conditions
b
u∗ 0, 0+ = = u∗ l, l− ,
u∗ l, 0+ = 0 = u∗ l, l− ,
2
bl
bl3
u∗ 0, 0+ =
= u∗ 0, l− , u∗ l, 0+ = −
= u∗ l, l− ,
2
12
is given by
u∗ (x, s) = λl3 2 + |ρ|3 − 3|ρ|2 ,
(2.54)
where ρ = r/l, λ = b/12, and r = x − s.
Example 2.17. The Heaviside function H(x) defines the distribution H (x) =
δ(x). Note that since H(x) is not differentiable at x = 0 in the classical sense, the
distributional definition (2.17) for H (x) when formally integrated gives
∞
0, x < 0
H(x) =
δ(x) dx =
1, x > 0.
0
It is obvious from (2.17) that the Heaviside function H(x) is a fundamental solution
for the basic differential operator ∂ on R .
Example 2.18. In Rn the radial solutions v(r) of the potential equation ∇m u =
0, which depends only on the distance r(= 0) of the point x from a fixed point x so
that r = |x − x |, are given by the equation
d2 v n − 1 dv
+
= 0.
dr 2
r dr
44
2. THE CONCEPT OF GREEN’S FUNCTIONS
This equation has the solutions
v(r) =
c1 + c2 r 2−n ,
c1 + c2 log r,
n > 2,
n = 2,
where c1 , c2 are arbitrary constants.; These solutions exhibit the so-called characteristic singularity at r = 0. The function



1
|x − x|2−n , n > 2,
(n
−
2)
Sn (1)
s (x, x ) =

 − 1 log |x − x |, n = 2,
2π
(2.55)
is called the singularity function for ∇n u = 0. The function s(x, x ) satisfies the
equation ∇n s = 0 for x = x and has the singularity at x = x . Every solution
u∗ (x, x ) of the potential equation ∇n u = 0 in D of the form
u∗ (x, x ) = s(x, x ) + φ(x),
x ∈ D,
is called a fundamental solution in D for the operator ∇n with a singularity at
x = x , where φ(x) ∈ C 2 (D), φ(x) ∈ C 1 (D ∪ ∂D), and ∇n φ = 0 in D.
We have provided in Appendix B a list of fundamental solutions for some differential operators. For an extensive study of fundamental solution for differential
operators, see Kythe [1996]. Fundamental solutions play a significant role in the
boundary element method. A detailed study of this subject can be found in Brebbia
and Dominguez [1992], and Kythe [1995].
2.6. Exercises
2.1. Show that the following functions define the Dirac δ-function as a limiting case:
1
2
(a) δ(x) = lim √
e−x /a , |x| < a;
a→0
πa
1
a
(b) δ(x) = lim
, |x| < a;
2
a→0 π a + x2

a

0, x < − ,


2


1
a
a
(c) δ(x) = lim
, − <x< ,
a→0 
 a
2
2


 0, x > a ;
2
sin ax
1 ∞ i xt
(d) δ(x) = lim
= lim
e dt.
a→∞ πx
a→∞ 2π −∞
2.6. EXERCISES
45
∞
a→0 −∞
2.2. Let δ(x) be defined as in Exercise 2.1(c). Show that lim
f (0).
2.3. Show that
∞
−∞
f (x)δ(x) dx =
δ(t) dt = 1.
Hint. Use property (2.9) (iii).
∞
2.4. Show that t δ(t) = 0. Hint. Set tφ(t) = ψ(t). Then −∞ δ(t) tφ(t) dt =
∞
ψ(t)δ(t) dt = ψ(0) = 0.
−∞
∞
φ(t )
2.5. Show that the integral I =
δ (f (t)) φ(t) dt =
, where f is a
|f (t )|
−∞
monotone function that vanishes at t = t , and φ is an arbitrary function.
∞
Solution. Set y = f (t). Then I = −∞ δ(y)ψ(y) dy, where ψ(y) =
φ(y)/|f (y)|.
δ(x, s)
∞
2.6. Show that δ f (x) = . Hint. Use −∞ δ(x − s)φ(x) dx = φ(s).
|f (s)|
∞
2.7. Show that −∞ δ(ax − b) φ(x) dx = |a|−1 φ ba−1 . Hint. Set ax − b = y.
2.8. Show that δ(x) = δ(−x).
∞
∞
2.9. Show that −∞ δ (x)φ(x) dx = − −∞ δ(x)φ (x) dx = −φ (0),
∞ ∞
δ (x)φ(x) dx = − −∞ δ (x)φ (x) dx = φ (0).
−∞
∞
∞
Hint. Use −∞ f (x)φ(x) dx = − −∞ f (x)φ (x) dx. (*)
∞
2.10. Show that −∞ H (t)φ(t) dt = φ(0). Hint. Use formula (*) of Exercise
2.9.
∞
∞
2.11. Show that −∞ f (x)φ(x) dx = −∞ sgn(x) φ(x) dx, where sgn(x) is the


 1, x > 0,
signum function defined by sgn(x) = 0, x = 0,


−1, x < 0.
Hint. Use formula (*) of Exercise 2.9.); then left-side integral
0
∞
0
=−
(−x)φ (x) dx −
xφ (x) dx = −
φ(x) dx +
−∞
= right-side.
0
−∞
∞
φ(x) dx
0
2.12. Show that f (x) = |x| = sgn(x). Hint. Follows from Exercise 2.11.
d
sgn(x) = 2δ(x).
dx
1
2.14. Evaluate I = −1 |x| ψ (x) dx. Ans. I = ψ (−1) − ψ (−1) − ψ(1) −
ψ(−1) + 2ψ(0).
2.13. Show that
46
2. THE CONCEPT OF GREEN’S FUNCTIONS
2.15. Solve L[K(x, t)] = δ(x − t), where L ≡
d2
, and K is the kernel function.
dx2
d2
Solution. Integrating
K(x, t) = δ(x − t) and using H (x) = δ(x),
dx2
d
we get
K(x, t) = H(x − t) + a(t), where a(t) is an arbitrary function.
dx
∞
Integrating again we find that K(x, t) = −∞ H(x − t) dx + x a(t) + b(t) =
(x − t)H(x − t) + xa(t) + b(t). where b(t) is an arbitrary function.
2.16.
support, show that u(x) =
∞If f (t) is an integrable function with compact
d2 u
K(x, t) f (t) dt satisfies the equation
= f (x), where K is the kernel
dx2
−∞
function of Exercise 2.15.
∞
2.17. Show that
δ (x)f (x) dx = −f (0), assuming that f (x) is continuous at
x = 0.
−∞
2.18. Let the δ-function be defined by δa (x) =
Find a representation for δ(x).
0,
x < 0,
−ax
ae
, x > 0.
Hint. The function ∆a (x) has the singularity at x = 0+ .
Ans. δ(x) = lim a e−ax .
a→∞
3
Sturm-Liouville Systems
The simple case of determining Green’s functions for ordinary differential equations
of second order is important because it brings out the methods involved with different
types of boundary conditions and the shape of the boundary, and it provides a framework for constructing Green’s functions for different ordinary and partial differential
operators.
3.1. Ordinary Differential Equations
We will discuss the theory and methods of solving nonhomogeneous ordinary differential equations of order n, n = 1, 2, . . . , of the form P (D)[u](x) = f (x), where
P (D) is the linear ordinary differential operator defined by
P (D) = a0 (x) Dn + a1 (x) Dn−1 + · · · + an−1 (x) D + an (x), a0 = 0,
(3.1)
which represents a polynomial of degree n, where D ≡ d/dx, D ≡ d /dx , and
the coefficients ak (x), k = 0, 1, 2, . . . , n, are continuous on some interval I on the
x-axis, and the function u(x) ∈ C n (I).
n
n
n
3.1.1. Initial and Boundary Conditions. An ordinary (or partial) differential
equation subject to prescribed conditions in the form of initial or boundary conditions
is known as an initial value or a boundary value problem. The initial conditions, also
known as Cauchy conditions, are the values of the unknown function u and an
appropriate number of its derivatives at the initial point.
The boundary conditions fall into the following three categories:
(i) Dirichlet boundary conditions (also known as boundary conditions of the first
kind), when the values of the unknown function u are prescribed at each point of
the boundary ∂D of a given domain D.
(ii) Neumann boundary conditions (also known as boundary conditions of the
second kind), when the values of the normal derivatives of the unknown function
47
48
3. STURM-LIOUVILLE SYSTEMS
u are prescribed at each point of the boundary ∂D.
(iii) Robin boundary conditions (also known as boundary conditions of the third
kind, or mixed boundary conditions), when the values of a linear combination of
the unknown function u and its normal derivative are prescribed at each point of
the boundary ∂D.
The classical two-point boundary value problem of ordinary differential equations
involves a second-order equation, an initial condition which is a boundary condition
at the initial point, and a terminal condition which is another boundary condition at
the end-point of the interval of definition of the problem. For example, the boundary
conditions imposed on an equation P (D)[u] = 0 defined on an interval (a, b), can
be one of the following forms: (i) u(a) = u0 , u(b) = u1 (Dirichlet); (ii) u (a) =
u0 , u (b) = u1 (Neumann); or (iii) a1 u(a) + b1 u(b) = c1 , b1 u (a) + b2 u (b) = c2
(Robin).
3.1.2. General Solution. The general solution of an ordinary differential equation of the form P (D)[u](x) = f (x), where the linear differential operator P (D)
is defined by (3.1), is determined by first considering the associated homogeneous
equation P (D)[u](x) = 0 and finding the set of its primitives which are linearly
independent solutions of the given equation, denoted by {u1 (x), u2 (x), . . . , un (x)}.
The following result holds:
Theorem 3.1. The set of primitives {u1 (x), u2 (x), . . . , un (x)} for the
homogeneous equation P (D)[u](x) = 0, is linearly independent on the interval
I = {x : x0 < x < +∞} iff the Wronskian W (x1 , . . . , xn ) = 0 for all x ∈ I.
Moreover, every solution of the homogeneous equation P (D)[u](x) = 0 is a
linear combination of the set of primitives for this equation.
The proof of this theorem is easy and left as an exercise; it is also available in
most textbooks on ordinary differential equations. The Wronskian is defined by
W (u1 , u2 , . . . , un ) = u1 (x)
u2 (x)
u1 (x)
u2 (x)
u1 (x)
u2 (x)
···
···
(n−1)
(n−1)
u1
(x) u2
(x)
···
···
···
···
···
un (x) un (x) un (x) = 0.
···
(n−1)
un
(x)
! ! For the sake of brevity we will henceforth denote the Wronskian by W (x).
The complementary function for the nonhomogeneous equation P (D)[u](x) =
f (x) is uc (x) = c1 u1 (x)+c2 u2 (x)+· · ·+cn un (x), where c1 , c2 , . . . , cn are arbitrary
constants. If f (x) = 0, then the function uc (x) provides the general solution of the
given homogeneous equation. If f (x) = 0, then a particular solution (known as
particular integral) up is usually determined by using the following methods: (i) The
method of undetermined coefficients which is used if the given equation has constant
coefficients ak , k = 0, 1, 2, . . . , n; (ii) the method of variation of parameters which
is used if the given equation has variable coefficients ak (x), k = 0, 1, 2, . . . , n; and
3.1. ORDINARY DIFFERENTIAL EQUATIONS
49
sometimes (iii) the particular solution up can be determined by inspection. Then the
general solution of the equation P (D)[u](x) = f (x) is given by
u(x) = uc (x) + up (x) = c1 u1 (x) + c2 u2 (x) + · · · + cn un (x) + up (x),
(3.2)
where the coefficients cj , j = 1, 2, . . . , n, are determined from the n prescribed
initial or boundary conditions.
Example 3.1. Consider the ordinary differential equation D2 u − 2Du − 3u =
2e − 10 sin x. The set of primitives is {e3x , e−x }, and the Wronskian is
3x
e
e−x W (x) = 3x
= −3e2x = 0.
3e
−e−x x
Thus, uc (x) = c1 e3x + c2 e−x . Using the method of undetermined coefficients,
we take the particular integral as up (x) = Aex + B sin x + C cos x, which after
substituting into the given equation and solving the identity so obtained gives A = − 12 ,
B = 2, C = −1. Hence, the general solution is u(x) = uc (x) + up (x) = c1 e3x +
c2 e−x − 12 ex + 2 sin x −cos x. If the initial conditions are prescribed as u(0) = 2 and
u (0) = 4, then we have an initial value problem. Using these conditions on u(x), we
find that c1 + c2 = 72 and 3c1 + c2 = 52 , which give c1 = 32 and c2 = 2. Hence, the
solution of this initial value problem is u(x) = 32 e3x +2e−x − 12 ex +2 sin x−cos x.
3.1.3. Method of Variation of Parameters. We will discuss this method
in some detail since it leads to the development of the so-called Green’s function
method for initial value and boundary value problems involving ordinary differential
equations. Consider the general second order equation
a0 (x)u + a1 (x)u + a2 (x)u = f (x),
a0 = 0,
(3.3)
where u ∈ C 2 on some interval I on the x-axis. Let the set of primitives be
{u1 (x), u2 (x)}, where u1 and u2 are linearly independent so that the Wronskian
W (x) = W (u1 , u2 ) = 0. Thus, the complementary function is uc (x) = c1 u1 (x) +
c2 u2 (x), where c1 , c2 are arbitrary constants. Now, using the method of variation of
parameters we take up (x) = v1 (x)u1 (x) + v2 (x)u2 (x). Then up = v1 u1 + v2 u2 +
v1 u1 + v2 u2 . Imposing the restriction that v1 u1 + v2 u2 = 0, which becomes the first
condition, a further differentiation gives up = v1 u1 + v2 u2 + v1 u1 + v2 u2 . Then
substituting these values of up , up and up into Eq (3.3) we find that
a0 v1 u1 + v2 u2 + v1 u1 + v2 u2 + a1 v1 u1 + v2 u2 + a2 v1 u1 + v2 u2 = f (x).
This equation, after using the fact that u1 and u2 satisfy the homogeneous equation
a0 (x)u + a1 (x)u +a2 (x)u = 0, simplifies to v1 u1 +v2 u2 = f (x)/a0 , which is the
second condition. From these two conditions we have the system of two equations to
determine v1 and v2 :
u1 v1 + u2 v2 = 0,
u1 v1 + u2 v2 =
f (x)
,
a0
50
3. STURM-LIOUVILLE SYSTEMS
This system can be written as
u
a0 1
u1
u2 v1
0
=
.
u2 v2
f (x)
The determinant on the left side is the Wronskian W (x). Using Cramer’s rule (§1.6)
this system is solved to yield
0 u2 u1 0 f u u f f u2
f u1
2
1
v1 =
=−
, v2 =
=
,
a0 W (x)
a0 W (x)
a0 W (x)
a0 W (x)
which on integration with respect to x gives
x
f (s)u2 (s)
v1 (x) = −
ds, v2 (x) =
a0 (s)W (s)
x
f (s)u1 (s)
ds.
a0 (s)W (s)
Thus, the particular integral is given by
x
x
∆(x, s)
up (x) =
f (s) ds =
g(x, s)f (s) ds,
a0 (s)W (s)
u (s) u2 (s) , g(x, s) = ∆(x, s) ,
where
∆(x, s) = 1
u1 (x) u2 (x) a0 (s)W (s)
(3.4)
(3.5)
and g(x, s) is the kernel in the integral (3.5) for up . Then the general solution of Eq
(3.3) is
x
u(x) = uc (x) + up (x) = c1 u1 (x) + c2 u2 (x) +
g(x, s)f (s) ds.
(3.6)
This analysis provides another form for the variation of parameters method to determine the particular integral up .
Example 3.2. Consider the equation u +u = tan x, for which the complementary function is uc (x) = c1 cos x+c2 sin x, where c1 , c2 are arbitrary constants. Using
the variation of parameters method, we find that v1 = − sin x tan x = cos x − sec x
and v2 = cos x tan x = sin x, which after integration give v1 = sin x − ln | sec x +
tan x| + c3 and v2 = − cos x + c4 . Then
up (x) = A cos x + B sin x − cos x ln | sec x + tan x|,
where A and B are the arbitrarily assigned values of c1 and c2 , respectively. Hence,
the solution can be written as
u(x) = C1 cos x + C2 sin x − cos x ln | sec x + tan x|,
where C1 = c1 + A and C2 = c2 + B. Alternatively, using formula (3.5) we have
a0 (s) = 1, W (s) = 1, and
∆(x, s) = u1 (s)u2 (x) − u1 (x)u2 (s) = sin x cos s − cos x sin s = sin(x − s).
3.2. INITIAL VALUE PROBLEMS
51
Then g(x, s) = sin(x − s), and the particular integral is given by
x
up (x) =
sin(x − s) tan s ds = − cos x sin x + sin x − ln | sec x + tan x| cos x
= − cos x ln | sec x + tan x|,
which is the same as found above by the variation of parameters method.
3.2. Initial Value Problems
We will develop the method for determining Green’s function for the nth-order ordinary differential equation P (D)[u] = f subject to prescribed n initial conditions.
The existence and uniqueness theorem for the general second-order ordinary differential equation F (x, u, u , u ) = 0, or u = f (x, u, u ) states that if f, fu , fu are
continuous in an open region R of the 3-D xuu -space and if the point (x0 , u0 , u0 )
is in R, then in some interval about x0 there exists a unique solution u = φ(x) of
the above general second-order differential equation that satisfies the prescribed initial conditions u(x0 ) = u0 , u (x0 ) = u0 . The following uniqueness theorem holds
for the general nth-order initial value problem involving either the space variable x
or time variable t (for proof see Coddington [1989], or Coddington and Levinson
[1955]).
Theorem 3.2. Let α1 , . . . , αn be any n constants, and let x0 be any real
number. In any interval I = (a, b) containing x0 there exists at most one solution φ(x) of P (D)u = 0 satisfying φ(x0 ) = α1 , φ (x0 ) = α2 , . . . , φ(n−1) (x0 ) =
αn .
3.2.1. One-Sided Green’s Functions. The following results are direct generalization of formula (3.6), which is used to find the particular integral for the
nonhomogeneous equation P (D)[u](x) = f (x) with variable coefficients.
Theorem 3.3. (For initial value problems.) Let {u1 (t), . . . , un (t)} be the
set of primitives for the homogeneous equation P (D)[u](t) = 0, and let W (t)
denote the Wronskian for this primitive. If ∆(t, s) denotes the determinant
···
un (s) u1 (s)
···
un (s) u1 (s)
∆(t, s) = ···
···
···
,
(n−2)
(n−2)
(s) · · · un
(s) u1
u1 (t)
···
un (t)
and if t0 is the initial point of the interval t > t0 , then
t
∆(x, s)
up (t) =
f (s) ds
a
t0 0 (s) W (s)
(3.7)
is the unique particular integral of the initial value problem P (D)[u](t) = f (t),
subject to the n initial conditions u(j) (t0 ) = uj (t0 ) for j = 0, 1, . . . , n − 1 .
52
3. STURM-LIOUVILLE SYSTEMS
Proof. Let up be of the form up = v1 (t)u1 (t) + · · · + vn (t)un (t). Impose the
following n − 1 conditions on the derivatives v1 (t), . . . , vn (t):
v1 u1 (t) + · · · + vn (t)u(k)
n (t) = 0. k = 0, 1, 2, . . . , n − 2,
(k)
and substitute into the nonhomogeneous equation P (D)[u] = f , which provides the
additional condition
v1 (t)u1
(n−1)
(t) + · · · + vn (t)u(n−1)
(t) = f (t).
n
Then using Cramer’s rule (see §1.6) solve the n equations for v1 , . . . , vn , integrate
each with respect to t to determine v1 (t), v2 (t), . . . , vn (t). The result follows after
substituting the values of vk (t), k = 1, , 2, . . . , n, into the assumed form of up (t).
Formula (3.7) is remarkable in that the kernel
g(t, s) =
∆(t, s)
,
a0 (s) W (s)
t > t0 ,
(3.8)
which is known as the one-sided Green’s function for the problem, can be determined
as soon as the primitives {u1 (t), . . . , un (t)} of solutions are known. Sometimes the
one-sided Green’s function (3.8) is written as
g(t, s) = H (t − t0 )
∆(t, s)
,
a0 (s) W (s)
where H(t − t0 ) is the Heaviside function (see §2.1.1). This method is also valid
if the independent variable is the space variable x instead of the time variable t, in
which case the factor H(x − x0 ) provides the non-zero solution for x > x0 .
For the non-homogeneous equation P (D)u(t) = f (t) with prescribed initial
conditions, note that the one-sided Green’s function g(t, s) is always obtained from
the associated homogeneous equation. Thus, for a fixed s the one-sided Green’s
function g(t, s) is the solution of the homogeneous initial value problem:
P (D) g(t, s) = 0, for t > t0 ,
dg 1
g(t, s)
= 0,
=
.
dt t=s
p(s)
t=s
(3.9)
Example 3.3. Consider the initial value problem (D2 + k 2 )x(t) = 0 such
that (a) x(0) = x0 , x (0) = 0, t > 0, and (b) x(0) = x0 , x (0) = u0 , t > 0.
This problem describes the free motion of a mass m which is attracted by a force
proportional to its distance from the origin and (a) starts from rest at x = x0 , and (b)
starts with an initial velocity u0 at x = x0 away from the origin. The primitive set is
{cos kt, sin kt}; thus, xc (t) = c1 cos kt + c2 sin kt. In both cases we find from (3.6)
H(t)
that the one-sided Green’s function is g(t, s) =
sin k(t − s), which satisfies the
k
properties (3.9).
3.2. INITIAL VALUE PROBLEMS
53
1 u
u + 2 = t3 , t > 1, subject to the
t
t
initial conditions u(1) = u0 and u (1) = v0 . Since the primitive set is {t, t ln t}, we
have u1 (t) = t, u2 (t) = t ln t. Also, the Wronskian is W (t) = t = 0, and a0 (t) = 1.
The one-sided Green’s function is given by
Example 3.4. Consider the equation u −
H(t − 1) s
g(t, s) =
t
s
s ln s = H(t − 1) t (ln t − ln s) ,
t ln t which satisfies the properties (3.9). By formula (3.7)
t
t
t5
t
t
up (t) =
g(t, s)f (s) ds =
s3 t (ln t − ln s) ds =
−
− ln t,
16
16
4
1
1
and the solution is u(t) = c1 t + c2 t ln t + up (t), where the initial conditions yield
t5
1
c1 = u0 , c2 = v0 − u0 , and thus, u(t) =
+ u0 − 16
t + v0 − u0 − 14 t ln t,
16
which satisfies the prescribed initial conditions.
Note that formula (3.5) for ordinary differential equations of order greater than 2
always yields up , and hence the general solution, as the following example shows.
Example 3.5. To solve the initial value problem: x3 u − 3x2 u + 6xu −
6u = x5 , x > 0, such that u(1) = 1, u (1) = 0, u (1) = 2, by one-sided Green’s
function method, note that the primitive set is {x,x2 , x3 }; thus, u
1 (x) = x, u2 (x) =
x x2 x3 x2 , u3 (x) = x3 , and the Wronskian is W (x) = 1 2x 3x2 = 2x3 = 0. Also,
0 2
6x s s2 s3 ∆(x, s) = 1 2s 3s2 = x3 s2 − 2x2 s3 + xs4 , and a0 (x) = x3 . Hence, by
x x 2 x3 formula (3.5)
g(x, s) =
∆(x, s)
x3 s2 − 2x2 s3 + xs4
=
.
a0 (s)W (s)
2s6
With f (x) = x5 , the particular integral up (x) is given by
x
up (x) =
g(x, s)f (s) ds =
0
0
x
x3 s − 2x2 s2 + xs3 ds =
x5
.
24
The general solution of this initial value problem is u(x) = c1 u1 (x) + c2 u2 (x) +
x5
c3 u3 (x) + up (x) = c1 x + c2 x2 + c3 x3 + . Applying the initial conditions we find
24
that c1 = 31
, c2 = − 14
, c3 = 74 , and the final solution of this initial value problem
8
3
14 2
7 3
1
5
is u(x) = 31
8 x − 3 x + 4 x + 24 x . Note that this problem can also be solved
by the method of variation of parameters, but it may sometimes require more work.
However, this may not always be the case, as is shown below in Example 3.12.
54
3. STURM-LIOUVILLE SYSTEMS
Theorem 3.4. The one-sided Green’s function for a constant-coefficient
operator P (D) = a0 Dn + a1 Dn−1 + · · · + an is given by g(t − s), a0 = 0,where
g(u) is a solution of
P (D)[g(u)] = 0, such that g(0) = 0, . . . , g n−2 (0) = 0, g n−1 (0) = 1.
Example 3.6. Consider the constant-coefficient P (D) = D 3 + 4D. Then the
one-sided Green’s function for the related initial vale problem is found by solving
g (u) + 4g (u) = 0,
which is g(y) =
1
4
g(0) = g (0) = 0,
g (0) = 1,
(1 − cos 2y), y = t − s. Hence, one-sided Green’s function is
g(t − s) =
H(t) [1 − cos 2(t − s)] .
1
4
It can verified that this Green’s function satisfies the given initial conditions, as
follows:
dg(t − s) g(0) = g(t − s)t=s = 0; g (0) =
= 12 sin 2(t − s)t=s = 0; and
t=s
dt
d2 g(t − s) g (0) =
= cos 2(t − s) t=s = 1.
t=s
dt2
3.2.2. Wronskian Method. An alternate effective method to determine the particular integral up for the initial value problems involving the equation P (D)[u](t) =
f (t) is provided by the following theorem (see Brauer and Nohel [1989]).
Theorem 3.5. Let {u1 (t), u2 (t), . . . , un (t)} be the primitive set for the
equation P (D)[u](t) = 0. Then the general solution for the nonhomogeneous
equation P (D)[u](t) = f (t), t ∈ I = {t : t0 < t <= +∞}, satisfying the
initial conditions u(t0 ) = α1 , u (t0 ) = α2 , . . . , u(n−1) (t0 ) = αn , where αk , k =
1, . . . , n, are real numbers, is given by
u(t) = uc (t) + up (t),
uc (t) = c1 u1 (t) + . . . + cn u2 (t)
is the complementary function for the associated homogeneous equation, and
n
up (t) =
t
uj (t)
j=1
t0
Wj (s)
f (s) ds, t > t0 ,
a0 (s)W (s)
(3.10)
W (t) = W (u1 , . . . , un ) is the Wronskian, and Wj (t) is obtained from W (t)
by replacing the jth column by [0 0 · · · 1]T .
It can be verified that formula (3.10) is the same as (3.7), and it is left as an
exercise. Note that either formula (3.7) or formula (3.10) is useful for determining
the particular integral up (x), since both of these formulas are equivalent to each other.
1
1
Example 3.7. Consider u + u − 2 u = f (t), t > 0, where the function
t
t
f is continuous for all t ∈ I. The primitive set for the associated homogeneous
3.3. BOUNDARY VALUE PROBLEMS
55
t
2
1/t equation is {t, 1/t}, so the Wronskian is W (t) = 2 = − , t > 0. Here,
1
−1/t
t
0
t 0
1
1/t W1 (t) = = − , and W2 (t) = = t, and thus,
1 −1/t2 1 1
t
t
s t
s
up (t) =
−
f (s) ds.
2
s
t
t0
s
2
The one-sided Green’s function is given by g(s, t) =
satisfies the properties (3.9)
t
s
−
, t > t0 , which
s
t
3.2.3. Systems of First-Order Differential Equations. One-sided Green’s
function for a system of n first-order linear ordinary differential equations can be determined by generalizing the above method, and is explained in Appendix F.
3.3. Boundary Value Problems
The method of variation of parameters is generally used to solve the second order
ordinary differential equation
a0 (x) u + a1 (x) u + a2 (x) u = F (x),
a < x < b,
(3.11)
where it is assumed that a0 (x) is continuously differentiable, and a1 (x), a2 (x) are
continuous functions on the interval I = (a, b). Further, it is assumed that a0 (x) > 0
for x > a, and it does
approachzero as x → a. If we multiply
equa not
x the above
x
1
a1 (s)
a1 (s)
tion by
exp
ds , and set p(x) = exp
, q(x) =
a0 (x)
a a0 (s)
a a0 (s)
a2 (x)
F (x)
p(x), and f (x) =
p(x), the above differential equation becomes
a0 (x)
a0 (x)
d
du
p(x)
+ q(x) u = f (x),
dx
dx
a < x < b,
(3.12)
where the function p(x) is continuously differentiable and positive on the interval
I = (a, b), and q(x) and f (x) are continuous on I. Eq (3.12) is known as the
self-adjoint Sturm-Liouville equation.
Theorem 3.6. The Sturm-Liouville operator of Eq (3.12) is self-adjoint.
Proof. Consider a general linear, second-order, differential equation of the form
L[u](x) ≡ a0 (x)
d2 u(x)
du(x)
+ a1 (x)
+ a2 (x) u(x), a0 (x) = 0,
2
dx
dx
(3.13)
where a0 , a1 and a2 are real functions of x defined over the interval a ≤ x ≤ b, such
that the first (2 − j) derivatives of aj (x) are continuous, and the zeros of a0 (x) are the
56
3. STURM-LIOUVILLE SYSTEMS
singular points. This equation is a Sturm-Liouville equation if a(x) = a1 (x)/a0 (x),
and q(x) = a2 (x)/a0 (x). An adjoint operator, denoted by L∗ , is defined by
L∗ [u] =
d2
d
[a0 u] −
[a1 u] + a2 u
2
dx
dx
d2 u
du
= a0 2 + (2a0 − a1 )
+ (a0 − a1 + a2 ) u.
dx
dx
(3.14)
Comparing (3.13) and (3.14) we obtain the necessary and sufficient condition for
L = L∗ , which is a0 (x) = a1 (x), so that when this condition is satisfied the operators
L and L∗ are said to be self-adjoint.
Self-adjoint operators are also called Hermitian operators. An important property is:
b
b
v L[u](x) dx =
u L[v](x) dx,
(3.15)
a
a
which is a consequence of the above theorem. Further, if we integrate vL[u] with
respect to x over an interval [a, x], where v(x) is any function of x, we obtain
a
x
x
vL[u] dx = (va0 )u − (va0 ) u + (va2 )u
a
x
+
(va) ) − (va1 ) + (va2 ) u.
(3.16)
a
If we differentiate both sides with respect to x , then
vL[u] − uL∗ [v] =
d
[a0 (u v − uv ) + (a1 − a0 )uv] ,
dx
(3.17)
which is known as Lagrange’s identity for the Sturm-Liouville operator L. Again,
if we integrate (3.16) from a to b, we get
b
b
(vL[u] − uL∗ [v]) dx = a0 (u v − uv ) + (a1 − a0 )uv .
a
a
(3.18)
For the self-adjoint Sturm-Liouville operator L formula (3.18) yields
b
b
(vL[u] − uL[v]] dx = a0 (u v − uv ) .
a
a
(3.19)
3.3.1. Sturm-Liouville Boundary Value Problems. Consider the nonhomogeneous, second-order boundary value problem on x ∈ (a, b):
p(x) u +q(x) u = f (x), a < x < b,
(3.20a)
a1 u(a)+a2 u (a) = 0,
b1 u(b)+b2 u (b) = 0,
a ≤ x ≤ b.
(3.20b)
3.3. BOUNDARY VALUE PROBLEMS
57
Before we solve this boundary value problem, we will first consider the associated
homogeneous equation [p(x) u ] + q(x) u = 0 with the boundary conditions (3.20b).
This equation has the primitives (linearly independent solutions) u1 (x) and u2 (x),
such that their Wronskian
u1 (x) u2 (x) = u1 (x)u2 (x) − u1 (x)u2 (x) = 0 on [a, b].
W (x) = u1 (x) u2 (x) The general solution of the homogeneous equation is u(x) = c1 u1 (x) + c2 u2 (x).
The first condition in (3.20b) is satisfied if
a1 u(a) + a2 u (a) = a1 c1 u1 (a) + a1 c2 u1 (a) + a2 c1 u1 (a) + a2 c2 u2 (a) = 0,
which implies that
a1 a2
u1 (a) u2 (a)
u1 (a) u2 (a)
c1
c2
=
0
.
0
The determinant of the matrix on the left side is W (a) = 0. Thus, this system has a
unique solution unless a1 or a2 is zero. Suppose a1 = 0 and a2 = 0. Then, we solve
c1 u1 (a) + c2 u2 (a) = 0.
This equation has at least one solution. Let u1 (x) be that solution of the homogeneous
equation satisfying the first condition in (3.20b). Similarly, there will always be a
solution of this equation, say u2 (x), which satisfies the second condition in (3.20b).
Let these two solutions u1 (x) and u2 (x) be such that they are not multiples of each
other. Then there is no single function, except the trivial solution u(x) ≡ 0, that
satisfies the homogeneous boundary value problem, and that 0 is not an eigenvalue
of this boundary value problem (see §3.4).
To find the solution of the nonhomogeneous boundary value problem (3.20a,b),
note that a particular solution of Eq (3.20a) is given by
x
x
u2 (x)u1 (s)
u1 (x)u2 (s)
up (x) =
f (s) ds −
f (s) ds.
(3.21)
p(s)W (s)
p(s)W (s)
a
a
Thus, the general solution of Eq (3.20a) is u(x) = c1 u1 (x) + c2 u2 (x) + up (x). The
first condition in (3.20b) gives
a1 u(a) + a2 u (a)
= a1 [c1 u1 (a) + c2 u2 (a)] + a2 [c1 u1 (a) + c2 u2 (a)] + a1 up (a) + a2 up (a)
= c2 [a1 u1 (a) + a2 u2 (a)] = 0.
If a1 u1 (a) + a2 u2 (a) = 0, then u2 (x) would satisfy the boundary value problem
(3.20a,b), i.e., there would exist no function u2 (x), contrary to our assumption.
Hence, a1 u1 (a) + a2 u2 (a) = 0, and so c2 = 0. Similarly, the second condition in
(3.20b) gives
b1 u(1) + b2 u (b) = c1 b1 u1 (b) + b2 u2 (b) + c2 b1 u1 (b) + b2 u2 (b) +
58
3. STURM-LIOUVILLE SYSTEMS
x u2 (b)u1 (s)
u1 (x)u2 (s)
+ b2
f (s) ds −
f (s) ds
a p(s)W (s)
a p(s)W (S)
b
u1 (b)u2 (s)
= c1 b1 u1 (b) + b2 u1 (b) − b1
f (s) ds
a p(s)W (s)
b u1 (b)u2 (s)
− b2
f (s) ds
a p(s)W (s)
b
u2 (s) ds
= c1 b1 u1 (b) + b2 u1 (b) − b1 u1 (b) + b2 u1 (b)
= 0.
a p(s)W (s)
x
b
u2 (s) ds
= 0, and c2 = 0. Hence, using (3.21) we find that solution
p(s)W
(s)
a
of the problem as
or c1 =
u(x) = c1 u1 (x) + up (x)
b
x
u1 (x)u2 (s)
=
f (s) ds +
a p(s)W (s)
a
x
b
u2 (x)u1 (s)
=
f (s) ds +
p(s)W (s)
a
x
where
u2 (x)u1 (s)
f (s) ds −
p(s)W (s)
u1 (x)u2 (s)
f (s) ds =
p(s)W (s)

u2 (x)u1 (s)


 p(s)W (s) ,
G(x, s) =

u (x)u2 (s)

 1
,
p(s)W (s)
x
a
u1 (x)u2 (s)
f (s) ds
p(s)W (s)
b
G(x, s) f (s) ds,
a
a ≤ x < s,
(3.22)
s < x ≤ b,
is Green’s function for the system (3.20a,b), and {u1 (x), u2 (x)} is the set of primitives
each of which satisfy the respective boundary conditions. Hence, the solution of the
Sturm-Liouville system is given by
b
u(x) =
G(x, s) f (s) ds.
a
3.3.2. Properties of Green’s Functions. For each fixed s Green’s function
G(x, s) satisfies the following five properties:
d
dG
(i)
p(x)
+ q(x) G = 0, for x = s;
dx
dx
(ii) G(a, s) = G(b, s) = 0 (boundary conditions);
(iii) G
− G
= 0 (continuity condition);
x=s+0
x=s−0
dG dG 1
(iv)
−
=−
(jump condition); and
dx x=s+0 dx x=s−0
p(s)
(v) G(x, s) = G(s, x) (symmetry condition).
3.3. BOUNDARY VALUE PROBLEMS
59
3.3.3. Green’s Function Method. We will summarize the method for constructing Green’s function and finding solution for boundary value problems involving
the Sturm-Liouville equation (3.20a), where p(x) = 0 and a ≤ x ≤ b, subject to the
boundary conditions (3.20b), or generally for ordinary differential equations of the
form P (D)u(x) = f (x) subject to appropriate boundary conditions.
Step 1. Find the primitive {u1 (x), u2 (x)} of the associated homogeneous
equation, so that the complementary function is uc (x) = c1 u1 (x) + c2 u2 (x).
Step 2. Since Green’s function satisfies the prescribed boundary conditions, let
g1 (x) be the value of uc (x) at the left boundary point x = a, and g2 (x) the value
of uc (x) at the other endpoint x = b. The set {g1 (x), g2 (x)} is called the basis of
solutions. Determine the Wronskian W (x) = W (g1 (x), g2 (x)).
Step 3. Since Green’s function must satisfy the continuity condition at x = s,
and since the jump discontinuity in Gx at x = s must be of magnitude −1/p(x), take
g1 (x) g2 (s), a ≤ x < s,
G(x, s) =
g1 (s) g2 (x), s < x ≤ b.
Step 4. Since the jump in the derivative Gx at x = s is g1 (s) g2 (s) −
= W (x), divide the right side of G(x, s) in Step 3 by p(s)W (s). This
gives the required Green’s function as

g1 (x) g2 (s)


 p(s)W (s) , a ≤ x < s,
G(x, s) =
(3.23)
 g1 (s) g2 (x)


, s < x ≤ b.
p(s)W (s)
The significance of determining the basis of solutions in Step 2 is evident from
Step 4, where G(x, s), split into two parts, is such that the first part contains g1 (x)
which satisfies the boundary condition at x = a, and the second part contains g2 (x)
which satisfies the boundary condition at x = b. Hence, Green’s function constructed
in Step 4 satisfies all five properties of §3.3.2.
g1 (s) g2 (s)
This method does not apply to initial value problems. To confirm we give a
counter-example by solving Example 3.4 using this method. The prescribed initial
conditions yield g1 (t) = u0 t, g2 (t) = (v0 − u0 )t ln t, so that the Wronskian is
W
(t) = u0 (v0 − u0 )t = 0, and p(t) = 1. Then by (3.23), we have G(t, s) =
st ln t, t < s,
t5
t
, which will give the solution as u(t) = 16
− 16
− 14 t ln t, that
st ln s, t > s
does not satisfy the initial conditions.
Example 3.8. Consider the boundary value problem
u + u = f (x),
u(0) = 0 = u(1).
The homogeneous equation u + u = 0 has the general solution u = c1 sin x +
c2 cos x. The boundary condition at x = 0 gives c2 = 0, so u = c1 sin x. Choosing
c1 = 1 gives u1 (x) = sin x. The condition at x = 1 gives c1 sin(1) + c2 cos(1) = 0,
where by taking c1 = cos(1) and c2 = sin(1), we get u2 (x) = sin(x − 1). Thus, the
60
3. STURM-LIOUVILLE SYSTEMS
basis of solutions is {sin x, sin(x − 1)}. The Wronskian of the solutions is
sin x sin(x − 1) = sin (1).
W (x) = cos x cos(x − 1) Hence, by (3.23) Green’s function is

sin(x − 1) sin s


,

sin (1)
G(x, s) =
 sin x sin(s − 1)


,
sin (1)
0 ≤ x < s,
s < x ≤ 1.
In particular, if we take f (x) = x, then
1
u(x) =
G(x, s)f (s) ds =
0
0
x
sin(x − 1) sin s
s ds +
sin 1
1
x
sin x sin(s − 1)
s ds
sin 1
sin x
=x−
.
sin 1
This result can be verified, for example,
by the variation
of parameters method, as
sin x cos x = −1, and a0 (x) = 1, so
follows: The Wronskin W (x) = cos x − sin x sin s
∆(x, s) = sin x
and up (x) =
cos s = sin s cos x − sin x cos s = g(x, s),
cos x x
(sin s cos x − sin x cos s) s ds = x, which leads to u(x) = uc (x)+
up (x) = c1 sin x + c2 cos x + x, and the boundary conditions yield c2 = 0, c1 =
sin x
sin x
−
; this finally gives the solution as u(x) = −
+ x, which is the same as
sin 1
sin 1
given above. Compare this example with Example 2.9.
Example 3.9. Consider the nonhomogeneous boundary value problem
1 + x2 u + 2xu − 2u = f (x), u(0) = 0, u(1) + u (1) = 0.
The general solution of the homogeneous equation is u = c1 x + c2 (1 + x arctan x).
The boundary condition at x = 0 gives c2 = 0, and by choosing c1 = 1, we
have
g1 (x) = x. The condition
at x = 1 gives c1 + c2 (1 + arctan 1) + c1 +
c2 12 + arctan 1 = 2c1 + c2 32 + π2 . Choosing c1 = π + 3 and c2 = −4, we find
the other linearly independent solution g2 (x) = (π + 3) x − 4 (1 + x arctan x). The
Wronskian is given by
x
(π + 3)x − 4 (1 + x arctan x)
x
W (x) = 1 π + 3 − 4 1 + x2 + arctan x
4
=
1 + x2 = 0.
3.3. BOUNDARY VALUE PROBLEMS
Hence, Green’s function is

π+3


xs − s (1 + x arctan x) ,
4
G(x, s) =

 π + 3 sx − x (1 + s arctan s) ,
4
0 ≤ x < s,
s < x ≤ 1.
Then the solution of this problem is given by u(x) =
instance, with f (x) = 1 we get
1
u(x) =
G(x, s) ds
61
1
G(x, s)f (s) ds. For
0
0
1
π+3
π+3
=
xs − s (1 + x arctan x) ds +
sx − x (1 + s arctan s) ds
4
4
0
x
x
π+1
= arctan x −
x,
2
8
x
where we have used the formula x arctan x dx = 12 (x1 + 1) arctan x − . This
2
result is verified by the variation
of
parameters
method
or
the
Wronskin
method,
x
1 + x arctan x 1
x
as follows: Since W (x) = 1
, and a0 (x) =
= −
+
arctan
x
1 + x2
2
1+x
s 1 + s arctan s = x(1 + s arctan x) − s(1 +
1 + x2 , we obtain ∆(x, s) = − x 1 + x arctan x x arctan x). Thus,
x
x
up (x) =
[x(1 + s arctan s) − s(1 + x arctan x)] ds = arctan x,
2
x
so that the general solution is u(x) = uc (x) + up (x) = c1 x + c2 (1 + x arctan x) +
x
arctan x. After applying the prescribed boundary conditions we find that c2 = 0
2
π+1
x
π+1
and c1 −
, which gives u(x) = arctan x −
x, same as above.
8
2
8
Example 3.10. Consider the problem u = f (x), 0 < x < 1, such that
u(0) = u(1) = 0. In view of the above properties (i) and (ii), Green’s function is
a1 (s) x, 0 ≤ x < s,
G(x, s) =
a2 (s) (1 − x), s < x ≤ 1,
where a1 and a2 are constants which depend on s only. By the above condition (v),
we have a1 (s) = A (1 − s), a2 (s) = A s, where A is a constant independent of s.
Then the continuity condition (iii) is automatically satisfied, and the jump condition
(iv) gives
dG dG = A s (−1),
= A (1 − s).
dx x=s+0
dx x=s−0
62
3. STURM-LIOUVILLE SYSTEMS
Thus, A = 1, and Green’s function is
(1 − s) x,
G(x, s) =
s (1 − x),
0 ≤ x < s,
s < x ≤ 1,
and the particular integral is
x
1
x(1 − s) f (s) ds +
up (x) =
(1 − x) s f (s) ds.
0
x
Example 3.11. Consider the transverse displacement of a string of unit length,
0 < x < 1, under tension at fixed x = 0, and connected to a spring with spring
constant b at x = 1, and under a unit transverse force applied at x = 1 and a transverse
force f (x) per unit length applied along the string. This problem is defined by
u = f (x),
u(0) = 0, b u(1) + u (1) = 1.
Under the above properties (i) and (ii) Green’s function is given by
G(x, s) =
a1 (s) x,
0 ≤ x < s,
a2 (s) [1 + b (1 − x)] ,
s < x ≤ 1.
By the symmetry condition (v), we have a1 (s) = A [1 + b (1 − s)] , a2 = As,
where A is a constant not dependent on s. By condition (iv), we have As(−b) −
1
A [1 + b (1 − s)] = −1, which gives A =
. Hence,
1+b
 x [1 + b (1 − s)]

,

1+b
G(x, s) =

 s [1 + b (1 − x)] ,
1+b
0 ≤ x < s,
s < x ≤ 1,
and
u(x) =
x
1 + b (1 − x)
+
1+b
1+b
+
x
1+b
x
s f (s) ds
0
1
[1 + b (1 − s)] f (s) ds.
x
If b = 0, we obtain Green’s function for the boundary value problem u = f, 0 <
x < 1, such that u(0) = 0, u (1) = 1.
Sometimes it becomes easier to find the solution of a boundary value problem
if it is possible to get a particular solution simply by inspection. This method is
useful if we need to find the solution u(x) without determining Green’s function.
The following example illustrates this short cut.
3.3. BOUNDARY VALUE PROBLEMS
63
Example 3.12. Consider the problem
u − u = 2x,
0 < x < 1,
u (0) = 0, u(1) = 0.
By inspection, a particular solution of the differential equation is u = 2x, and the
basis of solutions is {ex , e−x }. Then the general solution is
u(x) = 2x + A ex + B e−x .
Using the boundary conditions, we find that A = −
Hence,
u(x) = 2x −
2(1 + e)
2e (1 − e)
, B =−
.
2
1+e
1 + e2
2(1 + e) x 2e(1 − e) −x
e −
e .
1 + e2
1 + e2
Note that it will a lengthier process to find this solution through Green’s function
method.
Example 3.13. (a) L[u] = u = 0, 0 < x < 1, such that u(0) = 0 = u(1).
Green’s function is
(1 − s) x, if 0 ≤ x < s,
G(x, s) =
(1 − x) s, if s < x ≤ 1.
(b) L[u] = u = 0, 0 < x < 1, such that u(0) = 0, u (1) = 0. Green’s function is
x, if 0 ≤ x < s,
G(x, s) =
s, if s < x ≤ 1.
(c) L[u] = u = 0, −1 ≤ x ≤ 1, such that u(−1) = 0 = u(1). Green’s function is
G(x, s) = −
1
|x − s| + xs − 1 .
2
(d) L[u] = u = 0, 0 ≤ x ≤ 1, such that u(0) = u(1), u (0) = −u (1). Green’s
function is
1
G(x, s) = − |x − s| + 14 .
2
Example 3.14. L[u] = xu + u = 0, 0 ≤ x ≤ 1, such that u(1) = 0, u(0)
finite. Green’s function is
− ln s, if 0 ≤ x < s,
G(x, s) =
− ln x, if s < x ≤ 1.
This problem is associated with the Bessel function
J0 (x) of order zero. Note that
Bessel’s equation of order n is x2 y + xy + x2 − n2 y = 0, n ≥ 0, and thus
Bessel’s equation of order zero is xy + y + y = 0.
64
3. STURM-LIOUVILLE SYSTEMS
Example 3.15. L[u] = (xu ) −
n2
u = 0, 0 ≤ x ≤ 1, such that u(1) = 0,
x
u(0) finite. Green’s function is

1 x n
n


− (xs) ,
 n
s
G(x, s) =

1 s n


− (xs)n ,
n
x
if 0 ≤ x ≤ s,
if s < x ≤ 1.
This problem is associated with the Bessel function
Jn (x) of order n. Note that
Bessel’s equation of order n is x2 y + xy + x2 − n2 y = 0, n ≥ 0.
Example 3.16. (Linear Oscillator.) The linear oscillator equation for a vibrating string of length l is defined by
u (x) + λu(x) = 0,
u(0) = 0 = u(l).
The boundary conditions imply that the string is clamped at both end points. To
determine Green’s function, first we find the solutions of the corresponding homogeneous equation which is L[u](x) = u (x) = 0 under the prescribed conditions: A
solution is u1 (x) = x which satisfies the condition at x = 0, while the other solution
is u2 (x) = 1 − x which satisfies the condition at x = 1. We say that the primitive set
of solutions is {x, 1 − x}. Also, since the Wronskian W (x) = u1 u2 − u1 u2 = −1,
and p(x) = 1, we get A = −1. Thus, Green’s function is
x(1 − s), 0 ≤ x < s,
G(x, s) =
s(1 − x), s < x ≤ 1,
and the solution of of this problem is u(x) = λ
wavelength of the linear oscillator.
1
G(x, s) ds. Note that λ is not the
0
3.4. Eigenvalue Problem for Sturm-Liouville Systems
Consider the Sturm-Liouville system:
d du p(x)
+ λρ(x) u = 0, a < x < b,
dx
dx
a1 u(a) + a1 u (a) = 0, b1 u(b) + b2 u (b) = 0,
(3.24)
where λ is real, ρ(x) is the weight function, p and q are the functions as defined
before, and ai , bi , i = 1, 2, are real constants such that a1 and b1 , or a2 and b2 are
both not zero and a21 + a22 > 0 and b21 + b22 > 0. It is obvious that the system (3.24)
always has a trivial solution u = 0. The nontrivial solution of this problem consists
of the eigenfunctions φn (x) and the corresponding values of λ which are called as
the eigenvalues λn for the problem. The pair (λn , φn ) is known as the eigenpair .
3.4. EIGENVALUE PROBLEM FOR STURM-LIOUVILLE SYSTEMS
65
For example, the set 1, cos x, sin x, cos 2x, sin 2x, . . . of (orthogonal) functions consists of the eigenfunctions for the problem u + λ u = 0, u(−π) =
u(π), u (−π) = u (π), with the corresponding eigenvalues λn = n2 for n =
1, 2, . . . . Another example is as follows: For the eigenvalue problem u + λ u = 0,
where 0 < x < L, and (a) subject to the (Dirichlet) boundary conditions u(0) = 0 =
u(L), the eigenpair is
nπ
λn =
, φn (x) = sin λn x, n = 1, 2, . . . ;
L
and (b) subject to the (Neumann) boundary conditions u (0) = 0 = u (L), the
eigenpair is
nπ
λn =
, φn (x) = cos λn x, n = 0, 1, 2, . . . .
L
Theorem 3.7. Let the functions p, q, ρ, and p in Eq (3.24) be realvalued and continuous on the interval a ≤ x ≤ b. Let um (x) and φn (x) be
the eigenfunctions of the Sturm-Liouville problem (3.24) with corresponding
eigenvalues λm and λn , respectively, such that λm = λn . Then
b
φm (x) φn (x) ρ(x) dx = 0, m = n,
a
i.e., the eigenfunctions φm and φn are orthogonal with respect to the weight
function ρ(x) on the interval a ≤ x ≤ b.
Proof. Let λn and µm be two different eigenvalues with the corresponding
eigenfunctions φn and ψm , n = 1, 2, . . . . Then
(p φn ) − q φn + λn ρφn = 0,
(p ψn ) − q ψn + µm ρψn = 0.
If we subtract φn times the second equation from ψm times the first equation and
integrate from a to b, we get
b
ψm (p φ ) − φn (p ψm
) + (λn − µm ) ρφn ψm dx = 0.
a
Since ψm (p φn ) − φn (p ψm
) = (ψm p φn − φn p ψm
), and since φn and ψm are
zero at both end points, then first two terms of the integrand are zero and we have
b
(λn − µm ) a ρφn ψm dx = 0.
b
Thus, since λn = µm , this yields a ρφn ψm dx = 0, which means that the eigenfunctions corresponding to different eigenvalues are orthogonal with respect to the
weight function ρ(x)
Example 3.17. Consider u + λu = 0, x ∈ (0, π), and u(0) = 0 =
u(π). For λ = 0, the problem has a trivial solution. For λ < 0 the primitive
66
3. STURM-LIOUVILLE SYSTEMS
√
√
√
is {cos√ λx, sin λx}, and the solution of this problem is u(x)
√ = c1 cos λx +
c2 sin λx. The boundary conditions give c1 = 0, and √
c2 sin λπ = 0, so c2 = 0
would yield the trivial solution. Hence c2 = 0, so sin λπ = 0, which gives the
eigenvalues λn = n2 , n = 1, 2, . . . , where λn → ∞ as n → ∞ and the corresponding eigenfunctions are φn (x) = sin nx. Green’s function is given by
√
√

sin λx cos λs


√
, 0 ≤ x < s,

λ
G(x, s) =
√
√

cos λx

 sin λs
√
, s < x ≤ π.
λ
Example 3.18. (Cauchy-Euler Equation) Consider the problem
x2 u + xu + λu = 0,
1 ≤ x ≤ e;
u(1) = 0 = u(e).
This√equation
reduces to the
√
√ Sturm-Liouville
√ equation (3.11) which has the primitive
{xi λ , x−i λ } = {cos( λ ln x), sin( λ ln x)}. Green’s function for this problem
can be easily determined. The solution of this problem is
√
√
u(x) = c1 cos( λ ln x) + c2 sin( λ ln x),
√
The first condition gives c1 = 0, so taking c2 = 0, we get sin λ = 0, which yields
the eigenvalues λn = n2 π 2 , n = 1, 2, . . . ; the corresponding eigenfunctions are
φn (x) = sin(nπ ln x), n = 1, 2, . . . .
3.4.1. Eigenpairs. Consider the differential equation
L[u](x) + λ ρ(x)u(x) = 0,
(3.25)
where L is the Sturm-Liouville operator, λ a constant, and the weight function ρ is
a known function of x such that ρ(x) > 0 except possibly at isolated points where
ρ(x) = 0. For a given choice of λ, the eigenfunction φλ (x) satisfies Eq (3.25) and the
prescribed boundary conditions and it corresponds to the eigenvalue λ. The existence
of an eigenfunction φλ (x) does not necessarily depend on an arbitrary choice of the
eigenvalue λ. This restricts the acceptable eigenvalues λ to a discrete set. A list of
the values of p(x), q(x), ρ(x), and λ for ordinary differential equations encountered
in this book is given in Table 3.1.
Properties of Hermitian (self-adjoint) operators are as follows:
(i) Their eigenvalues are real;
(ii) Their eigenfunctions are orthogonal; and
(iii) Their eigenfunctions for a complete set. However, this property holds only for
linear, second-order Sturm-Liouville differential operators. Proofs of these properties
can be found in most textbooks on differential equations.
3.4. EIGENVALUE PROBLEM FOR STURM-LIOUVILLE SYSTEMS
67
Table 3.1.
Equation
p(x)
q(x)
ρ(x)
λ
Legendre’s
1 − x2
0
1
n(n + 1)
1
n(n + 1)
m2
1 − x2 −
1 − x2
n2
x
−
x
2
e−x
0
xe−x
0
xk+1 e−x 0
Associated Legendre’s
Bessel’s
Hermite’s
Laguerre’s
Associated Laguerre’s
a2
x
e−x
e−x
xk e−x
2
α
α
α−k
The eigenfunctions φn (x) of a self-adjoint (Hermitian) operator form a complete
set. This means that any continuous or at least piecewise continuous function f (x)
can be approximated to any degree of accuracy by a series of the form
∞
f (x) =
an φn (x),
(3.26)
n=0
where the coefficients an are determined by
am =
b
f (x)φm (x)ρ(x) dx.
(3.27)
a
The orthogonality of the eigenfunctions φm (x) implies that only the mth term survives. Thus, the set {φn (x)} is a complete set if the limit of the mean square error is
zero:
b
m
2
lim
f (x) −
an φn (x) ρ(x) dx = 0.
m→∞
a
n=0
3.4.2. Orthonormal Systems. A function f is said to be normalized with
b
respect to the weight function ρ on the interval a ≤ x ≤ b iff
|f (x)|2 ρ(x) dx = 1.
a
2
For example, the function f (x) =
sin x is normalized with respect to ρ(x) = 1
π
π 2
on [0, π] since
sin x dx = 1. Let {φn }, n = 1, 2, . . . , be an infinite set
π
0
of functions defined on the interval a ≤ x ≤ b. The set {φn } forms an orthonormal
system with respect to the weight function ρ(x) on a ≤ x ≤ b if (i) it is an orthogonal
system with respect to the weight function ρ(x) on a ≤ x ≤ b, and (ii) every function
of this system is normalized with respect to ρ(x) on a ≤ x ≤ b; i.e., the set {φn } is
orthonormal with respect to ρ(x) on a ≤ x ≤ b if
68
3. STURM-LIOUVILLE SYSTEMS
b
φm (x)φn (x)ρ(x) dx =
a
For example, the set {φn } =
2
π
0,
1,
m = n,
m = n.
sin x , ρ = 1, is an orthonormal system on
[0, π].
Consider the following boundary value problem involving the Sturm-Liouville
equation:
p(x)u + [q + λρ] u = 0,
A1 u(a) + A2 u (a) = 0,
a ≤ x ≤ b,
B1 u(b) + B2 u (b) = 0,
where A1 , A2 , B1 , B2 are real constants such that A1 and A2 are not both zero and
B1 and B2 are not both zero. Let {λn } denote the infinite set of eigenvalues arranged
in a monotone increasing sequence λ1 < λ2 < · · · . If φn , n = 1, 2, . . . , is one of
the eigenfunctions corresponding to the eigenvalue λn for each n = 1, 2, . . . , then
the infinite set of eigenfunctions {φ1 , φ2 , . . . } is an orthogonal system with respect
to ρ(x) on a ≤ x ≤ b. If φn is an eigenfunction, then µn φn is also an eigenfunction
for arbitrary scalar µn = 0 which corresponds to the eigenvalue λn . So we form
a new set of functions {µ1 φ1 , µ2 φ2 , . . . } which is orthogonal with respect to ρ(x)
on a ≤ x ≤ b. Let this set be denoted by E. We can choose constants µ1 , µ2 , . . .
such that every eigenfunction in the set E is also normalized with respect to ρ(x)
on a ≤ x ≤ b. Then the set E of eigenfunctions {µ1 φ1 , µ2 φ2 , . . . } will be an
orthonormal system with respect to ρ(x) on a ≤ x ≤ b. Now, to determine the
constants µ1 , µ2 , . . . so that the set E of eigenfunctions becomes orthonormal with
respect to ρ(x) on a ≤ x ≤ b, note that no eigenfunction φn , n = 1, 2, . . . , is
identically zero on a ≤ x ≤ b. Let
b
2
[φn (x)] ρ(x) dx = Mn > 0, n = 1, 2, . . . ,
a
which implies that,
2
1
φn (x) ρ(x) dx = 1, n = 1, 2, . . . .
Mn
a
1
1
Hence, the set √
φ1 , √
φ2 , . . . is an orthonormal set with respect to ρ(x)
M1
M2
on a ≤ x ≤ b, i.e., the set E = {µ1 φ1 , µ2 φ2 , . . . } is orthonormal with respect to
ρ(x) on a ≤ x ≤ b, if µn is chosen as
1
1
µn = √
= b
, n = 1, 2, . . . .
(3.28)
2
Mn
[φn (x)] ρ(x) dx
b
√
a
Example 3.19. Let φn (x) = cn sin nx, n = 1, 2, . . . on 0 ≤ x ≤ π,
where cn , n = 1, 2, . . . , are nonzero constants. We will form the set of orthonormal
eigenfunctions E = {µn φn }, where µn is given by (3.28). Here,
3.4. EIGENVALUE PROBLEM FOR STURM-LIOUVILLE SYSTEMS
π
Mn =
2
[cn sin nx] (1) dx =
0
69
c2n π
,
2
1
1
2
µn = √
=
;
c
π
Mn
n
1
2
2
thus, µn φn (x) =
cn sin nx =
sin nx, n = 1, 2, . . . .
cn π
π
3.4.3. Eigenfunction Expansion. Assuming that an expansion of a function f
in an infinite series of orthonormal eigenfunctions {φ1 , φ2 , . . . } exists, let
∞
f (x) =
cn φn (x),
(3.29)
n=1
for each x ∈ [a, b]. To determine the coefficients cn , we proceed formally, and
multiply both sides of (3.29) by φk (x)ρ(x), to obtain
∞
f (x)φk (x)ρ(x) =
cn φn (x)φk (x)ρ(x),
n=1
Integrating both sides from a to b we get
b
b ∞
f (x)φk (x)ρ(x) dx =
cn φn (x)φk (x)ρ(x) dx
a
a
∞
n=1
b
=
cn φn (x)φk (x)ρ(x) dx
n=1
a
∞
=
b
cn
φn (x)φk (x)ρ(x) dx
a
n=1
= ck ,
since {φn } is an orthonormal set with respect to ρ(x) on a ≤ x ≤ b. This result is
true for the kth coefficient ck for k = 1, 2, . . . . Hence
b
cn =
f (x)φn (x)ρ(x) dx, n = 1, 2, . . . .
a
In fact it can be shown that the series
∞
cn φn (x) converges uniformly to f (x) on
n=1
a ≤ x ≤ b, and thus the above formal procedure is justified.
Example 3.20. Obtain the eigenfunction expansion for f (x) = πx − x2 ,
0 ≤ x ≤ π, in the series of orthonormal eigenfunctions {φn } of the Sturm-Liouville
2
system u + λu = 0, u(0) = 0 = u(π). Since the set {φn } =
sin nx ,
π
n = 1, 2, . . . , is orthonormal with respect to ρ(x) = 1 on [0, π], let
70
3. STURM-LIOUVILLE SYSTEMS
∞
f (x) =
(3.30)
cn φn (x),
n=1
where
2
cn =
(πx − x )
sin nx (1) dx
π
0
π
π
2
2
=
π
x sin nx dx −
x sin nx dx
π
0
0
π
2
π
πx
=
sin nx −
cos nx −
π
n2
n
0
π
2
2x
2
x2
− 2 sin nx + 3 cos nx −
cos nx
n
n
n
2
2
π
π 0
2
π
2
π
2
=
− cos nx −
cos nx −
cos nx − 3
3
n
n
n
n

 2 4
2 2
if n is odd,
=
(1
−
cos
nπ)
=
π n3
3

π n
0 if n is even.
Thus, the series (3.30) becomes
∞
f (x) =
=
8
sin nx if n is odd
πn3
n=1
8
π
∞
sin(2n − 1)x
,
(2n − 1)3
n=1
and we have formally
πx − x2 =
8
π
∞
sin(2n − 1)x
, 0 ≤ x ≤ π.
(2n − 1)3
n=1
It can be shown that this series converges uniformly and absolutely to πx − x2 on the
interval 0 ≤ x ≤ π.
The following theorem provides a bilinear eigenfunction expansion of a Green’s
function for the regular Sturm-Liouville system.
Theorem 3.8. If G(x, s) is Green’s function for a regular Sturm-Liouville
system, defined by
L[u] ≡ [p(x)u (x)] + q(x)u = f (x), a ≤ x ≤ b
a1 u(a) + a2 u (a) = 0, b1 u(b) + b2 u (b) = 0,
(3.31)
3.4. EIGENVALUE PROBLEM FOR STURM-LIOUVILLE SYSTEMS
71
and the associated eigenvalue problem L[ψ] = λ ψ, a ≤ x ≤ b, subject to
the boundary conditions a1 ψ(a) + a2 ψ (a) = 0, b1 ψ(b) + b2 ψ (b) = 0, has
countably many eigenvalues λn with the corresponding eigenfunctions ψn (x),
then Green’s function can be represented as an eigenfunction expansion
∞
G(x, s) =
1
ψn (x)ψn (s).
λ
n=1 n
(3.32)
Proof. Assume that the solution u(x) of the regular Sturm-Liouville system
with the forcing term f (x) can be represented as an eigenfunction expansion
∞
u(x) =
∞
an ψn (x),
f (x) =
n=1
fn ψn (x),
(3.33)
n=1
where the coefficients an are to be determined, while the coefficients fn are the
b
Fourier coefficients determined from fn =
ψn (s) ds. Substituting (3.33) into the
Sturm-Liouville equation (3.31), we get
∞
L
a
an ψn (x) =
n=1
∞
fn ψn (x).
n=1
Since the left side of this equality is
∞
∞
L
an ψn (x) =
an L [ψn (x)] =
n=1
n=1
∞
an λn ψn (x),
n=1
then equating the right side of these two equations we get
b
1
1
an =
fn =
ψn (s)f (s) ds.
λn
λn a
Hence, from (3.33) we find that
b
∞
1
u(x) =
ψn (s)f (s) ds ψn (x)
λ
a
n=1 n
b ∞
b
1
=
ψn (x)ψ(s) f (s) ds =
G(x, s)f (s) ds,
λ
a
a
n=1 n
(3.34)
which, in view of (3.8) with the upper limit x = b, yields (3.32).
Example 3.21. Consider the problem x2 u + λ u = 0, 0 < x < 1, such
that u(1) = 0 and u is bounded. Green’s function is

1


for s ≤ x,
x
G(x, s) =

 1 for s ≥ x.
s
72
3. STURM-LIOUVILLE SYSTEMS
1
1
2
[G(x, s)] ρ(s) dx ds = ∞. The fact that the
Since ρ(x) = 1, we find that
0
0
finiteness of the double integral in (3.42) does not hold shows that the condition (3.42)
is sufficient but not necessary for completeness of the set of eigenfunctions. Moreover,
the solution of the differential equation
at x = 1 is
subject to the boundary condition
1 2
−1/2
u(x) = x
sin
λ + 1/4 ln x , which is unbounded, and 0 ρ u dx = +∞.
Hence, this problem has no eigenfunctions.
3.4.4. Data for Eigenvalue Problems. We provide some useful data for the
solution of eigenvalue problems with the three types of boundary conditions. Consider
u + λu = 0, 0 < x < L, such that a1 u(0) + b1 u (0) = 0, a2 u(L) + b2 u (L) = 0,
and let the eigenfunction expansion of an arbitrary function f (x) be
b
∞
1
f (x) =
cn φn (x), where cn =
f (x)φn (x) dx.
φn 2 a
n=1
Set hi = ai /bi , i = 1, 2. Then the nine cases for the three types of boundary
conditions are as follows.
Case 1 . Dirichlet at x = 0 and at x = L: a1 = 0, b1 = 0 a2 = 0, b2 = 0.
L
Then φn (x) = sin λn x; φn 2 = ; and λn are the roots of sin nL = 0, i.e.,
2
nπ
λn =
, n = 0, 1, 2, . . . .
L
Case 2. Dirichlet at x = 0 and Neumann at x = L: a1 = 0, b1 = 0 a2 = 0,
L
b2 = 0. Then φn (x) = sin λn x; φn 2 = ; and λn are the roots of cos λL = 0,
2
(2n − 1)π
i.e., λn =
, n = 1, 2, . . . .
2L
Case 3. Dirichlet at x = a and Robin at x = L: a1 = 0, b1 = 0 a2 = 0, b2 = 0.
λn L − sin λn L cos λn L
Then φn (x) = sin λn x; φn 2 =
; and λn are the roots of
2λn
b2
λ + h2 tan λL = 0. If L = −
> 0, then (0, x) is the eigenpair.
a2
Case 4. Neumann at x = 0 and Dirichlet at x = L: a1 = 0, b1 = 0 a2 = 0,
L
b2 = 0. Then φn (x) = cos λn x; φn 2 = ; and λn are the roots of cos λL = 0,
2
(2n − 1)π
i.e., λn =
, n = 1, 2, . . . .
L
Case 5. Neumann at x = 0 and Neumann at x = L: a1 = 0, b1 = 0 a2 = 0,
L
b2 = 0. Then φn (x) = cos λn x; φn 2 = , where L is replaced by 2L for n = 0;
2
nπ
and λn are the roots of sin λL = 0, i.e., λn =
, n = 0, 1, 2, . . . .
L
Case 6. Neumann at x = 0 and Robin at x = L: a1 = 0, b1 = 0 a2 = 0,
λn L + sin λn L cos λn L
b2 = 0. Then φn (x) = cos λn x; φn 2 =
; and λn are the
2λn
3.5. PERIODIC STURM-LIOUVILLE SYSTEMS
73
roots of λ tan λL = h2 .
Case 7. Robin at x = 0 and Dirichlet at x = L: a1 = 0, b1 = 0 a2 = 0, b2 = 0.
λn L − sin λn L cos λn L
Then φn (x) = sin λn (L − x); φn 2 =
; and λn are the
2λn
roots of λ cot λL = h1 .
Case 8. Robin at x = 0 and Neumann at x = L: a1 = 0, b1 = 0 a2 = 0,
b2 = 0. Then φn (x) = cos λn (L − x);
λn L + sin λn L cos λn L
;
2λn
and λn are the roots of λ tan λL = −h1 .
φn 2 =
Case 9. Robin
at x = 0 and Robin at x = L: a1 = 0, b1 = 0 a2 = 0, b2 = 0.
Then φn (x) = λn cos λn x − h1 sin λn x);
1
h2
2
2
2
φn =
(λn + h1 ) L + 2
− h1 ;
2
λn + h22
and λn are the roots of tan λL = −
(0, x − 1/h1 ) is the eigenpair.
λ(h1 − h2 )
1
1
. If L =
−
> 0, then
2
λ + h1 h2
h1
h2
3.5. Periodic Sturm-Liouville Systems
This system is defined by
d
du
p(x)
+ [q(x) + λρ(x)] u = 0,
dx
dx
a ≤ x ≤ b,
(3.35)
p(a) = p(b),
u(a) = u(b), u (a) = u (b).
In this case the solution is of period (b − a).
Example 3.22. Consider the Sturm-Liouville system
u + λu = 0,
−π < x < π;
u(= π) = u(π), u (−π) = u (π).
Here p(x) = 1, and
√ so p(−π)
√ = p(π). The system is periodic, and the primitive
√ of
solutions
is
{cos
λx,
sin
λx}.
For
λ
>
0,
the
solution
is
u(x)
=
c
cos
1
√
√
√ λx +
c2 sin λx, and the boundary conditions give
c
sin
λπ
=
0,
c
sin
λπ = 0.
2
1
√
Since c1 = 0, c2 = 0, we must have sin λπ = 0, or λ = n2 , n = 1, 2, . . . .
Thus, for λ > 0, and for every eigenvalueλn = n2 , there are two linearly independent eigenfunctions un (x) = cos nx and sin nx. Further, for λ = 00, we get a
single eigenvalue λ = 0 and the corresponding eigenfunction u(x) = 1. Lastly, for
λ < 0, there are no eigenvalues. Hence, the eigenvalues are λn = 0, n2 , with the
corresponding eigenfunctions {1, cos nx, sin nx}.
74
3. STURM-LIOUVILLE SYSTEMS
3.6. Singular Sturm-Liouville Systems
A Sturm-Liouville equation is said to be singular when the interval I is semi-infinite
or infinite, or when the functions p and q vanish, or when one of the functions p and
q become unbounded at one or both end points of the a finite interval I. A singular
Sturm-Liouville equation with appropriate boundary conditions is called a singular
Sturm-Liouville system. The boundary conditions in this system are not like those
in a regular Sturm-Liouville system. The boundary conditions imposed on a singular
Sturm-Liouville system differ from those imposed on the regular Sturm-Liouville
systems in that they are not separated boundary conditions but are often prescribed
as a bounded function u(x) at the singular boundary point. For example, consider
(vL[u] − uL[v]) dx = p(b) u (b)v(b) − u(b)v (b)
a+ε
− p(a + ε) u (a + ε)v(a + ε) − u(a + ε)v (a + ε) ,
b
(3.36)
whre ε > 0 is a small number. When this equation is subjected to the boundary
conditions
lim p(x) u (x)v(x) − u(x)v (x) = 0,
(3.37)
+
x→a
p(b) u (a + ε)v(a + ε) − u(a + ε)v (a + ε) = 0,
(3.38)
then (3.36) yields
b
(vL[u] − uL[v]) dx = 0.
(3.39)
a
In particular, if p(a) = 0, the boundary conditions (3.37) and (3.38) are replaced by
the conditions: (i) u(x) and u (x) are finite as x → a, and (ii) b1 u(b) + b2 u (b) = 0,
where b1 , b2 are constants. Thus, a singular Sturm-Liouville system is self-adjoint if
the condition (3.39) holds such that the functions u(x) and v(x) satisfy the boundary
conditions.
Consider the equation
d
∂u
p(x)
− q(x) u + λ ρ(x)u = 0,
dx
∂x
0 < x < 1,
(3.40)
where p, q ∈ C, p ∈ C ∞ ; p > 0, q > 0, q ≥ 0 for 0 ≤ x ≤ 1. The set of
eigenfunctions {un (x)} is complete in the interval [0, 1] in the following sense: The
eigenfunctions un (x) are orthonormal
in L2 (0, 1) if there is no nontrivial u ∈ L2 (0, 1)
!
"
orthogonal to all un , i.e., u, un = 0 for n = 0, 1, 2, . . . implies that u = 0 almost
everywhere. Thus, if {un } is an orthonormal set in L2#(0, 1), and if the #expansion
N
of u ∈ L2 (0, 1) has the Fourier coefficients cn , then #u − n=0 cn un # → 0 as
N → ∞ iff {un } is complete.
3.6. SINGULAR STURM-LIOUVILLE SYSTEMS
75
In many physical problems the conditions p, q, ρ and q are satisfied in the open
interval (0, 1) but not at one or both end points, where the functions p and ρ may
approach zero or infinity, and q may become infinite (at x = 0). Since u(0) =
0 = u(1), we require that u remain bounded at x = 0 and xu → 0. The integral
1 1
2
[G(x, s)] ρ(x)ρ(s) dx ds is finite since Green’s function G(x, s) is continu0
0
ous in both x and s for 0 ≤ x, s ≤ 1.
Example 3.23. (Legendre’s equation) Let the singular Sturm-Liouville system
involving Legendre’s equation
L[u] =
d 1 − x2 u + λu = 0,
dx
−1 < x < 1,
be subject to boundary conditions that both u and u remain finite as x → ±1.
Since p(x) = 1 − x2 vanishes at x = ±1, q(x) = 0, and ρ(x) = 1, this problem
has the eigenvalues λn = n(n + 1) and the eigenfunctions φn (x) = Pn (x) for
n = 0, 1, 2, . . . , which are the Legendre polynomials of order n, finite at x = ±1, and
orthogonal to each other with respect
 to the weight function ρ(x) = 1 on −1 ≤ x ≤ 1;
1
n = m
 0,
that is.
Pn (x)Pm (x) dx =
. The function Pn (x) represents
2

, n=m
−1
2n + 1
a surface zonal harmonic or a Legendre’s coefficient, or simply a Legendrian, of order
n. A few surface zonal harmonics are given in §C.1.1 and §C.1.2. In general, any
surface zonal harmonic is obtained from the two of the preceding orders by using the
formula (n + 1)Pn+1 (x) = (2n + 1)xPn (x) − nPn−1 (x).
There are four different forms of surface zonal harmonics, namely,
(i) a polynomial in x, given in §C.1.1.;
(ii) an expression involving cosines of multiples of θ, where x = cos θ (see §C.1.1);
1 dn 2
n
(iii) a form involving derivative with respect to x: Pn (x) = n
x −1 ;
n
2
n!
dx
and
1 π
n
(iv) a form involving a definite integral: Pn (x) =
x + x2 − 1 · cos s ds.
π 0
The problems involving surface zonal harmonics Pn (x) start with the determination of the value of a potential function in terms of powers of x. Thus, a given
function f (x) can be developed into a Fourier series involving Pn (x) of the form
∞
2n + 1 1
f (x) =
An Pn (x), −1 < x < 1, where An =
f (x) Pn (x) dx
2
n=0
−1

(2n + 1)m(m − 1)(m − 2) · · · (m − n + 1)


 (m − n + 1)(m − n + 3) · · · (m + n + 1)
=
if n < m + 1 and n + m is even,



0
if n > m or if n + m is odd.
The Sturm-Liouville boundary value problem involving the associated Legendre’s
76
3. STURM-LIOUVILLE SYSTEMS
equation of order n
1 − x2 u + λ −
m2
u = 0, m = 0, 1, 2, . . . , 0 ≤ x ≤ 1,
1 − x2
has the eigenvalues λn = n(n + 1) and the eigenfunctions φn (x) = Pnm (x) for
n = 0, 1, 2, . . . , which are called associated Legendre functions of first kind and order
dm
n, defined by Pnm (x) = (1 − x2 )m/2 m Pn (x) and known as spherical harmonics
dx
(see Chapter 9). If the above equation is subject to the boundary conditions that u(0)
and u(1) are finite, Green’s function for the interval [0, 1] when n = 0 is given by

m/2
1 1+x 1−s



, if x < s,
 2m 1 − x 1 + s
G(x, s) =
m/2


1 1+s 1−x


, if s < x.
2m 1 − s 1 + x
However, this result fails for m = 0, since the equation L[u] ≡ 1 − x2 u −
m2
u = 0 has the solution u = 1 which is regular everywhere and satisfies the
1 − x2
boundary conditions.
Example 3.24. (Bessel’s equation) Consider the singular Sturm-Liouville
system involving Bessel’s equation
d
dx
x
du
dx
u(a) = 0,
+ λx −
n2
x
u = 0,
0 < x < a, λ ≥ 0,
(3.41)
and u, u are finite as x → 0+ ,
where n ≥ 0 is constant. In this equation p(x) = x, q(x) = −n2 /x, and ρ(x) = x;
also p(0) = 0, q(x) → −∞ as x → 0+ , and ρ(0) = 0. Hence, it is a singular
Sturm-Liouville system. If λ = k 2 , k = 0, 1, 2, . . . , the eigenfunctions are φn (x) =
Jn (kx), n = 1, 2, . . . , which are the Bessel functions of the first kind and order
n; the eigenvalues are λn = kn2 . Also, Jn (0+ ) < +∞, Jn (0+ ) < +∞. The
eigenfunctions are orthogonal with respect to the weight function ρ(x) = x. For
λ = 1 Eq (3.41) is Bessel’s equaion of order n. For λ = 0 Bessel’s equation has two
solutions: xn and x−n for n > 0. Green’s function subject to the conditions that G
is bounded and x G → 0 at x = 0, and u(1) = 0, is given by
 n −n
x (s − sn )


for x < s,
2n
G(x, s) =
n
−n
n

 s (x − x ) for s < x.
2n
(3.42)
This Green’s function is not bounded and approaches infinity as x → 0. Hence,
the convergence is uniform in every interval ε ≤ x ≤ 1, where ε > 0, but not
3.6. SINGULAR STURM-LIOUVILLE SYSTEMS
77
in the whole interval [0, 1]. Note that Green’s function G(x, s) does not depend
on ρ(x). For λ = 1 and n = 0 Eq (3.41) reduces to Fourier’s equation which is
d2 u
1 du
+
+ u = 0 with general solution u = AJ0 (x) + BK0 (x), where the
dx2
x dx
function J0 (x) is called a cylindrical harmonic or Bessel function of order zero,
and K0 (x) another cylindrical harmonic or Bessel function of the second kind and
of order zero. In general, the function u = Jn (x) is a particular solution of Bessel’s
equation (3.41) and is called a cylindrical harmonic or Bessel function of order n,
and unless n is an integer, we have u = AJn (x) + BJ−n (x) as the general solution
of Bessel’s equation (3.41). If n is an integer, then Jn (x) = (−1)n J−n (x), and
then u = AJn (x) + BKn (x) is the general solution of Bessel’s equation (3.41). For
λ = 0 and n = 0 Eq (3.41) reduces to xu + u = 0, which has linearly independent
solutions 1 and − ln x, and with p(x) = x and W (x) = −1/x = 0 for x = 0 gives
the Green’s function that is the same as in Example 3.14.
Example 3.25. (Hermite’s equation) Consider Hermite’s equation
u − 2xu + λu = 0,
−∞ < x < ∞,
which is not self-adjoint. If we set v(x) = e−x
2
/2
v + (1 − x2 ) + λ v = 0,
(3.43)
u(x), Eq (3.43) becomes
−∞ < x < ∞,
(3.44)
which is self-adjoint. The eigenvalues of Eq (3.44) are λn = 2n for nonnegative
2
integers n, and the corresponding eigenfunctions are φn (x) = e−x /2 Hn (x), where
Hn (x) are the Hermite polynomials of order n, which are solutions of Eq (3.43). If
we impose the boundary conditions v → 0 as x → ±∞, the the polynomials Hn (x)
2
satisfy these conditions because xn e−x /2 → 0 as x → ±∞. Since Hn (x) are
square-integrable, we have the orthogonality relation
∞
−∞
−x2 /2
Hn (x)Hk (x) e
dx =
0, n = k,
√
2n n! π, n = k.
From Eqs (3.43) and (3.34) it is obvious that the Hermite polynomials u(x) =
2
Hn (x) and the corresponding orthogonal function v(x) = e−x /2 Hn (x) are characterized as solutions of the eigenvalue problem
e−x
2
/2
u
+ λ e−x
2
/2
u = 0,
(3.45)
and Eq (3.44) respectively, where the eigenvalues are λ = 0, 2, 4, . . . , −∞ < x < ∞,
and subject to the boundary condition that the eigenfunctions u must remain finite at
both x = ±∞ of order that of a finite power of x. If the series solution of Eq (3.45) is
∞
2n − λ
taken as u =
an xn , then the recurrence relation is an+2 =
an .
(n + 1)(n + 2)
n=1
Hence, either the series breaks off in the case when λ = 2n is a non-negative even
78
3. STURM-LIOUVILLE SYSTEMS
integer and thus represents the Hermite polynomials Hn (x), or the series has infinitely
many nonvanishing coefficients and converges for all x. All the coefficients an which
occur as soon as (2n − λ) becomes positive have the same sign. In the second case
the terms an xn occur for arbitrary large n. Thus, u(±∞) becomes infinite of an
order greater than any finite power of x. Hence, u cannot be an eigenfunction for the
problem and the Hermite polynomials are the only solutions of the above eigenvalue
problem.
For λ = 0 the function u(x) = e−x /2 is a solution of Eq (3.43) in u instead
of v, i.e., of L[u] ≡ u + (1 − x2 ) u = 0. Assuming a solution of the form
2
/2
u = w e−x
, we obtain w + 2w x = 0, which
has the solutions w = const and
2
x
e−t dt. Thus, u(x) = c1 e−x
x
e−t dt. Hence, the particular
c2
c2
∞
2
2
solutions which vanish at x = ±∞ are given by g1 (x) = a e−x /2
e−t dt and
x
x
√
−x2 /2
−t2
g2 (x) = b e
e
dt, respectively. Since p(x) = 1 and W (x) = ab π,
w = c1
2
2
/2
2
−∞
using (3.23) we find that Green’s function is given by
∞

2
1 −(x2 +s2 )/2 x −t2

√

e
e
dt
e−t dt, if x ≤ s,
 π
−∞
s
G(x, s) =
∞

2
1 −(x2 +s2 )/2 s −t2

 √ e
e
dt
e−t dt, if x > s,
π
−∞
x
∞
√
2
where we have used
e−t dt = π (Courant and Hilbert [1968:374]).
−∞
Example 3.26. (Laguerre’s equation) Laguerre’s equation is
xu + (1 − x)u + nu = 0,
n = 0, 1, 2, . . . ,
and its solutions are the Laguerre polynomials u = Ln (x) = ex
eigenvalue equation for the Laguerre polynomials Ln (x) is
xu + (1 − x)u + λu = 0,
dn
(xn e−x ). The
dxn
(3.46)
with eigenvalues λ = n, where n is a positive integer. In self-adjoint form this
equation is
−x x e u + λe−x u = 0,
(3.47)
subject to the boundary conditions that u remains finite at x = 0 and u(∞) does not
become infinite of an order greater than a positive power of x. For the associated
orthogonal functions v = e−x/2 Ln (x), we have the Sturm-Liouville eigenvalue
equation
1 x
(xv ) +
−
v + λv = 0,
(3.48)
2 4
3.7. EXERCISES
79
where we require regularity at x = 0. Note that w = x−1/2 Ln satisfies the selfadjoint eigenvalue equation
2 x w
−
x2 − 2x − 1
w + λxw = 0,
4
where this solution must vanish at x = 0. The corresponding eigenvalues are the
positive integers λ = n, and it leads to the associated Laguerre functions of the first
dm
kind u = Lm
Ln (x).
n (x) =
dxm
The Laguerre polynomials e−x/2 Ln (x) are solutions of the equation (3.48) in u
instead of v:
1 x
xu + u +
−
+ λu = 0.
2 4
1 x
+
u. Then the
2 4
x/2
equation L[u] = 0 has the particular solution
ux =−te . Assuming the general solue
tion of the form u = w ex/2 , we get w = c1
dt. The two particular solutions
t
c2
∞ −t
e
which are regular at x = 0 and vanish at x = ±∞ are g1 (x) = bex/2
dt
t
x
and g2 (x) = aex/2 , respectively. Since W (x) = ab/x and p(x) = x, using (3.23)
Green’s function for this problem is
∞ −t

e
(x+s)/2

e
dt, if x < s,

t
s
G(x, s) =
∞ −t

e

 e(x+s)/2
dt, if s < x.
t
x
For a particular value λ = −1, we define L[u] = xu + u −
Example 3.27. No Green’s function exists for the equation L[u] = u in
−∞ < x < ∞, subject to the condition that u remain finite for all x, because the
homogeneous equation u = 0 has solution u = const, which is regular at x = ±∞.
3.7. Exercises
3.1. Find one-sided Green’s function for the following initial value problems:
(b) t2 u + 2tu − 2u = 0, t > 0; (c) u + 2u − u − 2u = 0.
(d) t2 u + 4t u + 2 + t2 u = f (t), t > 1.
(a) u(n) = 0;
Ans. (a) g(t, s) =
(c) g(t − s) =
1
6
(t − s)n−1
;
(n − 1)!
et−s −
1
2
(b) g(t.s) =
e− (t − s) +
1
3
t3 − s 3
;
3t2
e−2(t−s) ;
80
3. STURM-LIOUVILLE SYSTEMS
sin(t − s)
.
t2
3.2. Determine Green’s Function and the solution for the following initial value
problems:
1
1
(1) u + u − 2 u = f (x), u(1) = 0, u (1) = 1;
x
x
1
(2) u + 2 u = f (x), u(0) = u (0) = 1/2;
4x
2 (3) x u − 7xu + 15u = f (x), u(1) = 0 = u (1);
(d) Hint. Primitive: {t−2 cos t, t−2 sin t}. Ans. g(t, s) = H(t−1)
(4) x2 u − xu + u = 0,
u(0) = 0, u (0) = 1;
(5) xu − (x + 1)u + u = f (x), u(1) = −1 = u (1);
(6) x2 1 − x2 u − 2xu + 2u = 0, u(1/2) = 0, u (1/2) = 1;
(7) u + u − 2u = ex , u(0) = 1, u (0) = 0;
(8) x2 u + 4xu + 2 + x2 u = f (x), x > 0, u(π/2) = 1 = u (π/2);
(9) 3x2 − 1 u + 3(3x − 1)u − 9u = f (x), x > 1/3, u(1) = 0 = u (1);
2
u = f (x), 0 < x < ∞;
x2
Hint. Primitives are: (1) {x, x−1 }; (2) {x1/2 , x1/2 ln x}; (3) {x3 , x5 };
x
1 + x
(4) {x, x ln x}; (5) {ex , x + 1}; (6) x, ln
; (7) {ex , e−2x };
2
1−x
(8) {cos x/x2 , sin x/x2 }; (9) {3x − 1, (3x − 1)−1 }; (10) {x2 , x−1 }.
2
1
3.3. Solve the initial value problem u − 2 u = , u(1) = 0, u (1) = 1.
x
x
Hint. Primitive: {x2 , x−1 }; W (x) = −3; g(x, s) = 13 x2 s−1 − x−1 s2 .
x
x
Ans. up (x) = − ; u(x) = 23 x2 − 16 x−1 − .
2
2
3.4. Obtain one-sided Green’s functions for the initial value problem ẍ(t)+k 2 x(t) =
0 such that (a) x(0) = x0 , ẋ(0) = 0, t > 0, and (b) x(0) = x0 , ẋ(0) = u0 , t >
0. This problem describes the free motion of a mass m which is attracted by a
force proportional to its distance from the origin and (a) starts from rest at x = x0 ,
and (b) starts with an initial velocity u0 at x = x0 away from the origin.
(10) u −
Hint. The primitive is {cos kt, sin kt}, and in both cases one-sided Green’s
function is g(t, s) = H(t) sin k(t − s).
3.5. Consider the problem of vibration of a spring which hangs vertically with its
upper end fixed, and a mass of 16 lb is attached to its lower end. After coming to
rest the mass is pulled down 2 inches and released. Assuming that the medium
offers a resistance of v/64 lb, where v is expressed in ft/sec, the equation of
3.7. EXERCISES
81
16 2
1
D x(t) +
Dx(t) + 48x(t) = 0, and the initial conditions are:
g
64
x(0) = 0 and Dx(1/16) = 0. Using g = 32, show that Green’s function is
H(t) g(t, s) ≈
v1 (t) v2 (s) − v1 (s) v2 (t) ,
a0 (s)W (s)
motion is
where v1 (t) = 16 e−0.0156t cos 9.8t and v2 (t) = 0.000265e−0.0156t sin 9.8t.
0.000265(9.8) −19.8t
e
.
6
3.6. Show that if u1 (x) and u2 (x) are linearly independent solutions of the homo
geneous equation [p(x) u ] + q(x) u = 0, then the Wronskian W (u1 (x), u2 (x))
is a constant multiple of 1/p(x).
Hint. a0 (t) = 0.5, W (v1 , v2 ) =
Hint. Differentiate p(x) W (u1 (x), u2 (x)) with respect to x and show that
p(x) W (u1 (x), u2 (x)) = const.
3.7. Consider the Sturm-Liouville equation L[u] = u = 0, 0 < x < 1, subject to
the boundary conditions: (a) u(0) = 0 = u(1); (b) u(0) = 0, u (1) = 0;
(c) u(−1) = 0 = u(1); and (d) u(0) = u(1), u (0) = −u (1). Determine
Green’s function in each case.
(1 − s) x, if x ≤ s,
x, if x < s,
Ans. (a) G(x, s) =
; (b) G(x, s) =
(1 − x) s, if x > s.
s, if s < x.
(c) G(x, s) = − 12 |x − s| + xs − 1 ; (d) G(x, s) = − 12 |x − s| + 14 .
3.8. Find the eigenpairs of the following Sturm-Liouville problems:
(a) u + λ u = 0, 0 < x < 1; u(0) = 0, u(1) + u (1) = 0;
(b) u + λ u = 0, 0 < x < 1; u (0) + u (1) = 0;
(c) u + λ u = 0, 0 < x < 1; u(0) = u(1), u (0) + u (1) = 0;
(d) u + 2u + 4λ u = 0, 0 < x < a; u(0) = 0 = u(a).
Ans. (a) For λ > 0: λ = kn2 ≈
1
4
(2n + 1)2 π 2 ; un (x) = sin kn x, n = 1, 2, . . . ;
(b) For λn = n2 π 2 ; un (x) = cos nπx, n = 0, 1, 2, . . . ;
(c) For λn = 4n2 π 2 ; un (x) = {sin 2nπx, cos 2nπx}, n = 0, 1, 2, . . . ;
(d) For 1−4λ = −kn2 < 0: λn = 14 1 + n2 π 2 /a2 ; un (x) = e−x sin (nπx/a), ds
n = 1, 2, . . . .
3.9. Show that Green’s function for 1-D potential equation1 , defined by
d
du
a
= f (x), 0 ≤ x ≤ l,
dx dx
1
This equation is found in problems of transverse deflection of a cable, axial deformation of a bar, heat transfer along a fin in heat exchangers, flow though pipes, laminar
incompressible flow through a channel under constant pressure gradient, linear flow through
porous media, and electrostatics.
82
3. STURM-LIOUVILLE SYSTEMS
in a homogeneous isotropic medium (a = const), is given by
1
G(x, s) =
(l − r), r = |x − s|.
2a
d2 G
Hint. Solve a 2 = δ(x, s). Green’s function is also the fundamental
dx
solution for this potential operator.
3.10. Solve: (x + 1)2 u − u = f (x), 0 < x < 1, subject to the boundary
conditions u(0) = 0 = u (1).
√
√
x
5/2
5/2 1
1+x
1+s
−1/2
−1/2
Ans. u(x) = √
(1+x)
(1+s)
−
·
1+s
1+x
5 0
√
√
1
·f (s) ds +
(1 + x) 5/2 − (1 + x)− 5/2 .
5(1 + x)
3.11. Solve x2 u + xu + u = ln x, x > 1, subject to the conditions u(1) =
0, u (1) = 0. Hint. x±i = cos(ln x) ± i sin(ln x). Find a particular solution
by inspection.
Ans. u(x) = ln x.
√
3.12. Solve 1 + x2 u + u = x, x > 0, subject to the condition u(0) = 0.
Hint. Reduce the equation to the form [p(x) u] = f (x).
−1 1
2
2
Ans. u(x) =
x− x+ 1+x
ln x + 1 + x
.
2
3.13. Solve u − u = ex , 0 < x < 1, such that u(0) = 1, u (1) = 0.
e sinh x cosh(1 − x)
+
.
cosh 1
cosh 1
Hint. The primitive of solutions is {ex , e−x }, and Green’s function is

sinh(x − 1) sinh s


, x < s,
cosh 1
G(x, s) =
.

 sinh x sinh(s − 1) , s < x.
cosh 1
3.14. Show that the initial value problem ẋ = x1/3 , x(0) = 0 has infinitely many
solutions passing through the origin of the (x, t)-plane,
$
0, 0 ≤ t ≤ ν,
Ans. the solutions are given by xν (t) = 2
3/2
, ν < t ≤ 1,
3 (t − ν)
Ans. u(x) =
1
2
x ex −
where ν ∈ [0, 1] is a real number.
d
du 3.15. The homogeneous Sturm-Liouville equation
p(x)
= 0 is satisfied by
x
dx
du
1
=
. Use this to obtain a second solution for Legendre’s equation.
dx
p(x)
3.7. EXERCISES
Ans. u2 (x) =
83
1 1+x
ln
.
2 1−x
d2 u
3.16. Determine Green’s function for the Sturm-Liouville system 2 = 0, u(0) =
dx
x, x < s,
0 = u (1), 0 < x < 1. Ans. G(x, s) =
.
s, s < x
3.17. Find Green’s function for the system
d2 u(x)
e−x
2
−
k
u(x)
+
λ
u(x) = 0, u(0) = 0 = u(∞), 0 < x < ∞.
dx2
x

1

 e−ks sinh kx, x < s,
k
Ans. G(x, s) =

 1 e−ks sinh kx, s < x.
k
3.18. Obtain the expansion for the function f (x) = 1, 1 ≤ x ≤ eπ , in a series of
λ
eigenfunctions {φn } of the Sturm-Liouville system [xu ] + u = 0, u(1) =
x
0 = u (eπ ).
Ans. 1 =
2
π
∞
1 − (−1)n
4
sin(n ln |x|) =
n
π
n=1
∞
sin [(2n − 1) ln |x|]
.
2n − 1
n=1
4
Bernoulli’s Separation Method
In the study of linear partial differential equations and determination of Green’s
functions there are two methods that are mostly used. They are: Bernoulli’s separation
method, commonly known as the method of separation of variables, and the integral
transform method. In this chapter we will present Bernoulli’s separation method by
means of some useful examples, which will explain the working of the method in
detail.
4.1. Coordinate Systems
Let a point P be located by the polar cylindrical coordinates (r, θ, z) as well as
rectangular cartesian coordinates (x, y, z) (see Fig. 4.1).
z
z
P
• (r ,θ, z )
•
z
0
θ
P (r , θ , φ )
θ r
0
y
φ
r
x
x
Fig. 4.1.
Fig. 4.2.
The transformation between these two coordinate systems is defined by
x = r cos θ
r = x2 + y 2
y = r sin θ
or
θ = arctan(y/x)
z=z
z = z.
84
y
4.2. PARTIAL DIFFERENTIAL EQUATIONS
85
Similarly, the point P (x, y, z) is defined in the spherical coordinate system (r, θ, φ)
by the transformations (see Fig. 4.2)
x = r sin θ cos φ
y = r sin θ sin φ
z = r cos θ
or
r = x2 + y 2 + z 2
θ = arccos(z/r)
φ = arctan(y/x).
We will study linear partial differential equations of the form
∇2 u + k 2 u = 0,
(4.1)
where the 3-D Laplacian ∇2 is expressed in different coordinate systems as follows:
In rectangular cartesian :
∂2
∂2
∂2
+
+
;
∂x2
∂y 2
∂z 2
∂2
1 ∂
1 ∂2
∂2
+
+ 2 2 + 2;
2
∂r
r ∂r r ∂θ
∂z
1 ∂
∂
1
∂
∂
1 ∂2
In spherical : 2
r2
+
sin θ
+
.
r ∂r
∂r
sin θ ∂θ
∂θ
sin2 θ ∂φ2
We will analyze the cases in which Eq (4.1) is separable. Thus,
In polar cylindrical :
if k 2 = 0, then Eq (4.1) is Laplace’s equation;
if k 2 = a positive constant, then Eq (4.1) is Helmholtz’s equation;
if k 2 = a negative constant, then Eq (4.1) is diffusion equation;
if k 2 = constant × kinetic energy, then Eq (4.1) is Schrödinger’s wave equation.
4.2. Partial Differential Equations
We will discuss the classification of second-order nonhomogeneous partial differential
equations of the form
a11 uxx + 2a12 uxy + a22 uyy + b1 ux + b2 uy + c0 u = f (x, y).
(4.2)
If f = 0 in Eq (4.2), the most general form of a second-order homogeneous equation
is
a11 uxx + 2a12 uxy + a22 uyy + b1 ux + b2 uy + c0 u = 0.
(4.3)
To show a correspondence of this equation with an algebraic quadratic equation, we
replace ux by α, uy by β, uxx by α2 , uxy by αβ, and uyy by β 2 . Then the left side
of Eq (4.3) reduces to a second degree polynomial in α and β:
P (α, β) = a11 α2 + 2a12 αβ + a22 β 2 + b1 α + b2 β + c0 .
86
4. BERNOULLI’S SEPARATION METHOD
It is known from analytical geometry and algebra that the polynomial equation
P (α, β) = 0 represents a hyperbola, parabola or ellipse according as its discriminant
a212 − a11 a22 is positive, zero or negative. Thus, Eq (4.3) is classified as hyperbolic,
parabolic or elliptic according as the quantity a212 −a11 a22 is positive, zero or negative.
An alternative approach to classify the types of Eq (4.3) is based on the following
theorem:
Theorem 4.1. The relation φ(x, y) = C is a general integral of the ordinary differential equation
a11 dy 2 − 2a12 dx dy + a22 dx2 = 0
(4.4)
iff u = φ(x, y) is a particular solution of the equation
a11 u2x + 2a12 ux uy + a22 u2y = 0.
(4.5)
Eq (4.4), or (4.5), is called the characteristic equation of the partial differential
equation (4.2) or (4.3); the related integrals are called the characteristics. Eq (4.4),
regarded as a quadratic equation in dy/dx, yields two solutions:
dy
a12 ±
=
dx
a212 − a11 a22
.
a11
The expression under the radical sign determines the type of the differential equation
(4.2) or (4.3). Thus, as before, Eq (4.2) or (4.3) is of the hyperbolic, parabolic or
elliptic type according as the quantity a212 − a11 a22 0.
Most frequently encountered partial differential equations of mathematical physics
are:
1. Laplalce’s equation: ∇2 u = 0. This equation occurs in the study of (i) electromagnetics including electrostatics, dielectrics, steady currents, and magnetostatics; (ii)
hydrodynamics (irrotational flow of perfect fluids and surface waves); (iii) heat
flow; and (iv) gravitation.
2. Poisson’s equation: ∇2 u = −ρ/ε0 .
3. Helmholtz’s (wave) and time-dependent diffusion equations: ∇2 u ± k 2 u = 0.
These equations occur in (i) elastic waves in solids including vibrating strings,
bars, membranes; (ii) acoustics; (iii) electromagnetic waves; and (iv) nuclear
reactors.
1 ∂u
4. The time-dependent diffusion equation: ∇2 u = 2
, and the corresponding
a ∂t
∂2
∂2
4-D analog of the Laplacian in Minkowski space: 2 = ∇2 +
=
+
∂x24
∂x2
∂2
∂2
∂2
+ 2+
.
2
∂y
∂z
(ic)2 ∂t2
4.2. PARTIAL DIFFERENTIAL EQUATIONS
5. The time-dependent wave equation: 2 u = 0, where =
87
∂2
− c2 ∇2 .
∂t2
6. The scalar potential equation: 2 u = −ρ/ε0 .
7. The Klein-Gordon equation: 2 u = µ2 u, and the associated vector equation in
which the scalar function u is replaced by a vector function.
8. Schrödinger’s wave equation: −
2 2
∂u
∇ u+V u = i , and the time-independent
2m
∂t
2 2
∇ u + V u = Eu.
2m
9. Equations for elastic waves and for viscous fluids, and telegrapher’s equation.
equation: −
10. Maxwell’s coupled equations for electric and magnetic fields, and Dirac’s equations for relativistic electron wave equations.
All these equationscan be written in the form
L[u] = F , where L is the linear
∂ ∂ ∂ ∂
differential operator L
,
,
, , x, y, z , F is a known function, and u is the
∂x ∂y ∂z ∂t
unknown scalar (or vector) function. Also, all these equations are second-order partial
differential equations except Maxwell’s and Dirac’s equations which are first-order
but involve two unknown functions. Sometimes higher order equations are found,
as in the slow motion of a viscous fluid and the theory
body which
4of an elastic
are
∂
∂4
∂4
2 2
governed by the fourth-order equation (∇ ) u =
+2 2 2 + 4 u =
∂x4
∂x ∂y
∂y
0. The techniques mostly used to solve these equations are: (i) Bernoulli’s separation
method which separates the equations into ordinary differential equations; (ii) integral
transform methods; (iii) integral solutions using Green’s functions; and (iv) numerical
computation. We will henceforth study different methods based on these techniques.
A important aspect of Green’s functions in the case of partial differential equation is similar to the case of ordinary differential equations where Green’s function
method is used to determine the solution of a given linear partial differential equation
L[u](x) = f (x), as proved in the following theorem.
Theorem 4.2. If f (x) is a continuous or piecewise continuous function
of x, x0 ≤ x ≤ x1 , then the function
x1
u(x) =
G(x, x )f (x ) dx
(4.6)
x0
is a solution of the equation L[u] = −f (x), which satisfies the prescribed
boundary conditions.
x1
Proof. Let u(x) =
G(x, x )f (x ) dx . Then
u (x) =
x0
x1
x0
G (x, x )f (x ) dx ,
88
4. BERNOULLI’S SEPARATION METHOD
x1
u (x) =
G (x, x )f (x ) dx
x0
+G (x, x − 0)f (x) − G (x, x + 0)f (x)
x1
=
x0
x1
=
G (x, x )f (x ) dx + [G (x, x + 0) − G (x, x − 0)] f (x)
G (x, x )f (x ) dx −
x0
x1
G (x, x )f (x ) dx +
x0
Thus, pu + p u − qu =
x1
f (x)
.
p(x)
(pg + p G − qG) f (x ) dx − f (x). Since Green’s
x0
function G(x, x ) satisfies the equation L[G(x, x )] = δ(x
x ), we find by us x−
1
ing shifting property (2.10) of the δ-function that L[u] = x0 f (s)L[G(x, s)] ds =
x1
f (x)δ(x − s) ds = f (x).
x0
To prove the converse in 2-D case, let D denote a domain with a rectifiable
boundary ∂D, and let L[u] be any of the partial differential operators listed in §4.2.
2
In particular, we consider the boundary
value problem: L[u] ≡ ∇ u = f (x, y)
subject to the boundary conditions u
= 0. If B represents the region
(x,y)∈∂D
(x − x )2 + (y − y )2 ≤ ε2 , then lim
∇2 G(x, y; x , y ) dx dy = −1. We use
ε→0
B
Green’s second identity (1.16) with f = 1 and g = G(x, x ), and obtain
∇2 G(x, y; x , y ) dx dy =
lim
ε→0
B
∂B
∂G
ds,
∂n
where n is the directionof the outward normal to the boundary ∂B. Obviously, n
can be replaced by r = (x − x )2 + (y − y )2 , and we have
lim
ε→0
Cε
∂G(x, y; x , y )
ds = −1,
∂r
where Cε is the circle with radius ε and center at (x , y ). In terms√of the polar
cylindrical coordinates: x = x + r cos θ, y = y + r sin θ, with ds = r 2 + r 2 dθ,
2π
∂G
ε dθ
we find that
ds =
= 2π. Thus, if we choose for G(x, y; x , y ) a
ε
Cε ∂r
0
1
function such that its derivative has a singularity of the type −
, then we can take
2πr
1
G(x, y; x , y ) =
log r + g(x, y; x , y ), where g(x, y; x , y ) ∈ C 2 (D). Hence,
2π
∂ 1
lim
log r + g(x, y; x , y ) ds = 1.
ε→0 C ∂r 2π
ε
Now, if we substitute u(x) as defined by (4.6) in the equation L[u] ≡ ∇2 u = f (x),
4.3. BERNOULLI’S SEPARATION METHOD
89
we get
G(x, y; x , y )f (x , y ) dx dy =
∇2 u = ∇
∇G(x, y; x , y )f (x , y ) dx dy 1
=
∇2 (log r) f (x , y ) dx dy , since ∇2 g = 0 everywhere
2π
D
1
ds
=
lim f (x̄, ȳ)
= f (x, y).
2π ε→0
Cε r
D
D
The 3-D case can be similarly completed.
4.3. Bernoulli’s Separation Method
We will present this method by using certain examples which involve simple partial
differential equations of mathematical physics.
4.3.1. Laplace’s Equation in a Cube. Consider Laplace’s equation in the
cube D = {0 < x, y, z < π}:
∇2 u ≡
∂2u ∂ 2u ∂ 2u
+ 2 + 2 = 0,
∂x2
∂y
∂z
in D,
(4.7)
such that u(0, 0, 0) = 0 = u(π, π, π), and u(x, y, 0) = g(x, y). This problem arises
in electrostatics where u is the potential which has the prescribed value g(x, y) on
the face z = 0 of the cube D, while the other faces are kept at zero potential. The
solution also defines an equilibrium temperature distribution when the faces of the
cube are kept at temperature zero and g(x, y) as defined above. Since the maximum
principle1 holds for Laplace’s equation in 3-D as well as 2-D, it implies that a 3D boundary value problem for Laplace’s equation has at most one solution, which
vanishes continuously with the prescribed boundary values. We will find this solution
by Bernoulli’s separation method: Let u = X(x) Y (y) Z(z), and substituting it in
X Y Z Eq (4.7) we get
+
=−
. Since the left-hand side is independent of z and
X
Y
Z
the right-hand side depends only on z, both sides must be equal to a constant, i.e.,
X Y Z +
=−
= k1 ,
X
Y
Z
X Y = k1 −
= k2 .
X
Y
Thus, we have X − k2 X = 0, Y − (k1 − k2 ) Y = 0, Z + k1 Z = 0, with the
respective boundary conditions: X(0) = X(π) = 0, Y (0) = Y (π) = 0, Z(π) = 0.
The boundary value problem for X is an eigenvalue problem, for which we must have
1
This principle states that a nonconstant function which is harmonic inside a bounded
domain D with boundary ∂D and continuous on the closed region D = D ∪ ∂D attains
its maximum and minimum values only on the boundary ∂D. Note that u is harmonic iff
∇2 u = 0 (see Theorem 8.1).
90
4. BERNOULLI’S SEPARATION METHOD
k2 = −n2 , where n ∈ Z+ , and the corresponding eigenfunction is X = sin nx. The
boundary value problem for Y is also an eigenvalue problem, for which k1 − k2 =
−m2√
, m ∈ Z+ , and thus, Y = sin my. But k1 = −m2 − n2 , so Z is a multiple of
sinh m2 + n2 (π − z). Hence, we have a solution of the form
∞
∞
u(x, y, z) =
cnm sin nx sin my sinh
m2 + n2 (π − z).
n=1 m=1
The boundary condition at z = 0 gives
∞
∞
g(x, y) =
cnm sin nx sin my sinh
m2 + n2 π.
n=1 m=1
If we set
cnm sinh
m2 + n2 π = dnm =
4
π2
π
π
g(x, y) sin nx sin my dx dy,
0
0
then the solution is given by
∞
∞
g(x, y) =
dnm
n=1 m=1
√
sinh m2 + n2 (π − z)
√
sin nx sin my
,
sinh m2 + n2 π
(4.8)
where the Fourier series of g(x, y) converges uniformly for z ≥ 0.
4.3.2. Laplace’s Equation in a Cylinder. Consider the problem
∂ 2 u 1 ∂u 1 2 ∂ 2 u ∂ 2 u
+
+
+ 2 =0
∂r2
r ∂r
r ∂θ2
∂z
(4.9)
in the cylinder 0 < r < a, 0 < θ < 2π, 0 < z < π, such that u(r, θ, 0) =
u(r, θ, π) = 0, u(a, θ, z) = g(θ, z). Using Bernoulli’s separation method, we take
u = R(r)Θ(θ)Z(z), and substituting it in Eq (4.9) we get
1 R
1 Θ
Z r
+ 2
=−
= k1 ,
R
r Θ
Z
R +
or
r2 R + rR
Θ
− r 2 k1 = −
= k2 .
R
Θ
Thus, solving Z + k1 Z = 0, Z(0) = Z(π) = 0, we get k1 = n2 , n ∈ Z+ ,
and Zn = sin nz. Again, Θ + k2 Θ = 0, where Θ is periodic with period 2π,
i.e., Θ(0) = Θ(2π). Then k2 = m2 , m = 0, 1, 2, . . . , with the corresponding
4.3. BERNOULLI’S SEPARATION METHOD
91
eigenfunctions Θ0 = 1, Θm = {sin mθ, cos mθ}, m = 1, 2, . . . . Finally, for R we
have the differential equation
2
1 m
2
Rmn + Rmn
−
+
n
Rmn = 0,
r
r2
which has the singularity at r = 0. The solution at r = a is
Rmn (r) =
Im (nr)
,
Im (na)
where Im is the Bessel function with imaginary argument, which converges for all r.
Hence, the solution is given by
1
u(r, θ, z) =
2
∞
cn0
n=1
∞
+
I0 (nr)
sin nz
I0 (na)
∞
Im (nr)
(cnm sin mθ + dnm cos mθ) sin nz,
I (na)
n=1 m=1 m
where
cnm =
dnm =
2
π2
2
π2
π
(4.10)
2π
g(θ, z) cos mθ sin nz dθ dz,
0
0
π
2π
g(θ, z) sin mθ sin nz dθ dz,
0
0
and the series in (4.10) converges uniformly for r < a, and 0 < z < π. Also see
Example 4.10 below.
4.3.3. Laplace’s Equation in a Sphere. Consider the boundary value problem
in a sphere of radius a:
∂ 2 u 2 ∂u
1
∂
∂u
1
∂2u
+
+
sin
θ
+ 2 2
= 0, r < a,
2
2
∂r
r ∂r
r sin θ ∂θ
∂θ
r sin θ ∂φ2
(4.11)
u(a, θ, φ) = f (θ, φ).
Using Bernoulli’s separation method, this Laplace’s equation has a solution of the
form u = R(r)Θ(θ)Φ(φ) if
R + 2 R /r (sin θ Θ )
Φ
+
+
= 0.
R
Θr 2 sin θ
Φ r 2 sin2 θ
Multiplying this equation by r 2 sin2 θ and transposing the last term, we see that
Φ /Φ must be a constant. Since Φ is a periodic function of period of 2π, we have
Φ = cos mφ or Φ = sin mφ, m = 0, 1, 2, . . . . Thus,
R + 2 R /r (sin θ Θ )
m2
+
−
= 0.
R
Θ r 2 sin θ
r 2 sin2 θ
92
4. BERNOULLI’S SEPARATION METHOD
Multiplying by r2 and transposing the first term, we obtain two equations:
2
r 2 R + R − λ R = 0,
r
m2
(sin θ Θ ) =
Θ + λ sin θ Θ = 0,
sin θ
(4.12)
(4.13)
where λ is a constant. The equation for Θ is singular at both end points θ = 0, π.
If we impose the condition that Θ and Θ remain bounded at these end points, this
becomes an eigenvalue problem in Θ with two singular points. If we use the new
variable t = cos θ and let Θ(θ) = P (cos θ), then Eq (4.13) becomes
d
dt
dP
1−t
dt
2
−
m2
P + λ P = 0,
1 − t2
−1 < t < 1.
(4.14)
where λ = n(n + 1) yields the associated Legendre’s equation
m2
1 − t2 Θ − 2t Θ + n(n + 1) −
Θ = 0,
1 − t2
(4.15)
which has the solutions Θ(t) = Pnm (t), or Θ(θ) = Pnm (cos θ), where Pnm (cos θ) are
the associated Legendre polynomials of order m = 0, ±1, ±2, . . . , ±n, and degree
n = 0, 1, 2, . . . . Also see Example 4.12 below.
4.3.4. Helmholtz’s Equation in Cartesian Coordinates. We rewrite Eq
(4.1) as
∂ 2u ∂ 2u ∂2u
+ 2 + 2 + k 2 u = 0,
∂x2
∂y
∂z
(4.16)
and solve it by using Bernoulli’s separation method: Set u(x, y, z) = X(x)Y (y)Z(z)
and substitute in Eq (4.16). Then we obtain
YZ
d2 X
dY
d2 Z
+ XZ 2 + XY 2 + k 2 XY Z = 0,
2
dx
dy
dz
or dividing by XY Z and rearranging terms we have
1 d2 X
1 dY
1 d2 Z
= −k 2 −
−
.
2
2
X dx
Y dy
Z dz 2
At this stage we have only one separation of variables: The left side is a function of
x only, while the right side depends on y and z. This means that we are equating
a function of x to a function of y and z, although x, y and z are all independent
variables. Thus, the behavior of x as an independent variable cannot be determined
4.3. BERNOULLI’S SEPARATION METHOD
93
by y and z. To resolve this situation we introduce another arbitrary constant l2 and
choose
1 d2 X
= −l2 ,
X dx2
1 d2 Y
1 d2 Z
+
= l2 − k 2 ,
2
Y dy
Z dz 2
(4.17)
(4.18)
where the sign of l2 will be fixed in specific problems subject to prescribed boundary
conditions. Now, we rewrite Eq (4.18) as
1 dY
1 d2 Z
= l2 − k 2 −
,
2
Y dy
Z dz 2
where we arrive at a second separation of variables by equating a function of y to a
function of z, which is accomplished by equating each side of the above equation to
another constant of separation −m2 , i.e.,
1
Y
1
Z
d2 Y
= −m2 ,
dy 2
d2 Z
= l2 − k 2 + m2 = −n2 .
dz 2
(4.19)
(4.20)
As a result of this separation method we obtain a set of three ordinary differential
equations (4.17), (4.19) and (4.20), which can be solved by known methods. Hence,
the solution of Eq (4.16) is given by
ulmn (x, y, z) = Xl (x) Ym (y) Zn (z),
(4.21)
subject to prescribed boundary conditions; the constants l, m, n are chosen such that
the condition k 2 = l2 + m2 + n2 holds. Then, the general solution of Eq (4.16) is
given by
u(x, y, z) =
clmn ulmn (x, y, z),
(4.22)
l,m,n
where the coefficients clmn are chosen so that u satisfies prescribed boundary conditions of the problem.
4.3.5. Helmholtz’s Equation in Spherical Coordinates. This equation
with positive k 2 in spherical coordinates is
1 ∂
1 ∂
∂
1 ∂2
2 ∂
r
+
sin
θ
+
+ k 2 u = 0.
(4.23)
2
r ∂r
∂r
sin θ ∂θ
∂θ
sin2 θ ∂φ2
As before, we set u(r, θ, φ) = R(r)Θ(θ)Φ(φ), substitute it in Eq (4.23), and divide
throughout by RΘΦ, to obtain
1 d
1
d
dΘ
1
d2 Φ
2 dR
r
+
sin
θ
+
= −k 2 .
2
2
2
2
Rr dr
dr
Θr sin θ dθ
dθ
Φr sin θ dφ2
94
4. BERNOULLI’S SEPARATION METHOD
Multiplying this equation by r 2 sin2 θ, we separate the Φ part and get
1 d2 Φ
1 d
1
d
dΘ
2
2
2
2 dR
=
−r
sin
θ
k
+
r
+
sin
θ
.
Φ dφ2
r 2 R dr
dr
dθ
r2 sin2 θΘ dθ
(4.24)
This equation equates a function of φ to a function of r and θ. But since r, θ, and φ
are independent variables, each side of Eq (4.24) must be equal to a constant. Again,
since in all physical problems φ denotes the azimuth angle, we assume a periodic
solution, and take the separation constant as −m2 . Then after separation we obtain
1 d2 Φ
= −m2 ,
(4.25)
Φ dφ2
1 d
1
d
dΘ
m2
2 dR
r
+
sin
θ
− 2 2 = −k 2 .
2
2
2
r R dr
dr
dθ
r sin θ Θ dθ
r sin θ
(4.26)
Multiplying Eq (4.26) by r2 and rearranging the terms we get
1 d
dR
1
d
dΘ
m2
r2
+ k2 r2 = −
sin θ
+
.
R dr
dr
sin θ Θ dθ
dθ
sin2 θ
(4.27)
Next, separating the remaining two variables and equating each side of Eq (4.27) to
a constant l, we obtain
1 d
l
2 dR
2
r
+ k − 2 R = 0,
(4.28)
r2 dr
dr
r
1 d
dΘ
m2
sin θ
+ l−
Θ = 0.
(4.29)
sin θ dθ
dθ
sin2 θ
The three equations (4.25), (4.28) and (4.29) are ordinary differential equations, which
can be solved by known methods. Note that Eq (4.29) is Legendre’s equation with
the constant l = n(n + 1). Thus, the general solution of Eq (4.23) is given by
u(r, θ, φ) =
Rl (r)Θlm (θ)Φm (φ).
(4.30)
l,m
The constant k 2 is a function of r in Schrödinger’s wave equation, and a function of
r, θ and φ in the problem of the hydrogen atom (see §7.8).
4.3.6. Wave Equation. Consider the wave equation for the space-dependent
amplitude of an electromagnetic wave in a cylindrical wave guide: ∇2 u + k 2 u = 0.
Taking the central axis of the cylindrical wave guide as the z-axis we find that in
a wave guide with perfectly conducting walls there is no attenuation and hence, no
dependence on z. Then the equation to be solved by Bernoulli’s separation method
for this problem is
∂ 2 u 1 ∂u
1 ∂2u
+
+
+ k 2 u = 0.
∂r2
r ∂r
r2 ∂θ2
(4.31)
4.4. EXAMPLES
95
Set u(r, θ) = R(r)Θ(θ) in this equation, which gives
1 d
dR
1 d2 Θ
r
+
+ k 2 = 0.
Rr dr
dr
Θr 2 dθ
Multiplying by r2 we separate the θ variable and get
1 d2 Θ
= −m2 ,
Θ dθ
(4.32)
where θ is the azimuth angle in this case; thus, the solution of E (4.32) will be periodic
of the form sin mθ, cos mθ. The equation in r is given by
1 d
dR
m2
r
+ k 2 − 2 R = 0,
(4.33)
r dr
dr
r
which is Bessel’s equation, for which see Example 3.24.
4.4. Examples
Some useful examples follow.
Example 4.1. (Hyperbolic equation.) The problem of a vibrating string is
governed by the one-dimensional wave equation. Now, consider the boundary value
problem
2
∂ 2u
2∂ u
=
c
, 0 < x < l,
∂t2
∂x2
u(0, t) = 0 = u(l, t), t > 0,
u(x, 0) = f (x),
(4.34)
(4.35)
ut (x, 0) = h(x),
0 < x < l,
(4.36)
where f ∈ C 1 (0, l) is a given function. We seek a solution of the form u(x, t) =
X(x)T (t), where X is a function of x only and T a function of t only, and assume
that a solution of this form exists. We will carry out the details to see if the method
∂2u
∂2u
works for this problem. Note that 2 = X T , and 2 = X T , where the primes
∂t
∂x
denote the derivative with respect to its corresponding independent variable. Then
Eq (4.34) reduces to X T = c2 X T , or, after separating the variables, it becomes
T X = c2
.
T
X
(4.37)
Since the left side of Eq (4.37) is a function of t only and the right side a function of
x only, the only situation where X(x) and T (t) have solutions for all x and all t is
c2 X (x)
T (t)
when
=
= const. Hence, from (4.37) we can write
X
T
1 T X =
= k,
c2 T
X
k = const.
(4.38)
96
4. BERNOULLI’S SEPARATION METHOD
The set of equations (4.38) is equivalent to two ordinary differential equations:
T − k c2 T = 0,
(4.39)
X − k X = 0.
(4.40)
It is apparent from Eq (4.38) that the constant k in Eqs (4.39) and (4.40) has the same
value. The general solution of Eq (4.39) is
√
√

c kt

+ c2 e−c kt for k > 0
 c1 e
T (t) =
(4.41)
c1 t + c2 for k = 0

√
√

c1 cos c −kt + c2 sin c −kt for k < 0,
and of Eq (4.40) is
√
√

kx

+ d2 e− kx for k > 0
 d1 e
X(x) =
d1 x + d2 for k = 0

√
√

d1 cos −kx + d2 sin −kx for k < 0.
(4.42)
In view of the boundary conditions (4.35) we must have
X(0)T (t) = 0 = X(l)T (t), for all t ≥ 0.
(4.43)
Using these conditions in (4.42) for k > 0 we get the system of equations
X(0) = d1 + d2 = 0,
√
kl
X(l) = d1 e
√
+ d2 e −
kl
= 0.
(4.44)
The system (4.44) is has a nontrivial solution iff the determinant of its coefficients
vanishes. But since
√
√
1√
kl 1−√kl = e− kl − e kl = 0,
e
e
a nonzero solution for X(x) in (4.42) for k > 0 is not possible. Next, for k = 0,
the boundary conditions (4.43) imply that d1 = 0 and d2 = 0. Hence, there is no
nonzero solution for k = 0. Finally, for k < 0, let us set k = −λ2 . Then the general
solution (4.42) in this case becomes X(x) = d1 cos λx + d2 sin λx, which, under the
boundary conditions (4.43), yields X(0) = d1 = 0 and X(l) = d2 sin λl = 0. To
avoid a trivial solution in this case, we choose λ such that λl is a positive multiple of
nπ
π, i.e., λl = nπ, or λ =
. The positive values of λ are chosen because the negative
l
multiples give the same eigenfunctions as the positive ones. This result leads to an
nπx
infinite set of solutions which are denoted by Xn (x) = d2,n sin
, where each
l
nπx
eigenfunction sin
corresponds to the eigenvalue
l
k=−
n2 π 2
.
l2
(4.45)
4.4. EXAMPLES
97
The solutions for T (t) for the choice of k < 0, as in (4.45), are then obtained from
(4.41) as
nπct
nπct
Tn (t) = c1,n cos
+ c2,n sin
.
l
l
Then the infinite set of solutions is
un (x, t) = Xn (x)Tn (t) = An cos
nπct
nπct
nπx
+ Bn sin
sin
,
l
l
l
(4.46)
where the constants An and Bn are determined from the initial conditions. The
eigenfunctions are contained in the solution (4.46), whereas the eigenvalues for this
boundary value problem are given by (4.45).
The next step is to obtain the particular solution which satisfies the initial conditions (4.36). At this point it may so happen that none of the solutions (4.46) will
satisfy (4.36). In view of the superposition principle, any finite sum of the solutions
(4.46) is also a solution of this boundary value problem. We should, therefore, find a
linear combination of those solutions, which also satisfies the initial conditions (4.36).
Even if this technique fails, we can always try an infinite series of solutions (4.46);
that is,
∞
u(x, t) =
∞
Xn (x)Tn (t) =
n=1
n=1
nπct
nπct
nπx
An cos
+ Bn sin
sin
. (4.47)
l
l
l
Now, we can take this series expansion formally and verify that the boundary conditions (4.35) are still satisfied. We will use the initial conditions to find the constants
An and Bn . Using the first of the initial conditions (4.36) we get
∞
u(x, 0) =
An sin
n=1
nπx
= f (x).
l
(4.48)
This infinite series is a Fourier sine series. Hence, f (x) can be regarded as an odd
function with period 2l. Thus, we expand this function f (x) on the interval 0 ≤ x ≤ l
such that f (−x) = −f (x) on the interval −l ≤ x ≤ 0, and f (x + 2l) = f (x) for all
x. Then the coefficients An for n = 1, 2, . . . are given by
2
An =
l
l
f (x) sin
0
nπx
dx.
l
Taking the time-derivative of (4.47), we get
ut =
∞
πc
nπct
nπct
nπx
n Bn cos
− An sin
sin
,
l n=1
l
l
l
98
4. BERNOULLI’S SEPARATION METHOD
which, in view of the second of the initial conditions (4.36), gives
∞
ut (x, 0) =
where
Bn =
2
nπc
πc
nπx
nBn sin
= h(x),
l n=1
l
l
h(x) sin
0
nπx
dx,
l
n = 1, 2, . . . .
Hence, the solution (4.47) is completely determined.
d’Alembert form of the solution. We will now derive the d’Alembert
solution for this problem. From (4.47) we have
∞
1
nπ(x + ct)
nπ(x − ct)
u=
An sin
+ sin
2 n=1
l
l
∞
1
nπ(x − ct)
nπ(x + ct)
(4.49)
+
Bn cos
− cos
2 n=1
l
l
1
1
=
f (x + ct) + f (x − ct) +
− g(x + ct) + g(x − ct) ,
2
2
∞
∞
nπz
nπz
where f (z) = n=1 An sin
, as in (4.48). Let g(z) = n=1 Bn cos
.
l
l
1
πc
nπz
1
∞
Then g (z) = −
= − h(z), and from (4.49) we obtain
n=1 nBn sin
c l
l
c
the formal solution, known as the d’Alembert solution for this problem, as
u(x, t) = φ(x − ct) + ψ(x + ct),
(4.50)
where c is the wave velocity, and
1
f (x − ct) + g(x − ct)
2
1
ψ(x + ct) = f (x + ct) − g(x + ct) .
2
φ(x − ct) =
An interpretation of the solution of this problem is as follows: At each point x of the
∞
nπc
nπx
string we have u(x, t) =
αn cos
(t + δn ) sin
. This equation describes
l
l
n=1
nπx
a harmonic motion with amplitudes αn sin
. Each such motion of the string
l
is called a standing wave, which has its nodes at the points where sin(nπx/l) = 0;
these points remain fixed during the entire process of vibration. But the string vibrates
with maximum amplitudes αn at the points where sin(nπx/l) = ±1. For any t the
∞
nπx
structure of the standing wave is described by u(x, t) =
Cn (t) sin
, where
l
n=1
4.4. EXAMPLES
99
nπc
Cn (t) = αn cos ωn (t + δn ), ωn =
. At times t when cos ωn (t + δn ) = ±1, the
l
velocity becomes zero and the displacement reaches its maximum value.
Example 4.2. (Parabolic Equation) Consider the one-dimensional heat conduction equation
∂u
∂2u
= k 2 , 0 < x < l,
∂t
∂x
subject to the boundary and initial conditions
t > 0,
(4.52a)
0 < x < l,
(4.52b)
u(0, t) = 0 = u(l, t),
u(x, 0) = f (x),
(4.51)
where f ∈ C 1 is a prescribed function. In physical terms, this problem represents
the heat conduction in a rod when its ends are maintained at zero temperature while
the initial temperature u at any point of the rod is prescribed as f (x). We assume
the solution in the form u(x, t) = X(x)T (t), which after substitution into Eq (4.51)
yields the set of equations
1 T
X =
.
(4.53)
k T
X
As in Example 4.1, the only situation where these two expressions can be equal is for
each of them to be constant, say each equal to m. Then Eq (4.53) yields two ordinary
differential equations
T − mk T = 0,
(4.54)
X − m X = 0,
(4.55)
where the boundary conditions (4.52a) reduce to
or X(0) = 0 = X(l),
X(0) T (t) = 0 = X(l) T (t),
(4.56)
except for the case when the rod has zero initial temperature at every point. This
situation, being uninteresting, can be neglected. We notice that for a nonzero solution
of the problem (4.55)-(4.56) we must choose negative values of m. Hence we set
m = −λ2 , and find that the eigenvalues m = −n2 π 2 /l2 have the corresponding
nπx
kn2 π 2
eigenfunctions Xn (x) = An sin
. Eq (4.54) then becomes T +
T =0
l
l2
2 2
2
whose general solution for each n is given by Tn (t) = Bn e−kn π t/l . Hence, we
consider an infinite series of the form
∞
u(x, t) =
∞
Xn (x)Tn (t) =
n=1
Cn sin
n=1
nπx −kn2 π2 t/l2
e
.
l
(4.57)
Now, we use the initial condition (4.52b) in (4.57) and obtain
∞
u(x, 0) =
Cn sin
n=1
nπx
= f (x),
l
(4.58)
100
4. BERNOULLI’S SEPARATION METHOD
which shows that f (x) can be represented as a Fourier sine series, by extending f
as an odd, piecewise continuous function of period 2l with piecewise continuous
derivatives. Equation (4.58) gives the coefficients Cn as
Cn =
2
l
l
f (x) sin
0
nπx
dx, n = 1, 2, . . . .
l
(4.59)
Hence, the solution (4.57) is completely determined for this problem. Note that the
series in (4.58) converges since u(x, 0) does, and the exponential expression in (4.57)
is less than 1 for each n and all t > 0 and approaches zero as t → ∞.
An interesting situation, considered in the next example, arises if the function
f (x) is zero in the initial condition (4.52b), but the boundary conditions are nonhomogeneous.
Example 4.3. Consider the dimensionless partial differential equation governing the plane wall transient heat conduction
ut = uxx ,
0 < x < 1,
(4.60)
with the boundary and initial conditions
u(0, t) = 1,
u(1, t) = 0,
u(x, 0) = 0,
0 < x < 1.
t > 0,
(4.61)
(4.62)
Since the nonhomogeneous boundary condition in (4.61) does not allow us to compute
the eigenfunctions, we proceed as follows: First, we find a particular solution of the
problem, which satisfies only the boundary conditions. Although there is more than
one way to determine the particular solution, we can, for example, take the steadystate case, where the equation becomes uxx = 0, which is independent of t, and
after integrating twice, has the general solution u(x) = c1 x + c2 , with the boundary
conditions u(0) = 1, u(1) = 0. Thus, c1 = −1, c2 = 1, and the steady-state
solution is u(x) = 1 − x. Next, we formulate a homogeneous problem by writing
u(x, t) as a sum of the steady-state solution u(x) and a transient term v(x, t), i.e.,
u(x, t) = u(x) + v(x, t), or
v(x, t) = u(x, t) − u(x).
(4.63)
Thus, the problem reduces to finding v(x, t). If we substitute v from (4.63) into
(4.60), we get
vt = vxx ,
(4.64)
where the boundary conditions (4.61) and the initial condition (4.62) reduce to
v(0, t) = u(0, t) − u(0) = 0,
v(1, t) = u(1, t) − u(1) = 0,
(4.65)
4.4. EXAMPLES
101
and
v(x, 0) = u(x, 0) − u(x) = x − 1.
(4.66)
Notice that the problem (4.64)-(4.66) is the same as in Example 4.2 with k = 1,
l = 1, f (x) = x − 1, and u replaced by v. Hence, its general solution from (4.57) is
given by
∞
Cn e−n
2
v(x, t) =
π2 t
sin nπx,
n=1
and the coefficients Cn are determined from (4.59) as
1
(x − 1) sin nπx dx = −
Cn = 2
0
2
.
nπ
Thus,
v(x, t) = −
2
π
∞
1 −n2 π2 t
e
sin nπx,
n
n=1
and finally from (4.63)
u(x, t) = 1 − x −
2
π
∞
1 −n2 π2 t
e
sin nπx.
n
n=1
In general, if the thickness of the plate is l, the solution is
u(x, t) = 1 −
x
2
−
l
π
∞
1 −n2 π2 t/l2
nπx
e
sin
.
n
l
n=1
(4.67)
The solution for the half-space is derived by letting l → ∞. Since
2
lim u(x, t) = 1 −
l→∞
π
∞
l −n2 π2 t/l2
e
sin
nπ
n=1
nπx
l
·
π
,
l
let nπ/l = ξ and π/l = dξ. Then
2
lim u(x, t) = 1 −
l→∞
π
∞
0
x x 1 −ξ2 t
e
sin ξt dξ = 1 − erf √ = erfc √ .
ξ
2 t
2 t
Example 4.4. (Elliptic equation.) Consider the steady-state heat conduction
or potential problem for the rectangle R = {0 < x < a, 0 < y < b}:
uxx + uyy = 0,
x, y ∈ R,
(4.68)
102
4. BERNOULLI’S SEPARATION METHOD
subject to the Dirichlet boundary conditions
u(0, y) = 0 = u(a, y), u(x, 0) = 0, u(x, b) = f (x).
(4.69)
Physically, this problem arises if three edges of a thin isotropic rectangular plate are
insulated and maintained at zero temperature, while the fourth edge is subjected to a
variable temperature f (x) until the steady-state conditions are attained throughout R.
Then the steady-state value of u(x, y) represents the distribution of temperature in the
interior of the plate. As before, we seek a solution of the form u(x, y) = X(x)Y (y),
which, after substitution into Eq (4.68), leads to the set of two ordinary differential
equations :
X − c X = 0,
(4.70)
(4.71)
Y + c Y = 0,
where c is a constant, as in Example 4.2. Since the first three boundary conditions in
(4.69) are homogeneous, they become
X(0) = 0,
X(a) = 0,
Y (0) = 0,
(4.72)
but the fourth boundary condition which is nonhomogeneous must be used separately.
Now, taking c = −λ2 , as before, the solution of (4.70) subject to the first two boundary
conditions in (4.72) leads to the eigenpairs as
λ2n =
n2 π 2
,
a2
Xn (x) = sin
nπx
,
a
n = 1, 2, . . . ,
while for these eigenvalues the solutions of (4.68) satisfying the third boundary condition in (4.72) are
Yn (y) = sinh
nπy
,
a
n = 1, 2, . . . .
(4.73)
Hence, for arbitrary constants Cn , n = 1, 2, . . . , we get
∞
u(x, y) =
Cn sin
n=1
nπx
nπy
sinh
,
a
a
(4.74)
where the coefficients Cn are determined by using the fourth boundary condition in
(4.69). Thus,
∞
u(x, b) = f (x) =
Cn sin
n=1
nπx
nπb
sinh
, 0 < x < a,
a
a
which, in view of formula (4.59), yields
nπb
2 a
nπx
Cn sinh
=
f (x) sin
dx,
a
a 0
a
n = 1, 2, . . . .
(4.75)
4.4. EXAMPLES
103
This solves the problem completely. In particular, if f (x) = f0 = const, then
Cn sinh
nπb
2f0 [1 − (−1)n ]
=
.
a
nπ
Thus, from (4.74), we have
u(x, y) =
2f0
π
∞
1 − (−1)n sin(nπx/a) sinh(nπy/a)
.
n
sinh(nπb/a)
n=1
(4.76)
Example 4.5. Consider the problem of transverse vibrations of a stretched
string of length l fastened at both ends, initially distorted into some given curve and
then allowed to vibrate. Let the equilibrium state of the string be along the x-axis
and one end at the origin, and suppose that the string is initially distorted into a given
curve defined by u = f (x). The problem is to find an expression for u which will be
the solution of the equation
2
∂2u
2 ∂ u
=
c
,
(4.77)
∂t2
∂x2
subject to the following conditions: (i) u(0, t) = 0, (ii) u(l, t) = 0, (iii) u(x, 0) =
du
f (x), and (iv)
(x, 0) = 0. The condition (iv) signifies the fact that the string starts
dt
from rest. To solve this problem, we assume u = eax+bt , b > 0, and substituting it
into Eq (4.77) we find that b2 = a2 c2 , which gives the general solution of Eq (4.77)
for any value of a as u = eax+act . To convert this solution into a trigonometric
form we replace a by ±ai. The u = e(x+ct)ai and u = e−(x+ct)ai become two
solutions of Eq (4.77). Adding these values and dividing by 2 we have cos a(x ± ct),
while subtracting them and dividing by 2i we have sin a(x + ct). This means that
u = cos a(x + ct), u = cos a(x − ct), u = sin a(x + ct), and u = sin a(x − ct) each
is a solution of Eq (4.77). Again, using trigonometric identities these four solutions
can be rewritten as u = cos ax cos act, u = sin ax sin act, u = sin ax cos act, and
u = cos ax sin act. Notice that the third form satisfies conditions (i) and (iv) for any
kπ
value of a and it will satisfy condition (ii) if a =
, where k is an integer. Then, if
l
we take
∞
kπx
kπct
u=
Ak sin
cos
,
(4.78)
l
l
k=1
where Ak , k = 1, 2, . . . , are undetermined coefficients, we have the solution of Eq
(4.77) which satisfies conditions (i), (ii), and (iv). Using condition (iii), we find from
(4.78) that
∞
kπx
2 l
kπx
y=
Ak sin
= f (x), where Ak =
f (x) sin
dx.
l
l 0
l
k=1
Example 4.6. To find the potential function u due to a circular wire ring of
radius a of small cross section, first notice that the potential satisfies the relation
104
4. BERNOULLI’S SEPARATION METHOD
M
, where M is the mass of the ring, and x is the distance from the
+ x2
center of the ring. Using spherical coordinates with the center of the ring as the origin
of the coordinates and the axis of the ring as the polar axis, and using symmetry we
find that the potential, being independent of φ, is the solution of the equation
∂ 2 (ru)
1 ∂
∂u
r
+
sin
θ
= 0,
(4.79)
∂r 2
sin θ ∂θ
∂θ
u = √
a2
M
when θ = 0. We assume u(r, θ) =
+ x2
n
r P (θ), where n > 0 is an integer and P is a function of θ, and substituting it in
(4.79) we find that
rn ∂
∂P
n
n(n + 1)r P +
sin θ
= 0,
sin θ ∂θ
∂θ
subject to the condition that u = √
a2
or dividing by r n ,
n(n + 1)P +
1 ∂
sin θ ∂θ
sin θ
∂P
∂θ
= 0.
To obtain a solution for this equation we set x = cos θ. Then this equation becomes
dP
d
1 − x2
+ n(n + 1)P = 0,
(4.80)
dx
dx
where P is now a function of x. Assuming a series solution P (x) =
ak xk and
k=0
substituting in Eq (4.80) we get
∞
∞
k(k − 1) ak xk−2 − k(k + 1) ak xk + n(n + 1) ak xk = 0.
k=0
Since this relation is true for any x, the coefficients of any given power of x, for
example, of xk , must vanish and thus,
(k + 2)(k + 1) ak+2 − [k(k + 1) − n(n + 1)] ak = 0,
which gives the recurrence relation
ak+2 = −
n(n + 1) − k(k + 1)
ak .
(k + 1)(k + 2)
Now, if k = n, then ak+2 = 0, ak+4 = 0, and so on. Rewriting this relation as
ak = −
(k + 2)(k + 1)
ak+2 ,
(n − k)(n + k + 1)
4.4. EXAMPLES
105
we begin with k = n − 2 and obtain
n(n − 1)
an ,
2 · (2n − 1)
n(n − 1)(n − 2)(n − 3)
=
an ,
2 · 4 · (2n − 1)(2n − 3)
n(n − 1)(n − 2)(n − 3)(n − 4)(n − 5)
=−
an , and so on.
2 · 4 · 6 · (2n − 1)(2n − 3)(2n − 5)
an−2 = −
an−4
an−6
If n is even, then we will end with a0 ; if n is odd, then we will end with a1 . Thus, a
solution of Eq (4.80) is
n(n − 1) n−2 n(n − 1)(n − 2)(n − 3) m−4
n
P (x) = an x −
x
+
x
− ··· ,
2 · (2n − 1)
2 · 4 · (2n − 1)(2n − 3)
(2n − 1)((2n − 3) · · · 1
, which
n!
acts like a normalizing factor to give P (1) = 1. This function P which is a particular
solution of Eq (4.80), is usually denoted by Pn (x) or Pn (cos θ) and is called the
Legendre polynomials of order n, or Legendre’s coefficients or surface zonal harmonics. A particular solution of Eq (4.79) is u(r, θ) = rn Pn (cos θ) and is sometimes
called a solid zonal harmonic. This solution can be written in the form of an infinite
series as
where an is arbitrary. It is convenient to take an =
∞
Ak r k Pk (cos θ),
u(r, θ) =
(4.81)
k=0
where Ak are arbitrary constants to be determined from the condition at θ = 0, i.e.,
∞
M
Ak r k = √
. Since by binomial theorem
2
a + r2
k=0
M
M
1 r2
1 · 3 r4
1 · 3 · 5 r6
√
=
1−
+
−
+
·
·
·
, r < a,
a
2 a2
2 · 4 a4
2 · 4 · 6 a6
a2 + r 2
the required solution for r < a is
u(r, θ) =
M
1 r2
1 · 3 r4
P0 (cos θ) −
P
(cos
θ)
+
P4 (cos θ)
2
a
2 a2
2 · 4 a4
1 · 3 · 5 r6
−
P6 (cos θ) + · · · .
2 · 4 · 6 a6
The formula to compute a few values of Pn (x) is
Pn (x) =
(2n − 1)((2n − 3) · · · 1 n
n(n − 1) n−2
x −
x
n!
2 · (2n − 1)
n(n − 1)(n − 2)(n − 3) n−4
+
x
2 · 4 · (2n − 1)(2n − 3)
n(n − 1)(n − 2)(n − 3)(n − 4)(n − 5) n−6
−
x
+··· .
2 · 4 · 6 · (2n − 1)(2n − 3)(2n − 5)
(4.82)
106
4. BERNOULLI’S SEPARATION METHOD
The values of Pn (x) and Pn (cos θ) for n = 1, . . . , 5 are given in §C.1.1 and §C.1.2.
We will now justify the value of the coefficient an which we have taken above
(2n − 1)((2n − 3) · · · 1
as equal to
in the definition (4.82) for Pn (x). Consider
n!
the potential u = 1/r, where r = |x = x1 |, x = (x, y, z), x1 = (x1 , y1 , z1 ).
In the spherical coordinates x = r sin θ cos φ, y = r sin θ sin φ, z = r cos θ, we
choose x = (r, θ, φ) and x1 = (r1 , θ1 , φ), where the point x1 is chosen so as to be
independent of φ, and θ1 = 0. Then we find that
u=
1
1
=
,
2
r
r − 2rr1 cos θ + r12
is a solution of Eq (4.79). This solution can be written as
(a) :
or
(b) :
r −1/2
1
r1
1
1 − 2 cos θ +
;
r
r
r
−1/2
1
r
r
u=
1 − 2 cos θ +
.
r1
r1
r1
u=
r1
r
in (a), and t =
in (b). Then, since x = cos θ, we will now discuss
r
r1
−1/2
in these two cases the expression 1 − 2t cos θ + t2
, which is known as the
generating function for the Legendre polynomials Pn (x).
We set t =
In case (a), this generating function can be developed into a convergent series
involving positive integral powers of t for t < 1:
1 − 2t cos θ + t2
−1/2
∞
pn (cos θ)tn ,
=
n=0
where the coefficients pn (cos θ) are yet to be determined. Thus, in this case we have
∞
r n
1
1
u=
pn (cos θ)
. Substituting it in Legendre’s equation (4.15) yields
r n=0
r
∞
n=0
r1n
r n+1
1 d
n(n + 1)pn (cos θ) +
sin θ dθ
dpn
sin θ
= 0, r > r1 .
dθ
Since the coefficients of each power of r must be zero, we obtain
1 d
dpn
sin θ
+ n(n + 1) pn = 0.
sin θ dθ
dθ
With x = cos θ, this equation reduces to
1 − x2
d2 pn
dpn
− 2x
+ n(n + 1) pn = 0.
2
dx
dx
4.4. EXAMPLES
107
Hence, t = pn (x) = pn (cos θ) is a solution of Legendre’s equation, and so pn (x) ≡
Pn (x). Similarly, in case (b) the generating function can again be developed into
a convergent series
positive integral powers of t for t < 1. Then u =
n
involving
∞
1 r
pn (cos θ)
, which leads to
r1 n=0
r1
∞
n=0
rn
r1n+1
1 d
n(n + 1)pn (cos θ) +
sin θ dθ
dpn
sin θ
= 0, r < r1 ,
dθ
and this again shows that t = pn (x) = pn (cos θ) is a solution of Legendre’s equation,
and so pn (x) ≡ Pn (x). But since pn is the coefficient of the tn in the development of
−1/2
−1/2
the above generating function, we have 1 − 2tx + t2
= [1 − t(2x − t)]
,
which when expanded by the binomial theorem gives
(2n − 1)((2n − 3) · · · 1 n
n(n − 1) n−2
x −
x
n!
2(2n − 1)
n(n − 1)(n − 2)(n − 3) n−4
+
x
− ··· ,
2 · 4 · (2n − 1)(2n − 3)
and this is precisely (4.82) which defines Pn (x).
The constant in formula (4.82) for Pn (x) can be found as follows: It is obviously
the value of Pn (x) when x = 0. Thus, it is the coefficient of tn in the development
−1/2
of 1 + t2
, which is
1 + t2
−1/2
=1−
1 2 1·3 2 1·3·5 6
t +
t −
t − ··· ,
2
2·4
2·4·6
and the coefficient of tn is

1 · 3 · 5 · · · (n − 1)

, if n is an even number,
 (−1)n/2
2 · 4 · 6···n

 (−1)(n−1)/2 1 · 3 · 5 · · · n , if n is an odd number.
2 · 4 · 6 · · · (n − 1)
Example 4.7. Consider the problem of temperature distribution within a solid
sphere of radius 1, one half of the surface of which is kept at the constant temperature
zero and the other half at a constant temperature unity. Taking the diameter perpendicular to the plane separating the two halves as the axis of spherical coordinates, we
solve the equation
1 ∂
1 ∂
∂u
1 ∂ 2u
2 ∂u
r
+
sin
θ
+
= 0,
r2 ∂r
∂r
sin θ ∂θ
∂θ
sin2 θ ∂φ2
which, because of symmetry, is independent of φ, and reduces to Eq (4.79), subject
to the conditions that u = 1 from θ = 0 to θ = π/2, and u = 0 from θ = π/2 to
108
4. BERNOULLI’S SEPARATION METHOD
θ = π. Then a particular solution of this equation is u = rn Pn (cos θ), where n > 0
∞
is an integer, and u(r, θ) =
Ak rk Pk (cos θ), where Ak are yet undetermined
k=0
coefficients. For r = 1, we have u =
∞
Ak Pk (cos θ), or equivalently, u =
k=0
∞
Ak Pk (x), where u = 0 from x = −1 to x = 0 and u = 1 from x − 0 to
1
x = 1. Thus, since P0 (x) = 1 (see §C.1), we have A0 = 12 0 P0 (x) dx = 12 , and
2n + 1 1
An =
P (x) dx for n ≥ 1, or since
0 n
2

1
 0 if n is even,
1
1 · 3 · 5···n
Pn (x) dx =
if n is odd,
 (−1)(n−1)/2
0
n(n + 1) 2 · 4 · 6 · · · (n − 1)
k=0
we obtain the desired solution as
1 3
7 1
11 1 · 3
+ P1 (x) −
P3 (x) +
P5 (x) + · · ·
2 4
8 2
12 2 · 4
1 3
7 1
11 1 · 3
= + P1 (cos θ) −
P3 (cos θ) +
P5 (cos θ) + · · · .
2 4
8 2
12 2 · 4
u=
∂2u
∂2u
+
= 0. Using
∂x2
∂y 2
Bernoulli’s separation method, let u = X(x)Y (y), which after substitution in the
Y X above equation, leads to
=−
= m2 , or equivalently, to
Y
X
Example 4.8. The 2-D Laplace’s equation is
Y − m2 Y = 0,
X + m2 X = 0.
The general solution of the first equation is Y = A emy + B e−my , and that of
the second X = C sin mx + D cos mx. Thus, there are four particular solutions of the 2-D Laplace’s equation, namely, u(x, y) = emy sin mx, u(x, y) =
emy cos mx, u(x, y) = e−my sin mx, u(x, y) = e−my cos mx.
2
∂2u
2∂ u
=
c
. Using Bernoulli’s
∂t2
∂x2
separation method, let u = T (t)X(x), which after substitution in the above equation,
T X leads to 2 =
= −m2 , or equivalently, to
c T
X
Example 4.9. The 1-D wave equation is
X + m2 X = 0,
T + c2 m2 T = 0.
The general solution of the first equation is X = A sin mx+B cos mx, and that of the
second T = C sin mct + D cos mct. Hence, there are four particular solutions of the
4.4. EXAMPLES
109
1-D wave equation: u = sin mx cos mct, u = sin mx sin mct, u = cos mx cos mct,
and u = cos mx sin mct.
Example 4.10. Laplace’s equation in cylindrical coordinates is given by
∂ 2 u 1 ∂u
1 ∂ 2u ∂ 2u
+
+ 2 2 + 2 = 0.
2
∂r
r ∂r
r ∂θ
∂z
(4.83)
Consider the following problems:
(a) The convex surface and one base of a cylinder of radius a and length b are
kept at the constant temperature zero, and the temperature at each point of the other
base is a preassigned function f (r) of the distance from the center of the base. It
is required to determine the temperature at any point of the cylinder after the steady
state is attained. Thus, we must solve Eq (4.83) subject to the following conditions:
u(r, θ, 0) = 0, u(a, θ, z) = 0, and u(r, θ, b) = f (r). Since, because of the symmetry
∂2u
= 0, we use Bernoulli’s separation method by taking u = R(r)Z(z), which
∂θ2
breaks up Eq (4.83) into two equations
d2 R 1 dR
+
+ m2 R = 0,
dr2
r dr
d2 Z
− m2 Z = 0.
dz 2
Solving these equations, we find the particular solutions of the problem as
u(r, z) = J0 (mr) sinh(mz),
u(r, z) = J0 (mr) cosh(mz).
If mk is a zero of J0 (ma), then u(r, z) = J0 (mk r) sinh(mk z) satisfies Laplace’s
∞
equation and the first two conditions. If f (r) =
Ak J0 (mk r), where mk are
zeros of J0 (ma), then the required solution is
∞
u(r, z) =
Ak
k=1
k=1
sinh(mk z)
J0 (mk r).
sin(mk b)
(4.84)
(b) If instead of keeping the convex surface of the cylinder at temperature zero,
we surround it it by an insulation impervious to heat, the condition u(a, z) = 0 is
dJ0 (mr) then replaced by ur (a, z) = 0, or by
= 0, i.e., by m J0 (ma) = 0, or
dr
r=a
by J1 (ma) = 0. Then (4.84) will the solution of this problem where mk are now the
zeros of J1 (ma).
(c) If instead of keeping the convex surface of the cylinder at temperature zero,
we allow it to cool in air at the temperature zero, the condition u(a, z) = 0 will
be replaced by ur (a, z) + hu(a, z) = 0, i.e., u(r, z) = J0 (mr) sinh(mz) will be
replaced by mJ0 (ma) + h J0 (ma) = 0, that is, by ma J0 (ma) + ah J0 (ma) = 0
110
4. BERNOULLI’S SEPARATION METHOD
or by ma J1 (ma) − ha J0 (ma) = 0. Then (4.84) will the solution of this problem
where mk are now the zeros of ma J1 (ma) − ha J0 (ma) = 0.
Example 4.11. Consider Laplace’s equation (4.11) in spherical coordinates,
independent of φ:
∂ 2 u 2 ∂u
1
+
+ 2
∂r2
r ∂r
r sin θ
∂u
sin θ
= 0.
∂θ
(4.85)
Using Bernoulli’s separation method, let u = R(r)Θ(θ), which after substitution in
Eq (4.85) gives
r d2 (rR)
1 d
dΘ
=
−
sin
θ
.
(4.86)
R dr 2
sin θ dθ
dθ
Since each side of Eq (4.86) must be equal to a constant, say α2 , this equation is
equivalent to the following two equations:
d2 (rR)
− α2 R = 0,
dr2 1 d
dΘ
sin θ
+ α2 Θ = 0.
sin θ dθ
dθ
r
(4.87)
(4.88)
Eq (4.87) can be written as
r2
d2 R
dR
+ 2r
− α2 R = 0,
dr 2
dr
and has the general solution as R(r) = Arn + Br m , where n = − 12 + α2 + 14
and m = − 12 − α2 + 14 . Thus, m = −n − 1, and we can write α2 = n(n + 1),
where n is arbitrary. Then R(r) = Ar n + Br−n−1 . Thus, R = r n and R = r −n−1
are particular solutions of Eq (4.87) with α2 = n(n + 1), i.e., of the equation
r2
d2 R
dR
+ 2r
− n(n + 1)R = 0.
dr 2
dr
(4.89)
Again, with this value of α2 Eq (4.88) becomes
1 d
sin θ dθ
dΘ
sin θ
+ n(n + 1)Θ = 0,
dθ
where n > 0 is an integer (see Example 4.6), so that its particular solution is Θ(θ) =
1
Pn (cos θ). Hence, u(r, θ) = rn Pn (cos θ) and u(r, θ) = n+1 Pn (cos θ), where n
r
is a positive integer, are particular solutions of Eq (4.85), where the first of these
solutions is known as a solid zonal harmonic.
4.4. EXAMPLES
111
We will now determine a series solution of the equation
x2
d2 u
du
+ 2x
− n(n + 1)u = 0,
dx2
dx
(4.90)
where n is a positive integer. This equation is same as Eq (4.89) with r replaced
∞
by x and R by u. Assume that u =
ak xk . Then substituting it in Eq (4.90)
k=0
∞
[k(k + 1) − n(n + 1)] ak xk = 0. This, being an identity equation, gives
we get
k=0
[k(k + 1) − n(n + 1)] ak ≡ 0. Hence, ak = 0 for all values of n except n = k and
n = −k − 1. Then, u = Axn + Bx−n−1 is the general solution of Eq (4.90), and
1
u = xn and u = n+1 are its particular solutions. If n is not a positive integer, this
x
method fails and does not yield any result.
Example 4.12. (Circular drum) Consider the problem of vibrations of a
stretched circular membrane of radius a fastened on its circumference, which can be
referred to as a circular drum, subject to the conditions that the membrane is initially
distorted into a given form which has circular symmetry about the axis through the
center perpendicular to the plane of the boundary, and then allowed to vibrate. Recall
that a function of the coordinates of a point has circular symmetry about an axis when
its value is not affected by rotating the point through an angle about the axis, and a
surface has circular symmetry about an axis when it is a surface of revolution about
the axis. Using the spherical coordinates the distortion u due to the symmetry is
independent of φ, and is governed by the equation
∂2u
= c2
∂t2
∂ 2 u 1 ∂u
+
∂r2
r ∂r
,
(4.91)
subject to the conditions u(r, 0) = f (r), ut (r, 0) = 0, and u(a, t) = 0. Using
Bernoulli’s separationmethod, let u(r, t) = R(r)T (t). Substituting it in Eq (4.91)
1 2
we get RT = c T R + R . Since the left side of this equation does not
r
involve r while the right side does not contain t, each side must be a constant, say
−m2 , where where m is an yet undetermined constant. Then this equation can be
written as
T 1
R
=
R +
= −m2 ,
c2 T
R
r
which is equivalent to the following two equations:
T + m2 c2 T = 0,
1
R + R + m2 R = 0.
r
(4.92)
(4.93)
112
4. BERNOULLI’S SEPARATION METHOD
The particular solutions of Eq (4.92) are T = cos mct and T = sin mct. To solve Eq
(4.93), set r = x/m. Then it becomes
d2 R 1 dR
+
+ R = 0.
dx2
x dx
(4.94)
This equation was first studied by Fourier while considering the cooling of a cylinder,
and therefore, it is sometimes known as Fourier’s equation. We will find a series
solution for Eq (4.94): Assume that R can be expressed in the form of an infinite
∞
series in powers of x, i.e., take R =
ak xk . Then, substituting it in Eq (4.94), we
k=0
obtain
∞
k(k − 1)ak xk−2 + kak xk−2 + ak xk = 0.
k=0
This equation must be true for all values of x. The coefficient of any power of x, say of
xk−2 , must vanish, which gives k(k − 1)ak + kak + ak−2 = 0, or k 2 ak + ak−2 = 0,
or
ak−2 = −k 2 ak .
(4.95)
If k = 0, then ak−2 = 0, ak−4 = 0, and so on. Thus, we start with k = 0, and (4.95)
gives the recurrence relation
ak−2
ak = − 2 ,
k
a0
a0
a0
whence we get a2 = − 2 , a4 = 2 2 , a6 = − 2 2 2 , and so on. Hence,
2
2 ·4
2 ·4 ·6
x2
x4
x6
R = a0 1 − 2 + 2 2 − 2 2 2 + · · · ,
2
2 ·4
2 ·4 ·6
where a0 is arbitrary, provided the series is convergent. If we take a0 = 1, then
R = J0 (x), where J0 (x) is the Bessel function of order zero, and the corresponding
series is convergent for all real or purely imaginary values of x, since the series
J0 (r) = 1 −
r2
r4
r6
+ 2 2 − 2 2 2 + ···
2
2
2 ·4
2 ·4 ·6
composed of the moduli of x, is convergent
for all r = |x|. Using the ratio test for
an+1 r2
this series, we find that the ratio =
→ 0 as n → ∞. Hence, J0 (x) is
an
4n2
absolutely convergent. The function J0 (x) is called a cylindrical harmonic or the
Bessel function of order zero.
Note that Eq (4.94) was obtained by substituting x = mr; thus,
R = J0 (mr) = 1 −
(mr)2
(mr)4
(mr)6
+
−
+ ··· .
22
22 · 42
22 · 42 · 62
4.4. EXAMPLES
113
Hence, u(r, t) = J0 (mr) cos mct or u(r, t) = J0 (mr) sin mct is a particular solution
of Eq (4.91). But since u(r, t) = J0 (mr) cos mct satisfies the prescribed second
condition for any value of m, in order to satisfy the first condition, m must be taken
such that J0 (ma) = 0, i.e., m must be a zero of J0 (ma). It is known that J0 (x) has
an infinite number of real positive zeros (see Abramowitz and Stegun [1965]), any
one of which can be obtained to any required degree of approximation. Let these
x1
x2
x3
zeros be x1 , x2 , x3 , . . . . Then, if
= m1 ,
= m2 ,
= m3 , and so on, we
a
a
a
get
∞
u(r, t) =
Ak J0 (mk r) cos mk ct,
(4.96)
k=1
where Ak are constants, as a solution of Eq (4.91) which satisfies second and third
prescribed conditions. Now, to satisfy the first prescribed condition, for t = 0 the
∞
solution (4.96) reduces to u(r, 0) =
Ak J0 (mk r). If f (r) can be expressed as a
k=1
series of this form, the solution of the problem can be obtained by substituting the
coefficients of that series for Ak , k = 1, 2, . . . .
Example 4.13. (Cooling ball ) Consider the boundary value problem
ut = ∇2 u,
0 ≤ r < 1,
0 ≤ φ ≤ π,
t > 0,
subject to the conditions u(r, φ, 0) = f (r, φ), and u(1, φ, t) = 0, where the function
u = u(r, φ, t) is independent of θ. The problem describes the temperature distribution
in the interior of the unit ball dropped in cold water. The first condition implies that
the temperature u is not uniform but depends only on r, φ and t. Thus, the solution
is assumed formally as
u(r, φ, t) = R(r) Φ(φ) T (t),
(4.97)
which, after separating the variables, gives
2 R
Φ + cot φ Φ
T
2
r
+
=
−α
=
= λ.
R
r2 Φ
T
R +
(4.98)
The left side of Eq (4.98) can be expressed as
R + 2 R
Φ + cot φ Φ
r
+ α2 r 2 = −
.
R
Φ
In order that this equation be satisfied, the terms on each side must be constant. Thus,
R + 2 R
Φ + cot φ Φ
r
+ α2 r 2 = µ = −
.
R
Φ
(4.99)
114
4. BERNOULLI’S SEPARATION METHOD
It is known that µ = n(n + 1) (Courant and Hilbert [1968]). Then Eq (4.99) yields
Φ + cot φ Φ + n(n + 1) Φ = 0,
(4.100)
r 2 R + 2r R + α2 r 2 R − n(n + 1) R = 0.
The first equation reduces to Legendre’s equation by using the substitution z = cos φ
(see Example 4.8), and its bounded solution is given by Φ = Pn (cos φ). If we set
x = αr in the second equation, it becomes
x2 R + 2x R + x2 − n(n + 1) R = 0,
which under the transformation w =
√
x R reduces to Bessel’s equation
x2 w + x w + x2 − (n + 12 )2 w = 0,
and has a bounded solution w = Jn+1/2 (x). Hence, R(r) =
Jn+1/2 (αr)
√
. The
αr
Jn+1/2 (αmn r)
Pn (cos φ) form an orthogonal set. Using the
√
αmn r
boundary condition at r = 1, we get Jn+1/2 (α) = 0. Let αmn denote the positive
zeros of Jn+1/2 (α). Then the solution for the temperature distribution in the unit ball
is given by
eigenfunctions ψmn =
∞
Cmn
m=1
n=0
1
where, after using
−1
Cmn
e−αmn t Jn+1/2 (αmn r)
Pn (cos φ),
√
αmn r
2
u(r, φ, t) =
Pn2 (x) dx =
2
,
2n + 1
π 1
√
(2n + 1) αmn
=
r 3/2 Jn+1/2 (αmn r)Pn (cos θ)f (r, φ) sin φ dr dφ,
β(αmn )
0 0
and
β(αmn ) = 2
0
1
2
2
r Jn+1/2
(αmn r) dr = Jn+1/2
(αmn ) .
Example 4.14. Consider the steady-state heat conduction or potential problem
uxx + uyy = 0,
0 < x < π,
0 < y < 1,
subject to the mixed boundary conditions
u(x, 0) = u0 cos x,
u(x, 1) = u0 sin2 x,
ux (0, y) = 0 = ux (π, y).
(4.101)
4.4. EXAMPLES
115
The separation of variables technique leads to the following set of ordinary differential
equations:
X + λ2 X = 0,
X (0) = 0 = X (π),
Y − λ2 Y = 0.
(4.102)
(4.103)
The eigenvalues and the corresponding eigenfunctions for (4.102) are
λ0 = 0,
X0 (x) = 1,
2
λn = n ,
Xn (x) = cos nx,
n = 1, 2, . . . ,
and subsequently the solutions of (4.103) are
Yn (y) =
A0 + B0 y, n = 0,
An cosh ny + Bn sinh ny,
n = 1, 2, . . . .
(4.104)
Hence, using the superposition principle, we get
∞
u(x, y) = A0 + B0 y +
[An cosh ny + Bn sinh ny] cos nx.
(4.105)
n=1
Now, the first boundary condition in (4.101) leads to
∞
u(x, 0) = A0 +
An cos nx = u0 cos x.
(4.106)
n=1
By comparing the coefficients of similar terms on both sides of (4.106), we find that
A0 = 0, A1 = u0 , and An = 0 for n ≥ 2. Hence, the solution becomes
∞
u(x, y) = B0 y + u0 cosh y cos x +
Bn sinh ny cos nx.
(4.107)
n=1
Similarly, using the second boundary condition in (4.101) we find from (4.107) that
∞
u(x, 1) = B0 + u0 cosh 1 cos x +
Bn sinh n cos nx
n=1
= u0 sin2 x = u0
1 − cos 2x
,
2
from which, after comparing the coefficients of similar terms on both sides, we get
B0 =
u0
u0 cosh 1
u0
, B1 = −
, B2 = −
, Bn = 0 for n ≥ 3. Hence,
2
sinh 1
2 sinh 2
116
4. BERNOULLI’S SEPARATION METHOD
from (4.107) the general solution is given by
u0
cosh 1 sinh y
sinh 2y
u(x, y) =
y + u0 cosh y −
cos x − u0
cos 2x
2
sinh 1
2 sinh 2
(4.108)
1
sinh(1 − y)
sinh 2y
= u0 y +
cos x −
cos 2x .
2
sinh 1
2 sinh 2
4.5. Exercises
4.1. Show that the general equation of the hydrogen atom
∇2 u(r, θ, φ) + k 2 + f (r) +
12
1
g(θ) + 2 2 h(φ) u(r, θ, φ) = 0
r
r sin θ
is separable, where the functions f , g and h are functions of the variable shown,
and k 2 is a constant.
Hint. Use Bernoulli’s separation method: take u = R(r)Θ(θ)Φ(φ), and separate the variables.
4.2. Show that if x = 1, then Pn (x) = 1.
−1/2
Hint. The generating function is 1 − 2t + t2
= (1 − t)−1 = 1 + t + t2 +
t3 + · · · . The coefficients of each power being 1, we get Pn (1) = 1.
4.3. Compute the value of potential u(r, θ) for the circular ring (see Example 4.6),
using the following data: M = 1, a = 1, and
(a) r = 0.2, θ = 0,
(b) r = 0.2, θ = π/4,
(c) r = 0.2, θ = π/2,
(d) r = 0.6, θ = 0,
(e) r = 0.6, θ = π/6,
(f) r = 0.6, θ = π/3,
(g) r = 0.6, θ = π/2.
Ans. (a) 0.9806, (b) 0.9948, (c) 1.0102, (d) 0.8540, (e) 0.8941,
(f) 1.0037, (g) 1.0763.
4.4. Show that Laplace’s equation is not completely separable in toroidal coordinates.
Note. The toroidal coordinates are defined by
x=
a sinh v cos φ
a sinh v sin φ
a sinh u
, y=
, z=
,
cosh v − cos u
cosh v − cos u
cosh v − cos u
h21 = h22 =
a2
a2 sinh2 v
2
,
h
=
,
3
(cosh v − cos u)2
(cosh v − cos u)2
and Laplace’s equation is
(cosh v − cos u)3 ∂
1
∂Φ
a2
∂u cosh v − cos u ∂u
3
(cosh v − cos u) ∂
sinh v
∂Φ
(cosh v − cos u)2 ∂Φ
+
+
.
2
a sinh v
∂v cosh v − cos u ∂v
∂φ
a2 sinh2 v
∇2 Φ =
4.5. EXERCISES
117
This coordinate system is seldom used although it has some very rare physical
applications, such as vortex rings.
4.5. Show that Bernoulli’s separation method applied to Laplace’s equation in parabolic cylindrical coordinates leads to Hermite’s equation y − 2xy + 2ny = 0.
Hint. Parabolic cylindrical coordinates are defined by x = 12 u2 − v 2 , y =
u v, z = z, and Laplace’s equation is
2
1
∂ w ∂2w
∂2w
∇2 w = 2
+
+
.
2
2
2
u +v
∂u
∂v
∂z 2
4.6. If Jn (x) is a solution of Bessel’s equation, verify that u(x) = xa Jn (bxc )
satisfies the equation
1 − 2a a2 − n2 c2
c−1 2
u +
u + bcx
+
u = 0.
x
x2
4.7. Solve ut = a2 uxx + f (x, t), 0 < x < l, subject to the boundary conditions
u(0, t) = 0 = u(l, t) for 0 ≤ x ≤ l, and the initial condition u(x, 0) = 0 for
t > 0.
∞
nπx
Ans. u(x, t) =
un (t) sin
, where
l
n=1
t
un (t) =
n2 π 2 a2 (τ −t)/l2
e
fn (τ ) dτ,
0
2
fn (t) =
l
l
f (ξ, t) sin
0
nπξ
dξ.
l
4.8. Find the interior temperature of the cooling ball of Example 4.13, if
1, 0 ≤ φ < π/2
f (r, φ) =
0, π/2 ≤ φ < π.
∞
Ans. u(r, φ, t) =
Cmn
m,n=1
Cmn
Jn+1/2 (αmn r)Pn (cos φ) −α2 t
e mn ,
√
αmn r
1
αmn 0 r 3/2 Jn+1/2 (αmn r) dr
=
1 2
2n(n + 1) 0 rJn+1/2
(αmn r) dr
√
1
(2n + 1)Pn (0) αmn 0 r 3/2 Jn+1/2 (αmn r) dr
=
,
2
n(n + 1) Jn−1/2
(αmn )
(2n + 1)Pn (0)
√
where αmn are the consecutive positive roots of Jn+1/2 (α) = 0.
4.9. Solve ut = uxx , −π < x < π, subject to the conditions u(x, 0) = f (x),
u(−π, t) = u(π, t), and ux (−π, t) = ux (π, t), where f (x) is a periodic function
118
4. BERNOULLI’S SEPARATION METHOD
of period 2π. This problem describes the heat flow inside a rod of length 2π,
which is shaped in the form of a closed circular ring.
Hint: Assume X(x) = A cos ωx + B sin ωx.
∞
Ans. ωn = n; u(x, t) =
e−n t (an cos nx + bn sin nx), where
n=0
1 π
1 π
1 π
an =
f (x) cos nx dx, bn =
f (x) sin nx dx, a0 =
f (x) dx.
−π
π −π
π
2π −π
2
4.10. Solve the problem ut = ∇2 u, r < 1, 0 < z < 1, such that u(r, z, 0) = 1,
u(1, z, t) = 0, and u(r, 0, t) = 0 = u(r, 1, t). This problem describes the
temperature distribution inside a homogeneous isotropic solid circular cylinder.
∞
Ans. u(r, z, t) =
Cmn e−(λm +n
2
m,n=1
2
π 2 )t
J0 (λm r) sin nπz, where λm are
4 (1 − (−1)n )
.
nπλm J1 (λm )
the zeros of J0 , and Cmn =
4.11. Solve the steady-state problem of temperature distribution in a half-cylinder
0 ≤ r ≤ 1, 0 ≤ θ ≤ π, 0 ≤ z ≤ 1, where the flat faces are kept at 0◦ and the
curved surface at 1◦ .
∞
16
Im (nπr) sin nπz
Ans. u(r, θ, z) = 2
sin mθ.
π m,n=1 Im (nπ) mn
m,n odd
4.12. Solve utt = c2 (uxx + uyy ) in the rectangle R = {(x, y) : 0 < x < a, 0 <
y < b}, subject to the condition u = 0 on the boundary of R for t > 0, and
the initial conditions u(x, y, 0) = f (x, y), ut (x, y, 0) = g(x, y). This problem
describes a vibrating rectangular membrane.
∞
mπx
nπy
Ans. u(x, y, t) =
(Amn cos λmn t + Bmn sin λmn ) sin
sin
,
a
b
m,n=1
where
b a
4
mπx
nπy
Amn =
f (x, y) sin
sin
dx dy,
ab 0 0
a
b
b a
4
mπx
nπy
Bmn =
g(x, y) sin
sin
dx dy,
abλmn 0 0
a
b
m2
n2
for m, n = 1, 2, . . . ; the eigenvalues are λmn = cπ
+
.
a2
b2
4.13. Solve uxx − utt = e−a π t sin aπx, subject to the conditions u(x, 0) = 0,
ut (x, 0) = 0, and u(0, t) = u(1, t) = 0, where a is an integer.
2 2
1
Ans. 2 2
cos aπt − e−a π t − aπ sin aπt sin aπx.
2
2
a π (1 + a π )
2
2
4.14. Solve uxx +uyy = 0, under the conditions u(x, 0) = 0 = u(x, π), u(0, y) = 0,
4.5. EXERCISES
and u(π, y) = cos2 y.
∞
Ans. u =
Cn sinh nx sin ny, where Cn =
n=1
119
2
π sinh nπ
π
cos2 y sin ny dy =
0
1
1
n
1 − (−1)n
−
, n = 2, and C2 = 0.
π sinh nπ
n n2 − 4
4.15. Solve Poisson’s equation uxx + uyy = −1, 0 < x, y < 1, subject to the
Dirichlet boundary conditions u(0, y) = 0 = u(1, y) = u(x, 0) = u(x, 1).
∞
16
sin jπx sin kπy
Ans. u(x, y) = 4
.
π
jk (j 2 + k 2 )
j,k=1
j,k odd
1
4.16. Solve urr + ur = utt , subject to the conditions u(r, 0) = u0 r, ut (r, 0) = 0,
r
and u(a, t) = au0 , lim u(r, t) < ∞, where u0 is a constant.
r→0
∞
Ai J0 (αi r) cos αi t + u0 a, where αi are the consecutive positive
Ans. u =
i=1
roots of J0 (αa) = 0, and
a
a
r(r − a) J0 (αi r) dr
r(r − a) J0 (αi r) dr
0 A i = u0
= 2u0 0
.
a
2
a2 J12 (αi a)
r J0 (αi r) dr
0
1
1
4.17. Solve urr + ur + 2 uθθ = 0, subject to the conditions u = 0 for θ = 0 or
r
r
π/2, and ur = sin θ at r = a.
∞
4(−1)n+1
Ans. u =
Cn r 2n sin 2nθ, where Cn =
.
π(4n2 − 1)a2n
n=1
4.18. Solve the problem of vibrations of a plucked string of length l, stretched
and fixed at both ends. The string is raised to a height b at x = a and then
released to vibrate freely. The problem to be solved consists of the wave equation
utt = c2 uxx , 0 < x < l, t > 0, subject to the boundary
 bxconditions u(0, t) =

, 0 ≤ x ≤ a,

a
0 = u(l, t) and the initial condition u(x, 0) = f (x) =

 b(l − x) , a ≤ x ≤ l.
l−a
Hint. See Example 4.1; note that h(x) = 0, so Bn = 0, and
2 l
nπx
An =
f (x) sin
dx
l 0
l
a
l
2
bx
nπx
b(l − x)
nπx
2bl2
nπa
=
sin
dx+
sin
dx =
sin
.
2
2
l 0 a
l
l−a
l
a(l − a)π n
l
a
Ans. u(x, t) =
2bl2
a(l − a)π 2
∞
1
nπa
nπx
nπct
sin
sin
cos
.
2
n
l
l
l
n=1
120
4. BERNOULLI’S SEPARATION METHOD
4.19. Show that the temperature distribution at any time t in a long cylinder of radius
a, which is kept inside at temperature 1 throughout and whose convex surface out∞
2
2
side is kept at temperature zero, is given by u(r, t) =
Ak e−c mk t J0 (mk r),
where mk are the zeros of J0 (ma), and 1 =
∞
k=1
Ak J0 (mk r).
k=1
4.20. Use Bernoulli’s separation method to solve the problem of transverse vibrations
of a beam governed by utt + a2 uxxxx = 0, subject to the conditions u(0, t) =
u(L, t) = uxx (0, t) = uxx (L, t) = ut (x, 0) = 0, and u(x, 0) = f (x).
X (4)
T Solution. Let u = X(x)T (t), then we have
= − 2 = λ, where λ is
X
a T
a parameter. By standard arguments it can be shown that the relevant values of
λ are positive. Let λ = α4 . Then the solutions for X and T are given by X =
A cos αx + B sin αx + C cosh αx + D sin αx, and T = E cos α2 t + F sin α2 t.
The condition X(0) = 0 means A + C = 0, and X(L) = 0 yields A cos αL +
B sin αL + C cosh αL + D sin αL = 0, and Xxx (0) = 0 implies 2α2 A = 0,
which gives A = C = 0, and Xxx (L) = 0, which yields α2 (B sinh αL −
D sin αL) = 0. Thus, we have a pair of two homogeneous equations:
B sinh αL − D sin αL = 0,
B sinh αL + D sin αL = 0.
For B and D to have nontrivial values, we must havesinh αL sin αL = 0, i.e.,
αL = nπ, and B = 0 and T (0) = 0 are equivalent to F = 0. Absorbing E in D
and using the initial condition, we get:
∞
nπx
a n2 π 2 t
2 L
nπx
u=
Dn sin
cos
, where Dn =
f (x) sin
dx.
2
L
L
L
L
n=1
0
5
Integral Transforms
The integral transform method is very useful in the study of linear partial differential
equations and determination of Green’s functions. A function f (x) may be transb
formed by the formula F (s) = a k(s, x)f (x) dx, provided F (s) exists, where s
denotes the variable of the transform, F (s) is the integral transform of f (x), and
k(s, x) is the kernel of the transform. An integral transform is a linear transformation, which, when applied to a linear initial or boundary value problem, reduces the
number of independent variables by one for each of its application. Thus, a partial
differential equation can be reduced to an algebraic equation by repeated application of integral transforms. The algebraic problem is generally easy to solve for the
function F (s), and the solution of the problem is found by obtaining the function
f (x) from F (s) by inversion formula of the transform applied to the problem. In this
chapter we will discuss basic definitions and some important result of the Laplace,
Fourier, finite Fourier, multiple Laplace and Fourier, and Hankel transforms, which
are used in this book. Tables containing certain important pairs for these transforms
are provided in Appendix D.
5.1. Integral Transform Pairs
Definitions of some transforms and their inverses are given below.
∞
1. Laplace transform: L{f (t)} ≡ F (s) = f (s) =
f (t) e−st dt, {s} > 0,
and its inverse: L−1 {F (s)} ≡ f (t) =
1
2πi
0
c+i∞
F (s) est ds, c > 0.
c−i∞
2. Fourier cosine transform: Fc {f (x)} ≡ fc (α) =
and its inverse: F −1 {fc (α)} ≡ f (x) =
2
π
3. Fourier sine transform: Fs {f (x)} ≡ fs (α) =
121
∞
0
2
π
2
π
∞
f (x) cos(xα) dx,
0
fc (α) cos(xα) dα.
∞
f (x) sin(xα) dx,
0
122
5. INTEGRAL TRANSFORMS
and its inverse: F
−1
{fs (α)} ≡ f (x) =
2
π
∞
fs (α) sin(xα) dα.
0
∞
1
√
4. Fourier complex transform: F{f (x)} = f (α) =
f (x) eixα dx,
2π −∞
∞
1
and its inverse: F −1 {f(α)} ≡ f (x) = √
f(α)e−ixα dα.
2π −∞
5. Hankel transform of order n:
∞
Hn {f (x)} ≡ Fn (σ) = fn (σ) =
xf (x)Jn (σx) dx,
0
and its inverse: Hn−1 {Fn (σ)} ≡ f (x) =
∞
σFn (σ) Jn (σx) dσ.
0
These definitions are not unique, particularly in the case of Fourier transforms.
5.2. Laplace Transform
Some basic properties of the Laplace transform are:
(i) L eat f (t) = F (s − a), and L−1 {F (s − a)} = eat f (t).
(ii) L {H(t − a)f (t − a)} = e−as F (s), and L−1 e−as F (s) = H(t − a)f (t − a).
(iii) Convolution Theorem: L−1 {G(s)F (s)} ≡ F G
t
t
= 0 f (t − u)g(u) du = 0 f (u)g(t − u) du.
(iv) L
dn f (t)
dtn
= sn F (s)−sn−1 f (0)−sn−2 f (0)−· · ·−sf (n−2) (0)−f (n−1) (0),
dn
and L−1 {sn F (s)} = n f (t).
dt
t
1
(v) L−1
F (s) =
f (u)du.
s
0
(vi) L {tn f (t)} = (−1)n
n
dn F
−1
nd F
,
and
L
(−1)
dsn
dsn
= tn f (t).
(vii) If L {f (x, t)} = F (x, s), then
L
∂f (x, t)
∂x
=
∂F (x, s)
,
∂x
and
L−1
∂F (x, s)
∂x
=
∂f(x, t)
.
∂x
The last two results, based on the Leibniz theorem, are extremely effective. This
theorem states that if g(x, t) is an integrable function of t for each value of x, and the
∂g(x, t)
partial derivative
exists and is continuous in the region under consideration,
b ∂x
b
∂g(x, t)
and if f (x) =
g(x, t)dt, then f (x) =
dt.
∂x
a
a
5.2. LAPLACE TRANSFORM
123
Some frequently used Laplace transforms are as follows.
1
1
L {eat } =
, L {teat } =
,
s−a
(s − a)2
n!
. After
(s − a)n+1
replacing a by ib, choosing an appropriate n, and comparing the real and imaginary
parts on both sides, we get the Laplace transforms of functions tn cos bt and tn sin bt,
and then combining with Property (i), we obtain the Laplace transforms of functions
tn eat cos bt and tn eat sin bt. For example, if we choose n = 2, then
2!
L t2 eat =
, which after setting letting a = ib gives L t2 eibt =
3
(s − a)
2!
2(s + ib)3
, and L t2 (cos bt + i sin bt) = 2
. Expanding the numerator
3
(s − ib)
(s + b2 )3
2(s3 + 3is2 b − 3sb2 − ib3 )
on the right side we get L t2 (cos bt + i sin bt) =
.
(s2 + b2 )3
Equating the real and imaginary parts in the above formula we have
2(s3 − 3sb2 )
2(3s2 b − b3 )
L t2 cos bt =
, L t2 sin bt =
,
2
2
3
(s + b )
(s2 + b2 )3
and transforms of L eat t2 cos bt and Leat t2 sin bt can now be easily obtained.
An important Laplace inverse is
which by repeated differentiation n-times gives L {tn eat } =
√
e−a s
a
L
= erfc √ ,
(5.1)
s
2 t
2 x −u2
2 ∞ −u2
where erf(x) = √ 0 e
du, and erfc(x) = 1 − erf(x) = √
e
du.
π
π x
−1
√
e−a s
√
Example 5.1. L
is obtained by differentiating formula (5.1) with
s√
e−a s
1
2
respect to a. Again, L−1 √
= √ e−a /4t is obtained after differentiating
s
πt
(5.1) with respect to a and canceling
out the negative sign on both sides. The classical
√
e−a s
method of deriving L−1 √
is by contour integration, or an interesting method
s
is given in Churchill [1972], which is explained in the next example.
−1
√
√
e−a s
dy
Example 5.2. Define √
= y and e−a s = z. Then y =
=
s
ds
√
1
a −a√s
− 3/2 e−a s −
e
, which yields 2sy + y + az = 0. Similarly, z =
2s
2s
√
a
− √ e−a s yields 2z + ay = 0. Taking the inverse transform of these equations,
2 s
we obtain aG − F − 2tF = 0, and aF − 2tG = 0, where L−1 {y} = F (t) and
1 a2 F
L−1 {z} = G(t). From these two equations in F and G, we get F =
−F ,
2t 2t
124
5. INTEGRAL TRANSFORMS
A
aA
2
2
whose solution is F = √ e−a /4t , which gives G = √ e−a /4t . Note that if
3
t
2 t
1
1
1
a = 0, then y = √ , and F (t) = √ implies that A = √ . Hence,
s
π
πt
2
2
1
a
F (t) = √ e−a /4t , G = √
e−a /4t .
3
πt
πt
√
e−a s
√
s
2
1
= √ e−a /4t with respect to a from 0 to a and
πt
√
−a s
1
π
−1 e
obtain L
. In this problem we have assumed that L √ =
.
s
s
t
√
2
a
Example 5.3. L−1 e−a s = √
e−a /4t is obtained by differentiating
3
2 √πt
−a s
−1 e
√
with respect to a the formula for L
in the previous example and canceling
s
out the negative sign.
Then we integrate L
−1
√
e−a s
a
Example 5.4. If we integrate the formula L
= erfc √ with
s
2 t
√
a
a
−x s
x
−1 e
respect to a from 0 to a, we get
L
dx =
erfc √ dx. Now,
s
2 t
0
0
after changing the order of integration and the Laplace inversion and carrying out the
integration on the left side, we get
√
a
−x s
√
−1 e
L
dx = L−1 (s−3/2 − s−3/2 e−a s ),
(5.2)
s
0
−1
and the right side yields
a
a
2
x
x a
1
√
√
√
erfc
dx = [x erfc
+
x e−x /4t dx
0
2 t
2 t
πt 0
0
a
t −a2 /4t
t
= a erfc √ − 2
e
+2
.
π
π
2 t
t
−1
−3/2
Since L
s
=2
, we get
π
√
t −a2 /4t
a
−1
−3/2 −a s
L
s
e
=2
e
− a erfc √ .
π
2 t
We state a very useful theorem without proof.
Theorem 5.1. If G(s) =
∞
Gk (s) is uniformly convergent, then
1
L−1 G(s) = g(t) =
where L
−1
Gk (s) = gk (t).
∞
k=1 gk (t),
(5.3)
5.2. LAPLACE TRANSFORM
125
Example 5.5. Since
L−1 s−3/2 e−1/s
1
1
1
n 1
1
−
+
−
+
·
·
·
+
(−1)
+
·
·
·
s 2!s2
3!s3
n!sn
s3/2
√
∞
∞
√
(−1)n
1 (−1)n (2 t)2n+1
1
√
= L−1
=
= √ sin(2 t),
n+3/2
(2n + 1)!
π 0
π
n! s
0
= L−1
1
this result and second part of Property (iv) give
√
1
L−1 s−1/2 e−1/s = √ cos(2 t).
πt
(5.4)
Example 5.6. Solve the heat conduction equation k Txx = Tt in the semiinfinite medium 0 ≤ x < ∞, −∞ < y, z < ∞, which has an initial zero temperature,
while its face x = 0 is maintained at a time-dependent temperature f (t), and find the
temperature T for t > 0. By applying the Laplace transform to the equation we get
s
T xx = T , where T = L{T }. The solution of this equation is T = Aemx +Be−mx ,
k where m = s/k. Since T remains bounded as x → ∞, we find that A = 0. The
boundary condition at x = 0 in the transform domain yields B = f¯(s), where
f¯(s) is the Laplace transform of f (t). Thus, the solution in the transform domain is
T = f¯(s) e−mx . To carry out Laplace inversion, we use the convolution property
and Example 5.2 and get
t
T =
0
2
x e−x /4kτ
√
f (t − τ ) dτ.
2τ πkτ
x e−x /4kt
√
If f¯(s) = 1, then the solution for T reduces to T =
. This solution is
2t πkt
the fundamental solution for the heat conduction equation for the half-space. In the
special case when f (t) = T0 , the solution is given by
2
T = T0 erfc
x
√
2 kt
.
5.2.1. Definition of Dirac Delta Function. It is known that L{δ(t)} = 1.
We will explain the existence of the δ-function heuristically. Consider the Heaviside
unit step function H(t) which is defined in §2.1.1. Since the Laplace transform of H(t)
1
is H(s) = , then by Property (iv) of the Laplace transforms, LH (t) = sH(s) = 1.
s
Obviously, H (t) vanishes for |t| > 0 and does not exist for t = 0. From the graph of
H(t) (Fig. 2.3), it is clear that there is a vertical jump at t = 0. Therefore, it seems
ε
reasonable to assume that lim H (t) → ∞. But since −ε H (t) dt = 1, it is obvious
t→0
126
5. INTEGRAL TRANSFORMS
that a function like H (t) exists only in the generalized sense. The function H (t) is
a generally denoted by δ(t) such that
δ(t) =
0
for |t| > 0,
∞
for t = 0,
ε
and
δ(t) dt = 1.
−ε
(5.5)
For other properties of δ-function, see §2.1.
Example 5.7. Solve the wave equation utt = c2 uxx subject to the initial
condition u = ut = 0 for t ≤ 0, and the boundary conditions u(0) = 0 and ux (l) =
T . By applying the Laplace transform to the wave equation, we get ūxx = c−2 s2 ū,
which has the solution: ū = A e−sx/c + B esx/c . Applying the boundary conditions
cT
in the transform domain, we get A +B = 0 and −A e−sl/c + B esl/c = 2 . Solving
s
for A and B and substituting their values in the solution for ū, we get
sx/c
c T sinh(sx/c)
cT
e
− e−sx/c
=
s2 cosh(sl/c)
s2
esl/c + e−sl/c
∞
T = 2
(−1)n e−s[(2n+1)l−x]/c − e−s[(2n+1)l+x]/c ,
s 0
ū(x, s) =
which, after inversion, gives
u(x, t) = T
∞
0
(2n + 1)l + x
(2n + 1)l + x
(−1)
t−
H t−
c
c
(2n + 1)l − x
(2n + 1)l − x
− t−
H t−
.
c
c
n
5.3. Fourier Integral Theorems
The following three theorems are the basis of Fourier transforms:
Theorem 5.2. (Fourier integral theorem) If f (x) satisfies the Dirichlet conditions on the entire real line and is absolutely integrable on (−∞, ∞), then
1
1
f (x + 0) + f (x − 0) =
2
2π
∞
−iαx
e
−∞
∞
dα
−∞
f (u)eiαu du.
(5.6)
Theorem 5.3. (Fourier cosine theorem) If f (x) satisfies the Dirichlet conditions on the non-negative real line and is absolutely integrable on (0, ∞),
then
∞
1
2 ∞
f (x + 0) + f (x − 0) =
dα
f (u) cos(αu) cos(αx) du.
(5.7)
2
π 0
0
5.3. FOURIER INTEGRAL THEOREMS
127
Theorem 5.4. (Fourier sine theorem) If f (x) satisfies the Dirichlet conditions on the non-negative real line and is absolutely integrable on (0, ∞),
then
∞
1
2 ∞
f (x + 0) + f (x − 0) =
dα
f (u) sin(αu) sin(αx) du.
(5.8)
2
π 0
0
1
[f (x + 0) + f (x − 0)] = f (x).
2
These three integrals form the basis of the Fourier transforms. For proof, see
Sneddon [1957].
If f (x) is continuous, then
5.3.1. Properties of Fourier Transforms. Let F{f (x)} = f(α). Then
(1) F{f (x − a)} = eiαa f(α).
1 (2) F{f (ax)} =
f (α/a).
|a|
(3) F{eiax f (x)} = f(α + a).
(4) F{f(x)} = f (−α).
dn f (α).
dαn
1 α+b
(6) F{f (ax)ei bx } =
f
.
|a|
a
(5) F{xn f (x)} = (−i)n
5.3.2. Fourier Transforms of the Derivatives of a Function. We assume
that f (x) is differentiable n times and the function and its derivatives approach zero
as |x| → ∞. Then
f(p) (α) = (−iα)f(p−1) ,
where f(p) is the Fourier transform of f (p) (x), which is the p-th derivative of f (x)
π
(p)
(p)
for 0 ≤ p ≤ n. If lim f (x) = 0, and lim f (x) =
cp , then
x→∞
x→0
2
fc(p) = −cp−1 + αfs(p−1) ,
(5.9)
fs(p) = −αfc(p−1) .
(5.10)
and
5.3.3. Convolution Theorems for Fourier Transform. The convolution
(or Faltung) of f (t) and g(t) over (−∞, ∞) is defined by
∞
∞
1
1
f g = √
f (η)g(x − η) dη = √
f (x − η)g(η) dη.
(5.11)
2π −∞
2π −∞
128
5. INTEGRAL TRANSFORMS
Theorem 5.5. Let f(α) and g(α) be the Fourier transforms of f (x) and g(x),
respectively. Then the inverse Fourier transform of f(α) g (α) is
∞
1
F −1 f(α)
g (α) = √
f (η)g(x − η) dη.
2π −∞
Theorem 5.6. Let f(α) and g(α) be the Fourier transforms of f (x) and g(x),
respectively. Then
∞
∞
f (α)
g (α)dα =
f (−η)g(η) dη.
(5.12)
−∞
−∞
Example 5.8. Find the Fourier transform of f (x) = e−k|x| , k > 0.
∞
1
F{f (x)} = f(α) = √
e−k|x| eixα dx
2π −∞
0
∞
1
=√
ekx eixα dx +
e−kx eixα dx
2π −∞
0
√
1
1
1
k 2
=√
−
= √
.
π(k 2 + α2 )
2π k + iα −k + iα
√
k 2
Note that by Property 4, F{f (x)} = √
yields f(α) = e−k|α| .
π(k 2 + x2 )
Example 5.9. Find the Fourier transform of f (x) = e−kx , k > 0.
∞
2
1
f(α) = √
e−kx eixα dx
2π −∞
∞
2
2
2
2
2
1
=√
e−k(x −ixα/k−α /(4k )+α /(4k )) dx
2π −∞
∞
∞
2
2
2
2
2
1
1
=√
e−k((x−iα/k) −α /(4k )) dx = √
e−α /(4k)
e−u du
2π −∞
2πk
−∞
√
2
2
1
1
=√
e−α /(4k) π = √ e−α /(4k) .
2πk
2k
2
Example 5.10. Find the Fourier transform of
f (x) =
1
F{f (x)} = √
2π
∞
for x < 0,
0
−ax
xe
for x > 0, a > 0.
1
f (x)eiαx dx = √
2π
−∞
∞
0
xe−ax eiαx dx
5.3. FOURIER INTEGRAL THEOREMS
129
∞
xe−ax+iαx ∞
1
=√
e−ax eiαx dx
−√
2π(iα − a) 0
2π(iα − a) 0
∞
1
1
= −√
e−ax eiαx dx = √
.
2π(iα − a) 0
2π(iα − a)2
Example 5.11. Find the Fourier transform of f (x) = 0 if x < b and f (x) =
2 2
e−a x if 0 < b < x. The solution is
∞
∞
2
2
2
2
2
1
1
−a2 x2 ixα
f (α) = √
e
e dx = √
e−a (x−iα/(2a )) −α /(4a ) dx
2π b
2π b
∞
−α2 /(4a2 )
e
1
iα
−u2
−α2 /(4a2 )
√
=
e
du = √ e
erfc ab −
.
2a
a 2π
2a 2
(ab−iα/2a)
Example 5.12. Find the solution of Laplace’s equation uxx + uyy = 0 in the
domain |x| < ∞ and y ≥ 0, subject to the conditions that u → 0 as |x| → ∞ or
as y → ∞, and u(x, 0) = δ(x). After applying the Fourier transform to the given
differential equation with respect to x, we get u
yy − α2 u
= 0, whose solution is
−|α|y
u
= Ae
. Applying the boundary condition at y = 0 in the transform domain,
1
1
we get u
(α, 0) = A = √ . Hence, u
= √ e−|α|y . On inverting, we obtain
2π
2π
u(x, y) =
1
y
.
π x2 + y 2
Now, we use the convolution theorem to obtain the solution to the problem with
arbitrary condition u(x, 0) = f (x). Then the solution is
1 ∞
yf (η)
u(x, y) =
dη,
(5.13)
π −∞ (x − η)2 + y 2
which is known as the Poisson integral representation for the Dirichlet problem in the
half-plane.
Example 5.13. In this example we use both Laplace and Fourier transforms
to solve the boundary value problem
utt − c2 uxx = e−|x| sin t,
u(x, 0) = 0,
ut (x, 0) = e−|x| .
Note that this equation in the Laplace transform domain becomes
s2 u(x, s) − c2 uxx (x, s) =
2 + s2 −|x|
e
.
1 + s2
Applying the complex Fourier transform, we get
2 2
c α +s
2
2
s) = 2 + s
u(x,
1 + s2
2
1
.
π 1 + α2
130
5. INTEGRAL TRANSFORMS
Thus,
2
s) = 2 + s
u(x,
1 + s2
2
1
2
π (1 + α ) (c2 α2 + s2 )
2 + s2
2
1
c2
=
−
.
(1 + s2 ) (s2 − c2 ) π 1 + α2
c2 α2 + s2
On inverting the Fourier transform, we have
2 + s2
c −s|x|/c −|x|
e
−
e
(1 + s2 ) (s2 − c2 )
s
1
1 1
= B
−
−
e−|x|
s−c s+c
(1 + c2 ) (1 + s2 )
2
1
1 cs
+
−B
+
−
cs
s−c s+c
(1 + c2 ) (1 + s2 )
u(x, s) =
where B =
e−s|x|/c ,
2 + c2
. After taking the Laplace inverse, we find that
2c (1 + c2 )
1
u(x, t) = B ect − e−ct −
sin
t
e−|x|
1 + c2
2
c
+ H (ct − |x|)
− B ect−|x| + e−ct+|x| −
cos
(t
−
|x|/c)
.
c
1 + c2
5.4. Fourier Sine and Cosine Transforms
Some important properties and theorems are given below.
5.4.1. Properties of Fourier Sine and Cosine Transforms.
(a) Fc {f (x)} = fc (α),
Fs {f (x)} = fs (α),
(b) Fc {fc (x)} = f (α), Fs {fs (x)} = f (α),
1
α
(c) Fc {f (kx)} = fc ( ), k > 0,
k
k
1 α
(d) Fs {f (kx)} = fs ( ), k > 0,
k
k
1 α+b
α−b (e) Fc {f (kx) cos bx} =
fc
+ fc
, k > 0,
2k
k
k
1 α+b
α−b (f) Fc {f (kx) sin bx} =
fs
− fs
, k > 0,
2k
k
k
1 α+b
α−b (g) Fs {f (kx) cos bx} =
fs
+ fs
, k > 0,
2k
k
k
1 α−b
α+b (h) Fs {f (kx) sin bx} =
fc
− fc
, k > 0,
2k
k
k
5.4. FOURIER SINE AND COSINE TRANSFORMS
131
d2n fc (α)
,
dα2n
d2n+1 fs (α)
(j) Fc {x2n+1 f (x)} = (−1)n
,
dα2n+1
d2n fs (α)
(k) Fs {x2n f (x)} = (−1)n
,
dα2n
d2n+1 fc (α)
(l) Fs {x2n+1 f (x)} = (−1)n+1
.
dα2n+1
5.4.2. Convolution Theorems for Fourier Sine and Cosine Transforms. These theorems are as follows:
Theorem 5.7. Let fc (α) and gc (α) be the Fourier cosine transforms
of f (x) and g(x), respectively, and let fs (α) and gs (α) be the Fourier sine
transforms of f (x) and g(x), respectively. Then
∞
1
Fc−1 fc (α) gc (α) = √
g(s) [f (|x − s|) + f (x + s)] ds.
(5.14)
2π 0
Theorem 5.8. If fs (α), fc (α), and gs (α), gs (α) are the Fourier sine and
cosine transforms of f (x) and g(x), respectively, then the following results
hold:
∞
1 ∞
(i)
fc (α) gs (α) sin αx dα =
g(s)[f (|x − s|) − f (x + s)] ds,
(5.15)
2 0
0
∞
1 ∞
(ii)
fs (α) gc (α) sin αx dα =
f (s)[g(|x − s|) − g(x + s)] ds,
(5.16)
2 0
0
∞
(iii)
fs (α) gs (α) cos αx dα
(i) Fc {x2n f (x)} = (−1)n
0
1
=
2
1
=
2
(iv)
∞
0
∞
g(t) H(t + x) f (t + x) + H(t − x) f (t − x) dt
∞
f (t) H(t + x) g(t + x) + H(t − x) g(t − x) dt,
0
0
fc (α) gc (α) dα =
∞
∞
f (s)g(s) ds =
0
fs (α)
gs (α) dα.
(5.17)
(5.18)
0
Proofs of all above theorems can be found in Sneddon [1978] or Kythe et al. [2003].
Example 5.14. Find the Fourier cosine transform of f (x), where


 x for 0 < x < 1,
f (x) = 2 − x for 1 < x < 2,


0 for 2 < x < ∞.
1
2
2
Here Fc [f (x)] =
x cos x dx +
(2 − x) cos x dx
π 0
1
132
5. INTEGRAL TRANSFORMS
2 x sin αx 1 1 1
(2 − x) sin αx 2 1 2
=
sin αx dx +
sin αx dx
−
+
π
α
α 0
α
α 1
0
1
2 2 cos α − 1 − cos 2α
=
.
π
α2
Example 5.15. We will use the Fourier transform to solve Example 5.6, which
was earlier solved by the Laplace transform. The partial differential equation and the
boundary and initial conditions are:
k uxx = ut ,
u=0
for t ≤ 0;
u → 0 as x → ∞;
and
u = T0 at x = 0 for t > 0.
Applying the Fourier sine transform to the partial differential equation, we get
∂
us
2
2
+ kα u
s =
k α T0 ,
∂t
π
where u
s is the Fourier sine transform of u. Its solution is given by
2 T0
−kα2 t
u
s = Ae
+
.
π α
2 T0
After using the initial condition at t = 0, we get A +
= 0. This yields
π
α
2 T0
2
u
s =
1 − e−kα t . Thus,
π α
∞
2
2
1
u(x, t) = T0
1 − e−kα t sin αx dα
π
α
0
∞
∞
2
sin αx
sin αx −kα2 t
= T0
dα −
e
dα
π
α
α
0
0
2
π π
x
x
= T0
− erf √
= T0 erfc √ .
π
2
2
2 kt
2 kt
5.5. Finite Fourier Transforms
When the domain of the physical problem is finite, it is generally not convenient to
use the transforms with an infinite range of integration. In many cases finite Fourier
transform can be used with advantage. We define
a
nπx Fs {f (x)} = fs (n) =
f (x) sin
dx
(5.19)
a
0
as the finite Fourier sine transform of f (x) for n = 1, 2, . . . . The function f (x) is
then given by
∞
2
nπx Fs−1 {fs (n)} = f (x) =
fs (n) sin
.
(5.20)
a n=1
a
5.5. FINITE FOURIER TRANSFORMS
133
Similarly, the finite Fourier cosine transform is defined for n = 0, 1, 2, . . . by
Fc {f (x)} = fc (n) =
a
f (x) cos
0
nπx dx,
a
(5.21)
and its inverse by
∞
fc (0) 2 nπx Fc−1 {fs (n)} = f (x) =
+
fc (n) cos
.
a
a n=1
a
(5.22)
From the theory of Fourier series it is known that the Fourier cosine series for
∞
1
nπx
f (x) on the interval 0 < x < a, which is f (0) +
an cos
, converges to the
a
a
n=1
1
value f (x) at each point of continuity in the above interval to the value
f (x + 0) +
2
f (x − 0) at each point of x of finite discontinuity (and also at all points of continuity)
in this interval. Similarly, the Fourier sine series for a function f (x) on the interval
∞
2
nπx
0 < x < 0, which is
bn sin
, converges to the value f (x) at each point of
a n=1
a
1
continuity in the above interval to the value
f (x + 0) + f (x − 0) at each point of
2
x of finite discontinuity (and also at all points of continuity) in this interval. Thus, we
can alternatively define the finite Fourier sine and cosine transforms fc and fc by an
and bn , respectively, as is done in Weinberger [1965]. An example of this approach
is given in §6.8, 8.7, 8.8.1 and 8.11. The finite Fourier sine transform is the general
choice because of its simplicity of use.
πx
For a = π set
= z. Then the above definitions become
a
a
nπx a π
az a
ax fs (n) =
f (x) sin
dx =
sin(nz)f
dz = Fs f
,
a
π 0
π
π
π
0
a
nπx a π
az a
ax fc (n) =
f (x) cos
dx =
cos(nz)f
dz = Fc f
.
a
π 0
π
π
π
0
Example 5.16. (a) For f (x) = 1 we have
a
nπx
a
Fs {1} = fs (n) =
sin
dx =
[1 − (−1)n ]; and
a
nπ
0
a
a, n = 0,
nπx
Fc {1} = fc (n) =
cos
dx =
.
a
0, n = 0.
0
(b) For f (x) = x, we have
a
(−1)n+1 a2
nπx
Fs {x} = fs (n) =
x sin
dx =
; and
a
nπ
0
134
5. INTEGRAL TRANSFORMS
 2
a


, n = 0,
a
nπx
2
Fc {x} = fc (n) = 0 x cos
dx =
.

a
 a 2 [1 − (−1)n ] , n = 0.
nπ
5.5.1. Properties. The following properties are useful:
nπ (i) Fs {f (x)} = −
fc (n),
a
nπ 2 nπ (ii) Fs {f (x)} = −
fs (n) +
f (0) + (−1)n+1 f (a) ,
a
a
nπ n
(iii) Fc {f (x)} =
fs (n) + (−1) f (a) − f (0), and
a
nπ 2 (iv) Fc {f (x)} = −
fc (n) + (−1)n f (a) − f (0).
a
5.5.2. Periodic Extensions. A function g(x) is said to be an odd periodic
extension of a periodic function f (x) of period 2π if
g(x) =
f (x) for 0 < x < π,
−f (−x) for −π < x < 0.
That is, g(x) = f (x) for 0 < x < π, and g(−x) = −g(x) and g(x + 2π) = g(x)
for −∞ < x < ∞.
A function h(x) is said to be an even periodic extension of a periodic function
f (x) of period 2π if
h(x) =
f (x)
for 0 < x < π,
f (−x) for −π < x < 0.
That is, h(x) = f (x) for 0 < x < π, and h(−x) = h(x) and g(x + 2π) = h(x) for
−∞ < x < ∞.
A theorem on odd and even extensions of a periodic function f is given below
without proof (Chruchill [1972]).
Theorem 5.9. (i) If g(x) is the odd periodic extension of f (x) of period
2π, then for any constant α
Fs {g(x − α) + g(x + α)} = 2 cos nα Fs {f (x)},
(5.23)
Fc {g(x + α) + g(x − α)} = 2 sin nα Fs {f (x)}.
(5.24)
(ii) If h(x) is an even periodic extension of f (x) of period 2π, then for any
constant α
Fc c{h(x−α)+h(x+α)} = 2 cos nα Fc {f (x)},
(5.25)
Fc {h(x − α) + h(x + α)} = 2 sin nα Fc {f (x)}.
(5.26)
5.5. FINITE FOURIER TRANSFORMS
135
5.5.3. Convolution. Let f (x) and g(x) be two piecewise continuous periodic
functions defined on −π < x < π. Then their convolution is defined by
π
f (x) g(x) =
f (x − u) g(u) du,
(5.27)
−π
which is a continuous periodic function of period 2π. The following result hold
(Churchill [1972]):
Theorem 5.10. Let f1 (x) and g1 (x) be the odd periodic extensions of
f (x) and g(x), respectively on 0 < x < π, and f2 (x) and g2 (x) their even
periodic extensions on 0 < x < π. Then
Fs {f1 (x) g2 (x)} = 2fs (n) gc (n),
Fs {f2 (x) g1 (x)} = 2fc (n) gs (n),
Fc {f1 (x) g1 (x)} = −2fs (n) gs (n), Fc {f2 (x) g2 (x)} = 2fc (n) gc (n),
or inversely,
Fs−1 {fs (n) gc (n)} =
1
{f1 (x) g2 (x)},
2
Fs−1 {fc (n) gs (n)} = 2{f2 (x) g1 (x)},
1
Fs−1 {fs (n) gs (n)} = − {f1 (x) g1 (x)},
2
1
Fc−1 {fc (n) gc (n)} = {f2 (x) g2 (x)}.
2
−1
Example 5.17. To determine F −1 n2 − a2
, n = 0, let
(5.28)
1
=
n2 − a2
n(−1)n+1
(−1)n+1
fs (n) gs (n), where fs (n) = 2
and gs (n) =
, which gives f (x) =
2
n −a
n
sin ax
x
, and g(x) = , so that
sin aπ
π
1
sin ax
x
= fs (n)
gs (n) = Fs
Fs
2
2
n −a
sin aπ
π
1
1
−1
−1 =F
=F
fs (n) gs (n) = − f1 (x) g1 (x),
2
2
n −a
2
x
where f1 (x) is the periodic extension of the odd function f (x) and g1 (x) = . Thus,
π
using (5.28)
π
1
1
F −1 2
=
−
f1 (x − u) g1 (u) du
n − a2
2π −π
π
1
cos[a(π − x)]
=−
f1 (x − u) u du = −
.
2π −π
a sin aπ
In the case of a function of two variables, f (x, y), 0 < x < a, 0 < y < b, these
finite transforms are denoted by fs (n, y) and fc (n, y), respectively, as the following
example shows.
136
5. INTEGRAL TRANSFORMS
Example
5.18. Consider Laplace’s equation in the rectangle 0 < x < a, 0 <
y<b
(5.29)
uxx + uyy = 0,
with the boundary conditions u(0, y) = u(a, y) = u(x, b) = 0, and u(x, 0) = f (x).
After applying the finite Fourier sine transform to u(x, y) with respect to x from 0 to
a, we have
(
us )xx (n) =
a
uxx sin
0
nπx
dx
a
nπ
n2 π 2
=
[u(0, y) − (−1)n u(a, y)] − 2 u
s (n, y).
a
a
2
d
n2 π 2
Then Equation (5.29) becomes
− 2
u
s (n, y) = 0. Solving for u
s (n, y),
dy 2
a
we get u
s (n, y) = Aenπy/a
+ Be−nπy/a . Since u
s (n,
nπ(y−b)/a
b) = 0, we can express
u
s (n, y) as u
s (n, y) = An e
− e−nπ(y−b)/a . After applying the boundary condition at y = 0, we get
An e−nπb/a − enπb/a = f s (n),
which, after solving for An and substituting its value into u
s (n, y), yields
u
s (n, y) = −
sinh[nπ(y − b)/a]
f s (n).
sinh(nπb/a)
Hence,
∞
u(x, y) =
2 sinh[nπ(b − y)/a]
f s (n) sin(nπx/a),
a 1
sinh(nπb/a)
a
where f s (n) =
f (t) sin
0
nπt
dt.
a
Other examples are available in §6.8.
5.6. Multiple Transforms
Multiple transforms used in this book are defined below.
5.6.1. Multiple Laplace Transform. Double Laplace transform is defined by
∞
L{f (x, t)} = F (κ, s) = f (κ, s) =
0
c+i∞
c−i∞
f (x, t) e−(κ+s)t dκ dt,
(5.30)
5.7. HANKEL TRANSFORMS
137
and its inverse by
L
−1
1
{F (κ, s)} = f (x, t) =
=
2π i
∞
0
c+i∞
F (κ, s) e(κ+s)t dt dκ,
(5.31)
c−i∞
where {z} > 0, {κ} > 0 and c > 0. For an application see §6.9.
5.6.2. Multiple Fourier Transform. Let x = (x1 , . . . , xn ) ∈ Rn . Then
under the same assumptions on f (x) as on f (x) ∈ R, the multi-dimensional Fourier
transform of f (x) is defined by
∞
∞
1
F{f (x) = f(ςς ) =
·
·
·
exp{−i (ςς ·x)} f (x) dx,
(5.32)
(2π)n/2 −∞
−∞
where ς = (ς1 , . . . , ςn ) denotes the n-dimensional transform variable and ς · x =
(ς1 x1 + · · · + ςn xn ).
The inverse Fourier transform is defined by
∞
∞
1
−1 ς
F {f (ς ) = f (x) =
···
exp{−i (ςς ·x)} f (x) dx, (5.33)
(2π)n/2 −∞
−∞
In 2-D and 3-D cases we will use ς = (α, β) and ς = (α, β, γ), respectively.
Thus, the double Fourier transform is defined by
∞ ∞
1
F{f (x, y)} =
e−i (αx+βy) f (x, y) dx dy,
(5.34)
2π −∞ −∞
and its inverse by
1
F −1 {f(α, β) =
2π
∞
−∞
···
∞
e−i (αx+βy) f (x, y) dα dβ,
−∞
and the 3-D Fourier transform by
∞ ∞
1
F{f (x, y, z)} =
e−i (αx+βy+γz) f (x, y, z) dx dy dz,
2π −∞ −∞
and its inverse by
F
−1
1
{f(α, β, γ)} =
2π
∞
−∞
···
∞
−∞
(5.35)
(5.36)
e−i (αx+βy+γz) f (x, y, z) dα dβ dγ, (5.37)
Example 5.19. For f (x, y, z) = δ(x)δ(y)δ(z) the 3-D Fourier transform is
f (ζ) = (2π)−3/2 .
Other examples are available in §6.3 and §7.10.1.
5.7. Hankel Transforms
The nth order Hankel transform of a function f (x) is defined by
∞
Hn {f (x)} = fn (σ) =
xJn (σx)f (x) dx,
0
(5.38)
138
5. INTEGRAL TRANSFORMS
where Jn (σx) is the Bessel function of order n, and the integral on the right side is
convergent. The inverse Hankel transform of order n is given by
Hn−1 {fn (σ)}
∞
= f (x) =
σJn (σx)fn (σ) dσ,
(5.39)
0
provided the integral exists.
Some basic properties of this transform are:
(i) (Scaling) If Hn {f (x)} = fn (σ), then Hn {f (ax)} =
1 σ
fn
.
a2
a
(ii) (Parseval’s Relation) If Hn {f (x)} = fn (σ) and if Hn {g(x)} = gn (σ), then
∞
∞
xf (x) g(x) dx =
σ fn (σ) gn (σ) dσ.
(5.40)
0
0
(iii) (Transform of derivatives) If Hn {f (x)} = f(σ), then
σ Hn {f (x)} =
(n − 1)fn+1 (σ) − (n + 1)fn−1 (σ) ,
2n
H1 {f (x)} = −σ f0 (σ),
n ≥ 1.
(5.41)
(5.42)
provided xf (x) → 0 as x → 0 and x → ∞.
(iv) If Hn {f (x)} = f(σ), then
Hn {f (x)} = Hn
1 d
x dx
2
df
n
x
− 2 f (x)
dx
x
= −σ 2 fn (σ),
(5.43)
provided both xf (x) and xf (x) vanish as x → 0 and x → ∞.
(v) The zero-order Hankel transform (n = 0) is defined by
∞
H0 {f (x)} = f (x) =
xJ0 (σx)f (x) dx.
0
Example 5.20. The following results hold:
(a) H0
δ(x)
x
= 1;
(b) H0 {H(a − x)} =
(c) H0
e−ax
x
=
(d) H1 {e−ax } =
e−ax
x
xJ0 (σx) dx =
0
a
a
J1 (aσ);
σ
e−ax J0 (σx) dx = √
x e−ax J0 (σx) dx =
a2
1
;
+ σ2
σ
;
3/2
(a2 + σ 2 )
a
−1/2 1 =
e−ax J1 (σx) dx =
1 − a a2 + σ 2
;
σ
0
0
(e) H1
0
a
a
(5.44)
5.9. EXERCISES
(f) Hn xn H(a − x) =
2
(g) Hn xn a−ax =
a
xn+1 Jn (σx) dx =
0
∞
an+1
Jn+1 (aσ);
σ
xn+1 Jn (σx) e−ax dx =
2
0
139
2
σn
e−σ /(4a) .
n+1
(2a)
Other examples are available in §6.7, §7.4 and §7.10.3.
5.8. Summary: Variables of Transforms
Laplace transform: L{f (t)} = f (s); variable of transform: s.
Double Laplace transform: L{f (x, t)} = f (κ, s); variable of transform: κ, s
Fourier transform: F{f (x)} = f(α); variable of transform: α.
Fourier sine transform: Fs {f (x)} = fs (α); variable of transform: α.
Fourier cosine transform: Fc {f (x)} = fc (α); variable of transform: α.
Finite Fourier transforms: n.
Fourier 2-D transform: F{f (x)} = f(ς); variable of transform: ς = (α, β).
Fourier 3-D transform: F{f (x)} = f(ζ); variable of transform: ζ = (α, β, γ).
Zero-order Hankel transform: H0 {f (x)} = f(σ); variable of transform: σ.
5.9. Exercises
5.1. Using the properties of §5.2 and L eat =
1
, derive the Laplace transform
s−a
of sin at, cos at, ebt sin at, ebt cos at, tn eat , sinh bt, and cosh bt.
√
2
a
5.2. Show that L−1 e−a s = √
e−a /4t .
2 πt3
√
1
5.3. Show that (a) L−1 s−3/2 e−1/s = √ sin(2 t); and
π
√
1
(b) L−1 s−1/2 e−1/s = √ cos(2 t).
πt
√
√
cosh a s
−1
−1 sinh a s
√ , and (b) L
√ , b > a > 0.
5.4. Find (a) L
s cosh b s
sinh
b s Hint. Use cosh x = ex + e−x /2, sinh x = ex − e−x /2, and (1 + z)−1 =
∞
(−1)n z n .
n=0
Ans. (a)
∞
n=0
(−1)n
erfc
(2n + 1)b − a (2n + 1)b + a √
√
+ erfc
,
2 t
2 t
140
5. INTEGRAL TRANSFORMS
(b)
∞ (2n + 1)b − a −[(2n+1)b−a]2 /(4t) (2n + 1)b + a −[(2n+1)b+a]2 /(4t)
√
√
e
−
e
.
4πt3
4πt3
n=0
5.5. Using the Laplace transform method, solve the partial differential equation
(1 − x)2
utt = uxx , with the initial conditions u(x, 0) = −
, ut (x, 0) = 0, and
2
the boundary conditions ux (0, t) = 1 and ux (1, t) = 0.
1
(1 − x)2
Ans. u = − t2 −
.
2
2
5.6. Using the Laplace transform method, solve the partial differential equation
ut = uxx , with the initial condition u(x, 0) = 0 and the boundary conditions
ux (0, t) = 0 and u(1, t) = 1.
√
cosh x s
√ .
Hint. The solution in the transform domain is ū =
s cosh s
∞
2n + 1 − x
2n + 1 + x
n
√
√
Ans. u =
(−1) erfc
+ erfc
, or
2 t
2 t
0
∞
4 cos(2n + 1)πx/2 −(2n+1)2 π2 t/4
u=1−
(−1)n
e
.
(2n + 1)π
0
5.7. Use Laplace transform to solve the wave equation utt = c2 urr , subject
to the initial conditions u(r, 0) = ut (r, 0) = 0 and the boundary conditions
∂u(ρ, t) = f (t) and u → 0 as r → ∞.
∂ρ
ρ=a
t
F (s)
c
Hint. Use L−1
= 0 e−k(t−x) f (x) dx ≡ φ(t), k = . Then
s+k
a
ac F (s) −s(ρ−a)/c
u = L−1 −
e
, then use Property (ii). Ans. u(r, t) =
s + c/a
ρ
ca
ρ−a
ρ−a
− H t−
φ t−
.
ρ
c
c
Γ(p + 1)
5.8. Show that L{tp } =
, where Γ(x) is the gamma function defined by
sp+1
∞
∞
Γ(x) =
tx−1 e−t dt, x > 0. Hint. Set st = x in L{tp } = 0 tp e−st dt.
∞
0
1
∞
sp+1
0
Γ(p + 1)
.
sp+1
5.9. Find the complex Fourier transform of the following functions:
Then
0
tp e−st dt =
(a) f (x) =
!
(b) f (x) =
0
e−ax
if x < 0
if x > 0.
xp e−x dx =
Ans. √
1
.
2π(a − iα)
0
if x < 0,
1 −ax
−bx
[e
− e ] if x > 0 and b > a > 0.
x
5.9. EXERCISES
141
1
b − iα
Hint. Use part (a) and integrate with respect to a. Ans. √ ln
.
2π a − iα

if |x| > a,
0
2 1
(c) f (x) =
Ans.
(1 − cos αa).
 1 − |x| if |x| < a.
π aα2
a
(d) f (x) = cos ax2 and f (x) = sin ax2 .
Hint: Use Example 5.9 and define k as ia.
1
α2
α2
1
α2
α2
Ans. √
sin
+ cos
and √
sin
− cos
.
2 a
4a
4a
2 a
4a
4a
√
1 if |x| < 1
2
(e) f (x) =
Ans. √ sin α.
α π
0 if |x| > 1.
sin kx if |x| < 1
i
sin(k − α) sin(k + α)
(f) f (x) =
Ans. √
−
.
k−α
k+α
0
if |x| > 1.
2π

 √1
if x > 0
x
(g) f (x) =
.

= 0 if x < 0
eiπ/4
Hint: Substitute αx = v 2 eiπ/2 . Ans. √
sgn α.
2α

 1
if x < 0
e−iπ/4
|x|
(h) f (x) =
Ans. √
sgn α.

2α
0
if x > 0.
1
1
(i) f (x) = . Ans. .
|x|
|α|
1
2
Ans. √ e−α /(4k) .
2k
for x < 0,
1
Ans. √
.
for x > 0, a > 0.
2π(iα − a)2
(j) f (x) = e−kx , k > 0.
2
(k) f (x) =
0
xe−ax
5.10. Prove the following properties of the Fourier convolution:
d (i)
f (x) g(x) = f (x) g(x) = f (x) g (x),
dx
d2 (i) 2 f (x) g(x) = f (x) g (x) = f (x) g(x),
dx
(m+n)
(iii) (f (x) g(x))
= f (m) (x) g (n) (x) ,
∞
∞
∞
(iv)
f (x) g(x) dx =
f (u) du g(v) dv,
−∞
−∞
−∞
1
(v) If g(x) =
H(a − x), then f (x) g(x) is the average value
2a
142
5. INTEGRAL TRANSFORMS
of f (x) in [x − a, x + a],
1
2
(vi) If gt (x) =
e−x /(4kt) , then gt (x) gs (x) = gt+s (x).
4π kt
5.11. Compute the Fourier cosine transform of f (t), where


 x if 0 < x < 1,
4
2
f (x) = 2 − x if 1 < x < 2, Ans. f(α) = 2
cos(α) sin2 (α/2).

α
π

0 if x ≥ 2.
5.12. Solve the partial differential equation utt −c2 uxx = 0, subject to the conditions
u(x, 0) = f (x) + g(x) and ut (x, 0) = c (f (x) − g (x)), where u(x, t), f (x),
g(x), u (x, t), f (x), and g (x) all go to zero as |x| → ∞.
Ans. u = f (x + ct) + g(x − ct).
1
5.13. Find the Fourier sine and cosine transforms of f (x) = √ e−ax .
x
√
√
a2 + α2 + a
a2 + α2 − a
√
√
Ans. Fc f (x) =
, Fs f (x) =
.
a2 + α2
a2 + α2
1 −ax
5.14. Find the Fourier cosine transform of e
− e−bx , {a}, {b} > 0.
x
2
1
b + α2
Ans. √ ln
.
a2 + α2
2π
1 −ax
2
α
5.15. Find the Fourier sine transform of e
, {a} > 0. Ans.
tan−1 .
x
π
a
√
5.16. Find Fourier sine and cosine transforms of f (x) = xe−ax , a > 0.
1
Hint: fc (α)+ifs (α) = ; then express (a−iα) in the polar form.
2(a − iα)3
−1
e(3i/2) tan (α/a)
Ans. fc (α) + ifc (α) = √
.
2(a2 + α2 )3/4
5.17. Find the Fourier sine transforms of the following functions: (i) f (x) =
x e−ax , a > 0; (ii) f (x) = e−ax cos x, a > 0.
∞
a
Hint. Differentiate both sides of the integral 0 e−ax sin kx dx = 2
a
+
α2
a
2aα
with respect to a. Ans. (ii) fs (α) =
.
π (a2 + α2 )2
5.18. Show that H0 {(a2 − x2 )H(a − x)} =
4a
2a2
J1 (aσ) − 2 J0 (aσ).
3
σ
σ
6
Parabolic Equations
Diffusion is one of the basic processes by which material moves. It is a result of
continuous motion of atoms, molecules and particles moving from regions of high to
low concentration. Besides, there are other processes that produce inhomogeneity,
and the combined result of diffusion and other processes is generally not precisely
defined. Diffusion depends on three main parameters: (i) temperature, (ii) mass (size)
of the diffusing particles and (iii) viscosity of the system. While the temperature is a
measure of the average kinetic energy, which is due to the movement of particles in
the system, in most systems the energy is equal to a constant times the temperature,
and a higher kinetic energy means a higher viscosity, yet the speed of diffusion always
increases with temperature. Since heavy particles have a lower velocity for a given
kinetic energy, they diffuse more slowly than lighter, smaller ones. Diffusion is most
rapid in a gas, slower in a liquid, and very slow or almost zero in a solid.
The flow of a quantity, such as heat, is described by the diffusion equation, where
the total amount of the quantity does not change. Thus, both heat and mass are
conserved, although their distributions may undergo slight changes. The first-order
time derivative in the diffusion equation signifies a complete determination of the
time development of the solution subject to prescribed initial temperature or mass
distribution, and presence of any sources or sinks. It also means the irreversibility
of the equation in the sense that the solution is always in the direction of time; thus,
a change in the sign of time also changes the behavior of the solution. Green’s
functions and fundamental solutions for the parabolic operator play an important role
in problems of heat conduction and neutron diffusion.
∂
In this chapter we will derive Green’s functions for the transient operator L D,
,
∂t
3
+
defined in the space R × R of the independent variables (x, t) = (x, y, z, t), where
L(D) is the 3-D Laplacian operator in the space variables. This transient operator is
also known as the 3-D diffusion operator or the heat conduction operator. It is often
143
144
6. PARABOLIC EQUATIONS
∂
denoted simply by
− k ∇2 , where k denotes the thermal diffusivity. The problem
∂t
of finding Green’s functions for these operators is solved by different methods, but
mostly by the integral transform method.
6.1. 1-D Diffusion Equation
We will consider boundary value problem for 1-D diffusion equation, which are solved
by two different methods.
6.1.1. Sturm-Liouville System for 1-D Diffusion Equation. Consider
the 1-D diffusion system
∂u
∂
∂u
=
p(x)
+ q(x)u + F (x, t), a ≤ x ≤ b, t > 0;
∂t
∂x
∂x
a1 u(a, t) + a2 ux (a, t) = 0, b1 u(b, t) + b2 ux (b, t) = 0,
u(x, 0) = f (x),
(6.1)
a < x < b,
where p, q, F are continuous functions on the interval I = [a, b], and a1,2 and b1,2
are real positive constants such that a21 + a22 > 0 and b21 + b22 > 0. Using the
Sturm-Liouville operator L, Eq (6.1) becomes
ut = L[u] + F.
(6.2)
First, we take F = 0 and use Bernoulli’s separation method: Let the solution of Eq
(6.2) be of the form u(x, t) = X(x) T (t) = 0. Then this equation reduces to
L[X] = λX, a ≤ x ≤ b,
dT
= λT, t > 0,
dt
(6.3)
where λ is the separation constant. The boundary conditions and the initial condition
in (6.1) become
a1 X1 (a) + a2 X (a) = 0, b1 X1 (b) + b1 X (b) = 0,
T (0) = f (x),
a < x < b.
(6.4)
The system (6.3)–(6.4) is called the associated Sturm-Liouville system for the
1-D diffusion equation. The solution u(x, t) of this system can be found in terms
of the eigenpair (λn , Xn (x)) as
u(x, t) =
∞
n=1
Xn (x) Tn (t),
(6.5)
6.1. 1-D DIFFUSION EQUATION
145
where the functions Xn (t) are determined as follows: Assume that the forcing term
F (x, t) is expandable in terms of the eigenfunctions as
F (x, t) =
∞
(6.6)
Xn (t) fn (t),
n=1
where fn (t) are Fourier coefficients given by
b
fn (t) =
F (x, t) Xn (x) dx.
(6.7)
a
Substituting (6.5) and (6.6) into Eq (6.2), we get
∞
∞
∞
Ṫn (t)Xn (t) = L
Tn (t)Xn (x) +
fn (t)Xn (x)
n=1
n=1
=
∞
n=1
[λn Tn (t) + fn (t)] Xn (x),
n=1
which yields the ordinary differential equation
Ṫn (t) − λn Tn (t) = fn (t),
where the dot denotes derivative with respect to t. To solve this equation, we apply
Laplace transform, which after inverting gives the solution as
t
λn t
Tn (t) = Xn (0) e
+
eλn (t−τ ) fn (τ ) dτ,
(6.8)
0
where u(x, 0) = f (x) =
∞
Xn (x)Xn (0) =
n=1
∞
Xn (x)
n=1
b
Xn (x )f (x ) dx .
a
Substituting (6.8) into (6.5) we have
t
∞ b
λn t
λn (t−τ )
u(x, t) =
Xn (x )f (x ) dx e
+
e
fn (τ ) dτ Xn (x)
=
a
n=1
b
∞
0
Xn (x) Xn (x ) eλn t f (x ) dx
a
n=1
t
+
0
∞ b
Xn (x)Xn (x ) eλn (t−τ ) F (x , τ ) dτ,
(6.9)
a n=1
where the integration and summation are interchanged. If we define Green’s function
for the operator L by
G(x, x ; t) =
∞
n=1
Xn (x) Xn (x ) eλn t ,
(6.10)
146
6. PARABOLIC EQUATIONS
then the solution (6.9) can be written as
b
u(x, t) =
t
b
G(x, x ; t)f (x ) dx +
0
a
G(x, x ; t−τ )F (x , τ ) dx dτ.
(6.11)
a
6.1.2. Green’s Function for 1-D Diffusion Equation. Consider the
boundary value problem defined by the nonhomogeneous 1-D diffusion equation.
ut − k uxx = f (x),
x ∈ R, t > 0,
(6.12)
subject to the boundary and the initial conditions1
u(x, t) → 0 as |x| → ∞,
and u(x, 0) = 0 for x ∈ R.
(6.13)
Note that the diffusion operator is not self-adjoint in (−∞, ∞) × (0, ∞). However,
in the Laplace transform domain it is self-adjoint. The equation and the boundary
and initial conditions to be satisfied by Green’s function G(x, x ; t, t ) are
∂G
∂ 2G
−k
= δ(x − x ) δ(t − t ),
∂t
∂x2
G(x, x ; 0, t ) = 0,
lim G(x, x ; t, t ) = 0.
|x|→∞
(6.14a)
(6.14b)
Applying Laplace transform to Eq (6.14a), we have
sG − k
d2 G
= δ(x − x ) e−st ,
2
dx
where the boundary condition (6.14b) becomes lim G(x, x ; s, t ) = 0. By prop|x|→∞
erty (iv) §2.3.2, the discontinuity condition at x is
dG(x+ , x ; s, t ) dG(x− , x ; s, t )
1
−
= − e−st .
dx
dx
k
(6.15)
The solution in the transform domain is
G(x, s) =
A1 ecx + B1 e−cx
if x < x ,
−cx
if x > x ,
cx
A2 e
+ B2 e
where c = s/k and s is the variable of the transform. The condition at x → −∞
yields B1 = 0 and at x → ∞ yields A2 = 0. The continuity of G(x, s) at x
1
If the variable t denotes time in the defining equation, then an initial condition must
be prescribed.
6.1. 1-D DIFFUSION EQUATION
147
gives A1 ecx = B2 e−cx , and the jump condition (6.15) on Gx (x, s) at x gives
1 −st
B2 e−cx + A1 ecx =
e
; thus,
kc
1 −st −cx
1 −st +cx
A1 =
e
, B2 =
e
.
2kc
2kc
Hence,

1 −st −c(x −x)


e
, x < x ,
2kc
G(x, s) =

 1 e−st −c(x−x ) , x > x ,
2kc
1 −st −c|x−x |
that is, for all x we have G(x, s) =
e
, which on inversion gives
2kc
G(x, x ; t, t ) =
2
H(t − t )
e−(x−x ) /(4k(t−t )) .
2 πk(t − t )
(6.16)
This Green’s function is also the fundamental solution for the diffusion operator.
Method 2. Apply Fourier transform followed by Laplace transform to Eq
(6.14a). This gives
+ k α2 G
= − √1 e−st +iαx ,
sG
2π
where the tilde and the bar over G denote its Fourier and Laplace transforms, respec is
tively. The solution for G
=√
G
1
e−st +iαx .
2
2π (s + kα )
Inversion of Laplace transform yields
− t ) −kα2 (t−t )+iαx
α) = H(t
√
G(x,
e
,
2π
which, after inverting Fourier transform, gives the Green’s function G which is the
same as (6.16).
Method 3. If we apply Laplace transform followed by Fourier transform to the
problem (6.12), we get
f(α)
u(α, s) =
,
s + kα2
s) → 0 as |α| → ∞. Then first inverting Laplace transform we have
such that u(α,
2
u
(α, t) = f(α) e−kα t .
148
6. PARABOLIC EQUATIONS
Note that if g(α) = e−kα t , then by inversion of Fourier transform we have g(x) =
2
1
√
e−x /(4kt)) . Then applying inverse Fourier transform we find that
2πkt
2
∞
2
1
u(x, t) = √
ei αx f(α) e−kα t dα
2π −∞
∞
∞
1
1
−(x−ξ)4 /(4kt)
= √
f (ξ) e
dξ =
f(α) G(x, t; ξ) dξ,
2π −∞
2 kπt −∞
where Green’s function G(x − x , t) is given by
2
1
G(x, t; x ) = √
e−(x−x ) /(4kt) ,
2 kπt
which is the same as (6.16).
The graphs for this Green’s function G(x, t) for some values of t > 0 are shown
in Fig. 6.1.
G
t = 0.25
0.5
t = 0.5
0.4
t=1
t=4
0.3
t = 10
0.2
0.1
-3
-2
-1
0
1
2
3
x
Fig. 6.1. Green’s Function for 1-D Diffusion Operator.
6.2. 2-D Diffusion Equation
∂
− k ∇2 in R2 , t > 0, we consider the
∂t
function u(x; t) which satisfies the following equation and the boundary conditions:
To find Green’s function for the operator
ut − k∇2 u = f (x, y) δ(t),
u(x, y, 0) = 0
−∞ < x, y < ∞, t > 0,
for all (x, y) ∈ R2 ,
u(x, y, t) → 0 as r =
x2 + y 2 → ∞.
(6.17)
6.2. 2-D DIFFUSION EQUATION
149
Applying Laplace transform, followed by a double Fourier transform to the above
system, we get
f(α, β)
u(α, s) =
, γ = (α, β),
s+kγ
which after the Laplace transform inversion gives
2
u
(α, β) = f(α, β) e−kγ t ≡ f(α, β) g (α, β),
where g(α, β) = e−kγ t , which after inverting yields
2
2
1 −(x2 +y2 )/(4kt)
g(x, y) = F −1 e−kγ t =
e
.
2kt
(6.18)
Then, by using the convolution theorem for Fourier transform (§5.3.3), we get
∞ ∞
1
(x − ξ)2 + (y − η)2
u(x, y, t) =
f (ξ, η) exp −
dξ dη
4πkt −∞ −∞
4kt
∞ ∞
=
f (ξ) G(r, ζ) dξ,
−∞
−∞
where r = |x − y|, ζ = |ξ − η|, and
G(r, ζ) =
H(t) −|r−ζ|2 /(4kt)
e
,
4πkt
(6.19)
which is the required Green’s function (and the fundamental solution) for the 2-D
diffusion operator.
6.2.1. Dirichlet Problem for the General Parabolic Equation in a
Square. Consider
y2
∂ 2u
∂2u
∂2u
∂u
∂u
− 2xy
+ x2 2 − x
−y
− u = 0,
2
∂x
∂x ∂y
∂y
∂x
∂y
(6.20)
subject to the boundary conditions u(x, ±1) = x2 , u(±1, y) = y 2 , on the square
bounded by the lines x = ±1, y = ±1 (Fig. 6.2). In view of $ 4.2 and Theorem 4.1,
Eq (6.20) is parabolic since a212 − a11 a22 = 0, and its characteristics are given by
dy
x
y 2 (dy)2 + 2xy dx dy + x2 (dx)2 = 0, with the solution
= − , or x2 + y 2 = c,
dx
y
which is a single parameter family of concentric circles with center at the origin.
If we transform Eq (6.20) into polar cylindrical coordinates by taking x = r cos θ,
y = r sin θ, we obtain
∂ 2u
− u = 0,
∂θ2
(6.21)
150
6. PARABOLIC EQUATIONS
which is independent of the radial direction r. Since the solution on any circular path
inside the square is required to be single-valued, it must return to the same values as
θ changes by 2π. Both solutions of Eq (6.21) are real exponential functions, but they
will be periodic only if u = 0 on this circular path. That is, u = 0 within and on the
largest circle C inside the square, which is defined by x2 + y 2 = 1 (Fig. 6.2). For the
circular paths in the corners we solve a simple two-point boundary value problem.
y
1
C
−1
x
0
1
−1
Fig. 6.2. The Square x = ±1, y = ±1.
Hence, the solution is given by

0, if x2 + y 2 ≤ 1,




|x| |y| 


exp arcsin
+ exp arccos
r r ,
u(x, y) =
x2 + y 2 − 1
1
1


exp arcsin
+ exp arccos



r
r


if x2 + y 2 > 1, −1 < x < 1, −1 < y < 1,
(6.22)
where r = x2 + y 2 . This solution is identically zero within and on the largest
circle C inside the square but nonzero in the corners only. In fact, the boundary
conditions and the solution at the points (±1, 0), (0, ±1) merge smoothly into the
solution u = 0 at the core which is the region on and interior to the circle C defined
by x2 + y 2 ≤ 1. This is because the prescribed boundary conditions tend to zero as
the points (±1, 0), (0, ±1) are approached since at these points the boundary of the
square and the largest inscribed circle C meet. However, if the boundary conditions
were prescribed arbitrarily on the square such that they (these boundary conditions)
do not approach zero at the above four points, then the solution would break down in
the neighborhood of these points, and we would no longer have a general solution.
This shows that this Dirichlet problem in not well posed.
6.3. 3-D DIFFUSION EQUATION
151
6.3. 3-D Diffusion Equation
∂
To find Green’s function for the operator
− k ∇2 in R3 , t > 0, we first assume
∂t
for the sake of simplicity that the singularity is at the origin and at t = 0. Green’s
function G(x; t) satisfies the following equation and the boundary conditions:
∂G
− k ∇2 G = δ(x) δ(t),
∂t
G(x; t) → 0 as x → ∞ and for t > 0.
(6.23)
Applying the triple Fourier transform with respect to the space variables, followed by
Laplace transform with respect to t, to Eq (6.23), we get
s G(α,
β, γ, s) − k (α2 + β 2 + γ 2 ) G(α,
β, γ, s) =
1
,
(2π)3/2
where α, β, and γ are the variables of Fourier transform with respect to x, y, and z,
respectively. Hence,
G(α,
β, γ, s) =
(2π)3/2
1
.
s − k (α2 + β 2 + γ 2 )
Inverting first with respect to Laplace transform, we have
H(t) −k(α2 +β 2 +γ 2 )t
G(α,
β, γ, s) =
e
.
(2π)3/2
Then inverting the triple Fourier transform, we get
G(x; t) =
2
2
2
2
H(t)
H(t)
e−(x +y +z )/4kt =
e−|x| /(4kt) .
(2π)3/2 (2kt)3/2
8(πkt)3/2
By translating the singularity to x and t , Green’s function for this operator with
singularity at (x , t ) is
G(x − x ; t − t ) =
2
H(t − t )
e− |x−x | /(4k(t−t )) .
8[πk(t − t )]3/2
(6.24)
6.3.1. Electrostatic Analog. Gauss’s law for an electric charge q applied at the
qr0
origin that produces an electric field E, given by E =
, has two parts: One,
4πε0 r 2
defined by the surface integral
q
, if S the closed surface includes the origin,
ε0
E · dσ =
S
0, otherwise,
152
6. PARABOLIC EQUATIONS
which by the divergence theorem (see §1.7) gives the second part
r q
r0 · dσ
q
0
=
∇
·
dV = 0.
4πε0 S r 2
4πε0 V
r2
z
r0
S
S’
q
r0’
y
x
Fig. 6.3.
The first part can be resolved by indenting the origin by a small sphere S of radius
δ. Let the volume outside the outer surface S and the volume inside the surface S be connected by a small hole, which joins the two surfaces S and S and combines
them into one simply-connected closed surface (Fig. 6.3). Since the radius of this
(imaginary) hole may be made arbitrarily small which can vanish in the limit, the
contribution by this hole to the surface integral is zero, and we obtain
q
r0 · dσ
r0 · dσ +
= 0.
4πε0
r2
δ2
S
S
In the second integral the element dσ = −r0 δ 2 dω, where dω is the volume element
and the minus sign is taken because the positive normal r0 is directed outward from the
volume. Since the outward r0 is in the negative radial direction, we have r0 = −r0 .
Thus, the second integral becomes
r0 · dσ r0 · dr0 δ 2 dω
=
−
= −4π.
2
δ
δ2
S
S
Hence, Gauss’s law becomes
E · dσ =
S
q
q
4π = .
4πε0
ε0
(6.25)
Also, if q denotes the total distributed charge enclosed by the surface S so that
2
q = V 2 dV , where 2 denotes the charge density, then S E · dσ = V
dV , which
ε0
6.3. 3-D DIFFUSION EQUATION
by Gauss’s theorem gives
V
∇ · E dV =
V
the two integrands must be equal, and thus
∇·E=
153
2
dV . Since the volume V is arbitrary,
ε0
2
,
ε0
(6.26)
which is one of Maxwell’s equations. Form (6.26) we can derive Poisson’s equation
by replacing E by −∇u; thus, Poisson’s equation is
∇ · ∇u = ∇2 u = −
2
,
ε0
(6.27)
which for 2 = 0 (absence of electric charge) reduces to Laplace’s equation ∇2 u = 0.
If the charges are point charges qi , then the solution of Poisson’s equation is
u=
1 qi
,
4πε0
ri
(6.28)
i
which is a superposition of single-point charge solutions.2 Then
1
r(r)
u
=
dr,
4πε0 r 2
r
r=0
or for the potential at r = r1 away from the origin and the charge r = r2 , we have
1
2(r2 )
u (r1 ) =
dr2 .
(6.29)
4πε0 r 2
|r1 − r2 |
Since by Gauss’s law,
∇2 (1/r) dV =
V
0
if the volume V does not include origin,
−4π
if the origin is included,
1
= −δ(r), or if there is a translation of the charge from
4πr
origin to to the position r = r2 , we have
1
2
∇
= −δ(r1 − r2 ),
4πr12
we can write ∇2
where r12 = |r1 − r2 |, and the Dirac δ-function δ(r1 − r2 ) = 0 for r1 = r2 . Thus,
in view of (6.29) the solution of the equation is
G(r1 , r2 ) =
2
1
,
4π|r1 − r2 |
(6.30)
This follows from Coulomb’s law for the force between two point charges q1 and q2 ,
q1 q2 r0
which is E =
.
4πε0 r2
154
6. PARABOLIC EQUATIONS
where G(r1 , r2 ) is known as Green’s function for Poisson’s equation (6.27). Green’s
function behaves like an influence function which enlarges or reduces the effect of
the charge element 2(r2 ) dr2 depending on its distance from the field point r1 . Thus,
it represents the effect of a unit point source at r2 in producing a potential at r1 . Also,
the solution (6.29) can be written as
1
u (r1 ) =
G(r1 , r2 ) 2(r2 ) dr2 .
(6.31)
ε0 r 2
Using the ‘cause and effect’ terminology of physicists, the charge density 2(r2 )
might be labeled as ‘cause’ that produces the ‘effect’ u(r1 ), which means that a
potential field is produced by the charge distribution, although the ‘effectiveness’ of
the charge distribution to produce this potential depends on the distance between the
element of charge 2(r2 ) dr2 and the point at r1 . Thus, the ‘influence’ of the charge
is represented by the quantity {4π|r1 − r2 |}−1 , and for this reason this quantity,
although designated as Green’s function, is sometimes called an influence function.
6.4. Schrödinger Diffusion Operator
If we consider the nonhomogeneous case, then the Fourier heat equation
1 ∂
− ∇2 u(x, t) = f (x, t)
k ∂t
(6.32)
has two interpretations:
(i) If k > 0 is a real constant, depending on the specific heat and thermal conductivity of the medium, then u(x, t) represents a temperature distribution. The source
function f (x, t) on the right side describes local heat production minus absorption.
(ii) The function u(x, t) represents a particle density and k is the diffusion coi
efficient. If k is purely imaginary such that k =
, where m is the mass of the
2m
−34
quantum particle, and = 1.054 × 10
joules-sec (or 6.625 × 10−27 erg-sec) is
the Planck’s constant, then Eq (6.32) for the 1-D steady-state Schrödinger equation
for a free particle becomes
i
∂u(x, t)
2 2
+
∇ u(x, t) = 0.
∂t
2m
(6.33)
In order to solve this equation subject to the boundary and initial conditions
u(x, t) → 0
as |x| → ∞, t > 0,
and u(x, 0) = f (x),
we apply Fourier transform to this system and get
u
t = −
i α2
u
,
2m
u
(α, 0) = f (α).
−∞ < x < ∞,
6.4. SCHRÖDINGER DIFFUSION OPERATOR
155
The solution of this system in the transform domain is
u
(α, t) = f (α) e−i bα t ,
2
b=
2m
which after inverting gives the solution as
∞
1
u(x, t) = √
f (α) ei α(x−bαt) dα
2π −∞
∞
∞
1
−i αy
= √
f (y) e
dy
ei α(x−αbt) dα
2π −∞
−∞
∞
∞
1
= √
f (y) dy
ei α(x−y−bαt) dα.
2π −∞
−∞
(6.34)
Since
i α(x−y−bαt)
e
2 2 x−y
x−y
x−y
= exp − i bt α − 2α
+
−
2bt
2bt
2bt
2 x−y
i (x − y)2
exp − i bt α −
exp
2bt
4bt
2
2
x−y
= ei (x−y) /(4 bt) e−i bq t , where q = α −
,
2bt
2
the function u(x, t) in (6.34) becomes
∞
∞
2
2
1
u(x, t) =
f (y) ei (x−y) /(4 bt) dy
e−i bq t dq
2π −∞
−∞
∞
2
1−i
= √
f (y) ei (x−y) /(4 bt) dy,
2 2πbt −∞
(6.35)
which, in view of (4.6), gives Green’s function for the 1-D steady-state Schrödinger
operator as
(1 − i) H(t) i (x−y)2 /(4 bt)
√
G(x; t) =
e
.
(6.36)
2 2πbt
Note that in view of (6.16), Green’s function in this case is
2
m
G(x; t) = H(t)
eimx /(2t) ,
2iπ t
which reduces to (6.36). We have assumed in this problem that the source is of
positive unit strength. Sometimes the notation h is used for the Plank’s constant; it is
related to by h = 2π .
The argument in the notation G(x, t; x , t ) is composed of two parts: The first
part x, t denotes the field point where the effect of the impulsive heat source located
156
6. PARABOLIC EQUATIONS
at the source point signifies the temperature at the point x at time t. The second
part x , t denotes the cause which is the impulsive heat source situated at the point
x generating an instantaneous (impulsive) heat at an earlier time t . The combined
notation has the physical significance of an entire space-time process which can be
visualized as G(effect; cause) ≡ G(x, t; x , t ).
6.5. Min-Max Principle
Let R = {(x, t) : 0 ≤ x ≤ l, 0 ≤ t ≤ T } denote a closed rectangle in the (x, t)plane. Then we have the following result:
Theorem 6.1. If u(x, t) is continuous on R and satisfies the 1-D diffusion
equation in R, then u(x, t) attains its maximum and minimum values on the
sides x = 0 and x = l or at time t = 0.
Proof. First, we will prove the maximum principle by contradiction. Let M be
the maximum value of u in R, and, contrary to the hypothesis, we will assume that
the maximum value on the vertical sides or the at time t = 0 is M − ε, where ε > 0.
Let us further assume that the maximum value is attained at a point (x0 , t0 ) ∈ R, i.e.,
u (x0 , t0 ) = M , where 0 < x0 < l, t0 > 0. Now, consider a function
ε
2
v(x, t) = u(x, t) + 2 (x − x0 ) ,
4l
which approaches u as ε → 0. Then on the vertical sides of R and at t = 0 we
ε
3ε
have v(x, t) ≤ M − ε + = M −
, since v (x0 , t0 ) = M . Thus, the maximum
4
M
value is not attained on the vertical sides of R or at t = 0. Again, let (x1 , t1 ),
0 < x1 < l, 0 < t1 < T be a point in R where v(x, t) attains its maximum value.
Then, for the function v to satisfy the necessary condition for a maximum at (x1 , t1 )
we must have vt = 0 if t1 < T or vt = 0 if t1 = T , and vxx ≤ 0, i.e., we must have
ε
vt − vxx ≥ 0 at (x1 , t1 ). But, since vt − vxx = ut − uxx − 2 < 0, we arrive at a
2l
contradiction, which proves the maximum part of the theorem. The minimum part is
then proved because −u satisfies the maximum principle.
6.6. Diffusion Equation in a Finite Medium
Consider the 1-D diffusion equation ut = k uxx , 0 < x < a, t > 0, subject to
the initial and boundary conditions u(x, 0) = 0, 0 < x < a; u(0, t) = U0 , t >
0, and ux (a, t) = 0, t > 0. Applying Laplace transform with respect to t, we
obtain the system:
d2 u
s
− u = 0, 0 < x < a,
dx2
k
du U0
u(0, s) =
,
= 0,
s
dx x=a
6.7. AXISYMMETRIC DIFFUSION EQUATION
157
where s is the variable of the transform. The solution of this system in the transform
domain is
cosh (a − x) s/k
.
u(x, s) = U0
s cosh a s/k
Using the Table D.1, the inverse Laplace transform gives the solution as
∞
4 (−1)n
(2n − 1)(a − x) π −(2n−1)2 π2 kt/(4a2 )
u(x, t) = U0 1 +
cos
e
π n=1 2n − 1
2a
∞
4 1
(2n − 1)πx −(2n−1)2 π2 kt/(4a2 )
= U0 1 −
sin
e
.
(6.37)
π n=1 2n − 1
2a
Note that the same result is obtained if this problem is solved by the Bernoulli’s
separation method.
6.7. Axisymmetric Diffusion Equation
Consider the axisymmetric diffusion equation in the polar cylindrical coordinate system
1
ut = k urr + ur , 0 ≤ r ≤ ∞, t > 0,
r
which is subject to the initial condition u(r, 0) = f (r) for 0 < r < ∞. Applying the
zero-order Hankel transform we get
du
+ k σ 2 u = 0,
dt
u(σ, 0) = f (σ),
where σ is the variable of the Hankel transform. The solution of this problem in the
transform domain is
2
u(σ, t) = f (σ) e−kσ t ,
which after inversion gives the solution as
∞
2
u(r, t) =
σ f (σ) J0 (σr) e−kσ t dσ
0 ∞ ∞
2
=
σ
zJ0 (σz) f (z) dz e−kσ t J0 (σr) dσ
0
0 ∞
∞
2
=
z f(z) dz
σ J0 (σz)J0 (σr) e−kσ t dσ,
0
0
where we have interchanged the order of integration. Since
∞
2
1 −(r2 +z2 )/(4kt) rz σJ0 (σz) J0 (σr) e−kσ t dy =
e
I0
,
2kt
2kt
0
158
6. PARABOLIC EQUATIONS
(see Gradsteyn and Rhyzik [2007; §6.626]), where I0 is the modified Bessel function,
the above solution becomes
∞
rz 2
2
1
u(r, t) =
e−(r +z )/(4kt) z f (z) I0
dz.
(6.38)
2kt 0
2kt
Let f (r) represent a heat source concentrated at r = 0. This situation can be visualized
by considering the heatsource concentrated in a circle of radius ε and letting ε → 0,
ε
1
1 δ(r)
and assuming that lim
r f (r) dr =
, or equivalently, f (r) =
. Then
ε→0 0
2π
2π r
the solution of this problem due to the heat source concentrated at r = 0 is
∞
rz 2
2
2
1
1
u(r, t) =
δ(r) I0
e−(r +z )/(4kt) =
e−r /(4kt) . (6.39)
4πkt 0
2kt
4πkt
6.8. 1-D Heat Conduction Problem
This example illustrates an application of finite Fourier sine transform directly by
using Fourier sine series as mentioned in §5.5.3 Consider the nonhomogeneous heat
conduction problem with Dirichlet boundary conditions:
ut − uxx = f (x, t),
0 < x < π, t > 0,
u(0, t) = u(π, t) = 0,
u(x, 0) = 0.
(6.40)
The solution of this problem, if any, can be formally written in a Fourier series for
each fixed t as
∞
!
u(x, t) =
bn (t) sin nx,
where bn (t) =
n=1
π
u(x, t) sin nx dx for an integer n . To determine the coefficients
0
bn , we take the finite Fourier sine transform of both sides of (6.40), which gives
bn (t) + n2 bn (t) = Bn (t),
where
Bn (t) =
(6.41)
π
f (x, t) sin nx dx.
(6.42)
0
Since the initial condition gives bn (0) = 0, so, solving Eq (6.41) subject to this
t
2
condition, we get bn (t) =
e−n (t−t ) Bn (t ) dt . Now, if the problem (6.40) has
0
a solution u ∈ C 2 , then we can formally write
∞ t
2 −n2 (t−t )
u(x, t) =
e
Bn (t ) dt sin nx.
π n=1 0
To use the finite Fourier sine transform, simply replace bn (t) and Bn (t) by u
s (n, t)
and fs (n, t) in the solution provided in below for this boundary value problem.
3
6.8. 1-D HEAT CONDUCTION PROBLEM
By Schwarz’s inequality, we have
t
2
bn (t)2 ≤
e−n (t−t ) dt
0
=
t
159
Bn2 (t ) dt
0
t
t
1 1
−2n2 t
2 1
−
e
B
(t
)
dt
≤
B 2 dt .
n
2n2
2n2 0 n
0
Then, using the Schwarz’s inequality for sums and the Parseval’s equality (§1.9) we
get
N
2
≤1
b
(t)
sin
nx
n
2
N
n=M +1
1
π
≤
t0
Since
π
n=M +1
N
n=M +1
1
n2
1
n2
N
Bn2 dt
0 n=M +1
t
0
t
π
[f (x, t ] dx dt .
2
0
2
[f ] dx dt converges for some t0 > 0, the series
0
!
bn (t) sin nx
0
converges uniformly for 0 ≤ x ≤ π, 0 ≤ t ≤ t0 . Hence, we can write the solution
as
∞ t
2 −n2 (t−t )
u(x, t) =
e
Bn (t ) dt sin nx,
(6.43)
π n=1 0
where u ∈ C 2 and satisfies the given conditions in (6.40). Under suitable conditions
t
2
on f , like f (0, t) = 0 = f (π, t) and
fxx
dx being uniformly bounded, the series
0
in (6.43) can be differentiated term-by-term, and hence, u defined by (6.43) is the
required solution.
Note that (i) Under some less restrictions on f , we use (6.42) and by formally
interchanging integration and summation, obtain
t π
2
u(x, t) =
G(x, x ; t, t ) f (x , t ) dx dt ,
(6.44)
π 0 0
where
∞
G(x, x , t) =
e−n
2
t
sin nx sin nx
(6.45)
n=1
is Green’s function for this problem.
(ii) If the initial condition in (6.40) is changed to u(x, 0)
π= g(x), then the above
initial condition bn (0) = 0 must be replaced by b∗n (0) = 0 g(x) sin nx dx. Then
bn (t) =
0
t
e−n
2
(t−t )
Bn (t ) dt + b∗n (0) e−n t ,
2
160
6. PARABOLIC EQUATIONS
and thus, the solution becomes
∞ t
∞
2
2
−n2 (t−t )
u(x, t) =
e
Bn (t ) dt sin nx +
b∗n (0) e−n t sin nx.
π n=1 0
n=1
(6.46)
Green’s function in this case, however, remains the same as (6.45).
6.9. Stefan Problem
We will consider the one-phase Stefan problem with constant applied heat flux, which
can be stated as follows: A solid, initially at the melting temperature, is heated by
prescribing either the temperature or the heat flux at a boundary. Let the melting
interface be denoted by X(t). Then the problem is to determine the interface X(t) as
well as the temperature u(x) in the liquid. The governing equation and the boundary
conditions for the case of a prescribed heat flux at a boundary are
ut = uxx ,
0 < x < X(t), t > 0,
dX
,
dt
− ux (0, t) = g(t); and X(0) = 0.
u(x, t) = 0 for x ≥ X(t);
u(0, t) = f (t),
or
−ux (X, t) =
(6.47)
This problem is solved by applying a double Laplace transform L{u(x, t)} = u (κ, s),
defined by
∞
u (κ, s) =
e
−st
0
X(t)
e−κt u(x, t) dx dt = L{u(x, t)},
(6.48)
0
where κ and s are the variables of Laplace transform, {s} > 0, {κ} > 0. Assuming that X(t) is an increasing function and 0 < X(t) < t so that X has an inverse,
which we will denote by ξ, i.e., ξ (X(t)) = t, we find that
L{ut (x, t)} = s u,
∞
−st
L{uxx (x, t)} =
e
0
e
−κx
uxx dx dt
0
X
dX −κX
−κx
=
e
−
e
+ g(t) + κ
e
ux dx dt
dt
0
0
∞
= g(s) + κ2 u − κ f (s) −
e−s ξ(y)−κ y dy,
∞
−st
X
0
where f (s) and g(s) are Laplace transforms of f and g, respectively. Thus,
∞
κ f (s) − g(s) +
es ξ(y)−κy dy
0
u(κ, s) =
.
κ2 − s
(6.49)
6.9. STEFAN PROBLEM
161
√
√
Notice that u is defined for κ = s as well as for κ = − s, which implies that the
numerator in the above expression must vanish in either case, i.e., we must have
∞
√
√
g(s) = s f (s) +
e−sξ(y)−y s dy,
0
∞
√
√
g(s) = − s f (s) +
e−sξ(y)+y s dy,
0
Hence,
∞
−sξ(y)
f (s) =
e
0∞
e−sξ(y)
g(s) =
√
sinh (y s)
√
dy,
s
√ cosh y s dy.
(6.50)
0
Inverting u with respect to κ, we get
√
√ g(s) sinh (x s)
√
u(x, s) = f (s) cosh x s −
s
√
∞
sξ(y) sinh ((x − y) s )
√
+
e
dy
s
0
√
∞
sinh ((y − x) s )
√
=
esξ(y)
dy.
s
0
(6.51)
The above result, after integrating by parts and inverting, gives the nontrivial solution
as
∞
1 dn
2n
u(x, t) =
[X(t) − x] ,
(6.52)
n
(2n)!
dt
n=1
and
∞
1 dn
2n
[X(t)] ,
n
(2n)!
dt
n=1
f (t) =
∞
1
dn
2n−1
g(t) =
[X(t)]
.
n
(2n
−
1)!
dt
n=1
(6.53)
Next, we will find an approximate solution of the integral equation for the prescribed flux g(t) by considering the second integral equation in (6.50):
∞
√ g(s) =
e−sξ(y) cosh y s dy.
0
Assuming that ξ(y) is quadratic, i.e., taking ξ(y) = A y + By 2 , we get
∞
√ 2
g(s) =
e−s(A y+By ) cosh y s dy
0
=
0
∞
√
e−
s(A y+By 2 )
√
s
cosh y dy,
162
6. PARABOLIC EQUATIONS
which on inversion gives
1
g(t) = √
πt
∞
2
1
et/(A +4Bt) .
A2 + 4Bt
0
(6.54)
Since the time derivative of t = AX + BX 2 is 1 = Ẋ (A + 2BX), and 4t B + A2 =
(A + 2BX)2 = 1/Ẋ 2 , we find for the case when t is a quadratic function of X that
e−(By
2
+A2 y 2 )/(4t)
cosh y dy = √
g(t) = Ẋ exp t Ẋ 2 .
(6.55)
√
a ea
Note that g(t) = et for X = t; g(t) = √ for X = 2a t; and g = 1 when
t
Ẋ exp t Ẋ 2 = 1, so that the precise value of X can be determined as follows:
dX
Let c = tẊ 2 , so Ẋ = e−c , t = c e2c , which gives
= (2c + 1) ec . Integrating
dc
this equation by parts with respect to c and using the initial condition X(0) = 0 we
find the parametric solution of this differential equation as X(t) = 1 + (2c − 1) ec ,
t = c e2c .
2
The algorithm to compute values of X uses the following series for X(t) when
g = 1:
X(t) = t −
1 2
5
51
827 5 18961 6
t + t3 − t4 +
t −
t
2!
3!
4!
5!
6!
574357 7 21995899 8 1032666859 9
+
t −
t +
t − ··· ,
7!
8!
9!
where the remainder after n terms is bounded and approaches ≈ 0.04 as n → ∞.
By using parametric solutions Ruehr [2002] has computed the values of X at t =
10k , k = 0, 1, 2, . . . , 7, as follows:
X(1) = 0.772443774949597096137987895609995394,
X(10) = 4.56549276368834105587472,
X(100) = 21.29139510280473148,
X(1000) = 87.161250425814,
X(104 ) = 332.089233225,
X(105 ) = 1212.124436,
X(106 ) = 4302.9380, X(107 ) = 14984.1.
MATLAB solution of the equation Ẋ exp t Ẋ 2 = 1 such that X(0) = 0 is given
by
X = −1 + W (2t) exp
1
2
W (2t) + 1 ,
6.10. 1-D FRACTIONAL DIFFUSION EQUATION
163
where W (t) is the Lambert W-function, and it is presented in Fig. 6.4.
X( t )
5
4.5
4
3.5
3
2.5
2
1.5
1
0.5
0
2
4
6
8
10
t
Fig. 6.4. Solution of Equation Ẋ exp t Ẋ 2 = 1, X(0) = 0.
6.10. 1-D Fractional Diffusion Equation
Consider the general boundary value problem of the fractional diffusion equation
Dtp u − ∇2 u = 0 in Rn × R+
(6.56)
∂p
where Dtp ≡ p , 0 < p ≤ 1, is the Caputo time derivative of the fractional order
∂t
p (see Appendix E). For p = 1, we get the classical diffusion equation (see §6.1 for
n = 1), and p = 0 gives Helmholtz’s equation (see §4.3.4, 4.3.5, 8.5, 8.5.1, 8.6, and
8.12.4). We will solve the boundary value problem involving Eq (6.56) for n = 1,
which is defined by
∂pu
∂2u
=
k
, x ∈ R, t > 0,
∂tp
∂x2
u(x, t) → 0 as |x| → ∞,
Dtp−1 u(x, t)
= f (x), x ∈ R,
(6.57)
t=0
where k is the diffusivity constant. Applying Fourier transform to (6.57) with respect
to x and subject to the given boundary condition, we get
Dtp u
(α, t) = −k α2 u
,
Dtp−1 u
(α, t)
= f(α),
t=0
(6.58)
164
6. PARABOLIC EQUATIONS
where u
(α, t) is Fourier transform of u(x, t) and α is the variable of Fourier transform.
Now, taking Laplace transform of (6.58), we have
u
(α, s) =
sp
f(α)
,
+ k α2
whose inverse Laplace transform gives
u
(α, t) = f(α) tp−1 Ep,p −k α2 tp ,
(6.59)
where Ep,q is the Mittag-Leffler function defined by
Ep,q (z) =
∞
zm
,
Γ(pm + q)
m=0
p > 0, q > 0.
Again, taking the inverse Fourier transform of (6.59) we obtain
∞
u(x, t) =
G(x − x , t) f (x ) dx ,
−∞
where
1
G(x, t) =
π
∞
−∞
tp−1 Ep,p −k α2 tp cos αx dα.
(6.60)
(6.61)
(6.62)
To evaluate this integral we take Laplace transform of G(x, t). Using the formula
[Costabel, 2004]
(m)
L tmp+q−1 Ep,q
(±atp ) =
m! sp−q
(sp ∓ a)
m+1 ,
dm
Ep,q (z), we obtain
dz m
∞
1
cos αx
1
G(x, s) =
dα = √ s−p/2 exp
π −∞ sp + k α2
2 k
(6.63)
(m)
where Ep,q
=
|x|
− √ sp/2 .
k
Taking the inverse Laplace transform of this result finally we get
1
p p
G(x, t) = √ tp/2−1 W −x , − ,
,
2 2
2 k
where x =
(6.64)
|x|
√ and W (z, p, q) is the Wright function defined by
k
tp/2
W (z, p, q) =
∞
zn
.
n! Γ(pn + q)
n=0
(6.65)
6.10. 1-D FRACTIONAL DIFFUSION EQUATION
165
For p = 1, the boundary value problem (6.56) reduces to 1-D diffusion equation
discussed in §6.1.2, and Green’s function (6.64) reduces to (6.16), that is,
2
1
x
1
G(x, t) = √ W − √ , − 12 , 12 = √
e−x /(4kt) .
2 kt
kt
2 πkt
Another Method. (Kemppainen and Ruotsalainen [2008:141]) Green’s
function G(x, t) for the boundary value problem (6.57) can be obtained by taking
Laplace transform in the time variable and Fourier transform in the space variable of
the fractional equation
∂2
Dtp −
G(x, t) = δ(x, t),
(6.66)
∂x2
which yields G(α,
s) =
1
. Using Laplace transform of the Mittag-Lefler
α2 + sp
function, defined by (6.60), we get Fourier transform of Green’s function as
G(α,
t) = tp Ep,p −α2 tp ,
which on inversion gives Green’s function in terms of the Fox H-function (see Notations; Fox [1961], Prudnikov et al. [1990:626], Costabel, [2004]) as
p,p
p−1
−2
20
2 −p G(x, t) = t
|x| H12 |x| t .
(6.67)
(1,1),(1,1)
6.10.1. 1-D Fractional Diffusion Equation in Semi-Infinite Medium.
Consider the boundary value problem defined by the heat equation in the semi-infinite
medium x > 0, where the boundary x = 0 is kept at a temperature U0 f (t) and the
initial temperature in the whole medium is zero:
∂pu
∂2u
=
k
, 0 < x < ∞, t > 0,
∂tp
∂x2
u(x, 0) = 0, x > 0,
u(0, t) = U0 f (t),
t > 0,
u(x, t) → 0 as x → ∞.
After applying Laplace transform to this problem we get
d2 u sp
+
u(x, s) = 0, x > 0,
dx2
k
u(0, s) = U0 f (s), and u(x, s) → 0 as x → ∞.
(6.68)
The solution of this problem in the transform domain is
−ax
u(x, s) = U0 f (s) e
,
where a =
sp
,
k
(6.69)
166
6. PARABOLIC EQUATIONS
which on inversion gives the solution as
t
f (t − τ ) g(x, τ ) dτ = U0 f (t) A g(x, t),
u(x, t) = U0
(6.70)
0
"
#
where g(x, t) = L−1 e−ax . For α = 1, the solution (6.69) in the Laplace transform
domain is
√
u(x, s) = U0 f (s) e−x s/k ,
which on inversion gives
t
f (t − τ )g(x, τ ) dτ = U0 f (t) A g(x, t),
u(x, t) = U0
0
where
√ #
"
2
x
g(x, t) = L−1 e−x s/k = √
e−x /(4kt) .
3
2 πkt
(6.71)
In particular, for p = 1 and f (t) = 1, the solution (6.69) in the Laplace domain is
u(x, s) =
U0 −x√s/k
e
,
s
which on inversion gives the classical solution
u(x, t) = U0 erfc
x
√
2 kt
(6.72)
.
6.11. 1-D Fractional Schrödinger’s Diffusion Equation
Consider the boundary value problem involving 1-D fractional Schrödinger’s equation
for a fine particle of mass m defined by
∂pu
2 ∂ 2 u
=
−
, x ∈ R, t > 0,
∂tp
2m ∂x2
u(x, 0) = f (x), x ∈ R,
i
(6.73)
u(x, t) → 0 as |x| → ∞,
where u(x, t) is the wave function, the Plank’s constant (see §6.4), and f (x) an
arbitrary function. Applying Laplace transform followed by Fourier transform to the
problem (6.73), we get
p−1 f (α)
s) = s
u(α,
,
sp + kα2
where k =
i
,
2m
(6.74)
6.12. EIGENPAIRS
167
and α and s are the variables of Fourier and Laplace transforms, respectively. Inverting
(6.74) and using the convolution theorem of Fourier transform, we obtain
∞
1
u(x, t) = √
f(α) Ep,1 −kα2 tp dα
2π −∞
(6.75)
∞
−1 2 p
=F
f (α) Ep,1 −kα t
=
G(x − x , t) f (x ) dx ,
−∞
where Green’s function G(x, t) is given by
1
1
G(x, t) = √ F −1 Ep,1 −kα2 tp
=
2π
2π
∞
−∞
Ep,1 −kα2 tp ei αx dα,
(6.76)
where Ep,1 is the Mittag-Leffler function defined by (6.60). For p = 1, the solution
(6.75) reduces to
∞
u(x, t) =
G(x − x , t) f (x ) dx ,
(6.77)
−∞
where Green’s function G(x, t) is
1
G(x, t) =
2π
∞
−∞
Ep,1 −kα t
2 p
1
dα =
2π
∞
2
ei αx−kα t dα
−∞
2
1
= √
e−x /(4kt) .
2 πkt
(6.78)
This solution matches with the classical solution defined by (6.36).
6.12. Eigenpairs and Dirac Delta Function
First, we will find eigenfunctions for boundary value problems, and then represent the
Dirac δ-function as a sum of eigenfunctions. This not only leads to certain methods
of explicitly finding fundamental solutions for some boundary value problems but
also provides a different approach to define the Dirac δ-function.
Let a linear boundary value problem be defined in a domain D by
∂u
+ k(s)u = h(s) on ∂D,
∂n
and let the eigenfunctions φn be given by L(φn ) = λn φn , where λn are called
the eigenvalues associated with the eigenfunctions φn . The form of the eigenpairs
(λn , φn ) depends on the coordinate system used for the boundary value problem. For
example, if the coordinate system is rectangular cartesian, the eigenfunctions are of
the form cos nx, sin nx, or einx , and the eigenvalues depend on n if the linear operator
L contains constant coefficients. The eigenfunctions φn must$be a complete set$in the
!N
sense that for any function u there exists an integer N such that $u− n=1 cn φn $ ≤ ε,
L(u) = 0,
u = g(s),
168
6. PARABOLIC EQUATIONS
where ε > 0 is a preassigned
%arbitrarily small quantity and cn are constants, and the
norm is defined by u =
u ū dV , where ū denotes the complex conjugate of
V
u. A sufficiently continuous function u possesses an eigenfunction expansion to any
degree of accuracy; thus
u = lim
N →∞
N
cn φn .
n=0
Also, a complete set of eigenfunctions forms an orthonormal set with the property that
< φm , φn >= δmn , where δmn is the Kronecker delta. Now, we can define the Dirac
δ-function on a region D in terms of the complete orthonormal set of eigenfunctions
for the region D, as
∞
δD (x, x ) =
φn (x) φ̄n (x ),
(6.79)
n=0
where x and x are two points in D, which are called the field and the source point,
respectively.
Example 6.1. Consider the 1-D transient problem:
∂2u
1 ∂u
=
,
∂x2
k ∂t
−a < x < a,
(6.80)
with the initial and boundary conditions as u(x, 0) = F (x) for −a < x < a, and
u(−a, t) = 0 = u(a, t) for t > 0. Using Bernoulli’s separation method by assuming
u(x, t) = X(x)T (t), we get
1 d2 X
1 dT
=
= −λ2 ,
2
X dx
kT dt
where λ is a real parameter. The boundary conditions and initial condition become:
X(−a) = 0 = X(a), and T (0) = F (x). It will be found that only the choice of
λ > 0 gives meaningful results, leading to the solution of the following two equations
d2 X
+ λ2 X = 0,
dx2
dT
+ kλ2 T = 0,
dt
which eventually leads to the solution
u(x, t) =
∞
An sin
−∞
where
2aAn =
nπx −n2 π2 kt/a2
e
,
a
∞
F (x) sin
−∞
nπx
dx.
a
(6.81)
(6.82)
6.12. EIGENPAIRS
169
nπ 1
nπx
The orthonormal set of spatial eigenpairs for this problem are
,
sin
,
a 2a
a
1 inπx/a
and the orthonormal eigenfunctions are φn (x) = 2a
e
. Hence, the Dirac δfunction in the region −a < x < a for the steady–state (as t → ∞) 1-D Laplace’s
equation over the interval −a < x < a is
∞
∞
1 inπx/a −inπx /a
1 inπ(x−x )/a
δ(x, x ) = 2
e
e
= 2
e
.
4a n=−∞
4a −∞
(6.83)
Note that in (6.82) we have assumed that the eigenfunction expansion is periodic
with period 2a. Hence the Dirac δ-function (6.83) also has period 2a. The solution
(6.81) can also be obtained by using the Laplace transform method. We have graphed
the real part of this function for the basic interval −a < x < a in Fig. 6.5, using
n = 5, 10, and 25 terms in (6.83), with a = 3, and x = 0. The graphs, however,
repeat outside this interval with a period 2a. From the graphs in Fig. 6.5, it is obvious
that the peak becomes infinitely higher and narrower at x = 0 as n increases.
0.6
0.3
0.5
0.25
0.4
0.2
0.15
0.3
n =5
0.1
0.1
0.05
-3
-2
1
-1
n = 10
0.2
2
3
x
-3
-2
-1
2
1
-0.1
-0.05
1
0.8
0.6
n = 25
0.4
0.2
x
-3
-2
-1
1
2
-0.2
Fig. 6.5. Dirac δ -Function with 5, 10, and 25 Terms.
Example 6.2. Consider the 2-D Laplace’s equation
∂ 2u ∂ 2u
+ 2 =0
∂x2
∂y
3
3
x
170
6. PARABOLIC EQUATIONS
over the rectangle R = {−a < x < a, −b < y < b}. The eigenpairs for this region
are given by
n2 π 2 1
1
λn = − √
+
,
b2
2 ab a2
Note that
1
φn (x, y) = √ einπx/a einπy/b .
2 ab
∂ 2 φn
∂ 2 φn
n2 π 2 1
1
+
=− √
+ 2 φn .
2
2
2
∂x
∂y
b
2 ab a
It can be easily verified that these eigenvalues are orthonormal by showing that
a b
&
'
φm , φn =
φm φ̄n dy dx = δmn . The solution in the complex form of
−a
−b
a double Fourier series over the rectangle is given by
∞
∞
1
u(x, y) = √
cnm einπx/a eimπy/b .
2 ab n=−∞ m=−∞
(6.84)
This representation of u is periodic both in x and y with periods 2a and 2b, respectively.
The Dirac δ-function for this region is
δ(x, x ) = δ(x, x ) δ(y, y )
=
∞
1 inπ(x−x )/a inπ(y−y )/b
e
e
,
4ab n=−∞
(6.85)
where x = (x, y) is the field point and x = (x , y ) the source point. The solution
of the problem is given by
u(x) = u(x, y) =
∞
∞
cn1 n2 φn1 n2 .
(6.86)
n1 =−∞ n2 =−∞
6.13. Exercises
∂
∂2
− k 2 in
∂t
∂x
the region {0 < x < a; t > 0} subject to the conditions G (0, t; x , t ) =
G (a, t; x , t ) = 0 and G (x, t; x , t ) = 0 for t < t .
∞
2 2
2
2
nπx
nπx
Ans. G (x, t; x , t ) = H(t − t )
e−n π k(t−t )/a sin
sin
.
a
a
a
n=1
6.1. Find Green’s function G (x, t; x , t ) for the operator L ≡
∂u
∂2u
− k 2 = 0,
∂t
∂x
subject to the boundary conditions u(x, 0) = 0, u(0, t) = T0 , and u(a, t) = 0.
6.2. Use Green’s function to find the solution of the problem
6.13. EXERCISES
171
6.3. Use Green’s function to find the solution of the problem ut − uxx = 0, x ∈ R,
0 < t, subject to the conditions u(x, 0) = H(x) − H(−x).
x Ans. u(x, t) = sgn(x) erf √ .
2 t
∂
∂ ∂
6.4. For the Fokker-Planck operator
−
+ x find Green’s function in R.
∂t ∂x ∂x
Hint. Use the transformation u = v et , X = x et and 2T = e2t . Then the given
operator reduces to the diffusion operator.
2
− x − x e−(t−t )
H(t − t )
.
Ans. G (x, t; x , t ) = % exp
2 1 − e−2(t−t )
2π 1 − e−2(t−t )
6.5. Solve the diffusion problem with a source function f (x, t), defined by
ut = k uxx + f (x, t),
x ∈ R, t > 0, such that u(x, 0) = 0 for x ∈ R.
t
∞
2
(t − τ )−1/2 dτ
f (z, τ ) e−(x−z) /(4k(t−τ ) dz.
1
Ans. u(x, t) = √
2 πk 0
−∞
6.6. Solve the boundary value problem: ut = k uxx − h u, 0 < x < ∞, t > 0,
where h is a constant, such that u(x, 0) = u0 for x > 0, and u(0, t) = 0 and
ux (0, t) → 0 uniformly in t as x → ∞, t > 0.
x
−ht
Ans. u(x, t) = u0 e
erf
.
4kt
6.7. Solve the boundary value problem: ut = k uxx , 0 < x < ∞, t > 0, such
that u(x, 0) = 0 for x > 0, and u(0, t) = v0 and ux (0, t) → 0 uniformly in t as
x → ∞, t > 0.
x
Ans. u(x, t) = v0 erfc
.
4kt
6.8. Solve uxx − ut = A e−αx , A ≥ 0, α > 0, where u(0, t) = 0 = u(l, t) for
t > 0, and u(x, 0) = f (x) for 0 < x < l.
x
A
A A
Ans. u = v − 2 − 2 e−αl − 1
+ 2 e−αx , where
α
α
l
α
∞
nπx −n2 π2 t/l2
2 l
nπx
v=
An sin
e
, An =
g(x) sin
dx, and
l
l
l
0
n=1
x
A g(x) = f (x) + 2 1 + e−αl − 1
− e−αx .
α
l
2
6.9. Solve ut = k uxx + f (x, t), 0 < x < a, subject to the boundary conditions
u(0, t) = 0 = u(a, t) for 0 ≤ x ≤ a, and the initial condition u(x, 0) = 0 for
t > 0.
∞
nπx
Ans. u(x, t) =
un (t) sin
, where
a
n=1
172
6. PARABOLIC EQUATIONS
2 a
nπy
f (y, t) sin
dy.
a 0
a
0
6.10. Find the interior temperature of the cooling ball of Example 4.13, if
un (t) =
t
2
en
π 2 k2 (τ −t)/a2
fn (τ ) dτ,
f (ρ, φ) =
∞
Ans. u(ρ, φ, t) =
Cmn
m,n=1
fn (t) =
1,
0 ≤ φ < π/2
0,
π/2 ≤ φ < π.
Jn+1/2 (αmn ρ)Pn (cos φ) −α2 t
e mn ,
√
αmn ρ
1
αmn 0 ρ3/2 Jn+1/2 (αmn ρ) dρ
1 2
2n(n + 1) 0 ρJn+1/2
(αmn ρ) dρ
√
1
(2n + 1)Pn (0) αmn 0 ρ3/2 Jn+1/2 (αmn ρ) dρ
=
,
2
n(n + 1) Jn−1/2
(αmn )
Cmn =
(2n + 1)Pn (0)
√
where αmn are the consecutive positive roots of Jn+1/2 (α) = 0.
6.11. Determine the steady-state temperature inside a solid hemisphere 0 ≤ ρ ≤ 1,
0 ≤ φ ≤ π/2 (a) when the base φ = π/2 is at 0◦ and the curved surface ρ = 1,
0 ≤ φ < π/2, is at 1◦ ; and (b) when the base φ = π/2 is insulated, but the
temperature on the curved surface is f (φ).
1
∂u
∂u sin φ ∂u
Hint:
= cos φ
−
= 0 on the base. Also use
Pn (x) dx =
∂z
∂ρ
ρ ∂φ
0
1
P (0), and
n(n + 1) n
n(n + 1) (x2 −1) Pn (x) =
Pn+1 (x)−Pn−1 (x) , where Pn (x) are the Legendre
2n + 1
polynomials.
Ans. (a) u(ρ, φ) =
∞
ρ2n+1 [P2n (0) − P2n+2 (0)] P2n+1 (cos φ).
n=0
(b) u(ρ, φ) =
∞
cn ρ2n P2n (cos φ), where
n=0
π/2
cn = (4n + 1)
f (φ) P2n (cos φ) sin φ dφ.
0
6.12. Solve ut = uxx , −π < x < π, subject to the conditions u(x, 0) = f (x),
u(−π, t) = u(π, t), and ux (−π, t) = ux (π, t), where f (x) is a periodic function
of period 2π. This problem describes the heat flow inside a rod of length 2π,
which is shaped in the form of a closed circular ring.
Hint: Assume X(x) = A cos ωx + B sin ωx.
6.13. EXERCISES
173
∞
2
Ans. ωn = n; u(x, t) =
e−n t (an cos nx + bn sin nx), where
n=0
π
1 π
1 π
1
an =
f (x) cos nx dx, bn =
f (x) sin nx dx, a0 =
f (x) dx.
π −π
π −π
2π −π
6.13. Solve the problem ut = ∇2 u, r < 1, 0 < z < 1, such that u(r, z, 0) = 1,
u(1, z, t) = 0, and u(r, 0, t) = 0 = u(r, 1, t). This problem describes the
temperature distribution inside a homogeneous isotropic solid circular cylinder.
∞
!
2
2 2
Ans. u(r, z, t) =
Cmn e−(λm +n π )t J0 (λm r) sin nπz, where λm are
m,n=1
the zeros of J0 , and Cmn =
4 (1 − (−1)n )
.
nπλm J1 (λm )
6.14. Find the steady-state temperature in a solid circular cylinder of radius 1 and
height 1 under the conditions that the flat faces are kept at 0◦ and the curved
∞
I0 (nπr) sin nπz
surface at 1◦ . Ans. u(r, z) = 4
.
I0 (nπ) mn
n=1
n odd
6.15. Solve the steady-state problem of temperature distribution in a half-cylinder
0 ≤ r ≤ 1, 0 ≤ θ ≤ π, 0 ≤ z ≤ 1, where the flat faces are kept at 0◦ and the
∞
16 Im (nπr) sin nπz
curved surface at 1◦ . Ans. u(r, θ, z) = 2
sin mθ.
π m,n=1 Im (nπ) mn
m,n odd
∂u
∂
∂u
6.16. Solve
=
x
, 0 < x < 1, t > 0, subject to the conditions
∂t
∂x
∂x
u(x, 0) = f (x) and u(1, t) = 0.
Hint: Set 4x = r 2 .
∞
√
2
Ans. u(x, t) =
Cn e−λn t/4 J0 (λn x), where λn are the zeros of J0 (λ), and
1
n=1
√
f
(x)
J
0 (λn x) dx
Cn = 0 1
.
√
[J0 (λn x)]2 dx
0
6.17. Solve the dimensionless partial differential equation governing the plane wall
transient heat conduction ut = uxx , 0 < x < 1, subject to the boundary conditions u(0, t) = 1, u(1, t) = 0 for t > 0, and the initial condition u(x, 0) = 0
for0 < x < 1. Also find the half-space solution.
Hint. Since the nonhomogeneous boundary condition does not allow us to
determine the eigenfunctions, as in Exercise 6.15 above, we proceed as follows: First, we find a particular solution of the problem, which satisfies only
the boundary conditions. Although there is more than one way to determine
the particular solution, we can, for example, take the steady-state case, where
the equation becomes u
xx = 0, which, after integrating twice, has the general
solution u
(x) = c1 x + c2 , with the boundary conditions u
(0) = 1, u
(1) = 0.
174
6. PARABOLIC EQUATIONS
Thus, c1 = −1, c2 = 1, and the steady-state solution is u
(x) = 1 − x. Next, we
formulate a homogeneous problem by writing u(x, t) as a sum of the steady-state
solution u
(x) and a transient term v(x, t), i.e., we take u(x, t) = u
(x) + v(x, t),
or v(x, t) = u(x, t) − u
(x), thus reducing the problem to finding v(x, t); that
is, we solve the following problem: vt = vxx , with the boundary conditions
v(0, t) = u(0, t) − u
(0) = 0, v(1, t) = u(1, t) − u
(1) = 0, and the initial condition v(x, 0) = u(x, 0) − u
(x) = x − 1, by the Bernoulli’s separation method as
in the above Exercise 6.15.
∞
2 1 −n2 π2 t
Ans. v(x, t) = −
e
sin nπx, and
π n=1 n
∞
x 2 1 −n2 π2 t/l2
nπx
−
e
sin
, where l is the thickness
l π n=1 n
l
of the plate. The solution for the half-space is derived by letting l → ∞. Since
∞
2 !
l −n2 π2 t/l2
nπx π
lim u(x, t) = 1−
e
sin
· , let nπ/l = ξ and π/l = dξ.
l→∞
π n=1 nπ
l l
∞
x x 2
1 −ξ2 t
Then lim u(x, t) = 1−
e
sin ξt dξ = 1−erf √ = erfc √ .
l→∞
π 0 ξ
2 t
2 t
6.18. Derive Eq (6.21).
u(x, t) = 1 −
6.19. Solve the linearized Bergers equation ut = ν uxx , −∞ < x < ∞, t > 0,
where ν is the kinematic viscosity, such that
u(x, 0) =
f1 (x), x < 0,
f2 (x), x > 0.
Hint. Use Fourier transform with respect to x.
x
[f1 (x) − f2 (x)] erf √ .
2 νt
6.20. Show that the solution u(x, t) for the equation uxx = ut + xu, x ∈ R, t > 0,
and satisfying the initial condition u(x, 0) = f (x) is
2
∞
x − s − t2
1
t3 /3−xt
u(x, t) = √ e
f (s) exp −
ds.
4
2 πt
−∞
Ans. u(x, t) =
1
2
[f1 (x) + f2 (x)] −
1
2
6.21. Show that the solution of the diffusion equation ut = uxx , t ≥ 0 and satisfying the initial conditions u(x, 0) =
1,√x> 0; u(x, 0) = −1, x < 0, with
discontinuity at t = 0 is u(x, t) = erf 4x t .
6.22. Solve the heat conduction problem ut = k uxx , 0 ≤ x ≤ a, t > 0, subject to
the Dirichlet boundary conditions u(0, t) = 0 = u(a, t), and the initial condition
u(x, 0) = f (x), 0 ≤ x ≤ a.
∞
2 −k n2 π2 t/a2
nπx
nπs
Ans. u(x, t) =
e
sin
f (s) sin
ds.
a n=1
a
a
7
Hyperbolic Equations
The homogeneous equation for the wave operator (or the d’Alembertian ) c ≡
∂2
− c2 ∇2 describes, e.g., the dynamics of a vibrating string in R × R+ , of a
∂t2
drum membrane or surface of a lake in R2 × R+ , or of a sound wave in air, or an
electromagnetic wave in vacuum in R3 × R+ . Here c denotes the speed of sound
or light. The wave equation propagates signals with velocities less than or equal to
c. Unlike the diffusion equation, the wave equation is not affected by a change in
the sign in the time variable, and so in this sense it is reversible. We will discuss
Sturm-Liouville system for 1-D wave equation, which is followed by finding Green’s
functions for 1-D, 2-D and 3-D wave equations, and other applications.
7.1. 1-D Wave Equation
We will discuss some boundary value problems for 1-D wave equation by different
methods.
7.1.1. Sturm-Liouville System for 1-D Wave Equation. Consider the
problem
∂2u
∂ ∂u =
p(x)
− q(x)u + F (x, t), a ≤ x ≤ b, t > 0,
∂t2
∂x
∂x
a1 u(a, t)+a2 ux (a, t) = 0, b1 u(b, t)+b2 ux (b, t) = 0,
(7.1)
u(x, 0) = f (x), ut (x, 0) = g(x),
a < x < b,
where the function p, q and F , and the constants a1,2 and b1,2 are the same as in Eq
(6.1). Let L denote the Sturm-Liouville operator. Then Eq (7.1) can be written as
∂
(p(x) ux )+q(x)u.
(7.2)
∂x
Using Bernoulli’s separation method, let the solution of Eq (7.2) with F = 0 be of
the form u(x, t) = X(x) T (t) = 0. Then Eq (7.2) reduces to
utt = L[u]+F, where L[u] ≡
175
176
7. HYPERBOLIC EQUATIONS
d2 T
= λ T, t > 0,
dt2
where λ is the separation constant. The boundary conditions on X are
L[u] = λX,
x ∈ I,
aa X(a) + a2 X (a) = 0,
b1 X(b) + b2 X (b) = 0.
(7.3)
(7.4)
The system (7.3)-(7.4) is known as the associated Sturm-Liouville system for
1-D wave equation, which can be solved by finding the eigenpair (λn , Xn (x)),
n = 1, 2, . . . . Using the superposition principle, the solution of Eq (7.2) can be
formally written as
∞
u(x, t) =
Xn (x) Tn (t),
(7.5)
n=1
where Tn (t) are yet to be determined, as follows: Assume that the forcing term
F (x, t) can be expanded in terms of the eigenfunctions Xn (x) as
∞
F (x, t) =
(7.6)
Xn (x) fn (t),
n=1
where fn (t) are the generalized Fourier coefficients of F (x, t), given by
b
fn (t) =
(7.7)
F (x, t) Xn (x) dx.
a
Then substituting (7.5) and (7.6) into (7.2), we find that
∞
∞
∞
T̈n (t)Xn (t) = L
Tn (t)Xn (x) +
Xn (x) fn (t)
n=1
n=1
=
∞
n=1
∞
Tn (t) L [Xn (x)] +
n=1
Xn (x) fn (t)
n=1
= [λn Tn (t) + fn (t)] Tn (x),
which yields the second-order ordinary differential equation
T̈n (t) + µ2 Tn (t) = fn (t),
Tn (0) = f (x), T (0) = g(x),
where µ2n = −λn . Applying the Laplace transform method to this equation we find
that
Tn (t) = Tn (0) cos (µn t) +
1
1
Ẋ(0) sin (µn t) +
Ẋ(0) sin (µn t) ,
µn
µn
where from the initial conditions in (7.1) we have
b
Tn (0) =
Xn (s) f (s) ds, Ṫn (0) =
a
a
b
Xn (s) g(s) ds.
(7.8)
7.1. 1-D WAVE EQUATION
177
Hence, from (7.5) and (7.7) the solution u(x, t) of this problem is given by
u(x, t) =
=
∞
∞ a
n=1
b
=
a
Xn (x) Tn (t)
n=1
b
∞
b
1
Xn (s) f (s) ds cos (µn t) +
Xn (s) g(s) ds
µn
a
b
1
+
sin µn (t − τ ) fn (τ ) dτ
µn a
Xn (x)X(s)
n=1
∞
b
sin (µn t)
cos (µn t) f (s) ds
1
Xn (x)X(s) sin (µn t) g(s) ds
µ
a
n=1 n
t b
∞
1
+
Xn (x)X(s) sin (µn t) F (s, τ ) ds dτ.
µ
0
a
n=1 n
+
(7.9)
Define two Green’s functions G1 and G2 as
G1,2 (x, x ; t) =
∞
Xn (x)X(x )
n=1
cos (µn t)
sin (µn t)
.
(7.10)
Then the solution (7.9) can be written as
b
u(x, t) =
G1 (x, x ; t) f (x ) +
a
t
b
+
0
a
1
G2 (x, x ; t) g(x ) dx
µn
1
G2 (x, x ; t − τ ) F (x , τ ) dx dτ.
µn
(7.11)
7.1.2. Vibrations of a Variable String. Consider the problem of vibrations
−2
of a string of density proportional to (1 + x) and defined by
1
∂2u ∂2u
−
= 0,
(1 + x)2 ∂t2
∂x2
0 < x < 1, t > 0,
(7.12)
∂u
(x, 0) = 0, and the boundary
∂t
conditions u(0, t) = 0 = u(1, t). By Bernoulli’s separation method (u = X(x) T (t))
we get
λ
X +
X = 0,
T + λ T = 0,
(1 + x)2
(7.13)
X(0) = 0 = X(1),
T (0) = f (x), T (0) = 0,
and subject to the initial conditions u(x, 0) = f (x),
178
7. HYPERBOLIC EQUATIONS
where λ is the separation constant. This is an eigenvalue problem. The equation
inX has
of the form (1 + x)a , where a(a − 1) + λ = 0, or a =
√ the solutions
1
2 1 ± 1 − 4λ . Then for X(0) = 0 we get
√
√
X(x) = (1 + x)1+ 1−4λ)/2 − (1 + x)1− 1−4λ)/2 .
√
√
√
1+ 1−4λ)/2
1− 1−4λ)/2
1−4λ)
The condition X(1)
=
0
gives
2
−
2
=
0,
or
2
= 1.
√
If λ < 1/4, then 1 − 4λ is real, and this equation has no solution in this case. If
λ = 1/4, the two solutions are (1+x)1/2 and (1+x)1/2 ln(1+x), which are no longer
linearly independent; although the latter solution satisfies the condition X(0) = 0, it
fails to satisfy X(1) = 0,√thus showing that λ = 1/4 is not an eigenvalue. Finally, if
λ > 1/4, the expression 1 − 4λ is imaginary, but it has the solutions
i√
= (1 + x)1/2 exp
4λ − 1 ln(1 + x)
2 1
1
1/2
= (1 + x)
cos
λ − ln(1 + x) + i sin
λ − ln(1 + x) ,
4
4
(1 + x)(1+
√
1−4λ)/2
and the two linearly independent solutions are given by the real and imaginary parts
of this expression. To satisfy the condition X(0) = 0 we take
1
1/2
X(x) = (1 + x)
sin
λ − ln(1 + x) .
4
1
1
1/2
Then the condition X(1) = 0 gives 2 sin
λ − ln 2 = 0, so λ − ln 2
4
4
1
must be an integral multiple of π, i.e., λ − ln 2 = π, which gives the eigenvalues
4
n π 2 1
λn =
+ , n = 1, 2, . . . ,
ln 2
4
and the corresponding eigenfunctions are
Xn (x) = (1 + x)1/2 sin
We take the Fourier series f (x) =
∞
n π ln(1 + x)
.
ln 2
cn Xn (x), where, since p(x) = (1 + x)−2 ,
n=1
1
n π ln(1 + x)
dx
ln 2
cn =
1
n π ln(1 + x)
(1 + x)−1 sin2
dx
0
ln 2
1
2
n π ln(1 + x)
=
f (x)(1 + x)−3/2 sin
dx.
ln 2 0
ln 2
0
f (x)(1 + x)−3/2 sin
7.1. 1-D WAVE EQUATION
179
The Fourier series converges absolutely and uniformly if f (0) = f (1) = 0 and
1 2
f dx < +∞. Since Eq (7.13) with T (0) = 1, T (0) = 0 has the solution
0
√
T (t) = cos λ t, the problem (7.12) has the solution
u(x, t) =
∞
n=1
cn cos
n2 π 2
1
+ t
2
(ln 2)
4
(1 + x)1/2 sin
n π ln(1 + x)
.
ln 2
(7.14)
This series converges uniformly, but to ensure continuous derivatives of u we must
1
assume that f ∈ C 2 , and f (0) = f (1) = 0, f (0) = f (1) = 0, and 0 f 2 dx <
+∞.
7.1.3. Green’s Function for 1-D Wave Equation
We assume that the singularity is at the origin and at t = 0. In R, Green’s function
G1 (x; t) satisfies the equation
∂ 2 G1
∂ 2 G1
− c2
= δ(x) δ(t),
2
∂t
∂x2
such that G1 (x; t) → 0 as |x| → ∞, t > 0. Applying the Laplace and Fourier
transforms to this equation, we get
1 = √1 ,
s2 + α2 c2 G
2π
or
1 = √
G
1
,
2π (s2 + α2 c2 )
which after the Laplace inversion gives
1 =
G
1
√
sin αct.
αc 2π
After inverting the Fourier transform, we obtain
G1 (x; t) =
1 H(x + ct) − H(x − ct) .
2c
However, if we reverse the process by carrying out the Fourier inversion prior to the
Laplace inversion, we get
G1 (x, t) =
1
1
H (t − |x|/c) =
H(ct − |x|).
2c
2c
It is easy to check that the two solutions are equivalent. Thus, if a source is at the
point (x ; t ), the above solution becomes
G1 (x, t; x , t ) =
1 H c(t − t ) − (|x − x |) .
2c
(7.15)
180
7. HYPERBOLIC EQUATIONS
7.2. 2-D Wave Equation
In R2 , Green’s function G2 satisfies the equation
2
∂ 2 G2
∂ G2
∂ 2 G2
2
−
c
+
= δ(x − x ) δ(y − y ) δ(t − t ).
∂t2
∂x2
∂y 2
Applying the Laplace transform, we get
2
∂ G2
∂ 2 G2
2
2
s G2 − c
+
= e−st δ(x − x ) δ(y − y ),
∂x2
∂y 2
or, using axial symmetry in polar cylindrical coordinates with r 2 = (x − x )2 + (y −
y )2 , we have
2
d G2
1 ∂G2
e−st δ(r)
s2 G2 − c2
+
=
.
(7.16)
dr 2
r ∂r
2πr
e−st δ(r)
in Eq (7.16) can be considered as representing a source
2πr
of strength e−st /c2 at the point r = 0. The solution of the homogeneous part of Eq
(7.16) is A I0 (kr) + B K0 (kr), where I0 and K0 are the modified Bessel functions of
the first and the second kind of order zero, respectively, and k = s/c. Since I0 → ∞
as r → ∞, we take A = 0, and then B is obtained by equating the flux across a circle
Cε of radius ε to e−st /c2 . Thus,
s sε ∂K0 (sr/c)
e−st
− lim Cε B
ds = lim 2πε B
K1
= 2π B = 2 .
ε→0
ε→0
∂r
c
c
c
−st
−st
e
e
sr
Hence, B =
, and G2 (r, s; t ) =
K0
. Since L−1 {K0 (αs)} =
2πc2
2πc2
c
H(t − α)
√
(see Erdelyi et al. [1954], or Abramowitz and Stegun [1965]), on inversion
t2 − α 2
we get

1

for r < c(t − t )
2
2πc c (t − t )2 − r 2
G2 (r, t; t ) =
(7.17)

0
for r > c(t − t )
The forcing term
=
H(c(t − t ) − r)
.
2πc c2 (t − t )2 − r 2
7.3. 3-D Wave Equation
In R3 , Green’s function G3 satisfies the equation
∂2G
∂ 2 G3
∂ 2 G3
∂ 2 G3 3
− c2
+
+
= δ(x − x )δ(y − y )δ(z − z )δ(t − t ).
2
2
2
∂t
∂x
∂y
∂z 2
7.3. 3-D WAVE EQUATION
181
Applying the Laplace transform, we get
s2 G3 − c2
∂2G
3
∂x2
+
∂ 2 G3
∂ 2 G3 +
= e−st δ(x − x )δ(y − y )δ(z − z ),
2
2
∂y
∂z
or, using the spherical symmetry with r 2 = (x − x )2 + (y − y )2 + (z − z )2 , we
have
s2 G3 − c2
d2 G
dr 2
2 ∂G3 e−st δ(r)
=
.
r ∂r
4π r 2
3
+
(7.18)
Set r G3 = V in Eq (7.18). Then it reduces to
s2 V − c2
∂2V
e−st δ(r)
=
,
∂r2
4π r
r = |x − x |.
The solution of this equation is V = A esr/c + B e−sr/c . Since esr/c becomes
unbounded as r → ∞, we take A = 0. Thus, G3 = B e−sr/c /r. To evaluate B, the
e−st δ(r)
forcing term, which is
in Eq (7.18), can be considered as representing a
4π r2
−st 2
source of strength e
/c at the point r = 0. Then by equating e−st /c2 to the flux
across a sphere Sε of radius ε and letting ε → 0, we have
B
ε→0
or
∂ e−sr/c e−st
dS = 2 ,
∂r
r
c
− lim
∂Sε
s e−sr/c
e−sr/c e−st
−4πB lim ε2 −
−
=
,
ε→0
c
r
r2
c2
r=ε
e−st −sr/c
which gives 4πB = e
/c , or G3 =
. On inversion, this yields
4π rc2
1
r
G3 (r; t, t ) =
δ
t
−
t
−
,
4π rc2
c
−st
2
or G3 (x, y, z, t; x , y , z , t )
1
1
2
2
2
=
δ t−t −
(x − x ) + (y − y ) + (z − z )
4π rc2
c
1
=
δ (c(t − t ) − r) , r = |x − x |.
4π rc
(7.19)
The graphs of Green’s functions G1 , G2 and G3 are presented in Figs. 7.1, 7.2 and
182
7. HYPERBOLIC EQUATIONS
7.3, respectively.
G1
G3
G2
δ
1/(2c)
1/(2 π c )
0
ct
Fig. 7.1.
|x|
|x|
|x|
0
0
ct
Fig. 7.2.
ct
Fig. 7.3.
7.4. 2-D Axisymmetric Wave Equation
Consider the 2-D wave equation in the polar cylindrical coordinates:
1 ∂
∂
utt − c2 ∇2 u = f (r, t), 0 < r < ∞, where ∇2 ≡
r
,
r ∂r
∂r
subject to the initial conditions u(r, 0) = 0 = ut (r, 0). The Green’s function G(r, t)
satisfies the system
1
Gtt − c2 ∇2 G =
δ(r)δ(t), G(r, 0) = 0 = Gt (r, 0).
2πr
Applying the zero-order Hankel transform with respect to r, followed by the Laplace
transform with respect to t, to this system, we get
G(σ, s) =
c2
,
2π (s2 + c2 σ 2 )
where s and σ are the variables of the Laplace and Hankel transform, respectively.
Inverting the Laplace transform we get
c
G(σ, t) =
sin(c σt),
2πσ
which after inverting the Hankel transform gives
G(r, t) =
c H(t) −1 1
H0
sin(c σt)
2π
σ
=
c H(t) H(ct − r)
√
.
2π c2 t2 − r 2
(7.20)
7.5. Vibrations of a Circular Membrane
Consider the problem of the transverse vibrations of a stretched circular membrane
of unit radius fixed on its boundary, defined by
2
∂ 2u
1 ∂u 1 2 ∂u
2 ∂ u
−
c
+
+
= 0, r < 1, t > 0,
(7.21)
∂t2
∂r2
r ∂r
r ∂θ
∂u
u(r, θ, 0) = f (r, θ),
(r, θ, 0) = 0, u(1, θ, t) = 0.
∂r
7.6. 3-D WAVE EQUATION IN A CUBE
183
Using Bernoulli’s separation
√ method (u = R(r)Θ(θ)T
√ (t)), the solutions are of the
form R(r) cos mθ cos c λ t and R(r) sin mθ cos c λ t, where m, n = 0, 1, 2, . . . ,
and R is a solution of the eigenvalue problem
m2
R + λrR = 0,
(7.22)
r
R(1) = 0; R and R bounded.
!
(m)
(m)
The eigenfunctions of this problem are Jm
λk
, while the eigenvalues λk
!
(m)
are the solutions of the equation Jm
λk
= 0. Thus,
(r R ) −
!
!
∞
1 (0)
(0)
Ck0 J0
λk r cos c λk t +
2
k=1
!
!
∞ ∞
(m)
(m)
+
(Ckm cos mθ + Dkm sin mθ) Jm
λk r cos c λk t ,
u(r, θ, t) =
n=1 m=1
(7.23)
where
Ckm =
1
am
0
1
π
−π
f (r, θ) cos mθ Jm
!
(m)
λk r r dr dθ,
!
1 π
1
(m)
Dkm =
f (r, θ) sin mθ Jm
λk r r dr dθ,
am 0 −π
2
1 !
(m)
am = π
Jm
λk r
r dr,
0
and the series in (7.23) converge uniformly to the solution u(r, θ, t). The special case
for a circular drum is discussed in Examples 4.12 on a circular drum.
7.6. 3-D Wave Equation in a Cube
Consider the problem in a cube D = {0 < x, y, z < π}:
2
∂2u
∂2u
∂ u ∂ 2u ∂ 2u
2 2
2
−c ∇ u≡ 2 −c
+ 2 + 2 =0
∂t2
∂t
∂x2
∂y
∂z
u(0, y, z, t) = u(π, y, z, t) = 0,
in D for t > 0,
u(x, 0, z, t) = u(x, π, z, t) = 0,
u(x, y, 0, t) = u(x, y, π, t) = 0,
u(x, y, z, 0) = f (x, y, z),
∂u
(x, y, z, 0) = g(x, y, z).
∂t
(7.24)
184
7. HYPERBOLIC EQUATIONS
We begin with the identity
2
∂u ∂ 2 u
∂ 1 ∂u
c2
∂u
2 2
2
−
c
∇
u
=
+
|∇u|
−
∇
·
c2
∇u ,
∂t ∂t2
∂t 2 ∂t
2
∂t
and integrate it over a 4-D domain bounded by the sides of the cube D, the initial
plane t = 0, and a 3-D surface S which passes through a point (x0 , y0 , z0 , t0 ) (see
Fig. 7.4).
(x0 , y0 , z0 , t 0 )
t
t0
z
S
y
D
x
0
Fig. 7.4. 4-D Cube.
Applying the divergence theorem (see §1.7) and using the conditions in (7.24), we
find that for f = g = 0,
S
1
2
∂u
∂t
2
+ c |∇u|
2
2
∂u ∂u
∂u
∂u
ni − c
nx +
ny +
nz
∂t ∂x
∂y
∂z
2
dS = 0,
(7.25)
where nx , ny , nz are the components of the upward normal n to the surface S such
that nt > 0 on S. Completing the squares, the integrand in (7.25) becomes
2
1 ∂u
∂u
c2 2
2
nt − c
+
n − c2 n2 |∇u|2 +
2nt ∂t
∂x
2nt t
2 2 2
c4 n2
∂u
1 ∂u
∂u
1 ∂u
∂u
1 ∂u
− 2
nx +
− 2
ny +
− 2
nz
,
2nt
∂x n ∂n
∂y
n ∂n
∂z
n ∂n
∂u
∂u
∂u
∂u
=
nx +
ny +
nz and n2 = n2x + n2y + n2z . This integrand is
∂n
∂x
∂y
∂z
nonnegative if
where
n2t − c2 n2 ≥ 0.
(7.26)
If the surface S has the property (7.26), then (7.25) implies that the integrand is
zero. Thus, for the steepest surface S with property (7.26) through the given point
7.6. 3-D WAVE EQUATION IN A CUBE
185
(x0 , y0 , z0 , t0 ), we must have
n2t − c2 n2x + n2y + n2z = 0 on S,
(7.27)
except at the point (x0 , y0 , z0 , t0 ). Thus, the surface S is the right cone with slope
1/c, and
!
1
2
2
2
t = t0 −
(x − x0 ) + (y − y0 ) + (z − z0 ) .
(7.28)
c
This cone is known as the characteristic cone with (x0 , y0 , z0 , t0 ) as its vertex.
Its cross-section by a plane t = const is the sphere of radius c (t − t0 ) and center
(x0 , y0 , z0 ). Now, on this cone S Eq (7.25) becomes
2
2
1
∂u
∂u
∂u nx ∂u
nt − c2
+ c4 n2
− 2
+
∂t
∂n
∂x n ∂n
S 2nt
2 2
∂u ny ∂u
∂u nz ∂u
− 2
+
− 2
dS = 0.
∂y
n ∂n
∂z
n ∂n
Since the integrand is the sum of continuous nonnegative functions, each term must
vanish everywhere on S. Also, since nt − c2 n2 = 0 and n2t + n2 = 1, we find from
the first term (which vanishes) that
∂u
∂u
∂u
∂u
2
=c 1+c
nx +
ny +
nz
on S.
∂t
∂x
∂y
∂z
Since at the vertex (x0 , y0 , z0 , t0 ) of the cone S the spatial part (nx , ny , nz ) of the
outward normal has all possible limiting directions, and therefore no fixed direction,
∂u
∂u
∂u
we conclude that
=
=
= 0 at (x0 , y0 , z0 , t0 ). Again, repeating this
∂x
∂y
∂z
argument at each point (x0 , y0 , z0 , t), t ≤ t0 , we conclude that u (x0 , y0 , z0 , t0 ) =
0. Hence, the solution u at the vertex (x0 , y0 , z0 , t0 ) is uniquely determined by
the boundary conditions imposed below and on the characteristic cone S through
(x0 , y0 , z0 , t0 ). Any surface whose outward normal satisfies the condition (7.27) is
known as the characteristic surface.
By using Bernoulli’s separation method it can be shown that the formal solution
of this problem is
∞ ∞ ∞ u(x, y, z, t) =
Clmn cos l2 + m2 + n2 ct+
l=1 m=1 n=1
Dlmn
where
π π π
8
f (x, y, z) sin lx sin my sin nz dx dy dz,
π3 0 0 0
π π π
8
= 3
g(x, y, z) sin lx sin my sin nz dx dy dz.
π 0 0 0
Clmn =
Dlmn
√
sin l2 + m2 + n2 ct
√
sin lx sin my sin nz,
l 2 + m 2 + n2
(7.29)
186
7. HYPERBOLIC EQUATIONS
7.7. Schrödinger Wave Equation
According to de Broglie’s postulation1 ‘every corpuscular motion can be treated as a
wave motion such that the wavelength λ is inversely proportional to the product of the
h
electronic mass m and velocity c of a particle’, i.e., λ =
, where h is the Plank’s
mc
constant (see §6.4). The total energy E of the particle is related to the wave frequency
ν by E = h ν. Since the kinetic energy is given by E = mc2 /2, the potential energy
V is related to E by E = V + mc2 /2, or
hν = V +
h2
.
2mλ2
(7.30)
Consider a wave function U defined by
U (x, t) = A cos 2π
x
λ
−νt−τ ,
where τ denote the phase shift such that it is a phase lag if t ≥ −2π + τ and a phase
advance if t ≤ 2π + τ . Then
x
x
Ut = 2π νA sin 2π
− ν t − τ , Utt = −4π 2 ν 2 A cos 2π
−νt−τ ,
λ
λ
x
x
2π
4π 2
Ux = −
A sin 2π
− ν t − τ , Uxx = − 2 A cos 2π
−νt−τ ,
λ
λ
λ
λ
which gives
Utt = −4π 2 ν 2 U,
Uxx = −
4π 2
U.
λ2
Hence,
h
Utt ,
4π 2 ν
2
h
h2
U
=
−
Uxx .
4mλ2
8π 2 m
hν U = −
(7.31)
(7.32)
Multiplying Eq (7.30) by U and substituting into Eqs (7.31) and (7.32), we get
∂2U
8π 2 m
2 ∂ 2U
−
VU= 2 2
,
2
2
∂x
h
λ ν ∂t2
which is 1-D Schrödinger’s wave equation. In 3-D this equation becomes
∇2 U −
1
8π 2 m
2 ∂2U
V
U
=
.
h2
λ2 ν 2 ∂t2
(7.33)
As mentioned in Sagan [1989:292], this postulation was made by de Broglie in 1924
and provided a basis for quantum mechanics.
7.8. HYDROGEN ATOM
Now, let U = A cos 2π
x
−νt−τ
λ
∇ u + 8π
2
Since E =
2
187
u(x, y, z). Then Eq (7.33) gives
1
m
− 2V
λ2
h
u = 0.
h2
1
mE
, or 2 = 2 , the above equation becomes
mλ2
λ
h
8π 2 m
(E − V ) u = 0,
(7.34)
h2
∞
where u(x, y, z) ≡ u(x) → 0 as cos |x| → ∞ such that
u2 dx dy dz < +∞,
−∞
∞
∞
and the norm u is defined by
u2 (x) dx dy dz = 1, (or
|u|2 dx dy dz
∇2 u +
−∞
−∞
= 1 in the case when u is complex). This condition can be expressed in spherical
coordinates as an iterated integral
r
2
π
2
r R dr
0
2π
2
Θ2 sin θ dθ = 1.
Φ dφ
0
(7.35)
0
The function
u is sometimes called a probability wave; that is, it is a function for
which
∞
u2 dx dy dz represents the probability that a particle is found in the
−∞
volume element dx dy dz.
7.8. Hydrogen Atom
The hydrogen atom consists of one nucleus of charge e and one electron of charge
−e with the electronic mass m. Hence, the potential energy is V = −e2 /r, where r
is the distance between nucleus and electron. Thus, Eq (7.34) becomes
8π 2 m
∇ u+
h2
2
e2
hν +
u = 0,
r
(7.36)
where we have chosen spherical coordinates such that the nucleus is at the origin,
i.e., the source point is (r , θ , φ ) = 0. The spherical coordinate system (see Fig.
7.5) is defined by x = r sin θ cos φ, y = r sin θ sin φ, z = r cos θ. Using Bernoulli’s
separation method, with u = R(r)Θ(θ)Φ(φ), where r > 0, 0 ≤ θ ≤ π, 0 ≤ φ < 2π,
Eq (7.36) becomes
R + 2 R /r cot θ Θ
1 Θ
1
Φ
8π 2 m
+ 2
+ 2
+ 2 2
+
R
r
Θ
r Θ
h2
r sin θ Φ
e2
hν +
= 0,
r
188
7. HYPERBOLIC EQUATIONS
or
r2 R 2r R
8π 2 m
+
+ r2
R
R
h2
e2
Θ Θ
1 Φ
hν +
= − cot θ
−
−
. (7.37)
r
Θ
Θ
sin2 θ Φ
z
( x ,y , z )
• ( r , θ, φ)
θ
r
y
φ
x
Fig. 7.5 Spherical Coordinates.
Since both sides of this equation must be equal to a constant, which we take as κ, we
get
8π 2 m 2
e2
2 r R + 2rR − κ R +
r hν +
R = 0,
(7.38)
h2
r
To separate equations in Θ and Φ, we take the constant as −µ2 . Then the right side
of Eq (7.37) gives two equations:
sin2 θ Θ + sin θ cos θ Θ + κ sin2 θ Θ − µ2 Θ = 0,
2
Φ + µ Φ = 0,
(7.39)
(7.40)
To solve Eq (7.39), let t = cos θ, Θ(θ) = P (t), and κ = n(n+1). Then this equation
becomes
dP
d µ2
1 − t2
+ n(n + 1) −
P = 0.
(7.41)
dt
dt
1 − t2
µ2
. The
1 − t2
µ
solutions Eq (7.41) are associated Legendre polynomials Θ = Pn (cos θ), where
Pnµ (t), and µ = 0, ±1, ±2, . . . . For associated Legendre’s equation and its properties,
see §C.2. The solution of Eq (7.40) are Φ = A cos µφ + B sin µφ, where φ is a
This equation would be Legendre’s equation if it were not for the term
7.8. HYDROGEN ATOM
189
periodic function of period 2π, i.e., Φ(φ) = Φ(2π + φ). Now, to solve Eq (7.38),
first we divide it by r 2 , and get
2 b n(n + 1)
R + R + a+ −
R = 0,
r
r
r2
√
8π 2 mν
ae2
and b =
. Introducing a new variable ρ = 2r −a, and
h
hν
b
2πme2
taking K = √
= √
, the above equation becomes
2 −a
h −2mhν
d2 R
2 dR
1 K
n(n + 1)
++
−
−
+
R = 0,
dρ2
ρ dρ
4
ρ
ρ2
where a =
or
d 2 ρ2 − 4Kρ + 4n(n + 1)
ρ R −
R = 0.
dρ
4
(7.42)
This equation is closely related to Laguerre’s equation, and its solution is
R(ρ) = ρn e−ρ/2 L2n+1
K+n (ρ),
(7.43)
where n ∈ Z, K is any number, and Lσk (ρ) are the generalized Laguerre polynomials
ρ−σ eρ dk
defined by Lσk (ρ) =
(ρn+σ e−ρ ) (for Laguerre polynomials and associk! dρk
ated Laguerre polynomials, see §C.4 and §C.5). If we impose the condition that K be
2π 2 m e4
an integer N , then we find that E = − 2 2 , N = 1, 2, . . . , are the possible valh N
ues of the kinetic energy such that these values represent a discrete energy spectrum
with a limit point at E = 0. For N = 1 we get a normal quantum state (or ground
state, as it is commonly called) of the hydrogen atom, which is the most stable state
occupied by the spectrum. For N > 1, we have a manifold of higher quantum states,
described as follows: Assuming that m ≤ k, we have 2n + 1 ≤ N + n, that is,
n ≤ N − 1, which means that for every N there are N possible values of n, namely
n = 0, 1, 2, . . . , N − 1. Since µ ≤ n, we will have µ = 0, 1, 2, . . . , n for every n.
N (N + 1)
Hence, for every N there are
independent quantum states. Hence, the
2
solutions of Eq (7.37) are given by
h2 N
2π 2 m e2
2 µ
uN,nµ (r, θ, φ) = 2
r
P
(cos
θ)
exp
−
r ×
n
4π m e2
h2 N
2
4π m e2
× L2n+1
r (A cos µφ + B sin µφ) , (7.44)
N +n
h2 N
where the constants A and B can be determined from the boundary conditions (to
be provided), and N is called the principal quantum number, n the azimuthal
quantum number, and µ the magnetic quantum number.
190
7. HYPERBOLIC EQUATIONS
7.8.1. Harmonic Oscillator. For a vibrating object the governing equation
involves the potential energy V which is defined by v = 12 kt2 = 12 mω 2 t2 , where ω
is the angular frequency of vibrations (= k/m ), k being the spring constant or force
1 unit. The classical vibration frequency is ν =
k/m, which gives ω = 2πν.
2π
The total energy is the sum of the potential and kinetic energies. We will now find
the form of the Hamiltonian operator. The kinetic energy T can be written in the
2 d2
operator form as T = −
, whereas the potential energy is 2π 2 ν 2 mt2 . Thus,
2m dt2
the Hamiltonian operator is
H=−
If we set b =
2 d2
2 d2
2 2
2
+
2π
ν
mt
=
−
+
2m dt2
2m dt2
1
2
mω 2 t2 .
2πνm
, then Schrödinger’s equation is Hu = Eu, or
d2 u
2mE
mω 2 2
+
−
t u = 0.
dt2
2
(7.45)
(7.46)
This equation is nonlinear as the potential varies with t2 . We seek a solution of the
form
u(t) = c e−bt ,
2
(7.47)
where b and c are constants. Then since
2
2
du
d2 u = −2bct e−bt , and
= −2bc + 4b2 ct2 e−bt ,
2
dt
dt
the term
2
2mE
mω 2 2
2mcE
m2 cω 2 2
−
t = − 2 +
t
e−bt .
2
2
d2
in Eq (7.46) and equating the
dt2
2
m2 cω 2 2 −bt2
mω
terms in t2 , we obtain
t e
= 4b2 ct2 e−bt , which yields b =
.
2
2
2
Again, equating the constant terms in Eq (7.46) we find that E = b , which gives
m
2
E = 12 ω2 . Hence, u(t) = ce−mωt /(2) , which is the solution of the Schrödinger
equation (7.46) for the harmonic oscillator in its lowest energy state.
Substituting the value of this term and that of
7.9. 1-D Fractional Nonhomogeneous Wave Equation
Consider the boundary value problem
2
∂pu
2 ∂ u
−
c
= F (x, t), x ∈ R, t > 0,
∂tp
∂x2
u(x, 0) = f (x), ut (x, 0) = g(x), x ∈ R,
(7.48)
7.9. 1-D FRACTIONAL NONHOMOGENEOUS WAVE EQUATION
191
where c is the wave constant and 1 < p ≤ 2 (for fractional derivatives, see Appendix
E). Applying the Laplace transform followed by the Fourier transform to the problem
(7.48), we get
(α, s)
p−1
g (α) sp−2
F
s) = f (α) s
u(α,
+
+
,
sp + c2 α2
sp + c2 α2
sp + c 2 α2
(7.49)
where s is the variable of the Laplace transform and a the variable of the Fourier
transform. First, inverting the Laplace transform we obtain
u
(α, t) = f(α) L−1
sp
sp−1
+ c2 α2
+
g (α) L−1
+ L−1
sp
sp−2
+ c2 α2
(α, s)
F
,
sp + c2 α2
which, in view of the formula (6.59), gives
u
(α, t) = f(α) Ep,1 −c2 α2 tp + g(α) Ep,2 −c2 α2 tp
t
+
F(α, t − τ ) τ p−1 Ep,p −c2 α2 tp dτ,
(7.50)
0
where Ep,1 is the Mittag-Leffler function defined by (6.60). Next, taking inverse
Fourier transform, we get
∞
1
u(x, t) = √
f(α) Ep,1 −c2 α2 tp ei αx dα
2π −∞
∞
1
+√
g (α) t Ep,2 −c2 α2 tp ei αx dα
2π −∞
t
∞
1
+√
τ p−1 dτ
F(α, t − τ ) Ep,p −c2 α2 tp ei αx dα.
2π 0
−∞
(7.51)
In the case when p = 2, the fractional wave equation in (7.48) reduces to the classical
1-D wave equation discussed in §7.1. In this case since
E2,1 −c2 α2 tp = cosh (i cαt) = cos(cαt),
t sinh(i cαt)
1
tE2,2 −c2 α2 tp =
=
sin(cαt),
i cαt
cα
the solution (7.51) reduces to
∞
1
u(x, t) = √
f(α) cos(cαt) ei αx dα
2π −∞
∞
1
sin(cαt) i αx
+√
g(α)
e
dα
cα
2π −∞
192
7. HYPERBOLIC EQUATIONS
t
∞
1
sin cα(t − τ ) i αx
+ √
dτ
F(α, τ )
e
dα
α
c 2π 0
−∞
"
1"
1 x+ct
= "f (x − ct) + f (x + ct)" +
g(y) dy
2
2c x−ct
x+c(t−τ )
1 t
+
dτ
F (y, τ ) dy.
2c 0
x−c(t−τ )
(7.52)
If we take c2 = k and g(x) = 0 in the problem (7.48), then it reduces to 1-D fractional
diffusion equation discussed in §6.10, such that Eq (7.49) reduces to
(α, s)
f(α) sp−1
F
u(α, s) = p
+
,
s + kα2
sp + kα2
which, in view of formula (6.59), gives
u
(α, s) = f(α)Eα,1 −kα2 tp
1
+
(t − τ )p−1 Ep,p −kα2 (t − τ )p F(α, t) dτ.
0
Taking the inverse Fourier transform of this result and applying convolution theorem
of the Fourier transform, we get
∞
1
√
u(x, t) =
f(α) Ep,1 −kα2 tp ei αx dα
2π −∞
t
∞
1
+√
dτ
F (α, τ ) (t − τ )p−1 Ep,p −kα2 (t − τ )p ei αx dα
2π 0
−∞
∞
=
G1 (x − x , t) f (x ) dx
−∞
t
(t − τ )p−1 dτ
+
0
where
∞
−∞
G2 (x − x , t − τ ) F (x , τ ) dx ,
∞
1
Ep,1 −kα2 tp ei αx dα,
2π −∞
∞
1
G2 (x, t) =
Ep,p −kα2 tp ei αx dα.
2π −∞
G1 (x, t) =
(7.53)
(7.54a)
(7.54b)
For α = 1, the solution (7.53) reduces to the classical solution:
∞
u(x, t) =
G1 (x − x , t − τ ) f (x ) dx
−∞
t
+
∞
dτ
0
−∞
(7.55)
G2 (x − x , t − τ ) F ((x , τ ) dx ,
7.10. APPLICATIONS OF THE WAVE OPERATOR
where
2
1
G1 (x, t) = G2 (x, t) = √
e−x /(4kt) .
2 πkt
193
(7.56)
7.10. Applications of the Wave Operator
We will discuss some examples which exhibit the importance of Green’s function in
certain physical problems.
7.10.1. Cauchy Problem for the 2-D and 3-D Wave Equation. The
Cauchy problem for this wave equation is governed by
utt = c2 (uxx + uyy ) ,
−∞ < x, y < ∞, t > 0,
subject to the initial conditions u(x, y, 0) = 0, ut (x, y, 0) = f (x, y) for −∞ <
x, y < ∞, where u and its first partial derivatives vanish at infinity, and c is a
constant. We apply the 2-D Fourier transform to this system, and get
d2 u
+ c2 σ 2 u
= 0, where σ 2 = α2 + β 2 ,
2
dt
d
u ""
u
(α, β, 0) = 0,
= f(α, β).
"
dt t=0
The solution of this system in the transform domain is
sin(c σt)
u
(α, β, t) = f(α, β)
,
cσ
which on inversion gives the solution
1
u(x, y, t) =
2πc
∞
∞
sin(c σt) i σ·r
f(σ)
e
dσ
cσ
−∞ −∞
∞ ∞ σ · r
1
f (σ)
=
exp i σ
+ ct
−
4 i πc −∞ −∞ σ
σ
σ · r
− exp i σ
− ct
dσ,
σ
(7.57)
where σ = (α, β) are the variables of the double Fourier transform and r = (x, y)
such that σ · r = αx + βy. In this solution the terms exp{i (σ · r ± γct) represent,
for any given σ, the plane traveling wave solution of the wave equation.
The solution of the Cauchy problem for the 3-D wave equation
utt = c2 (uxx + uyy + uzz ), −∞ < x, y, z < ∞, t > 0,
and subject to the initial conditions u(x, y, z, 0) = 0, ut (x, y, z, 0) = f (x, y, z),
194
7. HYPERBOLIC EQUATIONS
−∞ < x, y, z < ∞, can be similarly found as
f(ζ)
ζ·r
exp i ζ
+ ct
ζ
ζ
−∞ −∞ −∞
ζ ·r
− exp i ζ
− ct
dζ,
ζ
(7.58)
where ζ = (α, β, γ) are the variables of the triple Fourier transform and r = (x, y, z)
such that ζ · r = αx + βy + γz.
u(x, y, z, t) =
1
√
4π i c 2π
∞
∞
∞
Let k denote σ or ζ. Then for any given k the terms ei (k·r±kct) define a plane
traveling wave structure, since these terms remain constant on the planes k·r = ±kct
which move normal to themselves in the direction of ±k at constant speed c. This
wave structure, represented by the solution (7.58), is the sum total of all such plane
waves at different wave numbers k.
As a particular case, when f (x, y, z) = δ(x) δ(y) δ(z), we find that f(k) =
(2π)−3/2 , and the solution (7.58) becomes
u(x, y, z, t) =
1
8π 3
∞
−∞
∞
−∞
∞
−∞
sin(k ct) i (k·r)
e
dk,
kc
which, by transforming to spherical coordinates (r, θ, φ) where the z-axis is taken
along the direction of r such that k · r = kr cos θ, reduces to
u(r, t) =
=
=
=
=
2π
π
∞
1
sin ckt 2
dφ
dθ
ei kr cos θ
k sin θ dk
3
8π 0
ck
0
∞ 0
1
sin(ckt) sin(kr) dk
2π 2 cr 0
∞
1
ei k(ct−τ ) − ei k(ct+τ ) dk
2
8π cr
−∞
1 δ(ct − τ ) − δ(ct + τ )
4πcr
1
1
τ
δ(t − τ ) =
δ
t
−
,
4πcr
4πc2 r
c
(7.59)
where δ(ct + τ ) = 0 since ct + τ > 0 for t > 0.
7.10.2. d’Alembert Solution of the Cauchy Problem for the Wave
Equation. Method 1: (By Bernoulli’s separation method) First, we will consider
the homogeneous solution of the problem of a vibrating string governed by the 1-D
wave equation
2
∂2u
2∂ u
=
c
,
∂t2
∂x2
0 < x < l,
(7.60)
7.10. APPLICATIONS OF THE WAVE OPERATOR
195
subject to the boundary and initial conditions
u(0, t) = 0 = u(l, t),
u(x, 0) = f (x),
t > 0,
ut (x, 0) = g(x),
0 < x < l,
where f ∈ C 1 (0, l) is a given function. For a complete solution see Example 4.1.
Method 2: (By integral transform method) Let us now consider the nonhomogeneous Cauchy problem
utt − c2 uxx = f (x, t),
u(x, 0) = g(x),
−∞ < x < ∞, t > 0,
ut (x, 0) = h(x),
−∞ < x < ∞,
(7.61)
where f (x, t) represents the source term. Applying the Fourier transform, followed
by the Laplace transform, to this system, we get
u
(α, s) =
s g(α) + h(α) + f(α, s)
.
s2 + c2 α2
The inverse Laplace transform gives
h(α)
sin(cαt) +
cα
h(α)
= g(α) cos(cαt) +
sin(cαt) +
cα
u
(α, t) = g(α) cos(cαt) +
1 −!
cα
L
f(α, s)
cα
s2 + c2 α2
t
1
sin cα(t − τ ) f(α, τ ) dτ.
cα 0
which, after inverting the Fourier transform, gives the solution as
∞
i cαt
1
u(x, t) = √
e
+ e−i cαt ei αx g(α) dα
2 2π −∞
∞
i cαt
1
h(α)
√
+
e
− e−i cαt ei αx
dα
i cα
2 2π −∞
t
∞ i cα(t−τ )
1
e
+ e−i cα(t−τ ) i αx + √
dτ
e
f (α, τ ) dα
iα
2c 2π 0
−∞
1
1 x+ct
= g(x + ct) + g(x − ct) +
h(y) dy
2
2c x−ct
t
∞
x+c(t−τ )
1
+ √
dτ
f (α, τ ) dα
ei αy dy
2c 2π 0
−∞
x−c(t−τ )
x+ct
1
1
1
= g(x + ct) + g(x − ct) +
+
h(y) dy
2
2c 2c x−ct
x+c(t−τ )
1 t
+
dτ
f (y, τ ) dy.
(7.62)
2c 0
x−c(t−τ )
196
7. HYPERBOLIC EQUATIONS
7.10.3. Free Vibration of a Large Circular Membrane. To find the displacement function for the membrane, we consider the axisymmetric wave equation
1
ur = c2 utt , 0 ≤ r < ∞, t > 0,
r
u(r, 0) = f (r), ut (r, 0) = g(r), 0 ≤ r < ∞,
urr +
where f and g are arbitrary functions, and c2 = T /ρ = const, T being the tension in
the membrane, and ρ the surface density of the membrane. Applying the zero-order
Hankel transform to this system we get
u(σ, t) = f (σ) cos(cσt) +
g(σ)
sin(cσt),
cσ
which on inversion gives the solution as
∞
1 ∞
u(r, t) =
y f (y) cos(cyt) J0 (yr) dy +
y g(y) sin(cyt) J0 (yr) dy.
c 0
0
(7.63)
7.10.4. Hyperbolic or Parabolic Equations in Terms of Green’s
Functions. To solve the wave equation in terms of Green’s function, consider
utt − c2 ∇2 u = f (x, t) in D,
(7.64)
subject to the conditions u(x, 0) = 0 = ut (x, 0) and u(x, t) = 0 on the boundary
∂D. The Green’s function G (x, t; x , t ) for this problem satisfies the equation
Gtt − c2 ∇2 G = δ (x − x ) δ (t − t )
in D,
(7.65)
and the conditions G (x, x ; 0) = 0 = Gt (x, x ; 0) and G (x, t; x , t ) = 0 on ∂D.
Multiplying Eq (7.64) by G and Eq (7.65) by u and subtracting, we get
G utt − c2 G ∇2 u − u Gtt + c2 u ∇2 G = G f (x, t) − u δ (x − x ) δ(t − t ).
Integrating this equation with respect to t from 0 to t , we have
t
(G utt − u Gtt ) + c2 u ∇2 G − G ∇2 u dt
0
t
=
G f (x, t) − u δ (x − x ) δ (t − t ) dt,
0
which gives
"t "
(G ut − u Gt ) " −
0
t
0
=
0
t
(Gt ut − ut Gt ) − c2 u ∇2 G − G ∇2 u dt
G f (x, t) − u δ (x − x ) δ (t − t ) dt.
7.10. APPLICATIONS OF THE WAVE OPERATOR
197
Since, in view of the boundary and initial conditions, u(x, 0) = 0 = ut (x, 0), and
G (x, x ; 0) = G (x, x ; t , t ) = 0 = Gt (x, x ; 0) = Gt (x, x ; t , t ), we find that
t
2
c
2
u∇ G − G∇2 u dt =
0
t
G f (x, t) dt − u(x, t ) δ (x − x ) .
0
Integrating this equation over the region D, we obtain
t
2
c
D
2
u∇ G − G∇2 u dt dx
0
t
G f (x, t) dt −
=
D
0
u(x, t ) δ (x − x ) dx.
D
Applying Green’s second identity (1.16) after interchanging the time integral with
the space integral, we get
t
c2
0
t ∂G
∂u u
−G
dS dt =
G f (x, t) dx dt − u (x , t ) .
∂n
∂n
∂D
0
D
Since u and G both vanish on the boundary ∂D, we have
u (x , t ) =
t 0
G (x, x ; t, t ) f (x, t) dx dt.
D
Interchanging x and t with x and t , respectively, and noting the symmetry of
G (x, x ; t, t ) with respect to x and x and with respect to t and t , we obtain
t u (x, t) =
0
G (x, x ; t, t ) f (x , t ) dx dt ,
D
(7.66)
where D is the domain of integration in the x coordinate system. For the general
problem of Eq (7.64), subject to the conditions
u(x, 0) = φ(x),
ut (x, 0) = ψ(x),
α
∂u
+ β u(x, t) = g(x, t) on ∂D ,
∂n
the solution is given by
t u (x, t) =
G (x, x ; t, t ) f (x , t ) dx dt
0
∂G (x, x ; t, 0)
+
ψ(x ) G (x, x ; t, 0) − φ(x )
dx
∂t
D
c2 t
+
G (x, x ; t, t ) g (x , t ) dS dt , α = 0,
α 0
∂D
D
198
7. HYPERBOLIC EQUATIONS
t G (x, x ; t, t ) f (x , t ) dx dt
∂G (x, x ; t, 0)
+
ψ(x ) G (x, x ; t, 0) − φ(x )
dx
∂t
D
t c2
∂G (x, x ; t, t )
−
g (x , t ) dS dt , β = 0
β 0
∂n
∂D
(7.67)
Similarly, for the diffusion problem ut − k ∇2 u = f (x, t) in D, subject to the
∂u
conditions u(x, 0) = φ(x), α
+ β u(x, t) = g(x, t) on ∂D, α = 0, the solution
∂n
in terms of its Green’s function is given by
or u (x, t) =
D
0
t G (x, x ; t, t ) f (x , t ) dx dt
u (x, t) =
D
0
φ(x ) G (x, x ; t, 0) dx
t k
+
G (x, x ; t, t ) g (x , t ) dS dt ,
α 0
∂D
+
D
α = 0,
or
t G (x, x ; t, t ) f (x , t ) dx dt
u (x, t) =
0
D
φ(x ) G (x, x ; t, 0) dx
t k
∂G (x, x ; t, t )
−
g (x , t ) dS dt ,
β 0
∂n
∂D +
(7.68)
D
β = 0.
We will establish (7.68) by using the Laplace transform technique in §7.11 given
below.
We remark that the singularity solutions are synonymous to fundamental solutions
that are the free space Green’s functions, or Green’s functions for the whole space.
∂ ∂ On the other hand, the solutions of the equation L
,
= δ (x − x ) δ(t − t )
∂t ∂x
are known as the causal Green’s functions, or simply Green’s functions. We will not
discuss the fundamental solutions for partial differential operators; details about these
solutions are available in Friedlander [1982], Kythe [1996] and Vladimirov [1984].
7.11. Laplace Transform Method
This method can be used effectively to solve parabolic or hyperbolic equations. For
∂
∂
example, consider the equation L
−∇2 u = f (x, t), where L
is a linear
∂t
∂t
7.11. LAPLACE TRANSFORM METHOD
199
differential operator for t > 0. Applying the Laplace transform with respect to t to
this equation and assuming homogeneous initial condition, we get
L(s) − ∇2 u = F (x, s).
(7.69)
The transformed operator is self-adjoint. The Green’s function satisfies the equation
L(s) − ∇2 G = e−st δ(x − x ).
(7.70)
Multiplying (7.69) by G and (7.70) by u we obtain
u ∇2 G − G ∇2 u = GF (x, s) − u e−st δ(x − x ).
Integrating both sides of this equation over the region under consideration and using
the divergence theorem (§1.7), we have
∂G
∂u
u ∇2 G − G ∇2 u dx =
u
−G
dS
∂n
∂n
D
∂D
=
GF (x, s) − u e−st δ(x − x ) dx,
D
which yields
st
GF (x, s) dx −
u (x ) = e
D
∂u
∂G
u (x )
−G
dS .
∂n
∂n
∂D
The right side of this equation consists of either known functions or the expressions
disappears because of the boundary condition. Thus,
For the Dirichlet problem G = 0 and u is known on ∂D.
∂G
∂u
For the Neumann problem,
= 0 and
is known on ∂D. Hence, the solution
∂n
∂n
is known for the transform domain, and the solution to the problem can be found by
carrying out the Laplace inversion.
∂u
In the case of the Robin problem, let u + λ
= φ(xB , t) be satisfied on
∂n
the boundary, where xB is any point on the boundary. The corresponding Green’s
function is the solution of (7.70) with φ(xB , t) = 0. Hence
∂G
∂u
u
−G
dS
∂n
∂n
∂D
∂G
∂u
∂u
=
φ(xB , s) − λ
−G
dS
∂n
∂n
∂D ∂n
(7.71)
∂G
∂u
∂G ∂u
=
φ(xB , s) − λ
+λ
dS
∂n
∂n ∂n
∂D ∂n
∂G
=
φ(xB , s)
dS.
∂n
∂D
200
7. HYPERBOLIC EQUATIONS
Once again the right side of Eq (7.71) is defined in terms of known functions. The
final solution can be found by inverting the Laplace transform.
Example 7.1. Solve the boundary value problem
utt − c2 uxx = e−|x| sin t,
u(x, 0) = 0,
ut (x, 0) = e−|x| .
Applying the Laplace transform to this problem, we get
uxx (x, s) −
s2
2 + s2
u(x, s) = − 2
e−|x| .
2
c
c (1 + s2 )
Green’s function for this problem in the transform domain is
G(x, x ; s) = −
c −s|x−x |/c
e
,
2s
where s is the variable of the transform. This gives the solution in the transform
domain as
∞
2 + s2
u(x, s) =
e−|x |−s|x−x |/c dx .
2cs (1 + s2 ) −∞
Case 1. For x < 0, we have
u(x, s) =
2 + s2
2cs (1 + s2 )
x
−∞
∞
+
ex −s(x−x )/c dx +
0
ex +s(x−x )/c dx
x
e−x +s(x−x )/c dx
0
x
2 + s2
s e − c esx/c
2
2
2
s (1 + s ) (s − c )
1
1 x
ex
=B
−
e −
s−c s+c
(1 + c2 ) (1 + s2 )
2
1
1 cs
+
−B
+
−
esx/c ,
2
2
cs
s−c s+c
(1 + c ) (1 + s )
=
where B =
2 + c2
. Hence, after inverting the solution for x < 0 is
2c (1 + c2 )
ex sin t
u(x, t) = B ect+x − e−ct+x −
1 + c2
2
c
+ H(ct + x)
− B ect+x + e−ct−x −
cos
(t
−
x/c)
.
c
1 + c2
7.12. QUASIOPTICS AND DIFFRACTION
201
Case 2. For x > 0, the solution in the transform domain is
u(x, s) =
2 + s2
2cs (1 + s2 )
0
−∞
∞
+
ex −s(x−x )/c dx +
x
e−x −s(x−x )/c dx
0
e
−x +s(x−x )/c
dx
x
2 + s2 −x
s e − c e−sx/c
s (1 + s2 )
1
1 −x
e−x
=B
−
e −
s−c s+c
(1 + c2 ) (1 + s2 )
2
1
1 cs
+
−B
+
−
e−sx/c ,
cs
s−c s+c
(1 + c2 ) (1 + s2 )
=
which after inverting gives the solution for x > 0 as
e−x sin t
u(x, t) = B ect−x − e−ct−x −
1 + c2
2
c
+ H(ct − x)
− B ect−x + e−ct+x −
cos
(t
−
x/c)
.
c
1 + c2
7.12. Quasioptics and Diffraction
It is known that the electromagnetic radiation is represented by rays of light, and
reflections at plane and curved mirrors follow the rule: θin = θout , while refraction
follows Snell’s law: n1 sin θ1 = n2 sin θ2 , where n1 , n2 are different refraction
indices of the two media and θ1 , θ2 the angles from the normal of the incident and
refracted waves, respectively. In some problems diffraction is included, and one such
problem is discussed below.
7.12.1. Diffraction of Monochromatic Waves. The phenomenon of diffraction of waves plays an important role not only in the case of light waves but in all
kinds of waves, since diffraction is the result of all waves of given wavelength along
all obstructing paths. In the case of light waves it affects the visibility of objects
around us. For example, we cannot see the motion of a single molecule even under a
microscope, because it is much smaller than a polystyrene bead or a grain of pollen.
The objects are seen through a microscope since they block some of the light that
illuminates them from below as we look down upon them. If an object is smaller than
one half the wavelength of light, the light around it eliminates most of the shadow it
casts and we do not see it. However, some objects which are fluorescent and emit light
are visible and refraction no longer makes them invisible. For example, individual
DNA molecules are visible only when they are tainted with fluorescent dyes, but they
are not visible in normal microscope light because the width of the DNA helix is
much smaller than the wavelength of light.
202
7. HYPERBOLIC EQUATIONS
If we consider the contribution of an infinitely small neighborhood around a path
of waves, we obtain what is known as a wave front, and if we add all wave fronts (i.e.,
integrate over all paths from the source to the detector which is a point on the screen),
we obtain a procedure to study the phenomenon of diffraction. A diffraction pattern
can be determined by the phase and the amplitude of each wavefront, which enables
us to find at each point the distance for each source on the incoming wavefront. If
this distance for each source differs by an integer multiple of wavelength, then all
wavefronts will be in phase, which will result in constructive interference. But if
this distance is an integer plus one half of the wavelength, there will be a complete
destructive interference. Usually such values of minima and maxima are determined
to explain diffraction results.
The simplest type of diffraction occurs in a 2-D problem, like the water waves
since they propagate only on water surface. In the case of light waves we can neglect
one dimension if the diffracting object extends in the direction of light over a distance
far greater than its wavelength. However, in the case of light emanating through small
circular holes we must consider the 3-D problem. Some qualitative observations about
diffraction are as follows:
(i) The angular spacing in diffraction patterns is inversely proportional to the
dimension of the object causing diffraction, i.e., the smaller the diffracting object, the
wider the resulting diffraction pattern, and conversely. This pattern depends on the
sines of the angles.
(ii) The diffraction angles are invariant under scaling, i.e., they depend only on
the ratio of the wavelength to the size of the diffracting object.
(iii) If the diffracting object has a periodic structure, like a diffracting grating, the
diffraction becomes sharper.
The problem of computing the pattern of a diffracted wave depends on the phase
of each of the simple sources on the incoming wave front. There are two types of
diffraction: One is known as the Fraunhofer diffraction or far-field diffraction, in
which the point of observation is far from the diffracting object. It is mathematically
simpler that the other type, known as the Fresnel diffraction or near-field diffraction,
in which the point of observation is closer to the diffracting object.
The best way to study diffraction of light waves is presented by multi-slit arrangements which can be regarded as multiple simple wave sources if the slits are
very narrow. A light wave going through a very narrow slit extends to infinity in
one direction and this situation results in a 2-D problem. The simplest case is that
of a two-slit arrangement, where the slits are a distance a apart (Fig. 7.6.). Then the
maxima and minima in the amplitude are determined by the path difference to the
first slit and to the other slit.
(a) Fraunhofer Approximation. If the observer is far away from the slits,
the difference in path lengths to the two slits as seen from the image is ∆s = a sin θ.
The maxima in the intensity occur if this path length difference ∆s is an integer
7.12. QUASIOPTICS AND DIFFRACTION
203
multiple of wavelength, i.e., if a sinθ = m λ, where m is the number of the order of
each maximum, a the distance between the slits and θ the angle at which constructive
interference occurs.
z
slit
θ
nθ
slit
y′
slit
a si
− a/2
( x, 0 , z )
a/2
r
z
0
θ
y
x′
x
Fig. 7.6. Fraunhofer Diffraction.
Fig. 7.7. Fresnel Approximation.
A
cos(k r − ωt + φ), where k = 2π/λ,
r
ω the frequency and φ the phase of the wave at the slits. The function E(r) is the
A
real part of the complex wave Ψ(r) which is defined as Ψ(r) = ei (kr−ωt+φ) . The
r
amplitude of this complex wave is given by |Ψ(r)| = A/r. For M slits the radial
wave at a point x on the screen which is at a distance l away is given by
√2
2
M
−1
ei k l +(x−ma)
i (−ωt+φ)
EM (r) = A e
.
(7.72)
l2 + (x − ma)2
m=0
1/2
(x − ma)2
x2
max m2 a2
Since l2 + (x − ma)2 = l 1 +
≈l+
−
+
, we
2
l
2l
l
2l
can neglect the term of the order a2 /(2l), and then the complex wave is defined by
M
−1
A
x2 Ψ(r) =
exp i k l +
− ωt + φ
e−i kxma/l
l
2l
m=0
A radial wave is represented by E(r) =
M kax
(7.73)
2l ,
kax
sin
2l
where the above sum is a finite geometric series. If we set I0 = A/l, the intensity of
this wave is given by
A
=
exp i
l
x2 k l+
− ωt + φ
2l
sin
sin M kax 2l
I(x) = |Ψ|2 = I0
.
kax
sin
2l
(7.74)
204
7. HYPERBOLIC EQUATIONS
(b) Fresnel Approximation. Let Ψ(r) define a monochromatic complex
plane wave of wavelength α passing through a slit of width a centered at the point
(x, 0, z). The 3-D situation can be reduced to a 2-D problem if we consider a
coordinate system (x , y ) through the point (x, 0, z) so that the slit is defined by
{−a/2 < x < a/2, −∞ < y < ∞} (see Fig. 7.7). Notice that this slit is located
on the (x, z)-plane and the light beam is diffracted along the y -axis. This complex
wave is defined by
i
Ψ(r) =
Ψ0 e−i kr d(slit),
(7.75)
slit rλ
where Ψ0 is a constant, and
r = (x − x )2 + y 2 + z 2 = z
1+
(x − x )2 + y 2
(x − x )2 + y 2
≈z+
.
2
z
2z
Since the factor 1/r in (7.75) is non-oscillatory, its contribution to the magnitude of
the intensity is very small compared to the exponential factor. We will, therefore,
approximate 1/r by 1/z. Then (7.75) becomes
i Ψ0
Ψ(x) =
λz
=
a/2
∞
e
−a/2
i Ψ0 −i kz
e
λz
−i k z+
(x − x )2 + y 2 2z
dx dy −∞
a/2
2
e−i k(x−x )
−a/2
/(2z)
dx
∞
e−i ky
2
/(2z)
dy −∞
(7.76)
a/2
2
i Ψ0 −i kz −i kx2 /(2z)
2πz
=
e
e
ei kxx /z e−i kx /(2z) dx
λz
ik
−a/2
a/2
2
i −i kz −i kx2 /(2z)
= Ψ0
e
e
ei kxx /z e−i kx /(2z) dx ,
zλ
−a/2
∞
2πz
. Since the quanik
−∞
tity kx2 /z is small in the Fresnel approximation, we can approximate the factor
2
e−i kx /(2z) ≈ 1, and the terms e−i kz can also be neglected if we take z as a very
small positive quantity, say z = ε. Then (7.76) reduces to
i −i kx2 /(2ε) a/2 i kxx /ε Ψ(x) = Ψ0
e
e
dx
ελ
−a/2
i −i kx2 /(2ε) ei kax/(2ε) − e−i kax/(2ε)
= Ψ0
e
ελ
i kx/ε
kax
sin
i −i kx2 /(2ε)
i −i kx2 /(2ε)
kax
2ε
= Ψ0
e
= Ψ0
e
sinc
kx
ελ
ελ
2ε
2ε
where we have used the formula
e
−i ky 2 /(2z)
dy =
7.13. EXERCISES
= Ψ0
πax i −i kx2 /(2ε)
e
sinc
,
ελ
ελ
205
where k = 2π/λ.
(7.77)
Since limx→0 sinc(x) = 1, the sinc term in the above expression is almost equal to 1
in this near-field approximation, and hence, can be neglected. This leads to
2λ Ψ0
i −i kx2 /(2ε)
Ψ(x) =
e
,
π
ελ
which defines a diffracted wave similar to (2.26) in Example 2.8. The intensity of the
πax 2
|Ψ0 |
diffracted wave (7.77) is given by I = I0 sinc
, where I0 =
.
ελ
ελ
7.13. Exercises
7.1. Use Green’s function to solve the following problem (see §7.1):
utt − c2 uxx = e−x sin t, u(x, 0) = 0, ut (x, 0) = e−x for x > 0, t > 0,
u(0, t) = 0, and lim u(x, t) = 0.
x→∞
1
Ans. u(x, t) = B e−(x−ct) − e−(x+ct) −
e−x sin t − H(ct − x)
1 + c2
1
x c2 + 2
× B e−(x−ct) + e(x−ct) −
cos
t
−
, where B =
.
2
1+c
c
2c (1 + c2 )
7.2. For the wave operator show that
∞
G3 (x, y, z, t; x , y , z , t ) dz = G2 (x, y, t; x , y , t ) .
−∞
Hint. From (7.19), we have
∞
F (x, y, t; x , y , t ) =
G3 (x, y, z, t; x , y , z , t ) dz −∞
!
1
2
2
2
∞ δ t − t −
(x − x ) + (y − y ) + (z − z )
1
c
!
=
dz 4πc2 −∞
2
2
2
(x − x ) + (y − y ) + (z − z )
!
1
2
2
∞ δ t − t −
(x − x ) + (y − y ) + u2
1
c
!
=
du,
4πc2 −∞
2
2
(x − x ) + (y − y ) + u2
where u = z − z . Letting r 2 + u2 = c2 v 2 , r2 = (x − x ) + (y − y ) , and
noting that the integrand is an even function of u, we get
∞
1
δ (t − t − v)
1
√
!
F (x, y, t; x , y , t ) =
dv =
,
2πc 0
2
c2 v 2 − r 2
2πc c2 (t − t ) − r 2
2
2
206
7. HYPERBOLIC EQUATIONS
for t − t > r/c. Hence, F (x, y, t; x , y , t ) = G2 (x, y, t; x , y , t ).
7.3. Show that for the wave operator
∞
−∞
G2 (x, y, t; x , y , t ) dz = G1 (x, t; x , t ).
Hint. Follow the method in Exercise 7.2.
7.4. Derive (7.26) by using the Bernoulli’s separation method.
∂2
∂2
∂
∂
+
+ 2α
+ 2β
in
2
2
∂x
∂y
∂x
∂y
the region {0 < x < a, 0 < y < b} subject to Dirichlet boundary conditions.
7.5. Find Green’s function for the operator L ≡
Ans. G (x, y; x , y )
= −4 eα(x −x)+β (y −y)
m=1 n=1
Ans. u(x, t) = T0 1 −
x
a
mπx
nπy
mπx
nπy sin
sin
sin
a
b
a
b .
π 2 (m2 b2 + n2 a2 ) + a2 b2 (α2 + β 2 )
∞ ∞ ab sin
− 2T0
∞
1 −n2 π2 kt/a2
nπx
e
sin
.
nπ
a
n=1
7.6. Solve Dirichlet problem for Laplace’s equation in R3 .
Solution. Taking x1 = 0 = (0, 0, 0), Green’s function becomes symmetrical
about the origin. Hence, the differential equation becomes
∂ 2 G 2 ∂G
δ(r)
+
+ k2 G =
,
2
∂r
r ∂r
4πr2
lim G = 0,
r→∞
and the additional that the flux of G across a small sphere with center at the
origin be equal to −1. Substituting G = w/r, the differential equation for G
d2 w
δ(r)
becomes 2 + k 2 w =
. Applying Fourier sine transform to this equation,
dr
4πr
√
2
α
we obtain
α c0 + k 2 − α2 ws = 2π
, where c0 = limr→0 w. Thus,
π
4π
2
1
α
ws =
c0 −
. On inverting the sine transform, we have
π
4π α2 − k 2
1 −ikr
1 e−ikr
w = c0 −
e
, and therefore, G(r) = c0 −
. Using the flux
4π
4π
r
dG condition, we get − lim
dS = −1, where S is a sphere of radius ε
ε→0 S
dr r=ε
with center at the origin, and dS = ε2 sin φ dφ dθ is the infinitesimal element of
the surface. Then
1
1 2π π −ikr 1
lim c0 −
e
+ 2 ε2 sin φ dφ dθ = −1,
ε→0
4π 0
ε ε
0
7.13. EXERCISES
207
1
whence we have c0 −
4π = −1, which gives c0 = 0. Hence, G(r) =
4π
e−ikr
−
.
4πr
7.7. Use double Fourier transform to solve the telegrapher’s equation
utt + a ut + b u = c2 uxx , −∞ < x, t < ∞ subject to the conditions u(0, t) =
f (t), ux (0, t) = g(t), ∞ < t < ∞, where a, b, c are constants and f (t) and
g(t) are arbitrary functions of t.
∞
1
g(α)
Ans. u(x, t) =
f(α) cos(µx) +
sin(µx) ei αt dα , where
2π
µ
−∞
µ2 = α2 − i aα − b /c2 .
7.8. Show that Green’s function G(x, t) for the Bernoulli-Euler problem for an
elastic foundation, defined by
b Gxxxx + k G + m Gtt = a δ(x) δ(t), x ∈ R, t > 0, and subject to the initial
conditions G(x, 0) = 0 = Gt (x, 0), where b = EI, k and a are constants, is
given by
∞
√
a
sin pt i αx
G(x, t) =
e
dα, p = a2 α4 + ω 2 , a2 = b/m and ω 2 =
2πm −∞ p
k/m.
Hint. Use double Laplace and Fourier transforms.
7.9. Find Green’s function for 1-D Klein-Gordon’s equation
utt − c2 uxx + a2 u = f (x, t), x ∈ R, t > 0, where a and c are constants, and
subject to the initial conditions u(x, 0) = 0 = ut (x, 0) for all x ∈ R.
Hint. Apply Laplace transform followed by Fourier transform to Gtt −c2 Gxx +
1
1
√
a2 G = δ(x) δ(t), giving G(α,
s) = √
.
2π s2 + c2 α2 + a2
−1/2
∞
2
1
a2
a
1
2
Ans. G(x, t) =
α + 2
sin ct α2 + 2 = H(ct −
2πc −∞
c
c
2c
a |x|) J0
c2 t2 − x2 . Note that the above Green’s function reduces to that
c
for the wave equation in the limit as a → 0.
7.10. Show that Green’s function for the differential equation L[u](t) = f (t), t >
d2
d
0, where L = 2 + 2λ + ω0 2 , ω0 > λ > 0, is the differential operator for a
dt
dt
(damped) harmonic oscillator subject to an external force and initially at rest after
being subjected to a sudden impulse at t = 0, i.e., when f (t) = δ(t), is given by
H(t) −λt
G(t) =
e
sin αt, α = ω0 2 − λ2 .
α
Hint. Use the Laplace transform method to solve L[G](t) = δ(t).
7.11. Solve utt = c2 uxx , 0 < x < l, subject to the boundary conditions u(0, t) =
u1 , u(l, t) = u2 for t > 0, where u1 , u2 are prescribed quantities, and the initial
208
7. HYPERBOLIC EQUATIONS
conditions u(x, 0) = g(x), ut (x, 0) = h(x) for 0 ≤ x ≤ l.
x
Ans. u(x, t) = U (x) + v(x, t), where U (x) = u1 + (u2 − u1 ) , describes the
l
steady-state solution (static deflection), and v(x, t) is the solution of the problem
in Exercise 4.9, with f (x) replaced by g(x) − U (x).
7.12. Solve utt = c2 (uxx + uyy ) in the rectangle R = {(x, y) : 0 < x <
a, 0 < y < b}, subject to the condition u = 0 on the boundary of R for
t > 0, and the initial conditions u(x, y, 0) = f (x, y), ut (x, y, 0) = g(x, y).
This problem describes a vibrating rectangular membrane. Ans. u(x, y, t) =
∞
mπx
nπy
(Amn cos λmn t + Bmn sin λmn ) sin
sin
,
a
b
m,n=1
where
Amn =
4
ab
b
a
f (x, y) sin
0
b
a
mπx
nπy
sin
dx dy,
a
b
m2
n2
for m, n = 1, 2, . . . ; and the eigenvalues are λmn = cπ
+ 2.
2
a
b
7.13. Solve the problem of transverse vibrations of a semi-infinite string initially at
rest, defined by 1-D wave equation utt = c2 uxx , 0 < x < ∞, t > 0, and subject
to the boundary and initial conditions: u(0, t) = U0 f (t), t ≥ 0; u(x, t) → 0 as
x → ∞, t ≥ 0; and u(x, 0) = 0 = ut (x, 0) for 0 ≤ x < ∞, where U0 is a
constant.
x x
Hint. Use Laplace transform. Ans. u(x, t) = H t −
f t−
.
c
c
Bmn =
4
abλmn
0
mπx
nπy
sin
dx dy,
a
b
g(x, y) sin
0
0
7.14. Solve uxx − utt = e−a π t sin aπx, subject to the conditions u(x, 0) = 0,
ut (x, 0) = 0, and u(0, t) = u(1, t) = 0, where a is an integer.
2 2
1
Ans. 2 2
cos aπt − e−a π t − aπ sin aπt sin aπx.
2
2
a π (1 + a π )
2
2
8
Elliptic Equations
We will derive Green’s functions for Laplace’s and Helmholtz’s equations. In previous sections we have used a source (of strength +1) for parabolic and hyperbolic
equations, but now we will use a sink (of strength −1) for elliptic equations. The
method is to first develop Green’s function G (x) for the sink (source) at the origin,
and then obtain the general Green’s function G (x − x ) ≡ G (x; x ) for the sink
(source) at a point x = 0 by simply replacing x in G(x) by x − x . Note that in
R the Laplacian operator is an ordinary differential operator, which has already been
studied in Chapter 3. We will, therefore, determine Green’s function only in R2 and
R3 by different methods.
8.1. Green’s Function for 2-D Laplace’s Equation
In R2 , in order to solve Laplace’s equation ∇2 u = 0, we use polar cylindrical
coordinates and have
1 ∂ ∂u δ(r)
r
=
,
r ∂r
∂r
2πr
r 2 = x2 + y 2 .
Method 1. (By direct method) Multiply both sides of this equation by r and
∂u
1
integrate from 0 to r, to get
=
, which gives
∂r
2πr
u=
1
log r + B.
2π
Since it is not possible to normalize u by requiring that it must vanish as r → ∞,
we normalize it such that u = 0 at r = 1, which results in B = 0. Hence, Green’s
function in this case with singularity at the origin is given by
G(x) =
1
1
log r =
log x,
2π
2π
209
(8.1)
210
8. ELLIPTIC EQUATIONS
and for an arbitrary singular point x by
G(x − x ) =
1
log x − x .
2π
(8.2)
Method 2. (Based on the physical concepts) Since the operator ∇2 is invariant
under a rotation of the coordinate axes, we seek a solution that depends only on
r = x. For r > 0, the function G(r) satisfies the homogeneous equation ∇2 G = 0,
which in R2 , in polar cylindrical coordinates, is
1 ∂ ∂G r
= 0 for r = 0, where r = |x| = x2 + y 2 .
r ∂r
∂r
This equation has a solution G(r) = A ln r + B. To evaluate A, we consider the flux
across a circle Cε of radius ε with center at the origin. Then
∂G
−
ds = −2πA = −1,
∂r r=ε
Cε
which gives A = 1/2π. If we normalize as before, we have B = 0, and Green’s
function is the same as (8.1). If the source is at x , then Green’s function is the same
as given by (8.2).
Method 3. (Method of images) The problem of finding Green’s function
G(x; x ) = G(x, y; x , y ) inside some region D bounded by a closed curve Γ with
the homogeneous boundary condition G = 0 on Γ amounts to that of finding the
electrostatic potential due to a point charge at the point (x , y ) inside a grounded
conductor in the shape of the boundary Γ. Consider, for example, the region D to be
the half-plane x > 0. Then Green’s function is given by the point charge of strength
−1 at (x , y ) together with an equal but opposite charge (of strength +1) at the point
(x , −y ) which is the image of the point (x , y ) in the x-axis (Fig. 8.1). Thus,
Green’s function for the right half-plane
such that G = 0 on the x-axis and G → 0
as r → ∞, where r = |x − x | = (x − x )2 + (y − y )2 , is given by
G(x, y; x , y ) =
1
ln
4π
(x − x )2 + (y − y )2
.
(x − x )2 + (y + y )2
(8.3)
Since the algebraic sum of the charges over the entire (x, y)-plane is zero, the condition
G → 0 as r → ∞ holds. However, the method of images cannot be used to find
Green’s function for the upper half-plane y > 0 subject to the condition Gy = 0 on
the x-axis and G → 0 as r → ∞ (see the explanation at the end of §8.12.2: Neumann
problem for the half-space). But we can use the method of images to determine
Green’s function for the quarter-plane x > 0, y > 0, subject to the conditions G = 0
on y = 0 and Gy = 0 on x = 0. As seen in Fig. 8.1, we have the charge of strength
+1 at each (x , −y ) and (−x , y ), and the charge of strength −1 at each (x , y ) and
(−x , −y ). Thus, Green’s function for the quarter-plane is
G(x, y; x , y ) =
1
ln
4π
[(x − x )2 + (y − y )2 ][(x + x )2 + (y + y )2 ]
. (8.4)
[(x − x )2 + (y + y )2 ][(x + x )2 + (y − y )2 ]
8.2. 2-D LAPLACE’S EQUATION IN A RECTANGLE
y
.
(−x’, y’ )
strength
+1
211
.
strength
−1
(x’, y’ )
x
0
.
(−x’,− y’)
.
strength
−1
strength
+1
(x’,− y’)
Fig. 8.1.
8.2. 2-D Laplace’s Equation in a Rectangle
Consider ∇2 u = 0 in the rectangle R = {0 < x < a, 0 < y < b}. Then Green’s
function G (x, y; x , y ) for this rectangle satisfies the equation
∇2 G (x, y; x , y ) = δ (x − x ) δ (y − y ) ,
(8.5)
subject to the boundary conditions
G (0, y; x , y ) = G (a, y; x , y ) = G (x, 0; x , y ) = G (x, b; x , y ) = 0.
The double Fourier series representations for G (x, y; x , y ) and δ (x − x ) δ (y − y )
are of the form
∞
G (x, y; x , y ) =
∞
Amn sin
m=0 n=0
∞
mπx
nπy
sin
,
a
b
∞
δ (x − x ) δ (y − y ) =
Bmn
m=0 n=0
mπx
nπy
sin
sin
,
a
b
(8.6)
and satisfy the prescribed boundary conditions. To determine Amn , we first find
Bmn :
4
mπx
nπy Bmn =
sin
sin
.
ab
a
b
Then substituting (8.6) into Eq (8.5), we get
∞
∞
−
Amn
m2 π 2
a2
m=0 n=0
=
4
ab
∞
+
n2 π 2 mπx
nπy
sin
sin
2
b
a
b
∞
sin
m=0 n=0
mπx
nπy mπx
nπy
sin
sin
sin
.
a
b
a
b
212
8. ELLIPTIC EQUATIONS
Hence,
Amn = −
4ab
mπx
nπy sin
sin
,
m2 π 2 b2 + n2 π 2 a2
a
b
and
G (x, y; x , y )
∞
=−
m=0
∞
4ab
mπx
nπy mπx
nπy
sin
sin
sin
sin
.
2
2
2
2
2
2
m π b +n π a
a
b
a
b
n=0
(8.7)
8.3. Green’s Function for 3-D Laplace’s Equation
The functions u which satisfy Laplace’s equation ∇2 u (x) = δ (x) in a domain D
subject to homogeneous boundary conditions are known as Green’s functions for
Laplace’s equation.3 We will give some basic methods for finding Green’s function
for this operator.
Method 1. (By direct method) Assuming that the source point x is at the
origin, Green’s function G(x) ≡ G(x; 0) for the Laplacian ∇2 in R3 satisfies the
equation
∂2u ∂ 2u ∂ 2u
+ 2 + 2 = δ(x) δ(y) δ(z).
∂x2
∂y
∂z
(8.8)
Since the problem has radial symmetry, we transform Eq (8.8) into spherical coordinates, which becomes
∂ 2 u 2 ∂u
δ(r)
+
=
.
∂r2
r ∂r
4πr2
Multiplying both sides of this equation by r2 , we get
∂ 2 ∂u δ(r)
r
=
.
∂r
∂r
4π
Integrating both sides of this equation from 0 to r, we find that
r2
∂u
1
=
,
∂r
4π
or u = −
1
,
4πr
where the arbitrary constant due to integration is taken to be zero, in order to satisfy
the requirement that u must vanish as r → ∞. Hence, Green’s function in this case
with singularity (source) at the origin is given by
G(x) = −
1
1
=−
,
4πr
4π|x|
(8.9)
3 In some books the equation is taken as −∇2 u(x) = δ(x), to signify the source of
strength +1 at the origin. This practice is prevalent in physics, and to get the results in
this case Green’s functions derived here must be multiplied by −1.
8.3. GREEN’S FUNCTION FOR 3-D LAPLACE’S EQUATION
213
and for an arbitrary singular point x by
G(x − x ) = −
1
1
=−
.
4π r
4π |x − x |
(8.10)
Method 2. (Based on the physical concepts) Again, since the operator ∇2 is
invariant under
a rotation of the coordinate axes, we seek a solution that depends
only on r = x. For r > 0, the function G(r) satisfies the homogeneous equation
∇2 G = 0, which in the spherical coordinates is
1 ∂ 2 ∂G r
= 0.
r 2 ∂r
∂r
This equation has a solution G(r) = A/r + B. If we require this solution to vanish at
infinity, then B = 0, and thus, G(r) = A/r. To determine A, we consider a negative
source of unit strength at x = 0, which is given by the flux across a sphere of radius
ε as
2π π
∂G
A 2
−
dS =
ε sin φ dφ dθ = 4πA = −1,
(8.11)
2
∂r
ε
∂Sε
0
0
where ∂Sε is the surface of the sphere Sε . Physically, the above equation represents the
conservation of the charge, i.e., the flux of the electric field through the closed surface
∂Sε (of area 4πε2 ) is equal to the charge in the interior of Sε . Using Eq (8.11), we find
that A = −1/(4π), and thus, Green’s function of the three-dimensional Laplacian is
given by
1
1
G(x) = −
=−
.
4πr
4π|x|
If the source is at x , Green’s function is the same as in (8.10).
8.3.1. Laplace’s Equation in a Rectangular Parallelopiped. Green’s
function for Laplace’s equation in a rectangular parallelopiped D, defined by x =
±a/2, y = ±b/2, z = ±c/2, subject to the condition u = 0 on the boundary ∂D is
determined by constructing the lattices which correspond to the given domain D and
have the vertices (k + 12 )a, (m + 12 )b, (n + 12 )c , k, m, n = 0, ±1, ±2, . . . , which
reflect the point (x , y , z ) respectively across the lattice-planes.
This generates a
system of points ka + (−1)k x , mb + (−1)m y , nc + (−1)n z . We assume that at
each of these points a unit mass is connected, which is positive or negative according
as k + m + n is even or odd. Thus, the potential of such a mass distribution is zero
in the lattice-planes, since the contributions of the individual mass units cancel each
other there, and we get the following expression for Green’s function:
G(x, y, z; x , y , z ) =
1
4π
∞
∞
∞
(−1)k+m+n
,
N (k, m, n; x, y, z; x , y , z )
k=−∞ m=−∞ n=−∞
(8.12)
214
8. ELLIPTIC EQUATIONS
N (k, m, n; x, y, z; x , y , z )
2 2 2
= ka + (−1)k x − x + mb + (−1)m y − y + nc + (−1)n z − z .
where
The convergence in (8.12) can be at least conditional, and so we denote an expression of the form φ(k + 1) − φ(k), where φ(k) is any function of k, by ∆k φ(k). Then
for fixed k and m, and omitting the factor (−1)k+m , we can write the summation
over n in (8.12) as
1
1
∆n =−
∆n ,
N (k, m, n)
N (k, m, n)
n=±1,±2,...
n=0,±2,±4,...
N (k, m) =
since lim N (k, m, n) = ∞. We apply the same transformation to the sums over
|n|→∞
k and m. Since lim N (k, m) = 0, we have
|m|→∞
N (k) =
∆m N (k, m) = −
m=±1,±2,...
∆m N (k, m),
m=0,±2,±4,...
and so
G(x, y, z; x , y , z ) =
1
4π
∆k N (k) = −
k=±1,±3,...
1
4π
∆k N (k),
k=0,±2,±4,...
(8.13)
because lim N (k) = 0. Combining these results we get
|k|→∞
G(x, y, z; x , y , z ) = ±
1
4π
∆k ∆m ∆n m
k
n
1
N (k, m, n)
,
where each summation index runs over either all the even or all the odd integers from
−∞ to ∞, and the plus or minus sign in the beginning on the left side is chosen
according as the summation is taken over all even or all odd integers. The results
(8.12) and (8.13) are valid only if no N (k, m, n) vanishes.
Representation in terms of Riemann ϑ-functions. The sum (8.12)
can be represented as an integral overthe θ-functions, as follows: Since the Laplace
∞
2
2
1
2
transform of e−t is given by √
e−st dt = √ , we substitute for s the
π 0
s
expression N (k, m, n; x, y, z, x , y , z ) and get
∞
2
1
G(x, y, z; x , y , z ) = 3/2
∆k ∆m ∆n
eN t dt.
2π
0
m n
k
Interchanging the summations and integration, we find that
∞
2
1
G(x, y, z; x , y , z ) =
∆k ∆m ∆n eN t dt
2π 3/2 0
m n
k
∞
1
=
f1 f2 f3 dt,
2π 3/2 0
(8.14)
8.3. GREEN’S FUNCTION FOR 3-D LAPLACE’S EQUATION
∞
where
f1 =
215
2 (−1)k exp − t2 ka + (−1)k x − x
,
k=−∞
∞
f2 =
2
(−1)m exp − t2 [mb + (−1)m y − y] ,
m=−∞
∞
f3 =
2
(−1)n exp − t2 [nc + (−1)n z − z] .
n=−∞
We will express the function G in terms of the Riemann ϑ-functions defined by
∞
ei πr
ϑm (ζ, τ ) = ϑ(ζ, τ ) =
2
τ
e2i πνζ .
ν=−∞
But first we will show that the integral in (8.14) converges to zero on the interval
[0, ∞). The proof is as follows: Since
∞
∞ ∞ ∞
2 f1 f2 f3 dt =
f2 f3 exp − t2 ka + (−1)k x − x
dt,
0
0
k=−∞
0
we have
2 k
2
k < ea2 t2 /4
(−1) exp − t ka + (−1) x − x
|k|>p
−a2 t2 /4
<
2e
a2
(−1)k
|k|>p
(−1)k e−a
2 2 2
t k /2
|k|>p
2 2
1 −a2 t2 /4 1
4
e
< 2 e−a t /4 .
k2
p2
pa
Thus, the integral over the interval [0, ∞) converges to zero with increasing p, and
f2 , f3 remain uniformly bounded in the interval [0, ∞). This completes the proof.
Now, to represent Green’s function G in terms of ϑ00 -function, we find that
2at2 i (x − x ) 4a2 t2 i
f1 = e
ϑ00 −
,
π
π
2
2
2
2at
i
(x
+
x
)
4a2 t2 i
−t (x+x )
−e
ϑ00 −
,
π
π
2
2 2
2
2
2bt
i
(y
−
y
)
4b
t
i
f2 = e−t (y−y ) ϑ00 −
,
π
π
2
2
2
2bt
i
(y
+
y
)
4b2 t2 i
− e−t (y+y ) ϑ00 −
,
π
π
2
2 2
2
2
2ct i (z − z ) 4c t i
f3 = e−t (z−z ) ϑ00 −
,
π
π
2
2
2
2ct i (z + z ) 4c2 t2 i
− e−t (z+z ) ϑ00 −
,
.
π
π
−t2 (x−x )2
216
8. ELLIPTIC EQUATIONS
To each we apply the formula
e−i πζ /τ
ϑ00 (ζ, τ ) = √
ϑ00
−i τ
2
ζ
1
,− ,
τ
r
2
2 2
taking the principal value of the square root, and setting qx = e−π /(4a t ) , qy =
2
2 2
2
2 2
e−π /(4b t ) , and qz = e−π /(4c t ) . Then
√ π
x − x πi
x + x πi
ϑ00 −
, 2 2 − ϑ00 −
, 2 2
2at
2a 4a t
2a 4a t
√ ∞
∞
π
k2 −k(x−x )πi/a
k2 −k(x+x )πi/a
=
qx e
−
qx e
2at
k=−∞
k=−∞
√ ∞
2
π
k(x − x )π
k(x + x )π
=
qxk cos
− cos
at
a
a
k=1
√
∞
2
2π
kπx
kπx
=
qxk sin
sin
;
at
a
a
f1 =
k=1
and similar expressions for f2 and f3 . Thus, (8.14) becomes
∞
∞
∞
∞
kπx
kπx
sin
×
a
a
0
k=1 m=1 n=1
mπy
mπy nπz
nπz π k2
m2
n2
× sin
sin
sin
sin
exp − 2
+ 2 + 2
,
b
b
c
c
4t a2
b
c
4
G(x, y, z; x , y , z ) =
abc
1
t2
sin
which, after setting 1/t2 = τ , gives
G(x, y, z; x , y , z ) =
× sin
=
8
abcπ 2
2
abc
∞ ∞
0
∞
∞
k=1 m=1 n=1
exp −
π2 τ
4
k2
m2
n2
+ 2 + 2
2
a
b
c
×
kπx
kπx
kπx
mπy
mπy nπz
nπz sin
sin
sin
sin
sin
sin
a
a
a
b
b
c
c
∞
∞
∞
k=1 m=1 n=1
kπx
kπx
kπx
mπy
mπy nπz
nπz sin
sin
sin
sin
sin
sin
sin
a
a
a
b
b
c
c .
k2
m2
n2
+ 2 + 2
a2
b
c
8.4. HARMONIC FUNCTIONS
217
If τ = 1/t2 , then
G(x, y, z; x , y , z )
∞ 1
x − x πi τ
x + x πi τ
=
ϑ00 −
, 2 − ϑ00 −
, 2
32abc 0
2a
4a
2a
4a
y − y πi τ
y + y πi τ
ϑ00 −
, 2 − ϑ00 −
, 2
2b
4b
2b
4b
z − z πi τ
z + z πi τ
ϑ00 −
, 2 − ϑ00 −
, 2
.
2c
4c
2c
4c
(8.15)
8.4. Harmonic Functions
A real-valued function u(x, y) ∈ C 2 (D) is said to be harmonic in a region D iff
∇2 u = 0 in D. Some properties of harmonic functions in R2 are as follows:
1
1
(i) The function = is harmonic in a region that does
2
r
(x − x0 ) + (y − y0 )2
not contain the point (x0 , y0 ).
(ii) If u(x, y) is a harmonic function in a simply connected region D, then there
exists a conjugate harmonic function v(x, y) in D such that u(x, y) + i v(x, y) is an
analytic function of z = x + iy = (x, y) in D. In view of the Cauchy-Riemann
equations4 ,
(x,y)
v(x, y) − v(x0 , y0 ) =
(−uy dx + ux dy) ,
(x0 ,y0 )
where (x0 , y0 ) = z0 is a given point in D. This property is also true if D is multiply
connected. However, in that case the conjugate function v(x,
y) can be multiplevalued, as we see by considering u(x, y) = log r = log x2 + y 2 defined on a
region D containing the origin which has been indented by a small circle centered at
the origin. Then, in view of the Cauchy-Riemann equations, we get
v(x, y) − v(x0 , y0 ) = tan−1
y
± 2nπ + const,
x
n = 1, 2, . . . ,
which is multiple-valued.
(iii) Since derivatives of all orders of an analytic function exist and are themselves
analytic, any harmonic function will have continuous partial derivatives of all orders,
i.e., a harmonic function belongs to the class C ∞ (D), and a partial derivative of any
order is again harmonic.
(iv) A harmonic function must satisfy the mean-value theorem, where the mean
value at a point is evaluated for the circumference or the area of the circle around
4
The Cauchy-Riemann equations for a function w = f (z) = u(x, y)+iv(x, y), z = x+iy,
which is an analytic on a region D, are ux = vy , uy = −vx .
218
8. ELLIPTIC EQUATIONS
that point. If u is harmonic on a region containing the closed disk B̄(r, z0 ), where
z0 = x0 + iy0 , then
2π
1
u(x0 , y0 ) =
u z0 + r eiθ dθ.
2π 0
(v) In view of the maximum modulus theorem, the maximum (and also the minimum) of a harmonic function u in a region D occurs only on the boundary ∂D.
Theorem 8.1. (Maximum Principle) A nonconstant function that is harmonic inside a bounded region D with boundary Γ and continuous in the
closed region D = D ∪ Γ attains its maximum and minimum values only on
the boundary Γ of the region.
Thus, u has a maximum (or minimum) at z0 ∈ D, i.e., if u(z) ≤ u(z0 ) (or
u(z) ≥ u(z0 )) for z in a neighborhood B(ε, z0 ) of z0 , then u = const in B(ε, z0 ).
(v) The value of a harmonic function u at an interior point in terms of the boundary
∂u
values u and
is given by Green’s third identity (1.19).
∂n
(vi) A harmonic function satisfies the mean-value theorem, according to which
the value of a harmonic function at a point is equal to the average value of the function
on a circle in D with its center at the point.
Proofs of these results are available in textbooks on complex analysis, e.g., Carrier,
Krook, and Pearson [1966]. The functions u that satisfy Laplace’s equation ∇2 u = 0
are known as harmonic functions. We have the following result which is a corollary
of Theorem 2.2.
Theorem 8.2. Green’s function for the Laplacian is symmetric, i.e.,
G(x, y) = G(y, x).
Proof. Since the Laplacian is self-adjoint, its Green’s function is symmetric, by
Corollary 2.2.
8.5. 2-D Helmholtz’s Equation
The Helmholtz operator is defined as (∇2 + k 2 ), where k > 0 is real. Helmholtz’s
equation ∇2 + k 2 u = 0 is also called the reduced wave scattering equation and
k is known as the real (positive) acoustic wave number. Green’s function for this
operator satisfies the equation
(∇2 + k 2 )G(x − x ) = δ(x − x ).
(8.16)
In some books this operator is taken as − ∇2 + k 2 , which signifies a source of
strength +1.
For the operator ∇2 + k 2 in R2 we solve the equation
∇2 + k 2 G(x, y) = δ(x) δ(y).
8.5. 2-D HELMHOLTZ’S EQUATION
219
Since the right side is zero everywhere except at the origin, where a negative source
(sink) of unit strength exists, we find a solution with singularity at the origin such that
∂G
− lim
ds = −1,
(8.17)
ε→0 r=ε ∂n
where r is the distance from the origin. The solution is symmetric with respect to
the origin, and therefore, dependent only on r. Thus, the equation ∇2 + k 2 u = 0
reduces to
∂ 2 G 1 ∂G
+
+ k 2 G = 0 for r = 0.
∂r2
r ∂r
Its solution is
(1)
(2)
G(r) = AH0 (kr) + BH0 (kr),
(1)
(2)
where H0 (kr) = J0 (kr) + iY0 (kr) and H0 (kr) = J0 (kr) − iY0 (kr) are the
Hankel functions of the first and second kind, respectively. Applying the condition
(8.17), we find that
∂ ∂G
(1)
(2)
− lim
ds = − lim 2πε
AH0 (kr) + BH0 (kr)
ε→0 r=ε ∂n
ε→0
∂r
r=ε
= lim 2πε k (A + B) J1 (kr) + (A − B) iY1 (kr)
ε→0
r=ε
2i = 2π −
(A − B) = −1,
π
which gives A − B = −i/4. Since the contribution from the coefficient (A + B)
vanishes, we can assign it any value. Traditionally, B is taken zero. Hence, Green’s
function is
(1)
G(r) = −
i H0 (kr)
.
4
(8.18)
8.5.1. Closed-Form Green’s Function for Helmholtz’s Equation. Consider the equation ∇2 u(r) + k 2 u(r) = −f (r). The eigenfunctions φn (r) are the
solutions of the homogeneous equation, i.e.,
∇2 φn (r) + kn2 φn (r) = −f (r).
(8.19)
Green’s function G(r1 , r2 ) satisfies the (point source) equation
∇2 G(r1 , r2 ) + k 2 G(r1 , r2 ) = −δ(r1 − r2 ).
Expanding Green’s function in a series of eigenfunctions of the homogeneous equation
(8.19), we get
∞
G(r1 , r2 ) =
an (r2 )φn (r1 ),
n+0
(8.20)
220
8. ELLIPTIC EQUATIONS
which, after substituting into (8.20), gives
∞
(kn2
n=0
∞
− k ) an (r2 ) φn (r1 ) =
2
(8.21)
φn (r1 )φn (r2 ),
n=0
where we have replaced δ(r1 − r2 ) by its eigenfunction expansion (see §3.4). Then,
by using the orthogonality of the eigenfunction φn (r1 ) we determine an which, when
substituted into (8.21), yield the required Green’s function (in closed form):
∞
G(r1 , r2 ) =
φn (r1 ) φn (r2 )
,
kn2 − k 2
n=0
(8.22)
which is symmetric in r1 and r2 . Hence, the solution of the nonhomogeneous
Helmholtz’s equation is given by
u(r1 ) = G(r1 , r2 ) f (r2 ) dr1 .
(8.23)
If the given equation is L[u](r) + λ u(r) = −f (r), where L is a self-adjoint operator,
then the above Green’s function becomes
∞
G(r1 , r2 ) =
φn (r1 ) φn (r2 )
,
λ2n − λ2
n=0
(8.24)
where (λn , φn ) are orthonormalized eigenpairs of the associated homogeneous equation.
8.6. Green’s Function for 3-D Helmholtz’s Equation
In R3 , it is obvious from physical considerations that Green’s function for the Helmholtz
operator must be spherically symmetric. Then for x = 0 Green’s function G(x) in
the spherical coordinates satisfies the equation
1 d 2 dG δ(r)
r
+ k2 G =
.
r 2 dr
dr
4πr2
(8.25)
d2 w
δ(r)
If we substitute G = w/r, then this equation reduces to 2 + k 2 w =
, which,
dr
4πr
after applying the Fourier sine transform, gives
2
α c0 + k 2 − α 2 w
s =
π
2
π
=
∞
0
δ(r)
sin(αr) dr
4πr
2 1
sin(αr)
α
lim
=
π 4π r→0
r
4π
2
,
π
8.7. 2-D POISSON’S EQUATION IN A CIRCLE
221
2
1 α
where c0 = lim w. Then w
s =
c0 −
, which, on inverting yields
2
r→0
π
4π α − k 2
1
1 e−ikr
w = c0 −
e−ikr . Thus, G(r) = c0 −
. Using the flux condition,
4π 4π
r
dG
we get − lim
dS = −1, where S is a sphere of radius ε with center at
ε→0 S
dr r→ε
the origin and dS = ε2 sin φ dφ dθ is the infinitesimal element of this surface. Thus,
since
2π π 1
ik e−ikε
e−ikε
lim c0 −
+ 2
ε2 sin φ dφ dθ = −1,
ε→0
4π
ε
ε
0
0
1
we get c0 −
4π = −1, which gives c0 = 0. Hence,
4π
G(r) = −
e−ikr
.
4πr
(8.26)
The physical interpretation of this Green’s function is as follows: If we consider the
flux of the quantity G(r) across a sphere of radius r and take the limit as ε → 0, we
obtain the limiting value −1. This implies the presence of a sink of unit strength at
the origin. If instead of Eq (8.25) we were to find the solution of the equation
1 d 2 dG δ(r)
r
+ k2 G = −
,
2
r dr
dr
4πr2
(8.27)
e−ikr
the solution G(r) =
would then represent Green’s function due to a source of
4πr
unit strength at the origin. We point out that the flux of a quantity u away from the
∂u
source is always given by −
, where n is the outward unit normal to the surface.
∂n
Theorem 8.3. Green’s function for the Helmholtz Operator is symmetric,
that is,
G(x, y) = G(y, x).
The proof is similar to that of Theorem 8.2 and is left as an exercise.
8.7. 2-D Poisson’s Equation in a Circle
In the polar cylindrical coordinates the problem in a circle of radius a is defined by a
Poisson’s equation as
∂ 2 u 1 ∂u
1 ∂2u
+
+
= F (r, θ),
∂r2
r ∂r
r2 ∂θ2
u(a, 0) = 0.
∇2 u ≡
r < a,
(8.28)
222
8. ELLIPTIC EQUATIONS
This problem can always be solved by Bernoulli’s separation method by taking u =
R(r)Θ(θ), and is left as an exercise. However, we will use Fourier transform method
to solve this problem. Since the basis is {r n cos nθ, rn sin nθ}, the solution, if any,
can be formally written as a trigonometric Fourier series
1
a0 (r) +
an (r) cos nθ + bn (r) sin nθ ,
2
n=1
∞
u(r, θ) =
where
an (r) =
1
π
π
u(r, θ) cos nθ dθ,
−π
bn (r) =
1
π
π
u(r, θ) sin nθ dθ.
−π
Since u is a periodic function of period 2π, we take the finite Fourier transform of
1
n2
1
n2
(8.28) and get an + an − 2 an = An (r), bn + bn − 2 bn = Bn (r), where
r
r
r
r
π
π
1
1
An (r) =
F (r, θ) cos nθ dθ, Bn (r) =
F (r, θ) sin nθ dθ. (8.29)
π −π
π −π
The boundary conditions imply that an (a) = bn (a) = 0. At the singular point r = 0,
let us require that both an (r) and bn (r) be bounded. Then from (8.29) we have
(r an ) −
n2
an = r An ,
r
(r bn ) −
n2
b n = r Bn ,
r
an (a) = bn (a) = 0,
which is a Sturm-Liouville system. Using above conditions on an and bn at r = a
and r = 0, we get
r a a
a
a0 (r) =
ln
A0 (r ) r dr +
ln
A0 (r ) r dr ,
r
r
0
r
r n n n
1
a
r
r
an (r) =
−
An (r ) r dr 2π
r
a
a
0
a n n n r
a
r
+
−
An (r ) r dr , for n ≥ 1, (8.30)
a
r
a
r
r n n n
1
a
r
r
bn (r) =
−
Bn (r ) r dr 2π
r
a
a
0
a n n n r
a
r
+
−
Bn (r ) r dr ,
a
r
a
r
Now, by Schwarz’s inequality (§1.9.3), we have for n ≥ 2
a
a2
2
[An (r)] r dr ,
an (r)2 ≤
4n (n2 − 1) 0
a
a2
2
[Bn (r)] r dr .
bn (r)2 ≤
4n (n2 − 1) 0
8.7. 2-D POISSON’S EQUATION IN A CIRCLE
a
223
π
If
F (r, θ) r dr dθ < +∞, then the Fourier series solution of the problem
0
−π
(8.29) is
∞
1
a0 (r) +
[an (r) cos nθ + bn sin nθ] ,
2
n=1
u(r, θ) =
(8.31)
which converges uniformly, i.e., u ∈ C 2 and satisfies the boundary condition.
Note. If the boundary condition is changed to u(a, θ) = f (θ), and if F = 0,
then the solution by Bernoulli’s separation method is
r n
1
a0 (r) +
[an (r) cos nθ + bn sin nθ] ,
2
a
n=1
∞
u(r, θ) =
(8.32)
where an and bn are the Fourier coefficients of f (θ).
Green’s function for this problem is obtained as follows: Substituting (8.30) into
the solution (8.31), we get
n
∞ r
a
1 r
1 a n r n
r
u(r, θ) =
A0 (r ) ln
r dr +
−
·
2 0
r
2n
r
a
a
n=1 0
· [An (r ) cos nθ + Bn (r ) sin nθ] r dr +
n ∞ a
a 1 a
1 a n
r
r n
+
A0 (r ) ln
r dr +
−
·
2 r
r
2n
r
a
a
n=1 r
· [An (r ) cos nθ + Bn (r ) sin nθ] r dr ,
(8.33)
which after interchanging summation and integration gives
π a
u(r, θ) =
G(r, θ; r , θ ) F (r , θ ) r dr dθ ,
−π
0
where for r < r
n
1 a n r n
r
−
cos n(θ − θ ) ,
r
n
r
a
a
n=1
(8.34)
while for r > r we find that G(r, θ; r , θ ) = G(r , θ ; r, θ). The series in the above
solution are uniformly convergent in θ for r < r. If we use the identity
G(r, θ; r , θ ) =
1
2π
ln
a
∞
+
∞
1
1 − r2
+
r n cos nθ =
,
2 n=1
2 [1 − 2r cos θ + r 2 ]
then
∞
1 n
r cos nθ =
n
n=1
0
r
cos θ − z 1 dz = − ln 1 − 2r cos θ + r 2 .
2
1 − 2z cos θ + z
2
224
8. ELLIPTIC EQUATIONS
Thus, from (8.34) we find that
2 1
2rr rr
ln 1 − 2 cos(θ − θ ) +
−
r
4π
a
a2
1
2r r 2
−
ln 1 −
cos(θ − θ ) + 2
4π
r
r
2 2
1
r r
=
ln a2 − 2rr cos(θ − θ ) + 2 −
4π
a
− ln r2 − 2rr cos(θ − θ ) + r 2
G(r, θ; r , θ ) =
=
1
2π
1
ln
4π
ln
a
+
r 2 r2
a2
r2 − 2rr cos(θ − θ ) + r 2
a2 − 2rr cos(θ − θ ) +
,
for r < r.
(8.35)
This Green’s function is continuous at r = r; in fact, it belongs to the class C 2 and
satisfies Laplace’s equation ∇2 G = 0 at all points except the point (r , θ ). Moreover,
∂G ∂G and have finite (improper) integrals, in spite of the singularities, and so
∂r
∂θ
we can differentiate u under the integral sign and get
∂u
∂G
∂u
∂G
=
F (r , θ ) r dr dθ ,
=
F (r , θ ) r dr dθ , (8.36)
∂r
∂r
∂θ
∂θ
provided F (r , θ ) is bounded near (r, θ). For r > r, the Green’s function is
G(r, θ; r , θ ) = G(r , θ ; r, θ).
2
2
r dr dθ becomes
Notes. 1. The expression
ln r − 2rr cos(θ − θ ) + r
2
2
ln (x − x ) + (y − y ) dx dy in the rectangular cartesian coordinates. Thus,
the above Green’s function (8.35) in rectangular cartesian coordinates is
1
G(x, y; x , y ) =
ln (x − x )2 + (y − y )2 −
4π
x2 + y 2 x2 + y 2
− ln a2 − 2(xx + yy ) +
, (8.37)
a2
r 2 r 2
2. Since ∇ ln a − 2rr cos(θ − θ ) + 2
= 0 throughout the circle,
a
we conclude that for any bounded domain D the function
1
v(r, θ) =
ln r2 − 2rr cos(θ − θ ) + r 2 F (r , θ ) r dr dθ 4π
D
2
2
is a particular solution of Poisson’s equation ∇2 v = F (r, θ) in D, provided F ∈ C 2 .
In rectangular cartesian coordinates, this function becomes
1
v ∗ (x, y) =
ln (x − x )2 + (y − y )2 F ∗ (x , y ) dx dy ,
4π
D
8.7. 2-D POISSON’S EQUATION IN A CIRCLE
225
where v ∗ (x, y) = v ∗ (r cos θ, r sin θ) = v(r, θ) and F ∗ (x, y) = F ∗ (r cos θ, r sin θ) =
F (r, θ). Hence, we have the following result:
Theorem 8.4. For any bounded domain D with boundary ∂D, the solution
of the nonhomogeneous problem
∇2 u = F
in D, subject to the condition u = 0
is of the form
u(x, y) =
on ∂D,
G(x, y; x , y ) F (x , y ) dx dy ,
(8.38)
(8.39)
D
where G(x, y; x , y ) is Green’s function for Poisson’s equation in D.
Physically, this Green’s function represents the potential at a point (x, y) due to
a point source at (x , y ), or the displacement at (x, y) of a circular membrane due to
a point force at (x , y ).
From (8.37) it is clear that Green’s function in rectangular cartesian coordinates
is of the form
1 G(x, y; x , y ) =
ln (x − x )2 + (y − y )2 + γ(x, y; x , y ) ,
(8.40)
4π
where, for each (x , y ) ∈ D, the term γ is the solution of the problem
∇2 γ = 0
in D.
1
γ=
ln (x − x )2 + (y − y )2
4π
for (x, y) ∈ ∂D.
(8.41)
Example 8.1. Let D be the circle of radius a. Then the problem (8.40) in polar
cylindrical coordinates becomes
∇2 γ = 0
for r < a,
1
γ(a, θ; r , θ ) =
ln r 2 − 2r a cos(θ − θ ) + a2 .
4π
Since the Fourier series for the problem gives
n
∞
1
1
1
1 r
ln r 2 − 2r a cos(θ − θ ) + a2 =
ln a −
cos n(θ − θ ),
4π
2π
2π n=1 n a
we find that
γ(r, θ; r , θ ) =
1
2π
1
=
4π
=
n
1 r n r cos n(θ − θ )
n
a
a
n=1
1
rr
r2 r 2
ln a +
ln 1 − 2 2 cos(θ − θ ) + 4
4π
a
a
2 2
r r
ln a2 − 2rr cos(θ − θ ) + 2 ,
a
1
1
ln a −
2π
2π
∞
226
8. ELLIPTIC EQUATIONS
which when substituted in (8.40) gives (8.35).
Notes. 1. It is important to observe that the introduction of Green’s function
reduces the nonhomogeneous problem (8.28) to a 2-parameter family (8.41) of a
homogeneous problem.
2. The problem (8.28) can be reduced to a homogeneous problem by introducing
a particular solution v of the problem ∇2 v = F ; thus, for example, by introducing
the function v or v ∗ given above, and setting w = u − v or w = u − v ∗ . The function
w satisfies the boundary value problem
∇2 w = 0 on D, such that w = u − v on ∂D,
so that after finding w we obtain u = w + v or u = w + v ∗ .
Example 8.2. Let ∇2 u = 1 in the square S : {0 < x < π, 0 < y < π}, such
that u = 0 on the sides x = 0, x = π, y = 0, y = π. Let w = u − v. Then
∇2 w = 0 on S,
w(x, 0)w(x, π) = 12 x(x − π),
w(0, y) = w(π, y) = 0.
Using Bernoulli’s separation method, the solution of this system is
w(x, y) =
4
π
∞
k=1
sin(2k − 1)x cosh(2k − 1) y −
(2k − 1)2 cosh(2k − 1) π2
π
2
,
and thus,
u(x, y) = 12 x(x − π) +
4
π
∞
k=1
sin(2k − 1)x cosh(2k − 1) y −
(2k − 1)2 cosh(2k − 1) π2
π
2
.
8.8. Method for Green’s Function in a Rectangle
The above analysis leads to the following methods to determine Green’s function for
the parabolic or elliptic operators in compact domains.
8.8.1. Method 1: To solve the problem
∇2 u = F in R : {0 < x < a, 0 < y < b},
u = 0 on the sides x = 0, x = a, y = 0, y = b,
follow these steps:
Step 1. Use the Fourier sine transform:
2 a
2 a
bn (y) =
u(x, y) sin nx dx, Bn (y) =
F (x, y) sin nx dx.
a 0
a 0
8.8. METHOD FOR GREEN’S FUNCTION IN A RECTANGLE
227
Step 2. This reduces the problem to
bn − n2 bn = Bn
for 0 < x < a,
bn (0) = bn (a) = 0,
which has the solution
1
bn (y) =
n sinh nb
y
sinh n(b − y) sinh ny Bn (y )) dy 0
+
b
sinh ny sinh n(b − y )Bn (y ) dy .
y
Step 3. Write the Fourier series for u by substituting the definition of Bn , and
interchanging integration and differentiation. This yields the formal solution
a
b
u(x, y) =
0
G(x, y; x , y ) F (x , y ) dx dy ,
0
where
G(x, y; x , y ) =
 ∞ 2 sinh n(b − y) sinh by sin nx sin nx


,

na sinh nb
y ≤ y,
2 sinh ny sinh n(b − y ) sin nx sin nx
,
na sinh nb
n=1
y ≤ y ,



n=1
∞
or in terms of the elliptic functions
G(x, y; x , y ) =
1
2π
∞
1
n
n=1
+
e−n|y−y | +
e−n(2b+|y−y |) + e−n(2b−|y−y |) − e−2(2b−y−y ) − e−n(y+y )
×
1 − e−2nb
× cos n(x − x ) − cos n(x + x )
1
ln 1 + e−2|y−y | − 2 e−|y−y | cos(x + x ) −
4π
1
−
ln 1 + e−2|y−y | − 2 e−|y−y | cos(x − x ) +
4π
∞
−n(2b+|y−y |)
e
+ e2b−|y−y |) − e−n(2b−y−y ) − e−n(y−y )
+
sin nx sin nx ,
−2nb )
na
(1
−
e
n=1
=
where the infinite series and all its derivatives are uniformly convergent in x and y
for any fixed (x , y ) ∈ D.
228
8. ELLIPTIC EQUATIONS
8.8.2. Method 2. To find the solution of the homogeneous boundary value
problem ∇2 u = 0 in D, such that u = g on ∂D, in terms of the known Green’s
function, we follow these steps:
Step 1. Use the above Method 1 to reduce the nonhomogeneous problem to a
homogeneous problem.
Step 2. Choose any point (x0 , y0 ) ∈ D. Let h(x, y) ∈ C 2 (D) such that h = g
on ∂D and h = 0 in the neighborhood of (x0 , y0 ).
Step 3. Set w = u − h. Then w satisfies ∇2 w = −∇2 h in D, and w = 0 on
∂D. Hence,
w (x0 , y0 ) =
G(x, y; x , y ) ∇2 h(x , y ) dx dy .
D
Step 4. Since h = 0 near the point (x0 , y0 ), apply the divergence theorem (§1.7)
and the above relation for w (x0 , y0 ), which gives
∂G
w (x0 , y0 ) = −
(x0 , y0 ; x , y ) g(x , y ) ds,
∂D ∂n
∂G
where
is the directional derivative of G in the direction of the unit vector n
∂n
perpendicular to the boundary ∂D.
Step 5. Since h (x0 , y0 ) = 0, so w (x0 , y0 ) = u (x0 , y0 ). As (x0 , y0 ) is an
arbitrary point in D, the solution u(x, y) of this problem is given by
∂G
u(x, y) = −
(x0 , y0 ; x , y ) g(x , y ) ds.
(8.42)
∂n
∂D
Example 8.3. For the circle r < a, the Green’s function is given by (8.35).
Then
∂G
∂G
= (r, θ; a, θ )
∂n
∂r 1 2r 2 /a2 − 2r cos(θ − θ )
2a − 2r cos(θ − θ )
=
−
4π r 2 − 2ra cos(θ − θ ) + a2
r 2 − 2ra cos(θ − θ ) + a2
1
a2 − r 2
=
.
2
2aπ r − 2ra cos(θ − θ ) + a2
Thus, (8.42) gives
a2 − r2
u(r, θ) =
2π
π
−π
g(θ ) dθ ,
r 2 − 2ra cos(θ − θ ) + a2
which the Poisson’s integral formula for the circle.
r < a,
(8.43)
8.9. POISSON’S EQUATION IN A CUBE
229
8.9. Poisson’s Equation in a Cube
Consider the problem
∂2u ∂2u ∂ 2u
+ 2 + 2 = F (x, y, z), in the cube D : {0 < x, y, z < π},
∂x2
∂y
∂z
u(0, y, z) = u(π, y, z) = 0,
∇2 u ≡
u(x, 0, z) = u(x, π, z) = 0,
(8.44)
u(x, y, 0) = u(x, y, π) = 0,
where u ∈ C 2 . Expand u and F in triple Fourier sine series:
∞ ∞ ∞
u(x, y, z) =
almn sin lx sin my sin nz,
l=1 m=1 n=1
F (x, y, z) =
∞ ∞ ∞
Almn sin lx sin my sin nz,
l=1 m=1 n=1
substitute into Poisson’s equation in (8.44), which gives almn =
Thus,
∞
u(x, y, z) =
l=1
∞
Almn
.
l 2 + m 2 + n2
∞
Almn
sin lx sin my sin nz.
2 + m 2 + n2
l
m=1 n=1
(8.45)
This series for u(x, y, z) and its first and second derivatives converge absolutely and
uniformly, provided the series for F does so. By Parseval’s equality (§1.9.2), the
solution (8.45) can be written formally as
π π π
u(x, y, z) =
G(x, y, z; x , y , z ) F (x , y , z ) dx dy dz 0π 0π 0π
or
=
G(x, x ) F (x ) dx ,
(8.46)
0
0
0
where Green’s function G(x, x ) is of the form
G(x, x ) =
8
π3
∞
l=1
∞
∞
sin lx sin my sin nz sin lx sin my sin nz . (8.47)
l 2 + m 2 + n2
m=1 n=1
Let r = |x − x | = (x − x ) + (y − y ) + (z − z ) , and choose a constant a
depending on the fixed point x such that the sphere r ≤ a is inside the cube D. Define
the function

 1 5 1− r 4−3 1− r 5
for r < a,
2r
a
a
ψ(r) =

0 for r ≥ a.
2
2
2
230
8. ELLIPTIC EQUATIONS
It is easy to show that the function ψ(r) − 1/r has continuous second derivatives,
bounded third derivatives, and square integrable fourth derivatives. Let us change the
variables ξ = x − x, η = y − y, ζ = z − z, and set
π3
Blmn =
8
0
π
π
0
π
ψ(r) sin lx sin my sin nz dx dy dz 0
=
ψ
ξ2 + η2 + ζ 2 ×
ξ 2 +η2 +ζ 2 <a2
× sin l(x + ξ) sin m(y + η) sin n(z + ζ) dξ dη dζ
ψ (sin lξ cos lx + cos lξ sin lx) ×
=
ξ 2 +η2 +ζ 2 <a2
× (sin mη cos my + cos mη sin my) ×
× (sin nζ cos nz + cos nζ sin nz) dξ dη dζ
= sin lx sin my sin nz
ψ cos lξ cos mη cos nζ dξ dη dζ.
ξ2 +η 2 +ζ 2 <a2
Since ψ is even in ξ, η, ζ, the integrals of ψ sin lξ, ψ sin mη and ψ sin nζ are zero.
Further, since
1
cos( lξ + mη + nζ) + cos( lξ + mη − nζ)+
4
+ cos( lξ − mη + nζ) + cos(lξ − mη − nζ) ,
cos lξ cos mη cos nζ =
and since ψ is even in ξ, η, ζ, the integral of ψ times each of these cosines gives the
same result. Then
π3
Blmn =
8
ψ
ξ 2 + η 2 + ζ 2 cos (lξ + mη + nζ) dξ dη dζ.
ξ 2 +η2 +ζ 2 <a2
In terms of the spherical coordinates (r, θ, φ) with origin at (x, y,
z) and polar axis in
the direction of the vector with components √
(l, m, n), we have ξ 2 + η 2 + ζ 2 = r,
and lξ + mη + nζ = r µ cos θ, where µ = l2 + m2 + n2 , and thus,
π3
Blmn =
8
a
π
2π
ψ(r) cos(rµ cos θ) r 2 sin θ dr dθ dφ
0
0
0
π
a
sin(rµ cos θ)
= 2π
ψ(r) −
r2 dθ
rµ
0
0
4π a
=
ψ(r) r sin(rµ) dr.
µ 0
8.10. LAPLACE’S EQUATION IN A SPHERE
231
Then, using the definition of ψ(r), we get
32
60(2 + cos aµ) 180 sin aµ
Blmn = 2 2 1 −
+
sin lx sin my sin nz.
π µ
a4 µ4
a5 µ5
Thus,
∞
1
8
ψ(r) = 3
4π
π
∞
∞
l=1 m=1 n=1
1
60(2 + cos aµ) 180 sin aµ
−
+
×
r2
a4 µ6
a5 µ7
× sin lx sin my sin nz sin lx sin my sin nz .
Comparing this with (8.47) we find that
1
480
G(x, y, z; x , y , z ) =
ψ(r) − 3
4π
π
∞
∞
∞
l=1 m=1 n=1
2 + cos aµ 3 sin aµ
− 5 7 ×
a4 µ6
a µ
× sin lx sin my sin nz sin lx sin my sin nz .
(8.48)
1
∈ C 2 , and Green’s function G vanishes on the faces of the
4πr
1 d2
cube D. Also, ∇2 (1/r) = 0 for = 0, and ∇2 [ψ(r)] =
(rψ) for r = 0, and
r dr2
1 3
since the Laplacian of the series in (8.48) is the series of
∇ ψ, which implies that
4π
2
2
∇ G = 0. Hence, Green’s function G which satisfies ∇ G = 0 in the cube D, and
G = 0 on ∂D, is given by
The function G −
G(x, y, z; x , y , z ) =
4π
1
+ (y − y )2 + (z − z )2
1
− γ(x, y, z; x , y , z ) =
− γ(x; x ),
4πr
(x −
x ) 2
(8.49)
where r = |x − x |, and γ is a regular solution of Laplace’s equation, i.e., ∇2 γ = 0
1
in the cube D and γ =
on ∂D.
4πr
8.10. Laplace’s Equation in a Sphere
Consider the boundary value problem in a sphere of radius a:
∂ 2 u 2 ∂u
1
∂
∂u
1
∂2u
+
+
sin
θ
+
= 0, r < a,
∂r 2
r ∂r
r 2 sin θ ∂θ
∂θ
r 2 sin2 θ ∂φ2
u(a, θ, φ) = f (θ, φ),
(8.50)
which is the same as (4.11). This problem has been solved by Bernoulli’s separation
method in §4.3.3, which leads to Eqs (4.12), (4.13), and (4.14). The values of λ for
232
8. ELLIPTIC EQUATIONS
√
which there is a solution P of Eq (4.14) such that P and 1 − t2 P remain bounded
at both t = 1, −1, must be found in the following two cases:
Case 1. m = 0: Eq (4.14) with λ = 0 has two solutions: (1+t)m/2 (1−t)−m/2
and (1 − t)m/2 (1 + t)−m/2 , and Green’s function is given by

m/2
1 (1 + t )(1 − t)



,
 2m (1 − t )(1 + t)
G(t, t ) =

 1 (1 + t)(1 − t ) m/2


,
2m (1 − t)(1 + t )
1
, which implies that
2m
set of eigenfunctions is complete in this cases.
Obviously, G(t, t ) ≤
1
t ≤ t,
(8.51)
t ≥ t.
[G(t, t )] dt dt < +∞, and the
2
−1
Case 2. m = 0: Eq (4.14) with λ = 0 has only one solution: P = 1, which
is continuously differentiable in [−1, 1]. Hence, λ = 0 is an eigenvalue. Green’s
function G0 (t, t ) which is the solution of the problem:
1 − t2 G
−G=F
for −1 < t < 1,
G and G bounded,
is given by
G0 (t, t ) =
∞
uk (t) uk (t )
,
1
2
k=1 (1 + λ )
[uk (t)] dt
k
(8.52)
−1
where λk are the eigenvalues. The eigenfunctions for this case satisfy the equation
d
dP 1 − t2
+ λ P = 0,
dt
dt
(8.53)
a solution of which in the neighborhood of t = 1 in powers of (t − 1) is
∞
P = (t − 1)α
ck (t − 1)k .
k=0
Substituting this expression for P into Eq (8.53) we get a power series which must
be identically zero:
∞
− 2c0 α2 (t − 1)α−1 −
k=0
2(k + α + 1)2 ck+1 +
+ (k + α + 1)(k + α) ck − λ ck (t − 1)k+α = 0.
8.10. LAPLACE’S EQUATION IN A SPHERE
233
If c0 = 0, we must have α = 0. Since the above equation has a double root α = 0,
there is a second solution of the form ln(1 − t) at t = 1. But this solution, being
unbounded, must be discarded. Now, we set the coefficient of (t − 1)k equal to zero,
[k(k + 1) − λ]
and get ck+1 = −
ck , k = 0, 1, 2, . . . , which, by recursion, gives
2(k + 1)2
ck =
(−1)k [k(k − 1) − λ] [(k − 1)(k − 2) − λ] · · · [2 − λ] [−λ]
c0 , k ≥ 1.
2k (k!)2
∞
If we apply the ratio test to the series
ck (t − 1)k , we find that it converges for
k=0
|t − 1| < 2. This also holds for the derivatives of this series. Bur for k + 1 > λ, we
ck+1
1 k−1
have −
>
, which shows that this series will, in general, approach ±∞
ck
2 k+1
as t → 1, except in the case when the series terminates. Thus, if for some integer n
we have n cn = 0 but cn+1 = cn+2 = · · · = 0, then the above recursion formula
is true iff λ = n(n + 1), n = 0, 1, 2, . . . . Hence, we obtain a solution bounded
in [−1, 1] iff λ = n(n + 1). Setting c0 = 1, the eigenfunctions corresponding the
eigenvalues λn = n(n + 1) are given by
∞
Pn (t) =
n=0
(n + k)!
2k
2
(n − k)! (k!)
(1 − t)k ,
where Pn (t) are the Legendre polynomials of degree n in t; a few of these polynomials are: P0 (t) = 1, P1 (t) = t, P2 (t) = 32 t2 − 12 , P3 (t) = 52 t3 − 32 t.
!1
These polynomials satisfy the orthogonality condition: −1 tk Pk (t) dt = 0, for
k = 0, 1, 2, . . . , n − 1; they are normalized by taking Pn (t) = 1, and defined by
Rodrigue’s formula:
Pn (t) =
1 dn 2
t −1
n! dtn
2n
n
.
Hence, the formal solution of the problem (8.50) is given by
2π
π
u(r, θ, φ) =
0
G(r, θ, φ; a, θ , φ ) f (θ , φ ) sin θ dθ dφ,
(8.54)
0
where
G(r, θ, φ; a, θ , φ ) =
∞
n=0
r n
a
Gn (θ, φ; θ , φ ),
2n + 1
Pn (cos θ) Pn (cos θ )
4π
∞
(2n + 1)(n − m)! m
+
Pn (cos θ) Pnm (cos θ ) cos m(θ − θ ),
2π
(n
+
m)!
m=1
(8.55)
Gn (θ, φ; θ , φ ) =
(8.56)
234
8. ELLIPTIC EQUATIONS
∞
1
an0 P0 (cos θ) +
(anm cos mφ + nnm sin mφ) Pnm (cos θ) ,
2
n=0
m=1
2π π
(2n + 1)(n − m)!
anm =
f (θ, φ)Pnm (cos θ) cos mφ sin θ dθ dφ,
2π (n + m)!
0
0
(2n + 1)(n − m)! 2π π
bnm =
f (θ, φ)Pnm (cos θ) sin mφ sin θ dθ dφ.
2π (n + m)!
0
0
∞
f (θ, φ) =
For a fixed n, the function Gn (θ, φ; θ , φ ) has the property:
2π π
Gn (θ, φ; θ , φ ) r k Pkm (cos θ ) (a cos mφ + b sin mφ ) sin θ dθ dφ
0
0
=
r n Pnm (cos θ) (a cos mφ + b sin mφ) ,
0
for k = n,
for k = n.
m/2
Since Pn (1) = 1, and Pnm (t) has a factor 1 − t2
for m ≥ 1, so PnM (1) = 0 for
2n
+
1
m ≥ 1. Thus, Gn (0, φ ; θ , φ ) =
Pn (cos θ ), and
4π
2n + 1
Gn (θ, φ; θ , φ ) =
Pn (cos θ ),
(8.57)
4π
where θ is such that cos θ = cos θ cos θ + sin θ sin θ cos(φ − φ ). Substituting
the above value of Gn into (8.55) we get
G(r, θ, φ; a, θ , φ ) =
1
4π
1
=
4π
∞
(2n + 1)
n=0
r n
a
Pn (cos θ )
∞ ∂
r n
2r
+1
Pn (cos θ ).
∂r
a
n=0
(8.58)
Example 8.4. Consider the Laplace equation in spherical coordinates, independent of φ:
∂ 2 (ru)
1 ∂
∂u
r
+
sin
θ
= 0.
(8.59)
∂r 2
sin θ ∂θ
∂θ
Using Bernoulli’s separation method, let u = R(r)Θ(θ), which after substitution in
Eq (8.59) gives
r d2 (rR)
1 d
dΘ
=−
sin θ
.
(8.60)
R dr 2
sin θ dθ
dθ
Since each side of Eq (8.60) must be equal to a constant, say α2 , this equation is
equivalent to the following two equations:
d2 (rR)
− α2 R = 0,
dr2 1 d
dΘ
sin θ
+ α2 Θ = 0.
sin θ dθ
dθ
r
(8.61)
(8.62)
8.11. POISSON’S EQUATION AND GREEN’S FUNCTION IN A SPHERE
235
Eq (8.61) can be written as
r2
d2 R
dR
+ 2r
− α2 R = 0,
dr 2
dr
and has the general solution as R(r) = Arn + Br m , where n = − 12 +
and m = − 12 −
α2 +
1
4
α2 + 14 . Thus, m = −n − 1, and we can write α2 = n(n + 1),
where n is arbitrary. Then R(r) = Ar n + Br−n−1 . Thus, R = r n and R = r −n−1
are particular solutions of Eq (8.61) with α2 = n(n + 1), i.e., of the equation
r2
d2 R
dR
+ 2r
− n(n + 1)R = 0.
2
dr
dr
Again, with this value of α2 Eq (8.62) becomes
1 d
dΘ
sin θ
+ n(n + 1)Θ = 0,
sin θ dθ
dθ
where n > 0 is an integer (see Example 4.2), so that its particular solution is Θ(θ) =
1
Pn (cos θ). Hence, u(r, θ) = rn Pn (cos θ) and u(r, θ) = n+1 Pn (cos θ), where n
r
is a positive integer, are particular solutions of Eq (8.59), where the first of these
solutions, obtained in Example 4.2, is known as a solid zonal harmonic.
8.11. Poisson’s Equation and Green’s Function in a Sphere
Consider the problem in a sphere r < a, which is governed by Eq (8.50). Then
expanding u and F as Fourier series in Pnm (cos θ) cos mφ and in Pnm (cos θ) sin mφ
we find that
(2n + 1)(n − m)! π π
anm (r) =
u(r, θ, φ)Pnm (cos θ) cos mφ sin θ dθ dφ,
2π(n + m)!
−π 0
(2n + 1)(n − m)! π π
bnm (r) =
u(r, θ, φ)Pnm (cos θ) sin mφ sin θ dθ dφ,
2π(n + m)!
−π 0
(2n + 1)(n − m)! π π
Anm (r) =
F (r, θ, φ)Pnm (cos θ) cos mφ sin θ dθ dφ,
2π(n + m)!
−π 0
(2n + 1)(n − m)! π π
Bnm (r) =
F (r, θ, φ)Pnm (cos θ) sin mφ sin θ dθ dφ.
2π(n + m)!
−π 0
(8.63)
Then taking the finite Fourier transform of Eq (8.63) and integrating by parts we get
2 n(n + 1)
a −
anm = Anm ; ann (a) = 0, anm bounded,
r nm
r2
2
n(n + 1)
bnm + bnm −
bnm = Bnm ; bnn (a) = 0, anm bounded.
r
r2
anm +
236
8. ELLIPTIC EQUATIONS
These two boundary value problems have the solutions
a
anm (r) =
0a
bnm (r) =
Gn (r, r ) Anm (r) r 2 dr ,
(8.64)
Gn (r, r ) Bnm (r) r 2 dr ,
0
where
n 
1
r
r −n−1 r n

−
for r ≤ r,

 (2n + 1) a a
a
a
Gn (r, r ) =
r n r −n−1 r n 
1


−
for r ≥ r.
(2n + 1) a a
a
a
(8.65)
Then the solution of the problem (8.63) is
∞
u(r, θ, φ) =
n=0
1
an0 Pn (cos θ)+
2
m
anm cos mφ + bnm sin mφ Pn (cos θ) ,
n
+
(8.66)
m=1
!!! 2 2
where the series converge uniformly iff
F r sin θ dr dθ dφ < +∞. Now, if we
substitute the values of Anm (r) and Bnm (r) from (8.63) into (8.64) and interchange
integration and summation in (8.66), we obtain
π
π
a
G(r, θ, φ; r , θ , φ) F (r , θ , φ ) r 2 sin θ dr dθ dφ ,
u(r, θ, φ) =
−π
0
0
(8.67)
where for r < r
G(r, θ, φ; r , θ , φ.) =
∞
1
=
(2n + 1) a
n=0
r
a
n r −n−1 r n
−
Gn (θ, φ; θ , φ ),
a
a
(8.68)
where Gn (θ, φ; θ , φ) is defined in (8.57). Thus, by (8.58)
G(r, θ, φ; r , θ , φ ) =
∞
1
4πa n=0
r
a
n r −n−1 r n
−
Pn (cos θ ). (8.69)
a
a
For r > r, just interchange r and r in the above formulas.
8.12. APPLICATIONS OF ELLIPTIC EQUATIONS
237
8.12. Applications of Elliptic Equations
Some examples of these applications are as follows:
8.12.1. Dirichlet Problem for Laplace’s Equation. The solution of the
boundary value problem ∇2 u = f (x) in D such that u = g(xs ) on ∂D, where xs is
an arbitrary point on the boundary, is given by
u(x ) =
f (x)G(x, x ) dx +
g(xs )
D
∂D
∂G(x, x )
dS.
∂n
(8.70)
Proof. Let x denote an arbitrary point in D; then by (1.16) we have
∂v
∂u
u
−v
dS.
∂n
∂n
∂D
u∇2 v − v∇2 u dx =
D
If we replace v by G(x, x ) we get
D
u∇2 G(x, x ) − G(x, x )∇2 u dx
∂G(x, x )
∂u
=
u
− G(x, x )
dS.
∂n
∂n
∂D
which yields
(u δ(x, x ) − f (x)G(x, x )) dx =
D
g(xs )
∂D
∂G(x, x )
dS,
∂n
and the result follows.
8.12.2. Neumann Problem for Laplace’s Equation. Find the solution of
∂u
∇2 u = F (x) in D such that
= P (xs ) on ∂D, where xs is an arbitrary point on
∂n
the boundary. First, we note that the solution to the problem stated here is not always
possible, for if we apply the divergence theorem (§1.7) to ∇u, we have
∂u
∇ · ∇u dx =
∇2 u dx =
dS,
∂n
D
D
∂D
which implies that
F (x) dx =
D
P (xs )dS.
(8.71)
∂D
This consistency condition is necessary for the existence of the solution to the Neumann problem. We note that the homogeneous problem, i.e., when F (x) and P (x)
are both zero, always has a nontrivial solution u = C =(const). Hence, an arbitrary
constant can always be added to any solution of the Neumann problem. Green’s
238
8. ELLIPTIC EQUATIONS
function in this case needs to be modified such that it satisfies the consistency condition. If F (x) = δ(x − x ) and P (xs ) = 0, the consistency condition is not satisfied.
However, the consistency condition is satisfied if take F (x) = δ(x − x ) − 1/V ,
where V is the volume of D. This modified Green’s function, also known as the
Neumann’s function (see §10.3), satisfies
∇2 G(x, x ) = δ(x − x ) −
1
,
V
∂G(x, x )
= 0.
∂n
Proceeding as in §8.12.1, Green’s second identity (1.16) in the present case yields
1
u δ x − x −
− F (x)G(x, x ) dx
V
D
∂G (xs , x)
=
u(xs )
− P (xs )G(xs , x ) dS,
∂n
∂D
or
u
+ F (x)G(x, x ) dx −
P (xs )G(x, x ) dS.
(8.72)
D V
∂D
u
In the solution (8.72) the only unknown integral is
dx which is equal to the
V
D
average value of u in D and is a constant. Since we can add an arbitrary constant to
any solution of the Neumann problem, we can express the solution (8.72) as
u(x ) =
F (x)G(x, x ) dx −
P (xs )G(xs , x ) dS + C.
u(x ) =
D
∂D
This solution is valid only when the consistency condition (8.71) is satisfied.
In the one-dimensional case the corresponding problem is
y = f (x),
y (0) = α, y (l) = β.
If we integrate this equation from 0 to l, we get
l
l
y =
f (x) dx,
0
0
or β − α =
l
f (x) dx.
(8.73)
0
Thus, the solution to the one-dimensional Neumann problem does not always exist.
The condition (8.73) is a consistency condition required for the solution to exist.
The solution obtained by the method of images satisfying the differential equation ∇2 u (x, y; x , y ) = δ (x − x ) δ (y − y ), subject to the Neumann condition
uy (x, 0; x y ) = 0, is given by
u (x, y; x , y ) =
1
2
2
2
2
ln (x − x ) + (y − y )
(x − x ) + (y + y )
.
4π
8.12. APPLICATIONS OF ELLIPTIC EQUATIONS
239
It is clear that the algebraic sum of charges is not zero and u (x, y; x , y ) does not
approach zero as r → ∞ and does not satisfy the consistency condition (8.71) for the
Neumann problem in the half-space. Thus, we cannot use Green’s second identity
(1.16), as we have done above, for the half-space problem because in this case the
area is unbounded.
8.12.3. Robin Problem for Laplace’s Equation. Find the solution of
∂u
∇2 u = f (x) in D such that u + α
= g(xs ) on ∂D, where xs is an arbitrary point
∂n
on the boundary. The solution for the Robin problem is given by
1
u(x ) =
f (x)G(x, x )dx −
g(xs )G(x, x ) dS
α
D
∂D
∂G(x, x )
=
f (x)G(x, x ) dx +
g(xs )
dS.
∂n
D
∂D
Green’s function in this case is the solution of the following problem:
∇2 G(x, x ) = δ(x − x ) in D such that G(x, x ) + α
∂G(x, x )
= 0 on ∂D.
∂n
The proof is similar to the previous two cases and is left as an exercise.
8.12.4. Dirichlet Problem for Helmholtz’s Equation in Terms of
Green’s Function. We will solve the Dirichlet problem, which in this case is
(∇2 + k 2 )u(x) = F (x) in D, u(xs ) = K(xs )
on ∂D.
(8.74)
Green’s function satisfies the following conditions:
(∇2 + k 2 )G(x, y) = δ(x − y)
in D,
G(xs , y) = 0 on ∂D,
(8.75)
where xs is any point on the boundary ∂D. Multiplying the first equation in (8.74)
by G(x, y) and the first equation in (8.75) by u(x) and subtracting, we get
G(x, y)∇2 u(x) − u(x)∇2 G(x, y) = F (x)G(x, y) − u(x)δ(x − y).
(8.76)
Integrating both sides of Eq (8.76) over D and using the divergence theorem (§1.7),
we have
G(x, y)∇2 u(x) − u(x)∇2 (G(x, y) dx
D
∂u(x)
∂G(x, y)
=
G(x, y)
− u(x)
dS
∂n
∂n
∂D
=
[F (x)G(x, y) − u(x)δ(x − y)] dx
D
=
F (x)G(x, y) dx − u(y).
D
240
8. ELLIPTIC EQUATIONS
Thus,
u(y) =
F (x)G(x, y) dx +
D
K(xs )
∂D
∂G(xs , y)
dS,
∂n
(8.77)
since G(xs , y) = 0, and u(xs ) = K(xs ). The solution (8.77) implies that if both
F (x) = K(xs ) = 0, then only a trivial solution exists. But we note that if D is the
1
square R : {0 < x, y < a}, then u = φ(x, y) = sin(nπx/a) sin(nπy/a) is a
2
nontrivial solution of (∇2 + 2n2 π 2 /a2 )u(x, y) = 0, with u = 0 on the boundary. A
consequence of this result is:
Theorem 8.5. If u(x, y) is a solution of
(∇2 + 2n2 π 2 /a2 ) u(x, y) = F (x, y)
in R, and u = 0 on ∂R,
where R is the square {0 < x, y < a}, then φ(x, y) =
(8.78)
1
sin(nπx/a) sin(nπy/a),
2
and F (x, y) satisfies the condition
F (x, y)φ(x, y) dx dy = 0.
R
Proof. Let u(x, y) be a solution of Eq (8.78). Multiplying both sides of Eq
(8.78) by φ(x, y) and noting that (∇2 + 2n2 π 2 /a2 ) φ(x, y) = 0, we get
φ(x, y)(∇2 +2n2 π 2 /a2 )u(x, y)−u(x, y)(∇2 +2n2 π 2 /a2 )φ(x, y) = φ(x, y)F (x, y),
or φ(x, y)∇2 u(x, y) − u(x, y)∇2 φ(x, y) = φ(x, y)F (x, y). Integrating both sides
over R and applying the divergence theorem (§1.7) to the left side, we find that
∂u(x, y)
∂(φ(x, y) (φ(x, y)
− u(x, y)
ds =
φ(x, y)F (x, y) dx dy.
∂n
∂n
∂R
R
However, the left side is zero because both u(x, y) and φ(x, y) vanish on ∂R. Obviously, the solution is not unique, since u(x, y) + Cφ(x, y) is also a solution for any
arbitrary constant C.
The solution of the Neumann or the Robin boundary value problem can be derived
in the same manner as that for Laplace’s equation.
8.12.5. Dirichlet Problem for Laplace’s Equation in the Half-Plane.
Let D be the half-plane y > 0. Then Green’s function associated with the Dirichlet
boundary condition u = 0 on the boundary y = 0 is given by (8.3). Hence, for
u(x, 0) = f (x) we find from (8.70) that
∞
1
∂ (x − x )2 + (y − y )2 u(x , y ) = −
f (x)
ln
dx
4π −∞
∂y
(x − x )2 + (y + y )2 y=0
∞
1
−4y =−
f (x)
dx
4π −∞
(x − x )2 + y 2
y ∞
f (x)
=
dx.
π −∞ (x − x )2 + y 2
8.12. APPLICATIONS OF ELLIPTIC EQUATIONS
241
After interchanging (x , y ) and (x, y), we get
y ∞
f (x)
u(x, y) =
dx.
π −∞ (x − x )2 + y 2
Clearly the boundary condition at y = 0 is not satisfied. For an explanation, see
Exercise 8.9.
8.12.6. Dirichlet Problem
for Laplace’s Equation in a Circle. To
solve ∇2 u = 0, r = x2 + y 2 < 0, u(a, 0) = f (θ), let Γ be the circle r = a,
and let u(a, θ) = f (θ). In this case Green’s function is given in Exercise 8.2. Since
∂
∂
=
on the circle Γ, we find from (8.70) that
∂n
∂r
2π
2
1
∂ a2 [r 2 − 2rr cos(θ − θ ) + r ] u(r , θ ) =
f (θ)
ln 2 2
a dθ
4π 0
∂r
r r − 2rr a2 cos(θ − θ ) + a4 r=a
2π
1
2r − 2r cos(θ − θ )
=
f (θ) 2
4π 0
r − 2rr cos(θ − θ ) + r 2
2
2rr − 2a2 r cos(θ − θ ) − 2 2
a dθ
r r − 2rr a2 cos(θ − θ ) + a4 r=a
2 2π
a2 − r f (θ)
=
dθ.
2
2π
a − 2ar cos(θ − θ ) + r 2
0
After interchanging (r , θ ) and (r, θ), we get
a2 − r 2
u(r, θ) =
2π
2π
0
a2
−
2ar
f (θ)
dθ.
cos(θ − θ ) + r 2
Clearly the boundary condition at y = 0 is not satisfied. For an explanation, see
Exercise 8.9.
8.12.7. Dirichlet Problem for Laplace’s Equation in the Quarter
Plane. To find Green’s function for the problem
∂ 2φ ∂ 2φ
+ 2 = 0, x, y > 0,
∂x2
∂y
φ(x, 0) = f (x),
∂φ
(0, y) = g(y),
∂x
we use Green’s function to solve this problem. In this case Green’s function G =
G(x, y; x , y ) satisfies the system
∂ 2G ∂ 2G
+
= δ(x − x )δ(y − y ), x, y > 0,
∂x2
∂y 2
G(x, 0) = 0,
∂G
(0, y) = 0.
∂x
Method of Images. The boundary conditions imply that G = 0 on the
x-axis and that there is no flux across the y-axis. The distribution of sources and
242
8. ELLIPTIC EQUATIONS
sinks pertaining to these conditions is as follows: a source of strength 1 at the points
(x , −y ) and (−x , −y ) each, and a sink of strength −1 at the points (x , y ) and
(−x , y ) each (see Fig. 8.2).
y
.
strength
−1
(− x’, y’)
.
strength
−1
1
( x’, y’)
Β
x
0
.
strength
+1
.
strength
+1
(−x’,−y’)
(x’,−y’)
Α
0
Fig. 8.2.
1
Fig. 8.3.
Green’s function corresponding to the sources and sinks so marked in Fig. 8.2 is
given by
G(x, y; x , y ) =
1
[(x − x )2 + (y − y )2 ][(x + x )2 + (y − y )2 ]
ln
, (8.79)
4π
[(x + x )2 + (y + y )2 ][(x − x )2 + (y + y )2 ]
where
G(x, 0; x , y ) =
1
[(x − x )2 + y ][(x + x )2 + y ]
1
ln
2
2 = 4π ln(1) = 0,
2
2
4π
[(x + x ) + y ][(x − x ) + y ]
G(0, y; x , y ) =
1
[x + (y − y )2 ][x + (y − y )2 ]
ln
4π
[x 2 + (y + y )2 ][x 2 + (y + y )2 ]
2
2
2
2
1
[x + (y − y )2 ]
ln
,
2π
[x 2 + (y + y )2 ]
1
x − x
x + x
Gx (x, y; x , y ) =
+
π (x − x )2 + (y − y )2
(x + x )2 + (y − y )2
x+x
x − x
−
−
,
(x + x )2 + (y + y )2
(x − x )2 + (y + y )2
1
−x
x
x
Gx (0, y; x , y ) =
+
−
π x 2 + (y − y )2
x 2 + (y − y )2
x 2 + (y + y )2
−x
− 2
= 0,
x + (y + y )2
1
y − y
y − y
Gy (x, y; x , y ) =
+
2
2
2π (x − x ) + (y − y )
(x + x )2 + (y − y )2
2
=
8.12. APPLICATIONS OF ELLIPTIC EQUATIONS
243
y + y
y + y
−
,
(x + x )2 + (y + y )2
(x − x )2 + (y + y )2
y
1
1
Gy (x, 0; x , y ) = −
.
2 +
2
π (x − x ) + y
(x + x )2 + y 2
−
From §8.12.1, the solution of the problem is given by
φ(x , y ) =
∂G
∂φ
φ
−G
,
∂n
∂n
Γ
where Γ is the boundary of the quarter plane traversed in the counterclockwise direc∂
∂
∂
∂
tion and G is defined above. Note that
=−
on the y-axis,
=−
on the
∂n
∂x
∂n
∂y
∂φ
x-axis,
= g(y) on the y-axis, and φ(x, 0) = f (x). Also, since G(x, 0; x , y ) = 0
∂x
on OA, Gx (x, 0; x , y ) = 0 on BO (see Fig. 8.3), ds = −dy, and y varies from ∞
to 0, the solution is given by
φ(x , y ) =
lim
OA=OB→∞
y
=
π
OABO
∂G
∂y
dx + g(y) [G]x=0 (−dy)
y=0
1
1
f (x)
+
dx
(x − x )2 + y 2
(x + x )2 + y 2
∞
2
1
x + (y − y )2
+
g(y) ln
dy.
2π 0
x 2 + (y + y )2
∞
0
[−f (x)]
The solution φ(x, y) is obtained by interchanging (x , y ) and (x, y). Clearly the
boundary condition at y = 0 is not satisfied. For an explanation, see Exercise 8.9.
8.12.8. Vibration Equation for the Unit Sphere. Consider the unit sphere
x2 + y 2 + z 2 ≤ 1. Using spherical coordinates the eigenvalue problem is defined by
1
∂
∂ uφ ∂
2
∇ u + λu = 2
r ur sin θ +
+
(uθ sin θ) + λu = 0.
r sin θ ∂r
∂φ sin θ
∂θ
2
Using Bernoulli’s separation method by taking u = Y (θ, φ)R(r), we get
r 2 R
+ λ2 r 2 R
1
∂
Yφ
∂
=−
+
(Yθ sin θ) = k,
R
Y sin θ ∂φ sin θ
∂θ
where k is a constant, whose value must be determined in such a way that the differential equation
∇2 Y + k Y ≡
1
∂
Yφ
∂
+
(Yθ sin θ) + k Y = 0
sin θ ∂φ sin θ
∂θ
244
8. ELLIPTIC EQUATIONS
has a solution which is continuous on the entire surface of the sphere. Thus, this
solution must be periodic in φ, of period 2π, and regular at θ = 0 and θ = π,
i.e., at both these points it must approach a limit which is independent of φ. This
requirement is satisfied for the values of k = n(n + 1), n = 0, 1, 2, . . . , which are
the eigenvalues for this problem. Hence, the solutions are the spherical harmonics
Yn (θ, φ) (see Chapter 9).
The equation for R(r), r2 R − n(n + 1)R + λr2 R = 0, has the solutions
√
√
Jn+1 ( λ r)
√
Rn ( λ r) =
, which are regular at r = 0.
r
8.13. Exercises
8.1. Find Green’s function for the operator ∇2 − k 2 in R, R2 , and R3 .
1
1
Ans. − e−k|x−x | in R; −
K0 (kr) in R2 , where r = |x − x |; and
2k
2π
1 −kr
−
e
in R3 , where r = |x − x |.
4πr
8.2. Find Green’s function for the Laplace operator
in the circle r ≤ a subject to the
Dirichlet boundary conditions, where r = x2 + y 2 .
1
a r1
Ans. In rectangular cartesian coordinates: G (x, y; x , y ) =
ln , where
2π
r r2
a2 x 2 a2 y 2
2
2
2 2
2
2
2
r = x +y , r1 = (x−x ) +(y−y ) , and r2 = x− 2
+ y− 2
.
r
r
 r
 ln + an (r; r ) cos n(θ − θ ) for r < r,

a
In series form: G (r, θ; r , θ ) =

 ln r + an (r ; r) cos n(θ − θ ) for r > r,
a
r n a n r n
1
where an (r; r ) =
−
for n ≥ 1. Note that this
2nπ
a
r
a
result can be obtained from the first answer by using the formula
∞ n
r
a2 − r2
1+2
cos nθ = 2
.
a
a − 2ar cos nθ + r2
n=0
8.3. Find
Green’s function for the Laplace operator in the sphere r ≤ a, where
r = x2 + y 2 + z 2 .
1
1
a |x |
−
.
Ans. G (x; x ) = −
2
4π x − x |x | x − a2 x 8.4. Find Green’s function for the Laplace operator in the semicircle x2 + y 2 ≤ a2 ,
y ≥ 0.
1
r1 R2
Ans. G (x, y; x , y ) =
ln
, where x = r cos θ, y = r sin θ, x =
2π
r2 R1
8.13. EXERCISES
245
r cos θ , y = r sin θ ,
2
2
2
2
r = x + y , r12 = (x − x )2 + (y − y )2 = r 2 − 2rr cos(θ − θ ) + r ,
2 2
2 2
2
4
a x
a y
a r
a
r22 = x − 2
+ y − 2
= r 2 − 2 cos(θ − θ ) + 2 ,
r
r
r
r
2
R12 = (x − x )2 + (y − y )2 = r 2 − 2rr cos(θ + θ ) + r ,
2 2
2 2
2
a x
a y
a r
a4
R22 = x − 2
+ y + 2
= r 2 − 2 cos(θ + θ ) + 2 .
r
r
r
r
8.5. Use Green’s function to find the solution of the problem ∇2 u = 0, in the region
{x2 + y 2 ≤ a2 , y > 0}, subject to the boundary conditions u(x, y) = k on
x2 + y 2 = a2 and on y = 0 for x < 0, and u(x, y) = 0 on y = 0 for x > 0.
a − x
x2 + y 2 − ax k
Ans. u(x, y) = k −
tan−1
− tan−1
.
π
y
ay
8.6.Find Green’s function for the Laplace operator in the upper hemisphere r =
x2 + y 2 + z 2 ≤ a, z ≥ 0.
1
a |x |
1
+
Ans. G (x; x ) = −
+
4π |x − x | 4π |x |2 x − a2 x 4π |x − x1 |
−
a |x |
, where x1 = (x , y , −z ).
2
2
4π |x | x − a x 1
8.7. Use Green’s function for the half-plane (see §8.1) to solve the following problems: (a) ∇2 u = 0, x ∈ R, 0 < y, such that u(x, 0) = g(x).
(b) ∇2 u = 0, x ∈ R, 0 < y, such that u(x, 0) = k.
(c) ∇2 u = 0, x ∈ R, 0 < y, such that u(x, 0) = |x|/x.
∞
∂G(x, x ) Ans. (a) u(x ) =
g (xs ) −
dx; (b) u(x, y) = k; and
∂y
−∞
2
2θ
(c) u(x, y) = tan−1 (x/y) = 1 − .
π
π
2
∂ u 1 ∂u 1 ∂ 2 u
8.8. Solve Laplace’s equation 2 +
+
= 0, subject to the conditions
∂r
r ∂r r 2 ∂θ2
u(r, 0) = f (r) and u(r, π) = 0 for r > 0.
Hint. Reduce the partial differential equation to an ordinary differential equation
by applying the finite Fourier sine transform. Find Green’s function for the
transformed problem, i.e., solve the problem
d2 G (r, r )
G (r, r )
+
r
− n2 G (r, r ) = δ (r, r ) , lim G (r, r ) = 0,
r→0
dr 2
dr
dG
(r,
r
)
dG
(r,
r
)
1
lim G (r, r ) = finite,
−
= 2.
r→0
r
r
r
r +
r −
r2
246
8. ELLIPTIC EQUATIONS
Green’s function is given by
 r n

 r , r ≤ r ,
1
n
G(r, r ) = −
2nr 
 r
, r ≥ r .
r
Obtain the solution to the transformed problem by using Green’s function and
then invert the transform.
1 ∞
r sin θ f (r )
y ∞
f (x )
Ans. u(r, θ) =
dr
=
dx .
2
π 0 r 2 − 2rr cos θ + r π 0 (x − x )2 + y 2
Note that if we remove the restriction on u being zero on half of the x-axis, and
take u = f (x) on the entire x-axis, then we must add the solution for x < 0, to
the above solution. Hence, the solution under this condition is
y ∞
f (x )
u(x, y) =
dx .
π −∞ (x − x )2 + y 2
∂2u ∂ 2u
+ 2 = 0, y > 0, u(x, 0) = f (x), where f (x) is integrable for
∂x2
∂y
all x and approaches zero as x → ±∞.
d2 u
Hint. Apply Fourier transform with respect to x and solve 2 − α2 u
= 0.
dy
y ∞
f (η)
Ans. u(x, y) =
dη.
π −∞ (x − η)2 + y 2
Note that the solution does not directly satisfy the boundary condition at y = 0.
However, it can be shown that limy→0 u(x, y) = f (x). Since
y ∞
f (η)
u(x, y) =
dη
π −∞ (x − η)2 + y 2
∞
1
η−x
1 ∞ η−x
−1
−1
=
f (η) tan
−
f (η) tan
dη,
π
y
π −∞
y
−∞
8.9. Solve
the problem reduces to showing that
1 ∞ η−x
−1
− lim
f (η) tan
dη = f (x).
y→0 π −∞
y
The left side can be written as
x
∞
1
η−x
η−x
− lim
f (η) tan−1
dη +
f (η) tan−1
dη .
y→0 π
y
y
−∞
x
If we interchange the order of integration and the limit process
and let
y → 0,
η−x
π
−1
then in the first integral η − x < 0 and therefore, lim tan
=− ,
y
2
y→0
η−x
π
−1
and in the second integral lim tan
= . Hence,
y→0
y
2
8.13. EXERCISES
1
− lim
y→0 π
∞
f (η) tan
−1
−∞
η−x
y
1
dη =
2
=
247
x
−∞
f (η) dη −
∞
f (η) dη
x
1
[f (x) + f (x)] = f (x).
2
8.10. Find Green’s function for the boundary value problem in the rectangle {0 <
x < a, 0 < y < b}:
Gxx + Gyy = δ(x − x ) δ(y − y ), 0 < x, x < a, 0 < y, y < b, such that
G(0, y) = 0 = G(a, y), G(x, 0) = 0 = G(x, b).
mπx
nπy
Hint. Set G(x, y; x , y ) =
Amn sin
sin
, in the given equaa
b
m=1 n=1
tion and find Amn from
mπ 2 nπ 2
ab
mπx
nπy +
amn = sin
sin
.
a
b
4
a
b
8.11. Use Rodrigue’s formula and show that
π
1
(2n)!
2
(a)
Pn (cos θ) sin θ dθ = 2n
1 − t2
2 (n!)2 −1
0
π
2
(n + m)!
(b)
Pnm (cos θ)2 sin θ dθ =
.
2n
+
1
(n − m)!
0
n
dt =
2
.
2n + 1
8.12. Find Green’s function for the Laplacian in R.
d2 G
Ans. Green’s function for 1-D Laplace’s equation satisfies the equation
=
dx2
1
δ(x, x ), whose general solution for a fixed x is G(x) = |x − x | + Ax + B.
2
If we require the symmetry about x , i.e., G = G(|x − x |), then A = 0, and set
1
B = 0, then Green’s function is given by G(x) = |x − x |.
2
8.13. Derive Green’s function G(x, y) = G(x, y; x , y ) for the half-plane by Fourier
transform for Laplace’s equation.
Hint. Solve the problem
∂2G ∂2G
+
= δ(x − x )δ(y − y ),
∂x2
∂y 2
limx→±∞ G(x, y), Gx (x, y) = 0,
Ans. G(x, y) =
1
log
4π
G(x, 0) = 0,
limy→∞ G(x, y), Gy (x, y) = 0.
(x − x )2 + (y − y )2
, same as in §8.1, Method 2.
(x − x )2 + (y + y )2
8.14. Verify the solution given in §8.12.3 for the Robin problem ∇2 u = f (x) in
∂u
D such that u + α
= g (xs ) on ∂D, where xs is an arbitrary point on the
∂n
boundary ∂D.
8.15. Find Green’s function for the problem
248
8. ELLIPTIC EQUATIONS
∇2 u(r, θ) = F (r, θ) for r < a,
∂u(r, θ) u(a, θ) +
= 0, lim u(r, θ) < +∞.
r→0
∂r
r=a
Ans. G (r, θ; r , θ ) =
∞
Gn (r, θ; r , θ ), where
n=0

1


 2π
G0 (r, θ; r , θ ) =

1


2π
r
1
ln
−
, r < r,
a
a
,
r
1
ln
−
, r > r;
a
a
and
Gn (r, θ; r , θ )

r n
a n r n
1


−
(n
−
a)
+
(n
+
a)
cos nψ, r < r,
 2nπ(n + a)
a
r
a
=
n
a n r n

1
r

−
(n − a)
+ (n + a) cos nψ, r > r,
2nπ(n + a)
a
r
a
where ψ = θ − θ .
8.16. Find Green’s function for the Laplacian in the sector 0 < r < a, 0 < θ < π/3.
Hint. In complex coordinate system the sector is defined by 0 < |z| < a,
0 < arg{z} < π/3, z = x + i y. The conformal mapping w = z 3 , w = u + i v,
maps the sector onto a semi-circle. Use the result of Exercise 8.4.
1
AB
Ans. G (r, θ; r , θ ) =
ln
, where A = r 6 −2r3 r 3 cos 3 (θ − θ )+
4π
CD
r 6 , B = r 6 r 6 −2a6 r 3 r3 cos 3 (θ + θ )+a12 , C = r 6 −2r 3 r 3 cos 3 (θ + θ )+
r 6 , D = r 6 r6 − 2a6 r 3 r 3 cos 3 (θ − θ ).
8.17. Solve the following problem:
∂2φ ∂2φ
+ 2 = 0,
∂x2
∂y
x, y > 0,
φ(x, 0) = f (x),
∂φ
(0, y) = g(y),
∂x
by using the Fourier transform.
Hint. Break up the problem int two parts: Part (1) φxx + φyy = 0, φ(x, 0) = 0,
φx (0, y) = g(y).
Apply the cosine transform with respect to x and note that
αy
e
1 Fc−1
=
γ + ln x2 + y 2 − 2 ln x , where γ is the Euler gamma.
α
2π
Part (2) φxx + φyy = 0, φ(x, 0) = f (x), φx (0, y) = 0. Apply the sine transform
with respect to y. Add the two solutions.
∂2
8.18. Show that Green’s function G(x, y) for the operator L =
+ γ 2 in R2 is
∂x ∂y
√
given by G(x, y) = H(x)H(y) J0 2γ xy , which has a pole at the origin.
Hint. Use the double Fourier transform.
8.13. EXERCISES
249
√
Note that the classical solution for the equation L[u](x, y) = 0 is u = J0 (2γ xy)
for x, y > 0.
8.19. Solve uxx + uyy = 0, under the conditions u(0, y) = 0 = u(π, y), u(x, 0) =
sin x, lim u(x, y) < +∞.
y→∞
Ans. u = e−y sin x.
8.20. Solve r 2 urr +rur +uθθ = 0, such that u(b, θ) = f (θ), u(r, θ +2π) = u(r, θ),
and lim u(r, θ) < +∞ (circular disc problem).
r→0
Hint: Separate the variables
and show that the only relevant part of the solution
reduces to u(r, θ) = c0 + α r α (A(α) cos αθ + B(α) sin αθ). Note that under
the given conditions u(r, θ) must have a Fourier series representation in θ and,
therefore, α = n is a positive integer.
8.21. Solve uxx +uyy = 0, under the conditions u(x, 0) = 0 = u(x, π), u(0, y) = 0,
and u(π, y) = cos2 y.
π
∞
2
Ans. u =
Cn sinh nx sin ny, where Cn =
cos2 y sin nydy =
π
sinh
nπ
0
n=1
1
1
n n
1 − (−1)
−
, n = 2, and C2 = 0.
π sinh nπ
n n2 − 4
1
1
8.22. Solve urr + ur + 2 uθθ = 0, subject to the conditions u = 0 for θ = 0 or
r
r
π/2, and ur = sin θ at r = a.
∞
4(−1)n+1
Ans. u =
Cn r 2n sin 2nθ, where Cn =
.
π(4n2 − 1)a2n
n=1
8.23. Solve rurr +ur +ruzz = 0, u(a, z) = u0 , under the conditions u(a, 0) = u0 ,
u(r, 0) = 0 = u(r, h), and lim u(r, z) < +∞ (steady-state temperature in a finite
r→0
cylinder).
∞
4u0
I0 ((2n + 1)πr/h)
(2n + 1)πz
Ans. u =
sin
.
π n=0 (2n + 1) I0 ((2n + 1)πa/h)
h
8.24. Solve Poisson’s equation uxx + uyy = −1, 0 < x, y < 1, subject to the
Dirichlet boundary conditions u(0, y) = 0 = u(1, y) = u(x, 0) = u(x, 1).
∞
16
sin jπx sin kπy
Ans. u(x, y) = 4
.
π
jk (j 2 + k 2 )
j,k=1
j,k odd
er1 −r2
is Green’s function for the differential
4π |r1 − r2 |
G(r1 , r2 ) = −δ(r1 − r2 ).
8.25. Show that G(r1 , r2 ) =
equation ∇2 + k 2
Hint. Show that (i) G(r1 , r2 ) satisfies the homogeneous differential
!
0, r1 = r2 ,
equation for r1 = r2 , and (ii) ∇2 + k 2 G(r1 , r2 ) dr1 =
−1, r1 = r2 ,
for small |r1 − r2 |.
250
8. ELLIPTIC EQUATIONS
8.26. Show that Green’s functions for the radial part of 3-D Helmholtz’s equation,
1
with singularity or source at the origin, are: (a) G(r) = −
y0 (kr) for the
4π
i (1)
standing wave; and (b) G(r) =
h (kr) for the outgoing wave.
4π 0
Hint. Assume the time dependence ei ωt .
π
π
Note that yn (x) =
Yn+1/2 (x) = (−1)n
Jn−1/2 (x), and
2x
2x
π (1)
(1)
hn (x) =
H
(x) are the spherical Bessel functions.
2x n+1/2
9
Spherical Harmonics
We will start with a short historical account of the development of the spherical
harmonics. Laplace’s equation in a sphere was solved by Bernoulli’s separation
method in §4.3.3 and §8.10. The spherical harmonics are the angular portion (θ and
φ portions of the spherical coordinates) of a set of solutions of Laplace’s equation.
These harmonics are useful in many theoretical and physical applications, namely, in
physics, seismology, geodesy, spectral analysis, magnetic fields, quantum mechanics
and others. A detailed account of various approaches to spherical harmonics can be
found in Courant and Hilbert [1968] and MacRobert [1967].
9.1. Historical Sketch
The development of this subject started with Laplace’s own account of a function
Ynm (θ, φ) as a set of spherical harmonics that form an orthogonal system. This
special research was developed by Laplace in 1782 in connection with the Newtonian
potential for the law of universal gravitation in R3 , when he determined that the
gravitational potential P (x) associated with a set of point-masses mi located at points
mi
xi ∈ R3 is defined by P (x) =
, where each term in this summation is
|x − xi |
i
a Newtonian potential at the respective point mass. About the same time Legendre
had determined that the expansion for the Newtonian potential in powers of r = |x|
and r1 = |x1 | is given by
|x − x1 |−1 = P0 (cos γ)
1
r
r2
+ P1 (cos γ) 2 + P2 (cos γ) 3 + · · · ,
r1
r1
r1
where γ is the angle between x and x1 , and Pn are the well known Legendre polynomials, which are also a special case of spherical harmonics. The name ‘solid spherical
harmonics’ was introduced in 1867 by William Thomson (Lord Kelvin) and Peter G.
Tait in their book Treatise on Natural Philosophy to describe these functions which
251
252
9. SPHERICAL HARMONICS
are the solution of homogeneous Laplace’s equation ∇2 u = 0 in the sphere. The term
‘Laplace’s coefficients’ is generally used to describe the zonal spherical harmonics
as introduced by Laplace and Legendre.
9.2. Laplace’s Solid Spherical Harmonics
A function F (x, y, z) is said to be homogeneous of degree n if for any constant
factor λ we have F (λx, λy, λz) = λn F (x, y, z). The function F (x, y, z) is called
a solid spherical harmonic of degree n if F (x, y, z) is homogeneous of degree n
and if ∇2 F (x, y, z) = 0. By this definition the spherical harmonics are the angular
portion of the solution of Laplace’s equation in a sphere.
Recall that Laplace’s equation in spherical coordinates (r, θ, φ), r > 0, 0 ≤ θ ≤
π, 0 ≤ φ < 2π, is
1 ∂
1
∂
∂u
1
∂2u
2
2 ∂u
∇ u≡ 2
r
+ 2
sin θ
+ 2 2
= 0, (9.1)
r ∂r
∂r
r sin θ ∂θ
∂θ
r sin θ ∂φ2
which is the same as Eq (8.50). In §4.3.3 and §8.10 we have used Bernoulli’s separation
method to determine Green’s function for the sphere. Let the solution of Eq (9.1) be
of the form u = R(r)Θ(θ)Φ(φ). Substituting it into this equation we obtain
r sin2 θ d2 (rR) sin θ d
dΘ
1 d2 Φ
+
sin
θ
=
−
.
R
dr 2
Θ dθ
dθ
Φ dφ2
Since the left side of this equation does not contain φ and the right side contains φ
only, it must be a constant, say m2 . Then the above equation is equivalent to the two
equations
d2 Φ(φ)
+ m2 Φ = 0,
dφ2
r d2 (rR)
1
d
dΘ
m2
+
sin
θ
−
= 0.
R dr2
Θ sin θ dθ
dθ
sin2 θ
(9.2)
(9.3)
The general solution of Eq (9.2) is Φ = A cos mφ + B sin mφ. Since the first term
of (9.3) does not involve θ and the other two terms do not involve r, then each part
must be constant, which we take as n(n + 1) (see Example 4.11), and Eq (9.3) breaks
up into
d2 (rR)
− n(n + 1)R = 0,
dr2 1 d
dΘ
m2
sin θ
+ n(n + 1) −
Θ = 0.
sin θ dθ
dθ
sin2 θ
r
(9.4)
(9.5)
The solution of Eq (9.4) is is R = A1 r n + B1 r −n−1 . If in Eq (9.5) we set cos θ = t
and get
d
dΘ
m2
(1 − t2 )
+ n(n + 1) −
Θ = 0,
(9.6)
dt
dt
1 − t2
9.2. LAPLACE’S SOLID SPHERICAL HARMONICS
253
m
dm Pn (t)
m d Pn (t)
which has a particular solution Θ = (1 − t2 )m/2
=
sin
θ
, if
dtm
dtm
m and n are positive integers and m < n + 1. A second but less useful particular
dm Qn (t)
solution of Eq (9.6) is Θ = (1 − t2 )m/2
. Combining these solutions the
dtm
two particular solutions of Eq (9.1) are
dm
Pn (t),
dtm
dm
u = r −n−1 (A cos mφ + B sin mφ) sinm θ m Pn (t).
dt
u = r n (A cos mφ + B sin mφ) sinm θ
(9.7)
(9.8)
m
dm Pn (t)
2 m/2 d Pn (t)
The function sinm θ
or
(1
−
t
)
is represented by Pnm (t) and
dtm
dtm
is called the associated Legendre function of the first kind of order m and degree n.
By differentiating m-times the value of Pn (t) we get the formula
Pnm (t)
(2n)! sinm θ n−m (n − m)(n − m − 1) n−m−2
= n
t
−
t
2 n! (n − m)!
2 · (2n − 1)
(n − m)(n − m − 1)(n − m − 2)(n − m − 3) n−m−4
+
t
− ···
2 · 4 · (2n − 1)(2n − 3)
where the expression within the square brackets ends with the terms involving t0 if
n − m is even and with the term involving t if n − m is odd. Tables of some values
of Pnm (t) are given in Appendix C.
The functions cos mφPnm (t) = cos mφ sinm θ
dm Pn (t)
and sin mφPnm (t) =
dtm
dm Pn (t)
sin mφ sinm θ
are called tesseral harmonics of degree n and order m.
dtm
There are obviously 2n + 1 tesseral harmonics of order zero and degree n, viz.,
Pn (t),
dPn (t)
,
dt
d2 Pn (t)
cos 2φ sin2 θ
,
dt2
3
d Pn (t)
cos 3φ sin3 θ
,
dt3
···
cos φ sin θ
cos mφ sinm θ
dm Pn (t)
,
dtm
dPn (t)
dt
d2 Pn (t)
sin 2φ sin2 θ
dt2
3
d
P
n (t)
sin 3φ sin3 θ
dt3
···
sin φ sin θ
sin mφ sinm θ
dm Pn (t)
.
dtm
Note that Pn (t) is a tesseral harmonic of order zero and degree n and is sometimes
written as Pn0 (t). If each of these is multiplied by a constant and their sum taken, this
sum is called a surface spherical harmonic of degree n, and is a solution of Eqs (9.4)(9.5). This sum is represented by the functions Yn (t, φ) or Yn (θ, φ). The functions
r n Ynm (t, φ) and r−n−1 Ynm (t, φ), which are called solid spherical harmonics of
254
9. SPHERICAL HARMONICS
degree n, are solutions of Laplace’s equation (9.1). Note that a tesseral harmonic is a
special case of a surface spherical harmonic, and a zonal harmonic a special case of
a tesseral harmonic.
The solutions of the Θ and Φ equations (9.2)-(9.3) are combined by multiplication to give the complete angular dependence of the wave functions. These angular
functions are a product of the trigonometric functions, represented as a complex exponential ei mφ and the associated Legendre functions Pnm (cos θ) of the first kind and
order m, which can be represented as
Ynm (θ, φ) = N ei mφ Pnm (cos θ),
(9.9)
where N is a normalization factor which is defined by almost similar expressions,
but not uniquely, in different branches of science and technology (see §9.2.1), and θ
and φ represent colatitude (or polar angle) and latitude (or azimuth), respectively.
The colatitude ranges from 0 at the north pole to π at the south pole, and π/2 at the
equator, whereas the latitude assumes all values 0 ≤ φ < 2π. The function Ynm (θ, φ)
are called spherical harmonic functions of degree n and order m. It can be verified
that
n(n + 1) m
∇2 Ynm (θ, φ) = −
Yn (θ, φ).
(9.10)
r2
The solutions of the R-equation (4.12) involving r are called the radial wave functions Rnm (r), and thus, the overall solutions are u(r, θ, φ) = Rnm (r)YNm (θ, φ). If
Laplace’s equation (9.1) is solved on the surface of the sphere, the periodic boundary
conditions in φ and the regularity conditions at both north and south poles are satisfied. This in turn ensures that the degree n and the order m are integers such that
n ≥ 0 and |m| ≤ n.
If the function u in Eq (9.1) were defined for θ ≥ 0, the resulting spherical
harmonics would have been defined for integer order and non-integer degree. The
general solution of Laplace’s equation (9.1) in a sphere with center at the origin
and radius a is a linear combination of the spherical harmonic functions Ynm (θ, φ)
multiplied by a scaling factor rn :
u(r, θ, φ) =
∞ n
n m
cm
n r Yn (θ, φ),
(9.11)
n=0 m=−n
n m
where cm
n are constant and the factors r Yn (θ, φ) are known as solid spherical
harmonics. The representation (9.11) is valid in the sphere
r<a=
1
.
1/n
lim sup |cm
n|
n→∞
9.2.1. Orthonormalization. Different normalizations for Laplace’s spherical
functions are used in various fields of science and technology. Some of the common
conventions are as follows:
9.2. LAPLACE’S SOLID SPHERICAL HARMONICS
255
(i) In physics and seismology these functions are generally defined as
(2n + 1)((n − m)! i mφ m
e
Pn (cos θ).
4π (n + m)!
Ynm (θ, φ) =
(9.12)
These functions are orthonormal, i.e.,
π
θ=0
2π
Ynm Ynm dS = δnn δmm ,
φ=0
(9.13)
where Ynm denotes the complex conjugate of Ynm , dS = dθ dφ is an element of the
surface of the sphere, and δnn , δmm is the Kronecker delta.
(ii) In geodesy and spectral analysis these functions are defined as
Ynm (θ, φ) =
such that
1
4π
(2n + 1)
π
θ=0
2π
(n − m)! i mφ m
e
Pn (cos θ),
(n + m)!
φ=0
Ynm Ynm dS = δnn δmm
(9.14)
(9.15)
has unit value for n = n and m = m .
(iii) In magnetics and quantum mechanics, these functions are defined as
Ynm (θ, φ) =
(n − m)! i mφ m
e
Pn (cos θ),
(n + m)!
(9.16)
with the normalization
2n + 1
4π
π
θ=0
2π
φ=0
Ynm Ynm dS = δnn δmm .
(9.17)
The functions Ynm (θ, φ) defined by (9.16) are known as Schmidt semi-normalized
harmonics in magnetics, and the normalization (9.17) is known as Racah’s normalization in quantum mechanics.
Theorem 9.1. The integral over the surface of the unit sphere of the
product of two surface spherical harmonics of different degrees is zero, i.e.,
2π
1
dφ
0
−1
Yn (t, φ) Ym (t, φ) dθ = 0.
(9.18)
256
9. SPHERICAL HARMONICS
Proof. As we have seen, u = r n Yn (t, φ) and v = r m Ym (t, φ) are solutions
of Laplace’s equation Hence, by Green’s second identity (1.16) and formula (1.17),
where ∇2 u = 0 = ∇2 v, we have on the surface Γ of the unit sphere
∂v
∂u
u
−v
dS = 0,
∂n
∂n
Γ
∂v
∂v
∂u
∂u
=
= mrm−1 Ym (t, φ),
=
= nr n−1 Yn (t, φ),
∂n
∂r
∂n
∂r
∂v
∂u
−v
= (m − n)r n+m−1 Yn (t, φ)Ym (t, φ) = (m − n)Yn (t, φ)Ym (t, φ).
∂n
∂n
and thus,
(m − n)
Yn (t, φ)Ym (t, φ) = (m − n)
Γ
2π
1
dφ
−1
0
Yn (t, φ) Ym (t, φ) dt = 0,
which for n = m yields (9.18).
All the above normalized spherical harmonics satisfy the following relation:
Ynm (θ, φ) = (−1)m Yn−m (θ, φ).
(9.19)
The proof is left as an exercise.
9.2.2. Condon-Shortley Phase Factor. The associated Legendre functions
Pnm (cos θ), which occur in the spherical harmonic functions Ynm (θ, φ) as defined in
(9.9) and subsequent conventions in different branches of science and technology,
are related to the Legendre polynomials of the first kind Pn (x), x = cos θ, by the
relation:
Pnm (x)
= 1−x
2 m/2
1 − x2
dm
P
(x)
=
n
dxm
2n n!
m/2
dm+n
x2 − 1
dxm+n
n
. (9.20)
Since
Pn−m (cos θ) = (−1)m
(n − m)! m
P (cos θ),
n + m)! n
−n ≤ m ≤ n,
(9.21)
(see Exercise 9.6), the orthonormalization (9.12), and others, hold for both positive
and negative values of m, within a possible factor (−1)m . In view of (9.21) Yn−m
and Ynm differ in phase by (−1)m . This factor is known as the Condon-Shortley
phase factor that occurs in some definitions of spherical harmonics, while there is a
lack of inclusion of this factor in some other definitions of the associated Legendre
polynomials. In fact, in certain quantum mechanical calculations, particularly in
the quantum theory of angular momentum, this phase factor is usually associated
with positive m spherical harmonic, while in many other applications of spherical
9.2. LAPLACE’S SOLID SPHERICAL HARMONICS
257
harmonics it is a general practice to ignore these phase factor by using (9.12) for
m ≥ 0 and Yn−m = Ynm∗ for negative subscript.
Omitting the Condon-Shortley convention in the definition of spherical harmonic
function after using it in the definition of Pnm (cos θ) gives the representation (9.12),
whereas using it in this definition will yield
(2n + 1)((n − m)! i mφ m
m
m
Yn (θ, φ) = (−1)
e
Pn (cos θ).
(9.22)
4π (n + m)!
The Condon-Shortley phase is not necessary in the definition of spherical harmonics but its inclusion simplifies the treatment of angular moment in quantum mechanics
[Arfken, 1985:448]. The Condon-Shortley phase factor is never used in geodesy and
magnetics in their definition of spherical harmonics. For more results on associated
Legendre functions of the first kind with and without the inclusion of this factor, and
a list of spherical harmonics, see §C.3.1.
Instead of deciding on whether to include or ignore the Condon-Shortley phase
factor, a real basis of spherical harmonics is provided by setting
 0
Yn (θ, φ) if m = 0,



√

1


√ [Ynm (θ, φ) + (−1)m Yn−m (θ, φ)] = 2 N (n, m)Pnm (µ) cos mφ


 2
Ynm (θ, φ) =
if m > 0,


1

 √ [Y −m (θ, φ) − (−1)m Y −m (θ, φ)]= √2N (n, m)P −m (µ) sin mφ


n
n
n

 i 2

if m < 0,
where µ = cos θ, 0 ≤ θ ≤ π, and N (n, m) denote the normalization constants as
a function of n and m. The real form has the associated Legendre function Pnm for
m ≥ 0, while the imaginary form occurs only for m < 0. The spherical harmonics
for m > 0 are said to be of cosine type, and those with m < 0 of the sine type.
These functions have the same orthonormalization properties as described in §9.2.1.
9.2.3. Spherical Harmonics Expansion. Since Laplace’s special harmonics
form a complete set of orthonormal functions, they form an orthonormal basis of the
Hilbert space of square-integrable functions. On the unit sphere any square-integrable
function can be expanded as a linear combination of these special harmonics, i.e.,
u(θ, φ) =
∞ n
m
cm
n Yn (θ, φ),
(9.23)
n=0 m=−n
where cm
n are constants. This expansion is exact as long as n → ∞ and converges
uniformly to u(θ, φ). Thus, a square-integrable function can be expanded in terms
of the real harmonics Ynm (θ, φ) in the form of the series (9.23). However, truncation
errors arise when limiting the summation over n to a finite bandwidth of length N . The
258
9. SPHERICAL HARMONICS
expansion (9.23) is valid in the sense of mean-square convergence, i.e., convergence
in L2 , which implies that
2π
lim
n→∞
0
0
π
N
2
m
cm
Y
(θ,
φ)
u(θ, φ) −
sin θ dθ dφ = 0.
n n
n=0
In the above expansion the coefficients cm
n are analogous to the Fourier coefficients,
and can be obtained by multiplying the above equation by the complex conjugate
of Ynm (θ, φ), then integrating over the solid angle Ω, and using the orthogonality
relations of Ynm (θ, φ). Thus, in the case of orthonormalized harmonics we have
cm
n =
Ω
u(θ, φ)Ynm (θ, φ) dΩ =
2π
0
sin θ u(θ, φ)Ynm (θ, φ) dθ.
(9.24)
If the coefficients cm
n decay in n sufficiently rapidly, say exponentially, then the series
(9.23) also converges uniformly to u(θ, φ).
9.2.4. Addition Theorem. The addition theorem for spherical harmonics
is a generalization of the 2-D trigonometric identity cos(θ − θ ) = cos θ cos θ +
sin θ sin θ . In this theorem the function cos(θ − θ ) on the left side is replaced
by the Legendre polynomials whereas the trigonometric functions on the right side
are replaced by a product of the spherical harmonic functions and their complex
conjugates. Thus,
Theorem 9.4. Let x and y be two unit vectors with spherical angular coordinates (θ, φ) and (θ , φ ), respectively. Then
n
4π
Pn (x · y) =
Ynm (θ, φ) Ynm (θ , φ ),
2n + 1 m=−n
(9.25)
where Pn are the Legendre polynomials of first kind and order n. This result holds
for both real and complex harmonics. In fact, it is valid for any orthogonal basis of
special harmonics of degree n. For unit power harmonics the factor 4π appearing
in (9.25) must be removed. A proof of this theorem can be found in Whittaker and
Watson [1962:395], where the properties of the Poisson kernel in the unit sphere are
used, and then the right-hand side is calculated.
As a particular case when x = y, the above theorem reduces to Unsöld’s theorem
[Unsöld, 1927], which states:
n
m=−n
Ynm (θ, φ) Ynm (θ , φ ) =
2n + 1
.
4π
(9.26)
This result is a generalization of the 2-D trigonometric identity cos2 θ + sin2 θ = 1
for 3-D spherical harmonics.
9.2. LAPLACE’S SOLID SPHERICAL HARMONICS
259
9.2.5. Laplace’s Coefficients. We have seen that for a point (x1 , y1 , z1 ) the
function
−1/2
u = (x − x1 )2 + (y − y1 )2 + (z − z1 )2
is a solution of Laplace’s equation ∇2 u = 0. This function u in spherical coordinates
becomes
−1/2
u = r 2 − 2rr1 cos θ cos θ1 + sin θ sin θ1 cos(φ − φ1 ) + r12
,
which is a solution of Eq (9.1). If γ is the angle between the radius vectors r and r1 for
−1/2
the points (x, y, z) and (x1 , y1 , z1 ), then we can write u = r 2 −2rr1 cos γ +r12
.
Notice that cos γ = cos θ cos θ1 + sin θ sin θ1 cos(φ − φ1 ). Also, u = Pn (cos γ) is a
solution of Eq (9.6), and u = r n Pn (cos γ) and r −n−1 Pn (cos γ) are solutions of Eq
(9.1). If we transform our coordinates keeping the origin fixed and taking as our new
polar axis the radius vector of (x1 , y1 , z1 ), then γ becomes our new θ and Pn (cos γ)
reduces to Pn (cos θ), which is our new surface zonal harmonic, or a Legendrian, of
degree n. It is a Legendrian with its axis not the original polar axis but the radius
vector of (x1 , y1 , z1 ). Since a Legendrian is a surface spherical harmonic,
Pn (cos γ) = Pn [cos θ cos θ1 + sin θ sin θ1 cos(φ − φ1 )]
is a surface spherical harmonic of degree n. This harmonic is of a very special form
since, being a determinate function of t, φ, t1 and φ1 , it contains but two arbitrary constants if we regard it as a function of t and φ, instead of containing 2n+1. This function
is known as Laplace’s coefficient1 of degree n and denoted by Ln (t, φ, t1 , φ1 ). The
radius vector (x1 , y1 , z1 ) is called the axis of the Laplace’s coefficient and the point
where the axis cuts the surface of the unit sphere is called the pole of the Laplace’s
coefficient.
Theorem 9.2. If the product of a surface spherical harmonic of degree n
by a Laplace’s coefficient of the same degree is integrated over the surface of
4π
the unit sphere, the result is equal to
times the value of the spherical
2n + 1
harmonic at the pole of the Laplace’s coefficient, that is,
2π
π
4π
dφ
Yn (t, φ)Ln (t, φ, t1 , φ1 ) dθ =
Yn (t1 , φ1 ).
(9.27)
2n
+1
0
0
Proof. Transform to the axis of the Laplace’s coefficient as a new axis, and let
Zn (t, φ) be the transformed spherical harmonic. Then Ln (t, φ, t1 , φ1 ) will become
Pn (t) and (9.27) will be proved if we can show that
2π
1
dφ
0
1
−1
Zn (t, φ)Pn (t) dt =
4π
Zn (1, 0).
2n + 1
(9.28)
During the nineteenth century this function was also known as the Laplacian, but this
nomenclature was abandoned when Laplace’s operator ∇2 was referred to as ‘Laplacian’.
260
9. SPHERICAL HARMONICS
Now,
n
2 Zn (t, φ)Pn (t) = A0 Pn (t) +
(Am cos mφ + Bm sin mφ) Pnm (t)Pn (t),
m=1
2π
2
Zn (t, φ)Pn (t) dφ = 2πA0 [Pn (t)] ,
0
1
2π
dt
−1
Zn (t, φ)Pn (t) dφ =
0
4π
A0 .
2n + 1
But Zn (1, 0) = A0 , because Pn (1) = 1 and Pnm (1) contains (1 − 1)m/2 as a factor
and is equal to zero. Hence (9.28) is proved.
Theorem 9.3. The Laplace’s coefficient Ln (t, φ, t1 , φ1 ) has the following
series development in terms of spherical harmonics:
n (n − m)! m
m
Ln (t, φ, t1 , φ1 ) = Pn (t)Pn (t1 ) + 2
P (t)Pn (t1 ) cos m(φ − φ1 ) .
(n + m)! n
m=1
(9.29)
Proof. Since
Ln (t, φ, t1 , φ1 ) = Pn (cos γ) = Pn (cos θ cos θ1 + sin θ sin θ1 cos(φ − φ1 )
∞ n
m
=
A0,n Pn (t) +
(Am,n cos mφ + Bm,n sin mφ) Pn (t) ,
n=0
m=1
where
A0,n =
=
Am,n =
=
Bm,n =
=
1
2n + 1 2π
dφ
Ln (t, φ, t1 , φ1 )Pn (t) dt
4π
0
−1
2n + 1 4π
Pn (t1 ) = Pn (t1 ),
4π 2n + 1
1
2n + 1 (n − m)! 2π
dφ
Ln (t, φ, t1 , φ1 ) cos mφ Pnm (t) dt
2π (n + m)! 0
−1
2(n − m)!
cos mφ1 Pnm (t1 ),
(n + m)!
1
2n + 1 (n − m)! 2π
dφ
Ln (t, φ, t1 , φ1 ) sin mφ Pnm (t) dt
2π (n + m)! 0
−1
2(n − m)!
sin mφ1 Pnm (t1 ),
(n + m)!
and A0,n = Am,n = Bm,n = 0 for n = m. The result follows from these steps.
9.3. SURFACE SPHERICAL HARMONICS
261
Each term of the Laplace’s coefficient Ln involves a numerical coefficient, plus
a factor which is a function of t, a second factor which is the same function of t1 ,
and a third factor which is of the form cos m(φ − φ1 ). Table 9.1 gives a first few
Laplace’s coefficients, where C(m) denotes the coefficient of cos m(φ − φ1 ) for
m = 0, 1, 2, 3, 4.
Table 9.1.
L0 L1
C(0) 1
C(1)
C(2)
C(3)
C(4)
L2
1 2
t
4 (3 − 1)
1
2 12
(1 − t ) 3t(1 − t2 ) 2
3
2
4 (1 − t )
L3
L4
1
3
4 (5t − t)
1
3
(1 − t2 ) 2 (5t2
8
15
2
4 t(1 − t )
3
5
(1 − t2 ) 2
8
1
4
2
64 (35t − 30t + 3)
5
2 12
(1 − t ) (7t3 − 3t)
8
5
2
2
16 (1 − t )(7t − 1)
3
35
(1 − t2 ) 2
8
35
2 2
64 (1 − t )
− 1)
9.3. Surface Spherical Harmonics
Consider Laplace’s equation (9.1) which we multiply by r2 and substitute Θ(θ) for
u. Then we obtain
∂ 2 (rΘ)
1 ∂
∂Θ
1 ∂ 2Θ
r
+
sin
θ
+
= 0.
(9.30)
2
∂r
sin θ ∂θ
∂θ
sin2 θ ∂φ2
In order to get surface spherical harmonics, we will solve this equation by setting
Θ = r n Ψn .
(9.31)
Note that if Θ is a solid spherical harmonic of degree n, then the functions Ψn which
are related to Θ by the relation (9.31) are called surface spherical harmonics. The
term ‘surface’ signifies the fact that for a constant value of r, the function Θ reduces
to a function defined on the surface of a sphere of radius r, and therefore, Ψn defines
Θ on the surface of the unit sphere since Ψ = Θ/rn . Now, substituting (9.31) into
(9.30) and dividing by r n , we obtain the equation for the surface spherical harmonics
for a fixed r as:
1 ∂
∂Ψn
1 ∂ 2 Ψn
sin θ
+ n(n + 1)Ψn +
= 0.
(9.32)
sin θ ∂θ
∂θ
sin2 θ ∂φ2
As in solving Eq (9.3), we set x = cos θ in (9.32) where |x| ≤ 1, and get
∂
1 ∂ 2 Ψn
2 ∂Ψn
1−x
+ n(n + 1)Ψn +
= 0.
∂x
∂x
1 − x2 ∂φ2
An explicit expression for surface spherical harmonics is obtained if in this equation
we put Ψn (x, φ) = X(x)Φ(φ) and after separating the variables equate both sides to
−m2 . This leads to the equations
262
9. SPHERICAL HARMONICS
Φ + m2 Φ = 0,
d m2
2
1 − x X + n(n+1)−
X = 0.
dx
1 − x2
(9.33)
(9.34)
We will show that Eq (9.34) is the associated Legendre’s equation. If we make the
m/2
substitution X = 1 − x2
y, we obtain
1 − x2 y − 2(m + 1)xy + (n − m)(n + m + 1)y = 0,
(9.35)
which when differentiated m times with respect to x by using Leibniz’s rule1 gives
1 − x2 u − 2(m + 1)xu + (n − m)(n + m + 1)u = 0,
(9.36)
m
dm
2 m/2
2 d Pn (x)
where u =
P
(x).
Now
let
X(x)
=
(1
−
x
)
u(x)
=
(1
−
x
)
.
n
dxm
dxm
Then solving for u and differentiating, we get
mxX
u = X +
(1 − x2 )−m/2 ,
1 − x2
2mxX mX
m(m + 2)x2 X
u = X +
+
+
(1 − x2 )−m/2 .
1 − x2
1 − x2
(1 − x2 )2
Substituting into Eq (9.36), we find that the new function X satisfies the differential
equation (9.35), which is known as the associated Legendre’s equation. Eq (9.35)
reduces to Legendre’s equation for m = 0. If n is an integer, and y = Pn (x) are the
Legendre polynomials, then the solution of this equation, and therefore of Eq (9.35),
dm
are given by y =
Pn (x), and finally the solution of Eq (9.34) are given by the
dxm
associated Legendre functions of the first kind, i.e.,
X = Pnm (x) = 1 − x2
m/2
dm
Pn (x).
dxm
(9.37)
Note that the solutions (9.37) are not always real-valued for |x| > 1. From the above
form for Pnm (x) we will expect m to be non-negative. However, differentiating a
negative number of times being not defined, we use Rodrigue’s formula (C.2) to
express Pn (x); this limitation on m is relaxed and we may have −n ≤ m ≤ n; so
m is both positive and negative. A few values of Pnm (x) are given in §C.2.2. A
generating function for the associated Legendre functions is
(2m)! (1 − x2 )m/2
2m m! (1 − 2xt + t2 )
m+1/2
=
∞
m
Pm+k
tk ,
k=0
1 Leibniz’s rule for the nth derivative of a product is defined in §1.8.1(b). Note that in
the present case Dm (1 − x2 )y = (1 − x2 )(D m+2 y) − 2mx(D m+1 y) − m(m − 1)(D m y); and
D m xy = x(D m+1 y) + m(Dm y).
9.3. SURFACE SPHERICAL HARMONICS
263
but, unlike the generating function for the Legendre polynomials, it is not used much.
Recurrence formulas and orthogonality relation for Pnm (x) are given in §C.2. Using
the definition in (9.37) and Rodrigue’s formula (f.2) for Pn (x), we have, with ξ =
x2 − 1,
1
−1
Pkm (x)Pnm (x) dx
(−1)m
= k+n
2
k! n!
1
ξm
−1
dk+m k dn+m n
ξ
ξ dx.
dxk+m dxn+m
If k = n, let us assume k < n. Notice that since the superscript is m for both
polynomials, while integrating repeatedly all integrated parts will vanish so long as
there is a factor ξ = x2 − 1. If we integrate (n + m)-times, we obtain
k+m
dn+m
m d
k
ξ
ξ ξ n dx.
n+m
dxk+m
−1
−1 dx
(9.38)
If we expand the integrand on the right side by Leibniz’s formula, we get
1
Pkm (x)Pnm (x) dx
(−1)m (−1)n+m
=
2k+n k! n!
1
n+m
k+m
dn+m
(n + m)!
dn+m−j m dk+m+j k
m d
k
n
ξ
ξ
ξ
=
ξ
ξ
ξ .
dxn+m
dxk+m
j! (n + m − j)! dxn+m−j
dxk+m+j
j=0
n
(9.39)
Since the term ξ m contain no power of x greater than x2m , we must have n +m − j ≤
2m, or the derivative will vanish. Similarly, k + m + j ≤ 2k. Thus, the conditions
for a nonzero solution are: j ≥ n − m, or j ≤ k − m. If k < n, as we have assumed,
there is no solution and the integral vanishes. The same result is obtained for k > n.
Now, if k = n, we have only one term that corresponds to j = n − m. Substituting
Eq (9.39) into Eq (9.38), we obtain
1
−1
2
[Pnm (x)] dx =
(−1)n+2m (n + m)!
2n
2 (n!)2 (2m)!(n − m)!
1
ξn
−1
d2m m
ξ
dx2m
d2n n
ξ
dx.
dx2n
(9.40)
d2m
Since ξ m = (x2 − 1)m = x2m − mx2m−1 + · · · , we have 2m ξ m = (2m)!, and
dx
eq (9.40) becomes
1
−1
2
[Pnm (x)] dx =
(−1)n+2m (2n)! (n + m)!
22n (n!)2 (n − m)!
where the integral on the right equals (−1)
0
−1
Pkm (x)Pnm (x) dx =
ξ n dx,
−1
π
Combining these results we obtain the orthogonality relation
1
1
sin2n+1 θ dθ =
n
(−1)n 22n+1 (n!)2
.
(2n + 1)!
2 (n + m)!
δk,n ,
2n + 1 (n − m)!
264
9. SPHERICAL HARMONICS
which is the same as given in §C.2.4, or in spherical coordinates
π
2 (n + m)!
Pkm (cos θ)Pnm (cos θ) dθ =
δk,n .
2n + 1 (n − m)!
0
(9.41)
A similar orthogonality relation for the associated Legendre functions with the same
lower index but different upper indices is
1
−1
Pnm (x)Pnk (x)(1 − x2 )−1 dx =
(n + m)!
δm,k ,
m(n − m)!
(9.42)
where a new wight factor (x2 − 1)−1 appears. The proof of this result is left as an
exercise.
Example 9.1. We will solve the boundary value problem of steady temperature distribution in a spherical medium of unit radius. Let u(r, θ, φ) represent the
temperature, subject to the boundary condition u(1, θ, φ) = f (θ, φ). The solutions
of Eqs (9.33)-(9.34), written as
m
Ψm
n = (A cos mφ + B sin mφ) Pn (cos θ),
represents a surface spherical harmonic. Then the potential function u = r n Ψm
n can
be represented as
u(r, θ, φ) =
∞ n
r n (anm cos mφ + bnm sin mφ) Pnm (cos θ).
(9.43)
n=0 m=0
We will solve Eq (9.43) subject to the prescribed boundary condition at r = 1. The
coefficients anm and bnm in (9.43) will be determined by a Fourier-type expansion
of the function f (θ, φ):
f (θ, φ) =
∞ n
(anm cos mφ + bnm sin mφ) Pnm (cos θ).
(9.44)
n=0 m=0
We will assume that this expansion is possible and the series on the right side of (9.44)
converges uniformly to f (θ, φ). If we multiply (9.44) by cos kφ, k = 0, 1, 2, . . . , and
integrate from −π to π, we obtain
π
f (θ, φ) cos kφ dφ = π
−π
∞
ank Pnk (cos θ).
(9.45)
n=0
Similarly, multiplying (9.44) by sin kφ, k = 0, 1, 2, . . . , we get
π
f (θ, φ) sin kφ dφ = π
−π
∞
n=0
bnk Pnk (cos θ).
(9.46)
9.3. SURFACE SPHERICAL HARMONICS
265
Since Pnk are solutions of the associated Legendre’s equation (9.34), we have
k
d
k2
2 dPn
1−x
+ n(n + 1) −
P k = 0,
dx
dx
1 − x2 n
k
d
dPm
k2
1 − x2
+ m(m + 1) −
P k = 0.
dx
dx
1 − x2 m
k
Now, we multiply the first equation by Pm
and the second by Pnk and subtract one
from another, we get
k
dPnk
d
dPm
k d
k
Pm
1 − x2
−Pnk
1 − x2
= [m(m + 1) − n(n + 1)] Pnk Pm
.
dx
dx
dx
dx
If we integrate both sides of this equation from −1 to 1 is zero, we find that the integral
of the left side is zero, and we obtain
1
k
Pnk Pm
dx = 0 if n = m.
(9.47)
−1
Substituting x = cos θ in (9.47) we get
π
k
Pnk (cos θ)Pm
(cos θ) sin θ dθ = 0
for n = m.
(9.48)
0
Next, multiplying (9.45) and (9.46) by sin θPjk (cos θ) and integrating from 0 to π,
we find that
π
k
2
1 π π
f (θ, φ) cos kφPjk (cos θ) sin θ dphi dθ = ajk
Pj (cos θ) sin θ dθ,
π 0 −π
0
π
k
2
1 π π
f (θ, φ) sin kφPjk (cos θ) sin θ dphi dθ = bjk
Pj (cos θ) sin θ dθ.
π 0 −π
0
Since (see Appendix C)
π
k
2
Pj (cos θ) sin θ dθ =
0
1
−1
k 2
Pj (x) dx =
2 (j + k)!
,
2j + 1 (j − k)!
k ≤ j,
we find that
(2j + 1)(j − k)! π π
ajk =
f (θ, φ) cos kφPjk (cos θ) sin θ dφ dθ,
2π(j + k)!
0
−π
(2j + 1)(j − k)! π π
bjk =
f (θ, φ) sin kφPjk (cos θ) sin θ dφ dθ,
2π(j + k)!
0
−π
and the coefficients for k = 0 is
2j + 1 π π
aj0 =
f (θ, φ) cos kφPjk (cos θ) sin θ dφ dθ.
4π
0
−π
(9.49)
(9.50)
(9.51)
266
9. SPHERICAL HARMONICS
Having determined the coefficients we will use the trigonometric identity
cos mα cos mβ + sin mα sin mβ = cos m(α − β),
and the above values of the coefficients ajk and bjk , we finally get
u(r, θ, φ) =
+
∞
r
n 2n
+1
4π
n=0
∞ n
π
−π
0
0
π
−π
f (α, β)Pn (cos β) sin β dα dβ Pn (cos θ)
π
0
π
−π
f (α, β)Pnm (cos β) ×
× sin β cos m(φ − α) dα dβ Pnm (cos θ)
∞
π
π
2n + 1 (n − m)!
2π (n + m)!
rn
n=1 m=1
=
rn
n=0
2n + 1
f (α, β)Pn (cos β) sin β Pn (cos θ) ×
4π
m
× sin β cos m(φ − α) Pn (cos θ) dα dβ,
(9.52)
where we have interchanged the order of summation and integration.
9.3.1. Poisson Integral Representation. We start with Eq (9.52) and calculate the summation under the integral sign for θ = 0 = φ. Then we have
u(r, 0, 0) =
1
4π
π
π
∞
−π
0
r n (2n + 1)f (α, β)Pn (cos α) sin α Pn (1)
n=0
∞
n
1 n (2n + 1)(n − m)!
+
r
f (α, β)Pnm (cos α) ×
2π n=1 m=1
(n + m)!
× sin α cos mβ Pnm (1) dβ dα.
Since Pnm (1) = 0 for all n ≥ 1, the second part of the above sum vanishes. Also,
since Pn (1) = 1, we get
u(r, 0, 0) =
Since
∞
1
4π
π
π
∞
−π
n=0
0
2n rn Pn (x) =
(2n + 1) rn f (α, β)Pn (cos α) sin α dβ dα. (9.53)
2rx − 2r2
(1 − 2rx + r 2 )
adding these results we obtain
,
3/2
n=0
∞
n=0
(2n + 1) r n Pn (x) =
∞
r n Pn (x) =
n=0
1 − r2
(1 − 2rx + r 2 )
3/2
1 − 2rx + r 2
3/2
(1 − 2rx + r 2 )
.
,
(9.54)
9.3. SURFACE SPHERICAL HARMONICS
267
Substituting (9.54) into (9.53) we get
1 − r 2 π π f (α, β) sin α dβ dα
u(r, 0, 0) =
.
3/2
4π
0
−π (1 − 2r cos α + r 2 )
(9.55)
This result holds for any point (r, θ, φ) on the sphere. Since the expression ρ =
1/2
1 − 2r cos α + r 2
is equal to the distance between the reference point (1, 0, 0)
and the integration point (r, α, β), we choose the point (r, 0, 0) to represent the north
pole of the sphere. Hence, we have
1 − r 2 π π f (α, β) sin α dβ dα
u(r, θ, φ) =
,
(9.56)
4π
ρ3
0
−π
where ρ is the distance between the points (1, 0, 0) and (r, θ, φ). Let γ be the angle at the origin between the lines L1 and L2 (see Fig. 9.1), where the direction
cosines of the line L1 are (sin θ cos φ, sin θ sin φ, cos θ) and those of the line L2 are
(sin α cos β, sin α sin β, cos α). Then cos γ = cos θ cos α + sin θ sin α cos(φ − β).
z
(1, θ, φ)
•
L2
1
L1
α
θ
• ( r , θ, φ)
r
γ
y
φ
x
β
Fig. 9.1.
Thus, the Poisson integral representation of u(r, θ, φ) for the interior of the unit sphere
is given by
1 − r2 π π
f (α, β) sin α dβ dα
u(r, θ, φ) =
,
3/2
4π
0
−π [1 − 2 (cos θ cos α + sin θ sin α cos(φ − β)) r + r 2 ]
(9.57)
and for the exterior of the unit sphere by
r2 − 1 π π
f (α, β) sin α dβ dα
u(r, θ, φ) =
,
3/2
4π
0
−π [1 − 2 (cos θ cos α + sin θ sin α cos(φ − β)) r + r 2 ]
(9.58)
268
9. SPHERICAL HARMONICS
9.3.2. Representation of a Function f (θ, φ). We will consider the problem
of representing a function of θ and φ for all points on the unit sphere, i.e., −1 ≤ θ ≤ 1
and 0 ≤ φ ≤ 2π, in terms of surface spherical harmonics. Assume that
∞ n
m
m
f (t, φ) =
A0,n Pn (t) +
(Am,n cos mφPn (t) + Bm,n sin mφPn (t)) .
n=0
Since
2π
0
m=1
2π
sin2 mφ dφ = π and 0 dφ = 2π, we get
0
1
2n + 1 2π
A0,n =
dφ
f (t, φ)Pn (t) dt,
4π
0
−1
1
2n + 1 (n − m)! 2π
Am,n =
dφ
f (t, φ) cos mφPnm (t) dt,
4π (n + m)! 0
−1
1
2n + 1 (n − m)! 2π
Bm,n =
dφ
f (t, φ) sin mφPnm (t) dt,
4π (n + m)! 0
−1
cos2 mφ dφ =
2π
(9.59)
Which gives the expansion (9.59). This development holds for all values of t and φ
for all points on the unit sphere, provided only that the given function satisfies the
conditions that would have been satisfied if it were to be developed into a Fourier
series.
Example 9.2. To express sin2 θ cos2 θ sin φ cos φ in terms of surface spherical
harmonics, note that F (t, φ) = 12 t2 (1 − t2 ) sin 2φ, and
2π
2n + 1 1 2
2
A0,n =
t (1 − t )Pn (t) dt
sin 2φ dφ = 0,
8π
−1
0
2π
2n + 1 (n − m)! 1 2
Am,n =
t (1 − t2 )Pnm (t) dt
sin 2φ cos mφ dφ = 0,
4π (n + m)! −1
0
2π
2n + 1 (n − m)! 1 2
Bm,n =
t (1 − t2 )Pnm (t) dt
sin 2φ sin mφ dφ = 0, m = 2.
4π (n + m)! −1
0
For m = 2 we have
B2,n
2n + 1 (n − 2)!
=
4 (n + 2)!
1
−1
t2 (1 − t2 )
1 2n + 1 (n − 2)!
= n
2 n! 4 (n + 2)!
=
720 2n + 1 (n − 2)!
2n n! 4 (n + 2)!
d2 Pn (t)
dt
dt2
1
−1
1
−1
t2 (1 − t2 )
dn+2 (t2 − 1)n
dt
dtn+2
dn−4 (t2 − 1)n
dt,
dtn−4
where, by repeated integration by parts,
1 n−4 2
0 if n > 4,
d
(t − 1)n
dt = 1 2
256
n−4
dt
(t − 1)4 dt =
−1
−1
315
if n = 4,
9.3. SURFACE SPHERICAL HARMONICS
269
1
9 2! 4096
1
Thus, B2,n = 4 ·
=
. In the same manner, we have B2,3 = 0 and
2 4! 4 6! 7
105
1
B2,2 =
. Hence,
42
sin2 θ cos2 θ sin φ cos φ =
1
1
sin 2φ P22 (t) +
sin 2φ P42 (t),
42
105
which is the required answer. However, this can be verified by using the following
steps:
1
1
sin 2φ P22 (t) +
sin 2φ P42 (t)
42
105
1
d2 P2 (t)
1
d2 P4 (t)
=
sin 2φ sin2 θ
+
sin 2φ sin2 θ
2
42
dt
105
dt2
1
1
=
sin2 θ sin 2φ +
sin2 θ sin 2φ(7t2 − 1)
14
14
1
1
= t2 sin2 θ sin 2φ = cos2 θ sin2 θ sin 2φ.
2
2
9.3.3. Addition Theorem for Spherical Harmonics. The addition theorem
between two different directions in the spherical coordinate system is given in §9.2.3.
Another useful result is as follows.
Theorem 9.5. (Addition Theorem for spherical harmonics) The following
relations hold:
Pn (cos γ) =
n
4π
(−1)m Ynm (θ1 , φ1 )Yn−m (θ2 , φ2 ),
2n + 1 m=−n
(9.60)
n
4π
(−1)m Ynm (θ1 , φ1 )Ynm∗ (θ2 , φ2 );
2n + 1 m=−n
(9.61)
or equivalently,
Pn (cos γ) =
or in terms of the associated Legendre functions of the first kind:
Pn (cos γ) = Pn (cos θ1 )Pn (cos θ2 )
n
(n − m)! m
+2
Pn (cos θ1 )Pnm (cos θ2 ) cos m(φ1 − φ2 ).
(n
+
m)!
m=1
(9.62)
Proof. We will derive (9.61), and other result follow. Let a function f (θ, φ)
be expanded in a Laplace’s series
 m
 Yn (θ1 , φ1 ) relative to x1 , y1 , z1 ,
n
f (θ1 , φ1 ) =
(9.63)

anm Ynm (γ, ψ) relative to x2 , y2 , z2 ,
m=−n
270
9. SPHERICAL HARMONICS
where ψ is the azimuth angle, with any choice of the 0 of this angle (see Fig. 9.2).
At γ = 0 we have
2n + 1
f (θ1 , φ1 )
= an0
,
(9.64)
4π
γ=0
since Pn (1) = 1, and Pnm (1) = 0 for m = 0. Multiplying (9.63) by Yn0∗ (γ, ψ) and
integrating over the surface S of the sphere, we get
f (θ1 , φ1 )Yn0∗ (γ, ψ) dSγ,ψ = an0 ,
S
which in view of (9.64) can be written as
Ynm (θ1 , φ1 )Yn0∗ (γ, ψ) dS = an0 .
(9.65)
S
z1
θ2
θ1
γ
P
z2
x2
y1
φ1
x1
φ2
y2
Fig. 9.2.
Let us assume that the polynomial Pn (cos θ) has an expansion of the form
Pn (cos θ) =
n
bnm Ynm (θ1 , φ1 ),
(9.66)
m=−n
where bnm depend on θ2 , φ2 , i.e., on the orientation of the z2 -axis. Multiplying the
above integral
by Ynm∗ (θ1 , φ1 ) and integrating with respect to θ1 and φ1 over S, we
Pn (cos γ)Ynm∗ (θ1 , φ1 ) dSθ1 ,φ1 = bnm , which in terms of the spherical
obtain
S
harmonics becomes
4π
2n + 1
Yn0 (γ, ψ)Ynm∗ (θ1 , φ1 ) dS = bnm ,
S
(9.67)
9.3. SURFACE SPHERICAL HARMONICS
271
where the subscripts have been dropped from the solid angle element dS. Since the
range of integration is over all solid angles, the choice of the polar axis is relevant.
Then comparing (9.85) and (9.67) we find that
b∗nm = an0



4π
f (θ1 , φ1 )γ=0
4π
2n + 1
=
2n + 1 
 4π Y m (θ2 , φ2 )
2n + 1 n
by Eq (9.64),
by Eq (9.63).
Here the change of indices occurs because θ1 → θ2 and φ1 → φ2 for γ → 0.
Substituting into Eq (9.66) we obtain (9.61).
Example 9.3. (An application of Theorem 9.5.) To determine Green’s function
for the 3-D Laplace’s equation in spherical coordinates, let us assume that the source
is on the polar axis at the point r = 0, θ = 0, φ = 0. Then using the expansion of the
generating function for Legendre polynomials (see §C.1.5), i.e., (1−2xt +t2 )−1/2 =
∞
Pn (x) tn , |t| < 1, we obtain
n=0
 ∞
an


Pn (cos γ) n+1 ,

r
1
1
n=0
≡
=
∞
rn
R
|r − ka| 


Pn (cos γ) n+1 ,
a
n=0
r > a,
(9.68)
r < a,
where k denotes the unit vector in the z-direction. Now, if we rotate the coordinate
system such that the source is at (a, θ2 , φ2 ) and the observation point at (r, θ1 , φ1 ),
then, in view of (8.10) the required Green’s function is given by
1
G(r, θ1 , φ1 ; a, θ2 , φ2 ) =
4π |r − ka|
 ∞
n
1
an

Ynm∗ (θ1 , φ1 )Ynm (θ2 , φ2 ) n+1 ,


r
n=0 m=−n 2n + 1
=
∞
n
1
rn



Ynm∗ (θ1 , φ1 )Ynm (θ2 , φ2 ) n+1 ,
a
n=0 m=−n 2n + 1
r > a,
r < a.
(9.69)
9.3.4. Discrete Energy Spectrum. We have studied the Schrödinger wave
equation for the hydrogen atom in §7.8. The geometry of the hydrogen atom is
presented in Fig. 9.3. We will revisit that section and recall that we assumed a
solution of Eq (7.36) in the form u = R(r)Θ(θ)Φ(φ). Now we will find the solution
of Eq (7.36) in a closed form, which will provide us with some information about the
discrete energy spectrum and other quantum states of the hydrogen atom.
Let Φ(φ) = A cos mφ + B sin mφ, where m = 1, 2, . . . , and let Θ(θ) satisfy
Eq (4.88), or equivalently Eq (7.39) with the substitutions t = cos θ. Eq (7.41)
has the solutions in terms of the associate Legendre functions of the first kind, i.e.,
Θ(θ) = Pnm (θ), where m ≤ n, n being an integer (see Appendix C).
272
9. SPHERICAL HARMONICS
µ2
. For
1 − t2
Legendre’s equation and its properties, see §C.1. The solutions Eq (7.41) are associated Legendre polynomials Θ = Pnµ (cos θ), where Pnµ (t), and µ = 0, ±1, ±2, . . . .
For associated Legendre’s equation and its properties, see §C.2. The solution of Eq
(7.40) are Φ = A cos µφ + B sin µφ, where φ is a periodic function of period 2π,
i.e., Φ(φ) = Φ(2π + φ). Eq (7.38) has been solved in §7.8, and the solution for the
problem (7.37) is given by (7.44), namely,
h2 N
2π 2 m e2
2 µ
uN,nµ (r, θ, φ) = 2
r
P
(cos
θ)
exp
−
r
×
n
4π m e2
h2 N
2
4π m e2
2n+1
× LN +n
r (A cos µφ + B sin µφ) , (9.70)
h2 N
Eq (7.41) would be Legendre’s equation if it were not for the term
where the constants A and B can be determined from the boundary conditions (to
be provided), and N is called the principal quantum number, n the azimuthal
quantum number, and µ the magnetic quantum number. In fact, if we look at the
formal solution u(r, θ, φ) = R(r)Θ(θ)Φ(φ), which represents the wave function, the
term R(r) is related to the principal quantum number, the term Θ(θ) to the azimuthal
(orbital) quantum number, and the term Φ(φ) to the magnetic quantum number. The
spin quantum number s is a property of the electron since it has an intrinsic spin.
θ
mp +
r
−m
e
φ
Fig. 9.3. Hydrogen Atom.
Thus, Bernoulli’s separation method leads to three equations for the three spatial
variables, and their solutions produce the above three quantum numbers associated
with the hydrogen energy levels. The wave functions for certain spins of the hydrogen
2
atom are listed below, where a0 =
= 0.529 Å denotes the first Bohr radius which
me2
is the nuclear charge, and the subscript 1s, 2s, 3s indicate the first, second, and third
state of its radial density:
9.3. SURFACE SPHERICAL HARMONICS
273
1 −3/2
1
−3/2
u1s = √ a0 r e−r/a0 , u2s = √ a0
2 − a−1
e−r/(2a0 ) ,
0
π
4 2π
1
−3/2
2 −2
u3s = √ a0
27 − 18ra−1
e−r/(3a0 ) .
0 + 2r a0
81 3π
0.6
Most Probable Distance
Probability
0.5
0.4
Average Distance
0.3
0.2
0.1
0
1.5 a 0
a0
Distance in units of a 0
Fig. 9.4.
The probability of finding the electrons in terms of the three dimensions, i.e., its
radial density, is proportional to the square of the respective wave function. Fig. 9.4
shows a plot of the radial density which indicates that the distance of the highest
probability (most probable distance) is a0 .
To prove this statement, we find that squaring the 1s radial wave we obtain the
probability function P (r) as
P (r) =
1 −3 2 −2r/a0
a r e
.
π 0
(9.71)
Differentiating with respect to r and setting P (r) = 0 gives
P (r) = 0 =
2
−2r/a0
r(r − a0 )a−2
,
0 e
π
which yields r = 0, or r = a0 . Since r = 0 is the absolute minimum (see Fig. 9.4),
−2
we find that P (a0 ) = −8πa−3
, which is negative. Hence, P (r) has the relative
0 e
maximum at the distance r = a0 .
For the 2s wave function, we use the value of u2s given above. Then the probability
of finding the point where the electron has a node follows from setting the probability
P (r) = 0, and it occurs is at r = 2a0 , which can be seen from the 2 − r/a0 part of
the wave function. Similarly, for the 3s function, the only part of the wave function
274
9. SPHERICAL HARMONICS
u3s , and hence the square of this function, that can make P (r) = 0 is the term
2 −2
27 − 18ra−1
a
, or the polynomial 27 − 18x + 2x2 = 0 where x = r/a0 ,
0 + 2r
0
18 ± 182 − 4(2)(27)
i.e., at x =
≈ 7.1 or 1.9. Thus, the 3s wave has 2 nodes,
4
at r = 1.9 and r = 7.1. In general the probability has N − 1 nodes where N is the
principal quantum number.
Example 9.4. To calculate the average probability distance for the hydrogen
atom, we start with the normalized wave function u1s and let r denote the spatial
position coordinate of this distance. Then this distance, denoted by |r| is
|r| =
=
∞
0
a−3
0
π
a−3
= 0
π
=
π
2π
1
q
√ a−3/2 e−r/a0 m √ a−3/2
e−r/a0 r 2 sin θ dr dθ dφ
π
π 0
0
0
∞ π 2π
r 3 e−2r/a0 sin θ dr dθ dφ
0
0
∞
0
3 −2r/a0
r e
2π
sin θ dr dθ dφ
0
3
2
dr
0
−4
a−3
0 3! a0
4π =
π
16
π
0
a0 ,
∞
where we have used the formula
0
xn e−bx dx =
n!
.
bn+1
9.3.5. Further Developments. Schrödinger used the model of a standing wave
to represent the electron within the hydrogen atom and solved the resulting wave
equation. The reasoning behind using the spherical coordinate system is based on
the fact that the hydrogen atom has a spherically symmetric potential. The potential
energy is due to a point charge.
The hydrogen atom, being an atom of the chemical element H, is an electronically neutral atom which contains a single positively charged proton and a single
negatively-charged electron bound to the nucleus by the Coulomb force. In fact, the
1/r Coulomb factor leads to the Laguerre functions of the first kind. The hydrogen
atom is very significant in the study of quantum mechanics and quantum field theory
of a simple two-body problem. In 1914 Niels Bohr performed a number of simplifying
assumptions and obtained the spectral frequencies of the hydrogen atom (see [Griffiths 1995]). The Bohr model explained only the spectral properties of the hydrogen
atom, and his results were confirmed by an analytical solutions of the Schrödinger
wave equation (7.36) by Bernoulli’s separation method. These solutions use the fact
that the Coulomb potential produced by the nucleus is isotropic, i.e., it is radially
symmetric in space and depends only on the distance to the nucleus. They were
useful in calculating hydrogen energy levels and frequencies of the hydrogen spectral
lines. In this sense, the solutions of the Schrödinger equation are far more useful
than the Bohr model, because these solutions also contain the shape of the electron’s
wave function (orbital part) for various quantum mechanical states, which eventually
9.3. SURFACE SPHERICAL HARMONICS
275
explains the anisotropic character of atomic bonds. The hydrogen spectrum showing
the Bohr model is presented in Fig. 9.5.
Balmer Series
(Visible)
Lyman Series
(Ultrviolet)
410.2 nm
violet
.
434.1 nm
violet
n =1
n =2
656.3 nm
red
n =3
486.1 nm
bluegreen
Paschen Series
(Infrared)
n =4
n =5
Fig. 9.5. Hydrogen Spectrum.
Pauli [1926] solved the problem of the hydrogen atom by using a rotational
symmetry in four dimensions (the so called O(4)-symmetry) generated by the angular
momentum and the Laplace-Runge-Lenz vector. The non-relativistic problem of the
hydrogen atom was later solved by Duru and Kleinert [1979] using Feynman’s path
integral formulation of quantum mechanics. The energy eigenstates of the hydrogen
atom are summarized as follows:
−13.6 e V
α2
n
3
Energy Levels: Ejn =
1+ 2
−
, where α is the
n2
n
4
j + 12
fine-structure constant, and j denotes the total angular momentum eigenvalue, i.e.,
l = ± 12 depending on the direction of the electron spin; the value −13.6eV is called
the Rydberg constant that is determined in the Bohr model and it related to mass
me q 4
m and the charge q of the electron by the relation −13.6eV = − 2 e2 , or to the
8h <0
me c2 α2
0.51 M eV
fine-structure constant by the relation −13.6eV = −
=−
. The
2
2(137)2
normalized wave function is given in the spherical coordinates by
unlm (r, θ, φ)
3
l
2
(n − l − 1)! r/(na0 )
2r
2r
2l+1
=
e
Ln−l−1
Ylm (θ, φ), (9.72)
na0
2n(n + l)!
na0
na0
276
9. SPHERICAL HARMONICS
where a0 is the Bohr radius, L2l+1
n−l−1 are the associated Laguerre functions of the first
kind and degree n−l−1, and Ylm are the spherical harmonic functions of degree l and
order m, such that the quantum numbers take the following values: n = 1, 2, . . . ,
l = n − 1, n − 2, . . . , 1, 0 , and m = −l, . . . , l. The ground state is the state of
the lowest energy in which the electron is first found; it corresponds to the principal
quantum level n = 1, l = 0. The nodes of the wave function (9.72) are the spherical
harmonics Ylm (θ, φ). The quantum number determine the location of these nodes.
There are n − 1 total nodes, of which l are angular nodes, and n − l − 1 remaining
nodes are radial nodes. Out of the l angular nodes, m angular nodes go around the φ
axis in the plane z = 0, and the remaining l − m occur on the vertical θ axis.
The solutions of the Schrödinger equation fail to predict certain small but measurable deviations of the real spectrum of the hydrogen atom. For example, the
mean speed of the hydrogen electron is only 1/137th of the speed of light (c), and
since the electron’s wavelength is determined by its momentum, it is found that the
orbitals with higher speed electrons exhibit contraction due to smaller wavelength.
This requires a complete theoretical relativistic treatment of the problem which is not
possible with the Schrödinger equation. Another significant feature is the internal
magnetic field which exists due to moving electron and creates an electromagnetic
field of the nucleus even when there is no external magnetic field. The spin of the
electron creates an associated magnetic moment which interacts with this magnetic
field. This effect, in the language of special relativity, is known as the spin-orbit
coupling. These shortcomings in the Schrödinger equation treatment are overcome
by the Dirac’s relativistic equation which has predicted results that come closer to the
experimental data. Since this subject is beyond the scope of this book, the interested
reader is directed to related literature such as the books by Griffiths [1995], and Hecht
[2000].
9.4. Exercises
9.1. Show that u(x, t) = A cos
2π
(x − at − ϕ) is a solution of the equation
λ
2
∂2u
2∂ u
=
a
, where A is a constant, λ the wavelength and ϕ the phase of the
∂t2
∂x2
wave represented by the solution.
9.2. Prove Theorem 9.1. If the time required to go through a complete vibration,
called the period, is T , then set a = λ/T = λω, where ω = 1/λ is called the
angular frequency of the wave. Then show that the
x solution of the problem in
Exercise 9.1 can be written as u(x, t) = A cos 2π
− ωt − φ .
λ
9.3. Prove that
π
1
2 (n + m)!
2
2
m
[Pn (cos θ)] sin θ dθ =
[Pnm (x)] dx =
, m ≤ n,
2n
+ 1 (n − m)!
0
−1
where Pnm (x) are the associated Legendre polynomials.
277
9.4. EXERCISES
9.4. Write down all explicit representations of the associated Legendre polynomials
Pnm (x) for m ≤ n ≤ 3.
√
√
Ans. P11 (x) = 1 − x2 , P21 (x) = 3 1 − x2 ,
√
P31 (x) = 12 1 − x2 15x2 − 3 , P22 (x) = 3 1 − x2 ,
P32 (x) = 15x 1 − x2 , P33 (x) = 15 1 − x2
3/2
.
9.5. Prove the orthogonality relation (9.42) for associated Legendre polynomials,
1
(n + m)!
i.e., show that
Pnm (x)Pnk (x)(1 − x2 )−1 dx =
δm,k .
m(n
− m)!
−1
9.6. Prove that Pn−m (x) = (−1)m
(n − m)! m
P (x), where Pnm (x) is defined by
(n + m)! n
dn+m 2
(x − 1)n .
dxn+m
Hint. Apply Leibniz’s formula to (x + 1)n (x − 1)n .


 0, n even,
(−1)(n−1)/2 n!
9.7. Show that Pn (0) =

2 , n odd.

[(n − 1)/2]! 2(n−1)/2
Pnm (x) =
1
2n n!
(1 − x2 )m/2
Hint. Use either recurrence relation or expansion of the generating function.
2n + 1
9.8. Prove that Ynm (0, φ) =
δm,0 .
4π
Hint. Use (9.1).
9.9. Introduce a new function w(x) in Eq (9.75) by setting u = x(1−m))/2 ex/2 w,
and
(a) Show that this results in the following equation:
x2 w + rxw −
x2 + [2(m − 1) − 4n] x + m2 − 1
w = 0;
4
(b) Show that the equation in (a) has a regular singular point at x = 0; and
∞
(c) Find the series solution of this equation of the form w = xλ
ck xk , and
determine its radius of convergence.
k=0
[2(µ − 1) − 4m] ck−1 + ck−2
, µ = 2λ−1,
4k(k + µ)
= 0. The series converges everywhere on R.
Ans. (c) Recursion formula is ck =
c0 arbitrary, c−1
9.10. Prove Unsöld’s theorem (9.29).
9.11. The Schrödinger
equation for the harmonic oscillator in its lowest energy state
d2 u
2mE
mω 2 2
is 2 +
−
t u = 0, where u denotes the potential, and ω is
dt
2
278
9. SPHERICAL HARMONICS
the angular frequency of vibrations. Show that the solution of this equation is
2
u(t) = ce−mωt /(2) , where c is a constant.
Hint. The equation is nonlinear. Assume the solution of the form u = c e−bt ,
and substituting it in the equation and equating the coefficients of t2 and the
constant terms to zero, determine b.
2
9.12. Prove the relation (9.19).
Hint. Use Exercise 9.6.
9.13. Prove that Pn (cos θ) = (−1)
nr
n+1
∂n
n! ∂z n
1
.
r
Hint. Compare the Legendre polynomial expansion of the generating function
with a Taylor series expansion.
9.14. Show that Pn (cos θ) ≥ −1, and this result is sharp for Pn (−1), n odd.
Hint. Use Pn (−x) = (−1)n Pn (x).
1 dn
2 −1/2 (1
−
2t
cos
θ
+
t
)
to show that Pn (1) = 1
n! dtn
t=0
n
and Pn (−1) = (−1) .
d
n(n + 1)
9.16. Prove that Pn (1) =
Pn (x)
=
,
dx
2
x=1
1 m
1
9.17. Show that (i) −1 x Pn (x) dx = 0 for m < n, and (ii) −1 xn Pn (x) dx =
2n+1 (n!)2
.
(2n + 1)!
9.15. Use Pn (cos θ) =
Hint. Use Rodrigue’s formula.
9.18. The associated Lengendre function Pnm (x) satisfies the self-adjoint equation
d2 Pnm (x)
dPnm (x)
m2
(1 − x2 )
−
2x
+
n(n
+
1)
−
= 0.
dx2
dx
1 − x2
Use the equations for Pnm (x) and Pnk (x) to show that
1
−1
9.19. Show that sin θ
Pnm (x)Pnk (x)
dx
= 0,
1 − x2
for m = k.
d
Pn (cos θ) = Pn1 (cos θ).
dθ
9.20. Show that
π m
dPn dPnm
m2 Pnm Pnm
2n(n + 1) (n + m)!
+
sin θ dθ =
δnn ,
2
dθ
dθ
2n + 1 (n − m)!
sin θ
0
π 1
Pn dPn1
P 1 dPn1
+ n
sin θ dθ = 0.
sin θ dθ
sin θ dθ
0
279
9.4. EXERCISES
These integrals occur in problems of scattering of electromagnetic waves by
spheres.
2n + 1
m
9.21. Show that Yn (0, φ) =
δm0 .
4π
2n + 1
9.22. Prove that Pn+1 (x)Qn−1 (x) − Pn−1 (x)Qn+1 (x) =
x.
n(n + 1)
r − z r n+1 ∂ n 1
9.23. Prove that Qn (cos θ) = (−1)n
ln
.
n
n! ∂z 2r
r+z
9.24. Show that the integral over the surface of the unit sphere of the product of two
tesseral harmonics of the same degree but of different orders is zero.
2π
2π
Hint.
sin nφ cos mφ dφ =
sin nφ sin mφ dφ
0
=
0
2π
cos nφ cos mφ dφ = 0.
0
1
9.25. Show that
−1
Pnm (t)Pkm (t) dt =

0
if n = k,
2 (k + m)!

2k + 1 (k − m)!
if n = k.
9.26. Express cos2 θ sin2 θ sin φ cos2 φ in terms of surface spherical harmonics.
Hint. Use the method of Example 9.2.
1
1
2
Ans. 6930
P63 (t) + 1540
P43 (t) sin 3φ − 693
P61 (t) −
1
1
P
(t)
sin φ.
63 2
1
1
770 P4 (t)−
5
9 · 2! 2
13 · 4! 2
9.27. Show that cos 2φ = 2 cos 2φ P22 (t) +
P4 (t) +
P6 (t) + · · · .
4!
6!
8!
Hint. Follow the method of Example 9.2.
9.28. Show that if Sn (x, y, z) is a solid spherical harmonic of degree n, then
∇2 [r m Sn (x, y, z)] = m(2n + m + 1)r m−2 Sn (x, y, z).
2 2 2
2 ∂Sn
nSn
∂r
∂r
∂r
2
2
Hint. ∇ Sn = 0, ∇ r = ,
=
, and
+
+
=
r ∂r
r
∂x
∂y
∂z
1.

(n + m)!

 (−1)(n−m)/2 n − n n + n , n + m even,
2n
!
!
9.29. Show that Pnm (0) =

2
2

0, n + m odd.
9.30. Show that Pnn (cos θ) = (2n − 1)!! sin θ, n = 0, 1, 2, . . . .
280
9. SPHERICAL HARMONICS
∞
n
1
δ(θ1 − θ2 )δ(φ1 − φ2 )
sin θ1
n=0 m=−n
= δ(cos θ1 − cos θ2 )δ(φ1 − φ2 ). Hint. This is known as the spherical harmonic
closure relation.
1
φ(s)
9.32. Solve the integral equation f (x) =
ds, −1 ≤ x ≤ 1,
2 1/2
−1 (1 − 2xs + x )
for the unknown function φ(x) if (a) f (x) = x2k ; and (b) f (x) = x2k+1 .
9.31. Show that
Ans. (a) φ(x) =
Ynm (θ1 , φ1 )Ynm∗ (θ2, φ2 ) =
4k + 1
4k + 3
P2k (x); (b) φ(x) =
P2k+1 (x).
2
2
10
Conformal Mapping Method
Green’s functions are useful in solving Dirichlet problems of potential theory in itself
and, in the case of conformal mapping, of a region onto a circular disk. In the latter
case a relationship is needed between conformal mapping and Green’s function for
the circle. Although no unique expression is available for Green’s function for a
circle, yet an integral representation of Green’s function for the disk leads to the
Poisson integral representation. Besides certain Green’s functions determined for
Laplace’s equation in a circle, semi-circle, and a sector which have been discussed in
Chapter 8, there are other results for Green’s function for a circle, which are presented
in this chapter. Green’s functions for the ellipse, certain half-plane regions and the
parallel strip are determined, and an interpolation method for computation of Green’s
functions for convex regions is presented.
10.1. Definitions and Theorems
Let C denote the complex plane. If a ∈ C and r > 0, then we will denote by
B(r, a) = {z ∈ C : |z − a| < r}, B̄(r, a) = {z ∈ C : |z − a| ≤ r}, and
∂B(r, a) = {z ∈ C : |z − a| = r} an open disk, a closed disk, and a circle,
respectively, each of radius r and centered at a point a. Thus, B(1, 0) represents the
open unit disk, sometimes denoted by U . A connected open set D ⊆ C is called a
region (or domain ), and ∂D or Γ denotes its boundary.
10.1.1. Cauchy-Riemann Equations. Let z = x + i y be a complex number.
z + z̄
z − z̄
∂f
1
Then z̄ = x − i y, x =
, and y =
. Also ∂f =
= (fx − i fy ),
2
2i
∂z
2
∂f
1
¯
∂f =
= (fx + i fy ). The Cauchy-Riemann equations for the function f (z) =
∂ z̄
2
u(x, y) + i v(x, y) are: ux = vy , uy = −vx , or in polar form (z = rei θ ), ur =
1
1
vθ , vr = − uθ . and they satisfy the partial differential equations: ux vx +
r
r
281
282
10. CONFORMAL MAPPING METHOD
uy vy = 0, and ∇2 u = 0, ∇2 v = 0.
The function log z = ln |z| + i arg{z} is a multiple-valued function, unless the
principal values of arg{z} are taken, which range from −π to π. If z = rei θ , then
log z = ln r + i θ is a single-valued function for r > 0 and −π < θ < π.
¯ = 0, which is equivalent to
A function f : D → C is analytic on D iff ∂f
the Cauchy-Riemann equations for the function f (z) = u(x, y) + iv(x, y). Thus,
f (z) = ux + i vx = vy − i uy . The Cauchy-Riemann equations are the necessary
conditions for f (z) to be analytic on D. However, merely satisfying the CauchyRiemann equations alone is not sufficient to ensure the differentiability of f (z) at a
point in D. In view of the Cauchy-Riemann equations, property (ii) of §8.4 holds,
that is,
(x,y)
v(x, y)−v(x0 , y0 ) =
(−uy dx + ux dy),
(10.1)
(x0 ,y0 )
where (x0 , y0 ) = z0 is a given point in D. This property is also true if D is multiply
connected, although in that case the conjugate function
v(x, y) can be multiple-valued,
as we see by considering u(x, y) = log r = log x2 + y 2 defined on a region D
containing the origin which has been indented by a small circle centered at the origin.
Then, in view of (10.1),
v(x, y) − v(x0 , y0 ) = tan−1
y
± 2nπ + const,
x
n = 1, 2, . . . ,
which is multiple-valued.
A simple closed curve, or the Jordan contour, Γ in C is a path γ : [a, b] → C
such that γ(t) = γ(s) iff t = s or |t − s| = b − a. The Jordan curve theorem states
that if Γ is a simple contour, then C\Γ has two components, one called the interior
of Γ, denoted by Int (Γ), and the other called the exterior of Γ, denoted by Ext (Γ),
each of which has Γ as its boundary. Thus, if Γ is a Jordan contour, then Int (Γ) and
Ext (Γ) ∪ {∞} are simply connected regions.
Let L2 denotes the Hilbert space of all square-integrable analytic functions f in a
simply connected region D with boundary Γ. A function f (z) regular
in D is said to
belong to the class L2 (D), denoted by f ∈ L2 (D), if the integral D |f (z)|2 dSz <
+∞, where dSz = dx dy denotes an area element in D. If two functions f, g ∈
L2 (D), then their inner product is defined by
f, g =
f (z) g(z) dSz .
(10.2)
D
10.1.2. Conformal Mapping. A mapping f of a region D onto a region A
is called analytic iff it is differentiable. The mapping f is called conformal if it is
bijective and analytic. The conformal mapping theorem states that if a mapping
f : D → A is analytic and f (z0 ) = 0 for each z ∈ D, then f is conformal. The
mapping f is conformal if it is analytic with a nonzero derivative. Two important
properties are the following:
10.1. DEFINITIONS AND THEOREMS
283
(i) If f : D → A is conformal and bijective (i.e., one-to-one and onto), then
f −1 : A → D is also conformal and bijective.
(ii) If f : D → A and g : A → B are conformal, then the composition f ◦ g :
D → B is conformal and bijective.
Property (i) is useful in solving boundary value problems (e.g., the Dirichlet
problem) for a region D. The method involves finding a map f : D → A such that
A is a simply connected region on which the problem can be first solved, and then
the result for the original problem is provided by f −1 . Since the Dirichlet problem
involves harmonic functions, the following result on the composition of a harmonic
function with a conformal map is useful: If u is harmonic on a region A and if
f : D → A is conformal, then u ◦ f is harmonic on D.
Theorem 10.1. (Riemann mapping theorem) Let D ⊂ C be a simply connected region. Then there exists a bijective conformal map f : D → U , where
U is the open unit disk. Moreover, the map f is unique provided that f (z0 ) = 0
and f (z0 ) > 0 for z0 ∈ D.
This theorem implies that if D and A, both contained in C, are any two simply
connected regions, then there exists a bijective conformal map g : D → A. If
f : D → U and h : A → U, then g = h−1 ◦ f is bijective conformal. Thus, the
regions D and A are said to be conformal if there exists a bijective conformal map
between them. A bijective conformal map is also called a univalent map. A function
w = f (z) defining a univalent mapping is called a univalent function. Its inverse
image is also a univalent function defined on the image region.
10.1.3. Symmetric Points. The points z and z ∗ are said to be symmetric with
respect to a circle C through three distinct points z1 , z2 , z3 iff
(z ∗ , z1 , z2 , z3 ) = (z, z1 , z2 , z3 ).
(10.3)
z*
a.
z
R
Fig. 10.1. Symmetry with Respect to a Circle.
The mapping that carries z into z ∗ is called a reflection with respect to C. The points
z and z ∗ are also called the inverse points or symmetric points with respect to
284
10. CONFORMAL MAPPING METHOD
the circle C. If C is a straight line, then we choose z3 = ∞, and the condition for
z ∗ − z1
z̄ − z̄1
symmetry (10.3) gives
=
. Let z2 be any finite point on the line C.
z1 − z2
z̄1 − z̄2
∗
Then, since |z − z1 | = |z − z1 |, the points z and z ∗ are equidistant from the line
z∗ − z z−z 1
1
C. Moreover, since = −
, the line C is the perpendicular
z1 − z2
z1 − z2
∗
bisector of the line segment joining z and z . If C is the circle |z − a| = R, then,
recalling that the conjugates of sums and products are sums and products of the
conjugates, we find that
(z, z1 , z2 , z3 ) = (z − a, z1 − a, z2 − a, z3 − a)
R2
R2
R2
,
,
z1 − a z2 − a z3 − a
=
z̄ − ā,
=
R2
, z1 − a, z2 − a, z3 − a
z̄ − ā
=
R2
+ a, z1 , z2 , z3 .
z̄ − ā
R2
+ a are symmetric with respect
z̄ − ā
∗
2
∗
to the circle C, i.e., (z − a)(z̄ − ā) = R . Also, |z − a||z − a| = R2 , and since
z∗ − a
> 0, the points z and z ∗ are on the same ray from the center a (Fig 10.1).
z−a
Note that the point symmetric to a is z = ∞.
Hence, in view of (10.3), the points z and z ∗ =
10.1.4. Cauchy’s Integral Formula. Let f be analytic on a region D, and let
Γ be a simple closed contour in D that is homotopic to a point in D. Let z0 ∈ D be
a point not on Γ. Then
1
f (z)
f (z0 ) · I(Γ, z0 ) =
dz,
(10.4)
2i π Γ z − z0
where I(Γ, z0 ) denotes the index (or the winding number) of the curve Γ with respect
to a point z0 ∈ C; i.e., it is the integer n that expresses n-times Γ winds around z0 ,
and is defined by
1
dz
I(Γ, z0 ) =
,
2i π Γ z − z0
±n if z0 ∈ Int (Γ)
, where the plus or minus sign is
0 if z0 ∈ Ext (Γ)
chosen according as the path Γ is traversed in the counterclockwise or clockwise
direction. For n = 1, Cauchy’s integral formula becomes
1
f (z)
f (z0 ) =
dz,
(10.5)
2i π Γ z − z0
or, simply by I(Γ, z0 ) =
10.1.5. Mean-Value Theorem. The mean-value theorem for harmonic functions is discussed in §8.4. Thus, if we integrate in (10.1) on the circle |z − z0 | = r0
10.2. DIRICHLET PROBLEM
285
and if the function u(z) is harmonic on |z − z0 | ≤ r0 , then this integral must vanish,
which in polar coordinates is expressed as
2π
1
∂u
r0 dθ = 0.
(10.6)
2π 0 ∂r
dr
and integrating from r1 and r2 , we get
r0
2π
1
u(r1 , θ)dθ =
u(r2 , θ)dθ.
(10.7)
2π 0
After multiplying the integral (10.6) by
1
2π
2π
0
These are the mean values of u(z) taken on both circles |z − z0 | = r1 and |z| = r2 ,
where r1 < r2 . So long as u(z) is harmonic on the larger circle, these mean values
are equal. For r1 → 0 the left side of (10.7) takes the value u(z0 ) at the center, so
that we finally obtain
2π
1
u(z0 ) =
u(r, θ)dθ.
(10.8)
2π 0
This is the mean-value theorem of potential theory which states that for every function
that is harmonic on a circle the value at the center is equal to the mean value of the
function on the circumference (see §8.4.(iv)). An important consequence of the
mean-value theorem is that a nonconstant function u(z) harmonic in a region D takes
neither a maximum nor a minimum value in the interior of D. Since the real and
imaginary part of a regular analytic function w = f (z) = u(z) + iv(z) are harmonic
functions, the mean-value theorem also holds for analytic functions, and Cauchy’s
integral formula remains valid, i.e.,
f (z0 ) =
1
2iπ
|z|=r
f (z)
dz.
z − z0
(10.9)
10.2. Dirichlet Problem
Let z ∈ D be a fixed point (known as the source point). Green’s function for the
Dirichlet problem in the region D with a logarithmic singularity at z is the function
G(z, z ) with the following properties:
(i) As a function of z, G(z, z ) is harmonic everywhere in D except at the point z .
(ii) At the point z the function G(z, z ) is defined by1
G(z, z ) =
1
log
2π
1
r
+ g(z, z ),
r = |z − z |,
(10.10)
where g(z, z ) is harmonic everywhere in D.
1 If the point z is a sink of strength −1 , then the definition (10.10) becomes G(z, z ) =
1
log r + g(z, z ), r = |z − z |; see (8.1).
2π
286
10. CONFORMAL MAPPING METHOD
(iii) G(z, z ) = 0 if the point z lies on the boundary ∂D ≡ Γ.
A direct consequence of conformal mapping is that the Dirichlet conditions are
directly transferable from the physical model in the z-plane to the mathematical model
in the w-plane. The Dirichlet problem for the region D can be solved in an explicit
form by using Green’s function. Since, in view of the property (ii), the function G
becomes unbounded at z = z , we can indent the point z by a circle Γε of small
radius ε. Then the functions u and G become continuous in the region Dε bounded
by Γ and Γε . An application of Green’s second identity (1.16) (with g = G) yields
∂G
∂u
u
−G
ds = 0,
(10.11)
∂n
∂n
Γ+Γε
where n denotes the outward normal to the boundary which is normal to Γ, exterior
to Dε , and interior to Γε . Separating the integral (10.11) over two contours Γ and Γε
and using (10.10), we find that
1
∂ log
∂G
∂u
1
∂g
r
u
−G
ds =
u
ds +
u
ds
∂n
∂n
2π Γε
∂n
∂n
Γ
Γε
1
1 ∂u
∂u
−
log
ds −
g
ds ≡ I1 + I1 + I2 + I3 + I4 . (10.12)
2π Γε
r ∂n
∂n
Γε
The interior normal to Γε is along the radius r with its direction opposite to that of
∂
∂
increasing r, and hence,
= − , which yields
∂n
∂r
∂ log
∂n
1
r
∂ log
=−
∂r
1
r
=
1
.
r
Also, since r = ε on the circle Γε , we get, in view of (10.8),
1
I1 =
u ds = u(z )
2πε Γε
for any ε by the mean value theorem. In I3 , since r = ε and u is harmonic, we have
1
1
∂u
I3 =
log
ds = 0 for any ε.
2π
ε
∂n
Γε
∂u
The remaining two integrals I2 and I4 tend to zero as ε → 0. In fact, since u, g,
∂n
∂g
and
are bounded in the neighborhood of the point z , we find that
∂n
|I2 | ≤ max u
Γε
∂g
· 2πε,
∂n
and
|I4 | ≤ max g
Γε
∂u
· 2πε.
∂n
10.2. DIRICHLET PROBLEM
287
Hence, as ε → 0, the relation (10.12) yields
1
2π
u
Γ
∂G
∂u
−G
∂n
∂n
ds = u(z ).
(10.13)
Moreover, by condition (iii) the function G vanishes on Γ, and thus, from (10.13) we
obtain
1
∂G
u(z ) =
u
ds.
(10.14)
2π Γ ∂n
Note that this equation is also a consequence of Green’s third identity (1.19). If
Green’s function is known for the region D, then formula (10.14) can be used to
solve the Dirichlet problem for any continuous or piecewise continuous, boundary
values of the harmonic function u(z). An alternate form of formula (10.14) is
1
u(z ) =
u dG(z, z ).
(10.15)
2π Γ
The relationship between Green’s function and conformal mapping is established
as follows: Let w = f (z) = r ei θ map a simply connected region D conformally onto
the open unit disk U : |w| < 1 such that the point z ∈ D goes into the point w = 0,
f (z)
i.e., f (z ) = 0. If f (z) has a simple zero at z , the function F (z) =
= 0 for
z − z
all z ∈ D and is regular everywhere in D. Thus, log F (z) is also regular analytic in
D. Let us denote log F (z) = p + i q. Then
f (z) = (z − z ) ep+i q ,
which yields, with z − z = r ei φ ,
1
1
1
log
=
2π
|f (z)|
2π
log
1
−p .
r
(10.16)
1
1
log
satisfies all three properties for
2π
|f (z)|
Green’s function G(z, z ), namely, this function with a simple pole at z is harmonic
in D except at z , where it has a logarithmic singularity (in fact, it is a source of
1
1
1
strength +1), and it is equal to
log + g(z), where g(z) = −
p. Moreover,
2π
r
2π
1
1
|f (z)| = 1 on the boundary Γ, and, therefore,
log
= 0 there. Thus, we
2π
|f (z)|
have shown that Green’s function G(z, z ) and the mapping function f (z) which
produces the conformal map of the region D onto the unit disk U are related to each
other by (see Courant and Hilbert [1968:378])
It can easily be verified that the function
G(z, z ) =
1
1
1
log
=−
log |f (z)|,
2π
|f (z)|
2π
(10.17a)
288
10. CONFORMAL MAPPING METHOD
such that the point z ∈ D goes into the origin. If the singularity at z = z is a sink
of strength −1, then this formula becomes
G(z, z ) =
1
log |f (z)|.
2π
(10.17b)
Since we have used a source of strength −1 in Chapter 8, we will employ formula
(10.17b) whenever a relation between Green’s function G(z, z ) and the conformal
mapping function f (z) is required. If Green’s function for a region D is known, we
can use (10.17a or 10.17b) and construct the function f (z) which maps the region
D conformally onto the unit disk. The method of accomplishing this using (10.17b),
for example, is as follows: For each term in (10.16), we determine the respective
1
1
φ
conjugate harmonic function. The conjugate harmonic function for
log is
,
2π
r
2π
where φ = arg{z − z }. Let h(z) be conjugate to the function g(z). Then, in view
of (10.1),
z
∂p
∂p
h(z) =
dy −
dx + C,
(10.18)
∂x
∂y
z
where C is an arbitrary constant, which corresponds to the rotation of the unit disk
about w = 0. Hence, the required mapping function is given by
w = f (z) = r ei φ e−2π(g+i h) = (z − z ) e−2π(g+i h) .
(10.19)
Note that the construction of Green’s function G(z, z ) involves determining the harmonic function g(z), whose boundary values are determined from the third property,
namely, that G(z, z ) = 0 on the boundary Γ. This means that g(z) must take the
1
1
values
log on Γ. Hence, the conformal mapping problem of transforming the
2π
r
region D onto the unit disk reduces to the solution of the Dirichlet problem with the
boundary condition
1
g(x, y) Γ =
log r.
(10.20)
2π
A similar result is obtained, with a minus sign, if we use (10.17a) instead. If D can
be mapped conformally onto the unit disk, then the dependence of Green’s function
on z0 can be given explicitly. Thus, if w = f (z) is any function that maps D onto
the unit disk, then we can use the mapping
z → ei γ
w − w0
,
1 − w w0
where γ is an arbitrary real constant, which maps the unit disk onto itself such that
the point w(z0 ) goes into the origin and w0 = f (z0 ). Hence,
G(z, z ) = log
w − w0
.
1 − w w0
(10.21)
10.2. DIRICHLET PROBLEM
289
10.2.1. Dirichlet Problem for a Circle in the (x, y)-Plane. Let the circle
C be given by |z| = R, where z = x + i y. The problem is to find a solution for the
boundary value problem
∇2 u ≡
∂2u ∂2u
+ 2 = 0,
∂x2
∂y
u = f (θ) on C,
where θ is the angular coordinate on C, i.e., z = R ei θ on C. The constructive proof
for the existence of the solution consists in deriving an expression for the solution.
Let F (ζ) be an analytic function in the region enclosed by C such that {F (ζ)} on
|ζ| = R is equal to f (θ). Let z be a complex number in this region (at A in Fig 10.2).
The point symmetric to z with respect to the circle C is z ∗ , which lies outside C (at
B in Fig. 10.2), such that z ∗ z̄ = R2 . According to Cauchy’s integral formula (10.5)
1
F (z0 ) =
2π i
C
F (ζ)
dζ,
ζ −z
1
and 0 =
2π i
C
F (ζ)
dζ.
ζ − R2 /z̄
Subtracting we get
1
F (z) =
2π i
C
F (ζ) R2 − z z̄
dζ.
ζ (R2 + z z̄) − zζ 2 − zR2
B
•z *
o
C
A
•z
o
0•
|z| = R
Fig. 10.2. Inverse (Symmetric) Points with Respect to the Circle |z| = R.
Since ζ lies on C and z0 inside C, set ζ = R ei θ , z = r ei φ , r < ρ. Then
1
F r ei θ =
2π
2π
F Rei θ
0
R2
R2 − r 2
dθ.
− 2rR cos(θ − φ) + r 2
Equating the real parts on both sides, we get
u(x, y) =
R2 − r 2
2π
0
π
f (θ) dθ
,
R2 − 2rR cos(θ − φ) + r2
(10.22)
290
10. CONFORMAL MAPPING METHOD
where r 2 = x2 +y 2 and φ = arctan(y/x). This formula is called Poisson’s integral
representation in the (x, y)-plane, which is the same as (8.43). This completes the
proof of the existence of the solution for the Dirichlet problem.
A frequently encountered Green’s function for a circle is Poisson’s integral formula for the interior of a circle C of radius R. Let z = rei θ and z = ρei φ be two
distinct interior points of C. Then
1
R2 + r 2 ρ2 /R2 − 2ρ cos(θ − φ)
G(z, z ) =
log
,
4π
r 2 + ρ2 − 2rρ cos(θ − φ)
(10.23)
∂G
1
R2 − r 2
=−
,
∂ρ ρ=R
2πR R2 + r 2 − 2Rr cos(θ − φ)
where G applies to the Dirichlet condition and
[1961:530]).
∂G
to the Neumann condition (Korn
∂ρ
10.3. Neumann Problem
We will consider the case when D is a circle with center at the origin and radius R.
z
In this case we carry out the conformal mapping onto the unit disk by w = , and
R
Green’s function (10.21) becomes
G(z, z ) = log
R(z − z )
.
R2 − z z¯
1
z¯
2
+
dz. Since |z| = z z̄ = R2 on the boundary
z − z R2 − z z¯
z
z̄
of the circle and dz = i z dθ, we have dG(z, z ) = i
+
−
1
dθ.
z − z
z̄ − z¯
iθ
iφ
Setting z = R e and z = ρ e , we find that
Then dG(z, z ) =
dG(z, z ) = i
R2 − ρ2
dθ.
R2 + ρ2 − 2Rρ cos (θ − φ)
(10.24)
If we substitute (10.24) in (10.9), we obtain the Poisson integral
u(ρ ei φ ) =
1
2π
2π
u(R ei φ )
0
R2 − ρ2
dθ.
R2 + ρ2 − 2Rρ cos (θ − φ)
(10.25)
By a similar integral representation we can determine the harmonic function v(z)
which is conjugate to u(z). In view of (10.1)
v(z) − v(0) =
0
z
∂u
ds.
∂n
10.3. NEUMANN PROBLEM
291
When we apply this operation on (10.25) and follow through the corresponding integrations and differentiations, we get
v(ρei φ ) − v(0)
2π
ρei φ
1
∂
R2 + ρ2
iφ
=
u(R e )
ds dθ,
2π 0
∂u R2 + ρ2 − 2Rρ cos (θ − φ)
0
where the inner integral is taken on an arbitrary path that lies entirely in the interior
of the circle. Note that
R2 − ρ2
z
z̄
=
+
−1
R2 + ρ2 − 2Rρ cos (θ − φ)
z − z
z̄ − z¯
2z
=
−1 =
z − z
Thus,
z + z
z − z
=
z + z
z − z
−2Rρ sin (θ − φ)
,
R2 + ρ2 − 2Rρ cos (θ − φ)
.
(10.26)
and hence
v(ρei φ ) = v(0) −
1
2π
2π
u(R ei θ )
0
R2
2R sin(θ − φ)
dθ.
+ ρ2 − 2Rρ cos (θ − φ)
If we combine (10.25) and (10.27), we obtain Schwarz’s formula:
2π
1
R ei θ + ρ ei φ
iφ
f (ρe ) = i v(0) +
u(R ei φ )
dθ,
2π 0
R ei θ − ρei φ
(10.27)
(10.28)
which allows us to determine the value of a complex potential function f (z) =
u(z) + i v(z) in a circle with prescribed boundary values u(z) and v(0).
Note that the above formulas allows us to determine the value of a complex
potential function f (z) = u(z) + i v(z) in a circle with prescribed boundary values
u(z) and v(0), and find an explicit solution for the Neumann problem, for a disk, and
thereby for all regions with known conformal mapping onto the unit disk. In this case
∂u
the value of the normal derivative
on the boundary of the region is prescribed,
∂n
and we seek a function harmonic on D whose normal derivative takes this boundary
value. For this problem to have a solution we require that
Γ
∂u
ds = 0.
∂n
(10.29)
Since, in view of (10.1),
v(z2 ) = v(z1 ) −
z2
z1
∂u
ds,
∂n
292
10. CONFORMAL MAPPING METHOD
the value of v(z) is determined on the boundary up to an additive constant. With this
boundary value v(z) the value of u(z), which is a harmonic function conjugate to
v(z), can be determined in the interior of D by the formula (10.27). Thus,
2π
1
2Rρ sin(θ − φ)
u(ρei φ ) = u(0) +
v(R ei θ ) 2
dθ
2π 0
R + ρ2 − 2Rρ cos (θ − φ)
2π
1 = u(0) +
v(R ei θ ) log R2 + ρ2 − 2Rρ cos (θ − φ) 0 (10.30)
2π
2π
1
∂v(R ei θ )
−
log R2 + ρ2 − 2Rρ cos (θ − φ) dθ.
2π 0
∂θ
Since v(R ei θ ) and log R2 + ρ2 − 2Rρ cos (θ − φ) are periodic, the second term
on the right side of (10.30) vanishes. Also,
∂v
∂u
∂u
dθ = dv =
ds =
R dθ,
∂θ
∂n
∂n
so that we finally obtain the following integral formula:
2π
1
∂u
iθ
u(R e ) = u(0)−
R log R2 + ρ2 − 2Rρ cos (θ − φ) dθ, (10.31)
2π 0 ∂n
which establishes a relationship between the boundary values of
∂u
and the values
∂n
of u(z). We shall denote the expression
log R2 + ρ2 − 2Rρ cos (θ − φ) = log (z − z )(z̄ − z¯ ) ≡ N (z, z ) (10.32)
and call N (z, z ) the Neumann’s function. It plays the same role for the Neumann boundary value problem as Green’s function does for the Dirichlet problem.
This function represents a regular analytic function of z in D except for logarithmic
R2
singularities. Since z̄ =
, we have
z
(z − z )(R2 − z z¯ )
N (z, z ) = log (z − z )(z̄ − z¯ ) = log
.
z
(10.33)
This function can be regarded as a complex potential function for a flow, which has a
source at z and a sink at z and for which the circle |z| = R acts as an impermeable
boundary.
As in the case of Green’s function, Neumann’s function N (z, z ) can be characterized by the following conditions:
(i) N (z, z ) is a regular analytic function of z on a region D except for a logarithmic
singularity at z = z and at another fixed point z = z ∗ , i.e.,
N (z, z ) = log
z − z
+ n(z, z ),
z − z∗
(10.34)
10.4. GREEN’S AND NEUMANN’S FUNCTIONS
293
where n(z, z ) is a regular analytic function on D.
(ii) N (z, z ) as a function of z has a boundary value which is continuous everywhere on the boundary Γ of D, and {N (z, z )} = 0 on Γ. If these conditions are
satisfied, then
1
∂u
u(z ) = u(z ∗ ) −
N (z, z )
ds.
(10.35)
2π Γ
∂u
If the region D is mapped conformally onto the unit disk by the function w = f (z),
then N (z, z ) can also be defined by
N (z, z ) = log
[f (z) − f (z )]f (z )
,
f (z)
(10.36)
where z ∗ under this map goes into f (z ∗ ) = 0.
Note that the Neumann problem is solvable only if the condition (10.29) is satisfied. For a multiply connected region the contour Γ in (10.29) must include the
exterior and all interior paths. The reason why this condition does not hold for each
individual path is that the function u may be multiple-valued, which does not let the
integral of ∂u/∂n around Γ vanish.
10.4. Green’s and Neumann’s Functions
Analogous to Green’s functions, the solution of the Neumann problem of potential
theory is Neumann’s function which also possesses an integral representation, and
bears a close relationship with Green’s functions. We will also determine Green’s
and Neumann’s functions for an ellipse, and annulus, and a parallel strip.
The Laplacian ∇2 is invariant to translation as well as rotation of the coordinate
system. This result is proved in §10.3.1. Green’s function for the circle has been
extensively studied and only a few explicit expressions are known. One such expression, due to Lanczos [1961], is given below. Since, in view of the Riemann mapping
theorem, any simply connected region D with a piecewise smooth boundary can be
conformally mapped onto the unit disk, the pertinent Green’s function is also a conformal mapping according to the mapping equation (Courant and Hilbert, [1968:377]),
but only a few examples are known to be described by explicit formulas. In this
section we use some ideas developed in Borre [2001] with certain modifications.
10.4.1. Laplacian. Let a potential function φ(z) in the z-plane be mapped conformally onto a potential function ψ(w) in the w-plane such that ψ at (u(x, y), v(x, y))
equals φ at (x, y). The potential φ is harmonic, i.e., it satisfies Laplace’s equation
∇2 φ = 0. The question is whether ψ is also harmonic. We will prove the following
result:
Theorem 10.2. Laplace’s equation ∇2 φ = 0 remains invariant under
conformal transformation.
294
10. CONFORMAL MAPPING METHOD
Proof. (Based on Henrici [1974: Ch. 5]) If we set φ(x, y) = ψ(u, v), then
∂φ
∂ψ ∂u ∂ψ ∂v
=
+
.
∂x
∂u ∂x
∂v ∂x
(10.37)
In the z- and w-planes the complex gradients are
∇c φ =
∂φ
∂φ
+i
,
∂x
∂y
∇c ψ =
∂ψ
∂ψ
+i
.
∂u
∂v
(10.38)
Substituting (10.37) into (10.38) and using the Cauchy-Riemann equations (to replace
∂u
∂v
∂v
∂u
by −
and
by
), we get
∂y
∂x
∂y
∂x
∇c φ =
=
=
∂ψ
∂u
∂ψ
∂u
∂ψ
∂u
∂u ∂ψ ∂v
∂ψ ∂u ∂ψ ∂v
+
+i
+
∂x
∂v ∂x
∂u ∂y
∂v ∂y
∂u ∂ψ ∂v
∂ψ ∂v
∂ψ ∂u
+
+i −
+
∂x
∂v ∂x
∂u ∂x
∂v ∂x
∂ψ
∂u
∂v
+i
−i
= ∇c ψ · f (z).
∂v
∂x
∂x
(10.39)
To obtain the second partial derives, we start with (10.37) and get
∂ 2φ
∂2ψ
=
∂x2
∂u2
∂u
∂x
2
+2
∂ 2 ψ ∂u ∂v
∂2ψ
+
∂u∂v ∂x ∂x
∂v 2
A similar expression is obtained for
∂v
∂x
2
+
∂ψ 2
∂ψ 2
∇ u+
∇ v.
∂u
∂v
(10.40)
∂2ψ
. Then adding them we obtain
∂y 2
2
2
2
∂2φ ∂2φ
∂ 2ψ
∂u
∂u
∂ 2ψ
∂v
∂v
+
=
+
+
+
2
2
2
2
∂x
∂y
∂u
∂x
∂y
∂v
∂x
∂y
∂ 2 ψ ∂u ∂v
∂u ∂v
∂ψ
∂ψ
+2
+
+ ∇2 u
+ ∇2 v
,
∂u∂v ∂x ∂x ∂y ∂y
∂u
∂v
2
+
(10.41)
On the right side of (10.41) the first and second terms in the square brackets are
equal to |f (z)|2 ; the third term in the braces is zero because of the Cauchy-Riemann
equations, and ∇2 u = 0 = ∇2 v. Thus, (10.41) reduces to
∇2 φ(x, y) = ∇2 ψ(u, v) |f (z)|2 ,
or ∇2 ψ(u, v) = ∇2 φ(x, y) |f (z)|−2 .
(10.42)
Hence, ∇2 ψ = 0 when ∇2 φ = 0 so long as |f (z)| = 0, a condition which is
satisfied by analytic functions except at their singular points. Note that we can
1
substitute f (z) = in (10.42).
f (w)
10.4. GREEN’S AND NEUMANN’S FUNCTIONS
295
10.4.2. Green’s Function for a Circle. The Poisson integral representation
(10.25) is one form of Green’s function for a circle. Other representations are given
below by some examples.
Example 10.1. Consider the following problem with the Neumann boundary
condition:
∇2 h(z) = f (z) in D,
(10.43a)
∂h(z)
=0
∂n
(10.43b)
on ∂D.
Using the Gauss theorem (§1.7) we get
∇h dS =
D
∂D
∂h
ds = 0,
∂n
which after comparing with (10.43a) yields the compatibility condition
f (z) dS = 0.
(10.44)
D
Since the operator ∇2 is singular, Green’s function cannot be uniquely determined,
but any Green’s function yields
h(z ) =
G(z, z ) f (z) dS,
(10.45)
D
such that
∇2 G(z, z ) = δ(z, z ) − φ(z),
(10.46a)
∂G(z, z )
= 0,
∂nz
(10.46b)
where φ(z) is a function that must satisfy the condition
[δ(z, z ) − φ(z)] dS = 0, or
φ(z) dS = 1.
D
(10.47)
D
1
For the open unit disk U (see Fig. 10.3) we choose φ(z) =const = . Then the
π
explicit solution of Eqs (10.46a) and (10.47) is given by (see Lanczos [1961: (8.4.66)])
1 G(z, z ) =
log |z − z |2 + log |z − z |2 + |z|2 − 1 |z |2 − 1 −
4π
− |z|2 − |z |2 + 32
1 2
=
ln r − 2rr0 cos θ + r02 + ln 1 − 2rr0 cos θ + r 2 r02 −
4π
− r 2 − r02 + 32 .
(10.48)
296
10. CONFORMAL MAPPING METHOD
y
U
r
0
•z
x
θ
rο
1
•ζ
Fig. 10.3. Open Unit Disk U .
We will now consider the case of Green’s function in R2 for a domain D which
can be mapped conformally onto the unit circle. The following result holds.
Theorem 10.3. Let w = f (z) = u + i v, where z = x + i y, represent the
conformal mapping of a domain D in the z-plane onto the unit circle in the
w-plane, where f (z) is a simple analytic function of the complex variable z.
Then Green’s function for the domain D is given by
1
f (x + i y ) − f (x + i y)
G(x, y; x , y ) =
log
.
(10.49)
2π
f (x + i y ) f (x + i y) − 1
Proof. To show that G = 0 on ∂D, note that the boundary ∂D is mapped by
f onto the boundary of the unit circle in the z-plane. Thus, f (x + i y) = ei θ , where
the point (x, y) lies on ∂D. Set z = x + i y and z = x + i y . Then by (10.17b)
1
f (z ) − ei θ
On ∂D: G(x, y; x , y ) =
log
2π
f (z ) ei θ − 1
1
f (z ) − ei θ
=
log
2π
(f (z ) − ei θ ) ei θ
1
log(1) = 0.
2π
Next, we will show that the above function G(x, y; x , y) is a fundamental solution
of Laplace’s equation, i.e., we will show that it satisfies ∇2 G = 0 everywhere except
at (x, y) = (x , y ), and that in the neighborhood of this point we have
=
2πG(x, y; x , y ) = log (x, y) − (x , y ) + a harmonic function.
(10.50)
Since f is one-to-one, we have f (z) − f (z ) = 0 only at z = z . For any two pints
z = (x, y) and z = (x , y ) in D, z = z , we have |f (z)| < 1 and |f (z )| < 1. Thus,
f (z) f (z) − 1 = 0. The function
f (z ) − f (z)
log
(10.51)
f (z )f (z) − 1
10.4. GREEN’S AND NEUMANN’S FUNCTIONS
297
is analytical in z except for z = z , and the real part of the function (10.51), being a
harmonic function, satisfies Laplace’s equation. In the neighborhood of the point z we have the Taylor’s series
f (z) − f (z ) =
∞
g (n) (z )
(z − z )n
n!
n=1
= (z − z )
∞
g (n) (z )
(z − z )n−1 = (z − z ) g(z),
n!
n=1
where g(z) is analytic. Since the mapping is conformal, the function g(z) = 0 in a
neighborhood of z . Thus,
1
f (z) − f (z ) G(z, z ) =
log |z −z |+ log 1 − f (z)f (z ) −log
2π
z − z
1
=
log (x − x )2 + (y − y )2 + γ,
4π
where γ is the harmonic function defined by (8.41). Hence, the required Green’s
function is defined by (10.49).
10.4.3. Green’s Function for an Ellipse. As we have seen, Green’s function
G(z, z ) ≡ G(x, y; x , y ) and Neumann’s function N (z, z ) ≡ N (x, y; x , y ) are
both harmonic functions in any region D, and possess the following properties:
1. Both G(z, z ) and N (z, z ) are harmonic for fixed z ∈ D and z ∈ D̄ for all
z = z .
2. In the neighborhood of z =
z the functions G(z, z ) + ln r and N (z, z ) + ln r
are harmonic, where r = (x − x )2 + (y − y )2 .
3. At a fixed boundary point z ∈ ∂D, we have G(z, z ) = 0,
any z ∈ D.
∂N (z, z )
= const for
∂n
Note that the function N (z, z ) is normalized on the boundary ∂D by the condition
N (z, z ) ds = 0,
(10.52)
∂D
and any harmonic function h(z) defined on D̄ can be represented in the form
1
∂G(z, z )
h(z ) =
h(z) dsz ,
(10.53)
2π ∂D ∂nz
1
∂h(z)
or h(z) =
N (z, z )
dsz ,
(10.54)
2π ∂D
∂n
where dsz is an element of the arclength along
the boundary ∂D, and (10.54) holds
only if the function h(z) is normalized by
h(z) ds = 0.
∂D
298
10. CONFORMAL MAPPING METHOD
Let D denote the interior of an ellipse with eccentricity e. Consider the conformal
mapping z = cosh w, where z = x + i y and w = u + i v. Then
x = cosh u cos v,
(10.55)
y = sinh u sin v.
The lines u = const are mapped onto a confocal family of ellipses:
x2
y2
+
= 1,
cosh2 u sinh2 u
(10.56)
with foci at z = √
±1, semi-major axis a = cosh u, semi-minor axis b = sinh u, and
a2 − b2
1
eccentricity e =
=
. The lines v = const are mapped onto a family
a
cosh u
of hyperbolas:
x2
y2
−
= 1,
cos2 v
sin2 v
with foci at z = ±1. The element ds of the arclength is given by
2
ds2 = dx2 + dy 2 = (sinh u cos v du − cosh u sin v dv) +
+ (cosh u sin v du + sinh u cos v dv)
2
2
= cosh u − cos v du2 + dv 2 .
2
(10.57)
Let us consider a particular ellipse defined by u = u0 and a particular hyperbola by
v = v0 . Then, by (10.57), it is obvious that the ellipse with u = u0 and the hyperbola
with v = v0 are orthogonal to each other. Let dn and dt denote the arclength elements
at the node (u0 , v) along the normal and the tangent to the ellipse, respectively. Then
2
2
dn = cosh u0 − cos v du, dt = cosh2 u0 − cos2 v dv,
∂
1
∂
= ,
2
∂n
cosh u0 − cos2 v ∂u
∂2
∂2
1
∇2 =
+
=
2
∂x2
∂y 2
cosh u − cos2 v
∂2
∂2
+
∂u2
∂v 2
.
Thus, Green’s function for this particular ellipse, defined with u = u0 and w1,2 =
u1,2 + i v1,2 , is given by
G (w1 ; w2 ) = u0 − ln 2 − ln |z1 − z2 |−
∞
2 −nu0 cosh nu1 cosh nu2 cos nv1 cos nv2
−
e
+
n
cosh nu0
n=1
sinh nu1 sinh nu2 sin nv1 sin nv2
+
sinh nu0
∞
2 −nu0
≡ u0 − ln 2 − ln |z2 − z1 | −
e
[an + bn ] .
n
n=1
(10.58)
10.4. GREEN’S AND NEUMANN’S FUNCTIONS
299
We will now verify if this Green’s function satisfies the three properties listed at
the beginning of this section. The properties 1 and 2 are satisfied if we apply the
Laplacian, defined above, on both terms under the summation in (10.58) and show
that this sum is uniformly convergent. For the first term in the sum we get the estimate
∞
an ≤
n=1
∞
n=1
|an | ≤
∞
∞
2 −nu0 cosh nu1 cosh nu2
e
≡
αn .
n
cosh nu0
n=1
n=1
αn+1
< (1 + ε) e−(2u0 −u1 −u2 ) for all
αn
n ≥ n0 , i.e., for every u1 < u0 the series
an is uniformly convergent
for any
u2 ≤ u0 . A similar argument can be applied to the second term of series
bn . To
ex + e−x
prove the property 3, we use the following formulas: cosh x =
; ln(1 −
2
∞
n
x
x) = −
, −1 ≤ 1 ≤ 1; cosh(ix) = cos x, and sinh(ix) = i sin x. Since
n
n=1
1
z1 − z2 = cosh w1 − cosh w2 = − ew2 (1 − e−w1 −w2 ) (1 − ew1 −w2 ) , we have for
2
u2 > u 1 :
∞ 1 ln (cosh w2 − cosh w1 ) = − ln 2 + w2 −
en(−w1 −w2 ) + en(w1 −w2 ) .
n=1 n
For any ε > 0 we can find an n0 such that
Thus, for u2 > u1 ,
ln |z1 − z2 | = u2 − ln 2 −
∞
2 −nu2 e
cosh nu1 cos nv1 cos nv2 +
n
n=1
+ sinh nu1 sin nv1 sin nv2 .
In particular, for u2 = u0 , we get G (u1 , v1 ; u2 , v2 ) = 0, which proves that the
property 3 is satisfied.
Neumann’s function for the ellipse is given by
N (w1 ; w2 ) = u0 − ln 2 − ln |z1 − z2 |+
∞
2 −nu0 cosh nu1 cosh nu2 cos nv1 cos nv2
+
e
+
n
sinh nu0
n=1
sinh nu1 sinh nu2 sin nv1 sin nv2
+
.
cosh nu0
(10.59)
Using the same arguments as used above in the case of Green’s function we can
establish that this Neumann’s function satisfies the properties 1 through 3 (see Exercise
6.3).
10.4.4. Green’s Function for an Infinite Strip. If we keep the semi-minor
axis b of the ellipse in §6.3.3 fixed and take the limiting process as the semi-major axis
300
10. CONFORMAL MAPPING METHOD
a = cosh u0 → ∞, i.e., as u0 → ∞, then we obtain an infinite strip −∞ < x < ∞,
−b < y < b. This limiting process can be visualized in Fig. 10.4. However, we
will show that Green’s function, and Neumann’s function also, for this infinite strip
cannot be derived from (10.58) and (10.59), respectively, under this limiting process.
b
0
−b
Fig. 10.4. Limiting Process of Confocal Ellipses to an Infinite Strip.
To carry out this limiting process successfully, we will first modify the conformal
mapping z = cosh w by introducing a real scaling factor µ > 0, which need not
be equal to unity, such that x + i y = µ cosh(u + i v). Then a = µ cosh u0 and
b = µ sinh u0 . To maintain the geometry of the infinite strip, we keep b fixed; then
e−nu0
the limiting process will require that only u0 → ∞. Since lim
= 0 and
u0 →∞ cosh nu0
e−nu0
e−nu0
lim
= lim
= 0, Green’s function for the infinite strip is
u0 →∞ sinh nu0
u0 →∞
b
u0 − ln µ − ln 2 − ln |z1 − z2 |−
G (w1 ; w2 ) = lim
u0 →∞
∞
2 −nu0 cosh nu1 cosh nu2 cos nv1 cos nv2
−
e
+
n
cosh nu0
n=1
sinh nu1 sinh nu2 sin nv1 sin nv2
sinh nu0
= − ln µ − ln 2 + lim u0 − ln |z1 − z2 |,
+
u0 →∞
which does not exist because it is not finite. Similarly, Neumann’s function, given by
N (w1 ; w2 ) = lim
u0 →∞
u0 − ln µ − ln 2 − ln |z − z |−
∞
2 −nu0 cosh nu1 cosh nu2 cos nv1 cos nv2
−
e
+
n
sinh nu0
n=1
= − ln µ − ln 2 +
does not exists.
sinh nu1 sinh nu2 sin nv1 sin nv2
+
cosh nu0
lim u0 − ln |z − z |,
u+0→∞
10.4. GREEN’S AND NEUMANN’S FUNCTIONS
301
In order to determine Green’s function for the infinite strip, we will use the methods
of conformal mapping and first consider the case of Green’s function for the upper
half-plane, as in the following example.
Example 10.2. Green’s function for the upper half-plane y ≥ 0 is given by
G(x, x ) =
1
1
1
R2 log
−
log
x
−
x .
2π
|x − x | 2π
|x |2
(10.60)
Another interesting result for Green’s function for the upper half-plane {{z} ≥ 0}
is given in Wayland [1970:333] as
G(z, z ) = log
z + z̄ .
z − z
(10.61)
y
y=π
−∞ < x < ∞
w = ez
x
y= 0
0
z − plane
z = log w
v
v >0
u
v =0
0
w − plane
Fig. 10.5. Mapping of the Upper Half-Plane onto the Parallel Strip.
Example 10.3. To find Green’s function for the infinite strip 0 < y < π, note
that the function w = ez , z = x + i y, maps this strip conformally onto the upper
half-plane {w} > 0, w = u + i v (see Fig. 10.5). Thus, Green’s function for this
strip is obtained from (10.60) with r = 1 as
G(x, y; , x , y ) =
1
e2x − 2 ex+x cos(y + y ) + e2x
log 2x
.
2
e − 2 ex+x cos(y − y ) + e2x
(10.62)
If u is harmonic in this strip, then
sin y ∞
u(x, 0)
u(x, π)
u(x , y ) =
+
dx.
2π −∞ cosh(x − x ) − cos y cosh(x − x ) + cos y 302
10. CONFORMAL MAPPING METHOD
is the solution of Laplace’s equation ∇2 u = 0 in this strip.
10.4.5. Green’s Function for an Annulus. Consider an annulus D with the
outer radius r1 and the inner radius r2 , respectively. Then, using cylindrical polar
coordinates with z = r ei θ and z = ρ ei φ , Green’s function for the Dirichlet problem
for r < ρ is given by
(ln r − ln r2 ) (ln r1 − ln ρ)
G (r, θ; ρ, φ) =
+
ln r1 − ln r2
n
n
n
r2 n
r
ρ
r1
−
−
cos n(θ − φ) (10.63)
∞
r
r1
r1
ρ
r n r n =+
.
1
2
n
−
n=1
r
r
Example 10.4. (Green’s function for a sphere of radius R.) Let B(R, 0) be a
ball (sphere) of radius R and center at the origin in Rn , and let x ∈ B(R, 0) be an
arbitrary point. Then Green’s function for this sphere is given by
r
1
G(x, x ) = F (r) − F
|x | ,
R
n/2
2π
where Sn (1) =
is the surface area of B(1, 0) (unit sphere in Rn ), and
Γ(n/2)

1

2−n

, if n > 2,
 (n − 2) S (1) |x − x|
n
F (r) =
1
1



log
, if n = 2;
2π
|x − x |
r = |x − x |,
x
2
= x21 + · · · + x2n ,
r12 = x −
R2 x .
|x |2
Note that r1 is the distance of the point x from the reflected image of the point x
in the sphere. The function G(x, x ) satisfies all the requirements of being Green’s
function since (i) it is of the form s(x, x ) + φ(x), where s(x, x ) is the so-called
singularity function defined by (2.55), and φ(x) is regular in B(R, 0) and continuous
r1 in B ∪ ∂B, and (ii) it vanishes on ∂B, since r =
|x | on ∂B.
R
Example 10.5. We will determine Green’s function for the positive (right) halfplane x1 ≥ 0, n > 2, in Rn . Let x be an arbitrary point in the positive half-plane.
Then Green’s function is
G(x, x ) =
1
x − x
(n − 2)Sn (1)
2−n
− p(x),
n > 2,
where
p(x) =
(2−n)/2
n
1
2
2
(x1 + x1 ) +
(xk − xk )
.
(n − 2)Sn (1)
k=2
10.5. COMPUTATION OF GREEN’S FUNCTIONS
303
The function p(x) is obtained by taking the image of x in the boundary plane x1 = 0.
The solution of the Laplace equation ∇2 u = 0 in this half-plane is given by
∂
u(x ) =
f (x)
G(x, x )
dS.
∂x1
x1 =0
x1 =0
Example 10.6. (Green’s function for a cardioid)
Consider the conformal map√
ping z = f (w) = (w + 1)2 , with inverse w = z − 1, where −π < arg{z} < π.
Then f (w) = 2(w
0, but f (−1) = 0. The unit disk |w| < 1 goes into
√ + 1) =
2
the domain D : | z − 1| < 1, which in polar cylindrical coordinates is defined
√
θ
by r + 1 − 2 r cos θ2 < 1, or r < 4 cos2 ; thus, D : r < 2(1 + cos θ), which
2
is the interior of a cardioid with cusp at z = 0 that corresponds to w = −1, where
f (0) = 0 (Fig. 10.6). Green’s function for this cardioid D is
√
√
1 − ( z − 1)( z̄ − 1)
1
√
G(z, z ) =
log
,
√
2π
z − z
or in polar cylindrical coordinates
√
√
√
r + r + rr + 2 rr cos θ+θ
− r cos θ2 − r cos θ2
2
1
√
G(r, θ; r , θ ) =
log
.
4π
r + r − 2 rr cos θ−θ
2
(10.64)
y
2
x
0
4
−2
Fig. 10.6. Cardioid.
10.5. Computation of Green’s Functions
Numerical approximation methods are important as they have extended the applications of conformal mapping in various areas of practical significance. These methods
have a long history of development, starting with transforming complicated boundaries by successive mappings so that each approximation ultimately led to a near circle
in the final complex plane. This smoothing or osculation process was developed by
304
10. CONFORMAL MAPPING METHOD
Koebe [1915] who called the method ‘Schmiegungsverfahren’, but it suffered from
slow convergence for numerical computation, although it because a powerful technique for theoretical purposes. Many other methods were proposed during the late
half of the twentieth century, specially by Heinhold and Albrecht [1954], von Koppenfels and Stallmann [1959], Gaier [1964], and Henrici [1986]. Numerical methods
used today are based on either an approximation of the desired unknown mapping
function by power series expansions or variational techniques with orthogonalization,
or by methods of solving integral equations.
A domain D is said to be convex, if the line segment joining any two points in
the domain D is contained in D. A domain D is said to be starlike with respect to a
point z0 ∈ D if the line segment joining the point z0 to any point z ∈ D lies entirely
inside D. A convex domain D is starlike at every point in D. Also, if D is starlike
with respect a point z0 ∈ D and if z ∈ D, then t z ∈ D for 0 < t < 1.
10.5.1. Interpolation Method. We will develop a method, also known as
the method of simultaneous equations, developed by Kantorovich [1936] and Kantorovich and Krylov [1958], to numerically compute Green’s function for a simply
connected domain D ∈ R2 that has the origin as an interior point. If the origin is
outside D, we carry out a suitable translation to bring the origin inside D. We will
use the notation z = x + i y and z = x + i y instead of x = (x, y) and x = (x , y ),
respectively. The method is based on the following results from the theory of conformal mapping: If Green’s function G(z; z ) with a pole (singularity) at a point z ∈ D
and an analytic function f (z) each map the domain D conformally onto the unit disk,
then G(z; z ) and f (z) are related by (10.17b), i.e.,
1
G(z; z ) =
log |f (z)|, z = x + iy,
2π
and the mapping function f (z) is given by (10.19), that is,
f (z) = (z − z ) e−2π(g+ih) ,
where g and h are harmonic functions in D. Since log |f (z)| = log |z − z | − 2πg(z),
the construction of Green’s function G(z; z ) involves determining the harmonic
function g(z) ≡ g(x, y) such that G(z; z ) = 0 on the boundary ∂D. Thus, we solve
the Dirichlet problem for the domain D with the boundary condition g(x, y)
=
∂D
1
log r, r = |z − z |. Hence, Green’s function for the domain D is constructed by
2π
the formula2
1
G(z; z ) =
log r − g(x, y).
(10.65)
2π
The interpolation method can be developed as follows: Assuming the series
∞
k
representation for g(z) as g(z) = k=0 ck (z − z ) , ck = ak +i bk , we approximate
This formula is based on the assumption that a source of strength −1 exists at z . If
1
1
this strength is taken as +1, formula (10.65) becomes G(z, z ) =
log − g(x, y), which
2π
r
is the same as in Kantorovich and Krylov [1958], and Kythe [1998].
2
10.5. COMPUTATION OF GREEN’S FUNCTIONS
it by the harmonic polynomial pn (r, θ) = n
ck (z − z )
k
305
. Thus, in the polar
k=0
coordinates we have
g(x, y) ≈ pn (r, θ) = a0 +
n
r k (ak cos kθ − bk sin kθ) ,
k=1
where z −z = reiθ . Since this polynomial has (2n+1) coefficients, we take (2n+1)
arbitrary points z1 , . . . , z2n+1 on the boundary ∂D and then choose the coefficients
ak and bk such that at each of the points zj , j = 1, . . . , 2n+1, the polynomial pn (r, θ)
1
takes the same value as g(zj ). Since g(z) has the boundary value g(x, y) =
log r,
2π
the coefficients ak , bk are determined by solving the system of equations
a0 +
a0 +
n
k=1
n
r1k (ak cos kθ1 − bk sin kθ1 ) =
1
log r1 ,
2π
r2k (ak cos kθ2 − bk sin kθ2 ) =
1
log r2 ,
2π
k=1
···
a0 +
···
n
···
···
···
···
···
k
r2n+1
(ak cos kθ2n+1 − bk sin kθ2n+1 ) =
k=1
(10.66)
1
log r2n+1 ,
2π
where zj − z = rj ei θj , j = 1, . . . , 2n + 1. The determinant of the system (10.66)
depends on the choice of the points zj . Let us assume that these points lie on an
equipotential line of a harmonic polynomial Qn (r, θ) of degree at most n. This line
has the equation
Qn (r, θ) = a0 +
n
rk (ak cos kθ − bk sin kθ) = 0,
k=1
where the coefficients ak and bk are all not zero at the same time. Then this line
exists iff the homogeneous system of (2n + 1) algebraic equations
a0 +
n
rjk (ak cos kθj − bk sin kθj ) = 0,
j = 1, . . . , 2n + 1,
(10.67)
k=1
has a nonzero solution for a0 , a1 , b1 , . . . . Hence, Green’s function has the approximate representation
1
log r − a0 −
r k (ak cos kθ − bk sin kθ) .
2π
n
G(z; z ) ≈
k=1
(10.68)
306
10. CONFORMAL MAPPING METHOD
It is assumed here that the difference between pn (r, θ) and g(x, y) decreases as
n increases and the above interpolation method becomes justified, provided that
lim pn (r, θ) = g(z). In view of the Riemann mapping theorem, the function f (z)
n→∞
maps the domain D uniquely onto the unit disk iff f (0) = 0 and f (0) = 1. The latter
condition is used to compute the distortion in the mapping, and hence, the relative
error in the computation of G(z; z ).
Example 10.7. We will approximate Green’s function for the square {(x, y) :
−1 ≤ x, y ≤ 1} in the z-plane with the pole z , which is taken, without loss of
generality, at the origin (see Fig. 10.7). In view of the symmetry about both coordinate
axes, the values of g(x, y) are arranged symmetric to the y-axis; thus, all bk = 0 in
(10.67). Also, since the values of g(x, y) are symmetric to x-axis and bisectors
of coordinate angles, the series (10.67) will contain cosine terms of angles 4nθ,
n = 0, 1, 2, . . . . Hence, the series expansion for g(x, y) becomes
g(x, y) ≈ a0 + a4 r 4 cos 4θ + a8 r 8 cos 8θ + · · · .
2
We take the points zj on the boundary as follows: z1 = 1, z2 = √ eiπ/6 , z3 =
3
√ iπ/4
2e
. Then the coefficients a0 , a4 , and a8 are obtained by solving the system
(10.66), i.e.,



 
0 
1 1
1
 a0 

 1 − 8 − 128  a4 = 1
ln √2
,
9
81
  2π  √3 
1 −4
16
a8
ln 2
which gives a0 = 0.120286 = 0.075578/2π , a4 = −0.0117794 = −0.0740122/2π ,
and a8 = −0.000249209 = −0.00156583/2π. Note that additional points z4 =
2
√ eiπ/3 and z5 = eiπ/2 add nothing to the solution.
3
y
Y
1
0.5
x
-1
0
-0.5
0
0.5
1
1
-0.5
-1
Fig. 10.7. The Unit Square.
-1
Fig. 10.8. Cardioid for α = 1.
X
1.5
10.5. COMPUTATION OF GREEN’S FUNCTIONS
307
The approximate Green’s function is given by
1
log r − g(z)
2π
1
≈
log r − 0.075578 + 0.0740122 r 4 cos 4θ + 0.00156583 r 8 cos 8θ
2π
1
=
log z − 0.075578 + 0.0740122 z 4 + 0.00156583 z 8 .
2π
G(z; 0) =
Then, the function f (z) that maps the square onto the unit disk is found from
log |f (z)| = log z − 0.075578 + 0.0740122 z 4 + 0.00156583 z 8 ,
up to a purely imaginary additive constant which we will ignore. Hence,
f (z) = z e−0.075578+0.0740122 z
4
+0.00156583 z 8
.
Note that f (0) = e−2π a0 = e−0.075578 = 0.927207. The value is f (0) =
1
dζ
≈ 0.927037, which shows the error of about 0.017%.
1
+ ζ4
0
Example 10.8. To approximate Green’s function for the ellipse
x = 1 + λ2 cos t, y = 1 − λ2 sin t, 0 ≤ t < 2π,
or, in complex notation, z(t) = eit 1 + λ2 e−2it , 0 ≤ t < 2π, we consider the first
quadrant because of the axial symmetry and take the points zj for t = 0, π/8, π/4,
3π/8, and π/2. Thus,
g(x, y) ≈ a0 + a1 r1 cos 2t + a2 r2 cos 4t + a3 r3 cos 6t + a4 r4 cos 8t.
Then, for example, for λ = 0.1 we find that
g(z) = −0.0000159163 + 0.00159187 z 2 − 0.000023828 z 4
+ 5.31236 × 10−7 z 6 − 1.39292 × 10−8 z 8 ,
which gives
G(z; 0) =
1 log z + 0.0001 − 0.010002 z 2 + 0.00015 z 4
2π
− 3.3 × 10−6 z 6 + 8.75 × 10−8 z 8 .
Note that f (0) = e0.0001 ≈ 1.0001. For computational details for λ = 0.1 and also
for λ = 0.5. Note that if λ = 0, then g(x, y) = 0, and Green’s function G(z; 0) is
the same as (8.1).
308
10. CONFORMAL MAPPING METHOD
Example 10.9. Consider the curve
F (x, y, α) = (x − 0.5)2 + (y − α)2 1 − y 2 − (x − 0.5)2 = 0.1
for α = ∞, 1, 0.5, 0.3, and 0.2746687749, which is plotted in Fig. 10.9. We will
consider the curve for α = 1 which is a cardioid shown in Fig. 10.8, and approximate
Green’s function for this region which is a nearly circular cardioid. Since the region
is symmetric about the line x = 0.5, we shall choose seven boundary points zj
as follows: z1 = 0.5 − 0.98725 i, z2 = 0.5 + 0.60343 i, z3 = 0.8 − 0.94024 i,
z4 = 0.8+0.65588 i, z5 = 1.2−0.69296 i, z6 = 1.2+0.59722 i, and z7 = 1.474003.
Green’s function is given by
G(z; 0.5) ≈
1
log r − 0.0431944 + 0.0404596 r cos θ
2π
+ 0.0107145 r2 cos 2θ − 0.0114393 r 3 cos 3θ
− 0.0569173 r sin θ + 0.039137 r 2 sin 2θ − 0.00477567 r 3 sin 3θ,
where r = |z − 0.5|. Since the origin is at the point (−0.5, 0), the distortion, given
by f (0.5) = e0.0431944 = 1.04415.
α = 0.2746687749
α = 0.3
0.75
0.5
α=1
0.25
α = 0.5
-0.5
0
0.5
1
1.5
-0.25
-0.5
-0.75
-1
Fig. 10.9. Graph of the Curve in Example 10.9.
Example 10.10. Cassini’s ovals, defined by
F (x, y, α) = (x + α)2 + y 2 (x − α)2 + y 2 = 1,
are plotted for α = 0, 0.5, 0.9, 0.99, and 1 in Fig. 10.10. For α = 0 the curve becomes
the unit circle. The region does does not remain simply connected for α = 1. We will
10.6. EXERCISES
309
approximate Green’s function for the Cassini’s oval at α = 0.5 which is bounded by
a nearly circular ellipse. Because of the symmetry of the region about the coordinate
axes, the function g(z) has a series expansion of the form
g(z) = a0 + a2 r2 cos 2θ + a4 r4 cos 4θ + · · · .
Let the points zj on the boundary be chosen as z1 = 1.11803, z2 = 1.04942 eiπ/6 ,
z3 = 0.98399 eiπ/4 , z4 = 0.92265 eiπ/3 , and z5 = 0.866 eiπ/2 . Then compute
a0 = −0.00513, a2 = 0.02124, a4 = −0.00282, a6 = 0.0005, a8 = −0.000096.
Thus, Green’s function is given by
1 G(z; 0) ≈
log z − 0.0322236 + 0.133465 z 2 − 0.0177425 z 4
2π
+0.0031616 z 6 − 0.000605 z 8 ,
with f (0) = e0.0322236 ≈ 1.03276, which has an error of about 3%.
Y
1
α= 0
α = .5
α = .9
α = .99
α= 1
-1
0
1
X
-1
Fig. 10.10. Cassini’s Ovals.
10.6. Exercises
10.1. Determine Green’s function for the positive (right) half-plane x1 ≥ 0, in Rn ,
n > 2.
Solution. Let x be an arbitrary point in the positive half-plane. Then Green’s
1
2−n
function is G(x, x ) =
x − x
− p(x), n > 2, where
(n − 2)Sn (1)
(2−n)/2
n
1
2
2
p(x) =
(x1 + x1 ) +
(xk − xk )
.
(n − 2)Sn (1)
k=2
310
10. CONFORMAL MAPPING METHOD
The function p(x) is obtained by taking the image of x in the boundary plane
x1 = 0. The solution of Laplace’s equation
∇2 u = 0 in this half-plane is given
∂
by u(x ) =
f (x)
G(x, x )
dS.
∂x
1
x1 =0
x1 =0
10.2. Show that Neumann’s function (10.59) satisfies the properties listed in §10.3.
Hint. Properties 1 and 2 are satisfied by the same method as used for Green’s
∂N
function (10.58). Property 3 can be established by computing
and showing
∂nz
∂N
1
that
= −
, which is independent of (u1 , v1 ).
2
∂nz u2 =u0
cosh u0 − cos2 v2
2π
cos nv
Since all terms containing v2 are involve the terms
dv which
sin nv
0
have zero value, the normalization condition (10.52) is satisfied.
10.3. Compute Green’s function for an equilateral triangle of side 2 units, where the
x-axis is parallel to the base of the triangle and 1/2 units above it and the y-axis
passes through the top vertex.
Hint. Use the symmetry and assume g(x, y) ≈ a0 + a1 r 2 cos 2θ + a1 r 2 cos 2θ
+a2 r 4 cos 4θ + · · · . Consider the two sets of points zj :
√
√
(i) Four points: (0, −1/2), (1, −1/2), 1/2, ( 3 − 1)/2 , (0, 3 − 1/2); and
√
(ii) Seven points: (0, −1/2) , (1/2, −1/2) , (1, −1/2) , 3/4, ( 3 − 2)/4 ,
√
√
√
1/2, ( 3 − 1)/2 , 1/4, (3 3 − 2)/4 , and (0, 3 − 1/2).
10.4. Show that Neumann’s function for the annulus with the outer radius r1 and
the inner radius r2 , and using cylindrical polar coordinates with z = r ei θ and
ζ = ρ ei φ , for r < ρ is given by
r2 ln r − r1 ln ρ r12 ln r1 − r22 ln r2
−
+
2
r1 + r2
(r1 + r2 )
n n
n
r2 n
r
ρ
r2
+
+
cos n(θ − φ)
∞
r
r1
r2
ρ
+
.
n
n
r1
r2
n=1
n
−
r2
r1
N (r, θ; ρ, φ) =
Hint. See §10.4.4.
10.5.
Green’s
function for
Compute
the regions bounded by the Cassini’s ovals
(x + 1)2 + y 2 (x − 1)2 + y 2 = a4 (Fig. 10.11) for the following cases:
a = 1.1, 1.2, and 1.5.
10.6. Compute Green’s function for the square with round corners x4 + y 4 = 1 (Fig.
10.12).
10.7. Compute Green’s function for the Limaçon r = a − cos θ, a = 1.5, 2,
0 ≤ θ < 2π (Fig. 10.13).
10.6. EXERCISES
Y
a = 1.5
311
a = 1.4
a = 1.3
a =2
2
a = 2.5
−2
X
2
a =3
−2
a =1
a = 1.1
a = 1.2
Fig. 10.11. Cassini’s Ovals for a = 1, 1.1, 1.2, 1.3, 1.4, 1.5, 2, 2.5, 3, 5.
10.8. Compute Green’s
function for the upper half of the bean-shaped curve defined
by z = 2.25 0.2 cos s + 0.1 cos 2s − 0.1
+i (0.35 sin s + 0.1 sin 2s − 0.02 sin 4s) , −π ≤ s ≤ π,
and shown in Fig. 10.14.
10.9. Consider the region D bounded by two hyperbolic arcs
Γ1 = arc ABC : {z : z = x0 − 2 cosh s + i (y0 − sinh s) , s1 < s < s2 }
Γ2 = arc CDA : {z : z = −x0 + 2 cosh s + i (−y0 + sinh s) , s1 < s < s2 },
where x0 = cosh s1 + cosh s2 , y0 = (sinh s1 + sinh s2 ) /2 are the coordinates
of the center z0 of the arc ABC (see Fig. 10.15). Take s2 = 1, and choose s1
such that 4 tanh s1 · tanh s2 + 1 = 0. Compute Green’s function for this region
D.
10.10. Consider the quadrilateral in Fig. 10.16. Take the parametric equations of
the sides as
1
AC: z = 2 + √
t + zA , 0 ≤ t ≤ 2,
3
312
10. CONFORMAL MAPPING METHOD
4
CE: z = − √ (t − 3)2 e2iπ/3 + zE ,
3
EG: z = −(t − 3)2 + zE ,
2 ≤ t ≤ 3,
3 ≤ t ≤ 5,
GA: z = −2 (t − 6) i + zA ,
5 ≤ t ≤ 6,
and compute Green’s function for this quadrilateral.
Y
a = 1.1
y
2
a = 1.2
1
x
0
a =1
−2
−3
−1
1
0
X
-1
a = 1.5
−2
a=2
Fig. 10.12.
Fig. 10.13.
Y
Y
0.75
0.6
0.5
0.4
Γ2
0.25
-0.5
0
0.5
X
−0.6
C
B
0.2
−0.4 −0.2 0
0.2
X
0.4 0.6
−0.2
-0.25
-0.5
−0.4
D
Γ1
−0.6
-0.75
A
Fig. 10.14.
Fig. 10.15.
10.11. Consider the octagonal region in Fig. 10.17. Take the parametric equations
of the sides as
10.6. EXERCISES
313
F G : z = −2it3 + z2 ,
LA : z = 2it3 + z1 ,
−1 ≤ t ≤ 0;
BD : z = t zD + (1 − t)3 zB ,
3
0 ≤ t ≤ 1,
3
AB : z = 2t + z1 ,
DF : z = t zF + (1 − t) zD ,
GH : z = −2t + z2 ,
HK : z = t zK + (1 − t) zH ,
KL : z = t zL + (1 − t)3 zK ,
3
3
3
3
3
0 ≤ t ≤ 1,
0 ≤ t ≤ 1.
3
Compute Green’s function for this octagonal region.
Hint.This region is not convex; it is starlike with respect to any point on the line
segment AG, but since it has fourfold symmetry about the origin, consider either
half of the region.
F
G
E (2, 1)
2 π/3
D
0
H
A
C
B
(−2, −1)
Fig. 10.16.
10.12. Consider the circular sector of radius 1 and angle 3π/2 (Fig. 10.18). This
region is not convex; it is starlike with respect to any point on the line segment
AC, but it is symmetric about the origin along the line P1 P2 , and each part is
convex. Compute Green’s function for each part.
E
F
H
G
D (5, 5)
z2
− 3 , 3
2√ 2 2√ 2
)
C
0
C
•
0
z1
K
.(
p1
A
L
Fig. 10.17.
B
A
( 2√1 2 , − 2√1 2 )
B (5, −3)
D
Fig. 10.18.
.(
p
2
5 ,− 5
2√ 2
2√ 2
)
A
Adjoint Operators
A system of two first order linear homogeneous differential equations
y1 = a11 (x) y1 + a12 (x) y2 ,
(A.1)
y2 = a21 (x) y1 + a22 (x) y2 ,
where aij (x) are real valued functions of x, can be written in matrix form as
y = A(x) y,
a
where A(x) = 11
a21
a12
.
a22
(A.2)
A system of linear homogeneous differential equations is called adjoint to (A.1) if it
has the form y = B(x) y, where B = −AT , and AT denotes the transpose of A.
A second-order differential equation can always be reduced to a system of first
order equations. For example, the general second order differential equation
y + a1 (x) y + a2 (x) y = 0,
(A.3)
where ai , i = 1, 2, are real valued functions of x, is reduced to a system of first order
equations by setting y = y1 and y = y2 . Thus, (A.3) becomes
y1 = y2 ,
y2 = −a2 y − a1 y,
0
where A =
−a2
y = −AT y, or
1
0
T
. Since A =
−a1
1
or y = A(x) y,
−a2
, the adjoint system to (A.4) is
−a1
y1 = a2 (x) y2 ,
y2 = −y1 + a1 (x) y2 .
314
(A.4)
A. ADJOINT OPERATORS
315
After differentiating the second equation and substituting a2 y2 for y1 from the first
equation and writing y for y2 , this system reduces to
y − [a1 (x) y] + a2 (x) y = 0,
which is the adjoint equation to (A.3). In the case when the coefficient of y in Eq
(A.3) is a0 (x) = 0, i.e., if we have a0 (x) y + a1 (x) y + a2 (x) y = 0, then its adjoint
equation is
[a0 (x) y] − [a1 (x) y] + a2 (x) y = 0.
Definition. The operators
d2
d
+ a1
+ a2 ,
2
dx
dx
d2
d
M ≡ 2 (a0 ·) −
(a1 ·) + a2 ,
dx
dx
L ≡ a0
(A.5)
(A.6)
where the dot denotes the place holder for the function to which M is applied, are
called adjoint differential operators of second order.
Lagrange’s Identity. If L and M are adjoint differential operators of second
order, there there exists a function F (x, y, y , z, z ) such that
d
F (x, y, y , z, z ).
(A.7)
dx
Proof. zL[y] −yM [z] = z [a0 y + a1 y + a2 ] −y (a0 z) + y (a1 z) + a2 z
d =
a0 y z − y (a0 z) + y (a1 z) .
dz
Corollary. If L and M are adjoint operators of second order, then
z L[y] − y M [z] =
x2
zL[y] − yL[z] dx = a0 y z − y (a0 z) + y (a1 z)
x1
x2
x−1
.
(A.8)
A differential equation of second order can be written in the (Sturm-Liouville)
form
d
L[y] ≡
[p(x) y ] + q(x) y = 0.
(A.9)
dx
Lemma 1. A second order differential equation L[y] = 0 of the form (A.9) is
self-adjoint (i.e., adjoint to itself) iff the operator L has the form
L≡
where p(x) is differentiable.
d
d
p(x)
(·) + q(x),
dx
dx
(A.10)
316
A. ADJOINT OPERATORS
Proof. Since M is the adjoint operator to L, we have, in view of (A.6),
M≡
d2
d d2
d
[p(x) (·)] −
[p (x) (·)] + q(x) = p(x) 2 (·) + p (x) (·) + q(x),
2
dx
dx
dx
dx
d2
d
d2
i.e., L = M . Conversely, if L = a0 2 + a1
+ a2 and M =
(a0 ·) −
dx
dx
dx2
d
(a1 ·) + a2 are equal, then a0 = a1 = p (x).
dx
Lemma 2. The Lagrange’s identity for self-adjoint differential operators become
zL[y] − yL[z] =
d
[p(x) (y z − yz )] .
dx
(A.11)
B
List of Fundamental Solutions
B.1. Linear Ordinary Differential Operator with Constant Coefficients that satisfies the equation
Lu∗ (x) =
dn u ∗
dn−1 u∗
du∗
+
a
+
·
·
·
+
a
+ an u∗ = δ(x),
1
n−1
dxn
dxn−1
dx
where u∗ = u∗ (x, 0), has fundamental solution:
u∗ = H(x) w(x),
(B.1)
where H(x) is the Heaviside function, and w(x) ∈ C n (R) is the solution of the
homogeneous equation Lw = 0 with the initial conditions w(0) = w (0) = · · · =
w(n−2) (0) = 0, w(n−1) (0) = 1.
B.2. Fundamental Solutions for the Operators L1 =
d2
+ a2 are, respectively, given by
dx2
u∗1 (x) = H(x) e−ax ,
sin ax
u∗2 (x) = H(x)
.
a
d
+ a, and L2 =
dx
(B.2)
(B.3)
B.3. Elliptic Operator. Fundamental solution for the Laplace operator or the
Laplacian ∇2 in Rn , such that −∇2 u∗n (x) = δ(x), x = {x, . . . , xn }, is given by
u∗n (x) =
1
|x|2−n ,
(n − 2)Sn (1)
317
n ≥ 3,
(B.4)
318
B. LIST OF FUNDAMENTAL SOLUTIONS
2π n/2
denotes surface area of the unit ball.
Γ(n/2)
1
Special Cases: For n = 2:
u∗2 (x, x ) = −
ln |x − x |;
2π
1
For n ≥ 3:
u∗3 (x, x ) =
.
4π |x − x |
where Sn (1) =
d2 u∗1
= δ(x), which gives the fundamental
dx2
1
solution: u∗1 (x) = − |x|.
2
∗
B.4. Helmholtz Operator.
Let un (x) denote the fundamental solution for the
Helmholtz operator − ∇2 + k 2 in Rn , where k is complex, such that
For n = 1: solve −
− ∇2 + k 2 u∗n (x) = δ(x).
(B.5)
For n ≥ 2 and k 2 ∈ [0, ∞), the fundamental solution for (B.5) is given by
u∗n (r) =
i k (n/2)−1 (1)
H(n/2)−1 (kr),
4 2π
r = |x − x |,
n ≥ 2,
(B.6)
2
K(n/2)−1 (hr), h = −ik, where
iπ
(j)
H0 , j = 1, 2, are the Hankel functions, and K(n/2)−1 are the modified Bessel
functions, by
(1)
(1)
or, since H(n/2)−1 (kr) = H(n/2)−1 (ihr) =
u∗n (r) =
1 h (n/2)−1
K(n/2)−1 (hr),
2π 2πr
n ≥ 2,
(B.7)
which holds whenever −h2 ∈ [0, ∞), i.e., for all h with {h} > 0.
For n = 2, the fundamental solution is
u∗2 (r) =
(1)
i (1)
1
H0 (kr) =
K0 (hr).
4
2π
For n = 3, by using H1/2 (z) =
(B.8)
1 2 1/2 eiz
, we get
i π
z 1/2
u∗3 (r) =
eikr
e−hr
=
.
4πr
4πr
(B.9)
For n = 1, the fundamental solution is found directly as
u∗1 (x) =
1
sin k|x|.
2k
(B.10)
B. LIST OF FUNDAMENTAL SOLUTIONS
319
Note that if the Helmholtz equation is taken as ∇2 + k 2 u = 0, where k = 0, real,
denotes the wave number, then the fundamental solution in R is the same as (B.10),
but in R2 and R3 they are, respectively,
i (2)
u∗2 (r) = − H0 (kr),
4
and
u∗3 (r) =
e−ikr
.
4πr
(B.11)
Fundamental solution for the operator ∇2 − k 2 are given by

1


sinh kr for n = 1,


2k


1
u∗n (x, x ) =
K0 (kr) for n = 2,

2π




 1 eikr for n = 3,
4πr
(B.12)
where r = |x − x |. Note that lim u∗ (x, x ) = ∞. These fundamental solutions
x→±∞
are used in problems of neutron diffusion and reactor physics.
∂
B.5. Fundamental Solution for the Cauchy-Riemann Operator
=
∂ z̄
∂
∂
∂ ∗
∂ ∗
+ i , z = x + i y, which satisfies
u (z) =
u (x, y) = δ(x, y), is
∂x
∂y
∂ z̄
∂ z̄
1
u∗ (z) =
.
πz
∂
B.6. Fundamental Solution for the Diffusion Operator
− a∇2 , also
∂t
∂u∗
called the heat conduction operator, which satisfies
− a∇2 u∗ = δ(x, t), where
∂t
a > 0 is a constant determined by specific heat and thermal conductivity of the
medium, is
2
H(t)
u∗ (x, t) =
e−|x| /4at .
(B.13)
(4πat)n/2
Note that if the fundamental solution for the diffusion equation in R is taken
at initial time t and the instantaneous point heat source is located at x , then the
fundamental solution is written as
u∗ (x, x ; t, t ) =
1
4πa(t −
2
t )
e−(x−x )
/4a(t−t )
.
(B.14)
In this notation, the coordinates (x , t ) express the role of the source point, whereas
the coordinates (x, t) represent the field (or observation) point. In this sense, the value
of u∗ at the point x at time t is determined by the evolution of the distribution δ(x ) at
time t . Thus, the fundamental solution (B.14) can be regarded as the superposition
of the evolutions of initial distributions of the δ-functions at points x with amplitudes
f (x ).
320
B. LIST OF FUNDAMENTAL SOLUTIONS
B.7. Schrödinger Operator. The linear diffusion operator in Rn × R+ is
sometimes written as
n
1 ∂
1 ∂
− ∇2 =
−
a ∂t
a ∂t j=1
∂
∂xj
2
.
But the fundamental solution is still given by (B.13). If we consider the nonhomogeneous case, then the equation
1 ∂
− ∇2 u(x, t) = f (x, t)
a ∂t
(B.15)
is known as the Fourier heat equation. There are two interpretation of Eq (B.15).
Firstly, if a > 0 is a real constant depending on specific heat and thermal conductivity
of the medium, then u(x, t) determines the temperature distribution. The function
f (x, t) on the right side describes local heat production minus absorption. Secondly,
u(x, t) defines a particle density and a is the diffusion coefficient. If a is purely
i
imaginary such that a =
, where m is the mass of the quantum particle, and
2m
−27
= 1.054 × 10
erg-sec is the Planck’s constant, then Eq (B.15) defines the
Schrödinger equation
i
∂u(x, t)
2 2
+
∇ u(x, t) = f (x, t).
∂t
2m
(B.16)
If the quantum particle of mass m is an external force field with potential V (x),
2
and u(x, t) denotes the wave function of the particle such that |u(x, t| ∆u is the
probability of the particle being in the neighborhood u(x) of the point x at time t,
where ∆u denotes the volume of u(x, t), then f (x, t) = V (x)u(x, t), and Eq (B.16)
becomes
∂u
2 2
i
+
∇ u = V u.
(B.17)
∂t
2m
In the steady state, the energy E of the particle has a definite value, and the function
u has the form
u(x, t) = e−iEt/ u(x).
(B.18)
Thus, in view of (B.17), the wave function u(x) satisfies the steady state Schrödinger
equation
2 2
∇ u+Eu=V u
(B.19)
2m
If V = 0, the free particle Schrödinger equation reduces to homogeneous Helmholtz
equation. As in the case of the Helmholtz equation, so for the fundamental solution
we must require that Sommerfeld’s radiation condition
V (x) = O(|x|),
∂V (x)
−1
− i kV (x) = o |x|
∂|x|
(B.20)
B. LIST OF FUNDAMENTAL SOLUTIONS
321
√
2mE/, E ≥ 0. From (B.13) the fundamental
∂
2 ∂ 2
solution for the Schrödinger operator i +
in R × R+ is given by
∂t 2m ∂x2
1+i
m i mx2 /2t
u∗ (x, t) = −H(t) √
e
.
(B.21)
2 2πt
must be satisfied at infinity with k =
∂2
− c2 ∇2
∂t2
in Rn × R+ , which satisfies the equation c u∗n (x, t) = δ(x, t), where x ∈ Rn and
t ∈ (0, t) ⊂ R+ , is given by
1
In R: u∗1 (x, t) = H(ct − |x|),
2c
H(ct − |x|)
In R2 : u∗2 (x, t) =
.
2πc c2 t2 − |x|2
B.8. Fundamental Solution for the Wave Operator c =
H(t)
H(t)
δS (x) =
δ(c2 t2 − |x|2 ).
4πc2 t ct
2πc
B.9. Fundamental Solution for the Fokker-Plank Operator which is a
modification of the diffusion operator in §B.6, with a = 1, i.e., for the Fokker-Planck
∂u
∂
∂
equation
=
+ x u, is
∂t
∂x ∂x
In R3 :
u∗3 (x, t) =
u∗ (x, x ; t, t ) =
H(t − t )
2π(1 −
e−2(t−t ) )
−(t−t ) 2
e−(x−x e
) /2(1−e−2(t−t ) )
.
(B.22)
This equation is used in statistical physics to describe the evolution of probability
distribution functions.
B.10. Klein-Gordon Operator is defined by + m20 , where m0 is the mass of
3
+
a free relativistic pseudo-scalar particle. Then the wave
function
u(x, ct) in R ×R ,
2
where c is the speed of light, satisfies the equation + m0 u = 0. For m0 = 1,
the fundamental solution for the Klein-Gordon operator is a solution of the equation
∂ 2 u∗
∂ 2 u∗
−
+ u∗ = δ(x − x )δ(t − t ),
2
∂t
∂x2
(B.23)
such that u∗ (x − x , 0) = 0 = u∗t (x − x , 0). Fundamental solution for this operator
is
u∗ (x, x ; t, t ) =
1
J0 [ (t − t )2 − (x − x )2 ] H[t − t − (x − x )]×
2
× H[t − t + (x − x )].
(B.24)
C
List of Spherical Harmonics
C.1. Legendre’s Equation. Legendre’s equation of order n is
d2 u
du
1 − x2
−2x
+n(n+1) u = 0.
(C.1)
2
dx
dx
If n = 0, 1, 2, . . . , a solution of this equation is the Legendre polynomials Pn (x)
given by Rodrigue’s formula
1 dn 2
Pn (x) = n
x −1 .
(C.2)
n
2 n! dx
C.1.1. Special Legendre polynomials are:
P0 (x) = 1,
P4 (x) = 18 35x4 − 30x2 + 3 ,
P1 (x) = x,
P5 (x) = 18 63x5 − 70x3 + 15x ,
1
P2 (x) = 12 3x2 − 1 ,
P6 (x) = 16
231x6 − 315x4 + 105x2 − 5 ,
1
P3 (x) = 12 5x3 − 3x2 ,
P7 (x) = 16
429x7 − 693x5 + 315x3 − 35x .
C.1.2. If we set x = cos θ, then some of the above Legendre polynomials become
P0 (cos θ) = 1,
P1 (cos θ) = cos θ,
P2 (cos θ) =
1
4
(1 + 3 cos 2θ),
P3 (cos θ) =
1
8
(3 cos θ + 5 cos 3θ),
P4 (cos θ) =
1
64
P5 (cos θ) =
1
128
(9 + 20 cos 2theta + 35 cos 4θ),
(30 cos θ + 35 cos 3θ + 63 cos 5θ).
322
C. LIST OF SPHERICAL HARMONICS
323
C.1.3. The recurrence formulas for Legendre polynomials are:
(n + 1)Pn+1 (x) − (2n + 1)xPn (x) + nPn−1 (x) = 0,
Pn+1
(x) − xPn (x) = (n + 1)Pn (x),
xPn (x) − Pn−1
(x) = nPn (x),
Pn+1
(x) − Pn−1
(x) = (2n + 1)Pn (x),
2
x − 1 Ln (x) = nxPn (x) − nPn−1 (x),
C.1.4. Orthogonality. Orthogonality property in the interval −1 ≤ x ≤ 1:

1
 0, m = n,
Pm (x) Pn (x) dx =
2

, m = n.
−1
2n + 1
C.1.5. Generating Function:
∞
−1/2 1 − 2xt + t2
=
Pn (x) tn , |t| < 1 >
n=0
C.1.6. Some special results involving Legendre polynomials are:
Pn (1) = 1, Pn (−1) = (−1)n , Pn (−x) = (−1)n Pn (x),

 0, n odd,
Pn (0) =
 (−1)n/2 1 · 3 · 5 · · · (n − 1) , n even,
2 · 4 · 6···n
π
n
1
Pn (x) =
x + x2 − 1 cos φ dφ,
π 0
Pn+1 (x) − Pn−1 (x)
Pn (x) dx =
,
2n + 1
Pn (x) ≤ 1,
Pn (x) =
1
2n+1 π i
C
n
z2 − 1
dz,
(z − x)n+1
where C is a simple closed curve with x as an interior point.
There are Legendre functions of the second kind and order n, denoted by Qn (x),
but we will not discuss them since they do not concern us.
C.2. Associated Legendre’s Equation. We mentioned in §4.3.3 that Eq
(4.15) is the associated Legendre equation. We will rewrite it in the form
1 − x2 u − 2xy + n(n + 1) −
µ2
1 − x2
= 0.
(C.3)
Solutions of this equation are called the associated Legendre functions. We will
restrict our discussion to the case where µ and n are nonnegative integers.
324
C. LIST OF SPHERICAL HARMONICS
C.2.1. Associated Legendre Functions of the First Kind. These functions are defined by
n
Pm
(x)
2 m/2
= 1−x
m/2
1 − x2
dm
dm+n 2
P
(x)
=
x −1 ,
n
m
n
m+n
dx
2 n!
dx
(C.4)
where Pn (x) are Legendre polynomials discussed above. We have
Pn0 (x) = Pn (x);
Pnm (x) = 0 if m > n.
C.2.2. Special Associated Legendre Functions of the First Kind:
1/2
1/2
P11 (x) = 1 − x2
,
P31 (x) = 32 5x2 − 1 1 − x2
,
1/2
P21 (x) = 3x 1 − x2
,
P32 (x) = 15x 1 − x2 ,
3/2
P22 (x) = 3 1 − x2 ,
P33 (x) = 15 1 − x2
.
C.2.3. Recurrence Formulas:
m
m
(n + 1 − m)Pn+1
(x) − (2n + 1)xPnm (x) + (n + m)Pn−1
(x) = 0,
Pnm+2 (x) −
2(m + 1)x
(1 −
1/2
x2 )
Pnn+1 (x) + (n − m)(n + m + 1)Pnm (x) = 0.
C.2.4. Orthogonality: In the interval −1 ≤ x ≤ 1

1
 0 if k = n,
m
m
Pk (x)Pn (x) =
2
(n + m)!

if k = n.
−1
2n + 1 (n − m)!
There are associated Legendre functions of the second kind and order n, denoted
by Qm
n (x), but we will not discuss them since they do not concern us.
C.2.5. Generating Function:
∞
(2m)!(1 − x2 )m/2 tm
=
P m (x) tn ,
2m m! (1 − 2xt + t2 )m+1/2 n=m n
|t| < 1.
C.3. Relations with or without Condon-Shortley Phase Factor. Various kinds of relations concerning associated Legendre functions of the first kind with
and without the Condon-Shortley phase factor are presented here.
C.3.1. Some Orthonormalized Laplace’s Spherical Harmonics with
the Condon-Shortley Phase Factor:
1 1
Y00 (θ, ϕ) =
,
2 π
1
3 −i ϕ
−1
Y1 (θ, ϕ) =
e
sin θ,
2 2π
1 3
0
Y1 (θ, ϕ) =
cos θ,
2 π
C. LIST OF SPHERICAL HARMONICS
Y11 (θ, ϕ) = −
Y2−2 (θ, ϕ)
1
2
1
=
4
325
3 iϕ
e sin θ,
2π
15 −2i ϕ
e
sin2 θ,
2π
1 15 −i ϕ
=
e
sin θ cos θ,
2 2π
1 5 0
Y2 (θ, ϕ) =
3 cos2 θ − 1 ,
4 π
1 15 i ϕ
1
Y2 (θ, ϕ) = −
e sin θ cos θ,
2 2π
1 15 2i ϕ
Y22 (θ, ϕ) =
e
sin2 θ,
4 2π
1 7 0
Y3 (θ, ϕ) =
5 cos3 θ − 3 cos θ .
4 π
Y2−1 (θ, ϕ)
C.3.2. Recurrence Relations of the Associated Legendre Functions
Pnm without Condon-Shortley Phase Factor:
m
m
(n + 1 − m) Pn+1
(cos θ) − (2n + 1) cos θ Pnm (cos θ) + (n + m)Pn−1
(cos θ) = 0;
Pnm+2 (cos θ) −
2(m + 1) cos θ m+1
Pn (cos θ) + (n − m)(n + m + 1)Pnm (cos θ) = 0.
sin θ
C.3.3. Orthogonality: In the interval −1 ≤ x ≤ 1
1
Pnm (x) Pkm (x) dx = 0 if n = k;
−1
1
−1
2
[Pnm (x)] dx =
2
(n + m)!
, where x = cos θ.
2n + 1 (n − m)!
C.3.4. Without Condon-Shortley Phase Factor:
Recurrence relations of associated Legendre functions:
1.
Pn0 (sin θ) = 1;
2.
m−1
Pnm (sin θ) = (2n − 1) cos θ Pn−1
(sin θ) for all n ≥ 1;
3.
n−1
Pnn−1 (sin θ) = (2n − 1) sin θ Pn−1
(sin θ) for all n ≥ 1;
1
m
m
sin θ Pn−1
(sin θ) − (n + m − 1)Pn−2
(sin θ) for all
n−m
π
π
m ≥ 0 and m ≤ n − 2 and n ≥ 2; where | sin θ| ≤ 1, − 2 ≤ θ ≤ 2 , and θ is the
elevation angle.
4.
Pnm (sin θ) =
326
C. LIST OF SPHERICAL HARMONICS
C.3.5. Calculated Associated Legendre Functions:
1.
P00 (sin θ) = 1;
2.
P10 (sin θ) = sin θ;
3.
4.
P11 (sin θ) = cos θ;
P20 (sin θ) = 12 3 sin2 θ − 1 ;
5.
P21 (sin θ) = 3 cos θ sin θ;
6.
8.
P22 (sin θ) = 3 cos2 θ;
P30 (sin θ) = 12 sin θ 5 sin2 θ − 3 ;
P31 (sin θ) = 32 cos θ 5 sin2 θ − 1 ;
7.
P32 (sin θ) = 15 cos2 θ sin θ;
8.
P33 (sin θ) = 15 sin3 θ.
7.
C.4. Laguerre’s Equation. This equation is: xu + (1 − x)u + nu = 0. If
n = 0, 1, 2, . . . , then a solution of this equation is the Laguerre polynomial Ln (x)
defined by Rodrigue’s formula
dn Ln (x) = ex n xn e−x .
(C.5)
dx
C.4.1. Special Laguerre Polynomials:
L0 (x) = 1,
L3 (x) = −x3 + 9x2 − 18x + 6,
L1 (x) = −x + 1,
L5 (x) = x4 − 16x3 + 72x2 − 96x + 24,
L2 (x) = x2 −4x+2,
L3 (x) = −x5 +25x4 −200x3 +600x2 −600x+120,
L6 (x) = x6 − 36x5 + 450x4 − 2400x3 + 5400x2 − 4320x + 720.
C.4.2. Recurrence Formulas:
Ln+1 (x) − (2n + 1 − x)Ln (x) + n2 Ln−1 (x) = 0,
Ln (x) − nLn−1 (x) + nLn−1 (x) = 0,
xLn (x) = nLn (x) − n2 Ln−1 (x) = 0.
C.4.3. Orthogonality: On the interval [0, ∞)
∞
0 if m = n,
e−x Lm (x)Ln (x) dx =
2
(n!)
if m = n.
0
C.4.4. Special Results:
x
Ln (t) dt = Ln (x) −
Ln (0) = n!,
0
Ln+1 (x)
,
n+1
n2 xn−1
n2 (n − 1)2 x2
Ln (x) = (−1)n xn −
+
− · · · (−1)n n! ,
1!
2!
C. LIST OF SPHERICAL HARMONICS
∞
xk e−x Ln (x) dx =
0
n
k=0
327
0 if k < n,
(−1)n (n!)2 if k = n,
Lk (x)Lk (y)
Ln (x)Ln+1 (y) − Ln+1 (x)Ln (y)
=
,
2
(k!)
(n!)2 (x − y)
∞ k
t Lk (x)
k=0
(k!)2
√
= et J0 2 xt ,
∞
Ln (x) =
√
tn ex−t J0 2 xt dt.
0
C.5. Associated Laguerre’s Equation: xu +(m+1−x)u +(n−m)u = 0.
C.5.1. Associated Laguerre Polynomials: These are the solutions of the
associated Laguerre’s equation for integers m and n, defined by
Lm
n (x) =
dm
Ln (x),
dxm
where Ln (x) are the Laguerre polynomials (§C.4), such that
L0n (x) = Ln (x);
Lm
n (x) = 0 if m > n.
C.5.1.1. Special Results:
L11 (x) = −1,
L33 (x) = −6,
L12 (x) = 2x − 4,
L14 (x) = 4x3 − 48x2 + 144x − 96,
L22 (x) = 2,
L24 (x) = 12x2 − 96x + 144,
L13 (x) = −3x2 + 18x − 18,
L34 (x) = 24x − 96,
L23 (x) = −6x + 18,
L44 (x) = 24.
∞
(2n − m + 1) (n!)2
2
xm+1 e−x [Lm
(x)]
dx
=
.
n
(n − m)!
0
C.5.2. Recurrence Formulas:
n−m+1 m
2 m
Ln+1 (x) + (x + m − 2n − 1)Lm
n (x) + n Ln−1 (x) = 0,
n+1
d m L (x) = Lm+1
(x),
n
dx n
d m −x m x e Ln (x) = (m − n − 1)xm−1 e−x Lm−1
(x),
n
dx
d m m−1
x
L (x) = (x − m)Lm
(x),
n (x) + (m − n − 1)Ln
dx n
328
C. LIST OF SPHERICAL HARMONICS
Lm
n (x)
n!
n(n − m) n−m−1
= (−1)
xn−m −
x
(n − m)!
1!
n
+
n(n − 1)(n − m)(n − m − 1) n−m−2
x
+ ··· .
2!
C.5.3. Orthogonality: on the interval [0, ∞)

∞
 0 if k = n,
m
xm e−x Lm
(x)L
(x)
dx
=
(n!)2
n
k

if k = n.
0
(n − m)!
D
Tables of Integral Transforms
D.1. Laplace Transform Pairs
L{f (t)} = F (s) = f¯(s)
f (t)
1.
1
2.
eat
3.
sin at
4.
cos at
5.
sinh at
6.
cosh at
7.
eat sin bt
8.
eat cos bt
9.
tn
10.
tn eat
11.
H(t − a)
12.
H(t − a) f (t − a)
1
, s>0
s
1
, s>a
s−a
a
, s>0
2
s + a2
s
, s>0
2
s + a2
a
, s>0
2
s − a2
s
, s>0
2
s − a2
b
, s>a
(s − a)2 + b2
s
, s>a
(s − a)2 + b2
n!
, s>0
sn+1
n!
, s>a
(s − a)n+1
e−as
, s>0
s
e−as F (s) = e−as f¯(s)
13.
eat f (t)
F (s − a) = f¯(s − a)
14.
f (t) g(t)
F (s) G(s) = f¯(s) ḡ(s)
(n = 1, 2, . . . )
(n = 1, 2, . . . )
329
330
D. TABLES OF INTEGRAL TRANSFORMS
L{f (t)} = F (s) = f¯(s)
f (t)
15.
f (n) (t)
16.
f (at)
t
17.
0
f (t) dt
sn F (s) − sn−1 f (0) − · · · − f n−1 (0)
1 s
F
, a>0
a
a
1
1
F (s) = f¯(s)
s
s
18.
δ(t − a)
e−as
19.
t f (t)
−
a erf √
2 t
a erfc √
2 t
1 − e−a
s
20.
21.
22.
f (t) with period = T †
23.
J0 (at)
24.
I0 (at)
25.
Jn (at)
26.
In (at)
27.
t J1 (at)
28.
t I1 (at)
29.
†
tn Jn (at)
30.
†
tn In (at)
31.
t J0 (at)
32.
t I0 (at)
d
F (s)
ds
√
s
√
e−a s
s
T −st
e
f (t) dt
0
1 − e−T s
1
√
s2 + a2
1
√
2
s − a2
n
1
a
√
√
, n > −1
s2 + a2 s + s2 + a2
n
1
a
√
√
, n > −1
2
2
s −a
s + s2 − a2
a
, a>0
3/2
2
(s + a2 )
a
, a>0
3/2
(s2 − a2 )
(2n)! an
√
2n+1 , n > −1/2
2n n!
s2 + a2
2n Γ(n + 1/2) an
√ √
2n+1 , n > −1/2
π
s2 − a2
s
, a>0
3/2
(s2 + a2 )
s
, a>0
3/2
(s2 − a2 )
†f (t) is continuous in [0, T ] and periodic with period T , T > 0.
D. TABLES OF INTEGRAL TRANSFORMS
L{f (t)} = F (s) = f¯(s)
f (t)
(t − a)µ−1
H(t − a)
Γ(µ)
t (µ−1)/2
√ Jµ−1 2 at
a
t (µ−1)/2
√ Iµ−1 2 at
a
√
cos 2 at
√
πt
√
sin 2 at
√
πa
e−as
,
sµ
e−a /4t
√
πt
−a2 /4t
ae
√
2 π t3
e−bt − e−at
t
e−a
√
e−a
41.
δ(t)
1
42.
δ(t − a)
e−as ,
33.
34.
35.
36.
37.
2
38.
39.
40.
†
Note that {s} >
µ>0
e−a/s
, µ>0
sµ
ea/s
, µ>0
sµ
e−a/s
√ ,
s
a>0
e−a/s
,
s3/2
a>0
ln
√
s
s
√
s
,
a>0
,
a>0
s + a
s+b
,
a, b > 0
a>0
{a} (in 29), and {s} > {a} (in 30).
331
332
D. TABLES OF INTEGRAL TRANSFORMS
D.2. Fourier Cosine Transform Pairs
Fc {f (x)} = f˜c (α)
1 ˜ α
fc
a
a
f (x)
1.
f (ax)
2.
e−ax
3.
x−1/2
4.
e−ax
5.
x2
a
+ a2
x2 f (x)
7.
sin ax
x
8.
f (x)
9.
δ(x)
10.
H(a − x)
√
a + x2 + a 2
a 2 + x2
(a2 − x2 )1/2 ,
0, x > a
12.
13.
(x2 − a2 )−1/2 ,
14.
0, 0 < x < a
sin a2 x2
15.
cos a2 x2
a>0
1
√
α
1 −α2 /4a
√ e
, a>0
2a
π −aα
e
, a>0
2
−f˜ (α)
2
6.
11.
2
a
,
2
π α + a2
c
π
H(a − α)
2
−α2 f˜c (α) −
2 f (0)
π
2
π
2 sin aα
π α
0<x<a
a<x<∞
e−aα
√
α
π aJ1 (aα)
2
α
π
Y0 (aα)
2
π cos(α2 /4a2 ) − sin(α2 /4a2 )
4a
π cos(α2 /4a2 ) + sin(α2 /4a2 )
4a
−
D. TABLES OF INTEGRAL TRANSFORMS
333
D.3. Fourier Sine Transform Pairs
Fs {f (x)} = f˜s (α)
1 ˜ α
fs
a
a
f (x)
1.
f (ax)
2.
e−ax
2
α
2
π a + α2
3.
x−1/2
4.
x−1
5.
x
x2 + a 2
6.
arctan(a/x)
7.
x2 f (x)
1
√
α
π
2
π −aα
e
2
π 1 − e−aα
2
α
−f˜ (α)
8.
x
erfc √ ,
2 a
9.
f (x)
10.
H(a − x)
11.
x e−a
12.
13.
†
2
†
s
2 1 − e−aα
π
α
2
a>0
−α2 f˜s (α) +
x2
1
√ e−a/x
x
sin bx
x2 + a 2
| arg a| < π/2.
2
α f (0)
π
2 1 − cos aα
π
α
πα
2
2
√
e−α /4a
4 2 a3
√
√
√
1
√ e− 2aα (cos 2aα + sin 2aα ),
α







π
2
π
2
1 −ab
e
sinh(aα),
a
1 −aα
e
sinh(ab),
a
0<α<b
b<α<∞
334
D. TABLES OF INTEGRAL TRANSFORMS
D.4. Complex Fourier Transform Pairs
F{f (x)} = f˜(α)
f (x)
1.
f (n) (x)
2.
f (ax),
3.
f (x − a)
4.
δ(x − a)
5.
e−a|x|
6.
e−a
7.
2
(−iα)n f˜(α)
a>0
x2
1,
0,
|x| < a
|x| > a
8.
1
|x|
9.
f (x) g(x)
10.
H(x + a) − H(x − a)
11.
x e−a|x| ,
12.
a
x2 + a2
ax
(x2 + a2 )2
13.
14.
cos ax
15.
sin ax
16.
cos ax,
0,
a>0
|x| < π/2a
|x| > π/2a
1 ˜ α f
a
a
iaα ˜
e
f (α)
1
√ e−i aα
2π
2
a
, a>0
π a2 + α2
1
2
2
√ e−α /4a
a 2
2 sin aα
π α
1
|α|
√
2π f˜(α) g̃(α)
2 sin aα
π α
2
2iaα
π (α2 + a2 )2
π −a|α|
e
2
i π
α e−a|α|
2 2
π
δ(α + a) + δ(α − a)
2
π
i
δ(α + a) − δ(α − a)
2
2
a
cos(πα/2a)
π a2 − α2
D. TABLES OF INTEGRAL TRANSFORMS
17.
18.

 √ 1
, |x| < a
a 2 − x2

0, |x| > a
1 − |x|, |x| < 1
0, |x| > 1
335
π
J0 (aα)
2
2
2 sin(α/2)
2
π
α
D.5. Finite Sine Transform Pairs
The tables are for the interval [0, π]. If the interval is [a, b], then it can be transformed
into [0, π] by
y=
2.
sin mx,
∞
(D.1)
f˜s (n)
f (x)
1.
π (x − a)
.
b−a
m = 1, 2, . . .
an sin nx
π/2,
n=m
0,
n=m
π an /2
n=1
3.
π−x
4.
x
5.
1
π/n
−x, x ≤ a
π − x, x > a
x(π − a), x ≤ a
6.
7.
a(π − x),
8.
e
9.
sinh a(π − x)
sinh aπ
10.
f (x)
ax
x>a
π
(−1)n+1
n
1
1 − (−1)n
n
π
cos na, 0 < a < π
n
π
sin na, 0 < a < π
n2
n
1 − (−1)n eaπ
2
2
a +n
n
2
a + n2
−n2 f˜s (n) + n f (0) − (−1)n f (π)
336
D. TABLES OF INTEGRAL TRANSFORMS
D.6. Finite Cosine Transform Pairs
The tables are for the interval [0, π]. If the interval is [a, b], then it can be transformed
into [0, π] by formula (B.36).
f˜c (n)
f (x)
1.
cos mx,
3.
∞
a0
+
an cos nx
2
n=1
f (π − x)
4.
1
2.
π/2,
0,
m = 1, 2, . . .
π an /2
(−1)n f˜c (n)
π,
0,
5.
x
6.
7.
8.
9.
10.
x
2
1, 0 < x < a
−1, a < x < π
eax
a
cosh (c(π − x))
sinh(πc)
f (x)
n=m
n=m
n=0
n = 1, 2, . . .
π 2 /2, n = 0
1 (−1)n − 1 ,
2
n
n = 1, 2, . . .
π 3 /3, n = 0
2π
(−1)n , n = 1, 2, . . .
n2
2a − π, n = 0
2
sin na, n = 1, 2, . . .
n
(−1)n eaπ − 1
a2 + n2
c
2
c + n2
−n2 f˜c (n) + (−1)n f (π) − f (0)
D. TABLES OF INTEGRAL TRANSFORMS
337
D.7. Zero-Order Hankel Transform Pairs
f (x)
fˆ(σ) ≡ H0 (σ) =
1.
δ(x)
x
1
2.
H(a − x) †
3.
a2 − x2 H(a − x)
0
x J0 (σx) f (x) dx
a
J1 (aσ)
σ
4a
2a2
J1 (aσ) − 2 J0 (aσ)
3
σ
σ
2
−3/2
a a + σ2
4.
e−ax
5.
1 −ax
e
x
−3/2
a a2 + σ 2
2
−1/2
a + σ2
1
cos(ax)
x
1
sin(ax)
x
1
(1 − cos ax)
x2
1
J1 (ax)
x
2
−1/2
σ − a2
H(σ − a)
6.
7.
8.
9.
10.
11.
Y0 (ax)
12.
K0 (ax)
13.
14.
15.
†
∞
x2 + a 2
x2 + a 2
−1/2
e−aσ
2
−1/2
a − σ2
H(a − σ)
cosh−1
a
σ
H(a − σ)
1
H(a − σ), a > 0
a
−1
2 2
a − σ2
π
2
−1
a + σ2
√
x2 +b2
e−a
−1/2
sin x
x2
H denotes the Heaviside function.
2
−1/2 −b√a2 +σ2
a + σ2
e
1 −aσ
e
σ
 π

 2 , σ < 1,
1

 arcsin
,
σ
σ > 1.
E
Fractional Derivatives
In the case of the Cauchy problem, the space-time fractional diffusion equation is
derived from the standard diffusion equation by replacing the second-order space
derivative with a Riemann-Liouville derivative of order β, 0 < β ≤ 2, and the firstorder time derivative with Caputo derivative of order α, 0 < α ≤ 1. These derivatives
are defined as follows:
Riemann-Liouville derivative of order β, 0 < β ≤ 2, for a differentiable function
f on the interval [a, b]:
Dβ [f (x)] =
1
d
Γ(1 − β) dx
x
a
f (s)
ds,
(x − s)β
a < x < b, 0 < β ≤ 2.
Caputo derivative of order α, 0 < α ≤ 1, of a differentiable function f on the interval
[0, T ]:




t
1
f (s)
ds, 0 ≤ t ≤ T, 0 < α < 1,
Γ(1 − α) 0 (t − s)α
Dα [f (t)] =


 df (t) , 0 ≤ t ≤ T, α = 1.
dt
The semiderivative corresponds to α = 1/2. The Green’s function (or the fundamental solution) for the Cauchy problem can also be interpreted as a spatial probability
density function evolving in time. In many cases it leads to ill-posed problems which
are solved numerically by a suitable regularization method. Some formulas for fractional derivatives are given below.
D α [f (t)] = Dm D−(m−α) f (t) , m ≥ α is an integer;
D α tλ =
D α [c] =
Γ(λ + 1)
tλ−α , for λ > −1, α > 0;
Γ(λ − α + 1)
c t−α
,
Γ(1 − α)
where c is a constant;
D α ekx = k α ekx ;
D α [sin x] = sin x +
πα
πα
; D α [cos x] = cos x +
;
2
2
338
E. FRACTIONAL DERIVATIVES
πα
;
2
πα
πα
D α ei x = iα ei α = ei (x+i π/2)α) = sin x +
+ i cos x +
;
2
2
p!
Γ(p + 1)
D α [xp ] =
xp−n =
xp−α ;
(p − n)!
Γ(p − α + 1)
∞
∞
Γ(n + 1)
α
n
D
an x =
an
xn−α , by differentiating
Γ(n
−
α
+
1)
n=0
n=0
D α [sin(kx)] = k α sin kx +
πα
;
2
339
Dα [cos(kx)] = k α cos kx +
term-by-term;
∞
∞ x
x
, using ex =
;
Γ(n − α + 1)
n=0 n!
n=0
∞
∞
i nx
cn e
=
cn (i n)α ei nx , by fractionally differentiating
D α [ex ] =
Dα
n−α
n=−∞
n
n=−∞
term-by-term;
x
1
f (s)(x − s)n−1 ds;
(n − 1)! 0
x
1
f (s) ds
D α [f (x)] =
.
Γ(−α) 0 (x − s)α+1
D −n [f (x)] =
For integration from b to x, we have
x
1
f (s) ds
α
D
[f
(x)]
=
;
b x
Γ(−α) b (x − s)α+1
α
b Dx
[(x − c)p ] =
−1
b Dx
Γ(p + 1)
(x − c)p−α ;
Γ(p − α + 1)
x
[eax ] =
eax dx =
b
1 ax
1
e , when eab = 0, or when ab = −∞;
a
b
so if a is positive, then b = −∞. This kind of integral with a lower limit of
−∞ is known as the Weyl fractional derivative, and −∞ Dxα [eax ] = aα eax ;
x
xp+1
bp+1
−1
p
D
[x
]
=
xp dx =
−
; so if b = 0, we get
b x
p+1 p+1
b
α
0 Dx
[xp ] =
Γ(p + 1)
xp−α ;
Γ(p − α + 1)
Γ(p + 1)
(x − c)p−α ; and
Γ(p − α + 1)
x
1
f (s) ds
α
.
b Dx [f (x)] =
Γ(−α) b (x − s)α+1
α
b Dx
[(x − c)p ] =
We write 0 Dxα or 0 Dtα simply as Dα . Note that D(−α) D(−β) = D −(α+β) always,
340
E. FRACTIONAL DERIVATIVES
but Dα Dβ = D α+β sometimes. The functional derivative of an elementary function
is usually a higher transcendental function. For more details on fractional derivatives,
see Oldham and Spanier [1974], Miller and Ross [1993], and Debnath [2005].
F
Systems of Ordinary Differential Equations
F.1. Systems of First-Order Equations. We will discuss the one-sided
Green’s function for a system of n linear first-order ordinary differential equations
subject to n initial conditions. Consider a system of first-order homogeneous ordinary
differential equations
n
ui =
aij (x)uj , i = 1, . . . , n,
(F.1)
j=1
where aij (x) are defined and continuous on an interval I : [a, b]. The following result
holds:
Theorem F.1. There exist n solutions
u11 (x) · · · un1 (x)
u12 (x) · · · un2 (x)
···
···
···
u1n (x) · · · unn (x)
of the system (F.1) such that
u11 (x)
u (x)
W (x) = 12
···
u1n (x)
un1 (x) un2 (x) = 0 on a ≤ x ≤ b,
··· unn (x)
n
and W (x) satisfies the differential equation W −
aij (x) W = 0.
···
···
···
···
(F.2)
j=1
A proof can be found in Murray and Miller [1954:118].
Now, consider the nonhomogeneous system of equations
n
ui =
aij (x)uj +bi , i = 1, . . . , n.
(F.3)
j=1
An extension of the method of variation of parameters will enable us to solve this system and also lead us to the one-sided Green’s function matrix. Suppose aij (x), bi (x) ∈
C(I : [a, b]), and to solve Eq (F.3) subject to the initial conditions ui (x0 ) = ui,0 for
341
342
F. SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS
x0 ∈ I, we further assume that we have functions φi (x) such that φi (x0 ) = 0 and
φi =
n
aij (x)uj +bi ,
i = 1, . . . , n.
(F.4)
j=1
Thus, if the set {ujk (x)} is the primitive of solutions of the homogeneous equations
corresponding to Eqs (F.3) with the initial conditions ujk (x0 ) = δjk , then
ψi =
n
uk,0 uik (x) + φi ,
i = 1, . . . , n,
k=1
where uk,0 = uk (x0 ), is a solution of Eq (F.3) with the prescribed initial conditions.
Hence, if we have a solution φi (x) of the nonhomogeneous system (called the principal integral ) with the initial conditions φi (x0 ) = 0, i = 1, . . . , n, then the solution
of Eq (F.3) is φi (x) plus a linear combination of the primitives. This solution is
unique by Picard’s existence and uniqueness theorem.1
Theorem F.2. The solution φi (x) of Eq (F.3), where aij (x), bi (x) ∈ C(I :
[a, b]), with the initial conditions φi (x0 ) = 0, x0 ∈ I, can be expressed in
terms of quadratures involving the primitives of the corresponding homogeneous equations.
Proof. Let uij (x) be the primitives of the homogeneous equations (F.1). Let
hi (x) denote the n continuously differentiable functions of x to be determined later.
Consider the functions
n
φi (x) =
hm (x)uim (x), i = 1, . . . , n,
(F.5)
m=1
and substitute in Eq (F.3):
n
hm uim +
m=1
or
n
m=1
n
hm uim =
m=1
n
j=1
aij
n
hm ujm + bi ,
m=1
n
n
hm uim −
aij ujm +
hm uim = bi ,
j=1
m=1
Since uim are solutions of Eq (F.1), the terms within the parentheses are zero. Now,
we determine hm such that
n
hm uim = bi .
(F.6)
m=1
1
The theorem states: If f (u, x) is a real-valued function of two real variables u, x, which
are defined and continuous on an open domain D of the plane R2 and satisfy a Lipschitz
condition in D, then for every point (u0 , x0 ) ∈ D we can find a δ > 0 and a function
u(x) which has a continuous first derivative in the neighborhood |x − x0 | ≤ δ such that
du
= f (u(x), x) and u(x0 ) = u0 in this neighborhood, and u(x) is unique.
dx
F. SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS
343
Since uij are the primitives and their Wronskian is defined by (F.2), we can regard Eq
(F.6) as a system of linear algebraic equations on the unknowns hm with determinant
W (x) = 0 on I. Hence, by Cramer’s rule
u21 (x)
···
un1 (x) u11 (x)
···
···
···
···
u1,m−1 (x) u2,m−1 (x) · · · un,m−1 (x) 1 hm (x) =
b1
b2
···
bn
W (x) u1,m+1 (x) u2,m+1 (x) · · · un,m+1 (x) ···
···
···
···
u1n (x)
u2n (x)
···
unn (x)
n
1 =
bi (x)Ykj (x),
W (x) j=1
where Ykj is the cofactor2 of ujk in W (x). Thus,
hk (x) =
n
x
j=1
x0
1
bj (s)Ykj (s) ds + Ck ,
W (x)
where Ck are constants of integration. From Eq (F.5)
n
1
bj (s)Ykj (s) ds +
Cm uim (x)
W (x)
m=1
m=1
j=1 x0
n
n
n
x ujm (x)Ymj (s)
=
bj (s) ds +
Cm uim (x).
W (s)
m=1
j=1 x0 m=1
φi (x) =
n
hi (x)
n
x
The constants Cm can be determined by the initial conditions; thus, φi (x0 ) = 0, i =
n
1, . . . , n, we get 0 = φi (x0 ) = 0+
Cm uim (x0 ), i = 1, . . . , n, which is a system
m=1
of n linear algebraic equations on Cm with determinant W (x0 ) = 0. Hence Cm = 0
for m = 1, . . . , n, and the solution becomes
n
n
x ujm (x)Ymj (s)
φi (x) =
bj (s) ds, i = 1, . . . , n.
(F.7)
W (s)
j=1 x0 m=1
n
1
uim (x)Ymj (s), are called one-sided Green’s
W (s) m=1
functions. Note that Eq (F.7) can be written in matrix form as
The functions gij (x, s) =
x
φ (x) =
g(x, s) + b(s),
(F.8)
x0
2 Let A = [a ] be an n × n matrix. If the ith row and jth column of A are deleted,
ij
the remaining (n − 1) rows and (n − 1) columns form another matrix Mij , and det[M ] is
i+j det[M ], such that
called
n the minor of aij . The cofactor
n of [aij ] is defined as cij = (−1)
a
c
=
0
for
j
=
k,
and
a
c
=
0
for
i
=
k.
These
relations can also be
ij
ij
ik
kj
i=1
i=1 n
written as n
i=1 aij cik = det[A] δjk and
i=1 aij cik = det[A] δik .
344
F. SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS
where


φ1 (x) 


 φ2 (x) 

φ (x) =
,
..



 . 

φn (x)

G11 (x, s)
 G21 (x, s)
g(x, s) = 
···
Gn1 (x, s)


b1 (x) 


 b2 (x) 

b(x) =
,
..



 . 

bn (x)
···
···
···
···

G1n (x, s)
G2n (x, s) 
.
···
Gnn (x, s)
The matrix g is the one-sided Green’s function, which can also be written as
u21 (x)
u11 (x)
u12 (s)
u11 (s)
···
···
j−1 (−1)
gij (x, s) =
uj−1,1 (s) uj−1,2 (s)
W (s) uj+1,1 (s) uj+1,2 (s)
···
···
un1 (s)
un2 (s)
···
···
···
···
···
···
···
un1 (x) u1n (s) ···
uj−1,n (s) .
uj+1,n (s) ···
unn (s)
(F.9)
Formula (F.9) is similar to (F.8), although each is useful in its respective situation.
Example F.1. Consider the system u1 = u2 , u2 = 4u1 − 3u2 , which has the
general solution: u1 = c1 e−4x + c2 ex , u2 = −4c1 e−4x + c2 ex . The Wronskian is
u11 u21 e−4x −4e−4x = 5e−3x = 0.
W (x) = =
u12 u22 ex
ex Then, since a0 (x) = 1, by (F.9) the one-sided Green’s function is given by
−1 e−4x ex 1 x −s
g22 (x.s) =
e e − e4x e4s .
−4s
s=
−3s
e
s
5e
5
Tho check this result, note that the given system of equations reduces to the second
order equation u + 3u − 4u = 0, where by setting u = u1 , u = u2 we get the given
system. Since for this second-order equation a0 (x) = 1 and W (t) = 5e−3x , the
one-sided Green’s function g(x, s) obtained by using (3.8) is the same as the above
g22 (x, s).
Note that Green’s functions defined by (F.9) convert a system of first-order linear
ordinary differential equations together with the initial conditions into Volterra integral equations of the first kind. The classical Green’s functions which are developed
below for the Sturm-Liouville systems with boundary conditions lead to Fredholm
integral equations of the first kind. However, these developments into integral equations are outside the scope of this book; the interested readers should consult literature
on integral equations, some of which is available in the Bibliography.
Bibliography
Abramowitz, M. and I. A. Stegun (Eds.). 1965. Handbook of Mathematical
Functions. New York: Dover.
Akhiezer, N. I. 1990. Elements of the Theory of Elliptic Functions. AMS
Translation of Mathematical Monographs, Vol. 79. Providence, RI: American
Mathematical Society.
Arfken, G. 1985. Mathematical Methods in Physics, 3rd edition. Orlando, FL:
Academic Press.
Bateman, H. 1944. Partial Differential Equations. New York: Dover.
. 1959. Partial Differential Equations of Mathematical Physics. Cambridge: Cambridge University Press.
Borre, K. 2001. Plane Networks and Their Applications. Boston: Birkhäuser.
Boyce, W. E. and R. C. DiPrima. 1992. Elementary Differential Equations, 5th
ed. New York: Wiley.
Brebbia, C. A. and J. Dominguez. 1992 Boundary Elements. Southampton: Computational Mechanics Publications, and New York: McGraw-Hill.
Brauer, F. and J. A. Nohel. 1989. Qualitative Theory of Ordinary Differential
Equations: An Introduction. New York: Dover.
Brychkov, Y. A. 2008. Handbook of Special Functions: Derivatives, Integrals,
Series and Other Formulas. Boca Raton, FL: CRC Press.
Carrier, G., M. Krook and C. E. Pearson. 1966. Functions of a Complex Variable:
Theory and Technique. New York: McGraw-Hill.
Churchill, R. V. 1972. Operational Methods, 3rd edition. New York: McGraw-Hill.
Coddington, Earl A. 1989. An Introduction to Ordinary Differential Equations.
New York: Dover.
and N. Levinson. 1955. Theory Ordinary Differential Equations. New
York: McGraw-Hill.
Costabel, M. 2004. Time-dependent problems with the boundary integral equation
method. In Encyclopedia of Computational Mechanics, eds. E. Stein, R. Borst
and T. J. R. Hughes. New York: Wiley.
Courant, R. and D. Hilbert. 1968. Methods of Mathematical Physics, Vol. 1, 2.
New York: Interscience.
Davies, B. 1978. Integral Transforms and Their Applications. New York:
Springer-Verlag.
345
346
BIBLIOGRAPHY
Debnath, L. 2005. Nonlinear Partial Differential Equations for Scientists and
Engineers, 2nd edition. Boston: Birkhäuser.
Dirac, P. A. M. 1926-27. The physical interpretation of the quantum dynamics. Proc.
Royal Society, A, London 113: 621–41.
1947. The Principles of Quantum Mechanics. Oxford: Clarendon Press.
Duffy, D. G. 1994. Transform Methods for Solving Partial Differential Equations. Boca Raton, FL: CRC Press.
Duru, H. and H. Kleinert. 1979. Solution of the path integral for the H-atom. Physics
Letters B 84: 185–88.
Epstein, B. 1962. Partial Differential Equations. New York: McGraw-Hill.
Erdélyi, A., W. Magnus, F. Oberhettinger and F. G. Tricomi. 1954. Tables of
Integral Transforms, Vol. 1. New York: McGraw-Hill.
Farlow, S. J. 1982. Partial Differential Equations for Scientists and Engineers.
New York: Wiley.
Fox, C. 1961. The G and H-functions as symmetrical Fourier kernels. Trans. Amer.
Soc. 98: 395–429.
Friedlander, F. G. 1982. An Introduction to the Theory of Distributions. Cambridge: Cambridge Univ. Press.
Gaier, D. 1964. Konstruktive Methoden der konformen Abbildung. Berlin:
Springer-Verlag.
Gladshteyn, I. S. and I. M. Ryzhik. 2007. Tables of Integrals, Series and Products,
eds. Alan Jeffrey and Daniel Zwillinger, 7th edition. New York: Academic Press.
Goldsmith, Paul F. 1998. Quasioptical Systems. Piscataway, NJ: IEEE Press.
Griffiths, David, J. 1995. Introduction to Quantum Mechanics. Upper Saddle
River, NJ: Prentice Hall.
Groetsch, C. W. 1984. The Theory of Tikhonov Regularization for Fredholm
Integral Equations of the First Kind. Boston: Pitman Advanced Publishing
Program.
Hecht, K. T. 2000. Quantum Mechanics. New York: Springer-Verlag.
Heinhold, J. and R. Albrecht. 1954. Zur Praxis der konformen abbildung. Rendiconti Cirulo Math. Palermo 3: 130–48.
Henrici, P. 1974-1986. Applied and Computational Complex Analysis, Vols.
1–3. New York: John Wiley.
Kanwal, R. 1983. Generalized Functions: Theory and Technique. New York:
Academic Press.
Kantorovich, L. V. and V. I. Krylov. 1936. Methods for the Approximate Solution
of Partial Differential Equations (Russian). Leningrad-Moscow: Gostekhizdat.
and V. I. Krylov. 1958. Approximate Methods for Higher Analysis.
New York: Interscience.
Kemppainen, J. and K. Ruotsalainen. 2008. Boundary integral solution of the timedependent fractional diffusion equation. In Integral Methods in Science and
Engineering, ed. C. Constanda and S. Potapenko, 141-48. Boston: Birhäuser.
BIBLIOGRAPHY
347
Koebe, P. 1915. Abhandlungen zur Theorie der konformen Abbildung, I. Journal
reine und angew. Math. 145: 177–223.
Korn, G. A. 1961. Mathematical Handbook for Scientists and Engineers: Definitions, Theorems, and Formulas. New York: McGraw-Hill.
Kulshrestha, P. K. 1973. Generalized convexity in conformal mappings. J. Math.
Analysis and Appls. 43:441-49.
. 1974. Coefficient problem for alpha-convex univelaent functions. Archive
Ratnl. Mech. Anal. 54:204-11.
. 1976. Coefficient problem for a class of Mocanu-Bazilevic functions.
Annales Polon. Math. 31: 291-299.
and P. Puri. 1969. An exact solution of hydromagnetic rotating flow. In
Developments in Mechanics 5: 265-71. Proc. 11th Midwestern Conf. Ames:
Iowa State Univ. Press.
and P. Puri. 1983. Wave structure in oscillatory Couette flow of a dusty gas.
Acta Mechanica 46: 127–35.
Kythe, P. K. 1995. An Introduction to Boundary Element Methods. Boca Raton,
FL: CRC Press.
. 1996. Fundamental Solutions for Differential Operators and Applications. Boston: Birkhäuser.
, P. Puri and M. R. Schäferkotter. 1997. Partial Differential Equations
and Mathematica Boca Raton, FL: CRC Press, ISBN 1-58488-314-6.
. 1998. Computational Conformal Mapping. Boston: Birkhäuser.
and P. Puri. 2002. Computational Methods for Linear Integral Equations. Boston: Birkhäuser.
, P. Puri and M. R. Schäferkotter. 2003. Partial Differential Equations
and Boundary Value Problems with Mathematica, 2nd edition. Boca Raton,
FL: Chapman & Hall/CRC.
and M. R. Schäferkotter. 2005. Handbook of Computational Methods
for Integration. Boca Raton, FL: Chapman & Hall/CRC.
Lanczos, C. 1961. Linear Differential Operators. New York: Van Nostrand.
MacRobert, T. M. 1967. Spherical Harmonics: An Elementary Treatise on
Harmonic Functions, with Applications. New York: Pergamon Press.
Miller, K. S. and B. Ross. 1993. An Introduction to the Fractional Calculus and
Fractional Differential Equations. New York: John Wiley.
Murray, F. J. and K. S. Miller. 1954. Existence Theorems for Ordinary Differential Equations. New York: New York University Press.
Oberhettinger, F. 1990. Tables of Fourier Transforms and Fourier Transforms
of Distributions. Berlin: Springer-Verlag.
Oldham, K. B. and J. Spanier. 1974. The Fractional Calculus. New York: Academic Press.
Pauli, W. 1926. Über das Wasserstoffspektrum von Standpunkt der neuen Quantenmechanik. Zeitschrift für Physik 36: 336–363.
Pierpont, J. 1959. Functions of a Complex Variable. New York: Dover.
Porter, D. and D. S. G. Sterling. 1993. Integral Equations.Cambridge: Cambridge
348
BIBLIOGRAPHY
University Press.
Pontriagin, L. S. 1962. Ordinary Differential Equations. Reading, MA: AddisonWesley.
Prudnikov, A. P., Y. A. Brychkov and O. Marichev. 1990. Integrals and Series,
Vol. 3, Special Functions. Amsterdam: Overseas Publishers Association.
Puri, P. and P. K. Kythe. 1988. Wave structure in unsteady flows past a flat plate in
a rotating medium. In Proc. SECTAM XIV, Developments in Theor. Appl.
Mech. 14: 207–213.
and P. K. Kythe. 1988. Some inverse Laplace transforms of exponential
form. ZAMP 39: 150–156; 954.
Roach, G. F. 1982. Green’s Functions, 2nd edition. Cambridge: Cambridge University Press.
Ross, S. L. 1964. Differential Equations. Waltham: Bleisdell.
Ruehr, Otto G. 2002. Analytical-Numerical treatment of the one-phase Stephan
problem with constant applied heat flux. In Integral Methods in Science and
Engineering, ed. P. Schiavone, C. Constanda and A. Mioduchowski, 215–220.
Boston: Birkhäuser.
Sagan, Hans. 1989. Boundary and Eigenvalue Problems in Mathematical
Physics. New York: Dover.
Sneddon, I. N. 1957. Partial Differential Equations. New York: McGraw-Hill.
. 1978. Fourier Transforms and Their Applications. Berlin: SpringerVerlag.
Stakgold, I. 1968. Boundary Value Problems of Mathematical Physics, Vol. II.
New York: Macmillan.
. 1979. Green’s Functions and Boundary Value Problems. New York:
Wiley.
Unsöld, A. 1927. Beiträge zur Quantummechanik der Atome. Annalen der Physik
387: 355–393.
von Koppenfels, W. and F. Stallmann. 1959. Praxis der konformen Abbildung.
Berlin: Springer-Verlag.
Vladimirov, V. S. 1984. Equations of Mathematical Physics (English Translation).
Moscow: Mir Publishers.
Wayland, H. 1970. Complex Variables Applied in Science and Engineering.
New York: Van Nostrand Reinhold.
Weinberger, H. F. 1965. A First Course in Partial Differential Equations. New
York: John Wiley.
Whittaker, E. T. and G. N. Watson. 1962. A Course of Modern Analysis. 4th
edition. Cambridge: Cambridge University Press.
Download