INTRODUCTION TO THE SPECIAL FUNCTIONS OF MATHEMATICAL PHYSICS
with applications to the physical and applied sciences
John Michael Finn
April 13, 2005
CONTENTS
Contents iii
Preface xi
Dedication xvii
1. Infinite Series 1 1.1Convergence 1 1.2A cautionary tale 2 1.3Geometric series 6 Proof by mathematical induction 6 1.4Definition of an infinite series 7 Convergence of the chessboard problem 8 Distance traveled by A bouncing ball 9 1.5The remainder of a series 11 1.6Comments about series 12 1.7The Formal definition of convergence 13 1.8Alternating series 13 Alternating Harmonic Series 14 1.9Absolute Convergence 16 Distributive Law for scalar multiplication 18 Scalar multiplication 18 Addition of series 18 1.10Tests for convergence 19 iv Contents Preliminary test 19 Comparison tests 19 The Ratio Test 20 The Integral Test 20 1.11Radius of convergence 21 Evaluation techniques 23 1.12Expansion of functions in power series 23 The binomial expansion 24 Repeated Products 25 1.13More properties of power series 26 1.14Numerical techniques 27 1.15Series solutions of differential equations 28 A simple first order linear differential equation 29 A simple second order linear differential equation 30 1.16Generalized power series 33 Fuchs's conditions 34
2. Analytic continuation 37 2.1The Fundamental Theorem of algebra 37 Conjugate pairs or roots. 38 Transcendental functions 38 2.2The Quadratic Formula 38 Definition of the square root 39 Definition of the square root of -1 40 The geometric interpretation of multiplication 41 2.3The complex plane 42 2.4Polar coordinates 44 Contents v 2.5Properties of complex numbers 45
2.6The roots of z1/n 47 2.7Complex infinite series 49 2.8Derivatives of complex functions 50 2.9The exponential function 53 2.10The natural logarithm 54 2.11The power function 55 2.12The under-damped harmonic oscillator 55 2.13Trigonometric and hyperbolic functions 58 2.14The hyperbolic functions 59 2.15The trigonometric functions 60 2.16Inverse trigonometric and hyperbolic functions 61 2.17The Cauchy Riemann conditions 63 2.18Solution to Laplace equation in two dimensions 64
3. Gamma and Beta Functions 67 3.1The Gamma function 67 Extension of the Factorial function 68 Gamma Functions for negative values of p 70 Evaluation of definite integrals 72 3.2The Beta Function 74 3.3The Error Function 76 3.4Asymptotic Series 78 Sterling’s formula 81
4. Elliptic Integrals 83 4.1Elliptic integral of the second kind 84 4.2Elliptic Integral of the first kind 88 vi Contents 4.3Jacobi Elliptic functions 92 4.4Elliptic integral of the third kind 96
5. Fourier Series 99 5.1Plucking a string 99 5.2The solution to a simple eigenvalue equation 100 Orthogonality 101 5.3Definition of Fourier series 103 Completeness of the series 104 Sine and cosine series 104 Complex form of Fourier series 105 5.4Other intervals 106 5.5Examples 106 The Full wave Rectifier 106 The Square wave 110 Gibbs Phenomena 112 Non-symmetric intervals and period doubling 114 5.6Integration and differentiation 119 Differentiation 119 Integration 120 5.7Parseval’s Theorem 123 Generalized Parseval’s Theorem 125 5.8Solutions to infinite series 125
6. Orthogonal function spaces 127 6.1Separation of variables 127 6.2Laplace’s equation in polar coordinates 127 6.3Helmholtz’s equation 130 Contents vii 6.4Sturm-Liouville theory 133 Linear self-adjoint differential operators 135 Orthogonality 137 Completeness of the function basis 139 Comparison to Fourier Series 139 Convergence of a Sturm-Liouville series 141 Vector space representation 142
7. Spherical Harmonics 145 7.1Legendre polynomials 146 Series expansion 148 Orthogonality and Normalization 151 A second solution 154 7.2Rodriquez’s formula 156 Leibniz’s rule for differentiating products 156 7.3Generating function 159 7.4Recursion relations 162 7.5Associated Legendre Polynomials 164 Normalization of Associated Legendre polynomials 168 Parity of the Associated Legendre polynomials 168 Recursion relations 169 7.6Spherical Harmonics 169 7.7Laplace equation in spherical coordinates 172
8. Bessel functions 175 8.1Series solution of Bessel’s equation 175 Neumann or Weber functions 178 8.2Cylindrical Bessel functions 180 viii Contents Hankel functions 181 Zeroes of the Bessel functions 182 Orthogonality of Bessel functions 183 Orthogonal series of Bessel functions 183 Generating function 186 Recursion relations 186 8.3Modified Bessel functions 188 Modified Bessel functions of the second kind 190 Recursion formulas for modified Bessel functions 191 8.4Solutions to other differential equations 192 8.5Spherical Bessel functions 193 Definitions 194 Recursion relations 198 Orthogonal series of spherical Bessel functions 199
9. Laplace equation 205 9.1Origin of Laplace equation 205 9.2Laplace equation in Cartesian coordinates 207 Solving for the coefficients 210 9.3Laplace equation in polar coordinates 214 9.4Application to steady state temperature distribution 215 9.5The spherical capacitor, revisited 217 Charge distribution on a conducting surface 219 9.6Laplace equation with cylindrical boundary conditions 221 Solution for a clyindrical capacitor 225
10. Time dependent differential equations 227 10.1Classification of partial differential equations 227 Contents ix 10.2Diffusion equation 232 10.3Wave equation 236 Pressure waves: standing waves in a pipe 239 The struck string 240 The normal modes of a vibrating drum head 242 10.4Schrödinger equation 245 10.5Examples with spherical boundary conditions 246 Quantum mechanics in a spherical bag 246 Heat flow in a sphere 247 10.6Examples with cylindrical boundary conditions 250 Normal modes in a cylindrical cavity 250 Temperature distribution in a cylinder 250
11. Green’s functions and propagators 252 11.1The driven oscillator 253 11.2Frequency domain analysis 257 11.3Green’s function solution to Possion’s equation 259 11.4Multipole expansion of a charge distribution 260 11.5Method of images 262 Solution for a infinite grounded plane 263 Induced charge distribution on a grounded plane 265 Green’s function for a conducting sphere 266 11.6Green’s function solution to the Yakawa interaction 268
PREFACE
This text is based on a one semester advanced undergraduate course that I have taught at the College of William and Mary. In the spring semester of 2005, I decided to collect my notes and to present them in a more formal manner. The course covers se- lected topics on mathematical methods in the physical sciences and is cross listed at the senior level in the physics and applied sciences departments. The intended audience is junior and se- nior science majors intending to continue their studies in the pure and applied sciences at the graduate level. The course, as taught at the College, is hugely successful. The most frequent comment has been that students wished they had been intro- duced to this material earlier in their studies.
Any course on mathematical methods necessarily involves a choice from a venue of topics that could be covered. The empha- sis on this course is to introduce students the special functions of mathematical physics with emphasis on those techniques that would be most useful in preparing a student to enter a program of graduate studies in the sciences or the engineering discip- lines. The students that I have taught at the College are the gen- erally the best in their respective programs and have a solid foundation in basic methods. Their mathematical preparation xii Preface includes, at a minimum, courses in ordinary differential equa- tions, linear algebra, and multivariable calculus. The least expe- rienced junior level students have taken at least two semesters of Lagrangian mechanics, a semester of quantum mechanics, and are enrolled in a course in electrodynamics, concurrently. The senior level students have completed most of their required course work and are well into their senior research projects. This allows me to exclude a number of preliminary subjects, and to concentrate on those topics that I think would be most helpful. My classroom approach is highly interactive, with students pre- senting several in-class presentations over the course of the semester. In-class discussion is often lively and prolonged. It is a pleasure to be teaching students that are genuinely interested and engaged. I spend significant time in discussing the limita- tion as well as the applicability of mathematical methods, draw- ing from my own experience as a research scientist in particle and nuclear physics. When I discuss computational algorithms, I try to do so .from a programming language-neutral point of view.
The course begins with review of infinite series and complex analysis, then covers Gamma and Elliptic functions in some de- tail, before turning to the main theme of the course: the unified study of the most ubiquitous scalar partial differential equations of physics, namely the wave, diffusion, Laplace, Poisson, and Schrödinger equations. I show how the same mathematical me- thods apply to a variety of physical phenomena, giving the stu- Preface xiii dents a global overview of the commonality of language and techniques used in various subfields of study. As an interme- diate step, Strum-Liouville theory is used to study the most common orthogonal functions needed to separate variables in Cartesian, cylindrical and spherical coordinate systems. Boun- dary valued problems are then studied in detail, and integral transforms are discussed, including the study of Green functions and propagators.
The level of the presentation is a step below that of Mathemati- cal Methods for Physicists by George B. Arfken and Hans J. Weber, which is a great book at the graduate level, or as a desk- top reference; and a step above that of Mathematical Methods in the Physical Sciences, by Mary L. Boas, whose clear and sim- ple presentation of basic concepts is more accessible to an un- dergraduate audience. I have tried to improve on the rigor of her presentation, drawing on material from Arfken, without over- whelming the students, who are getting their first exposure to much of this material.
Serious students of mathematical physics will find it useful to invest in a good handbook of integrals and tables. My favorite is the classic Handbook of Mathematical Functions, With Formu- las, Graphs, and Mathematical Tables (AMS55), edited by Mil- ton Abramowitz and Irene A. Stegun. This book is in the public domain, and electronic versions are available for downloading on the worldwide web. NIST is in the process of updating this xiv Preface work and plans to make an online version accessible in the near future.
Such handbooks, although useful as references, are no longer the primary means of accessing the special functions of mathe- matical physics. A number of high level programs exist that are better suited for this purpose, including Mathematica, Maple, MATHLAB, and Mathcad. The College has site licenses for sev- eral of these programs, and I let students use their program of choice. These packages each have their strengths and weak- nesses, and I have tried to avoid the temptation of relying too heavily on proprietary technology that might be quickly out- dated. My own pedagogical inclination is to have students work out problems from first principles and to only use these pro- grams to confirm their results and/or to assist in the presenta- tion and visualization of data. I want to know what my students know, not what some computer algorithm spits out for them.
The more computer savvy students might want to consider using a high-level programming language, coupled with good numeric and plotting libraries, to achieve the same results. For example, the Ch scripting interpreter, from SoftIntegration, Inc, is availa- ble for most computing platforms including Windows, Linux, Mac OSX, Solaris, and HP-UX. It includes high level C99 scien- tific math libraries and a decent plot package. It is a free down- load for academic purposes. In my own work, I find the C# pro- gramming language, within the Microsoft Visual Studio pro- gramming environment, to be suitable for larger web-oriented Preface xv projects. The C# language is an international EMCA supported specification. The .Net framework has been ported to other plat- forms and is available under an open source license from the MONO project.
These notes are intended to be used in a classroom, or other academic settings, either as a standalone text or as supplemen- tary material. I would appreciate feedback on ways this text can be improved. I am deeply appreciative of the students who as- sisted in this effort and to whom this text is dedicated.
Williamsburg, Virginia John Michael Finn April, 2005 Professor of Physics [email protected]
DEDICATION
For my students at The College of William and Mary in Virginia
1. Infinite Series
The universe simply is. Existence is not required to explain itself. This is a task that mankind has chosen for himself, and the reason that he invented mathematics.
1.1 Convergence
The ancient Greeks were fascinated by the concept of infinity. They were aware that there was something transcendental beyond the realm of rational numbers and the limits of finite al- gebraic calculation, even if they did not fully comprehend how to deal with it. Some of the most famous paradoxes of antiquity, at- tributed to Zeno, wrestle with the question of convergence. If a process takes an infinite number of steps to calculate, does that necessarily imply that it takes an infinite amount of time? One such paradox purportedly demonstrated that motion was im- possible, a clear absurdity. Convergence was a concept that mankind had to master before he was ready for Newton and his calculus.
Physicists tend to take a cavalier attitude to convergence and limits in general. To some extent, they can afford to. Physical particles are different than mathematical points. They have a 2 Infinite Series property, called inertia, which limits their response to external force. In the context of special relativity, even their velocity re- mains finite. Therefore, physical trajectories are necessarily well-behaved, single-valued, continuous and differentiable func- tions of time from the moment of their creation to the moment of their annihilation. Mathematicians should be so fortunate. Nevertheless, physicists, applied scientists, and engineering pro- fessionals cannot afford to be too cavalier in their attitude. Un- like the young, the innocent, and the unlucky, they need to be aware of the pitfalls that can befall them. Mathematics is not re- ality, but only a tool that we use to image reality. One needs to be aware of its limitations and unspoken assumptions.
Infinite series and the theory of convergence are fundamental to the calculus. They are taught as an introduction to most intro- ductory analysis courses. Those who stayed awake in lecture may even remember the proofs—Therefore, this chapter is in- tended as a review of things previously learnt, but perhaps for- gotten, or somehow neglected. We begin with a story.
1.2 A cautionary tale
The king of Persia had an astronomer that he wished to honor, for just cause. Calling him into his presence, the king said that he could ask whatever he willed, and, if it were within his power to grant it, even if it were half his kingdom, he would. Infinite Series 3 To this, the astronomer responded: “O King, I am a humble man, with few needs. See this chessboard before us that we have played on many times, grant me only one gain of gold for the first square, and if it please you, twice that number for the second square, and twice that again for the third square, and so forth, continuing this pattern, until the board is complete. That would be reward enough for me.” The king was pleased at such a modest request, and commanded his money changer to fulfill the astronomer’s wish. Figure 1-1 shows the layout of the chess- board, and gives some inkling of where the calculation may lead.
… 262 263
1 2 22 23 24 25 …
Figure 1-1 Layout of the King’s chessboard. 4 Infinite Series The total number grains of gold is a sequence whose sum is giv- en by S=1+2+22+23+24+…, or more generally
63 S = ∑ 2n . (1.1) n=0
Note that mathematicians like to start counting at zero since x0 =1 is a good way to include a leading constant term in a pow- er series. Many present day computer programs number their arrays starting at zero for the first element as well.
The above is an example of a finite sequence of numbers to be summed, a series of N terms, defined by an , which can be writ- ten as
N −1 = SaNn∑ , (1.2) n=0
th where an denotes the n element in the sum of a series of N
terms, expressed as SN . The algorithm or rule for defining the constants in our chess problem is given by the prescription
== aaa011, and nn+ 2 . (1.3)
Note that a is used to compactly describe the progression. For infinite series, where it is physically impossible to write down every single term, the series must be defined by such a rule, of- ten recursively derived, for constructing the nth term in the se- ries. Infinite Series 5 Most of us are familiar with computers and know that they store data in binary format (see Figure 1-2). A bit set in the nth place represents the number 2n . Our chess problem corresponds to a binary number with the bit pattern of a 64 bit integer have all its bits set, the largest unsigned number that can be stored in 64 bits. Adding one to this number results in all zeroes plus the set- ting of an overflow bit representing the number 264 . Therefore, the answer to our chess problem would require ( 2164 − ) grains of gold. This is a huge number, considering that there are there are only 6.02⋅ 1023 atoms per gram-mole of gold.
11111111 11111111 11111111 11111111 11111111 11111111 11111111 11111111
Figure 1-2 A 64-bit unsigned-integer bit-pattern with all its bits set
I could continue the story to its conclusion, but it is more inter- esting to leave you to speculate as to possible outcomes. Here are some questions to ponder:
• What do you suppose the king did to the astronomer? Was this something to lose one’s head over?
• Most good stories have a point, a moral, or a lesson to be learnt. What can one learn from this story?
• If N goes to infinity does the series converge? If not, why not?
(Hint: Preliminary test: if the terms in the series an do not tend to zero as n →∞, the series diverges) 6 Infinite Series 1.3 Geometric series
The chess board series is an example of a , one where successive terms are multiplied by a constant ratio r. It represents one of the oldest and best known of series. A geometric series can be written in the general form as
N −1 ()=+++=,2 n GaraN 00, (1 rr ...) a 0∑ r. (1.4) 0
= = The initial term is a0 1 and the ratio is r 2 for the chess board problem. The sum of a geometric series can be evaluated giving solution with a closed form:
N −1 ⎛⎞1− r n Gar(,)== a rn a . (1.5) N 00∑ 0⎜⎟− 0 ⎝⎠1 r
Proof by mathematical induction
There are a number of ways that the formula for the sum a geo- metric series (1.5) can be verified. Perhaps the most useful for future applications to other recursive problems is Induction in- volves carrying out the following three logical steps.
• Verify that the expression is true for the initial case. Letting = a0 1 for simplicity, one gets
− 1 ()==⎛⎞1 r Gr1 ⎜⎟1. (1.6) ⎝⎠1− r Infinite Series 7 • Assume that the expression is true for the N th case, i.e., as- sume
N − ()= ⎛⎞r 1 GrN ⎜⎟. (1.7) ⎝⎠r −1
• Prove that it is true for the N +1 case:
⎛⎞N − =+=NNr 1 + GGrNN+1 ⎜⎟ r, ⎝⎠r −1 ⎛⎞NNN−−−+− =+rrrrr111(1)N ⎛⎞ = GrN +1 ⎜⎟⎜⎟ , (1.8) ⎝⎠rr−−11⎝⎠ r − 1 NNNN−+++11 − − − N + 1 ===rrrr1)11⎛⎞⎛⎞ r GN +1 ⎜⎟⎜⎟. rrr−−−111⎝⎠⎝⎠
1.4 Definition of an infinite series
The sum of an infinite series Sr() can be defined as the sum of a
series of N terms SrN () in the limit as the number of terms N goes to infinity. For the geometric series this becomes
= Gr() lim GN () r (1.9) N →∞
or
− ∞ ==⎛⎞1 r Gr() G∞ () r ⎜⎟, (1.10) ⎝⎠1− r
where 8 Infinite Series ⎛ 0if r <1, ∞ →∞⎜ > rfr⎜ 1, (1.11) ⎜ ⎝1++ 1 ...diverges .
Therefore, for a general infinite geometric series,
⎧ ∞ a arn =<0 forr 1, = ⎪∑ 0 − Ga(,)0 r ⎨ n=0 1 r (1.12) ⎪ ⎩undefined for r ≥ 1.
Convergence of the chessboard problem
Let’s calculate how much gold we could obtain if we had a chessboard of infinite size. First, let’s try plugging into the series solution:
N −1 N − ==n =21 SGNN∑ 2 (1,2) , = 21− n 0 (1.13) →=−1 SN 1. N →∞ 12−
This is clearly nonsense. One can not get a negative result by adding a sequence that contains only positive terms. Note, how- ever, that the series converges only if r <1. This leads to our first two major conclusions:
• The sum of a series is only meaningful if it converges.
• A function and the sum of the power series that it represents are mathematically equivalent within, and only within, the radius of convergence of the power series. Infinite Series 9 Distance traveled by A bouncing ball
Here is an interesting variation on one of Zeno’s Paradoxes: If a bouncing ball bounces an infinite number of times before com- ing to rest, does this necessarily imply that it will bounce forev- er? Answers to questions like this led to the development of the formal theory of convergence. The detailed definition of the problem to be solved is presented below.
Discussion Problem: A physics teacher drops a ball from rest
at a height h0 above a level floor. See Figure 1-3. The accelera- tion of gravity g is constant. He neglects air resistance and as- sumes that the collision, which is inelastic, takes negligible time (using the impulse approximation). He finds that the height of each succeeding bounce is reduced by a constant ratio r, so = hrhnn−1
• Calculate the total distance traveled, as a Geometric series.
• Using Newton’s Laws of Motion, calculate the time tn needed
to drop from a height hn .
• Write down a series for the total time for N bounces. Does this series converge? Why or why not? 10 Infinite Series
Figure 1-3 Height (m) vs. time (s) for a bouncing ball
Figure 1-3 shows a plot of the motion of a bouncing ball (h0=10 m, r=2/3, g=9.8 m/s2). The height of each bounce is reduced by a constant ratio. A complication is that the total distance tra- veled is to be calculated from the maximum of the first cycle, re- quiring a correction to the first term in the series.
The series to be evaluated turn out to be geometric series. The motion of the particle for the first cycle is given by
gt 2 yt()=− vt , 0 2 = vgh002, (1.14) Δ= = tvghg002/ 8/, 0
Δ where t0 is the time for the first cycle, and Infinite Series 11 2h Dhhhrhh=−=−=−22n 0 , ∑∑n 00 01− r 0 Δ= =Δ thrgtrn 8/00 , (1.15) ΔΔtt T =−00. 1− r 2
Both series converge, but the time series converges slower than the distance series as illustrated in Figure 1-4 below, since rr>< for r 1.
Figure 1-4 The distance and time traveled by a bouncing ball 2 (h0=10 m, g=9.8 m/s and r=2/3)
The bouncing ball undergoes an infinite number of bounces in a finite time. For a contrary example, an under-damped oscillator undergoes an infinite number of oscillations and requires an in- finite amount of time to come to rest.
1.5 The remainder of a series
An infinite series can be factored into two terms
=+ SSNN R, (1.16) 12 Infinite Series where
N −1 • Sa= is the , which has a finite sum of terms, and Nn∑n=0
• RN is the of the series, which has an infinite number of terms
∞ = RNn∑ a (1.17) nN=
1.6 Comments about series
• SN denotes the partial sum of an infinite series S , the part that is actually calculated. Since its computation involves a finite number of algebraic operations (i.e., it is a finite algo- rithm) computing it poses no conceptual challenge. In other words, the rules of algebra apply, and a program can happily be written to return the result.
• The remainder of a series S , denoted as RN , is an infinite se- ries. This series may, or may not, converge.
• The convergence of RN is the same as the convergence of the series S . The convergence of an infinite series is not affected by the addition or subtraction of a finite number of leading terms. Infinite Series 13 • Computers (and humans too) can only calculate a finite
number of terms, therefore an estimate of RN is needed as a measure of the error in the calculation.
• The most essential component of an infinite series is its re- mainder—the part you don’t calculate. If one can’t estimate or bound the error, the numerical value of the resulting ex- pression is worthless.
Before using a series, one needs to know
• whether the series converges,
• how fast it converges, and
• what reasonable error bound one can place on the remainder
RN .
1.7 The Formal definition of convergence
A series SS=+ R converges if NN (1.18) =− <εε > RSSNNfor all NN ( ).
Definition: A series of terms with alternating terms is an alternating series
Consider the alternating series A, given by 14 Infinite Series =−()n Aa∑ 1 n (1.19) n
An alternating series converges if
→→∞ ann 0, as , and <> (1.20) aann+10 , for all n N
An oscillation of sign in a series can greatly improve its rate of convergence. For a series of reducing terms, it is easy to define a
maximum error in a given approximation. The error in SN is
smaller that the first neglected term
< RNNa . (1.21)
Alternating Harmonic Series
An alternating harmonic series is defined as the series
n ∞ ()−1 111 =−1. + − + (1.22) ∑ + n=0 n 1234
This is a decreasing alternating series which tends to zero and so meets the preliminary test.
Example: Series expansion for the natural logarithm
In a book of math tables one can lookup the series expansion of the natural logarithm, which is
n+1 ∞ −−()x x2 x ln() 1+=xx =−+ (1.23) ∑ + n=0 n 123 Infinite Series 15 Setting x =1, allows one to calculate the sum of an alternating harmonic series in closed form
∞ ()−1 n ln(2)=+= ln() 1 1 = 0.6931471806 . (1.24) ∑ + n=0 n 1
Finding a functional representation of a series is a useful way of expressing its sum in closed form.
What about ln(0) ?
∞ 111 ln() 1−= 1 ==++ 1 ∑ + n=0 n 123 (1.25) ln(0) =−∞ diverges
So, to summarize:
⎛⎞()−1 n S= ⎜⎟ converges ∑⎜⎟n +1 ⎝⎠ but (1.26) ()−1 n Sdiverges= ∑ n +1
(1.27)
The series expansion for ln(1+ x ) can be summarized as
∞ −−()x n+1 ln() 1+=x , ∑ + n=0 n 1 (1.28) for02.<≤ x
This series is only conditionally convergent at its end points, de- pending on the signs of the terms. What is going on here? 16 Infinite Series Let’s rearrange the terms of the series so all the positive terms come first (real numbers are commutative aren’t they?)
SSS→++−, ∞ + 1 S =→∞∑ , = 2n n 0 (1.29) ∞ − 1 S =− →−∞, ∑ + n=0 21n S =∞−∞ is undefined.
The problem is that a series is an algorithm, the first series di- verges so one never gets around to calculating the terms of the second series (remember we are limited to a finite number of calculations, assuming we have only finite computer power available to us)
• Note that infinity is not a real number!
1.9 Absolute Convergence
A series converges absolutely if the sum of the series generated by taking the absolute value of all its terms converges. Let = Sa∑ n , then if the corresponding series of positive terms
′ = Sa∑ n (1.30)
converges, the initial series S is said to be Otherwise if S is convergent, but not absolutely convergent, it is said to be Infinite Series 17 Discussion Problem: Show that a conditional convergent se- ries S can be made to converge to any desired value.
Here is an outline of a possible proof: Separate S into two se- ries, one of which contains only positive terms and the second only negative terms. Since S is convergent, but not absolutely convergent, each of these series is separately divergent. Now borrow from the positive series until the sum is just greater than the desired value (assuming it is positive). Next subtract from the series just enough terms to bring it to just below the desired value. Repeat the process. If one has a series of decreasing terms, tending to zero, the results will oscillate about and even- tually settle down to the desired value.
What is happening here is easy enough to understand: One can always borrow from infinity and still have an infinite number in reserve to draw upon:
∞± =∞ an (1.31)
You can simply mortgage your future to get the desired result.
Some conclusions:
• Conditionally convergent series are dangerous, the commut- ative law of addition does not apply (infinity is not a num- ber).
• Absolutely convergent series always give the same answer no matter how the terms are rearranged. 18 Infinite Series • Absolute convergence is your friend. Don’t settle for any- thing less than this. Absolutely convergent series can be treated as if they represent real numbers in an algebraic ex- pression (they do). They can be added, subtracted, multip- lied and divided with impunity.
Distributive Law for scalar multiplication
A series can be multiplied term by term by a real number r without affecting its convergence
Scalar multiplication
Scalar multiplication of a series by a real number r is given by
= ra∑ nn∑ ra. (1.32)
Addition of series
Two absolutely convergent series can be added to each other term by term; the resulting series converges within the common interval of convergence of the original series.
ab±=() ab ±, ∑∑∑nnnn (1.33) ±= ± cS12ab cS∑ ca 12 nn cb. Infinite Series 19 1.10 Tests for convergence
Here is a summary of a few of the most useful tests for conver- gence. Advanced references will list many more tests.
Preliminary test
A series SN diverges if its terms do not tend to zero as N goes to infinity.
Comparison tests
Comparison tests involve comparing a series to a known series. Comparison can be used to test for convergence or divergence of a series:
• = Given an absolutely convergent series Saan∑ the series = Sbbn∑ converges if
< bann (1.34)
for nN> .
• = Given an absolutely divergent series Saan∑ , the series = > > Sbbn∑ diverges if bann for nN.
• = Given a absolutely convergent series Saan∑ ,the series = Sbbn∑ converges if 20 Infinite Series ba nn++11< (1.35) bann
For nN> . (This test can also be used to test for divergence)
The Ratio Test
This is a variant of the comparison test where the ratio of terms is compared to the geometric series: By comparison to a geome- = tric series, the series Sbbn∑ converges if the ratio of succeed- ing terms decreases as n →∞.
b define r = lim n+1 n→∞ bn ⎧r <1 the series converges, (1.36) ⎪ if⎨ r >1 the series diverges, ⎪ ⎩r = 0 the test fails.
The Integral Test
= The series Sbbn∑ converges if the upper limit of the integral
N N obtained by replacing bbndn→ () converges as N →∞, and ∑ n ∫ N 0 N0 it diverges if the integral diverges. The proof is demonstrated graphically in Figure 1-5, which demonstrates that a sum of pos- itive terms is bounded both above and below by its integral. The Infinite Series 21 integral can be constructed to pass through all the steps in the partial sums at either the beginning or the end of an interval.
6
Sum() x, r 4 , Iplus()xr , Iminus()xr 2
0 0 5 10 15 x
Figure 1-5 The Integral test
1.11 Radius of convergence
The series
∞ = n Sx() ∑ axn (1.37) n=0
defines a an absolutely convergent power series of x within its radius of convergence given by
a rx= Within its radius of convergence, the function and its power se- ries are identical. The power series expansion of Sx() is unique. 22 Infinite Series Example: Definition of the exponential function. The exponential function is defined as that function which is its own derivative de() x = ex(). (1.39) dx Let’s show that the series expansion for the exponential function obeys this rule: ∞ xn= eax∑ n n=0 x ∞ de = n−1 ∑ nan x dx n=1 letting n→+ n′ 1 on the RHS (1.40) ∞∞ nnx−1 =+′ ′ = ∑∑nann x(1) n a′+1 x e nn==10′ ∞∞ ′ +=nn′ ∑∑(1)naxnn′+1 ax . nn′==00 Test for radius of convergence: ()n +1! xn<==lim() 1 n! (1.41) x <∞. (In practice, the useful range for computation is limited, de- pending on the format and storage allocation of a real variable in one’s calculator.) Infinite Series 23 Evaluation techniques • if x ≤1 this series is converges rapidly • if f x >1 use eeeab+ = a b to show 1 ex = , e− x (1.42) ∞ ()−x n e− x = ∑ . n=0 n! Alternating series converge faster, it is easy to estimate the er- ror, and if the algorithm is written properly one shouldn’t get overflow errors. Professional grade mathematical libraries would use sophisticated algorithms to accurately evaluate a function over its entire useful domain. 1.12 Expansion of functions in power series The power series of a function is unique within its radius of cur- vature. Since power series can be differentiated, we can use this property to extract the coefficients of a power series. Let us ex- pand a function of the real variable x about the origin: 24 Infinite Series ∞ ()= n fx∑ axn ; n=0 define d n ()n ()= () ffx0;n dx x=0 then (1.43) ()= fa0,0 (1) ()= fa0,1 (2) ()= fa02,2 ()n ()= fna0!.n This results in the famous Taylor series expansion: () ∞ f n ()0 f ()xx= ∑ n . (1.44) n=0 n! Substituting x →−xa a, we get the generalization to a McLau- ren Series () ∞ fxan ()− f ()xxa=−∑ ()n . (1.45) n=0 n! Taylor’s expansion can be used to generate many well known se- ries, such as the exponential function and the binomial expan- sion. The binomial expansion The binomial expansion is given by Infinite Series 25 ∞ ()1(,)1,+=xBpmxxp ∑ n < (1.46) n=0 where p is any real number. For integer p , the series is a poly- nomial with p +1 terms. The coefficients of this series are known as the binomial coefficients and written as ⎛⎞⎛pp ⎞ p! Bnm(, )==⎜⎟⎜ ⎟ = . (1.47) ⎝⎠⎝npn− ⎠ ()pnn− !! For non-integer p , but integer m , the coefficients can be ex- pressed as the repeated product 1 n−1 B(,nm )=−∏ ( p m ). (1.48) n! m=0 Repeated Products occur often in solutions generated by iteration. A repeated product of terms rm is denoted by the expression N −1 =⋅⋅ ∏ rrrrrmN013 − 1. (1.49) m=0 The is an example of a repeated product n nm!.= ∏ (1.50) m=1 Discussion Problem: Sine and cosine series Euler’s theorem, given by 26 Infinite Series eiiθ =+cosθθ sin , (1.51) is counted among the most elegant of mathematical equations. Derive the series expansion of sin x and cos x using the power series expansion for ex and substituting x → ix , giving ∞ ()−1 n x21n+ sin xx=<∞ (1.52) ∑ + n=0 ()21!n and ∞ ()−1 n x2n cosxx=<∞∑ . (1.53) n=0 ()2!n Then use the definition of the exponential, as the function which is its own derivative ( dex / dx= ex ) to prove dxsin = cosx , dx (1.54) dxcos =−sinx . dx 1.13 More properties of power series • Power Series can be added, subtracted, and multiplied within their common radii of convergence. The result is another power series. • Power Series can also be divided, but one needs to avoid di- vision by zero. This may restrict the radius of convergence of the result. Infinite Series 27 • Power series can be substituted into each other to generate new power series. For example, one can substitute x →−x2 into the exponential function to get the power series of a Gaussian function: ∞ nn2 ()−x2 (1)− x ex=<∞∑ ,.2 (1.55) n=0 n! 1.14 Numerical techniques Example: Calculating the series for ln(1+ x ) using long division d 1 ln(1+=x ) . dx1+ x By long division: −+2 − 1 xx ∞ (1.56) ()1+=−xx 1 ∑ ( )n , n=0 + ∞∞n 1 x −−()x ∴+=ln(1xxdx ) ( − )n = . ∫0 ∑∑+ nn==00n 1 Example: Evaluation of indeterminate forms by series expan- sion: 11− ex− xn⎛⎞⎛∞ x2 ⎞ lim=− 1 =−−++ 1 1x ... , → ⎜⎟⎜∑ ⎟ x 0 xx= n!2 ⎝⎠⎝n 0 ⎠ (1.57) nn 1 ∞∞xx()−−() x ==, ∑∑++ xnnn==001! n 1! 28 Infinite Series 1.15 Series solutions of differential equations The equation N i ⎛⎞d = () ∑ Ai ()xyxSx⎜⎟ () (1.58) i=0 ⎝⎠dx defines an N th order linear differential equation for yx(). If the source term Sx()= 0 , the equation is said to be homogeneous. Otherwise, the equation is said to be inhomogeneous. We will concern ourselves with solutions to homogeneous equations at first. A linear homogenous equation has the general form N i ⎛⎞d = ∑ Axi ()⎜⎟ yx () 0 (1.59) i=0 ⎝⎠dx A N th order differential equation has N linearly-independent so- { } lutions yxi (), and by linearity, the general solution can be written as N = yx()∑ cyii () x (1.60) i=0 If the coefficients Axi () can be expanded in a power series about = the point x 0 , one can attempt to solve for yi ()x in terms of a power series expansion of the form ∞ = n yxiim() ∑ ax (1.61) n=0 Infinite Series 29 Since the function is linear in y , the resulting series expansion will be linear in the coefficients ain , and the self-consistent solu- tion will involve recursion relations between the coefficients of various powers of m . A simple first order linear differential equation Consider the first order differential equation dY() x =−Yx() (1.62) dx We already know the solution, it is given by = − x Yx() Ye0 (1.63) Since the equation is of first order, there is only one linearly in- dependent solution so the above solution is complete. Let’s try expanding this function in a power series ∞∞∞ ===+nn′′−1 () n′ Yx()∑∑∑ axnnn ; Y () x nax n 1 a′+1 x (1.64) nnn===010′ Where the last term involves making the change of va- riables nn=+′ 1 Substituting into equation (1.62) gives ∞∞ ()′ +=−nn′ ∑∑nax1 nn′+1 ax (1.65) nn′==00 But n′ and n are dummy variables and we can compare similar powers of x by setting nn′ = , giving the series solution 30 Infinite Series ∞ ()++n = ∑ ⎣⎦⎡⎤na10.nn+1 ax (1.66) n′=0 The above expression can be true for arbitrary x only if term by term the coefficients vanish: ()++= na10.nn+1 a (1.67) This gives rise to the recursive formula aann−1 aa+ =−; or =− , (1.68) nn1 ()nn+1 () with the solution n n a ()−1 aY=−()1.o = (1.69) n nn!!0 The series solution is given by ∞ ()−1 n ==nx− Yx() Y00∑ x Ye . (1.70) n=0 n! A simple second order linear differential equation Here is a second order differential equation for which we al- ready know the solution: d 2 Y′′() x==− Yx () Yx (). (1.71) dx2 Infinite Series 31 In this case, there are two linearly independent solutions and the general solution can be written as =+() () Yx() a01 cos x a sin x (1.72) However suppose we didn’t know the solution (or at least its se- ries expansion which amounts to the same thing.) How would go about finding two linearly independent solutions? Here symme- try comes to our help. The operator ddx22/ is an even function of x , so the even and odd parts of yx()are separately solutions to equation (1.71). This suggests that we try to find series solu- tions of the form ∞ ==2ns+ Yx()∑ a2ns+ x ; for s 0,1 (1.73) n=0 If s = 0 we get an even function of x ; and if s =1, an odd func- tion of x . Substituting this series into equation (1.71) gives ∞ ′′ =++−()()22ns+− Yx()∑ 2 ns 2 ns 1 a2ns+ x n=2 ∞ =++++()()′′2ns′+ =+ ′ ∑ 2ns 2 2 ns 1 a22ns′++ x (letting nn 2)(1.74) n′=0 ∞ =− =− 2ns+ Yax∑ 2ns+ n=0 Comparing terms of the same power of x gives the recursion formula ()()++ ++ =− 2221ns ns a22ns++ a 2 ns + (1.75) with solution 32 Infinite Series 2n as a + =−()1 (1.76) 2ns ()2!ns+ = If s 0, this gives a cosine series normalized to the value of a0 ; = and if s 1, a sine series normalized to a1 , with the sum yielding the general solution given by equation (1.72). By making the substitution x → mx , we get the differential equa- tion Yx′′()=− mYx2 () (1.77) with solutions =+() () Yxmm() a cos mxb m sin mx (1.78) Some quick comments: • In both of the above examples, we should have used the ratio test to find the radius of convergence of the series solutions, but we have already shown that the exponential, sine and co- sine functions converge for all finite x . • The power series expansion fails if the equation has a singu- larity at the expansion point. Using the Method of Forbenius in the next section, we will see how to extend the series tech- nique to solve equations that have nonessential singularities at their origin. Infinite Series 33 1.16 Generalized power series When a power series solution fails, one can try a generalized power series solution. This is an extension of the power series method to include a leading behavior at the origin that might in- clude a negative or fractional power of the independent variable. A second order linear homogeneous differential equation of the form yfxygxy′′++=() ′ () 0 (1.79) is said to be regular at x = 0 if xfx() and x2 gx()can be written in a power series expansion about x = 0 . That is, the singularity of f ()x is not greater than x−1 and the singularity of gx()is not greater than x−2 at the origin of the expansion. Such a differen- tial equation can be solve in terms of at least one generalized power series of the form ∞ = ns+ yx() ∑ axn (1.80) n=0 ≠ s where a0 0, The leading power x can be a negative or non in- teger power of x . The statement that a0 is the first nonvanishing term of the se- ries, requires that terms aa−−12,, vanish. This constraint defines an quadratic indicial equation for s that can be solved to deter- mine the two roots s1,2 . 34 Infinite Series Fuchs's conditions Given a regular differential equation of the form ′′++=() ′ () yfxygxy0 , with solutions s1,2 for the indicial equa- tion: • − If ss21 is non-integer, s1 and s2 define two linearly inde- pendent generalized power series solutions to the equation. • − If ss21 is integer-valued, the two solutions may or may not be linearly independent. In the second case, the larger of the two constants is used for the first solution y1()x and a second solution can be found by making the substitution =+ y21()xyxxbx ()ln() () (1.81) where bx() is a second generalized power series. Example: Solve by the method of Forbenius: xy22′′++=20. xy ′ xy (1.82) Note that the differential operator is an even function of x . This suggests that we try a solution of the form ∞ = 2ns+ yx()∑ a2n x , (1.83) n=0 where a0 is the first nonvanishing term. Substituting (1.83) into (1.82) gives Infinite Series 35 ∞∞ ()()++−+22ns++ () + ns ∑∑221ns ns ax22nn 22 nsax nn==00 ∞∞ (1.84) =−22ns++ =− 2 ns + ∑∑ax222nn a− x , nn==−01 which yields the recursion formula ()+++=− ⎣⎦⎡⎤2(21)ns ns a222nn a− . (1.85) Letting n = 0 gives the indicial equation () +=−= ⎣⎦⎡⎤ss(1) a02 a− 0 (1.86) or s =−0, 1. (1.87) Equation (1.85) can be rewritten as ()−1 n aa= , (1.88) 20n ()21!ns++ giving the solutions n ∞ ()−1 sin x yx()== a x2n a , (1.89) 00∑ + 0 n=0 ()nx1! ∞ ()−1 n ==21n− cos x yx−10() a∑ x a 0 . (1.90) n=0 ()nx! The general solution is given by sinx cos x yx()=+ A B (1.91) x x 36 Infinite Series In this case, the solutions can be expressed in terms of elemen- tary functions. A solution could have been found by substituting yuxx= ()/ and solving for ux() to obtain ux()=+ A sin x B cos x 2. Analytic continuation By venturing into the complex plane, the geometric sense of multiplying by -1 can be replaced from the operation of reflection, which is discrete, to that of rotation, which is continuous. The result is almost miraculous. 2.1 The Fundamental Theorem of algebra Complex variables were introduced into algebra to solve a fun- damental problem. Given a polynomial function of order N of a real variable x, how do we find its roots (zero crossings)? The equation to be solved can be written as N ==n fxNn()∑ ax 0 (2.1) n=0 The problem may not have a real-valued solution, it may have a unique solution, or it may have up to N distinct solutions, which are called the N roots of fN ()x . The problem can be reduced to the question of whether we can fully factor the function into N linear products . That is, does an algorithm exist that gives us uniquely 38 Analytic continuation =−−−()()()() − fxNN( ) axx012 xx xx ... xx N− 1 N (2.2) =−=() axxNm∏ 0 m=0 This problem does have a solution, but only if we allow for the possibility of complex roots. This is the Fundamental Theorem of Algebra, which asserts that a polynomial of order N of a complex variable z can always be completely factored into its N roots, which are complex in general N N ()==n ( − ) fNnNmzazazz∑ ∏ . (2.3) n=0 m=0 Conjugate pairs or roots. If the coefficients of the power series are all real, the non-real roots always come in complex conjugate pairs. (Note that ()++=+*2 +2 z zzz00() z2Re( zzz 00 ) has real parameters. ) Transcendental functions If the power series is infinite, it has an infinite number of com- plex roots. Such functions are said to be transcendental. 2.2 The Quadratic Formula Let’s apply this to the quadratic formula given by Analytic continuation 39 2 ax++= bx c a()()0 x − x+− x − x = where (2.4) −±bb2 −4 ac x± = 2a Here a, b, c are real coefficients. The roots x± are real if ()bac2 −≥40, and are complex if ()bac2 −<40. For the later case we can rewrite the equation as −±bi4 acb −2 x± = . (2.5) 2a Definition of the square root Consider the plot of the quadratic function y = x2 shown in Fig- ure 2-1. 9 7 yx() 5 0 5 3 − 1 1 1 3 3 11 3 x 40 Analytic continuation Figure 2-1 Plot of the quadratic y = x2 y = x2 is well defined for all x . However, the inverse function 1 x = y 2 , shown in Figure 2-1, has two real solutions for y > 0, one for y = 0, and none for y < 0 . 4 2 0 y^1/2 2 4 113579 y 1 Figure 2-2 Plot of the half- root of y y 2 =± y 1 The square root function is the principal branch of y 2 , which returns the positive branch of the function, i.e. y ≥ 0 for all positive y. Definition of the square root of -1 There are 2 roots of ()−1 1/2 . The roots are labeled as 1 ()−=±1,2 i (2.6) Analytic continuation 41 where i =−1 is considered to be the primary branch of the square root function. Since i is not a real number, it represents a new dimension (degree of freedom). Just as one cannot add meters and seconds, we can not add numbers on the real axis to those along the imaginary axis i .To fully understand the mean- ing of i , one first needs to appreciate the geometric interpreta- tion of multiplication. The geometric interpretation of multiplication The set of real numbers is isomorphic to a one dimensional vec- tor function x , called the number line. Every real number x cor- responds to a vector as labeled x on this line. Multiplication of the vector x by the positive number r , written as f ()xrx= , changes the length of the vector x by the ratio r (see Figure 2-3). Multiplication of x by −r : can be thought of as multiplica- tion by r , followed by multiplication by (1)− : f ()xd=− rx = (1()) − rx (see Figure 2-3) . The latter operation re- flects the orientation of the vector about the origin, an improper operation. By extending the number line into a two-dimensional plane, called the complex plane, a second interpretation of mul- tiplication by (1)− is possible, it can represent a rotation by π radians. This is important because rotations, unlike reflections, can be done continuously. The square (1)−=2 1 is understood as 42 Analytic continuation two rotations by π , which brings us back to starting point. −−()rr = . And i can be written as the phase rotation eiπ =−1. The geometric interpretation of ()±=−i 2 1 is that i is the rota- tion that when doubled produces a rotation by π radians. The possible answers are ei±iπ /2 =± , where ie==−iπ /2 1, is the prin- cipal branch of the square root function. Figure 2-3 Multiplication of a point a on the real number line by a real number( ±2 ) Shown in Figure 2-3 is multiplication of a vector a by(+2) and (- 2). The concept of multiplication on the real number line is one of a scale change plus a possible reflection (multiplication by -1). Reflection is an improper transformation as it is discontinuous. On the complex plane this reflection is replaced by a rotation of 180 . 2.3 The complex plane A complex number c can be thought of as consisting of a vector pair of real numbers (a, b) on a two dimensional plane called the complex plane. A complex vector c can be written as Analytic continuation 43 caib=+. (2.7) Addition of complex numbers is the same as addition of 2- dimensional vectors on the plane. The plane represents all poss- ible pairs of real numbers. Let x be an arbitrary number on the real axis, and y be a arbitrary number along the imaginary axis, then an arbitrary point on the complex plane can be referred to as zxiy=+. (2.8) The complex plane can be quite “real” in that the properties of “real” vectors constrained to a 2-dimensional plane can be quite well represented as complex numbers in many applications. The modulus z of a complex number z is its geometric length z =+xy22. zzz2 = ∗ , where z∗ is the complex conjugate of z (see Figure 2-4). Definition: The complex conjugate of z is the complemen- tary point on the plane given by changing the sign of i . zxiy* =− (2.9) Complex Conjugate Pairs y x 44 Analytic continuation Figure 2-4 Conjugate pairs of vectors in the complex plane x ± iy . 2.4 Polar coordinates Like any 2-dimensional vector pair, (,x y ), the transformation of a complex number into polar coordinates is given by the map- ping xr= cosθ , (2.10) yr= sinθ , using zr==iθ r()cosθθ + i sin . (2.11) Therefore, a complex number can be thought of as having a real magnitude r > 0 and an orientation θ wrt (with respect to) the x axis. Note that the phase angle is cyclic, i.e. periodic, on inter- val 2π eein(2)θπ+ = i θ. (2.12) Example: Using exixix =+cos sin , (2.13) • Derive the series expansion of sin x and cos x using the pow- x er series expansion for e and substituting x → ix . • Use the definition of the exponential, as the function which is its own derivative ( dex / dx= ex ), to prove Analytic continuation 45 dxsin = cosx , dx (2.14) dxcos =−sinx . dx 2.5 Properties of complex numbers Complex numbers form a division algebra, an algebra with a unique inverse for every non-zero element. Complex numbers form commutative, associative groups under both the opera- tions of addition and multiplication. The distributive law also applies. Definition: Addition and subtraction of complex numbers Given =+ =+ zxiyzxiy11 1and 2 2 2, (2.15) then, =±=()() + ± + z312zz xx 12 iyy 12. (2.16) Definition: Multiplication of complex numbers =⋅= + ⋅ + =() − + + z3121zz()() x iyx 1 2 iy 2 xxyy 12121221 ixyxy ( ), (2.17) which in polar notation becomes θ θθ θθ+ ==⋅=i 3 ii12() i() 12 zrerererre33 1 2 12 . (2.18) The geometric interpretation of complex multiplication is that it = represents a change of scale, scaling the length by rrr312, and a 46 Analytic continuation rotation of one number by the phase of the other, with the final θθ+ orientation being the sum of the two phases 12. Definition: Division is defined in terms of the inverse of a complex number =⋅−1 zz12/ zz 12 * −1 = z z * zz (2.19) in polar notation we get − θθ1 1 − ()reii= e r Example: Calculate ()2/(1)++ii. 2 + i let z = 1+ i 22−+ii 41 + then z== z* z ⋅= =5/2 −+ + 11ii 11 Example: Calculate zi2 = 2 First try it by brute force z222=+=−+()xiy2 () x y i()2 xy This leads to two real equations xy22−=0, 22.xy = Substituting yx=1/ into the first equation, we get Analytic continuation 47 2 2 ⎛⎞1 x −=⎜⎟ 0, ⎝⎠x x4 −=10, 1 ⎛⎞4 x ==±Re⎜⎟ 1 1, ⎝⎠ 1 y ==±1. x Therefore z =±()1. +i 1 2 Of course, the easy way is to calculate ()2i directly, the way to do so will be made clear in the next section below 2.6 The roots of z1/n 1 The principal root of 11n = , since 11n = for the identity element. Using polar notation, the nth distinct root of 1 is given by 1 1 ππn 1,n ==()eeim22/ imn with distinct roots for (2.20) mn=−0,1...() 1 . That is, the roots are unit vectors whose phase angles are equally spaced from the identity element 1 in steps of 2/π n as shown in Figure 2-5. This allows us to calculate the nth root of z 48 Analytic continuation let z= reiθ , then 111 θ π (2.21) zznn= ()⋅=1,n rein/2/nie m for m = 0,1,...(n − 1). Definition: The principal root of z1/n is defined as n zre= 1/niφ / n. All other roots are related by uniformly spaced phase rotations of magnitude 2/π n For example: zie3/2==88iπ has the solution ()()ππ++ ππ ze= {2,2iπ /6 eii/6 2 /3 ,2 e /6 4 /3 } , where eiπ /6 =+cos30 i sin 30 . The cube roots of1 +120 deg 0 deg -120 deg 1 Figure 2-5 The n roots of 1n on the unit circle Analytic continuation 49 1 Shown in Figure 2-5 The n roots of 1n on the unit circle are the cube roots of one. The concept of complex multiplication in- volves a phase rotation plus a change of scale. The identity ele- 1 ment 1 is the principle root of 1n . The other n −1 roots are equal- ly spaced vectors on the unit circle. The cube roots of 1 are those phase vectors that, when applied 3 times, rotate themselves into the real number 1. 2.7 Complex infinite series A complex infinite series is the sum a real series and an imagi- nary series. The complex series converges if both real series and the imaginary series separately converge ==() +=+ Sccn∑∑ aibaib nnn ∑ ∑ n. (2.22) A complex series converges absolutely if the series of real num- + bers given by aibnn absolutely converges. ≤ ≤ Proof: Clearly, by the comparison test, acnn and bcnn, so if Sc converges absolutely, then the component series Sa and Sb converge absolutely. n The radius of convergence r of a complex power series ∑czn is given by c rz= By analytic continuation ez is found by substituting z for x in the power series representation of ex ∞ xn ez = ∑ (2.24) n=0 n! The radius of convergence is given by 1 zn<=+=∞limn! lim 1 . (2.25) nn→∞1 →∞ n +1! 2.8 Derivatives of complex functions To understand the meaning of a complex derivative, first let us remind others of the definition of the derivative for a function f ()x of a real variable x . The derivative df/ dx of a real valued function of x exits iff (if and only if) the limit f ()()xfx+−ε df() x/lim dx = (2.26) x→0 ε exists, and the limit is the same whether approaches zero from below or above x . Analytic continuation 51 Example: The derivative of a function is undefined where the slope is undefined. Figure 2-6 shows a plot of the absolute value of x, where the derivative is undefined at the origin x=0. fx() abs value of x x x Figure 2-6 A plot of the absolute value of x. Definition: The derivative of a function of a complex vari- able at a point z0 is given by df() z f ()()zzfz+Δ − = lim 00 (2.27) Δ→z 0 dz= Δ z zz0 Provided that the limit exists and is independent of the path Δ taken by z in approaching z0 . This is much more stringent condition that for the real deriva- tive. There are an infinite number of paths that Δz can take in going to zero. This viewed as important enough so that the exis- tence of the complex derivative is given a special name: Definition: A function of f ()z who’s derivative exists in the vicinity of a point z0 is said to be analytic at z0 Note that z is an analytic function of itself: 52 Analytic continuation ()zzz+Δ − lim= 1 (2.28) Δ→z 0 Δz It follows (using the binomial expansion) that the derivative of z n is also analytic: n n ⎛⎞n nm−− m n n12 ()zz+Δ =∑⎜⎟ z Δ zznzzOz = + Δ +(), Δ m=0 ⎝⎠m (2.29) nn+Δn − n −1 dz()zz z nzΔ z − ===lim limnzn 1 . dzΔ→zz00ΔΔ z Δ→ z Clearly this means that all power series in z are analytic within their radius of convergence. Note that • Inverse powers z : z−n , are singular at the origin, so are not analytic in the vicinity of z = 0. • Inverse power series in z , can be thought of as power series in ()1/ z . ∞ ()−−1 = n f zcz∑ n . (2.30) n=0 c By the ratio test such series converge for limn+1 < 1, or for n→∞ czn c zr>=limn+1 , (2.31) n→∞ cn that is, they represent functions that are analytic outside of some radius of convergence. Analytic continuation 53 2.9 The exponential function The exponential function is unusual in to it has a special syntax ez()= ez ; some of its most important properties are listed below. dez = ez dz ()zz+ eezz12= e12 (2.32) ∞ z n ezz =<∞∑ n=0 n! ee1 ==2.718281828... Proof: The first equation is simply the definition of the expo- nential function as the function that is its own derivative. The power series for ez comes from substituting the series into the differential equation. We have already done this in the section on infinite series, just substitute x → z in the proof. The proof ()zz+ that eezz12= e12 can be derived by substituting the series for the function in the expression, then rearranging the terms. An outline of a proof follows: ∞∞nm 22 zz zz⎛⎞⎛⎞ zz zz ee12==++++++∑∑12⎜⎟⎜⎟1...1... 11 22 nm==00nm!!⎝⎠⎝⎠ 12 12 ⎛⎞zz+++ z222 zzz =+⎜⎟1...12 + 1 122 + ⎝⎠12 (2.33) ⎛⎞()zz++() zz2 =+⎜⎟112 + 12 + ... + ... ⎜⎟12 ⎝⎠ ∞ n ()zz+ ()+ ==∑ 12 e zz12. n=0 n! 54 Analytic continuation A more formal proof can be made using the binomial theorem. From that it follows that n n ()eeezznz==∏ (2.34) n=1 Then, by extension, for any number c , we define c ()eezcz= (2.35) These properties, given in (2.33) and (2.35), are the justification for using a power law representation for ez(). 2.10 The natural logarithm The natural logarithm is the inverse of the exponential function. Given we= z , ln(wz )= . (2.36) + ==zz12() zz 12 Multiplying ww12 e e e gives =+= + ln(ww12 ) z 1 z 2 ln w 1 ln w 2 . (2.37) ln(z ) can easily be evaluated using polar notation. Let zrere==iimθθπ(2+ ), then ln()reim(2θπ++ )=+ ln() r ln() e im (2 θπ ) . (2.38) Therefore, ln(z ) is defined as lnzriim=++ ln() θπ 2 for all m=0, ± 1,... (2.39) Analytic continuation 55 and the of the logarithm is defined as lnzri=+−<≤ ln() θπθπ , . (2.40) For example, ln()ieiim==+ ln()iπ /2 ππ / 2 2 . (2.41) For all integer m . Therefore, the logarithm is a multivalued function of a complex variable. 2.11 The power function The power function is defined by analytic continuation as w zewz==()ln( ) e wz ln . (2.42) For example, i iπ 2 −+()ππ ()ie==iiln e ie ln = e ii (ππ 2+ 2 m ) = e/2 2m . (2.43) Note that all the roots of ()i i are real. This definition results in the expected behavior for products of powers: ()ww++ln z () ww zzwwzwz112ww === eln e ln e12 z 12 (2.44) 2.12 The under-damped harmonic oscillator The equation for the damped harmonic oscillator is given by 56 Analytic continuation mx ++= bx kx 0, (2.45) where x ==dx/ dt and x d22 x / dt . This equation can be thought of as the projection onto the x axis of motion in a 2dimensional space given by zxiy=+ mz ++= bz kz 0. (2.46) Let’s try a solution of the form zt()= eut . This leads to the qua- dratic equation mu2 ++= bu k 0, (2.47) With solutions −±bb2 −4 mk u± = (2.48) 2m If bmk2 −<40, −±bi4 mkb −2 u± = (2.49) 2m The solution oscillates: xt()=+() ceitωω ce*/2−− it e btm =() 2αω cos t + 2 βω sin t e − btm /2 . (2.50) • The complex solutions are weighted sums of decaying spirals one of which rotates clockwise and the other counter- clockwise (Figure 2-7). This diagram could also represent a 2-dimensional phase space plot of position vs. momentum pmv= for a 1-dimensional problem. In that case the point Analytic continuation 57 that they decay into is the stable point of the equations of motion ( xp ==0 ) which is often called the attractor. Figure 2-7 Decaying spiral solutions to the damped oscillator in the complex plane. The total solution for z(t) is the weighted sum of the 2 complex solutions ut+− ut zt()=+ ce+− ce . (2.51) The solution can be made real by taking the projection onto the real axis zz+ * xt()= . (2.52) 2 This forces c± to be conjugate pairs, giving the solution =+()ωω ()−λt zt()⎣⎦⎡⎤ a cos t b sin t e . (2.53) The equation of the under damped oscillator can be rewritten as x()tA=+ cos()ωϕ t e−λt , (2.54) 58 Analytic continuation where A is the amplitude, ω is the oscillation frequency , λ is the decay rate, and ϕ is a phase angle that depends on the initial conditions (see Figure 2-8).The constants α and β are fixed by specifying the initial conditions == x(0)xdxdtv00 ; and / (0) . (2.55) Underdamped Hamonic Oscillator 1 − λ ⋅t A⋅ cos() w⋅φ t+ ⋅e − λ ⋅t Ae 0 − − λ ⋅t Amplitude A e 1 0123 t t (sec) Figure 2-8 The behavior of the damped oscillator x(t), on the real axis. 2.13 Trigonometric and hyperbolic functions All the trigonometric and hyperbolic functions are defined in terms of the exponential function ez . Analytic continuation 59 2.14 The hyperbolic functions cosh z and sinh z are defined as the even and odd parts of the exponential function: ⎛⎞eezz+ − ∞ z2 n cosh() z ==⎜⎟∑ , ⎝⎠22!n=0 n (2.56) ⎛⎞eezz− −+∞ z21 n sinh() z == . ⎜⎟∑ + ⎝⎠221!n=0 n A large number of identities have been tabulated for these func- tions, let’s look at a few cosh22() z−= sinh () z 1, dcoshz() = sinh(), z dz dsinhz() = cosh(), z (2.57) dz cosh(2 z )=+ cosh22 ( z ) sinh ( z ), sinh(2) z= 2 cosh () z sinh (). z The proofs all follow easily from the definitions of the functions. Some selected proofs follow: 22 Example: Prove coshzz−= sinh 1: zz++−−22 zz 22⎛⎞⎛⎞ee ee cosh z−= sinh z ⎜⎟⎜⎟ − ⎝⎠⎝⎠22 eeeeeeee2222zzzzzzzz++22−− −+ −− =− (2.58) 44 4eezz− ==1. 4 60 Analytic continuation Example: Prove that dzdzzsinh /= cosh : dsinh() z d⎛⎞⎛⎞ ezz−+ e−− e zz e ===⎜⎟⎜⎟cosh(). z (2.59) dz dx ⎝⎠⎝⎠22 Example: Prove that sinh 2z = 2sinhzz cosh ⎛⎞⎛⎞eezzzz+−−− ee e22 z − e − z 2()()2sinh z cosh z===⎜⎟⎜⎟ sinh (2). z (2.60) ⎝⎠⎝⎠22 2 2.15 The trigonometric functions The trigonometric functions are defined as the mapping eeziz→ giving ∞ ()iz n eziziz ==+∑ cos( ) sin( ), n=0 n! ∞ ()iz 2n cos(zcoshiz )== ( )∑ , (2.61) n=0 2!n + ∞ ()iz 21n sin(zisinhizi )=− ( ) =− . ∑ + n=0 21!n Again a large number of identities have been derived for these functions, and a few of these are Analytic continuation 61 cos22() z+= sin () z 1, dcos() z =−sin(), z dz dsin() z = cos(), z (2.62) dz cos(2 z )=− cos22 ( z ) sin ( z ), sin(2 z )= 2sin( z )cos( z ). The proofs are similar to the proofs for the hyperbolic functions. Here is an example proof, made by direct substitution: 22−=− 222 ⎣⎦⎣⎦⎡⎤⎡⎤cosh (iz ) sinh ( iz ) cos ( z ) i sin ( z ) (2.63) =+=22 ⎡⎤⎣⎦cos (zz ) sin ( ) 1. It is reassuring to know that all the familiar trigonometric iden- tities, that we commonly use in real analysis, carry over essen- tially unchanged into the complex plane. 2.16 Inverse trigonometric and hyperbolic functions The inverse trigonometric and hyperbolic functions can be ex- pressed in terms of the natural logarithm. However, it takes some practice to get good at this. 62 Analytic continuation arcsinh ()z Example: Find : eezz−+−− uu1 wz==sinh( ) = , 22 zw= arcsinh() , ue= z , 21,wu=− u2 (2.64) uwu2 −−=210, ()uw−=+2 w2 1, ue==±z w w2 +1, zww=±+ln( 2 1) . Solving for z zwww==±+sinh−12 ( ) ln( 1 ) (2.65) but which of the signs is correct? That depends on the problem to be solve. For example, one can find, in a book of math inte- grals, the following formula: x ⎛⎞dx − = sinh1 (x ) >0 for x>0, (2.66) ∫ ⎜⎟2 0 ⎝⎠x −1 implying that the positive branch is the correct one for the case x>0. Analytic continuation 63 2.17 The Cauchy Riemann conditions If a function f (,xy )=+ uxy (, ) ivxy (, ) is analytic in a region, the real and imaginary parts of the function satisfy the following constraints, called the Cauchy Riemann conditions: ∂∂uv ∂ u ∂ v ==−; (2.67) ∂∂x yy ∂ ∂ x By saying that f (,xy ) is analytic in a region we mean that the derivative exists and is unique at each and every point in the re- gion. The existence of the derivative implies, by the chain rule, the existence of the partial derivatives with respect to x and y. Let us consider df() z/ dz calculated two different ways, first by holding y constant, then by holding x constant. In the first case we get df(, x y ) ∂∂∂ f u v ==+i . (2.68) ∂∂∂ dz xy= const x x Secondly, holding x constant gives df(, x y )∂∂∂∂∂ f 1⎛⎞ u v v u ==+=−ii. (2.69) ∂∂∂∂∂⎜⎟ dz i yxconst= i⎝⎠ y y y y But the two expression are the same; therefore comparing the real and imaginary parts, we get the Cauchy-Riemann condi- tions. The Cauchy-Riemann conditions are both necessary and sufficient conditions for a function to be analytic in a region. Basically, the proof follows from rotational invariance: these 64 Analytic continuation conditions have to be met on every straight line path chosen to approach the limit. Any other well behaved path can be approx- imated by a straight line over a small enough interval. Example: Using the Cauchy Riemann conditions it is easy to show that zxiy∗ =− is not an analytic function of z , applying them we get ∂∂x y ≠− (2.70) ∂∂x y So z∗ is not an analytic function of z . 2.18 Solution to Laplace equation in two dimensions The Laplace equation in 2dimensions can be written as ∂∂22 ⎛⎞+Φ= ⎜⎟22(,xy ) 0. (2.71) ⎝⎠∂∂xy It is easy to show that, for real Φ the general solution take’s the form Φ=++−(,x yfxiygxiy ) ( ) ( ). (2.72) Where ∗ g()z∗ = ()f ()z (2.73) Proof: By direct substitution, show that f ()z is a solution: Analytic continuation 65 f ()()xiyfz+= ∂∂fdfzdffdfzdf ∂22 ∂ 2 ==,, = = ∂∂xdzxdzx ∂22 dzxdz ∂ 2 ∂∂f df z df ∂22 f d f ∂ z d 2 f (2.74) ==iii,, = =2 ∂∂y dz y dz ∂22 y d z ∂ y d 2 z ∂∂22f f df 22 df ∴+=−=0. ∂∂22x y dz 2 dz 2 The proof for gz()∗ , is similar, but, more directly, since the op- ()∗ = ′ ∗ erator is real, if f ()z is a solution, then f ()z g (z ) must al- so be a solution. If Φ(,x y ) represents a real potential, then the solution takes the self-conjugate form ∗ Φ=+(,xy ) f ()z ()f ()z . (2.75) 3. Gamma and Beta Functions A function that calls itself is like a dog chasing its tail. Where will this nonsense end? 3.1 The Gamma function The coefficients of infinite power series are often given in terms of recursive relations. For example the series solution x = n eax∑ n to the differential equation for the exponential func- tion dex / dx= ex leads to the following recursive formula: 1 aa=⋅− , (3.1) nnn 1 with a solution 1 aa= . (3.2) n n! 0 = Normalizing to a0 1 gives ∞ xn ex = ∑ , (3.3) n=0 n! where 68 Gamma and Beta Functions n nm!.= ∏ (3.4) m=1 The factorial function occurs in the definition of Taylor’s Expan- sion as well as in the definition of trigonometric and hyperbolic functions. This particular combinatory formula is so useful that it becomes desirable to extend its definition to non-integer val- ues of n . The key property of the factorial is its : nnn!1!.=⋅() − (3.5) It is this property that we wish to maintain as we extend it into the domain of real numbers. Extension of the Factorial function The Gamma function represents the extension of the factorial function. Its definition must satisfy two key requirements: Γ+=(pp 1) ! for integer p > 0, (3.6) Γ+=Γ(ppp 1) ( ) for all real p . It is sufficient to define Γ()p in the interval p = [1, 2] as recur- sion can be used to generate all other values. However, there ex- ists a definite integral that has all the required properties and which is valid for all positive p . This integral is what is used to define Γ()p for p > 0 . This integral is given by ∞ Γ=()p ∫tedtpt−−1 (3.7) 0 Gamma and Beta Functions 69 It is easy to demonstrate by integration by parts that ∞∞ ∞∞ Γ=(2)te1 −−tt dt = te + e − t dt =−= 0 e − t 1=1! (3.8) ∫∫00 00 And, also by integration by parts, ∞∞ ∞ Γ+=(1)p tedttept−− = pt + pt p −−1 edt t =Γ p (). p (3.9) ∫∫0 00 Therefore, by recursion, Γ=Γ=(3)2(2)2!, etc. The integral (3.7) meets all the necessary requirements to be the extension of the factorial function (3.4). By explicit integration, we find ∞ Γ=(1)∫ edt−t == 1 0! (3.10) 0 which defines 0!, and ∞ Γ=−=(0) 1!∫ tedt−−1 t =∞ . (3.11) 0 In fact, , since by using Γ=Γ+()ppp ( 1)/ 1 Γ−(1) = Γ− (1 + 1) =−Γ (0) =−∞ , −1 (3.12) 111⋅ Γ−( 2) = ⋅Γ− ( 1) = ⋅Γ (0) =∞ , et etc. −−⋅−2(2)(1) 70 Gamma and Beta Functions Gamma Functions for negative values of p Evaluation of the Gamma function for negative p is given by re- peated applications of the relationship 1 Γ=Γ+()pp() 1, (3.13) p until Γ+()pn returns a positive number. A plot of the Gamma function and its inverse is shown in Figure 3-1 20 Γ()x 0 20 40 4 x 5 1 0 Γ()x 5 40 4 x Figure 3-1 Plot of the Gamma function and its inverse Gamma and Beta Functions 71 The gamma function has a shallow minimum between 12< integers. The inverse of Γ()p , on the other hand, is quite well behaved. For positive p, it has a single maximum between 12< Figure 3-2 The Gamma function represents a recursive mapping of its value in the interval [1,2] of the real number line. An important identity of the Gamma function is π ΓΓ−=()(1)pp . (3.14) sinπ p This identity is useful in relating negative values of p to their positive counterparts. Note also that sinπ p = 0 for integer p . 72 Gamma and Beta Functions Γ=π Example: Show that (1 / 2) . Use (3.14), letting p =1/2, π ΓΓ−=Γ=()11()2 () 1 =π 221, 2 sin()π / 2 (3.15) ∴Γ()1 = π 2 . Γ− Example: Find (3/2). Use the recursive property of the gamma function: Γ=Γ+()ppp (1/ ) , 1114⎛⎞⎛⎞ (3.16) Γ−()3 = Γ−()11 = Γ() = π . 22−−−33⎜⎟⎜⎟1 2 222⎝⎠⎝⎠ 3 Evaluation of definite integrals An important use of Gamma functions is in the evaluation of de- finite integrals. In fact, its definition for positive p can be thought of as defining the normalization of a family of integrals. By making a change of coordinates, the normalization of many other useful integrals can be found. Example: Transformation of coordinates tx= 2. tx= 2 ⇒ x== t;2; dtxdx ∞∞ ()− 2 (3.17) Γ=()p ∫∫tedtpt−−1 = 2 x21p exdx − x . 00 Gamma and Beta Functions 73 Therefore, ∞ 2 Γ=()p 2∫ xedx21px−− . (3.18) 0 Example: Find the normalization of a the Normal Gaussian Dis- tribution The Normal Distribution of a statistical measurement of a quan- tity X , centered at a mean X 0 and having a random rms error spread σ , is given by 2 yx()= Ne−x /2 , (3.19) =−()σ where xXX0 / . This distribution is normalized such that ∞∞ 2 ∫∫y() x dx== N e−x /2 dx 1. (3.20) −∞ −∞ Since the integrand is symmetric we can rewrite this as ∞ − 2 1 ∫ edxx /2 = . (3.21) 0 2N Let’s make the change of variable xx′′= 2 /2⇒ x= 2 x , 1 − (3.22) dx′′′= xdx⇒ dx= x1/2 dx , 2 Then, by substitution, 74 Gamma and Beta Functions ∞∞ −−−2 111′ edxxedxxx/2==Γ=′′ 1/2 ()1 , (3.23) ∫∫ 2 00222N which leads to the following formula 21 1 N ==. (3.24) Γ()1 π 2 2 2 Therefore, the “normalized” is given by ∞ 11−−22 yx()== exx/2 ;∫ e /2 dx 1. (3.25) 22ππ−∞ 3.2 The Beta Function Another statistical combination that often reoccurs is represented by the binomial coefficients, given by ⎛⎞n n! ⎜⎟= . (3.26) ⎝⎠m ()nmm− !! In probability theory, they denote the number of ways one can arrange n objects taken m at a time. These coefficients have al- ready been seen in the Binomial formula for integer powers of n, n n ⎛⎞n nm− m ()AB+=∑⎜⎟ A B. (3.27) m=0 ⎝⎠m Again, one would like to extend this formula to the real number domain. This is done by defining the Β (Greek capital beta) function given by Gamma and Beta Functions 75 ΓΓ()()p q Β=(,)pq . (3.28) Γ+()p q Clearly, the Beta function is symmetric under interchange of in- dices Β=Β(,)pq (,), q p (3.29) and, for integer values of pn= and qm= , nm!( )! 1 nm !( )! Β+(1,1)nm += = (3.30) (nm++ 1)! nm ++ 1 ( nm + )! ⎛⎞nm+ 1 ⎜⎟= . (3.31) ⎝⎠m ()nm++Β+1( n 1,1) m + The Beta function is useful in the determining normalization of many common integrals. Among them are the canonical forms for p >>0,q 0 : − π /2 21p − Β=()p,2sincos,qd (θθθ )( )21q ∫0 − 1 p 1 − Β=()pq,1, x ( − x )q 1 dx (3.32) ∫0 − ∞ p 1 Β=y ()pq,.+ dx ∫0 ()1+ y pq These three integral forms are related to one another by the transformations x ==+sin2 θ yy /() 1 . (3.33) A number of other definite integrals that can be put into one of these forms. 76 Gamma and Beta Functions The following proof that the above functions are indeed equiva- lent to the defining equation for the Beta function makes use of the fact that the integration over the surface of a quadrant 0,0<<∞<<∞xy is integration over a quarter-circle. When we change to polar coordinates the range of angles is 0/2<<θπ , giving ∞∞ 22 ΓΓ=()()p qxedxyedy 4∫∫21px−− 21 qy −− . (3.34) 00 Letting() x== rcosθθ , y r sin , dxdy = rdrd θ : π /2 ∞ −− 2 (3.35) = 4cossin∫∫()()θθθ21pq 21dr2(pq+− ) 1 edr − r , 00 and therefore, π −− ΓΓ=Γ+()()p qpq2( )∫ ( cosθθθ )21pq ( sin ) 21 dpqBpq =Γ+ ( )(,).(3.36) 0 3.3 The Error Function The Error function (Figure 3-3) can be considered as an incom- plete integral over a gamma function or “”. are defined as the partial integrals of the form t γ (,tp )= ∫ tpt−−1 edt . (3.37) 0 Often, it is desirable to use the normalized incomplete gamma functions instead Gamma and Beta Functions 77 t tedtpt−−1 ∫ γ (,tp )= 0 . (3.38) Γ()p For example, the normalization of the Gaussian integral is given by ∞ 2 Γ=(1 / 2) 2∫ edx−x . (3.39) 0 and the error function is defined as the normalized incomplete integral x 2 2 edx− x ∫ x 2 − 2 Erf() x==0 ex dx . (3.40) Γ π ∫ (1 / 2) 0 This integral can be expanded in a power series, giving ∞ − n 2 ()1 + Erf() x= x21n , (3.41) ∑ + π n=0 nn!(2 1) which is useful for small x. The is defined as erfc() x=− 1 erf (). x (3.42) The Error function is closely related to the likelihood of error in a measurement of normally distributed data. However, like many standard mathematical functions, the normalization is slightly different from how physicists would like to see it de- 78 Gamma and Beta Functions fined. Its relation to the Gaussian probability distribution is giv- en by x 1 − ′2 ⎛⎞x Pxx()−=,. ex /2 dxerf′ = (3.43) π ∫ ⎜⎟ 22−x ⎝⎠ This returns the probability that a measurement of normally dis- tributed data falls in the interval [,]−x x . Books of mathematical tables will tabulate at least one of these two functions, if not both. 1 erf() x erfc() x 0 024 x Figure 3-3 The Error function and the complementary Error func- tion 3.4 Asymptotic Series Figure 3-3 shows that the error function converges to its sum rapidly. Indeed the challenge is to measure error probabilities that are small but non-zero at large x. For example, there is a major difference between saying “a proton never decays” and Gamma and Beta Functions 79 that of saying “a proton rarely decays”. Since the universe is still around, the probability must be very, very small; but if we want to quantify this probability, then we must be able to calculate small deviations from zero. Taylor’s expansion in terms of a convergent power series works well for small x. But, at large x, it is convenient to expand the function in inverse powers of x. However, in this case it turns out that that expansion doesn’t converge. For problems like these, the concept of an asymptotic series was created. Let’s look at how we might calculate the complementary func- tion from first principles. Define ∞ 2 − 2 erfc() x=− 1 erf () x = et dt . (3.44) π ∫ x We can try to solve this by integrating by parts by making the substitutions −t2 −−−−2222te 11− eedtedtdetttt===,,2 () (3.45) ttt22 in ∞∞∞ 2211−−−222⎛⎞− edtttt=+⎜⎟ e edt ππ∫∫⎜⎟22tt2 xx⎝⎠x (3.46) ∞ 21⎛⎞−−22 1 =+eedtxt, π ⎜⎟∫ 2 ⎝⎠22xtx where 80 Gamma and Beta Functions ∞∞ 11−−22 edttt< edt. (3.47) ∫∫22 xx22tx Therefore, after the first integration by parts, the fractional error 1 in the remainder is less than . This is a small error if x is 2x2 large enough. We can repeat the process if we are not satisfied, getting the asymptotic series − 2 ⎛⎞ e x 113135⋅⋅⋅ erfc() x ∼ ⎜⎟ 1−+ − + . (3.48) x π ⎜⎟2x2 ()2223() ⎝⎠22xx For a number of iterations, integration by parts improves the er- ror, but, after a while, the error begins to grow again (there is a double factorial hiding in the numerator). Therefore, there are an optimum number of integration by parts to make. The series, with the partial sum SN taken to infinity, does not converge. Nevertheless the error in the finite series (for fixed N ) goes to zero as x →∞. This is the difference between the definition of a convergent series and an asymptotic series. A convergent series is convergent for a given x, i.e., holding x constant, one takes the limit N →∞; The asymptotic series holds N constant and takes the limit x →∞. Definition: A series f(x) is said to be an asymptotic series −n in x , written as ∞ a fx()∼ n , (3.49) ∑ n n=0 x Gamma and Beta Functions 81 if the absolute value of the difference of the function and the partial sum goes to zero faster than x− N for fixed N as x →∞ N a limfx ( )−⋅→n xN 0. (3.50) x→∞ ∑ n n=0 x It is possible for as series to be both convergent and asymptot- ic—e.g., all power convergent series in x can be said to be asymptotic as ()x → 0 —but the non-convergent case is the most interesting one. Asymptotic series often occur in the solution of integrals of the following kind: ∞ I ()xefudu= −u () (3.51) 1 ∫ x Or of the type ∞ − ⎛⎞u I ()xefdu= u , (3.52) 2 ∫ ⎜⎟ x ⎝⎠x where f ()ux/ is expanded in a power series in ux/ . Sterling’s formula is a good example of a non-convergent asymptotic series. It is given by xx− ⎧ 1 1 139 ⎫ Γ+(1)2pxxe∼ π ⋅⋅⎨ 1 + + − +⎬ . (3.53) ⎩⎭12xx 28823 51840 x 82 Gamma and Beta Functions If p is as small as 10, stopping after the second term gives an error on the order of 50 ppm. For very large p , p!2∼ π ppe⋅ pp− (3.54) is a good approximation to the factorial function, where the ∼ indicates that the ratio of the two sides approaches 1 as p →∞. Discussion Problem: The Exponential Integral The integral − ∞ e t Ei() x= dt (3.55) ∫x t is called the Exponential Integral. Note that it diverges as x → 0 . • Find the asymptotic expansion for the exponential integral. x • Express 1/ln(1/tdt ) as an exponential integral. ∫0 4. Elliptic Integrals When Kepler replaced the epicycles of the ancients with ellipses, he was onto something special. Books of integral tables tabulate and catalog integrals in terms of families with a certain generic behavior. For example, a large number of integrals can be categorized as a rational function of x times a radical of the form ax2 ++ bx c The solution of inte- grals of this general form almost always can be expressed in terms of elementary trigonometric functions or hyperbolic func- tions. For example, substituting x = sinθ , θθsin 2 ∫∫1cos−=x22dxθθ d =++ C 24 (4.1) =+−+1 (arcsin(x )xxC 12 ) . 2 Or a similar example: 1 dx==θ arcsin x . (4.2) ∫ 2 1− x Elliptic functions were introduced to allow the solutions to the large class of problems of the form ∫ R()xydx, (4.3) 84 Elliptic Integrals where Rxy(), is any rational function of x and y , and y()xaxbxcxdxe=++++432 . (4.4) The more complete math tables will have many pages of exam- ples of integrals of this type, solved in terms of one of three standard forms of elliptic integrals, called the elliptic integrals of the 1st, 2nd, and 3rd kinds. We will look at some detail at the first 2 kinds of elliptic integral. The elliptic integral of the 3rd kind is less frequently seen in elementary physics texts. These integrals are usually expressed in one of two standard forms, called the form and the forms of the integrals. The Elliptic integral of the 2nd kind is related to the arc length of an ellipse, which lent its name to this class of integrals. Therefore we will examine the in- tegrals of the second kind first. 4.1 Elliptic integral of the second kind The Jacobi form for the incomplete elliptic integral of the 2nd kind is given by x 1− kx22 Ekx(,)=≤≤ dx for 0 k 1. (4.5) ∫ 2 0 1− x Note that ()()11−−kx22 x 2 has 4 real roots at x =±1, ± 1 /k . Letting x = sinφ , the Legendre form of the integral is given by Elliptic Integrals 85 φ Ek( ,φφφ )=−∫ 1 k22 sin d for 0 ≤≤ k 1. (4.6) 0 A plot of the Legendre form, shown in Figure 4-1, is illustrative of the general behavior. The integrand is periodic on interval [0,π ], so the functions are sometimes said to be “doubly- periodic” in φ . The integral is a repeating sum of the form E(,knφ +=π ) 2 nEkEk () + (,)φ (4.7) where E()k is the given by π 1/21− kx22 Ek()== Ek (,π /2) dx =− 1 k22 sinφφ d . (4.8) ∫∫2 001− x By symmetry about π /2, it is only necessary to tabulate the integral from [0,π / 2]. Ek(,φ +=π /2)2 Ek () − Ek (,),φ (4.9) ()0/2.≤≤φπ The integral over an even integrand is an odd function so Ek(),,.φφ=− Ek () (4.10) Combining the above results gives Ekn()()(),2πφ±= nEk ± Ek , φ for 0/2. ≤≤ φπ (4.11) 86 Elliptic Integrals Elliptic Integral of the Second Kind 2 2 1k− ⋅sin()φ Ek(), φ 4 0 0 90 180 270 360 φ deg phi(deg) Figure 4-1 Elliptic Integral of the second kind (k=sin(60 deg)) Example: Calculate the arc length of a segment of an ellipse. An ellipse has a semi-major axis of length a and a semi-minor axis of length b (see Figure 4-2). Aligning the ellipse with the semi-minor axis along the x-axis, it can be described by the fol- lowing two parametric equations: xb= cosφ , (4.12) ya= sinφ . ellispse 1 1 a⋅ sin()θ 0 − 1 1 10 1 − 1 b⋅cos(θ) 1 Figure 4-2 diagram of an ellipse with ab==1, 0.5 Elliptic Integrals 87 An element of arc-length can be written as ds=+= d22 x d y b 2222sinφφφ + a cos d 22 (4.13) ab− =−adakd1sin1sin,φφ =−2 φφ a2 2 where kbae=−1/() = is the eccentricity of the ellipse; k = 0 is a circle; and k =1 is a vertical line. Integrating along φ gives θ ∫ ak1sin−=22φφ daEe() ,. φ (4.14) 0 The circumference of the ellipse is found by integrating over a complete revolution CaEk==(),2π 4 aEk () . (4.15) Verify: kCaEa===0, circle, 4 (0) 2π , (4.16) kCaEa===1, straight line, 4 (1) 4 . Since the orbits of planets are ellipses, Ek(),φ is a very valuable function. There are several ways common ways of calculating Ek(),φ : • Look up the tabulated value in a book of integral tables (the common way, before the invention of personal computers). • Use a high level math program like Maple or Mathematica. 88 Elliptic Integrals • Use a scientific programming library, and your favorite pro- gramming language. • Expand the integral in a power series insin2 θ (converges ra- pidly for small k ). 4.2 Elliptic Integral of the first kind The elliptic integral of the first kind occurs in the solution of many classical mechanics problems, including the famous one of the simple pendulum. Whole books have be written about it. The Jacobi form of the integral is given by x 1 Fkx(,)=≤≤ dx for 0 k 1. (4.17) ∫ 222 0 ()()11−−xkx And, letting x = sinφ , the Legendre form is given by θ Fk(),φφφ=−∫ 1 k22 sin d for 0 ≤≤ k 1. (4.18) 0 The is given by π /2 1 Kk()== Fk (,π /2) dφ . (4.19) ∫ 22 0 1sin− k φ Figure 4-3 shows a plot of the elliptic integral of the 1st kind, for k = sin 60 . The same kind of symmetry arguments used in dis- cussing the Integrals of the 2nd kind apply here, F is doubly pe- Elliptic Integrals 89 riodic in φ , and, by symmetry, only the values between [0,π / 2] need to be tabulated. In general Fkn()()(),πφ±= 2 nKk ± Fk , φ for 0 ≤≤ φπ / 2. (4.20) Table 4-1 tabulates the values of the complete elliptic integrals of the first and second kind. When k = 0 , one has a circle and the value of the integral isπ /2. For k =1, the complete elliptic integral of the 1st kind diverges. Ellipitic Integra of the First Kind 1 2 2 1k− ⋅sin()φ Fk(), φ 4 0 0 90 180 270 360 φ deg phi (deg) Figure 4-3 Elliptic Integral of the first kind (k=sin(60 deg)) Integrands of Elliptic Integrals are periodic on interval 180 deg and are symmetric about half that interval; therefore, they are generally only tabulated in the interval [0, 90] deg. Table 4-1 Complete Elliptic Integrals of the first and second kind Complete Elliptic Integrals of the 1st and 2nd kind 90 Elliptic Integrals ψ E(sin(ψ )) K(sin(ψ )) 0 1.571 1.571 5 1.568 1.574 10 1.559 1.583 15 1.544 1.598 20 1.524 1.62 25 1.498 1.649 30 1.467 1.686 35 1.432 1.731 40 1.393 1.787 45 1.351 1.854 50 1.306 1.936 55 1.259 2.035 60 1.211 2.157 65 1.164 2.309 70 1.118 2.505 Elliptic Integrals 91 75 1.076 2.768 80 1.04 3.153 85 1.013 3.832 90 1 ∞ ()θ ==φ (θ )φ Example: By substituting sin / 2k sin sinmax / 2 sin , and using cosθθ=− 1 2sin2 () / 2 , show that F ()k,φ can be writ- ten as θ 1 dθ Fk(),φ = . (4.21) ∫ θθ− 2coscos0 max Proof: Work the problem backwards from the result: Begin by calculating the change in derivatives: 1 ddkdkdsin()θθθ / 2=== cos() / 2 sinφφφ cos , 2 (4.22) 2cosk φ ddθφ= . cos()θ / 2 Next, make the necessary substitutions into the integral: 92 Elliptic Integrals θ 1 dθ F = , 2 ∫ −−−22()θθ() 0 ()12sin /2() 12sinmax /2 θθ 11ddθθ ==, ∫∫22 22 22sin()()θθ / 2−− sin / 2k sin() θ / 2 00max (4.23) θθφ 1112cosdddkθθφφ === , ∫∫∫222 φφθ() 2000kk− sin φ 2kk cos 2 cos cos / 2 φ θθ ddφφ d φ == = . ∫∫()θ 222 ∫ 00cos / 2 1−− sinθϕ / 2 0 1k sin 4.3 Jacobi Elliptic functions A rich literature has grown around the topic of the elliptic integral of the 1st kind, with a specialized language and names for functions. To see where this language comes from consider the simpler circular integral which can be obtained by letting k = 0 in the Jacobi form: x 1 uF==()0, x dx ==φ arcsin x , ∫ 2 0 1− x (4.24) sn() u== x sinφ . Now let’s generalize this terminology for k ≠ 0 , in which case u ≠ φ . We call φ the amplitude of u amp() u = φ . (4.25) and denote the inverse function F −1 by the special name so that Elliptic Integrals 93 x − 1 usnxFk==1 (),φ = dx ∫ 222 0 11−−xkx φ 1 = dφ, (4.26) ∫ 22 0 1sin− k φ sn() u== xsinφ = sin() amp () u . sn is pronounced roughly as “ess-en”. (Try saying three times fast: “ess-en u is the sine of the amplitude of u”). A plot of sn() u vs. u looks very similar to a sine wave as seen in Figure 4-4. 1 sn() u, k 0 1 0510 u Figure 4-4 A plot of sn() u== x sinφ Just as a number of trigonometric identities have been devel- oped over the years, the same is true for the elliptic functions. Several of the more basic relationships are given by: cn() u==−=−=− cosφφ 1 sin22 1 sn() u 1 x 2 , (4.27) φ −1 ddu⎛⎞ 22 22 dn() u= =⎜⎟ =− 1sin1sn kφ =− k u , (4.28) du⎝⎠ dφ dd dφ sn() u== sinφφ cos = cn () u dn () u . (4.29) du du du 94 Elliptic Integrals Example: The simple pendulum Figure 4-5 Simple Pendulum The simple pendulum (Figure 4-5) satisfies the conservation of energy equation Iθ 2 +−=mgl ()cosθθ cos 0, (4.30) 2 max where I = ml 2. Solving, θ 1 dθ 1 = gldt/, (4.31) ∫∫θθ− 2coscos00max Elliptic Integrals 95 θ 1 dθ ωϕtFku===(, ) , 0 ∫ θθ− 2coscos0 max sinθϕ / 2= k sin , k = sinθ / 2, max (4.32) sinθ / 2 xsnusntk===()()ω , , θ 0 sinmax / 2 θω= () sin / 2ksni 0 tk , , θω= () 2arcsin()ksni 0 tk , . The depends on its amplitude and is given by ωθπθ==() = () 0maxmaxTFk()sin / 2 ,2 4 K() sin / 2 . (4.33) Let’s put in some numbers, choosing m = 1 kg, L = 1 m, ω ==2 = 0 gl/ 9.8/ s 3.13 rad/s, (4.34) θ = max 60 , ===θ k sinmax / 2 sin 30 1/ 2. Then, the period of the pendulum is ωπ==() 0TFk1/2,2 , 1 TK= 41/2,() (4.35) ω 0 1 T ==4() 1.686 2.15 s, 3.13 Compare this to the small amplitude limit 2π T ==2.01 s. (4.36) 0 smallOsc ω 0 96 Elliptic Integrals Note the 7% difference from the small oscillation behavior. Nev- er use a simple pendulum to tell time! The analytic solution to the simple pendulum for the conditions studied is shown in Fig- ure 4-6. simple pendulum K=1/2 AND K=1/4 90 45 0 angle (deg) 45 90 0123 time (s) Figure 4-6 Plot of angle vs. time for the simple pendulum. Note that the zero crossing time of the period depends on the ampli- tude. 4.4 Elliptic integral of the third kind For completeness, here is the definition of the Elliptic Integral of the 3rd kind: The Legendre form of the Incomplete Integral is φ dφ Π=(,,)knφ . (4.37) ∫ 222 0 ()1+−nk sinφφ 1 sin And, the Complete Integral is given by π /2 dφ Π=(,,knπ /2) . (4.38) ∫ 222 0 ()1+−nk sinφφ 1 sin Elliptic Integrals 97 5. Fourier Series Time is measured in cycles. The earth rotates around the sun. Atoms oscillate. Patterns repeat. 5.1 Plucking a string What happens when one plucks a string on a stringed instru- ment? The fundamental harmonic is given by the length of the string, its mass density and its tension. Depending on where we pluck the string, one can choose to emphasize different harmon- ics. After this point, it starts to get complicated, as the shape and nature of the sound board will further modify the sound. This problem will be analyzed in some detail when we study the wave equation. But in general, if one does a Fourier decomposition of the wave form, only multiples of the fundamental frequency will contribute. In the case of a plucked string, the boundary condi- tions are due to the clamping of the string, which removes all other frequencies. Is this effect real, or is it a mathematical con- nivance? It is definitely real, one can hear it. The human ear is a pretty good frequency analyzer. In this section we will explore how to decompose a periodic function into its Fourier series components. These components are a solution to an eigenvalue 100 Fourier Series problem. The eigenfunctions represent the possible normal modes of oscillation of a periodic function. In the case of the plucked string, the motion is periodic in time. In other cases, we might be dealing with a cyclic variable, say the rotation angle of a planet as it makes its path around its sun. We begin with the simplest of models: a one dimensional rotation angle. If one de- fines a field on a circle, consistency requires that a rotation by 2π must give the same field. 5.2 The solution to a simple eigenvalue equation In solutions to partial differential equations in cylindrical or spherical coordinates, the technique of separation of variables often leads to the following very simple equation for the azimul- thal coordinate φ df2 ()φ =−mf2 ()φ (5.1) dφ 2 where f ()φ satisfies periodic boundary conditions ff()()φ +=2π φ (5.2) The solutions of this equation are very well known —think “sim- ple harmonic oscillator”— The exponential form of the solution is ()φ ==±±imφ fcemmm,0,1,2, (5.3) Fourier Series 101 ()φ π The requirement that fm be periodic on interval 2 restricts the eigenvalues to positive and negative integers. The case m = 0 is a special case in that m =±0 represents the same eigenvalue. Often it is convenient to solve the equation in terms of sine and cosine functions. Using emim±imφ =±cosφφ sin , we find the real solutions to the eigenvalue equation to be ⎧ a 0 m=0 ()φ = ⎪ fm ⎨ 2 (5.4) ⎪ ()φφ+ () ⎩ambmmmcos sin m=1,2,3,… Here, the counting runs only over non-negative integers, since sin(−mφ ) and cos(−mφ ) are not linearly independent from sin(mφ ) and cos(mφ ) Orthogonality The eigenfunctions solutions of this equation are orthogonal to each other when integrated over interval 2π . First, let us prove this for the complex form of the series, normalizing the func- = tions by setting cm 1: ⎧ 2π mm= ′ ππ π ′ ()′− φ ⎪ π ffd* φφφ=== e−imφφ e im d eim m d im()′− mφ (5.5) ∫∫mm′ ∫ ⎨e =≠′ −−ππ − π ⎪ im()′− mφ 0 mm ⎩ −π The proof for sine and cosine series is slightly more complicated. If mm≠ ′, the sine and cosine functions can be re-expanded into 102 Fourier Series terms involving e±imφ , and orthogonality follows from the above equation: ππ ∫∫sin()(mmdmφφφφ sin′′ )= sin () cos ( md φφ ) −−ππ π (5.6) ==≠∫ cos()mmdmmφφφ cos (′′ ) 0 −π For mm= ′ , one can use the fact that sin() 2mφ is odd on interval [−ππ, ] to show ππ 1 ∫∫sin()mmdφφφ cos ()== sin() 2 md φφ 0 (5.7) −−ππ2 Also, by symmetry, using the fact that sine and cosine functions are the same up to a phase shift, ππ ∫∫()sin22mmddφφφφ+== cos() 1 2π −−ππ ππ (5.8) 2π ∫∫sin22mdφφ=== cos md φφ π −−ππ 2 Finally, normalize the m = 0 term to unity (1cos(0)= ), giving ππ ∫∫1sin()md′′φφ== 1cos () md φφ 0 −−ππ π (5.9) ∫ 12dφπ= −π In quantum mechanics (QM), it is conventional to normalize the square integrated eigenfunctions to unity. However, this is not Fourier Series 103 often the case in classical physics. The special functions of ma- thematical physics have a variety of sometimes bewildering normalizations, all of which made sense to the people who first studied them. The definition of Fourier series long predated QM. The above normalization is standard in the literature. 5.3 Definition of Fourier series A, , function f ()x which is periodic over interval [−ππ, ], and π 2 where the positive-definite integral ∫ f ()xdxis finite, can be −π expanded in a series of sine and cosine functions having the general form a ∞∞ ()=+0 () + () f xanxbnx∑∑nncos sin (5.10) 2 nn==11 The function has a finite number of maxima and minima The function has a finite number of step-wise discrete disconti- nuities The coefficients of the series are given by π 1 afxnxdxm==( )cos() for 0,1,2… n π ∫ −π π (5.11) 1 bfxnxdxm==( )sin() for 0,1,2… n π ∫ −π 104 Fourier Series The infinite series converges to the function where it is conti- nuous and to the midpoint of the discontinuity where it is step- wise discontinuous. Completeness of the series In what sense can the function, which may be discontinuous af- ter all, be said to be equal to a series consisting of only conti- nuous functions? Note the limitation of the discontinuities to a finite number of discrete points. These points have zero weight when the function is integrated. The series and the function can be said to be equivalent up to an interval of zero measure. That is, π −=2 limfx ( ) fN ( x ) dx 0 (5.12) N →∞ ∫ −π where fN ()x is the partial sum of the infinite series. Sine and cosine series If a piece-wise continuous, periodic function f(x) is an even function of x, it may be expanded in a Fourier Cosine series on interval [,−+ππ ] a ∞ ()=+0 () fcnxanx∑ cos (5.13) 2 n=1 Fourier Series 105 If a piece-wise continuous, periodic function f(x) is an odd func- tion of x, it may be expanded in a Fourier Sine series on interval [,−+ππ ] ∞ ()= ( ) fsnxbnx∑ sin (5.14) n=1 Complex form of Fourier series A real-valued function can also be represented as a complex in- = () finite series. Let cab±mmm∓ /2, then ∞ ()= inx f xce∑ n (5.15) n=−∞ With coefficients given by π 1 − cefxdx= inx () (5.16) n π ∫ 2 −π Note for m ≠ 0 −−aib−+ aib imx+= imx⎛⎞⎛⎞mm imx + mm imx cemm c− e⎜⎟⎜⎟ e e ⎝⎠⎝⎠22 imx+−−− imx imx imx =+⎛⎞⎛⎞ee ee abmm⎜⎟⎜⎟ (5.17) ⎝⎠⎝⎠22i =+ amxbmxmmcos sin 106 Fourier Series 5.4 Other intervals Often the interval is of arbitrary length 2L rather than 2π . Usually it is preferable to keep the interval symmetric over [−LL, ]. This involves making a change of variable x //xL′ = π (5.18) dx= ()π / L dx′ Making these changes we find a ∞∞nxππ′′ nx ()′ =+0 ⎛⎞ + ⎛⎞ fx∑∑ anncos⎜⎟ b sin ⎜⎟ 2 nn==11⎝⎠LL ⎝⎠ 1 L ⎛⎞nxπ ′ afxdx= ()cos′′ (5.19) n ∫ ⎜⎟ LL−L ⎝⎠ 1 L ⎛⎞nxπ ′ bfxdx= ()sin′′ n ∫ ⎜⎟ LL−L ⎝⎠ 5.5 Examples The Full wave Rectifier The full wave rectifier takes a sinusoidal wave at line frequency and rectifies it using a bridge diode circuit. Figure 5-1 shows a schematic of a full wave rectifier circuit. Positive and negative parts of the line cycle take different paths through this diode bridge circuit, but the current always flows through the resister in a unidirectional manner, rectifying the signal. The result is Fourier Series 107 usually filtered, by adding a capacitor to the output line, to give an approximation of a D.C. circuit, but for many purposes this first step is sufficient. R I Figure 5-1 Diagram of a full wave rectifier circuit The initial wave is a pure sine wave at the base line frequency. After rectification (Figure 5-2), this frequency disappears and one is left with a hierarchy of frequencies starting at double the base frequency. Let’s look at the base wave form as it appears on an oscilloscope, locked to the line frequency. It is given by the periodic function ==−<≤θπθπ fwtlinesin o sin (5.20) θ = where wto . After rectification, but before filtering, the mod- ified wave form is given by 108 Fourier Series = θ fout sin (5.21) Full Wave Rectifier 1 abs(sin(theta)) 0 20 2 theta Figure 5-2 The output of a full wave rectifier, before filtering Note that the original wave form was an odd function of θ , while the rectified function is an even function of θ . The original frequency has completely disappeared, and one is left with har- monics based on a new fundamental of double the frequency. Even symmetry under a sign change in θ implies that we can expand the solution in terms of a Fourier Cosine Series: ∞ ==θθ foutsin∑ an n cos (5.22) n=0 The amplitudes of the frequency components are given by π 1 afnd= cos θθ (5.23) noutπ ∫ −π However, since the integrand is even, we need calculate only the positive half cycle Fourier Series 109 ππ 22 afnd==cosθθ sin θ cos nd θθ (5.24) noutππ∫∫ 00 This integral can easily be solved by converting to exponential notation 2 1 ⎛⎞⎛⎞eeiθθ−+−− i e in θ e in θ ad= θ (5.25) n π ∫ ⎜⎟⎜⎟ 0 ⎝⎠⎝⎠22i However, it is even easier to look up the answer in a book of integral tables: π π ⎛⎞cos() (1−+nn )θθ cos() (1 ) sinθθθ cos nd =−⎜⎟ (5.26) ∫ 2(1−+nn ) 2(1 ) 0 ⎝⎠0 The integral vanishes for odd n, and the result can be written as ⎧ 1 −4 ⎪ n even a = 2 − (5.27) n π ⎨n 1 ⎩⎪ 0n odd The first term is positive and gives a DC offset π 12a ffd===≈θ 0 0.637 (5.28) outππ∫ out 22−π 1 L a fx()==∫ fxdx () 0 (5.29) 22L −L The evaluated series can be written as 110 Fourier Series 2 ∞ 4cos() 2nw t sin(wt ) =− 0 (5.30) 0 π ∑ 2 n=1 π ()()21n − And the allowed frequencies are = wnw2 0 (5.31) 2 Successively amplitudes fall off as 1/()() 2n − 1 , which means that the energy stored in the frequency components falls off as 2 1/()() 2n 2 − 1 . Clearly, the series is convergent since by the integral test 4/π 1 dn ⎯⎯⎯→→0 (5.32) ∫ 41nn2 − n→∞ π By adding a capacitor on the output side of the full wave rectifi- er, one can short circuit the high frequency components to ground. If the capacitor is large enough, the output of the circuit is nearly D.C. The Square wave The square wave (shown in Figure 5-4) and its variants (i.e., the step function, etc) are often found in digital circuits. The wave form is given by ⎧+<<10θπ f ()θ = ⎨ (5.33) ⎩−−<<10πθ Fourier Series 111 This function has stepwise singularities at θπ=±{0, } . By the Fourier Series Theorem the series will converge to the midpoint of the discontinuity at those points. This function is an odd func- tion of θ , so it can be expanded in a Fourier Sine series ∞ = ()θ fSquareWave∑bn n sin (5.34) n=1 Where πππ 122 bfndfndnd===sinθθ sin θθ sin θθ (5.35) nSWSWπππ∫∫∫ −π 00 The solution is π ⎧ 4 − n −−2() cos(nπ ) cos(0) n odd ==2 θ = = ⎪ π bdnn cos ⎨n (5.36) nnππ∫ 0 ⎩⎪ 0 n even The frequency decomposition of the square ware is shown in Figure 5-3. The Fourier series expansion of this wave form is given by ∞ 4 fn=+sin()() 2 1 θ (5.37) SW ∑ + π n=0 ()21n 112 Fourier Series Square wave frequency amplitudes 1.4 1.2 1 0.8 0.6 0.4 0.2 0 12345678910111213141516 Figure 5-3 Square wave frequency components Gibbs Phenomena The Square Wave converges slower than the first series we stu- died, and it is not uniformly convergent, as seen in Figure 5-4(4 and 20 terms are plotted). In fact, one can expect extreme diffi- culties getting a good fit at the discontinuous steps. Any finite number of terms will show in the vicinity of the discontinuity. The amplitude of this overshoots persists, but as the number of terms increases. As we approach an infinite number of terms, this overshoot covers an interval of negligible measure. This is the meaning of the expression that the series and the function are the same up to . Mathematically, this is expressed by π −=2 limfx ( ) fN ( x ) dx 0 (5.38) N →∞ ∫ −π Fourier Series 113 This behavior is not unique to the square wave. Similar over- shoots occur whenever there is a discontinuity. This behavior at stepwise discontinuities is referred to as the . Fourier Series Fit to a Square Wave 1.5 1 0.5 0 Amplidude 0.5 1 1.5 20 2 x (radians) Figure 5-4 The Gibbs Phenomena If one uses a good analog scope to view a square wave generated by a pulse generator, odds are that you won’t see any such beha- vior. In part, this is because the analog nature of the scope. But there are more fundamental reasons. These pertain to how the signal is measured and how it was originally generated. If one has a fast pulse generator, but a slow scope, then, at the highest 114 Fourier Series time resolutions, one sees a rise time in the signal due to the re- sponse of the scope. If one has a fast scope, but a limited genera- tor, one resolves the time structure of the source instead. Piece- wise step functions are not physical. They are approximations that allow us to ignore the messy details of exactly how a sudden change happened. In time dependent problems, this is called the impulse approximation. For another example, consider a spherical capacitor, with one conducting hemisphere at positive high voltage and the second at negative high voltage. The step in voltage at the interface ig- nores the necessary presence of a thin insulating barrier sepa- rating the two regions. Such approximations are fine, as long as one understands their limits of validity. Find the Fourier series expansion to the step function given by ⎧10< Hint: Note that it can be written as a sum of an even function and an odd function. Non-symmetric intervals and period doubling Although the interval for fitting the period is often taken to be symmetric, it need not be so. Consider the saw-tooth wave, shown in Figure 5-5, initially defined on the interval[0,L ]. f ()xx=<<;0 xL (5.40) Fourier Series 115 Figure 5-5 A linear wave defined on interval[0,L] One can double the period and fit it either as n even function (Cosine Transform) or as and odd function (Sine Transform). If we are interested only a fit within this region, there are several ways of fitting this function. The most common technique is called : The interval is doubled to the interval [−LL, ]and the function is either symmetrized or anti-symmetrized on this greater interval. The rate of convergence often depends on the choice made. • Symmetric option (Triangle wave) The symmetrized function represents a triangle wave of the form 116 Fourier Series f (xx )=−<< for LxL (5.41) Fitting this with a Cosine Series, gives 12LL⎛⎞nxππ ⎛⎞ nx ax==cos dxx cos dx (5.42) n ∫∫⎜⎟ ⎜⎟ LLLL−L ⎝⎠0 ⎝⎠ Integrate by parts using ∫ udv=−∫∫ d() uv vdu or look up in integral tables cosax x sin ax xaxdxcos =− (5.43) ∫ aa2 This gives the solution ⎛⎞(2nx+ 1)π ∞ cos⎜⎟ LL4 L fx()==− x ⎝⎠ (5.44) π 22∑ + 2(21)n=0 n Note that after the first term, which gives the average value of the function, only terms odd in n contribute. Figure 5-6 shows that after the addition of the first cosine term, the fit to a trian- gle wave is already a fair approximation. Fourier Series 117 Fourier Series Fit to a Triangle Wave 1 0 Amplidude 1 20 2 x (radians) Figure 5-6 Fit to a triangle wave (2 and 10 terms) • Antisymmetric option (sawtooth wave form) The antisymmetrized function is a sawtooth waveform f (xx )=−<< for LxL (5.45) The solutions can now be expressed as a Fourier Sine Series 12LL⎛⎞nxππ ⎛⎞ nx bx==sin dxxdx sin (5.46) n ∫∫⎜⎟ ⎜⎟ LLLL−L ⎝⎠0 ⎝⎠ Integration by parts gives sinax x cos ax xaxdxsin =− (5.47) ∫ aa2 With the solution 118 Fourier Series π ()− n+1 ⎛⎞nx ∞ 1sin⎜⎟ 2L L fx()== x ⎝⎠ (5.48) π ∑ n=0 n Figure 5-7 shows the Fourier series fit to the sawtooth wave form. Note that the Gibbs phenomenon has returned. Compar- 1 ing the two solutions, the triangle wave converges faster ∼ , n2 1 while the sawtooth wave converges only as ∼ . The principle n difference, however, is that the triangle wave is continuous, while the saw tooth has discontinuities at ±π . Given the choice of symmetrizing or anti-symmetrizing a wave form, pick the choice that leads to the best behaved function for the problem at hand. Fourier Series Fit to a Sawtooth Wave 1 0 Amplidude 1 20 2 x (radians) Fourier Series 119 Figure 5-7 Fourier series fit to sawtooth wave(4 and 20 terms) • Combinations of solutions If one sums the even and odd series, the wave form remains un- changed for positive x , but cancels for negative x . This allows us solve for the function ⎧x 0 < giving + ππ ⎛⎞(2nx 1) ()− n ⎛⎞ nx ∞∞cos⎜⎟ 1 sin ⎜⎟ LL2 LL L fx()=−⎝⎠ + ⎝⎠ (5.50) ππ22∑∑+ 4(21)nn==00nn Since the function no longer has definite parity under reflection, a combined Fourier (Cosine + Sine) Series is required. 5.6 Integration and differentiation Like Taylor series, Fourier Series can be differentiated or inte- grated. The effect is easiest to demonstrate using the complex form of the series. Differentiation One can take the derivative of a series within its radius of con- vergence, giving 120 Fourier Series ∞∞ ′()==dd() inx =() inx f xfxceince∑∑nn (5.51) dx dx nn=−∞ =−∞ Because of the added factor of n in the numerator, f ′()x con- verges slower than f ()x . Integration One can integrate a series within its radius of convergence, giv- ing ∞∞ceinx f ()x dx==+ c einx dxn const (5.52) ∫∫∑∑n nn=−∞ =−∞ in Because of the added factor of n in the denominator f ′()x con- verges faster than f ()x . The constant of integration can be tricky. it depends on where the lower limit of integration is placed, as that affects the aver- age value of the function. Often the integral is taken from the origin x = 0 , and the upper limit is either positive or negative x . Fourier Series 121 Example: Evaluate the integral of the square wave ⎧⎫+>10x ∞ 4 fn()θθ==sin()() 2 + 1 SW ⎨⎬−< ∑ ()+ π ⎩⎭10x n=0 21n θ ∞ 4 x fd()θθ== θ sin2()() n + 1 θθ d ∫∫∑ ()+ π 00n=1 21n (5.53) ∞ −4 θ =+cos()() 2n 1 θ ∑ 2 0 n=1 ()21n + π ∞ −+4cos()() 2n 1 θ =+ ∑ 2 C n=1 ()21n + π But this integral must be the triangle wave previously defined over the interval [,]−ππ ,therefore the answer should be π 4 ∞ cos() (2n + 1)θ fxx()==− (5.54) sawtooth π ∑ + 2 2(21)n=0 n Therefore, the constant of integration is π C = (5.55) 2 If we didn’t know the answer ahead of time, one can fix the con- stant by evaluation at a carefully chosen value of θ ∞ − ()==4 + f 00∑ 2 C n=1 ()21n + π (5.56) ∞ π ==4 C ∑ 2 n=1 ()21n + π 2 122 Fourier Series Reversing the procedure, we see that the method allows us to calculate the sum of a difficult looking series of constants in closed form. This is a common use of Fourier Series. Example: Find the value of the Zeta function ς ()m for m=2 The Zeta function is defined as the series of constants ∞ 1 ς ()m = (5.57) ∑ m n=1 n Therefore ∞ 111 π 2 ς ()==+++= 21∑ 2 (5.58) n=1 n 49 6 To prove this identity, the trick will be to find some Fourier se- ries that gives this as a constant series for some value of its pa- rameter. Let’s try fx()=−<< x2 ππ x (5.59) This the even series given by ππ 12 a== x22 cos() nx dx x cos() nx dx n ππ∫∫ −π 0 π π 222x32π axdx===2 0 ππ∫ 33 0 0 (5.60) π ⎛⎞2 =+−22xx()⎛⎞ 2 () anxnxn ⎜⎟cos⎜⎟ sin π ⎜⎟nnn23 ⎝⎠⎝⎠ o 4 n =−()1 n2 Fourier Series 123 or 2 ∞ 2π n f ()xx==2 + 4∑() − 1 cos nx (5.61) 6 n=1 Letting x = π , and using cosnπ =−() 1 n , we get ππ22∞ 1 f ()ππ==+2 442 =+ ς() ∑ 2 33= n n 1 (5.62) 2ππ22 ς ()2 == 34⋅ 6 5.7 Parseval’s Theorem We have already defined the mean (expectation) value of a Fourier series as π 1 a ffxdxx==−<<() 0 ππ (5.63) π ∫ 22−π Let’s now calculate the expectation value of f 2 .using the com- plex series notation L π ∗∗112 ff== f dx ffdx ∫∫π 22L −−L π π ∞∞ = 1 ∗−inθθ im ∑∑cenm ce dx 2π ∫ =−∞ =−∞ −π nm (5.64) ∞∞ π 1 ⎛⎞∗−θθ = cein ce im dx π ∑∑⎜⎟∫ nm 2 nm=−∞ =−∞ ⎝⎠−π ∞∞ ∞ 1 ∗∗ ==≥20πδ cc cc π ∑∑nm n m ∑ n n 2 nm=−∞ =−∞ n =−∞ 124 Fourier Series This is Parseval’s Theorem. For classical waves, the wave’s ener- gy density is proportional to the square of the amplitude of wave. Therefore, Parseval’s Theorem can be interpreted as meaning that the energy per cycle of a particular frequency is proportional to the square of its amplitude integrated over a pe- riod. In Quantum mechanics, the norm of a wave function is usually normalized to unit probability. For this case, the theo- rem is equivalent to saying that the partial probability of finding a particle in a frequency eigenstate n is given by the square of its amplitude. Definition: Parseval’s Theorem The expectation value of the square of the absolute value of a function when averaged over its interval of periodicity is given by LL∞ 1122∗ f * f=== f dx f fdx c (5.65) ∫∫∑ n 22LL−−LLn=−∞ If the function is real-valued, then the sum can be written as a2 ∞ 222=+0 () + f ∑ abnn (5.66) 2 n=1 f 2 ≥ 0 so ff* > 0 unless the function vanishes everywhere, except possibly on an interval of null measure. This definition is used to define the norm of a square-integrable function space. The norm of a function is just the sum of the norms of its com- ponent eigenfunctions. Fourier Series 125 Generalized Parseval’s Theorem Parseval’s theorem can be generalized by defining the as ππ∞∞ ∗∗==11 ∗−inθθ im f g f gdx∑∑ fnm e g e dx 22ππ∫∫=−∞ =−∞ −−ππnm (5.67) ∞ = ∗ ∑ fgnn n=−∞ 5.8 Solutions to infinite series We have already seen one way to find the sum of a series and π 2 have shown that ς ()2 = . Here is a second way, using Parsev- 6 al’s Identity. Note that π ππ 32 ∗ 1222x π xxxxdxxdx222== = = = (5.68) πππ∫∫ −π 0 330 However x is just the functional form of a sawtooth wave, and we have already solve that series. Letting the interval L = π , then the Fourier Sine series normalizes to ∞ ()−12n ==() = fx() x∑ bnn sin nx , b (5.69) n=0 n Then, by Parseval’s Identity 126 Fourier Series 2π 2 ∞ 22== =ς () xb∑ n 42 3 = n 1 (5.70) π 2 ∴=ς ()2 6 6. Orthogonal function spaces The normal modes of the continuum define an infinite Hilbert space. 6.1 Separation of variables The solution of complicated mathematical problems is facili- tated by breaking the problem down into simpler components. By reducing the individual pieces into a standard form having a known solution, the solution of the more complex problem can be reconstructed. For example, partial differential equations are often solvable by the technique of separation of variables. The resulting are that can be solved by the general techniques that we will explore in the following sections. The general solution to the original partial differential equation of interest can then be constructed from a summation over all product solutions to the eigenvalue equations that meet certain specified boundary re- quirements imposed by physical considerations. 6.2 Laplace’s equation in polar coordinates An illuminating example is the solution to in two space dimen- sions. The equation takes the form 128 Orthogonal function spaces ∇Ψ=2 0. (6.1) where ∇2 is and Ψ can be interpreted as a static potential func- tion in electromagnetic or gravitational theory, and as a steady state temperature in the context of thermodynamics. Let’s try separating this equation in polar coordinates. The equation can be rewritten as ∂∂∂22 ⎧⎫++11 Ψ()φ = ⎨⎬222r,0. (6.2) ⎩⎭∂∂∂rrrrφ Then, we look for product solutions of the form Ψ=Φ()rfr,φφ ()(). (6.3) Separation of variables leads to the following coupled set ordi- nary differential equations ∂∂2 λ ⎧⎫+−1 () = ⎨⎬22frλ 0 . (6.4) ⎩⎭∂∂rrrr ∂2 ⎧⎫+Φλφ() = ⎨⎬2 λ 0 . (6.5) ⎩⎭∂φ We are not interested in all solutions to this eigenvalue problem, only those that make physical sense. In this case, φ is a cyclic variable, and the requirement that the solution be single-valued (i.e. uniquely defined) imposes the periodic boundary condition Φ+()()φ 2mπ =Φφ . (6.6) This restricts the eigenvalues to the denumerable set Orthogonal function spaces 129 λ =≥∀mm2 0; integer =±± 0, 1, 2,.... (6.7) The solutions, therefore, are of the form ∞ Ψ=()φ imφ rfre,()∑ m . (6.8) m=−∞ For fixed r , the solution is a , which forms a complete function basis for periodic functions in φ . It is easy to show (by direct substitution) that the radial solu- tions are mm− ()=+⎛⎞rr ⎛⎞ ≠ frmm A⎜⎟ B m ⎜⎟ for m 0, ⎝⎠rr00 ⎝⎠ (6.9) =+() = AB00ln rr / 0 for m 0. Here r0 is some convenient scale parameter to allow the coeffi- { } cients Amm, B to all have the same units. The general solution to Laplace’s equation in 2-dimensions is therefore given by − ∞ ⎛⎞mm ⎛⎞rr ⎛⎞ φ Ψ=()φ ( ) +⎜⎟ + im rBrrA,ln/00∑ mm⎜⎟ B ⎜⎟ e. (6.10) =−∞ ⎜⎟rr m ⎝⎠⎝⎠00 ⎝⎠ The series solution can be interpreted as representing a multi- pole expansion of the potential function Ψ ()r,φ . We will return to this solution when we discuss partial differen- tial equations in more detail in the following chapters. In partic- ular, we will need to discuss what type of boundary conditions lead to consistent, sable, and unique solutions to the partial dif- 130 Orthogonal function spaces ferential equation of interest. The product solutions, neverthe- less, illustrates some general features of the separation of varia- ble technique that are worth pointing out at this point: • Separation of variables in partial differential equations natu- rally leads to ordinary differential equations that are solu- tions to an . • The allowed eigenvalues are constrained by the imposed on the equations. • The function basis generated by the eigenvalue equations forms a for the class of functions that satisfy the same boun- dary conditions. • The complete solution to the partial differential equation is a to the eigenvalue equations where the coefficients are chosen to match the physical . 6.3 Helmholtz’s equation Let us next consider a class of second order partial differential equations, which contains some of the most famous named sca- lar equations in physics. These include • Laplace’s equation ∇Ψ2 ()r =0. (6.11) • The wave equation Orthogonal function spaces 131 ⎛⎞∂2 ∇−2 1 Ψ() = ⎜⎟22 r,0.t (6.12) ⎝⎠vt∂ • The diffusion equation ∂ ⎛⎞2 1 ⎜⎟∇− Ψ()r,0.t = (6.13) ⎝⎠λ ∂t • The non-interacting Schrödinger’s equation −∂2 ⎛⎞ 2 ⎜⎟∇−it Ψ()r,0. = (6.14) ⎝⎠2mt∂ These equations differ in their time behavior, but their spatial behavior is essentially identical. They all are linear functions of Laplace’s operator ∇2 . After separating out the time behavior, they lead to a common differential equation, known as Helm- holtz’s equation ()∇+22k Ψ()r =0. (6.15) Or, in some cases, to the modified Helmholtz’s equation ()∇−22k Ψ()r =0. (6.16) Laplace’s equation refers to the special case k 2 = 0. The addition of a local potential term U ()r to Helmholtz’s equation changes the details of the solution, but not the general character of the boundary conditions. The Laplacian operator is special in that it is both translationally and rotationally invariant. It is also invariant under the discrete 132 Orthogonal function spaces symmetries of mirror reflection and parity reversal. As a conse- quence of its high degree of symmetry, separation of variables can be carried out in at least 18 commonly used coordinate frames. Here we will consider only the three most obvious ones, those employing Cartesian, spherical and cylindrical coordi- nates. The functions that result from its decomposition include the best known and studied equations of mathematical physics. Helmholtz’s equation can be written symbolically in the opera- tor form 222++ Ψ=±Ψ 2 ()DDDxyz k. (6.17) From which we see that the operator has an “elliptical” signature 222 ()()()Xa///++ Yb Zc . This signature totally determines the allowed choices of Boundary conditions. An Elliptic Differential Equation has a unique (up to a constant), stable solution if one or the other (but not both) of the following two sets of Boundary conditions are met. • The function is specified everywhere on a closed spatial boundary. (Dirichlet Boundary conditions), or • The derivative of the function is specified everywhere on a closed spatial boundary (Neumann Boundary conditions). Other choices of boundary conditions either under-specify or over-specify the constraints or lead to ambiguous or inconsis- tent results. The depends on the dimensionality of the equation: Orthogonal function spaces 133 • In three-dimensions, the boundary is a closed surface; • In two dimensions, it is an enclosing line; • In one dimension, it is given by the two end points of the line. If the volume to be enclosed is infinite, the enclosing surface is taken as the limit as the radius R →∞ of a very large boundary envelope. 6.4 Sturm-Liouville theory After separation of variables in the Helmholtz Equation, one is left with a set of second order ordinary differential equations having the following linear form: 2 {}()=++⎛⎞()dd() () () L yx⎜⎟ Ax2 Bx Cx yx ⎝⎠dx dx (6.18) = λWxyx() (). LL= ∗ denotes a real-valued, linear second order differential op- erator, A,,BC are real-valued functions of the dependent varia- ble x , λ is an eigenvalue, Wx()is a weight function, and yxλ () is the eigenfunction solution to the eigenvalue equation. This equation is assumed to be valid on a closed inter- val x ⊂ [,]ab. The eigenvalue solutions of the above equation are real if the linear operator can be written in the 134 Orthogonal function spaces ⎛⎞dd⎛⎞ Lyx{}()=+=⎜⎟⎜⎟ Ax() Cx() yx ()λ Wxyx () (). (6.19) ⎝⎠dx⎝⎠ dx This requires the constraint Bx()= A′ () x. (6.20) An equation that can be put in such a form is said to be a Sturm- Liouville differential equation. If the equation is not in a self- adjoint form, an integrating factor can often be found to put it into such a form. In the standard notation for Sturm-Liouville equations the functions A and C are referred to as the func- tions P and Q respectively, so that the equation is often written in the standard form ⎛⎞dd⎛⎞ Lyx{}()=+=⎜⎟⎜⎟ Px() Qx() yx ()λ Wxyx () (). (6.21) ⎝⎠dx⎝⎠ dx The Wx() usually arises from the Jacobean of the transforma- tion encountered in mapping from Cartesian coordinates to some other coordinate system. This weight is required to be pos- itive semi-definite. That is, it is except at a finite number of dis- crete points on the interval x ⊂ [,]ab where it may vanish. It de- fines a norm for a function space such that b NWxyydx=≥∫ ()∗ 0. (6.22) a The class of functions that have a finite norm N are said to be Orthogonal function spaces 135 Linear self-adjoint differential operators A linear differential operator is said to be self-adjoint on interval [ab, ] if it satisfies the following criteria bb y∗L{}y dx = y L{}y* dx (6.23) ∫∫21 12 aa with respect to any normalizable functions yi that meet certain specified boundary conditions at the end points of the interval. Sturm-Liouville differential operators are self-adjoint for Dirich- let, Neumann, and periodic boundary conditions (B.C.). First note that the term Qx()ς ()m where Q is a real-valued function is automatically self-adjoint bb y∗∗Qxydx()= yQxydx () , (6.24) ∫∫21 12 aa since functions commute. Next, integration by parts gives b ∗ ⎛⎞dd⎛⎞ yPxydx() ∫ 21⎜⎟⎜⎟ ⎝⎠dx⎝⎠ dx a (6.25) b b ∗ ∗ d dy⎛⎞ dy =−y Px() y21 Px () dx . 21∫ ⎜⎟ dxa a dx⎝⎠ dx Likewise, 136 Orthogonal function spaces b ⎛⎞dd⎛⎞∗ yPxydx() ∫ 12⎜⎟⎜⎟ ⎝⎠dx⎝⎠ dx a (6.26) b b ∗ d ∗ dy⎛⎞ dy =−y Px() y21 Px () dx . 12∫ ⎜⎟ dxa a dx⎝⎠ dx Subtracting (6.26) from (6.25) gives bb b ∗ b ∗∗dy dy yL{} y dx−=− yL{} y* dx yP12 yP . (6.27) ∫∫21 12 2 1 aa dxa dx a This clearly vanishes if the functions or their derivatives vanish at the limits [ab, ] (i.e., for Dirichlet or Neumann B.C.). It also vanishes if the upper and lower limits have the same value (i.e., for periodic B.C.) and also that Pa()== Pb () 0. (6.28) This latter case occurs for certain types of spherical functions, such as the Legendre polynomials. An important theorem is that the eigenvalues of a self-adjoint differential operator are real. The proof follows from the use of the conjugate of a Sturm- Liouville equation ∗∗∗∗⎛⎞dd⎛⎞ Ly{}() x=+=⎜⎟⎜⎟ Px() Qx() y () xλ Wxy () () x. (6.29) ⎝⎠dx⎝⎠ dx (Note that the operator L and the weight function W are real). Therefore, bb b yL∗∗∗{} y dx−=−= yL{} y* dx (λλ ) Wyydx 0. (6.30) ∫∫21 12 12 ∫ 21 aa a Orthogonal function spaces 137 = Letting y12y gives b ()λλ−=∗∗Wy y dx 0, (6.31) 11∫ 11 a b but the norm Wy∗ y dx 0 unless y ≡ 0 ; therefore, ∫ 11 1 a λλ= ∗ 11. (6.32) Orthogonality The eigenfunctions of different non-degenerate eigenvalues are orthogonal to each other with respect to weight W . The proof follows, from (6.30) b ()λλ−=Wy∗ y dx 0. (6.33) 12∫ 21 a λλ≠ If 12, this implies that b Wy∗ y dx = 0. (6.34) ∫ 21 a Given a set of linearly independent, but degenerate, eigenfunc- ψ tions n with the same eigenvalue, one can always construct a φ “diagonal” basis of eigenfunctions n that are orthogonal to each other with respect to weight W. One procedure, attributed to {φ } Schmidt, is to construct the basis n from the sequence 138 Orthogonal function spaces φψ= 11, φψ=− φ 22211c , φψ=− φ − φ 33311322cc, (6.35) n−1 φψ=− φ nn∑c nmm. m=1 th φ The coefficients of the n term is chosen such that n is ortho- gonal to all previously orthogonalized eigenfunctions. b Wdxφφ∗ =∀0. mn < (6.36) ∫ mn a For example, bb WdxWφφ∗∗=− φ() ψ cdx φ ∫∫12 1 2 211 aa bb =−WdxcWdxφψ∗∗ φφ =0, (6.37) ∫∫12 21 11 aa b Wdxφψ∗ ∫ 12 = a c21 b . Wdxφφ∗ ∫ 11 a When the functions are to be normalized as well as orthogona- lized, this algorithm is referred to as the . Orthogonal function spaces 139 Completeness of the function basis Any piece-wise continuous function, with a finite number of maxima and minima, that is normalizable on an interval [ab, ] and which satisfies the same B.C. as the eigenfunctions of a self- adjoint operator on that interval can be expanded in terms of a complete basis of such eigenfunctions. ()= () f xyx∑cnn (6.38) allλ If the basis is an orthogonal one with normalizations given by b NWxyydx= () ∗ (6.39) nnn∫ a one can invert the problem to solve for the coefficients, giving b 1 ∗ cWxfxydx= () () (6.40) nn∫ Nn a The proof of inversion can be obtained by using the orthogonali- ty condition b Wxy() ∗ ydxN= δ (6.41) ∫ mn nnm a Comparison to Fourier Series In retrospect, we see that our development of Fourier Series is a direct application of Sturm-Liouville Theory. Letting 140 Orthogonal function spaces Px()== Wx () 1; Qx () = 0; λ =− m2 (6.42) in the Sturm-Liouville equation and assuming periodic B.C. on interval [−ππ, ] gives ⎧⎫∂2 +Φ2 ()φ = ⎨⎬2 m m 0 (6.43) ⎩⎭∂φ having eigenfunctions emimφ , for integer (6.44) in terms of which, we can expand any piece-wise continuous, normalizable, periodic function as an infinite series ∞ = imφ f ()xce∑ n (6.45) m=−∞ By orthogonality, we can solve for the coefficients giving π 1 − φ cfed= ()φφim (6.46) n π ∫ 2 −π This generalization would be a lot of work to go through just to solve a single eigenvalue equation, but it saves time in the long run. We no longer need to prove reality of the eigenvalues, or- thogonality of the eigenfunctions, and completeness of the func- tion basis in an ad-hoc manner for every eigenvalue equation that we encounter. Discussion Problem: Generalization of Parseval’s theorem Show that Parseval’s theorem can be generalized to the form Orthogonal function spaces 141 b Wx() fx ()22 dx= N c . (6.47) ∫ ∑ nn a n And that the average of the norm of f ()x 2 with respect to weight W can be written as b ∫Wx() fx ()2 dx ∗ N 2 f fc==an, (6.48) b ∑ N n ∫Wxdx() n W a where b NWxdx= () . (6.49) W ∫ a = If NNnW/1, for all eigenfunctions, the eigenfunctions are said to be normalized. The norm of a function can then be written as ∗ = 2 f fc∑ n . (6.50) n Convergence of a Sturm-Liouville series Although we will not formally prove completeness here, it is use- ful to define the sense in which we mean that the function and its series expansion are equal. Partitioning the series into a finite partial sum SN of N terms and an infinite remainder RN , then the function and the series are the same in the sense that the norm of the remainder tends to zero as N →∞ 142 Orthogonal function spaces bb ()22=−=() () limWxR ( )Nn x dx lim Wx ( ) f x S x dx 0. (6.51) NN→∞∫∫ →∞ aa In plain English, this means that the series converges to the function wherever the function is continuous and the Weight function non-zero, and the function differs from the series only at a finite number of discrete points (i.e., only on intervals of null measure). This of course is essentially the same criteria that we applied to the convergence of Fourier Series. Vector space representation Those familiar with quantum mechanics will recognize that a is just a special case of a . Hermitian operators have real eigenva- lues and form complete function spaces. By treating the basis functions as representing independent degrees of freedom, one can define an infinite-dimensional vector space with coeffi- { } cients cn . This is, in fact, the classical origins of Hilbert space, which preceded the development of quantum mechanics. Let us expand functions with respect to a normalized eigenfunction ba- sis: b Wx()φφ∗ dx ∫ nm φφ∗ ==a δ nmb nm. (6.52) ∫Wxdx() a Orthogonal function spaces 143 Each eigenfunction can be thought of as defining an indepen- dent degree of freedom of the system, one which projects out an orthogonal state with normalization = δ nm nm. (6.53) An arbitrary state in this infinite dimensional space can be writ- ten as ⎡⎤c1 fcnc==⎢⎥, ∑ n ⎢⎥2 n (6.54) ⎣⎦⎢⎥ † ==() ∗∗∗ =⎡ ⎤ ff∑ ncccn ⎣ 12...⎦ , n where the basis vectors define projection operators for the am- plitude coefficients ==∗ nf cnn,, fn c (6.55) and the norm of an arbitrary vector is given by = 2 f fc∑ n . (6.56) n The connection between the and the is given by relating the in- ner product of the vector space with the weighted integral over the overlap of two functions. b ∫WxA()∗ () xBxdx () ===∗∗a AB∑ abnn AB b . (6.57) n ∫Wxdx() a 144 Orthogonal function spaces This last expression is recognizable as the extension of Parsev- al’s theorem applied to an arbitrary set of normalized orthogon- al functions. The interpretation of the meaning of the norm of a Hilbert space depends on the physical context. In quantum mechanics, the normalization has a probabilistic interpretation, and the total single particle wave function is normalized to unit probability. In classical wave mechanics, the square of the wave amplitude can be related to its energy density. Parseval’s theorem can then be interpreted to state that normal (orthogonal) modes of oscil- lation contribute independently to the energy integral. The total energy of a wave is the incoherent sum of the energies of each normal mode of oscillation. 7. Spherical Harmonics Consider a spherical cow —Introductory physics problem example We live on the surface of a sphere. The stars and planets are ap- proximate spheroids, as are atomic nuclei, at the other limit of the size scale. The three-dimensional character of space, along with the assumption that one is dealing with localized sources, leads one naturally into considering using spherical coordinates. Various irreducible classes of rotational symmetries arise out of the rotational invariance of three-dimensional space. It is no surprise, then, that the spherical functions are ubiquitous in ma- thematical physics. Of these, the most important are the spheri- ()θ φ cal harmonics Ylm , , which represent the angular eigenfunc- tions of Laplace’s operator in a spherical basis. These, in turn, are products of the Fourier series expansion eimφ for the cyclic azimulthal angle dependence and the associated Legendre poly- ()θ nomial eigenfunction expansion Plm cos for polar angle beha- vior. 146 Spherical Harmonics 7.1 Legendre polynomials The and their close cousins, the , arise in the solution for the po- lar angle dependence in problems involving spherical coordi- nates. The Legendre polynomials deal with the specific case where the solution is azimuthally symmetric; the associated Le- gendre polynomials deal with the general case. After separation of variables in the Helmholtz equation, using spherical coordi- nates ()r,,θ φ and assuming no φ dependence, one is left with the following differential equation dd ()1()1().−=−+x2 yx ll() yx (7.1) dx dx ll Here x = cosθ so the domain of the equation is the interval −≤11x ≤. We recognize this equation as being a Sturm-Liouville equation with Px()=− 1 x2 , Qx()= 0, Wx()= 1, having real ei- genvalues λ =−ll(1) + . Because P(1)0±= at the end points of the interval, any piecewise-continuous, normalizable function can be expanded in a Legendre’s series in the interval [1,1]− . By substituting x = cosθ , Legendre’s equation can be written in the form 1 dd sinθ ylly=−() + 1 . (7.2) sinθθdd θll Spherical Harmonics 147 Discussion Problem: Origin of the spherical harmonics and the associated Legendre equation: Starting with Helmholtz’s equation in spherical coordinates (see Figure 7-1 for a sketch of the coordinate system) ⎪⎧ dd22211⎡ d d 1 d⎤⎪⎫ ++sinθ + ⎨ 22⎢ θθ θ 22 θφ⎥⎬ ⎪⎩⎭dr r dr r⎣sin d d sin d ⎦⎪ (7.3) =−kr2 Ψ(,θφ , ) show that separation of variables leads to the angular equation 2 ⎧⎫11ddθθ+=−+ d()φ ()θ φ ⎨⎬sin22YllY , ( 1) , . (7.4) ⎩⎭sinθθdd θ sin θ d φ (You don’t need to solve the radial part to show this). Show by further separation of variables that ()θφ ∝ imφ YPxe,()lm , (7.5) where Pxlm () are the Associated Legendre polynomials given by dd m2 ()1()()1().−−=−+x2 Px Px ll() Px (7.6) dx dxlm1− x2 lm lm The ordinary Legendre polynomials are related to the associated = Legendre polynomials by Pxll() P0 () x 148 Spherical Harmonics Figure 7-1 A spherical coordinate system Series expansion Laplace’s differential operator is an even function of x . There- fore, for every l , there will be two linearly-independent solu- tions to the eigenvalue equation that can be separated into even and odd functions. It will turn out that only one of these series will converge for the allowed values of l . Let us rewrite the equ- ation, putting terms that couple to the same power of x on the right-hand side, y′′=+−+xy2 ′′2(1) xy ′ ll y. (7.7) = n Substituting y ∑ axn gives the series expansion n Spherical Harmonics 149 ∞∞∞∞ −=−+nnnn−2 −+ ∑∑∑∑nn( 1) axnnnn nn ( 1) ax 2 nax ll ( 1) ax (7.8) nnnn====2000 or ∞∞ ()++nn =[] +−+ ∑∑nnaxnnllax2 ( 1)nn+2 ( 1) ( 1) , (7.9) nn==00 giving the recursion relation [nn(1)(1)+− ll +] aa+ = , (7.10) nn2 ()nn++2( 1) which decouples even and odd powers of x . We can test the series to determine its radius of convergence, giving a ()nn++2( 1) x2 <=limn lim = 1 (7.11) nn→∞ →∞ []+− + annlln+2 (1)(1) Therefore the range is the open interval (1,1)−+ . However, the convergence of the series at the end points is still in doubt. A more careful analysis shows that the ratio rn approaches 1 from above for large n , and it turns out the series diverges at the end points x =±1. This appears to be a disaster, if one fails to ob- serve that the series terminates for integer values of l . More specifically, the even series terminates for even l , and the odd = series terminates for odd l . When nl, the coefficient an+2 and all further terms in the series vanish, see Eq. (7.10). Therefore, the boundary conditions at x2 =1 are satisfied by setting 150 Spherical Harmonics l = 0,1,2 . (7.12) The solutions that converge at the end points of the interval are finite polynomials of order l , called the Legendre polynomials, which have an even or odd reflection symmetry given by =−()l − Pxll() 1 P ( x ) (7.13) For historic reasons they are normalized to 1 at x =1 = Pl (1) 1 . (7.14) Figure 7-2 shows a plot of the first six Legendre polynomials. By direct substitution in the recursion relation (7.10) and using the normalization constraint (7.14), the first few polynomials can be written as = P0 1, = Px1 , =−1 2 (7.15) Px2 2 ()31, =−1 3 Pxx3 2 ()53. You should verify these expressions for yourselves. Let’s calcu- late Px2 () as an example. There are two nonzero terms in the expansion, aa02& . They are related by [−+ll(1)] −6 aaaa===−3. (7.16) 2000()2(1) 2 Therefore, Spherical Harmonics 151 ()=−2 + Px20()31, x a ()= =− Pa2011⇒ 2, (7.17) ∴=() 1 2 − Px2 2 ()31. x Note that a Legendre polynomial of order n is a power series in x of the same order n . The Legendre polynomials are bounded by ≤ Pxl () 1. (7.18) This can be useful in estimating errors in series expansion. A useful formula is ⎧ 0 for odd l ()= ⎪ Pl 0 ⎨ l /2 ()l −1!! (7.19) ⎪()−1 for even l ⎩ l!! Orthogonality and Normalization Since Legendre’s equation is a Sturm-Liouville equation, we don’t have to prove orthogonality, it follows automatically. The norm of the square-integral is given by 1 2 PPdx = δ . (7.20) ∫ ll′′+ ll −1 21l The proof will be left to a discussion problem. A Legendre series is a series of Legendre polynomials given by 152 Spherical Harmonics ∞ ()=≤ () fx∑ aPxnn ,1. x (7.21) n=0 By orthogonality, the series can be inverted to extract the coeffi- cients 21n + 1 afxPxdx= () () . (7.22) n ∫ l 2 −1 A polynomial of order N can be expanded in a Legendre series of order N : NN m = ∑∑bxmnn aP(). x (7.23) mn==00 The proof follows from the linear independence of the Legendre polynomials. Since a Legendre series of order N is a polynomial of order N , the above expression leads to N +1 linear equations relating the an and bm coefficients. By linear independence, the equations have a non-trivial solution. Since a Legendre series expansion is unique, the solution obtained is the only possible solution. Solving for an by brute force we get 21n + 1 N aPbxdx= m . (7.24) nnm∫ ∑ 2 −1 m=0 Spherical Harmonics 153 Example: Expand the quadratic equation ax2 ++ bx c in a Le- gendre series: 22++=++1 − ax bx c a01 a x a 22 ()31 x 31 (7.25) =++−ax2 ax a a. 222102 Therefore, 21 aaabaca==, , and =+ . (7.26) 2133 0 Discussion Problem: A spherical capacitor consists of two conducting hemispheres of radius r. The top hemisphere is held at positive voltage and the bottom hemisphere is held at nega- tive voltage. The potential distribution is azimuthally symmetric and is given by ⎧ +V for 1>x>0 Vx()= o (7.27) ⎨− ⎩ V0 for 0>x>-1 Calculate the Legendre series for this potential distribution. 154 Spherical Harmonics 1 Leg() 0, x Leg() 1, x 0.5 Leg() 2, x Leg() 3, x 0 Leg() 4, x Leg() 5, x 0.5 1 1 0.5 0 0.5 1 x Figure 7-2 Legendre Polynomials A second solution The second solution to Legendre’s equation for integer l is an infinite series that diverges on the z-axis, where x ==±cosθ 1 (Figure 7-3). Although not as frequently seen, it is permitted for problems with a line-charge distribution along the z-axis. The solutions are labeled Qxl () and have the opposite symmetry to the Pxl (), =−()l+1 − Qxll() 1 Q ( x ). (7.28) A closed form solution for Qxl () can be found by substituting Spherical Harmonics 155 + =+11⎛⎞x Qxll() Px ()ln⎜⎟ Bx l () (7.29) 21⎝⎠− x into Legendre’s equation, where Bl ()x is a second polynomial to be solved for. The first few terms are tabulated below: + = 11⎛⎞x Q0 ln⎜⎟ , 21⎝⎠− x + =−xx⎛⎞1 Q0 ln⎜⎟ 1, 21⎝⎠− x (7.30) 2 −+ =−311xxx⎛⎞ 3 Q0 ln⎜⎟ , 41⎝⎠− x 2 32−+ =−+15xx 3⎛⎞ 1 x 5 x 2 Q0 ln⎜⎟ . 41⎝⎠− x 23 2.8 2.4 , , Qlm() 0 0 x 2 Qlm() 1, 0, x 1.6 1.2 , , Qlm() 2 0 x 0.8 Qlm() 3, 0, x 0.4 0 , , Qlm() 4 0 x 0.4 Qlm() 5, 0, x 0.8 1.2 1.6 2 0 0.2 0.4 0.6 0.8 x Figure 7-3 Legendre’s polynomials of the second kind Legendre polynomials are a good starting point for the study of orthogonal functions, because a number of its properties can be generalized to other orthogonal functions. There exists a diffe- 156 Spherical Harmonics rential form called Rodriquez formula that can be used to gener- ate the polynomials. There is a generating function that serves the same purpose. Finally, there are recursion relations connect- ing Legendre polynomials to each other. Once one sees how these various identities apply for Legendre’s polynomials, one can easily accept the existence of other such formulae for other orthogonal functions at face value, and apply them in a similar manner. 7.2 Rodriquez’s formula Rodriquez’s formula for Legendre polynomials is given by l 1 d l Px()=−() x2 1 (7.31) l 2!llldx The proof uses Leibniz’s rule for differentiating products. Leibniz’s rule for differentiating products The nth derivative of a product of two terms is given by the bi- nomial expansion n − ⎛⎞ddUxdVxn ⎛⎞n nm() m () UxVx() ()= . ⎜⎟ ∑⎜⎟ nm− m (7.32) ⎝⎠dxm=0 ⎝⎠m dx dx Proof: Let D = ddx/ . Then, Spherical Harmonics 157 ()()=+=+ ()( ) D UV DU V U DV Duv D UV , (7.33) where Du denotes the derivative’s action on the function U , and Dv denotes the derivative’s action on the function V . Then n n ⎛⎞⎛⎞n − nnmm()(=+ )() = DUV Duv D UV⎜⎟∑ ⎜⎟ D uv D UV (7.34) ⎝⎠m=0 ⎝⎠m or nn ⎛⎞nn−− ⎛⎞ nnmmnmm()==()() ()() DUV∑∑⎜⎟ DUDVuv ⎜⎟ DUDV. (7.35) mm==00⎝⎠mm ⎝⎠ The proof that Rodriquez’s formula is correct involves • () Showing that Pxl is a solution to Legendre’s equation, and P ()11= • Showing that l . l To prove the first part let vx=−()2 1 , then dv l−1 ()x222−=12lx()() x −−= 112. x lxv (7.36) dx Differentiating this expression l +1 times by Leibniz’s rule gives ll++12⎛⎞dv 1 Dx⎜⎟()−=12 Dlxv()⇒ ⎝⎠dx +++ ⎛⎞lll11122lll++ ⎛⎞ 1 ⎛⎞ ⎜⎟()x −+122Dv ⎜⎟ xDv + ⎜⎟ Dv (7.37) ⎝⎠012 ⎝⎠ ⎝⎠ ++ ⎛⎞ll11ll+1 ⎛⎞ =+lxDvlDv⎜⎟22, ⎜⎟ ⎝⎠01 ⎝⎠ 158 Spherical Harmonics where ⎛⎞ll++11 ⎛⎞ ⎛⎞ l + 1ll()+1 ⎜⎟==+=1, ⎜⎟l 1, ⎜⎟ (7.38) ⎝⎠01 ⎝⎠ ⎝⎠ 22 which gives ddll2 (1)+ ()2 −+++()lll() () () x 12122 Dv l x Dv Dv dx dx 2 (7.39) d =++212.lx() Dvll l() l Dv dx Simplifying and changing signs dd2 −−()x2 12() Dvll − x() Dv ++ ll (1)0, () Dv l = dx2 dx (7.40) ⎛⎞dd2 l ⎜⎟()1(1)0.−++=xllDv() ⎝⎠dx dx This is the Legendre equation, which completes the proof of the first part. The second part of the proof involves factoring xxx2 −=111()() + − and applying Leibniz’s rule to the product, then setting the result to x =1. Only one term in the product survives: 1 ll Px() =+()() x1 Dl x −+ 1 terms of order () x − 1 l x=1 2!l l x=1 (7.41) 1 l 0 =−=()2!lx ( 1 ) 1. 2!l l Spherical Harmonics 159 Example: Calculate Px2 () from Rodriquez’s formula: 2 11dd2 Px()=−=−() x22141⎡ xx()⎤ 2 22 ⎣ ⎦ 22!dx 8 dx (7.42) 11d =−=−()xx32()31. x 22dx xm lm> Example: Show that is orthogonal to Pl if : Proof: One possible proof is to use Rodriquez’s formula and integration by parts. Direct integration is easier. Expanding xm in a Legendre series gives 11ml< Pxm dx== P a P dx 0, (7.43) ∫∫llmm∑ ′′ −−11m'0− since mlm''≠∀ . 7.3 Generating function Legendre polynomials can also be obtained by a Taylor’s series expansion of the generating function ∞ −1/2 Φ=−+=()()2 l x,12hxhhPh∑ l . (7.44) l=0 Note that, for x =1, we get 1 ∞∞ Φ==()1,hhPhll = (1) . (7.45) − ∑∑l 1 h ll==00 160 Spherical Harmonics = This will turn out to be related to why the normalization Pl (1) 1 was originally chosen. Before proving (7.44) it is useful to consider its physical origin. Consider a unit charge located at a point r0 , as shown in Figure 7-1. Its electrostatic potential is given by qq V (,rr )== , (7.46) 0 4π rr− π 22−+ 0 42rxrrr00 where x is the cosine of the angle between r andr0 Let = () rGreaterrr> ,,o = () rLesserrr< ,,o (7.47) r h =≤< 1. r> Then, l qq∞ ⎛⎞r Vrr(, )=Φ() xh , = Px ()< . (7.48) 0 ππ∑ l ⎜⎟ 44rrr>>>l=0 ⎝⎠ At large distances rr 0 ,The distribution approaches a pure 1/r potential. The generating function is the multipole expan- sion that arises from the fact that we didn’t think to place the = charge at the origin. The normalization of Pl (1) 1 was used to give all the angular moments equal weight at h =1. Now let’s turn to the proof of equation (7.44). Assume that the RHS of this equation is the definition of Φ . We want to multiply Φ by Legendre’s operator Spherical Harmonics 161 dd Lx=−()1 2 . (7.49) dx dx This gives ∞∞ ∂2 LLPhllPhhhΦ=()ll = −(1) + =− Φ (7.50) ∑∑ll∂ 2 ll==00 h This results in a second order partial differential equation ⎛⎞∂∂∂2 ()−+2 Φ=() ⎜⎟1,0xhhxh2 . (7.51) ⎝⎠∂∂∂xxh Now it is just a matter of substituting the LHS of equation (7.44) into the partial differential equation to verity that the PDE has − Φ=−+() 2 1/2 the closed form solution: xh,12⎣⎡ xh h⎦⎤ . This last step is straightforward, and is left as an exercise for the reader. z V(r,r0) r-r0 q θ r r0 y φ x 162 Spherical Harmonics Figure 7-4 Sketch of a point charge located on the z axis The generating function is useful for proving a number of recur- sion relations relating the Legendre Polynomials and their de- rivatives. 7.4 Recursion relations Just like there are a number of trig identities that are useful to keep at hand, so, too, there are a number of identities relating Legendre polynomials. The first relates Pl to PPll−−12& =−() −− () lPlll21 l xP−−12 l 1 P (7.52) = = Since we know that P0 1 and Px1 , this relationship can be used recursively to generate all the other Legendre polynomials from the first two in the sequence. Unlike Rodriquez’s Formula or the Generating Function, this doesn’t require taking any de- rivatives. Other useful recursion formulas are ′′−= xPPll−1 lP l, (7.53) ′′−= PxPlPll−−11 l, (7.54) −=−2 ′ ()1,x PlPlxPll−1 l (7.55) and ()+=−′′ 21lPPPll+−11 l . (7.56) Spherical Harmonics 163 Given two recursion relations, the others follow by substitution, so it is sufficient to prove the first two equations (7.52) and (7.53) are valid identities. The first relation Eq. (7.52) can be proven by taking partial de- rivatives of the generating function. Taking the derivative with respect to h gives ∂− Φ=xh = l−1 3/2 lPhl , ∂ 2 ∑ h ()12−+xh h l ()−Φ=−() +21l− xh12 xhh∑ lPhl , (7.57) l ()−=−+ll()21− x h∑∑ Phll l12 xh h Ph , ll−=+−11 l − l + l + 1 ∑∑xPhll Ph ∑ lPh l ∑2, lxPh l ∑ lPh l Collecting terms of the same power of h , one gets the first recur- sion formula (7.52): ()+−+−lll ()+−11 = ∑∑∑21l xPhlll l 1 Ph lPh 0, ()−−−−=lll−−−111 () ∑∑∑21lxPhlll−−12 lPh 1 lPh 0, (7.58) ⎡⎤()−−−−= () l−1 ∑ ⎣⎦21lxPlPlPhlll−−12 1 0, ∴=() − −− () lPlll21 l xP−−12 l 1 P . In a similar manner, the solution for the second recursion rela- tion, proceeds by taking the differential wrt x in the definition of the generating function: 164 Spherical Harmonics ∂ Φ=h = l ′ 3/2 ∑ hPl , ∂x ()12−+xh h2 hxhhhΦ=()12 − + 2 ∑ l , (7.59) Phll+12=−()12 xh + h h P′ ∑∑l l =−′′′lll++12 + ∑ Phlll2. xPh Ph Comparing coefficients of order hl+1 gives =−′′′ + PPll+−112. xPP ll (7.60) Now differentiate the first recursion relation to get ′′′ lP=−()21 l P−−− + x () 21 l − P −− () l 1 P or ll112 ll (7.61) ()+=+++−′′′ ( ) ( ) lP12121.llll+−11 lPxlPlP ′ Eliminating the Pl+1 terms in equations (7.60) and (7.61) gives the desired result for the second recursion formula (7.53): =−′′ lPlll xP P−1. (7.62) 1 2 = 2 Discussion Problem: Prove that Pdxl . using the re- ∫−1 21l + =−′′ cursion relation lPlll xP P−1. 7.5 Associated Legendre Polynomials Legendre polynomials represent the convergent solutions of the special case m = 0 of the associated Legendre Equation: Spherical Harmonics 165 dd m2 ()1()()1(),−−=−+x2 Px Px ll() Px (7.63) dx dxlm1− x2 lm lm where ()= () Pxll P0 x, (7.64) From the symmetry of the equation one sees that the substitu- tion ±m leads to the same equation, therefore, ∝ PPlm l− m , (7.65) Unfortunately, because of how these polynomials where origi- nally defined, they turn out not to be simply equal to each other, they differ in their norms. They also vary in sign conventions from text to text. Like the Legendre Polynomials, the associated Legendre func- tions are solutions to an eigenvalue equation of the Sturm Liou- ville form ⎛⎞dd⎛⎞ Lyx{}()=+=⎜⎟⎜⎟ Px() Qx() yx ()λ Wxyx () (), (7.66) ⎝⎠dx⎝⎠ dx which differs from the Legendre equation by the addition of a function of x : m2 Qx()= , (7.67) 1− x2 Therefore, for fixed azimulthal index m , the Plm also represent eigenvalue functions of −+ll(1). For positive integer 0 ≤≤ml the solutions are given by 166 Spherical Harmonics m m/2 d Px()=−() 1 x2 Px (), (7.68) lmdxm l which can be verified by direct substitution into the Associated Legendre equation. Substituting Rodriguez’s formula for Pxl (), gives the more general form, lm+ ml/2 1 d Px()=−() 1 x22() x − 1. (7.69) lm 2!llmldx+ This is the generalized form of Rodriquez’s formula. In this form, it can be applied to both positive and negative values of m . This is in fact how the associated Legendre are defined, and gives their normalization up to a sign convention of (1)− m em- ployed in some textbooks. Using this formula as it stands, one finds that the positive and negative m solutions are related by ⎛⎞()lm− ! Px()=− (1)m Px (), (7.70) lm− ⎜⎟+ lm ⎝⎠()lm! and one sees that the solutions for negative m are simply pro- portional to those of positive m . From formula (7.68) one finds that the stretched configuration of Pll is proportional to ∝−2 m/2 = l θ Pxll ()() 1 x sin. (7.71) Example: Calculate the Legendre polynomials for l =1 == Use Px110() P () x x, one needs to calculate only P11± . Use equa- tion (7.68) to calculate P11 Spherical Harmonics 167 l 1 m/2dd 1/2 Px=−()1()1()22 Pxx =−() Px 11dxl l dx1 1 (7.72) 1 1/2d 1/2 =−()11sin.xxx22 =−() = θ dx1 The negative values for m can be found from equation (7.70) ⎛⎞()lm− ! Px()=− (1)m Px (), 11− ⎜⎟+ lm ⎝⎠()lm! ⎛⎞()11!− =−(1)1 Px (), (7.73) ⎜⎟+ 11 ⎝⎠()11! −−111/2 =−()1sin,x2 = θ 22 There are ()21l + m−states associated with a given value of l . − The solution for the 3 m states of P1m are =−=2 θ Px11 1sin, == θ Px11 cos , (7.74) −− 112 Px− =−=1sin.θ 11 22 This illustrates one of the problems with using the Associated Legendre Polynomials. Because of the—too clever by far— substitution into Rodriquez’s formula, the normalization of the positive and negative m states differ. For this reason, it is better to work with the spherical harmonics directly for cases where m ≠ 0. 168 Spherical Harmonics Normalization of Associated Legendre polynomials The normalization of the Associated Legendre Polynomials are given by 1 2 ()lm+ ! dxP() x P () x = δ . (7.75) ∫ lm l′′ m+−() ll −1 21llm ! Therefore a series expansion of a function of x for fixed m takes the form ∞ = fxmlmlm()∑ APx (), l=0 (7.76) 21l + ()lm− ! 1 AfxPxdx= () () . lm()+ ∫ m lm 2!lm−1 Parity of the Associated Legendre polynomials Knowing the parity of the Associated Legendre Polynomials is useful. Reflection symmetry can often be used to identify terms that identically vanish, reducing computational effort. The pari- ty of a Legendre Polynomial of order ()lm, is ()()−=−lm+ () Pxlm1. Px lm (7.77) Another useful result is ()±= ≠ Pmlm 1 0, for 0. (7.78) Spherical Harmonics 169 Recursion relations There are a significant number of recursion relations for the As- sociated Legendre Polynomials. Here are a couple of examples: • For fixed l : 21()mx+ PPlmlmP++−+−++=()(1)0. (7.79) lm,21− x2 lm ,1 lm , • For fixed m : ()+− − + + + = lmPlxPlmP1(21)()0.lm+−1, lm , lm 1, (7.80) This latter relation reduces to equation (7.52) when m = 0. 7.6 Spherical Harmonics Legendre Polynomials do not appear in isolation. They represent the polar angle solutions to a spherical problem which also has azimulthal dependence. In particular, they are the solu- tions to the following angular equation which occurs when one separates Laplace’s equation in spherical coordinates. ∂∂22 ∂ ⎧⎫11θθφ+++=() () ⎨⎬sin222ll 1 Ylm , 0. (7.81) ⎩⎭sinθθ∂∂ θ sin θφ ∂ The operator is closely related to the square of the orbital angu- lar momentum operator in quantum mechanics: 170 Spherical Harmonics ⎧⎫11∂∂22 ∂ 22()θ φ =−θθ + + ()φ LYlm ,sin, ⎨⎬222 Ylm ⎩⎭sinθθ∂∂ θ sin θφ ∂ (7.82) =+()2 ()θφ ll1,. Ylm When there is no azimulthal symmetry, ( m ≠ 0 ), it is usually bet- ()θ φ ter to work directly with the product solutions Ylm , , which are called the spherical harmonics. The spherical harmonics are products of the associated Legendre Polynomials and the com- plex Fourier series expansion of the periodic azimulthal eigens- tates. They have the advantages of having a simple normaliza- tion: YYd* ()()θφ,, θφΩ= δδ , ∫ lm l′′ m ll ′ mm ′ 1 π (7.83) where ddxdΩ=φπ =4 . ∫∫∫−−1 π The normalization of a complex Fourier series is given by +π −imφφ im′ deφ e = 2.πδ ′ (7.84) ∫−π mm while the normalization of the associated Legendre polynomials is given by equation (7.75). Putting the two together and one finds m 21l + ()lm− ! φ YPe()()θφ,1=− () cos, θ im (7.85) lm4!π ()lm+ lm where ()−1 m is a commonly used phase convention. Because dif- ferent phase conventions are in common use, one has to be care- ful in using the spherical harmonics in a consistent manner. Spherical Harmonics 171 This is true as well for the Associated Legendre Polynomials in general. Any piecewise continuous function defined on a sphere f ()θ,φ can be expressed as sums over the spherical harmonics ∞=+ml ()θ φ = ()θ φ fCY,,,∑∑ lm lm (7.86) lml==−0 which, by orthogonality, gives C =Ωf ()()θφ,,.Yd∗ θφ (7.87) lm ∫ lm If f ()θ,φ is real, the spherical harmonics occur in complex con- jugate pairs. Calculating the spherical harmonics is not much more compli- cated than calculating the associated Legendre Polynomials. There are 21l + m -states for every irreducible spherical har- monic tensor of rank l . For l = 0, this reduces to a single spheri- cally symmetric state 1 Y = . (7.88) 00 4π where it is easy to verify that 2 YdΩ=1. (7.89) ∫ 00 For l =1, equation (7.85) reduces to 172 Spherical Harmonics 3 φ Ye=− sinθ i , 11 8π 3 Y = cosθ , (7.90) 10 4π 3 −iφ Ye− = sinθ . 11 8π Some useful formulas are =−()m ∗ YYlm,,− 1 lm and (7.91) l 2 21l + Y ()θφ,.= (7.92) ∑ lm, π ml=− 4 The completeness relation for spherical harmonics is given by ∞ l * ()()θ′′φ θ φ =−δθ θδ ′φφ − ′ ∑∑YYlm,,,,(coscos)(), lm (7.93) lml==−0 and the multipole expansion of a point charge gives 14∞ l π r l = < YY* ()()θ′′,,.φ θ φ (7.94) −+′ ∑∑ l+1 lm,, lm rr lml==−0 21lr> This latter equation is a generalization of the generating func- tion (7.48) to the case where the point charge is not restricted to be along the z-axis. 7.7 Laplace equation in spherical coordinates Laplace’s equation in spherical coordinates can be written as Spherical Harmonics 173 ⎛⎞1 ∂∂ r 2 + ⎜⎟rr2 ∂∂ r ∇Ψ2 ()rr =⎜⎟ Ψ() ⎜⎟11⎛⎞∂∂22 1 ∂ (7.95) sinθ + ⎜⎟2222⎜⎟θθ∂∂ θ θφ ∂ ⎝⎠r ⎝⎠sin sin = 0, Ψ=() () (θ φ ) which has product solutions of the form r frYllm, , , yielding the radial equation + ⎛⎞1(1)ddll2 −= ⎜⎟22rfrl () 0 or rdrdr r ⎝⎠ (7.96) dd rfrllfr2 ()=+ ( 1)(). dr dr ll = λ λ =−+ Letting fl ()rr gives ll,( 1) leading to the solutions ll−+(1) =+⎛⎞rr ⎛⎞ frll() A⎜⎟ B l ⎜⎟ , (7.97) ⎝⎠rr00 ⎝⎠ Where r0 is a scale parameter chosen such that the coefficients A and B all have the same units. The general solution to Lap- lace’s equation in spherical coordinates is then given by a sum over all product solutions l + ∞ l ⎡⎤⎛⎞r r (1)l Ψ=() +⎛⎞0 ()θ φ r ∑∑⎢⎥ABlm⎜⎟ lm⎜⎟ Y lm ,. (7.98) ==− rr lml0 ⎣⎦⎢⎥⎝⎠0 ⎝⎠ where the Alm coefficients are valid for the interior solution, → which includes the origin r 0 , and the Blm coefficients are va- lid for the exterior solution, which includes the point at infinity ( r →∞). 174 Spherical Harmonics 8. Bessel functions 8.1 Series solution of Bessel’s equation is a solution for the radial part of the Helmholtz equation in cy- lindrical coordinates. The equation can be written as ⎛⎞22 ddm+−1 () =−2 () ≤≤∞ ⎜⎟22yr kyr,0 r . (8.1) ⎝⎠dr r dr r Letting kk22→− would give us the . m is an integer for cylin- drical problems, but the equation is also useful for other cases so we will replace m with the arbitrary real number p in what fol- = lows, and yr() with Jxpp() Jkr ( ). Like the sine and cosine functions in the expansion for Fourier series, the eigenvalue k can be scaled away by setting x = kr , and the equation can be written in the standard form dd ⎛⎞−+22 () = ⎜⎟xx pxJxp 0 , (8.2) ⎝⎠dx dx which can be expressed in the explicitly self-adjoint form 2 ⎛⎞ddp−+ = ≤≤∞ ⎜⎟xxJxp 00 , (8.3) ⎝⎠dx dx x where 176 Bessel functions Px()= x , Wx()= x , p2 (8.4) Qx()=− ; x λ =1, This equation can be solved by the . Noting that the operator in equation (8.2) is an even function of x . Let’s try a generalized power series solution of the form ∞ 2ns+ ()= ⎛⎞x Jxpn∑ a⎜⎟ , (8.5) n=0 ⎝⎠2 where the factor of 2−+(2ns ) was arbitrarily inserted to simplify the normalization of the final answer. ao is the first non- vanishing term in the series. Regroup the equation to put terms with the same power of x on the same side of the equation ⎛⎞dd−=−22() () ⎜⎟x xpJxxJxpp (8.6) ⎝⎠dx dx and substitute in the generalized power series ∞∞22ns++ ns ⎛⎞⎛⎞⎛⎞dd−=−22 x x ⎜⎟⎜⎟⎜⎟xx p∑∑ ann x a , (8.7) ⎝⎠⎝⎠⎝⎠dx dx nn==0022 where expansion gives ∞∞222ns+++ ns ()+−2 2 ⎛⎞xx =− ⎛⎞ ∑∑()24ns pann⎜⎟ a ⎜⎟ nn==00⎝⎠22 ⎝⎠ (8.8) ∞ 2'ns+ =− ⎛⎞x 4.∑ an−1 ⎜⎟ n'1=− ⎝⎠2 Bessel functions 177 Comparing coefficients of the same power of x yields the recur- sion formula ()()2444ns+−2 pa22 =⎡⎤ n + nss +−() 22 p a =− a (8.9) nnn⎣⎦−1 Subject to the constraint that the a−1 term must vanish 22−=− ()spa014 a− . (8.10) This gives the indicial equation sp22= (8.11) or sp=± . (8.12) In this case Δ=sp2 , so if p is integer, or half-integer, there is a possibility that the two series won’t be linearly independent. (This in fact is what happens for integer pm= , but we are get- ting ahead of ourselves.) Substituting into equation (8.9) ()±=− nn pann a−1 (8.13) or ()−1 n aa= . (8.14) n nn!!()± p 0 For noninteger m , this can be written as ()−1 n aa= . (8.15) n Γ+Γ±+(1)nnp() 10 178 Bessel functions = So, if we choice to normalize to a0 1, we have the solutions ± n 2np ∞ ()−1 ⎛⎞x Jx()= (8.16) ± p ∑ Γ+Γ±+⎜⎟ n=0 (1)nnp() 12⎝⎠ This is the series solution for Bessel’s equation. In general, the series expansion for Bessel functions converges on the open in- terval ()0,∞ . However, Γ+()p 1 is infinite for negative integers p , so that, for integer p = m , the two series are not linearly independent. =−()m Jx−mm() 1 Jx (). (8.17) Neumann or Weber functions In the case of Bessel’s equation, a special technique is used to find a second linearly-independent solution. These are referred to variously in the literature as N p or Yp functions: cos()()π p Jx− J− () x Nx()== Yx() pp. (8.18) pp sin ()π p For noninteger p , N p and J p are linearly independent since → J ± p are linearly independent. As pmint one has a nonvanish- ing indefinite form to evaluate, which provides the second solu- tion to Bessel’s equation for integer pm= . Using L’Hospital’s rule, the Neumann functions for integer m can be written as Bessel functions 179 2 Nx()=+() ln(/2) xγ Jx () mmπ − (8.19) m−1 ()mk−−1! 2km − 1 ⎛⎞x π ∑ ⎜⎟ k =0 k!2⎝⎠ where γ = 0.5772156... is Euler’s constant. This second solution is often used instead of J − p even for noninteger p . The general solution to Bessel’s equation is therefore given by ()=+ () () ≥ ykrAJkrBNkrppp for all p 0 , (8.20) ()=+ () () ≠ ykrAJkrBJkrpp− p for p 0,1, 2,3... The main difference between the Bessel and Neumann functions is that the Bessel Functions for p ≥ 0 converge at the origin, while the Neumann functions diverge at the origin. Their re- spective leading order behavior for small kr is given by p 1 x + =+⎛⎞ ()p 2 limJxp ( ) ⎜⎟ Ox x→0 Γ+()p 12⎝⎠ ⎧−Γ() − p p ⎛⎞x 2− p (8.21) ⎪ ⎜⎟ +>Ox()for p 0 = ⎪ π ⎝⎠2 limNxp ( )⎨ x→0 2 ⎪ ln(xO )+=() 1 for p 0. ⎩⎪π For large kr , the asymptotic expansions of two functions behave like phase-shifted sine and cosine functions with a decay enve- lop that falls of as ()kr −1/2 : 180 Bessel functions 221p + − ⎛⎞−+π ()3/2 limJxp ( )∼ cos⎜⎟ x Ox x→∞ π x ⎝⎠4 (8.22) 221p + − ⎛⎞−+π ()3/2 limNxp ( )∼ sin⎜⎟ x Ox . x→∞ π x ⎝⎠4 8.2 Cylindrical Bessel functions Bessel functions of integer order 1 1 J0() x J1() x 0 J_n(x) Jn() 2, x − 1 1 0510 0 x 10 x = Figure 8-1 Cylindrical Bessel functions of order m 0,1,2,3 For integer m , the solutions to Bessel’s equation are the , Jkrm () and the cylindrical Neumann functions Nkrm (). Graphs for the first three Bessel functions of integer order are shown in Figure 8-1. Bessel functions 181 Neumann-Weber functions of integer order 1 0.521 Y0() x Y1() x 0 Y_n(x) Yn() 2, x − 1 1 0510 0.01 x 10 x = Figure 8-2 Neumann (Weber) functions of order m 0,1,2 All Bessel functions for positive m , except those of order m = 0, start off with a zero at the origin. A similar plot showing the first few Neumann (Weber) functions is shown in Figure 8-2. Note that they diverge to negative infinity at the origin. Hankel functions Closely related to the Bessel functions are the , which are de- fined by (1) =+ HxJxiNxppp() () (), (8.23) (2) =− HxJxiNxppp() () (). 182 Bessel functions Hankel functions are most often encountered in scattering prob- lems where the boundary conditions specify incoming or out- going cylindrical or spherical waves. Zeroes of the Bessel functions There are an infinite numbers of zeroes (zero-crossings) of the Bessel functions. The zeroes of the Bessel functions are impor- tant, since they provide the eigenvalues needed to find the inte- rior solution to a cylindrical boundary value problem, where one has either Dirichlet or Neumann boundary conditions. For Di- == richlet Boundary conditions, let x kr ar/ r0 , where r0 is the radius of a cylinder. Then limJarr()() /== Ja Jx() = 0 , (8.24) → ppppn0 rro th th where xpn represent the n zero of the p Bessel function. () Therefore the eigenvalues of Jkrp are restricted to x = pn k pn . (8.25) r0 For Neumann boundary conditions one has instead limJarr′′′() /== Jx() 0 , (8.26) → pppn0 rro ′ th th where xpn represent the n zero of the derivative of the p Bes- sel function. Bessel functions 183 Orthogonality of Bessel functions Bessel’s equation is a self-adjoint differential equation. There- fore, the solutions of the eigenvalue problem for Dirichlet or Neumann boundary conditions are orthogonal to either other with respect to the weight function x = kr . Like the Fourier se- φ = imφ ries fm () e , the eigenfunctions of Bessel’s equation for fixed p are the same function Jkrmnm() stretched to have a zero at the boundary. Let ab, be distinct zeroes of J p , then the square-integral nor- malization of a Bessel function is given by 1 221 xdx J() ax= J+ () a . (8.27) ∫0 pp2 1 = Substituting x rr/ o 2 r 0 22ro rdrJ() x r/ r= J+ () x (8.28) ∫0 ppn012 p pn Orthogonal series of Bessel functions Consider a piecewise continuous function f p ()r that we want to ≤≤ expand in a Bessel function series for the interval 0 rro . Let = xpn denote the zeroes of Jxrrppn(/)0 for rro . Then, the series expansion is given by 184 Bessel functions ∞ = () f pnppn()rAJxrr∑ /0 , (8.29) n=1 where the coefficients An are given by 2 1 AxdxfrJxx= () () pn2 ∫0 p pn Jxppn+1 () (8.30) 2 r0 = rdr f() r J() x r / r 22 ∫0 ppn 0 rJ01ppn+ () x = and x rr/ 0 . = Jx() Discussion Problem: Expand fr() 1 in a 0 Bessel func- tion expansion inside a cylinder of radius a , assuming Dirichlet boundary conditions. Note: A first glance, this problem does not appear solvable as a Bessel function series, since the function does not meet the re- = quired boundary conditions fr()0 0. But all this really means is that we have a stepwise discontinuity at ra= . Orthogonal functions are well suited to handle such discontinuities. (One can expect to see some version of the Gibbs phenomena at the discontinuous point however.) Since f ()r is nonzero at r = 0 , it is appropriate to try an expansion in terms of Jxra00(/)n . (Ex- ≠ pansions in Jxp () for p 0 would not work.) The function to be fitted can by rewritten as f ()rra=< 1 for , (8.31) f ()rra== 0 for . Bessel functions 185 Example: Plot the first few eigenfunctions of Jkr0 () that have = zeroes at rr/10 . The zeroes of the Bessel functions are transcendental numbers. One can find their values numerically, using a root finding algo- rithm. This gives the values following values for the roots of Jx0 (): { } x0n = 2.405, 5.52, 8.654, 11.792, 14.931, 18.071, . (8.32) Figure 8-3 shows the first four eigenfunctions of Jkr0 () satisfy- ==() ing Dirichlet Boundary conditions at rkx1, mn mn J0(k_n*r) 1 1 J0() k ⋅r 1 0.5 ⋅ J0() k2 r ⋅ J0() k3 r ⋅ J0() k4 r 0 − 0.403 0.5 0 0.2 0.4 0.6 0.8 0 r 1 Figure 8-3 First four eigenfunctions of Jkr0 ()satisfying Dirichlet boundary conditions at r=1. 186 Bessel functions () For fixed p , a function f p r that is finite at the origin and va- = nishes at cylindrical boundary rr0 () can be expanded as a Bessel function series ∞ ()= () f ppnppnrAJxrr∑ /,0 n=1 (8.33) ()= frp 0 0. If, instead, one were to use , one would use the expansion ∞ ()= ()′ f ppnppnrAJxrr∑ /,0 n=1 (8.34) ′ ()= frp 0 0. Generating function The generating function for Bessel functions of integer order is given by ∞ mt(1/)/2− t= m eJxt∑ m () . (8.35) =−∞ Recursion relations Like Legendre Polynomials, there are a large number of useful recursion formulas relating Bessel functions. Some of the more useful identities are 2 Jx+−()=− JxJx () (), (8.36) ppp11p Bessel functions 187 ′ 1 J ()xJxJx=−⎡ −+ () ()⎤ , (8.37) ppp2 ⎣ 11⎦ and ′ pp J ()xJxJxJxJx=− () +−+ () = () − (). (8.38) pppppx 11x Of particular importance are the raising and lowering ladder op- erators that relate a Bessel function to the next function on the ladder: d pp ⎡⎤x Jx()= xJ− () x (8.39) dx ⎣⎦pp1 and d −−pp ⎡⎤x Jx()= xJ+ (). x (8.40) dx ⎣⎦pp1 The Neumann functions satisfy the same relations as (8.36)- (8.40). These recursion relations can most readily be proven by direct substitution of the series expansion given by equation (8.16). For example, the proof of equation (8.39) is given by n dd∞ ()−1 ⎛⎞ x22np+ ⎡⎤xJp () x = ⎣⎦p ∑ Γ+Γ++()⎜⎟2np+ dx dxn=0 (1) n n p 12⎝⎠ (8.41) n ∞ ()(−+12np 2 )⎛⎞x221np+− = . ∑ Γ+Γ++()⎜⎟2np+ n=0 (1)nnp 1⎝⎠ 2 Using Γ++=(1)()()np np + Γ+ np gives 188 Bessel functions n dx∞ ()−1 ⎛⎞221np+− ⎡⎤xJp () x = ⎣⎦p ∑ Γ+Γ+()⎜⎟21np+− dxn=0 (1) n n p ⎝⎠ 2 n ∞ ()−1 ⎛⎞x21np+− = x p (8.42) ∑ Γ+Γ+()⎜⎟21np+− n=0 (1)nnp⎝⎠ 2 = p xJp−1. 8.3 Modified Bessel functions If one makes the replacement of kik→ in Bessel’s equation (8.1), one gets the modified Bessel equation. ⎛⎞dd−−22 () = ⎜⎟xx pxIxp 0 . (8.43) ⎝⎠dx dx where I p ()x denote the modified Bessel functions of the first kind. Their series solution is nearly identical to Bessel’s series (8.16), except that the coefficients no longer alternate in sign 2np± ∞ 1 ⎛⎞x Ix()= . (8.44) ± p ∑ Γ+Γ±+⎜⎟ n=0 (1)nnp() 12⎝⎠ Noting that the substitution kik→ is equivalent to the substitu- tion x → ix , so the solutions also can be written as ()= p ( ) I ppxiJix, (8.45) where the factor i p is included so that the series expansion (8.44) is a real-valued function. Bessel functions 189 If the Bessel functions could be said to be oscillatory in charac- ter, asymptotically involving decaying sinusoidal functions, the solutions to the modified Bessel equation are exponential in be- havior. For positive p , the solutions are finite at the origin and grow exponentially with increasing x as shown in Figure 8-4. They have the asymptotic behavior 2 I ()xe∼ x for large x. (8.46) p π x Modified Bessel functions In(x) 10 10 I0() x I1() x 5 I_n(x) In() 2, x 0 0 024 0 x 4 x Figure 8-4 Modified Bessel Functions of the first kind → () For integer pm, the solutions I± p x are not linearly inde- pendent, = I−mm()xIx (). (8.47) 190 Bessel functions Modified Bessel functions of the second kind To obtain a second linearly-independent solution, valid for all p , the linear combination π Kx()=−⎡ I− () x Ix ()⎤ (8.48) ppp2sinπ p ⎣ ⎦ is used. The modified Bessel functions of the second kind K p ()x diverge at the origin. They exponentially decay for large values of x , as shown in Figure 8-5, with the asymptotic behavior 1 − Kx()∼ ex for large x . (8.49) p 2π x = () For integer pm, the series expansion for Kxp , calculated using L’Hospital’s rule, is given by − m−1 2km ()=− ( )mk ( ) +γ () +1 ()( − − − )⎛⎞x Kxmm1ln/2⎡⎤⎣⎦ x Ix∑ 1 mk 1!⎜⎟ 22k =0 ⎝⎠ (8.50) m 2km+ ()−Φ+Φ+1 m−1 ()kkm ( )⎛⎞x + , ∑ + ⎜⎟ 2!()!2k =0 km k ⎝⎠ where n 1 Φ=()nn for ′ ≠ 0, ∑ ′ n′=1 n (8.51) Φ=()00. () () For small x , I p x and Kxp have the leading order expan- sions Bessel functions 191 p 1 x −+ ()=+⎛⎞ ()p 2 Ixp ⎜⎟ Ox , Γ+()p 12⎝⎠ − ⎧ − p ⎧ ()2 p > 1 ⎛⎞x ⎪Ox for p 1, (8.52) ⎪ Γ+()p ⎜⎟ ⎨ Kx()= 22⎝⎠ ()p << p ⎨ ⎩⎪Ox for 0 p 1, ⎪ −+ = ⎩⎪ ln()xO () 1 for p 0. Modified Bessel functions Kn(x) 5 5 K0() x K1() x 2.5 K_n(x) Kn() 2, x 0 0 024 0.01 x 4 x Figure 8-5 Modified Bessel functions of the second kind Recursion formulas for modified Bessel functions Unlike their close cousins, the Bessel functions of the first and second kind, the modified Bessel functions of the first and second kind satisfy different recursion formulas. Several of the 192 Bessel functions more useful of these are listed below, others can be found in standard compilations of mathematics tables. 2 IxIx+−()=− () Ix(), pp11p p (8.53) 2 KxKx+−()=+ () Kx(), pp11p p 1 ′ ()=+⎡⎤() () Ixppp⎣⎦ I−+11 x I x, 2 (8.54) ′ 1 K ()xKxKx=−⎡ −+() + ()⎤ , ppp2 ⎣ 11⎦ d ⎡⎤pp()= () ⎣⎦xIpp x xI−1 x, dx (8.55) d pp ⎡⎤x Kx()=− xKx− (), dx ⎣⎦pp1 d −− ⎡⎤pp()= () ⎣⎦xIpp x xI+1 x, dx (8.56) d −−pp ⎡⎤x Kx()=− xK+ () x. dx ⎣⎦pp1 8.4 Solutions to other differential equations A significant use of Bessel’s functions is in finding the solutions of other differential equations. For example, the second order differential equation of the form 222 12−−aapc⎡⎤− 2 ′′()+++ ′ () ()c 1 () = yx yx⎢⎥ bcx2 yx 0 (8.57) xx⎣⎦ has the solution Bessel functions 193 = ac y xZp () bx , (8.58) where Z p is any linear combination of the Bessel functions J p and N p , and abc,,, p are constants. 8.5 Spherical Bessel functions The spherical Bessel equation represents the radial solution to the Helmholtz equation in spherical coordinates. This equation can be written as ⎡⎤1(1)ddll+ rkyr22−+() = 0 (8.59) ⎣⎦⎢⎥rdrdr22 r where the values of l = 0,1,2 is restricted to integer values. The substitution x = kr is made to put the equation into dimension- less form and to scale away the eigenvalue k 2 . The resulting eq- uation can be rewritten in the self-adjoint form as ⎡⎤dd xllxyr22−++(1) ()0 =. (8.60) ⎣⎦⎢⎥dx dx Note that the equation has a weight factor Wx()== x2 () kr2 . This factor of r2 in the weight comes from the Jacobean of transfor- mation of an element of volume when expressed in spherical coordinates dV=Ω r2 drd . (8.61) 194 Bessel functions The solution to the equation can be given in terms of elementary functions, but it is usual to express the solution in terms of Bes- sel functions. Using (8.57) and (8.58) one finds the solution = −1/2 yl ()xxZ()l+1/2 () x (8.62) (try abcpl=−1/2, = = 1, = + 1/2). Discussion Problem: Show by mathematical induction that − l ()= l ⎛⎞d () jloxx⎜⎟ jx, (8.63) ⎝⎠xdx where ()= jxo sin xx / , (8.64) is a solution to the spherical Bessel equation (8.60). Definitions The spherical Bessel functions of the first and second kind are defined as π = j ()xJx()+ () l 2x l 1/2 (8.65) − l l ⎛⎞dxsin = x ⎜⎟ , ⎝⎠xdx x Bessel functions 195 π = nx() N()+ () x l 2x l 1/2 (8.66) −−l l ⎛⎞dxcos = x ⎜⎟ . ⎝⎠xdx x Like the cylindrical Bessel functions, the spherical Bessel func- tions of the first (Figure 8-6) and second (Figure 8-7) kind are oscillatory, with an infinite number of zero crossings. For large x their decay envelope falls off as 1/ x . Spherical Bessel functions 1 1 js() 0, x , js() 1 x 0 j_l(x) js() 2, x − 1 1 0510 0 x 10 x Figure 8-6 Spherical Bessel functions 196 Bessel functions Spherical Weber functions 0.5 0.337 ys() 0, x , ys() 1 x 0.25 y_l(x) ys() 2, x − 1 1 0510 0.01 x 10 x Figure 8-7 Spherical Neumann (Weber) functions The spherical Hankel functions are defined as π (1) ()1 hx()==+ H+ () x jxinx () (), (8.67) llll2x 1/2 π (2) ()2 hx()==− H+ () x jxinx () (). (8.68) llll2x 1/2 Lastly, the modified spherical Bessel functions are given by π = ix() I()+ () x l 2x l 1/2 l (8.69) l ⎛⎞dxsinh = x ⎜⎟ , ⎝⎠xdx x π = kx() K()+ () x l 2x l 1/2 (8.70) − l − x l ⎛⎞de = x ⎜⎟. ⎝⎠xdx x Bessel functions 197 Table 8-1 lists the first three l values of the most common spherical Bessel functions. The limiting behavior of these func- tions for small and large values of x are summarized in Table 8-2. Table 8-1 Spherical Bessel functions of order 0, 1, and 2 198 Bessel functions Table 8-2 Asymptotic limits for spherical Bessel Functions Recursion relations Some recursion relations for the spherical Bessel functions are summarized in (8.71) where fl can be replaced by any of the (1) (2) functions jlll,,nh , h l. 21l + fx−+()+= fx () fx (), ll11x l d nf−+() x−+ ( l 1) f () x =() 2 l + 1 f (), x (8.71) ll11dx l dl+1 l f ()xfx=−−+ () fxfx () =−+ () fx (). dxll11 x l l x l The ladder operators for the spherical Bessel functions are given by Bessel functions 199 d ++ ⎡⎤ll11= ⎣⎦x fxll() x f−1 (), x dx (8.72) d −−ll ⎡⎤x fx()=− xf+ (). x dx ⎣⎦ll1 The equivalent recursion relations for the modified spherical Bessel functions are summarized in (8.73) where fl can be re- l+1 placed by ikll or (-1) . 21l + fxfx−+()−= () fx (), ll11x l d nf−+()(1)()21 x++ l f x =() l + f (), x (8.73) ll11dx l dl+1 l f ()xfx=−−+ () fxfx () =+ () fx (). dxll11 x ll x l The ladder operators for the modified spherical Bessel functions are given by d ++ ⎡⎤ll11= ⎣⎦x fxll() x f−1 (), x dx (8.74) d −−ll ⎡⎤x fx()= xf+ (). x dx ⎣⎦ll1 Orthogonal series of spherical Bessel functions = Let x rr/ 0 , where r0 is the surface of a sphere. Assuming Di- richlet boundary conditions, the eigenfunctions of the spherical Bessel functions that vanish on this surface are given by = jalln(), 0 , (8.75) 200 Bessel functions th th where aln, denotes the n zero of the l spherical Bessel func- tion. Suppose one has a function fl ()r defined on the interior of this sphere. Assume one wants to expand fl ()r in a Bessel func- tion series of order l . The expansion would take the form ∞ = () fllnlln()rAjarr∑ ,,0 / . (8.76) n=0 (Figure 8-8 shows how the functions scale to fit in the nth zero at the boundary) Since this is a series of orthogonal functions, one can use the orthogonality relation, which is given by 2 1 ja() xdxjaxj2 ()() bx = l+1 δ , (8.77) ∫0 ll2 ab where ab, denote two zeroes of the lth Bessel function. There- fore, the coefficients Aln, are given by 2 1 Axdxfrjarr= 2 () ( / ) ln,,0ja2 ()∫0 l l ln l+1 (8.78) 2 1 = rdrfr2 () j ( a r / r ). 23∫0 llln,0 jarl+10() Bessel functions 201 j0(kr) 1 1 js() 0, k ⋅r 1 0.5 , ⋅ js() 0 k2 r , ⋅ js() 0 k3 r , ⋅ js() 0 k4 r 0 − 0.217 0.5 0 0.2 0.4 0.6 0.8 0 r 1 Figure 8-8 Eigenfunctions of j0 ()kr Example: Expand, in a series of spherical Bessel functions, the distribution ⎧ f for 0 ≤ ≤ Where the series solution is valid for rr0. The distribution is spherically symmetric, so one can expand the function in a series of l = 0 Bessel functions ∞ ()= () f rAjarr∑ nn00,0/ , (8.80) n=1 where 202 Bessel functions 2 f 1 A = 0 xdxja2 () x . (8.81) nn2 ∫0 00, ja1 ()n Using the ladder operators (8.72) d ⎡⎤x22jx()= xjx (), (8.82) dx ⎣⎦10 the integral can be evaluated, giving 1 1 a0,n xdxjax22()= x′′ dxj () x ′ ∫∫0000n a3 0,n (8.83) 11a0,n ==⎡⎤xjx2 () j() a , 3 ⎣⎦110,0 n aa0,nn 0, 22ffja10,()n ==00 An 2 , (8.84) j10,0,1()aann aja nn () yielding the result ∞ j ()arr/ ()= 00,0n fr2 f0 ∑ . (8.85) n=1 aja0,nn 1() 0, π The zeros of j0 are given by an0,n = . The results of the series approximation are shown in Figure 8-9. Because the distribu- tion is discontinuous, the overshooting effect that is characteris- tic of the Gibbs Phenomena is observed. The magnitude of the overshoot persists even when increasing the number of terms, but the area of the overshoot gets smaller. In the infinite series limit, the series and the function would agree except on a inter- val of null measure. Bessel functions 203 Figure 8-9 Spherical Bessel function fit to a distribution with a piecewise discontinuity. 9. Laplace equation 9.1 Origin of Laplace equation Laplace’s equation ∇Φ2 ()r =0 (9.1) occurs as the steady-state (time-independent) limit of a number of scalar second-order differential equations that span the range of physics problems. In electrostatics or Newtonian gravitation problems, the Φ field can be interpreted as defining a potential function (electrostatic or gravitational, respectively) in a source free region. The equation also occurs in thermodynamics, where Φ can be interpreted as the local temperature of a system in a steady state equilibrium. To understand Laplace’s equation, let’s derive it in the context of Gauss’s Law, which states that the net Electric flux crossing a closed boundary surface is proportional to the charge enclosed in the volume defined by the bounding surface: