Propagation of Electromagnetic Waves in Continuous Media
Stefan Zollner Department of Physics New Mexico State University December 18, 2018
Contents
1 Scalar and vector waves 2
2 Fourier Series of Periodic Functions 2
3 Fourier Transforms of Non-periodic Functions 4
4 Fourier Transforms and Fourier Series in Multiple Dimensions 5
5 Maxwell’s Equations in Vacuum (Microscopic) 6
6 Plane Wave Solutions to Maxwell’s Equations in Vacuum 8
7 Polarized Light in Vacuum; Jones vectors; Stokes parameters 10
8 Ellipsometry 11
9 Dielectric in a Static Electric and Magnetic Field 12
10 Magnetostatics and Magnetization 14
11 Dielectric Response Function 15
12 Maxwell’s Equations for Homogenous Continuous Media 17
13 Generalized Plane Waves 18
14 Generalized Plane Waves in Homogenous Continuous Media 19
15 Energy Density and Energy Flow of an Electromagnetic Wave 20
16 Dielectric Function: Drude and Lorentz Model 21
17 Macroscopic Optical Constants 28
18 Causality, Kramers-Kronig-Relations, and Sum Rules 29
19 Analytical properties of the dielectric function 29
20 Experimental Assignments 32
21 References 33
1 1 Scalar and vector waves
A sinusoidal scalar traveling wave as a function of position ~r and time t s (~r, t) = A cos ~k · ~r − ωt + φ (1.1)
described by an amplitude A, wave vector ~k, angular frequency ω, and phase φ has an infinite duration and spatial extension. It is difficult to plot or to use in numerical calculations. A more compact mathematical form is preferred. This can be achieved by Fourier series or Fourier transforms. Similarly to a scalar wave, we can write a vector wave in components as E0x cos ~k · ~r − ωt + φ ~ ~ ~ ~ E (~r, t) = E0y cos k · ~r − ωt + φ = E0 cos k · ~r − ωt + φ , (1.2) E0z cos ~k · ~r − ωt + φ
where the amplitude is now a vector with three components E~0 = (E0x,E0y,E0z).
Problems: 1.1. Show that the scalar wave in Eq. (1.1) is periodic in time with period T = 2π/ω. 1.2. The scalar wave in Eq. (1.1) is also periodic in real space with wavelength ~λ, written as a vector. Find the relationship between ~λ and ~k. 1.3. A scalar s is defined as s = ~a ·~b. A coordinate transformation defined by a matrix A transforms the unit vectors ~i, ~j, and ~k into a different set of right-handed unit vectors. Show that the scalar s is invariant under this coordinate transformation.
2 Fourier Series of Periodic Functions
A (real-valued scalar) function f (t) is called periodic with period T , if f (t) = f (t + T ) for all values of t. We see that the scalar wave in Eq. (1.1) is periodic with period T = 2π/ω. A periodic function f (t) with period T can be written as (Jackson 1975, 2.36-37) 1 ∞ 2πmt 2πmt f (t) = A + X A cos + B sin (2.1) 2 0 m T m T m=1 or 1 ∞ f (t) = A + X [A cos (mωt) + B sin (mωt)] (2.2) 2 0 m m m=1 with Fourier coefficients
T 2 2 Z 2πmt A = f (t) cos dt (2.3) m T T T − 2 T 2 2 Z 2πmt B = f (t) sin dt (2.4) m T T T − 2 or in terms of the angular frequency ω
π ω ω Z A = f (t) cos (mωt) dt (2.5) m π π − ω
2 π ω ω Z B = f (t) sin (mωt) dt (2.6) m π π − ω In plain English, Eq. (2.2) means that any periodic function with period T and angular frequency ω can be writen as a sum of its overtones. This is also called harmonic analysis. It is inconvenient that the Fourier series written above have two terms (sines and cosines) and two sets of coefficients. The equations can be simplified by moving to complex numbers. If we remember that
exp (−iωt) = cos (ωt) − i sin (ωt) , (2.7)
then we can show that the Fourier series becomes +∞ X f (t) = cm exp (−imωt) (2.8) m=−∞ with Fourier coefficients (check !)
π ω A0 Z 2 ; m = 0 ω 1 cm = f (t) exp (imωt) dt = 2 (Am + iBm); m > 0 (2.9) π 1 π (A−m − iB−m); m < 0 − ω 2 Since the function f (t) is a real function (any measurable physical quantity must be real!), the series (2.8) is not really a complex function. All imaginary terms in this series must cancel (not individually, but after summation). The Fourier coefficients cm are only a mathematical definition and not physical quantities. Therefore, the Fourier coefficients are usually complex quantities. Note: Following the usual ellipsometric convention (see also Born & Wolf, Jackson 1975, etc), we use a minus-sign in the exponential containing the time t. This is different from some math texts. The significance of this choice will be explained later.
Problems: 2.1. A scalar function f (t) is called even, if f (t) = f (−t). It is called odd, if f (t) = −f (−t). What can you say about the Fourier coefficients Am and Bm of even and odd functions. 2.2. Use the form of the coefficients cm in Eq. (2.9) and substitute into Eq. (2.8). Can you recover the Fourier coefficients Am and Bm in Eqs. (2.3) and (2.4)? (Let me know if you find a mistake!) 2.3. What can you say about the Fourier coefficients cm for even and odd functions f (t)? 2.4. Solve the following integrals for arbitrary integers m1 and m2:
π ω ω Z cos (m ωt) cos (m ωt) dt = ? (2.10) π 1 2 π − ω
π ω ω Z sin (m ωt) sin (m ωt) dt = ? (2.11) π 1 2 π − ω π ω ω Z sin (m ωt) cos (m ωt) dt = ? (2.12) π 1 2 π − ω q ω q ω The functions π cos (mωt) and π sin (mωt) (where m is an integer) form an orthonormal basis set for the infinite-dimensional vector space of periodic functions with period T (Jackson 1975, Sec. 2.8).
3 2.5. Show that the function f (t) in Eq. (2.8) is real, i.e., equal to its complex conjugate. 2.6. Check out the article on Fourier series in wikipedia and calculate the first few terms in the Fourier series for a square wave and a sawtooth wave (stop reading after example 1). Many compilations of mathematical tables (such as Bronstein-Semendjajew 1979) list Fourier series expansions of common functions.
3 Fourier Transforms of Non-periodic Functions
If the function f (t) is not periodic, then the period T becomes infinite and the frequency spacing ω between overtones becomes very small. Instead of needing a series of discrete overtones with spacing ω (as for finite period T ), we now need a continuous set of angular frequencies ω. The Fourier series (2.8) becomes a Fourier integral (Jackson 1975, 2.44) ∞ 1 Z f (t) = √ F (ω) exp (−iωt) dω. (3.1) 2π −∞ The function F (ω) is called the Fourier transform of f (t) ∞ 1 Z F (ω) = √ f (t) exp (iωt) dt. (3.2) 2π −∞ I have written the Fourier transform (3.2) and the inverse Fourier transform (3.1) with a symmetric prefactor √1 . This notation can vary (Jackson 1975, 6.55-56). For any function f (t) describing a physical 2π quantity, the imaginary terms in the inverse Fourier transform (3.1) must cancel to make the function real-valued. The Fourier transform F (ω) may be complex, since it is not a physical quantity. Just like the time t, the angular frequency ω is a real quantity in these transforms. The orthogonality and completeness relations are now (Jackson 1975, 2.46-47) ∞ 1 Z exp i ω − ω0 t dt = δ ω − ω0 (3.3) 2π −∞ and ∞ 1 Z exp iω t − t0 dω = δ t − t0 , (3.4) 2π −∞ where δ (t) is the Kronecker δ-function. These integrals (3.3) and (3.4) are not to be taken literally, since the δ-function is not a real function. These equations only become meaningful mathematically within the theory of functional analysis. Only integrals containing the δ-function as a kernel become meaningful. The functions √1 exp (iωt) form an orthonormal basis of a Hilbert space of functions. 2π Problems: 3.1. The convolution (f ∗ g)(t) of two functions f (t) and g (t) is defined by ∞ Z (f ∗ g)(t) = f t0 g t − t0 dt0. (3.5) −∞
Show that the Fourier transform of the convolution of two functions f√(t) and g (t) is equal to the product of their Fourier transforms F (ω) and G (ω) (up to a factor of 2π), i.e., ∞ 1 Z √ √ (f ∗ g)(t) exp (iωt) dt = 2πF (ω) G (ω) . (3.6) 2π −∞
4 √ 3.2. Can we choose√ a normalization factor other than 2π in the Fourier transform to get rid of the prefactor 2π in the convolution theorem? 3.3. If F (ω) is the Fourier transform of f (t), then the Fourier transform of the derivative f 0 (t) is iωF (ω). In other words, ∞ 1 Z √ f 0 (t) exp (iωt) dt = iωF (ω) . (3.7) 2π −∞ 3.4. If F (ω) is the Fourier transform of f (t), show that F (−ω) = F¯ (ω) (i.e., the complex conjugate of F (ω)).
4 Fourier Transforms and Fourier Series in Multiple Dimensions
The Fourier transforms described in Sec. 3 are written in one dimension. An example for this is the transform from time into frequency space. We can also transform scalar or vector functions in real space and move into wave vector space. (In quantum mechanics or solid-state physics, this would be called momentum space or reciprocal space.) A scalar field s (~r) in a crystal (Bravais lattice) is called periodic if s ~r + R~ = s (~r) for all Bravais lattice vectors R~ = u~a + v~b + w~c with integral coefficients u, v, and w and primitive lattice translations ~a, ~b, and ~c spanning the unit cell V = ~a · ~b × ~c . In this case (Ashcroft & Mermin 1976, Appendix D), the three-dimensional Fourier series is given by a sum over all reciprocal lattice vectors G~ = h~a∗ + k~b∗ + l~c∗ (with real-valued components) X ~ s (~r) = sG~ exp iG · ~r , (4.1) G~ where 2π 2π 2π ~a∗ = ~b × ~c , ~b∗ = (~c × ~a) , and ~c∗ = ~a ×~b (4.2) V V V and the Fourier coefficients are given by
1 Z s = s (~r) exp −iG~ · ~r d3~r. (4.3) G~ V C
The integral is over the primitive cell C spanned by ~a, ~b, and ~c with volume V . Compare the one- dimensional case in Eq. (2.8), which is the same except for a different sign and for a factor of 23 = 8. (For a periodic time-dependent function with period T , the unit cell volume is V = T = 2π/ω and 2π/V = ω is the reciprocal space unit.) This formalism can also be applied to functions periodic in reciprocal space. As always, the imaginary terms in the series (4.1) must cancel if the scalar field s (~r) is a real-valued physical quantity. The same equations apply to a periodic vector field
~ X ~ ~ E (~r) = EG~ exp iG · ~r , (4.4) G~ where the Fourier coefficients are given by
1 Z E~ = E~ (~r) exp −iG~ · ~r d3~r. (4.5) G~ V C
As always, the imaginary terms in the series (4.4) must cancel if the vector field E~ (~r) is a real-valued physical quantity.
5 Now let us turn to the Fourier transforms for non-periodic scalar and vector fields. The three- dimensional Fourier integral for a scalar field s (~r) becomes
∞ ∞ ∞ 1 3 Z Z Z s (~r) = √ S ~k exp i~k · ~r d3~k. (4.6) 2π −∞ −∞ −∞ The function S ~k is called the Fourier transform of s (~r)
∞ ∞ ∞ 1 3 Z Z Z S ~k = √ s (~r) exp −i~k · ~r d3~r. (4.7) 2π −∞ −∞ −∞
Similar equations hold for non-periodic vector fields E~ (~r)
∞ ∞ ∞ 1 3 Z Z Z E~ (~r) = √ E~ ~k exp i~k · ~r d3~k (4.8) 2π −∞ −∞ −∞ with the Fourier transform defined by
∞ ∞ ∞ 1 3 Z Z Z E~ ~k = √ E~ (~r) exp −i~k · ~r d3~r. (4.9) 2π −∞ −∞ −∞
As always, the imaginary terms in the inverse Fourier transforms (4.6) and (4.8) must cancel if the scalar field s (~r) and the vector field E~ (~r) are real-valued physical quantities. The vectors ~r and ~k also have real components.
5 Maxwell’s Equations in Vacuum (Microscopic)
In mechanics, our goal is to determine the position ~r of a particle with mass m exposed to a net force F~ , given the initial position and velocity. The equation of motion is Newton’s second law: F~ = m~r¨. In the theory of electricity and magnetism, the electric field strength E~ and the magnetic field strength H~ need to be determined using initial or boundary conditions for a given charge density ρ and current density ~j. Both fields E~ and H~ are vector fields and depend on position and time, but they can be Fourier-transformed and then become functions of wave vector ~k and angular frequency ω. Maxwell’s equations take the place of Newton’s second law. (The definition of the magnetic field strength H~ follows Serway 1990, Sec. 30.9. See also Holm 1991. This allows a symmetric treatment of the dielectric and magnetic susceptibilities involving the microscopic pair of quantities E~ and H~ turning into D~ and B~ in the macroscopic case.) For completeness, we also define the dielectric displacement D~ and the magnetic field B~ (also known as the magnetic flux density)
D~ = 0E~ (5.1)
B~ = µ0H.~ (5.2)
Let’s keep in mind that these expressions (5.1-5.2) are only valid in vacuum and need to be modified inside a continuous medium. In differential form, the general microscopic form of Maxwell’s equations is ρ Gauss’ law (electric field) ∇~ · E~ = (5.3) 0
6 Gauss’ law (magnetic field) ∇~ · H~ = 0 (5.4) ∂H~ Faraday’s law ∇~ × E~ = −µ (5.5) 0 ∂t ∂E~ Ampere’s law ∇~ × H~ = ~j + , (5.6) 0 ∂t
where 0 is the permittivity of free space and µ0 the permeability of free space. The first equation is Coulomb’s law and implies that the electric field strength from a point charge drops off like the square of the distance. The second equation states that there are no magnetic charges. Faraday’s law states that a flux change causes an induced voltage. (There is no magnetic current, because there are no magnetic charges.) Ampere’s law states that the magnetic field strength curls around the current. We can apply Gauss’ theorem and Stokes’ theorem to convert these equations into their integral form (but this is not necessary here). There’s a few other things to note about Maxwell’s equations: They are linear in the fields E~ and H~ , they are first order in space coordinates and time, and they have constant coefficients 0 and µ0. (Ac- tually, if we choose different units, the speed of light c is the only parameter in these equations. The speed of light couples the electric and magnetic field strenghts.) However, these equations are inhomo- geneous: The charge density ρ and the current density ~j are the inhomogeneous sources. Because of the n o n o linearity of the equations, if the systems E~1, H~ 1, ρ1,~j1 and E~2, H~ 2, ρ2,~j2 solve Maxwell’s equations, n o then E~1 + E~2, H~ 1 + H~ 2, ρ1 + ρ2,~j1 +~j2 is also a solution. This linearity allows us to Fourier-transform Maxwell’s equations. Because of linearity, the integrals commute with the partial derivatives. Finally, we see that the electric and magnetic field strengths E~ and H~ are only coupled through the partial time deriva- tives. Therefore, electrostatics and magnetostatics are independent and the static fields don’t mix. The electric and magnetic field strengths E~ and H~ are only coupled, once we consider dynamic (time-dependent) effects, for example in the theory of the propagation of electromagnetic waves. Since we are interested in the propagation of electromagnetic waves, we don’t have to worry about free charges (ρ=0) and currents flowing through a wire (~j = 0). Therefore, Maxwell’s equations in vacuum read as follows (Jackson 1975, 7.1):
Gauss’ law (electric field) ∇~ · E~ = 0 (5.7) Gauss’ law (magnetic field) ∇~ · H~ = 0 (5.8) ∂H~ Faraday’s law ∇~ × E~ = −µ (5.9) 0 ∂t ∂E~ Ampere’s law ∇~ × H~ = . (5.10) 0 ∂t While the current density ~j vanishes, we note that we still need to deal with the displacement current density on the right hand side of Ampere’s law. The speed of light c is defined by 1 c = √ . (5.11) 0µ0 Maxwell’s equations are now truly homogeneous without any source terms. Because of the linearity of n o n o the equations, if the systems E~1, H~ 1 and E~2, H~ 2 solve the homogeneous Maxwell’s equations, then n o E~1 + E~2, H~ 1 + H~ 2 is also a solution.
Problems: 5.1. Derive the wave equations (Holm 1991, Jackson 1975)
1 ∂2E~ ∇~ 2E~ − = 0 and (5.12) c2 ∂t2
7 1 ∂2H~ ∇~ 2H~ − = 0 (5.13) c2 ∂t2 from Maxwell’s equations in vacuum. See Good & Nelson 1971, 4.7. n o n o 5.2. Show the following: If the systems E~1, H~ 1, ρ1,~j1 and E~2, H~ 2, ρ2,~j2 solve Maxwell’s equations, n o then E~1 + E~2, H~ 1 + H~ 2, ρ1 + ρ2,~j1 +~j2 is also a solution. 5.3. What are the units of the electric and magnetic field strengths E~ and H~ in the MKSA system? The Poynting vector is defined as (Serway 1990, 34.23)
S~ = E~ × H.~ (5.14)
What is the direction and the unit of the Poynting vector?
6 Plane Wave Solutions to Maxwell’s Equations in Vacuum
As stated before, the purpose of Maxwell’s equations is to determine the electric and magnetic field strengths E~ (~r, t) and H~ (~r, t) for a given charge and current density. Maxwell’s equations are linear and therefore the integrals in the Fourier transformed fields commute with the differential operators in Maxwell’s equations. Let’s Fourier transform the electric field strength E~ (~r, t) in all four dimensions (position ~r and time t). The resulting Fourier transforms are given by
∞ ∞ ∞ ∞ 1 2 Z Z Z Z h i E~ ~k, ω = E~ (~r, t) exp −i ~k · ~r − ωt d3~rdt and (6.1) 2π −∞ −∞ −∞ −∞ ∞ ∞ ∞ ∞ 1 2 Z Z Z Z h i E~ (~r, t) = E~ ~k, ω exp i ~k · ~r − ωt d3~kdω (6.2) 2π −∞ −∞ −∞ −∞ with a corresponding expression for the magnetic field strength H~ (~r, t). We note that the exponential phase has the opposite sign in the Fourier transform (6.1) and the inverse Fourier transform (6.2). How we choose this phase (if the minus sign appears in the Fourier transform or in the inverse transform) is convention. In the Nebraska convention for ellipsometric measurements as modified by Aspnes (Muller 1969, Holm 1991, Huml´ıˇcek2005, Jackson 1975), Eq. (6.2) with the negative time dependence factor is the preferred choice. The form of the exponential in Eq. (6.1) with the minus sign applied to time is also the preferred choice for a quantum-mechanical wave packet to ensure that the kinetic energy of a free particle is positive (Holm 1991). This choice has profound consequences on the sign of the imaginary part of the dielectric constant (related to the extinction coefficient). Unfortunately, the positive sign of the time-dependence is almost universally in use in electrical engineering. This makes it difficult to transition from the optical to the microwave regime, if different conventions are used. Finally, we note that the terms involving time and space coordinates have opposite signs in the usual convention. (In a one-dimensional case with the wave moving along the z-direction, the wave moves towards increasing z with increasing time t.) A plane wave E~ ~k, ω with wave vector ~k and angular frequency ω is a δ-function in wave vector and angular frequency space and thus all four integrals collapse. The electric field strength for a plane wave looks like this: h i E~ (~r, t) = E~0 exp i ~k · ~r − ωt . (6.3)
There is a similar equation for the magnetic field strength H~ (~r, t) of a plane wave: h i H~ (~r, t) = H~ 0 exp i ~k · ~r − ωt . (6.4)
8 Both E~0 and B~ 0 are vectors with complex coefficients, but they do not depend on the space and time coordinates ~r and t. Unfortunately, plane waves are not physical solutions to Maxwell’s equations for two reasons: First, because the electric and magnetic field strengths are not real-valued, but complex quantities. Second, the fields are δ-functions in wave vector and frequency space, with infinitely sharp frequencies and wave vectors. To form a physical solution, we must do two things. First, we form a wave packet with smoothly varying envelope functions E~ ~k, ω and H~ ~k, ω in frequency and wave vector space. A Gaussian superposition is very popular, because it resembles the output of a well-tuned laser. Second, we add the complex conjugate wave (terms with negative frequencies and wave vectors) to achieve real-valued smoothly varying solutions E~ (~r, t) and H~ (~r, t). While they are not physical solutions, plane waves form an orthonormal basis of the Hilbert space containing all solutions to Maxwell’s equations. They are also very easy to deal with from a mathematical viewpoint. For a plane wave in Eqs. (6.3-6.4), the phase of the wave is a well-defined quantity for any given time or position. We say that the coherence length of a plane wave is infinite. If we form a Gaussian wave packet, then the frequency spread of the wave causes an uncertainty of the phase for times t 6= 0 or away from the origin. The phase is defined reasonably well only within one coherence length of the origin. At distances further away, only the magnitude of the wave (i.e., its intensity) is well-defined. Therefore, interference effects of two waves with the same frequency can only be observed if the two waves are less than one coherence lenght apart (i.e., if the path difference is less than a coherence length). In this case, we add electric and magnetic field strenghts. For path differences much greater than one coherence length, we add intensities, not the field amplitudes. The Fourier transform equations (6.1) and (6.2) say that any electric field can be constructed by superposition of plane waves (orthogonality and completeness). If we can show that the plane waves (6.3- 6.4) are a solution to Maxwell’s equations (5.7-5.10), then we can form the general solution by superposition. Substitution of the electric and magnetic fields (6.3) and (6.4) into Maxwell’s equations (5.7-5.10) in vacuum yields the following (see Good & Nelson 1971, page 384): (CHECK!)
Gauss’ law (electric field) ~k · E~0 = 0 (6.5)
Gauss’ law (magnetic field) ~k · H~ 0 = 0 (6.6)
Faraday’s law ~k × E~0 = ωµ0H~ 0 (6.7)
Ampere’s law ~k × H~ 0 = −ω0E~0. (6.8)
Therefore, plane waves indeed solve Maxwell’s equations in vacuum, but only under certain conditions (6.5-6.8) for the frequency and wave vector. The first two equations state that the electric and magnetic fields for an electromagnetic wave with non-vanishing amplitudes are perpendicular to the wave vector. We therefore say that electromagnetic waves are transverse waves, because the disturbance is perpendicular to the direction of propagation. The last two equations state that the electric and magnetic field stregnths are also perpendicular to each other. The unit vectors for the fields and the direction of propagation form a right-handed coordinate system. Finally, by substituting Faraday’s law (6.7) into Ampere’s law (6.8), we obtain the equivalent of the wave equation (5.12): (CHECK !)
ω2 k2 = (6.9) c2
This is called a dispersion relation, because it contains an expression between the wave vector ~k and the angular frequency ω. This dispersion relation defines the speed of the wave (in this case the speed of light). Looking at the simplicity of Maxwell’s equations and the wave equation for a plane wave, you will understand why we went through the trouble of defining the Fourier transformation and the inconvenience of a complex-valued solution in the form of a plane wave.
9 Faraday’s Law and Ampere’s Law allow us to compare the magnitude of the electric and magnetic field strengths. We find that (Fox, A.36-37)
rµ0 E0 = H0 = Z0H0. (6.10) 0
The quantity Z0=377 Ω is called the impedance of vacuum.
Problems: 6.1. For an electric field h i E~ (~r, t) = E~0 exp i ~k · ~r − ωt (6.11) show that (CHECK !) ∇~ · E~ (~r, t) = i~k · E~ (6.12) ∇~ × E~ (~r, t) = i~k × E~ (6.13) ∂E~ (~r, t) = −iωE~ (6.14) ∂t ∇~ 2E~ (~r, t) = −k2E~ (6.15) ∂2E~ (~r, t) = −ω2E~ (6.16) ∂t2 6.2. Derive the wave equation (6.9) from Faraday’s law and Ampere’s law in ~k-dependent form.
7 Polarized Light in Vacuum; Jones vectors; Stokes parameters
For an electromagnetic wave propagating along the z-direction, let’s write the wave vector as ~k = kkˆ, where k is the magnitude of the wave vector and kˆ the unit vector along the z-axis. Gauss’ law for plane waves (6.5)-(6.6) forces E0z = 0 and H0z = 0, another statement of the transverse character of electromagnetic waves. Together with the wave equation (dispersion relation) (6.9), we can write the magnetic field strength H~ as a function of the electric field strength E~ :
r 0 H0x = − E0y (7.1) µ0 r 0 H0y = E0x (7.2) µ0 Therefore, an electromagnetic plane wave is described by the following quantities: • the direction of propagation (wave vector), • the angular frequency ω (which defines the magnitude k of the wave vector), • the magnitudes of the electric field strength perpendicular to the wave vector (which is related to the magnetic field strengths through the equations above). If we fix the angular frequency ω, the electromagnetic wave is described by only two complex numbers Eox and E0y, i.e., four real quantities. As a function of position coordinate and time, the electric field strength of the wave can be written as a vector with two columns E E~ (~r, t) = 0x exp [i (kz − ωt)] . (7.3) E0y This wave is known as a plane wave, because the electric field strength does not depend on the x and y coordinates. In the xy-plane (where z=0), the electric field is E E~ (z = 0, t) = 0x exp (−iωt) . (7.4) E0y
10 Keeping in mind that the quantities E0x and E0y are complex numbers, they have an amplitude (real) and a phase. Let’s call the (real) magnitude of the electric field E0 and the normalized real amplitudes of the electric field strength in the x and y directions X and Y , respectively. Let’s call the corresponding phases ∆X and ∆Y . The electric field in the xy-plane then becomes X exp (i∆X ) E~ (z = 0, t) = E0 exp (−iωt) . (7.5) Y exp (i∆Y ) We can easily rewrite this as (see Humlicek 2005, Fujiwara 2007)
X exp (i∆) E~ (z = 0, t) = E exp [−iω (t − t )] , (7.6) 0 Y 0
if ∆ = ∆X − ∆Y and t0 = ∆Y /ω. The vector X exp (i∆) (7.7) Y
is called the Jones vector of the electromagnetic wave. Defining X tan ψ = , (7.8) Y the Jones vector can also be written as sin ψ exp (i∆) , (7.9) cos ψ where ψ and ∆ are called the ellipsometric angles. For ∆ = 0 or ∆ = π, we obtain linearly polarized light. For ψ=π/4 and ∆ = π/2 or ∆ = −π/2, we obtain circularly polarized light. In the most general case, where ψ ranges from 0 to π/2 and ∆ ranges from −π to π, we obtain elliptically polarized light. Show examples of Jones vectors for linearly and circularly polarized light. Compare Table 3.1 in Fujiwara 2007. Define Stokes parameters following Section 3.4 in Fujiwara 2007. Introduct Poincare sphere. Discuss totally and partially polarized light. Introduce depolarization.
8 Ellipsometry
In an ellipsometry experiment, we have an incident beam and a detected (reflected) beam, each with an electric field strength with complex amplitude
X exp (i∆) E . (8.1) 0 Y
In the experiment, the detected beam was somehow altered (for example by reflection by or transmission through a sample). Define Jones matrix as relating the Jones vectors for the incident and reflected (or transmitted beam) by a two-by-two matrix with complex coefficients. Off-diagonal elements vanish for isotropic materials, but can be non-zero in case of anisotropy. Define Mueller matrix as the relationship between incident and reflected (or transmitted) Stokes pa- rameters.
11 9 Dielectric in a Static Electric and Magnetic Field
After the propagation of electromagnetic waves (i.e., light) in vacuum has been described, we now turn to the response of materials to an electromagnetic field. Before we can discuss the response of materials to time-dependent fields, we must first describe their response to static electric and magnetic fields. Referring to Fig. 1, we first consider a dielectric exposed to a homogeneous externally applied electric field E~0. The charges (positively charged nuclei and negatively charged electrons) in the dielectric respond to the external field with small displacements. The positive charges in the material are pulled along the field lines of the applied electric field (shown by solid lines), while negative charges are forced in the opposite direction. The piece of material as a whole is still electrically neutral and the average macroscopic charge density is zero (look inside the green box, for example). However, these small displacements collectively create an induced electric field E~1 opposite to the applied external field. The electric field lines of the induced electric field (shown by the dashed lines) point from the positive charges to the negative charges. Also, the material has acquired a positive surface charge on the left and a negative surface charge on the right. The total (local) electric field E~ , sometimes called E~local, is the sum of the applied (external) electric field E~0 and the induced (depolarizing) electric field E~1, i.e, E~ = E~0 + E~1 (Nye 1985). The magnitude of the local electric field is usually smaller than that of the external applied field E~0. This is called screening. (See Fox 2010, Clausius-Mossotti relationship.)
Figure 1: A dielectric material is placed inside a homo- geneous static electric field E~0. The charges (positively charged nuclei and negatively charged electrons) in the di- electric respond to the external field with a small displace- ment. This displacement creates an induced electric field E~1 opposite to the applied external field. The total (local) elec-
tric field is therefore E~ = E~0 +E~1, which is called screening. The charge displacements create a dipole moment, which is proportional to the local electric field E~ . The dielectric po- larization is defined as the dipole moment per unit volume.
Each small charge displacement creates an electric dipole. We remember that the magnitude of the dipole moment ~p = q∆~r is equal to the distance between the charges times the magnitude of the charge (Young & Freedman 1987, Sec. 21.7). The dipole moment as a vector points from the negative charge to the positive charge. The dielectric polarization P~ is defined as the dipole moment per unit volume (average over the green box, for example). Now we are ready to define the dielectric constant and the dielectric displacement. The dielectric displacement D~ for a material is obtained as a sum, including the original electric field (in vacuum) and the dielectric polarization:
D~ = 0E~ + P.~ (9.1) If the electric field is not too large (below the dielectric breakdown strength of the material) and in the absence of ferroelectric effects (where a non-vanishing dielectric polarization remains even in the absence of an electric field), the dielectric polarization is proportional to the applied electric field. This defines the dielectric susceptibility χe as P~ = 0χeE.~ (9.2) In the ferroelectric case, this can be generalized as
P~ = P~r + 0χeE,~ (9.3)
12 where P~r is the remanent (ferroelectric) polarization at a vanishing electric field E~ . The dielectric constant is defined by
= 1 + χe. (9.4)
The dielectric displacement can therefore be written as
D~ = 0E~ + P~ = P~r + 0E~ (9.5) as a product involving the dielectric constant and the applied electric field (if P~r vanishes). So far, we have assumed that the charges in the dielectric will move the same distance (create the same magnitude of polarization) regardless of the direction of the applied electric field E~ . This is true for isotropic materials. In uniaxial or biaxial crystals, the charges may actually move in a direction different from the applied electric field, because the restoring force inside the crystal varies with direction. Therefore, the induced electric field E1, the dielectric polarization P~ , and also the dielectric displacement D~ may not be parallel to the applied electric field E~ . In mathematical terms, the dielectric susceptibility χe and the dielectric constant are not numbers, but 3×3 tensors. This makes calculations more difficult. Furthermore, a polarization of the medium in a paramagnet can also be caused by an applied magnetic field of strength H~ . Examples of such magneto-optical effects are the magneto-optical Faraday effect (in transmission) and Kerr effect (in reflection). Including magneto-electric effects, the displacement is written in an even more general form as (Schubert 2005)
D~ = 0E~ + P~ = P~r + 0E~ + 0δH.~ (9.6)
The term 0δH~ describes the polarization in a paramagnet. In a ferromagnet, the polarization can de- pend on the magnetization even in the absence of a magnetic field H~ . In the ferromagnetic case, the magneto-optical properties of the material are described by off-diagonal elements in the dielectric tensor (Mansuripur 1995, Schubert 2005). The ferroelectric remanence P~r will vary only slowly with time, especially at optical frequencies, because the charges must be moved across large ferroelectric domains, which takes longer than moving a single charge. Therefore, without loss of generality, we may assume that
∂P~ r = 0. (9.7) ∂t
Similarly, if the ferroelectric domains are infinitely large, then the spatial derivatives of P~r will vanish also, but this is not true across domain walls.
Problems:
9.1. What are the units of E~ , D~ , P~ , χ, 0, and , in MKSA units? 9.2. A dielectric contains bound charges with charge q and charge density n. We apply a static electric field E~ . Assume that the charges respond by moving in the same direction as the field (isotropic model) and 2 that the restoring force is proportional to the displacement with a spring constant k = mω0 (linear approximation for small electric fields). Show that the static dielectric susceptibility is (Jackson 1975, 4.72) 2 2 nq ωP χe = 2 = 2 . (9.8) m0ω0 ω0 For reasons that will become clear later, the term s nq2 ωP = (9.9) m0 is called the plasma frequency.
13 9.3. Let’s treat an ideal gas as a dielectric. What is the density of molecules at normal temperature and pressure? Each molecule has several types of charges. The nuclei (positive charges) contribute very little, because of the large mass in the denominator. Charges in completely filled (inner) shells contribute very little, because they are bound very tightly and have large values of ω0. Let’s only consider electrons in partially filled shells contributing to molecular bonds. What is the plasma frequency for such a gas (give a range, based on the number of charges per molecule)? Assuming that the resonance wavelength λ0 is 100 to 300 nm, estimate a range of expected values for the susceptibility of a gas (like hydrogen, oxygen, nitrogen, etc). Compare your result with literature values for the refractive indices of various gases. (See Jackson 1975, Sec. 4.6.)
10 Magnetostatics and Magnetization
While there are no magnetic charges, an applied magnetic field can align microscopic magnetic dipoles (e.g., spins) inside a material. The average magnetic dipole moment per unit volume is called magnetization M~ , similar to the dielectric polarization P~ . The magnetic field B~ in the magnetic material is a sum involving the applied magnetic field strength H~ and the magnetization M~
B~ = µ0H~ + M.~ (10.1)
When comparing the definition of the magnetization (10.1) with the definition of the dielectric displacement (9.1), we note a symmetry in the units, which motivates this convention (which is a bit unusual compared to the literature). In the absence of ferromagnetic effects (where a non-zero magnetization exists for a vanishing magnetic field strength) and for small magnetic field strengths, the magnetization is proportional to the magnetic field strength, leading to the definition of the magnetic susceptibility χm
M~ = µ0χmH.~ (10.2)
In the ferromagnetic case, this can be generalized as
M~ = M~ r + µ0χmH.~ (10.3) where M~ r is the remanent (ferromagnetic) magnetization (also called remanence) at a vanishing magnetic field H~ . The magnetic permeability µ is defined by
µ = 1 + χm. (10.4)
Finally, we get an expression of B~ as a function of H~ involving the magnetic permeability
B~ = µ0H~ + M~ = M~ r + µµ0H,~ (10.5) which is symmetric to the expressions for the dielectric constant. If we also consider that (through a different magneto-optical effect) the magnetization can be created by an electric field, then more general expression reads (Schubert 2005)
B~ = µ0H~ + M~ = M~ r + µ0µH~ + µ0γE,~ (10.6)
Just like in the electrostatic case, the magnetic susceptibility χm and the magnetic permeability µ are 3×3-tensors, since the magnetization may not be parallel to the applied magnetic field strength vector H~ . Therefore, B~ and H~ may not be parallel. Also, it is very difficult for the magnetization to follow the applied field at optical frequencies. Therefore, µ (ω) and γ (ω) are always zero at optical frequencies (from the infrared to the UV), but static electric
14 and magnetic fields can, of course, produce a magnetization, which implies that µ (0) and γ (0) may be non-zero. Similarly to the case of the ferroelectric polarization, we may assume without loss of generality that ∂M~ r = 0. (10.7) ∂t
The spatial derivatives of M~ r will vanish within a single ferromagnetic domain, but they will be non-zero across domain boundaries.
Problems:
10.1. What are the units of H~ , B~ , M~ , χµ, µ0, and µ, in MKSA units?
11 Dielectric Response Function
As described above, the charges (and permanent electric dipoles) in a material will respond to the applied electric field and produce a dielectric polarization (see Fig. 1), but there will be a time delay. In the linear optical case (ignoring ferroeletric and magneto-optical effects), the resulting dielectric polarization can be written as (Yu & Cardona, 1996, 6.1) Z 0 0 0 0 0 3 0 P~ (~r, t) = 0 χˆe ~r , ~r, t , t E~ r~ , t dt d ~r , (11.1) where the integral is a four-dimensional infinite integral (time and space as four coordinates t0 and ~r0). If the susceptibility tensorχ ˆe is a δ-function in the differences of the time and space coordinates, we recover the static expression (9.2). Let’s examine this integral (11.1): The dielectric polarization P~ (~r, t) depends not only on the electric field E~ (~r, t) applied at time t, but also on the electric fields E~ (~r, t0) applied previously (t0 < t). This is known as temporal dispersion (or simply dispersion) or temporal non-locality. Why is that? Because of the inertia of the charges and the related damping, it will take some time for them to respond to the applied electric field. One example is the charging of a capacitor: Because of the finite (non-zero) resistance of the wiring, the capacitor is not charged instantly, but the charge increases gradually over time and it takes an infinite amount of time for the capacitor to reach its full charge. Another example is the Lorentz model (see problems). Similarly, a charge located at position ~r is influenced not only by the electric field at position ~r, but also at positions ~r0. Why is that? Imagine that the applied electric field is non-zero at position ~r0 and zero everywhere else. The charge at position ~r0 will move in response to the applied electric field by an amount ∆~r0. If the distance between the charges is comparable to the magnitude of ∆~r0, then nearby charges will be affected by this displacement ∆~r0 and respond with their own displacement. (You can think of Fermi’s exclusion principle, which does not allow two identical charges to be at the same place at the same time. You can also think of a system of coupled pendulums, where one mass is displaced and thus causes a displacement of the other masses also.) This phenomenon is called screening. Its effect on the dielectric susceptibility χe is known as spatial dispersion or spatial non-locality (sometimes just non-locality). We conclude that the dielectric response of a material to an applied electric field is non-local in both space and time coordinates. Spatial dispersion is difficult to observe in bulk materials, but it has been found near the band-gap of semiconductors by resonant Raman scattering (Yu 1971) and even has implications for deep-UV lithography in deep-submicron CMOS lithography (Serebryakov 2003). Temporal dispersion is easily found in all materials, where the complex dielectric function usually varies with frequency. It is impossible for the charges to respond to an electric field that has not yet been applied. This is known as causality. There is no mechanism in physics to predict the future (clairvoyance). Therefore,