An Introduction to the Mathematical Aspects of Quantum Mechanics: Course Notes
Universidade Federal de Goiás Instituto de Física e Química
An Introduction to the Mathematical Aspects of Quantum Mechanics: Course Notes
Petrus Henrique Ribeiro dos Anjos Paulo Eduardo Gonçalves de Assis
Catalão,Go 2015
Contents
1 Quantum States and Observables 5 1.1 QuantumStates...... 6 1.1.1 Uncertainty Principle ...... 11 1.2 Inner Product and Hilbert Spaces ...... 12 1.3 Observables ...... 15 1.4 Probability and Functions of Observables ...... 18 1.5 Self-adjointoperators...... 20 1.6 RiezRepresentation...... 22 1.7 SpectralTheorem...... 24 1.8 Exercises...... 28
2 The Spectrum 29 2.1 SpectrumandResolvent ...... 30 2.2 Findingthespectrum...... 35 2.2.1 ThespectrumoftheHamiltonian ...... 36 2.3 Exercises...... 41
3 Quantum Dynamics 43 3.1 Time evolution and Schrödinger Equation ...... 43 3.2 Applications to Two-Level Systems ...... 49 3.3 Schrödinger’swaveequation ...... 52 3.4 Time dependence of Expected Values ...... 57 3.4.1 Newton’s 2nd Law and Quantum Mechanics ...... 58 4 CONTENTS
3.5 QuantumPictures ...... 60 3.6 Exercises...... 62
4 Approximation Methods 63 4.1 Non-perturbativemethods ...... 63 4.1.1 VariationalMethods ...... 63 4.1.2 ExtensiontoExcitedStates ...... 66 4.1.3 A 2 2 example ...... 67 × 4.1.4 Method os Successive Powers ...... 68 4.1.5 WKB - Semiclassical approximation ...... 68 4.2 Time-independent Perturbation Theory ...... 71 4.2.1 Time-independent perturbation: Non-degenerate . . . . 72 4.2.2 Time-independent perturbation: Degenerate ...... 75 4.3 TheAnharmonicOscillator ...... 77 4.4 Time-dependent Perturbation ...... 79 4.4.1 Time-dependent Perturbation Theory ...... 80 4.5 Furhter Applications and Fermi’s Golden Rule ...... 83 4.6 Dirac’sinteractionpicture ...... 84 4.7 Exercises...... 85
Referências 87 Chapter 1
Quantum States and Observables
In classical physics the mathematical description of a phenomenon is somewhat clear. From the early days of modern science, the movement of a macroscopic body could be completely characterized by the specification of its position at a given instant of time. This process was easily achieved with the use of simple tools such as rulers and clocks. Predictions could be com- puted based on formulations relying on simple mathematical objects called functions. The development of differential and integral calculus, in fact, owes a lot to the works of scientists like Isaac Newton.
With the advance of our understanding of the microscopic world, physi- cist were forced to abandon the classical formalism to match experiments. Intead, a new theory came to light and both the basic objects to describe it and the mathematical tools to formulate it had to be revised. Part of this mathematical toolbox was already well stablished at the beginning of the last century but some of it had to be developed alongside the results of experiments.
Here we present an introduction to the mathematical aspects of quantum mechanics.
5 6 CHAPTER 1. QUANTUM STATES AND OBSERVABLES 1.1 Quantum States
Definition 1.A. The state of a single particle 1-D quantum system is a complex valued continuous function ψ(x, t) such that
i. The probability (x I) of finding that the position of the particle Pψ ∈ belong to the interval I R at time t is given by ⊂ (x I)= ψ(x, t) 2dx Pψ ∈ | | ZI ii. The probability (p I) of finding that the momentum of the particle Pψ ∈ belong to te interval I R at time t is given by ⊂ 1 p (p I)= ψ˜( , t) 2dp, Pψ ∈ ~ | ~ | ZI where f˜ is the Fourier transform of ψ, given by 1 f˜(k)= f(x)eikxdx √2π R Z A general property of the Fourier transform called Parceval’s identity [1], show us that Note that p dp ψ(x, t) 2dx = ψ˜( , t) 2 . (1.1.1) R | | R ~ | ~ Z Z And to our probability interpretation hold, we that
ψ(x, t) 2dx =1, (1.1.2) R | | Z which means that the particle lies somewhere in the real line. As an exercise, we suggest that you show that the defined on 1.A Pψ actually obeys other probabilities rules. More precisely, you should show that with our definition, for any countable sequence of disjoint intervals I1,I2, ..., we have that (x I )= (x I ). (1.1.3) Pψ ∈ n Pψ ∈ ∪ n n n [ X 1.1. QUANTUM STATES 7
So the position x (or the momentum p) of the particle can assume any value in the real line, and to each interval we assign a probability (x I) Pψ ∈ that the value of x lies in I. We can show how to compute the mathematical expectation of x. As a warm up, assume that x is restricted to a bounded interval [a, b]. We can divide [a, b] into smaller subintervals Ik, and consider the following object:
xkIk, Xk where xk is an arbitrary point of Ik. We desire that this sum converge to a limit as the maximum length goes to zero, and furthermore the convergence is independent of our choices of intervals Ik and point xk. If all this holds, we call the limit x¯ the mathematical expectation of x. If x is not restricted to a bounded interval, we can fix an arbitrary bounded interval [a, b], calculate (as above) a limit for this bounded interval, and finally take the limit where [a, b] goes to the real line. Now, in the latter case, if this limit (a) exist, (b) is independent of our choice of interval [a, b] and (c) is independent of how we grow the interval; we call this limit x¯ the mathematical expectation of x. There are a lot of “if” ’s here, and in order to our physical theory works we need to start to prove some things. Fortunately the following holds:
Lemma 1.1.1. Let the system be in a state ψ and f(x) be a continuous function such that f(x) ψ(x, t) 2dx < + , R | || | ∞ Z then f¯ the mathematical expectation of f(x) is given by
f¯ = f(x) ψ(x, t) 2dx R | | Z Proof. Let I R be a bounded interval and I , ...I be a partition of I into ⊂ 1 N smaller non-overlaping intervals with maximum length δ. For let for each k pick an arbitrary x I , then k ∈ k 8 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
2 2 f(x) ψ(x, t) dx f(xk) ψ(x Ik) 6 f(x) f(xk) ψ(x, t) dx. I | | − P ∈ Ik | − || | Z k k Z X X The continuity of f(x) implies that we can make f(x) f(x ) 6 ǫ | − k | 2 by taking δ sufficiently small. Therefore f(x) f(xk) ψ(x, t) dx 6 Ik | − || | 2 ǫ ψ(x, t) dx, then R Ik | | R
2 2 f(x) ψ(x, t) dx f(xk) ψ(x Ik) 6 ǫ ψ(x, t) dx. I | | − P ∈ I | | Z k Z X
To conclude the demonstration we note that +l f(x) ψ(x) 2dx = lim f(x) ψ(x) 2dx R | | l + l | | Z → ∞ Z−
A direct consequence of the lemma above is the
Corolary 1.1.1. Let the system be in a state ψ, the following holds
x¯ = x ψ(x, t) 2dx R | | Z Analogue statements holds concerning the momentum, as follows. Their proof are left as exercises.
Lemma 1.1.2. Let the system be in a state ψ and g(p) denote a continuous function such that p dp g(p) ψ˜( , t) 2 < + , R | || ~ | ~ ∞ Z the following holds p dp g(p)ψ˜( , t) 2 =g ¯ R ~ | ~ Z Corolary 1.1.2. Let the system be in a state ψ, the following holds p dp p¯ = pψ˜( , t) 2 R ~ | ~ Z 1.1. QUANTUM STATES 9
Since we use Fourier transform to describe the probabilities concerning the momentum, it will be helpful to recall some properties of this integral transformation. We use Fourier transform here without too much concern on the conditions that f(x) must satisfy in order that the transforated function exists, for a complete discussion of Fourier transform see Ref. [1]. First we recall an useful property of Fourier transform:
∂ f(k)= ikf˜(k) (1.1.4) ∂x − g This property follows directly form integration by parts and the observation that if f˜ exists then f(x) 0 as x 0. We also need an inverse Fourier | |→ | |→ transform, and it is given by
1 ikx f(x)= f˜(k)e− dk, (1.1.5) √2π R Z . To have some insight into equation 1.1.5, it worth consider a very helpful identity: 1 ik(x x′) δ(x x′)= e− − dk, (1.1.6) − 2π R Z where δ(x) denotes the quite ubiquitous entity that physicist call “Dirac Delta Function. Formally, this object can be defined as a “function” such that for any continuous function f and ǫ> 0
+ǫ f(x)δ(x)dx = f(0). (1.1.7) ǫ Z− When dealing with “Dirac Delta Function”, we need to keep in mind that:
1. It is not a function: Even an initial analysis show that δ(x) is not a function, it is actually an entity that the mathematicians call a distri- bution, and we suggest to see ref. [2] for a rigorous treatment.
2. Dirac is not the first to use it: For exemple, an infinitesimal formula for an infinitely tall, unit impulse delta function explicitly appears in a work of Cauchy in 1827 (see ref [3]). And indeed, several other 10 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
authors have dealt with objects with similar characteristics (among them Poisson , Kirchoff , Lord Kelvin , again see ref. [3]). And in particular, in the late nineteenth century Heaviside had derived the main properties of δ(x).
Despite these considerations , it should be noted that the notation introduced by Dirac in his influential book The Principles of Quantum Mechanics ( see Ref [4]), is not only convenient but insightful and intuitive, what made it readily available to a much larger community. Putting these ideas into practice , we can use the Fourier transform to to connect x and p representations of the quantum state. This is done in the
Proposition 1.1. Let the system be in a state ψ, the following holds
∂ p =( i~) ψ(x, t) ψ†(x, t)dx − R ∂x Z Proof. It is a straightforward application of eqs. 1.1.5, 1.1.4 and 1.1.6. In fact, we find that
∂ ~ i(k k′)x i~ ψ(x, t) ψ†(x, t)dx = kψ˜(k)ψ˜†(k′)e− − dkdk′dx − R ∂x 2π = ~ kψ˜(k)ψ˜†(k′)δ(k k′)dkdk′ R R 2 − = ~ Rk R ψ˜(k) dk 2 ˜ p = Rp ψ ( ~ ) dp =p ¯
R
Note that the Proposition 1.1 gives us the expected value of the momen- tum, free from Fourier transform under the price of the introduction of partial differentials. This gives us an important hint about the nature of quantum mechanical systems: We can define a Momentum Operator Pˆ
~ ∂ Pˆ ψ = ψ (1.1.8) i ∂x 1.1. QUANTUM STATES 11
For instance, this became clear by a repeated application of the ideias leading to Proposition 1.1, gives us
pn = Pˆnψ, ψ , where we use the notation
(f,g)= f(x)g†(x)dx. (1.1.9) R Z Note 1.a. Despite being a well-known fact , it should worth remember , at least to our most inexperienced readers, that usually pn =p ¯n. 6
For exemple, eq. 1.1.2 can be written as (ψ, ψ)=1. We will further discuss this notation and the role of operators like Pˆ in quantum mechanics in section . But before that, we want to show an intriguing property of quantum systems: The famous Uncertainty Principle.
1.1.1 Uncertainty Principle
The famous Heisenberg Uncertainty Principle is a theorem about Fourier transforms, once we grant a certain model of quantum mechanics. That is, there is an unavoidable mathematical mechanism that yields an inequality which has an paradigm shit interpretation in physics. This First note that integrating by parts
∂ ∂ (ψ, ψ)= x (ψψ†)dx = 2Re xψ(x) ψ†(x)dx. − R ∂x − R ∂x Z Z That is ∂ (ψ, ψ) 6 2 xψ(x) ψ†(x) dx. R ∂x Z
Furthermore, Cauchy-Schwartz identity applies here implying that
∂ ∂ (ψ, ψ) 6 2 xψ(x) ψ†(x) 6 2 xψ ψ , R ∂x k kk∂x k Z
12 CHAPTER 1. QUANTUM STATES AND OBSERVABLES where f 2 =(f, f). Now we use eq. 1.1.4 and Parceval’s identity (eq. 1.1.1) k k to converts derivatives (in the position) to multiplication (by the momentum p ~ ): ∂ ∂ ψ = ψ = kψ˜ . k∂x k k∂x k k k Since for any quantum state ψ =(gψ, ψ)=1, so we obtain the Heisenberg k k inequality (using p = ~k) ~2 xψ pψ˜ > . k kk k 4 A similar argument gives, for any x ,p R 0 0 ∈ ~2 (x x )ψ (p p )ψ˜ > . (1.1.10) k − 0 kk − 0 k 4
Put x0 =x ¯ and p0 =p ¯, then (x x )ψ = x2 x¯2 = ∆x k − 0 k − . (p p )ψ˜ = p2 p¯2 = ∆p k − 0 k − Applying this in Heisenberg’s general inequality (i.e. eq. (1.1.10)) we obtain the famous Heisenberg Principle: ~ ∆x∆p > . 2 Roughly speaking the quantities ∆x and ∆p can interpreted as the error obtained in measurements of position and momentum respectively. That is Heisenberg’s uncertainty principle, is a fundamental limit to the precision with which the pair (x, p) can be known simultaneously, or in another words it gives a lower bound on how spread out the probability distributions of x and p must be. Note that the only relevant Physical assumptions are in Definition 1.A, namely: the probabilistic interpretation of the quantum state and that Fourier transform relates x and p.
1.2 Inner Product and Hilbert Spaces
In this section, we develop the notation introduced in eq. (1.1.9) 1.2. INNER PRODUCT AND HILBERT SPACES 13
Definition 1.B. For every states ψ,φ we define the scalar product on 2 L
(φ, ψ)= φ∗(x, t)ψ(x, t)dx R Z Lemma 1.2.1. Let a, b C and φ,ψ,ν 2. The scalar product on 2 has ∈ ∈L L the following properties
1. (φ, aψ + bν)= a(φ, ψ)+ b(φ, ν);
2. (ψ,φ)=(ψ,φ)∗, where the asterisk denotes complex conjugation.
3. ψ 2 =(ψ, ψ) > 0 unless ψ =0 k k Note 1.b. Note that statement c of Lemma 1.2.1 is not quite correct. To be precise, we should say that if ψ 2 =0 then ψ(x)=0 almost everywhere, i.e. k k (in a technical sense) the set for which the property holds takes up nearly all possibilities. However, to simplify our discussion, we identify two functions that agrees almost everywhere.
A scalar product obeying Lemma 1.2.1 is called an inner product, and a vetor space provided with an inner product is called an inner product space. The Lemma 1.2.1 has the following important consequences:
Corolary 1.2.1. Let ψ,φ belong to an inner product space, then the following holds
i. (ψ,φ) 6 ψ φ (Schwarz inequality). | | k kk k ii. ψ + φ 6 ψ + φ (Triangle inequality). k k k k k k Proof. It is clear that [i.] holds, when ψ = 0, so assume that both φ and ψ ane non-zero. Also, we assume that (ψ,ϕ) =0 since otherwise the inequality 6 is obviously true. (ψ,ϕ) Let σ = ψ ϕ 2 ϕ, therefore the linearity of the inner product implies − k k that (ψ,ϕ) (σ, ϕ)=(ψ,ϕ) (ϕ,ϕ)=0. − ϕ 2 k k 14 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
Now, 2 (ψ,ϕ) 2 ψ = σ + ϕ 2 ϕ k k k k k k 2 (ψ,ϕ) (ψ,ϕ) = σ + ϕ 2 > ϕ 2 . k k k k k k
To show [ii.] , just expand ψ + ϕ 2 =( ψ + ϕ, ψ + ϕ), and apply Schwarz k k inequality. We leave the details as an exercise.
We are now capable of addressing an important question, that we are postponing, that is what functions are acceptable as quantum states? Our Remark raise some of the issues. But since we need the Definition 1.A, we must insist that these functions are square integrable on the real line. So, they must belong to the well know space L2(R). We also desire that these functions satisfies some features: Suppose there is a sequence of functions ψ L2 such that k ∈ ψ ψ 0 as j, k + . (1.2.11) k k − jk→ → ∞ A sequence satisfying eq.1.2.11 is called a Cauchy sequence. To have things working properly, we want that all Cauchy sequences converge to some el- ement of L2. We call this property completeness. More precisely: Given a sequence satisfying eq. 1.2.11 then there is ψ L2 such that ∈ ψ ψ 0 as k + . (1.2.12) k − kk→ → ∞ We say that a sequence satisfying equation 1.2.12 converges strongly (or in the norm) to ψ and write ψ ψ. We say that a sequence of ψ ’s converges k → k weakly to ψ when
(ψ ,ϕ) (ψ,ϕ) for all ϕ L2. (1.2.13) k → ∈ We point out that, as a consequence of Schwarz inequality, every strong convergent sequence is also weak convergent, since
(ψ ψ,ϕ) 6 ψ ψ ϕ . | k − | k k − kk k 1.3. OBSERVABLES 15
Furthermore, we say that a subset S of L2 is closed if every convergent sequence of ψ S converge to some ψ S, (if S is not closed then there k ∈ ∈ are sequences of elements of S that converge to elements of L2 that are not in S). A subset S of L2 is called dense if for every ψ L2 and ǫ> 0, there ∈ is an φ S such that ψ φ < ǫ. That means that any function in L2 can ∈ k − k be approximated by functions in S. The following property is very useful
Lemma 1.2.2. If S is a dense subset of L2 and
(ψ,ϕ)=0 ϕ S ψ =0. (1.2.14) ∀ ∈ ⇒ Proof. Note that (ψ,ϕ)=0 imply that ψ ϕ 2 = ψ 2 + ϕ 2 > ψ 2. k − k k k k k k k Since ϕ S then it can be taken arbitrarily close to ψ so ψ 2 6 0 therefore ∈ k k ψ =0.
A vector space with an inner product, that is complete with respect to the norm induced by the inner product is called a Hilbert Space (that is a inner product vector space that satisfies the completeness property). The space L2 is a very important exemple of a Hilbert Space, and many of the statements we made here actually holds for general Hilbert Spaces. However, if you are not familiar with Hilbert Spaces do not despair, you still follow our notes without difficult considering everything in the L2 context.
1.3 Observables
Definition 1.C. A physical quantity that can be measured is called an Ob- servable.
Postulate 1. To every observable a there is an corresponding linear operator A with dense domain D(A) 2 such that a¯ the expected value of a for the ⊂L system at the state ψ D(A) is given by ∈ a¯ =(ψ, Aψ) . 16 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
Not all operators with the property described in 1 correspond to an ob- servable. First, we need that the observable can only assume real values.
Definition 1.D. A linear operator A is called Hermitian if (ψ,Aφ)=(Aψ,φ) for every ψ,φ D(A). ∈ Lemma 1.3.1. A is a Hermitian Operator (ψ, Aψ) R for every ψ ⇔ ∈ ∈ D(A).
Proof. For ψ D(A), and since A is Hermitian we note that (Aψ, ψ) = ∈ (ψ, Aψ)=(Aψ, ψ)∗. Conversely, assume that Im(Aψ, ψ)=0, we have that
(A[iϕ + ψ], iϕ + ψ)=(Aφ,φ)+(Aψ, ψ)+ i(Aϕ, ψ) i(Aψ,ϕ), − taking imaginary parts we have that Re(Aψ,ϕ)= Re(ψ,Aφ). Finally,
Im(Aψ,ϕ)= Im [ i(Aiψ, ϕ)] = Re(Aiψ, ϕ)= Rei(ψ,Aϕ)= Im(ψ,Aϕ). − − − Then (Aψ,ϕ)=(ψ,Aϕ).
Now consider, for example, the momentum operator defined on eq.1.1.8. The expression
∂ψ p¯ =(Pψ,ψˆ )= i~ (x)ψ†dx − R ∂x Z makes sense only if ψ is differentiable with respect to x and if the integral exists. That means D(Pˆ), the domain of the momentum operator, can not be the whole L2. In fact, there is many examples of elements of L2 functions that are not differentiable everywhere. We need ways to extend the domain of an operator to deal with these cases. To do this is enough to require that the domain of the operator to be dense in L2. Furthermore, we need that the domain of the operator is the largest possible. This motivates the
Postulate 2. If A is an operator corresponding to an observable a and B is a Hermitian operator such that D(A) D(B), and ⊂ a¯ =(ψ,Bψ), 1.3. OBSERVABLES 17
ψ D(A) then B = A (i.e. D(B)= D(A) and Aψ = Bψ). ∀ ∈ The ideia of this postulate is to give us a way to extend the domain of an operator. For instance, suppose that ψ D(A), but there is a sequence 6∈ ψ D(A) such that ψ ψ and Aψ f, then we define a Aψ = f. { n} ⊂ n → n → This only makes sense if f does not depend on the sequence ψ . Now, { n} suppose we have another sequence φ D(A) ψ such that Aφ f. n ∈ → n → Therefore, µ = ψ φ 0, and we need that Aµ 0 for that method n n − n → n → works. An operator with this property is called closable . When it works this allow us to construct an extended operator denoted by A¯, the closure of A.
Note 1.c. It is also, worth to note that A¯ can not be extended any further using the method outlined above. In fact, suppose that there is a sequence ψ D(A¯) such that ψ ψ, Aψ¯ f. So there are sequences ψ D(A) k ∈ k → k → jk ∈ such that ψ ψ and Aψ Aψ¯ . So we can take one element of each of jk → k jk → k this sequences, and construct a new sequence ψ where ψ ψ, Aψ f. kk kk → kk → Therefore ψ D(A¯) and Aψ¯ = f. ∈ As our previous discussion suggests, not all operators are closeable. Fortu- nately, for the quantum mechanical theory, we we are dealing with Hermitian operators with dense domain, and we have the
Lemma 1.3.2. A Hermitian operator with dense domain is closable.
Proof. Let D(A) ψ 0 and Aψ f. For any ϕ D(A) we have that ∋ n → n → ∈
(Aψn,ϕ)=(ψn, Aϕ).
Taking the limits, we have that
(f,ϕ)=(0, Aϕ)=0.
Since D(A) is dense Lemma 1.2.2 implies that f =0.
Finally, we can put these results together. This lead us to the 18 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
Proposition 1.2. If the observable a is real valued then the corresponding operator is a closed Hermitian Operator.
Proof. We already prove that A must be Hermitian. Since, we require that its domain is dense, then it is closable. We claim that the closure A¯ is also Hermitian. In fact, for ψ,ϕ D(A¯), there are sequences D(A) ψ ψ ∈ ∋ n → and D(A) ϕ ϕ, such that Aψ Aψ¯ and Aϕ Aϕ¯ , and therefore ∋ n → n → n →
(Aψ,ϕ¯ ) = lim(Aψn,ϕk) = lim(ψn, Aϕk)=(ψ, Aϕ¯ ). n,k n,k
A¯ can not be further extended so the proposition follows from Postulate 2.
1.4 Probability and Functions of Observables
It is fundamental in any scientific theory that one can make some sort of prediction of the outcome of an experiment. If the theory is not able to make predictions, then it is not of any use to science (and most important not scientific at all). In particular, in quantum mechanics one need to deter- mine the probability (a I) that the real observable a be in the interval Pψ ∈ I. Clearly this quantity must depend on the quantum state ψ and the cor- responding operator A. We now show how one can obtain this information. Our approach is based on the
Note 1.d. A real function of an observable is an observable. The reason we emphasize this point is that usually physical quantities are not measured directly , but calculated from other quantities. So this is a funda- mental requirement of the theory. It is also highly likely that a more careful author would prove this statment or include it as an postulate. That means for any observable a and f : R R, there is an observable f(a). Corre- → sponding to the observable f(a) there is a corresponding Hermitian operator that (despite the abuse of the notation) we denote f(A) (postulate 1). 1.4. PROBABILITY AND FUNCTIONS OF OBSERVABLES 19
Now let χI (λ) be the characteristic function of the interval I, i.e.
1 λ I χI (λ)= ∈ (1.4.15) 0 λ I 6∈
So let us consider the operator χI (A), where A is the operator correspond- ing to the observable a. Since that for any state ψ, a real valued observable a can only be or not be in I, we find that
χ (a)= (a I) 1+ (a I) 0 I Pψ ∈ × Pψ 6∈ × (ψ, χ (A)ψ)= (a I) ⇒ I Pψ ∈ Unfortunately, this hardly answers our question. since we still need to understand what is the operator χI (A) and given A how to construct it. We now will try to deal with these problems proving some properties of χI (A).
Lemma 1.4.1. The following holds
i. χ (A)ψ 6 ψ , for all ψ D [χ (A)]2 . k I k k k ∈ I
2 ii. D(χI (A)) = L
Proof. Since χI (A) is Hermitian, we have that
χ2(a)=(χ2(A)ψ, ψ)=(χ (A)ψ, χ (A)ψ)= χ (A)ψ 2 I I I I k I k , for all normalized ψ in the domain of χ2(a). Also χ2(a)= χ (a)= (a I I I Pψ ∈ ϕ I) < 1, therefore χI(A)ψ 6 1. Now, for ϕ = 0, write ψ = ϕ and this k k 6 k k leads to statement [i.].
Lemma 1.4.2. If A is a Hermitian Operator with dense domain such that
(Aψ, ψ)=0 ψ D(A) ∀ ∈ then A =0. 20 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
Lemma 1.4.3. For each I,J R, the following holds ⊂ 2 2 i. χI (A)= χI (A)
ii. χI J (A)= χI (A)+ χJ (A) χI J (A) ∪ − ∩
iii. χI J (A)= χI (A)χJ (A) ∩
2 2 Proof. Note that χI (a) = χI (a) implies that (χI (A)ψ, ψ)=(χI (A)ψ, ψ), then ([χ2(A) χ (A)]ψ, ψ)=0, for all ψ L2. So statement [i.] follows from I − I ∈ lemma 1.4.2. We leave the demonstration of the remaining statements as an exercise 2 (Hint: Note that χI J = χI + χJ χI J and that [χI + χJ ] = χI + χJ + ∪ − ∩ 2χI J ) ∩ We advanced quite a bit, but still we are not able to construct the operator
χI (A) given A and not even know if it is defined for every densely defined closed Hermitian operator. To address these issues, in the next section we shall derive more consequences of our postulates.
1.5 Self-adjoint operators
Our main aim in this section is to prove and discuss the
Proposition 1.3. If A is the operator corresponding to a real observable, then the operator (1 + A2) is onto.
The proposition 1.3 is instrumental. It tells us that when dealing with quantum mechanical operator, for each f L2 the equation (ψ + A2ψ = f ∈ has a solution ψ D(A2) D(A). We should note that given to operators ∈ ⊂ A and B, D(A+B)= D(A) D(B), then D(1+A2)= D(A2). Furthermore, ∩ note that (1+A2)=(A+i)(A i)=(A i)(A+i), therefore the proposition − − 1.3 tells us that for every f L2 there are ψ,ϕ D(A) such that ∈ ∈ (A i)ψ =(A + i)ϕ = f. (1.5.16) − 1.5. SELF-ADJOINT OPERATORS 21
Once noticed this fact, the proposition 1.3 also suggests that not all densely defined closed Hermitian operator can correspond to an observable. This lead us to the
Proposition 1.4. let A be an operator corresponding to a real observable and let ψ, f D(A) such that for all ϕ D(A) we have ∈ ∈
(ψ,Aϕ)=(f,ϕ).
Then ψ D(A) and Aψ = f. ∈ Proof. What need to be proved is that ψ D(A) (if ψ D(A), we can use ∈ ∈ that A is Hermitian then (Aψ f,ϕ)=0, and by the lemma 1.2.2 Aψ = f). − Now note that, for every ϕ D(A), we have ∈
(ψ, [A + i]ϕ)=(ψ,Aϕ) i(ψ,ϕ)=(f iψ,ϕ) (1.5.17) − −
From eq. 1.5.16, there a w D(A) such that (A i)w = f iψ, which leads ∈ − − to (ψ, [A + i]ϕ)=(f iψ,ϕ) = ([A i]w,ϕ)=(w, [A + i]ϕ). − − Again by eq. 1.5.16, there is a ϕ D(A) such that [A + i]ϕ = ψ w. ∈ − Therefore
0=(ψ w, [A + i]ϕ)=(ψ w, ψ w)= ψ w 2, − − − k − k which show us that ψ = w D(A) ∈ The property described proposition 1.4 is instrumental to build functions of observables, however (un)fortunately this property is not shared by all densely defined closed Hermitian operators. This is one of the crucial points of this chapter, then take a deep breath and allow yourself a moment of reflection on this issue. For any densely defined operator A with domain 2 D(A) L we can define A†, the adjoint of A,by A†ψ = f in the proposition ⊂ 22 CHAPTER 1. QUANTUM STATES AND OBSERVABLES
1.4, this makes perfectly sense because D(A) is dense. Otherwise there can be more then one function that satisfies proposition 1.4. That means
(ψ,Aϕ)=(A†ψ,ϕ). (1.5.18)
The linearity of the inner product implies that A† is also a linear operator. When A is Hermitian , we can be tempted to swap the position of A in the inner product and say that A† = A, however this is only true if ψ D(A), ∈ 2 and the proposition 1.4 holds for ψ L . Then it is clear that D(A) D(A†), ∈ ⊂ and restricted to D(A), we have A† = A. For an operator that satisfies the proposition 1.4, we say that the operator is self-adjoint, i.e. D(A†) = D(A) and A† = A. Putting all this ideas together we just prove the
Proposition 1.5. If a is a real observable then the corresponding observable is self-adjoint.
1.6 Riez Representation
When we are dealing with Hilbert Spaces, Adjoints and more important physicists it is important to have some rigorous mathematical results to guide us. Usually, when dealing with quantum mechanics, physicists use a very particular “language” that they call “Dirac Notation” or “Bra-Ket Notation” to denote abstract vectors and linear functionals (i.e linear operator from the Hilbert space to the scalars), leaving inner products to a some sort of secondary role. The name “Bra-Ket” is so because the inner product of two quantum states is denoted by ψ ϕ , where ψ , called a “Bra”, is a linear h | i h | functional and ϕ is a vector in the Hilbert space. | i Note 1.e. It is curious that (again) Dirac is not the first one to introduce this notation, the notation has its roots on Grassmann’s algebra calculation nearly 80 years before (see [5]). But, again Dirac has all the merit into dis- seminating the notation that nowadays is widespread in quantum mechanics: 1.6. RIEZ REPRESENTATION 23 almost every phenomenon that is explained using quantum mechanics (in- cluding a large portion of what is called modern physics) is usually explained with the help of bra-ket notation . Much beyond of all these bras, part of the (sex)appeal of “Dirac Notation” is the abstract representation-independence it encodes, together with its versatility in producing a specific representation (e.g. x, or p) without much ado, or excessive reliance on the nature of the linear spaces involved. So once again the notation disseminated by Dirac is not only convenient but insightful and intuitive, what made it readily available to a much larger community.
The important guide line that we want to discuss is the (Hilbert Space) Riez Representation lemma. This theorem establishes an important connec- tion between a Hilbert space and its linear functionals, that is a connection between “Bra´s” and “ket’s”. Here we will only describe and discuss the the- orem, but we will not prove it. For a demonstration, we refer to [6]. First note, that for each ϕ in the Hilbert Space , we can define a linear functional H F using the inner product by making, for all ψ , ϕ ∈ H
Fϕψ =(ϕ, ψ). (1.6.19)
A Bra-Ket enthusiast should say that to every Ket (i.e. an element of the Hilbert space) correspond a unique Bra (i.e. a linear functional). The Riez Representation lemma deals with the converse of this statement (i.e. it an- swers the question if for every Bra there is a unique Ket). We say that a linear functional F is bounded if there is a C > 0 such that for all ψ ∈ H Lψ Fxψ = ψ(x), 24 CHAPTER 1. QUANTUM STATES AND OBSERVABLES for all ψ L2, i.e. F is the linear functional that associates each state to its ∈ x value at a given point x. The Riez Representation lemma states that every bounded linear func- tional (a “bounded Bra”) correspond to a vector in the Hilbert Space. More precisely, Theorem. [ Riez Representation Lemma] Let F be a bounded linear functional defined everywhere on a Hilbert space . Then there is ϕ such that ϕ = F and, for all ψ , H ∈ H k k k k ∈ H F ψ =(ϕ, ψ). In other words, Riez Representation lemma say that a bounded linear functional on a Hilbert space is just a inner product. 1.7 Spectral Theorem At this point, after some simple physical-mathematical requirements, we understand that the operator A corresponding to real observable a is self- adjoint. These requirements are actually enough to caracterize the operators in quantum mechanics, that means if A is self-adjoint we can construct the operators χI (A) that allow us to calculate the probabilities that the observ- able lies in the interval I. This is due to the Spectral theorem. The Spectral theorem (for self-adjoint operators) basically sates that any self-adjoint op- erator is unitarily equivalent to a multiplication operator. This result is far from trivial and here we will only state the theorem, explain it and not try to give a proof. For a demonstration we refer to Ref [6]. An orthogonal projection O on a closed subspace S of is a linear op- H erator such that for all ψ S, P ψ = ψ and for all ψ S⊥, P ψ = 0. Any ∈ ⊥ ∈ arbitrary ϕ can be written as a unique decomposition ϕ = ψ + ψ , so ∈ H ⊥ that linearity extend O to hole space (and Oϕ = ψ). Clearly, O is idempotent 1.7. SPECTRAL THEOREM 25 (i.e. O2 = O), since for any ϕ = ψ + ψ , we have O2ϕ = O(Oϕ)= Oψ = ψ. ⊥ Given any two vector ϕi = ψi + ψi (i =1,2), we note that ⊥ (Oϕ1,ϕ2) = (ψ1, ψ2 )+(ψ1, ψ2)=(ψ1, ψ2) ⊥ = (ψ1,Oϕ2)+(ψ1 ,Oϕ2)=(ϕ1,Oϕ2), ⊥ which show that O is Hermitian. Moreover, Oϕ = ψ 6 ϕ , k k k k k k then O is bounded. We note that orthogonal projections satisfy many of the properties of the operators χI of section 1.4. For example, suppose that S and S′ are closed subspaces of . We want to consider O the orthogonal projection on H C = S S′. Let O and O ′ be the orthogonal projections on S and S′ ∩ S S respectively. Clearly, if ψ C, then Oψ = O ψ = O ′ ψ = ψ. Also C⊥ is a ∈ S S subset of both S⊥ and S′⊥, then Oψ = OSψ = OS′ ψ = 0. This leads to ⊥ ⊥ ⊥ the conclusion that O = OSOS′ , and we suggest the reader to compare this with the lemma 1.4.3. For each self-adjoint operator, the Spectral theorem garantes the existence of a family of orthogonal projection,that provides a canonical decomposition, called the spectral decomposition, of the underlying Hilbert space on which the operator acts. More precisely, Theorem. [Spectral Theorem] Let A be a self-adjoint operator on . H There is a family of ortogonal projection (λ) depending on a real parameter EA λ, called a spectral family such that i. λ <λ (λ ) (λ )= (λ ) 1 2 ⇒EA 1 EA 2 EA 1 ii. For ǫ> 0, (λ + ǫ) (λ) as ǫ 0. EA →EA → iii. For ψ ∈ H A(λ)ψ 0 as λ E (λ)ψ → ψ as λ → −∞+ EA → → ∞ 26 CHAPTER 1. QUANTUM STATES AND OBSERVABLES iv. ψ D(A) λ2d E (λ)ψ < + ∈ ⇔ R k A k ∞ R v. For ψ D(A) and φ , ∈ ∈ H (φ, Aψ)= λd (φ, A(λ)ψ) R E Z vi. If f is a complex valued function, the operator f(A) is given by f(A)= f(λ)d A(λ), R E Z defined D(f(A)) consisting of all ψ such that ∈ H 2 f(λ) d A(λ)ψ < + R | | kE k ∞ Z Now back to business (i.e. prove some results). If a is a real valued observable, with corresponding operator A then from the spectral theorem the mathematical expectation a¯ of the observable a for a system at state ψ D(A), is given by ∈ a¯ = λd(ψ, A(λ)ψ). R E Z Furthermore, the probability (a I) that the observable a lies in the Pψ ∈ interval I for a system at state ψ D(A) ∈ (a I)= d(ψ, (λ)ψ). Pψ ∈ EA ZI As we point out, this theorem gives us a prescription to construct the operator χI (A). In fact, using [vi.], we find that χ (A)= d (λ). I EA ZI Note 1.f. For I =( ,λ], we have (λ)= χ (A). −∞ EA I For a concrete example, consider the position of the particle. It is clearly a real valued observable, and therefore there is a corresponding self-adjoint operator Xˆ, given by [Xψˆ ](x)= xψ(x) (1.7.21) 1.7. SPECTRAL THEOREM 27 It is worth to remember that an operator is not defined unless its domain is specified. Here, it is of double importance since we the operators of interest arise form real observables, and therefore are self-adjoint. Even a superficial analysis should convince us that a slightest change in the domain can destroy the self-adjointness. So we need to specify a domain where Xˆ is self-adjoint. The simplest (an largest) domain D(Xˆ), we can choose in order to Xˆ be self-adjoint is is the set of those ψ L2 such that xψ L2. In fact , it is ∈ ∈ easy to see that Xˆ is Hermitian. Or more directly, take an f L2, we see ∈ f ˆ ˆ ˆ that ψ = x i D(X) and (X i)ψ = f, which show that (X i) is onto, ± ± ∈ ± ± ± so the proposition 1.4 holds. Moreover, Xˆ is densely defined, since for any 2 f f L we can take ψ = 2 D(Xˆ). But ∈ ǫ ǫx +1 ∈ f ψ f 6 ǫx2 | | , | ǫ − k ǫx2 +1 which implies that ψ 0 as ǫ 0. Therefore Xˆ is self-adjoint. ǫ → → Now, the Spectral theorem states that there is a family of orthogonal projections (x ) such that E 0 Xˆ = x0d (x0). R E Z Our last remark, show us that the orthogonal projections (x ), are actually E 0 given by [ (x )](x)= H(x x )ψ(x), E 0 − 0 where H(x) is the step function (H(x)=0 for x > 0 and H(x)=1 for x 6 0). We close this chapter stating a last postulate, necessary to the logical self-consistency of our physical theory. The postulate, usually referred “the wave function colapse” is the following: Postulate 3. If the measurement of the observable a of a system in a state ψ gives the result λ immediately after the measurement the new state of the 28 CHAPTER 1. QUANTUM STATES AND OBSERVABLES system became (λ)ψ φ = EA (λ)ψ kEA k This is to say that immediately after a measurement of the observable a, we know for sure the result if we perform the same measurement again. That means, that if we find that a I in our first measurement, the second one ∈ (performed immediately after) will also gives that a I. So the system can ∈ not remain in the same state ψ, otherwise we can only predict the probability of a I in the second measurement. It is the essence of measurement in ∈ quantum mechanics and connects the wave function with classical observables like position and momentum. The colapse is indeed one of the two process by which a quantum system evolve in time, the other one is given by the Schrödinger equation that we will discuss in chapter 3. For a more detailed discussion of the “wave function colapse” we refer to Ref. [7]. 1.8 Exercises 1. Show that eq. 1.1.3 holds. 2. Prove Lemma 1.1.2 and the corollary 1.1.2. 3. Prove the triangle inequality. 4. Give an example of a sequence in which ψ ψ weakly, but not k → strongly. 5. Prove [ii.] and [iii.] of the lemma 1.4.3. 6. Construct the operator χI (Pˆ). 7. Show that the linear functional Fxψ = ψ(x) is not bounded. Chapter 2 The Spectrum Having at our disposal the mathematical formalism capable of describing a quantum system, as discussed in the previous chapter, we are now in a position to aspire a connection between “our” theory and the results obtained in a laboratory. For this purpose it is not enough, altough necessary, to know the quantum state of an experimental object. This is because one hardly measures the state itself but, rather, infers information about it through a set of indirect measurements [8]. Therefore, if we have an electron confined to live in a certain region, a particle in a box, so to speak, in order to extract information about its condition we must design experiments to specify, for instance -and if possible- its position, its momentum, its energy, its (spin) angular momentum, etc. We have seen so far that there exist some restrictions on the outcomes of a measurement, be it due to the uncertainty principle, quantization conditions or the self-adjointness postulate. It becomes, then, paramount to investigate the possible results measured by an experimentalist. This set of allowed outcomes of an observable mea- surement consists what is called the spectrum of the associated operator. In this chapter we would like to shed light in this question. 29 30 CHAPTER 2. THE SPECTRUM 2.1 Spectrum and Resolvent We have seen how in quantum mechanics, we can predict the probability that a real observable a lies in some interval I. We can now ask what values an observable can assume or if it can attain any value. The set of possible values depends on the corresponding self-adjoint operator A, and we call it σ(A) the spectrum of A. To properly answer to this question, we shall apply the spectral theorem to our observables. To do this, we give the Definition 2.A. We say that a scalar λ is in the resolvent set ρ(A) of a closed operator A on if there is a bounded operator R on , called the H λ H resolvent of A, such that Rλ(λ A)ψ = ψ ψ D(A) (λ A−)R ϕ = ϕ, ϕ∈ . − λ ∈ H 1 The definition say that if exists then R =(λ A)− . With this definition λ − we have the Proposition 2.1. If A is a self-adjoint operator then all non-real numbers are in ρ(A) 1 Proof. For every non-real λ, f(x)= λ x , is bounded and continuous over the − real line. So by the spectral theorem 1 f(A)= d λ(A) R λ x E Z − is a bounded operator defined everywhere. More over f(A) satisfies the requirements on definition 2.A. The key point here is that if an real observable can not attain (in some sense) a certain value λ0 then λ0 lies in the resolvent of the corresponding operator. More precisely, we have the Proposition 2.2. let A be a self-adjoint operator and I an open interval of the real line. If χ (A)=0 then I ρ(A). I ⊂ 2.1. SPECTRUM AND RESOLVENT 31 Proof. Let λ I, since I is open, δ the minimum distance between λ and 0 ∈ 0 the ends of I is positive. So define 1 x λ λ0 x 6∈ f(x)= − 0 x λ ∈ 1 Clearly, in the real line, f(x) is piecewise continuous and bounded by δ . Thus by the spectral theorem 1 f(A)= d λ(A) R I λ0 x E Z − − is a bounded operator defined everywhere. Furthermore, f(A)(λ0 A)ψ = (1 χI (A))ψ = ψ, ψ D(A) (λ A)−f(A)ϕ = (1 − χ (A))ϕ = ϕ, ∀ϕ ∈ . 0 − − I ∀ ∈ H Therefore λ ρ(A). Since this holds for all λ I, we that I ρ(A). 0 ∈ ∈ ⊂ The converse of this theorem is also true. This leads us to the Proposition 2.3. For any real λ ρ(A) there is an open interval I R 0 ∈ ⊂ such that λ I and χ (A)=0. 0 ∈ I Proof. Suppose there is no such interval. Therefore there must be a sequence of open intervals I such that λ I and χ (A) =0. So there is a non- n → 0 ∈ n In 6 zero ψ , such that χ (A)ψ = ψ . So the Spectral theorem implies n ∈ H In n n (λ0 A)ψn = (λ0 A)χIn (A)ψn 6 sup λ0 λ 0. k − k k − k λ In | − |→ ∈ But since λ ρ(A) this is an absurd, because 0 ∈ 1 1 ψ = (λ A)− (λ A)ψ = (λ A)− (λ A)ψ 0, k nk k 0 − 0 − nk k 0 − kk 0 − nk→ which contradicts the fact that ψ > 0. k nk Corolary 2.1.1. If A is self-adjoint, then ρ(A) is an open set. 32 CHAPTER 2. THE SPECTRUM Proof. Let λ ρ(A). If λ is not real, then there is a small open disk D with ∈ λ center in λ which contain no real numbers. So by the proposition ??, we have that D ρ(A). Now, if z are real then the proposition 2.3 say that there λ ⊂ is an open interval I ρ(A) with center in λ. The length of this interval is ⊂ the diameter of the open disk D centered in λ, and clearly D ρ(A). I I ⊂ The points that are not in the resolvent set are in the spectrum. In other words: Definition 2.B. The set σ(A) , called the spectrum of A is the set C ρ(A). − That means that σ(A) consist of those points λ in that the operator (λ A) is not invertible. So a straightforward consequence of corollary is the − Corolary 2.1.2. If A is a self-adjoint operator, then the spectrum σ(A) is a closed set. The main reason we go through all this mess is that spectrum is a fun- damental object in quantum theory. Its role becomes clear in the Proposition 2.4. An observable can only assume values in the spectrum of the corresponding operator. Proof. Let A be the operator corresponding to the observable a. For λ 0 ∈ ρ(A), by Proposition 2.3 there is an interval I such that χi(A)=0. Following section 1.4, we have P (a I)=(χ (A)ψ, ψ)=0, ψ ∈ I for any ψ . Therefore, a can not assume values in I for any state ∈ H function. So when faced with an observable in quantum mechanics (e.g. we wish to make a measurement on a physical system) , we must consider the corre- sponding operator to this observable and (possibly ) the first thing we should 2.1. SPECTRUM AND RESOLVENT 33 look at is the spectrum of this operator. In fact, as the proposition above states, we can only measure values in the spectrum of the operator! Thus, a criteria for finding the spectrum of an self-adjoint operator is essential to quantum mechanics. To develop this criteria, we will need some preliminary results: Lemma 2.1.1. A closed vector subspace S of a Hilbert Space is it self a H Hilbert Space. Proof. We just net to prove that S is complete. Let ψ S be a ⊂ H k ∈ Cauchy sequence. Since is complete ψ ψ . But, since S is closed H k → ∈ H then it contains all limit points, therefore ψ S. So all Cauchy sequences of ∈ elements of S converges to an element in S. So S is a complete inner product space, i.e. a Hilbert space.. We now state a very useful theorem for determine the sepctrum of a self-adjoint operator: Proposition 2.5. Let A be a self-adjoint operator. λ R lies in σ(A) if ∈ and only if there is a sequence ψ D(A) such that, ψ =1 and k ∈ k kk (λ A)ψ 0. (2.1.1) k − kk→ Proof. Suppose λ σ(A) and eq. 2.1.1 holds. Then λ ρ(A) and therefore 6∈ ∈ R (λ A)ψ = ψ , λ − k k a contradiction with eq. 2.1.1. Now suppose that eq. 2.1.1 does not hold. Our strategy is to show that (λ A) is invertible, so λ ρ(A). To do this we claim that there is a constant − ∈ C such that ψ 6 C (λ A)ψ . (2.1.2) k k k − k If not, there must be a sequence ϕ D(A) such that k ∈ ϕ k kk . (λ A)ϕ →∞ k − kk 34 CHAPTER 2. THE SPECTRUM ϕk So we can take ψk = ϕ , and eq. 2.1.1 holds (which contradicts our as- k kk sumption). Form eq. 2.1.2, we have that (λ A) is injective. If it was not injective − then there must be ϕ = ϕ such that 1 6 2 (λ A)(ϕ ϕ )=0, − 1 − 2 which contradicts eq. 2.1.2. Furthermore, consider a sequence ψ R(λ A), such that ψ ψ . n ∈ − n → ∈ H Since (λ A) is injective, we call ϕ the unique solution of (λ A)ϕ = ψ . − n − n n Now, eq. 2.1.2 shows that the sequence ϕn is a Cauchy sequence, in fact, ϕ ϕ 6 C ψ ψ 0. k k − nk k k − nk→ Thus ϕ ϕ . So for f D(A), we have n → ∈ H ∈ (ϕ, (λ A)f) = lim(ϕ , (λ A)f) = lim(ψ , f)=(ψ, f). − n − n Therefore, ϕ D(A) and (λ A)ϕ = ψ. Then we conclude that R(λ A) ∈ − − is closed, and by proposition 2.1.1 it is a Hilbert Space. Let f be an arbitrary vector of and w R(λ A). We take v to be H ∈ − the (unique) solution of (λ A)v = w, and construct the linear Functional − F on R(λ A) by defining − Fw =(v, f). This is a bounded linear functional, since by eq. 2.1.2 , we have Fw 6 v f 6 C f w . | | k kk k k kk k So by Riez representation lemma, there is a u R(λ A) such that Fw = ∈ − (u,w), w R(λ A). Thus, we find taking vinD(A), ∀ ∈ − ((λ A)v),u)=(f, v). − The self-adjointness of A implies that u D(A) and (λ A)v = f. Therefore ∈ − f R(λ A) and = R(λ A). So (λ A) is one-to-one (injective) and ∈ − H − − surjective, therefore invertible, which implies λ σ(A). 6∈ 2.2. FINDING THE SPECTRUM 35 As a consequence we have Corolary 2.1.3. Self-Ajoint operators are closed. Proof. We left this proof as an exercise. Corolary 2.1.4. If A is a closed operator and v 6 C Av then A is k k k k injective and R(A) is closed. Proof. Do λ =0 in the proof of the proposition 2.5. There are some special elements in the subset of σ(A) called the eigen- values. Definition 2.C. λ is said to be an eigenvalue of A if there is a non-vanishing solution for Aψ = λψ. When this non-vanishing solution ψ exist, we call it an eigenvector of the operator A corresponding to the eigenvalue λ Note 2.a. The prefix eigen- is adopted from the German word eigen for “own- ”, “unique to”, “peculiar to”, or “belonging to” in the sense of “idiosyncratic” in relation to the originating object(operator). Therefore it is usual to carry it to more specific objects, for example: eigenfunctions in L2 spaces or eigenstate in Quantum mechanics ... With this definition it is clear that Corolary 2.1.5. If λ is an eigenvalue of A then λ σ(A). ∈ Proof. It is straightforward, if λ is an eigenvalue then there is ψ = 0, such 6 that (λ A)ψ =0. So (λ A) is not injective, and λ σ(A). − − ∈ 2.2 Finding the spectrum We now want to apply these ideias to find the spectrum of some operators of interest. we consider as an initial example the position operator, discussed 36 CHAPTER 2. THE SPECTRUM in eq. 1.7.21. The easiest elements are the eigenvalues. Thus we should look for solution of Xψˆ = λψ . xψ(x) = λψ(x) ⇒ Thus, there are no eigenvalues, since the only solution of the equation above is ψ = 0 for x = λ. That means ψ(x) almost everywhere, and to us 6 this means ψ = 0! So we need to use proposition 2.5. In order to do this, it worth to have some sort of motivation to construct the sequence in the statement of proposition 2.5. An object that can satisfies ψ = 0 for x = λ, 6 is the (infamous) Dirac’s Delta function. So we can think in a sequence such that ψ (x) δ(x λ), and ψ =1. These sequences are some times called k → − k kk a delta sequence. For example, we can take 2 1 k (x λ)2 ψ (x)= e− 2 − , k N(k) where N(k) is such that ψ =1, that means N 2(k)= √π . Clearly, k kk k 1 2 y2 λ Xˆ ψ y e− dy , ( ) k = 2 3 0 k − k N (k)k R → Z as k . So for any λ real we have that λ σ(Xˆ). That means the → ∞ ∈ spectrum of the position operator is the whole real line. The argument for the momentum operator is similar. But we shall use the Fourier transform and the identity of eq.1.1.6. We leave this as an exercise. You should find that the spectrum of the momentum operator is the whole real line. 2.2.1 The spectrum of the Hamiltonian We are now, interested in the spectrum of the Energy operator or the Hamiltonian P 2 H = + V (Xˆ). 2m To consider this operator we will need the following lemma. 2.2. FINDING THE SPECTRUM 37 Lemma 2.2.1. If A is a self-adjoint operator, then A2 is a self-adjoint op- erator. Proof. We have that A2 +1=(A + i)(A i), here we stress that the domain − of both operator are D(A2). If A is self-adjoint, then for each f there is ∈ H u D(A2) such that (A2 + 1)u = f. Furthermore since i ρ(A) for each ∈ ± ∈ f there are v D(A) such that (A i)v = f and (A + i)v+ = v . Since ± ∈ − − − both v are in D(A), it is also true for Av . Therefore, R(A2 + 1) is the ± − whole space and therefore A2 is self-adjoint. This show that P 2 H = 0 2m is self-adjoint. H0 is the (Hamiltonian) energy operator of a free particle. That is a particle that is not subject to a external potential. To study its spectrum, we write 1 (λ H )= (√2mλ P )(√2mλ P ), − 0 2m − − and observe that √2mλ R √2mλ ρ(P ). So for λ < 0, we have that ∈ ⇒ 6∈ 2 λ ρ(H ). Now, consider the case λ > 0. We write λ = ~k , again we search ∈ 0 2m for eigenvalues so ~k2 H0ψ = 2m ψ dψ 2 dx2 = k ψ ⇒ − +ikx ikx ψ(x) = c+e + c e− , ⇒ − 2 The only way this ψ(x) can lie in L is that c+ = c =0! Thus again, there − are no eigenvalues! Again we employ can employ proposition 2.5 to find the spectral points. Not surprisingly, the method is quite analogue the one you probably use to find the spectrum of the momentum operator. Let 1 x ψ (x) == ϕ eikx, n √n n 1 x2 where ϕ(x)= N e− , and N is such that φ(x)2dx =1. R Z 38 CHAPTER 2. THE SPECTRUM So 1 x 2 ψ = ϕ dx = ϕ(y) dy =1. k nk n n | | Z Z Moreover, d 1 x ikx dx ψn(x) = ikψn(x)+ n−3/2 ϕ′( n )e d2 2 2ik x ikx 1 x ikx 2 ψ (x) = k ψ (x)+ 3 2 ϕ′( )e + 5 2 ϕ′′( )e dx n − n n− / n n− / n So we have that 2ik x 1 ψ k2ψ ϕ ϕ . n′′ + n = 1/2 ′( ) + 3/2 ′′ k k kn− k n k n− k 2 And therefore, as n , ψ′′ + k ψ 0. Thus, λ > 0 lies in σ(H ). → ∞ k n nk → 0 Therefore the positive real axis is contained in the energy spectrum of the free particle. Since the spectrum is a closed set, 0 is also an element of σ(H0). This proves the Proposition 2.6. σ(H ) = [0, + ) 0 ∞ To consider the case of a particle subject to a potential, must consider the operator V (Xˆ). Again, D(V ) the simplest and largest domain of V is the subset of such that ψ D(V ) and V ψ . Once we fix this domain, H ∈ ∈ H we know the domain of H, which is given by D(H ) D(V ). And here comes 0 ∩ the problem. Is H self-adjoint ? In fact, there are many examples that show us that with no further re- strictions on V , it is not true that (in general) H is self-adjoint. At first this seems to be a terrible failure to our theory. But actually, this only raise questions and should be taken as an opportunity for research. This is one of the central problems of modern mathematical physics. One could fill libraries with the mathematics generated trying to answers this question. Mathematically, this is the question of when the sum of the Laplace operator and a multiplication operator on L2 is self-adjoint (more precisely, essentially self-adjoint). A more particular and immediate question is the following: Is there some potential such that H is self-adjoint? 2.2. FINDING THE SPECTRUM 39 (Un)Fortunately, the proposition above describes a large class of poten- tials such that H is self-adjoint. Proposition 2.7. If for all ψ D(H ) D(V ), there are constants a < 1 ∈ 0 ∩ and b such that, V ψ 6 a H ψ + b ψ , k k k 0 k k k then H = H0 + V is self-adjoint. To prove this theorem we use the two following results of general interest. Lemma 2.2.2. If B is a bounded linear operator on and B < 1, then H k k there is a bounded linear operator N on such that H N(1 B) = (1 B)N =1 − − 1 N 6 1 B − k k k − k Proof. The operator N is the inverse of the operator (1 B), an is given by − the Neumann series 1 2 3 N = (1 B)− =1+ B + B + B + .... − Lemma 2.2.3. Let A be a densely defined Hermitian operator on . Suppose H there is a complex number z such that R(z A) = R(z† A). Then A is − − self-adjoint. Proof. The hypotheses implies that (1 + A2) be onto, therefore A is self- adjoint. We leave the details as an exercise. Now, back to proposition 2.7. Proof. Follows from the Lemmas 2.2.2 and 2.2.3. In fact, for λ R and ∈ ψ D(H ) D(V ), we have ∈ 0 ∩ (H iλ)ψ 2 = H ψ 2 + λ2 ψ 2. k 0 − k k 0 k k k 40 CHAPTER 2. THE SPECTRUM The self-adjointness of H and proposition 2.1 implies that if λ =0, for every 0 6 ϕ , we can find ψ D(H ) such that (H iλ)ψ = ϕ. Thus ∈ H ∈ 0 0 − 2 1 2 2 1 2 ϕ = H (H iλ)− ϕ + λ (H iλ)− ϕ . k k k 0 0 − k k 0 − k We can see that 1 H0(H0 iλ)− ϕ 6 ϕ k − 1 k k k λ (H iλ)− ϕ 6 ϕ . k 0 − k k k If V satisfies the hypotheses of proposition 2.7 then 1 1 1 b V (H iλ)− ϕ 6 a H (H iλ)− ϕ + b (H iλ)− ϕ 6 (a + ) ϕ . k 0 − k k 0 0 − k k 0 − k λ k k | | Since a< 1, we can find a sufficiently large λ such that we have 1 b V (H iλ)− 6 (a + ) 6 1. k 0 − k λ k | | So we can write 1 1 H iλ = (1+ V (H iλ)− )− (H iλ). − 0 − 0 − The right hand side is onto,hence R(H iλ) = for λ sufficiently large. − H | | The adjointness of H follows from lemma 2.2.3. Proposition 2.7 gives a useful way to answer the question on if H is self- adjoint. To do this, we must know when the potential satisfies the hypotheses of the proposition. This can be done with the following criteria, which we state without demonstration. Theorem. [A Criteria for self-Adjointness] The following statements are equivalent i. D(H ) D(V ). 0 ⊂ ii. V ψ 2 6 C( H ψ 2 + ψ 2) k k k 0 k k k iii. C = sup x+1 V (x) 2dx < + 0 x x | | ∞ R 2.3. EXERCISES 41 iv. For ǫ> 0, there is a constant K such that V ψ 2 6 ǫ H ψ 2 + K ψ 2 k k k 0 k k k v. For ǫ> 0, there is a constant C such that V ψ 6 ǫ H ψ + C ψ k k k 0 k k k So if any of this statements holds then H is self-adjoint. In any case, the spectrum of H will depend on a fundamental way on the potential V . 2.3 Exercises 1. Prove Corollary 2.1.3 2. Show that the spectrum of the momentum operator is the real line. 3. Give a detailed proof of Lemma 2.2.3. 4. Show that λ< 0 lies in ρ(H0). 5. Prove that 0 is not an eigenvalue of H0. Chapter 3 Quantum Dynamics One important point we should stress is that time is just a parameter in quantum mechanics, that is it is not an operator. That means time is not an observable in the language of the previous chapters. So it makes no sense to treat time in the same way as we treat the position or momentum operators. In this chapter, we analyse how a quantum state ψ change in time, and how the probabilities and expected values depend on the time parameter. 3.1 Time evolution and Schrödinger Equation So we need to look for a time evolution operator U such that U(t t )ψ(t )= ψ(t) − 0 0 We expect that time is a continous parameter so that for a quantum state ψ lim ψ(t)= ψ(t0) t t0 → Therefore the time evolution operator U must be continous on t and lim U(t t0)= U(0) = I, t t0 → − Also, a quantum state reamains a quantum state as time goes by so ψ(t) 2 = ψ(t), ψ(t) = U(t)ψ(0), U(t)ψ(0) = U(t)ψ(0) 2 = ψ(0) 2 k k k k k k 43 44 CHAPTER 3. QUANTUM DYNAMICS Therefore U †(t)U(t)= I (3.1.1) An operator obeying eq.3.1.1 is called a unitary operator . Another necessary feature is that if we are interested in obtaining time evolution from t0 to t2, then we can obtain the same result by considering time evolution first from t0 to t1 and then from t1 to t2, ψ(t ) = U(t t )ψ(t ) 2 2 − 1 1 ψ(t ) = U(t t )U(t t )ψ(t ) 2 2 − 1 1 − 0 0 Therefore, for t2 > t1 > t0 U(t t )= U(t t )U(t t ). 2 − 0 2 − 1 1 − 0 An operator with this property is said to be strongly continous. Another physical requirement concerns energy conservation that if the potential V does not explicitly depend on the time then the total energy must be conserved. In our quantum mechanical setting this means that for a given state ψ the mathematical expectation of the energy and the probability that the energy lies in an interval I does not depend on time. Note 3.a. A final requirement is the inertia principle, which means that if the potential V is invariant under translations then the momentum must be conserved. In our quantum mechanical setting this means that for a given state ψ the mathematical expectation of the energy and the probability that the momentum lies in an interval I does not depend on time. We sumarize our demands on the Postulate 4. The time evolution of a 1-D quantum system from the time 0 to a time t is given by an operator U(t) such that 3.1. TIME EVOLUTION AND SCHRÖDINGER EQUATION 45 i. U(t) is continous on t. ii. U(t1 + t2)= U(t1)U(t2) iii. U †(t)U(t)= I iv. If the potential does not depend explicitly on t then ψ, U † H U ψ = ψ,Hψ , Now we can ask about U Lemma 3.1.1. Let A be a self-adjoint operator and let ∞ ( it)n f (A) = exp( itA)= − An t − n! n=0 X then D (f (A)) = 2 and the following holds t L i. ft1+t2 (A)= ft2 (A)ft1 (A) ii. f(t)ψ ψ as t 0. → → U(t)ψ ψ iii. − iAψ as t 0, for all ψ D(A). t → − → ∈ U(t)ψ ψ iv. if limt 0 − exists then ψ D(A). → t ∈ Proof. Follows directly from the spectral theorem. Our next result (also know as the Stone’s theorem on one-parameter unitary groups) establishes a one-to-one correspondence between self-adjoint operators on a Hilbert space and one-parameter families of unitary oper- H ator that are strongly continous. To do this we need the converse of Lemma 3.1.1, this result is due to Stone and we state it here without demonstration (for a proof see [9]). 46 CHAPTER 3. QUANTUM DYNAMICS Theorem. [Stone’s theorem] Let f(t)t R be a strongly continuous one- ∈ parameter unitary group on a Hilbert space . Then there exists a unique H self-adjoint operator A on such that f(t)= eiAt. H We know that the time evolution operator U(t) must have the form eiAt for some self-adjoint operator A. The energy conservation requiriment will gives us who is A. Lemma 3.1.2. If the potential does not explicitly depend on t, then H = U †(t)HU(t). Proof. From Stone’s Theorem, U(t)= eiAt where A is a self-adjoint operator. Now let H′ = U †(t)HU(t), it easy to see that the self-adjointness of H implies that H′ is self-adjoint. The energy conservation requairiment of Postulate 4 implies that the expected value E¯ does not depend on t, that is E¯(t) = (ψ(t), Hψ(t))=(U(t)ψ(0), HU(t)ψ(0)) = (ψ(0), U(t)†HU(t)ψ(0)) = (ψ(0),H′ψ(0)) must coincide with E¯(0) = (ψ(0), Hψ(0)). Thus (U(t)ψ,HU(t)ψ)=(ψ, U †(t)HU(t)ψ)=(ψ,H′ψ)=(ψ,Hψ), so for all ψ D(H) we find that (ψ, (H′ H)ψ)=0 therefore H′ = H. ∈ − In fact, taking ψ(0) as an eigenstate of the Hamiltonian, Hψ(0) = Eψ(0), then as we let it evolve in time, U(t) Hψ(0) = E U(t)ψ(0) = E ψ(t), 3.1. TIME EVOLUTION AND SCHRÖDINGER EQUATION 47 it should still remain an eigenstate of the system at a later time, if the energy is to be conserved, Hψ(t)= Eψ(t). Comparing these expressions we get U(t) Hψ(0) = Hψ(t)= H U(t)ψ(0), that is, the object U(t) must commute with the Hamiltonian, [H, U(t)] = H U(t) U(t) H =0. − Having in mind all these restrictions one has to impose on the evolution operator, it must be of a general form i A t U(t)= e− ~ , with A being a symmetry of the problem, [H, A]= HA AH =0. − The latter can, in principle, be considered a function of the Hamiltonian H and be expanded as n A = A(H)= cnH . n X For simplicity, we make use of the correspondence principle and argue that, similarly to what happens in classical mechanics where the evolution is governed by the Hamiltonian function itself, the quantum evolution operator can be given by i Ht Proposition 3.1. U(t)= e− ~ , where H is the Hamiltonian. It acts as a group element on the initial state and the Hamiltonian op- erator belongs to an algebra. Once the Hamiltonian is determined, from a physical point of view, in terms of relevant operators - such as those repre- senting the observables for position and momentum, X and P , for instance 48 CHAPTER 3. QUANTUM DYNAMICS - there is still freedom in the choice of representation. The evolution of the evolution operator itself is then given by ∂ i~ U(t)= H U(t). ∂t On the other hand, ∂ ∂ ∂U(t) i~ ψ(t)= i~ U(t)ψ(0) = i~ ψ(0) = H U(t) ψ(0) ∂t ∂t ∂t so a quantum state evolves deterministically according to the linear Schrödinger equation. ∂ Hψ = i~ ψ, (3.1.2) ∂t If the Hamiltonian is time-dependent one has to consider the following formal solution in terms of a time-ordered exponential H ψ(t) = p exp i dt ψ (0) − ~ | i Z t t t′ H(t) 1 H(t′) H(t′′) = 1 i dt dt′ dt′′ + ψ(0). − ~ − 2 ~ ~ ··· Z0 Z0 Z0 ! In situations where the Hamiltonian is time-independent, we have a re- duction to the simpler expression discussed above, so that the state evolves according to i Ht ψ(t)= e− ~ ψ(0). Note 3.b. It worth to note that any quantum-mechanical object is completely characterized by the state function and the time-evolution of state function is completely deterministic. Everything, the system, the equipment, the en- vironment, and the observer are part of a (quite complex) state vector of “universe”. The measurements with different results are part of state func- tions at different points of the spacetime, furthermore the measurement is a complicated process involving interactions between the system and equipment. And the equipment alone has something of order of 1023 degrees of freedom! 3.2. APPLICATIONS TO TWO-LEVEL SYSTEMS 49 That means we neither know nor are able to compute the states of the equip- ment that we use to make the measurement. That is why some people say that quantum mechanics is deterministic, but it is also probabilistic: that means you can deterministically calculate the probability of the outcome of an ex- periment. This is to distinguish it from non-deterministic (i.e. stochastic) systems where you do not generally have “one” solution but an entire family of solutions depending on random variables. Historically, two equivalent representations were introduced at the dawn of the quantum theory. The first was Heisenberg’s attempt to reproduce experiments with the launch of infinite dimensional matrices to describe co- ordinates, momenta and energies. Independently, Schrödinger proposed a formalism based on wave equations and differential operators to represent some observables. Later it became clear that Schrödinger’s differential opera- tors correspond to a neat way of representing Heisenberg’s more cumbersome matrices of infinite dimensions. There is also a difference in the perspective between both approaches, which will become more evident in what follows, making that of Schrödinger more closely related to equation (3.1.2). We start by presenting this formu- lation. Note 3.c. In Heisenberg’s formalism the observables vary in time whereas the vectors remain fixed. 3.2 Applications to Two-Level Systems Now that we have set out the basic framework to describe quantum phe- nomena and explored some relevant consequences of the theory, it is time to discuss in some detail a number of important and elementary physical sys- tems. From a mathematical point of view, it is simpler to tackle a quantum problem which admits a finite dimensional representation. The first example 50 CHAPTER 3. QUANTUM DYNAMICS to investigate is, thus, a two level problem for which the quantum state can be in only two states. A 2 2 Hilbert space can be used, for example, to reproduce the behaviour × 1 of a quantum spin 2 , which can be in the states up and/or down. Nonetheless it can also be used to describe, approximately only, a larger system but for which there are two states that are almost decoupled from the remaining states. For a finite dimensional vector, charactering a particular quantum state, the normalization of the probability modifies according to 2 2 ψ(x, t) dx =1 ψi(t) =1. R | | −→ | | i Z X A suitably normalized basis to describe vectors in this Hilbert space is, for example, 1 0 ψ = , ψ = , ↑ 0 ↓ 1 with ψ†, ψ , = 1 and ψ†, ψ , =0. If, in this orthonormal basis, the energy ↑ ↓ ↑ ↓ ↓ ↑ ↑ ↓ observable, specified by a Hamiltonian operator, has a general Hermitian form, (φ,Hψ)=(Hφ,ψ), α β eiδ H = iδ β e− γ then the evolution of a given state ψ is given by the Schrödinger equation ~ ∂ i ∂t ψ = Hψ. Taking the following combinations θ i δ θ i δ ψ+ = cos e 2 ψ + sin e− 2 ψ , 2 ↑ 2 ↓ θ i δ θ i δ ψ = sin e 2 ψ cos e− 2 ψ , − 2 ↑ − 2 ↓ 2β with the convenient reparametrization θ = arctan α γ , the equation of − motion can be written as ∂ i~ ψ = E ψ . ∂t ± ± ± The advantadge of introducing these states above, denoted eigenvectors, is that they evolve in a very simple way, E i ± t ψ (t)= e− ~ ψ (0), ± ± 3.2. APPLICATIONS TO TWO-LEVEL SYSTEMS 51 governed by E , the associated eigenvalues, ± α + γ α γ E = − sec θ. ± 2 ± 2 Therefore, if one performs a measurement the particle can only be in either of the two possible states, ψ+ with energy E+, or ψ with energy E . − − But if one does not measure its state it can be in linear combination of both eigenstates. The particle can evolve from one state to another and we can define the transition frequency between the states as E+ E ω = −~ − . If we let these up and down states evolve, it is convenient to express it in terms of the eigenvectors, iδ θ θ ψ = e− 2 cos ψ+ + sin ψ , ↑ 2 2 − + iδ θ θ ψ = e 2 sin ψ+ cos ψ , ↓ 2 − 2 − so the action of the evolution operator becomes almost trivialized, iδ i i θ ~ E+ t θ ~ E− t U(t) ψ = e− 2 cos e ψ+ + sin e ψ , ↑ 2 2 − iδ i i + θ E+ t θ E− t U(t) ψ = e 2 sin e ~ ψ+ cos e ~ ψ . ↓ 2 − 2 − Now we are in a position to compute the probability of a initial state ψ ↑ being in a state ψ+, or ψ , after an interval t, − 2 δ i 2 θ i 2 ~ E+ t 2 θ ψ+† U(t) ψ = cos e e = cos , ↑ 2 2 2 δ i 2 θ i E+ t 2 θ † 2 ~ ψ U(t) ψ = sin 2 e e = sin 2 , − ↑ respectively, as well as the probability of a initial state ψ transitioning into ↓ a state ψ+, or ψ , − 2 δ i 2 θ i 2 ~ E− t 2 θ ψ+† U(t) ψ = sin e− e = sin , ↓ 2 2 2 2 θ i δ i E t 2 θ † 2 ~ − ψ U(t) ψ = cos 2 e− e = cos 2 . − ↓ In both situations, once we have a initial state prepared, be it an up or a down state, and let it evolve, after a while any energy measurement can 52 CHAPTER 3. QUANTUM DYNAMICS only produce the values E , associated to states ψ , respectively. Thus the ± ± probability of being in either of these eigenstates sum to unity, as is clear from the expressions above. Note that the probability of having an up state or a down state transtioning to an eigenstate does not vary with time and depends only on how much how close the up and down states are close to being an eigenstate, or ultimately on α,β,γ. Moreover we can calculate the transition amplitudes of starting with up and down states, suitably prepared, letting them evolve under the current Hamiltonian operator and measuring, after some time has elapsed, if the particles are again in one of the up and down states, respectively, ω ω ω i 2 θ i 2 t 2 θ i 2 t i 2 t ~ E∓ t ψ†, U(t) ψ , = cos e + sin e− e e , ↑ ↓ 2 2 ↑ ↓ θ θ ω i ω t i E t † 2 ~ − ψ , U(t) ψ , = 2i sin 2 cos 2 sin 2 t e e , ↓ ↑ ↑ ↓ to show that the probability of being at the initial state after a while varies with time, P (t)=1 sin2 θ sin2 ω t (0, 1), ↑↑ − 2 ∈ and similarly for the probability of transmutation between states, P (t) = sin2 θ sin2 ω t (0, 1). ↑↓ 2 ∈ As expected, since there is no further degree of freedom in the system, the total probability of transition is conserved, P (t)+ P (t)=1, ↑↑ ↑↓ at any instante, but the interference pattern between initial and final states varies harmonically with time. 3.3 Schrödinger’s wave equation In the previous section we discussed one of the simplest quantum prob- lems, for which the states are described by vectors living in a two-dimensional 3.3. SCHRÖDINGER’S WAVE EQUATION 53 Hilbert space. Since the particles can be in only two linearly independent states, the physical observables correspond to 2 2 matrices. on the other × hand, the particle has infinitely many degrees of freedom, matrix represen- tations are not the most convenient and vectors are replaced by functions. For particles restricted to move in one dimension and the equation govern- ing the quantum phenomena is the Schrödinger equation given in eq. 3.1.2. Using the momentum operator of chapter 1, this can be written as a linear partial differential equation ∂ ~2 ∂2 i~ ψ(x, t)= + V (x, t) ψ(x, t), ∂t −2m ∂x2 which simply corresponds to a representation of the fundamental equation (3.1.2) in terms of space-time coordinates. The problem of determining the evolution of the probability distribution for a quantum particle is equiva- lent to solving a PDE with suitable boundary conditions. This is a fairly nontrivial task unless the interaction has a simple form. Our first result in this context is the Lemma 3.3.1. If the potential V is time-independent, the state function evolve as a (infinite dimensional) linear combination of i E t ψ(x, t) = ϕ(x)e− ~ , where ϕ and E are given by ~2 ∂2 ϕ + V (x)ϕ = Eϕ. −2m ∂x2 Proof. If the potentials are time-independent, V (x, t) = V (x), we can sepa- rate variables as ψ(x, t)= ϕ(x)φ(t) so that we have ~2 1 ∂2 1 ∂φ ϕ + V (x)= i~ = E. −2m ϕ ∂x2 φ ∂t 54 CHAPTER 3. QUANTUM DYNAMICS The time-dependence can be immediately obtained, i E t φ(t)= φ(0)e− ~ , and the complete solution has the general form i E t ψ(x, t) = ϕ(x)e− ~ . Note 3.d. We can see that the probability amplitude is not affected by the flow of time, since ψ(x, t) = ϕ(x, 0) . | | | | The solution of Schrödinger’s time-independent second order differential equation, ~2 ∂2 ϕ + V (x)ϕ = Eϕ, (3.3.3) −2m ∂x2 can be expanded in terms of two linearly independent functions, ϕ(x)= c1ϕ1(x)+ c2ϕ2(x), (3.3.4) complemented by appropriate boundary conditions, and the following re- quirements on the wave function: ϕ(x) and ϕ′(x) must be finite • ϕ(x) and ϕ′(x) must be single-valued • ϕ(x) and ϕ′(x) must be continuous • Note 3.e. As we discuss in chapter 1, to lie in L2 the state functions must vanish both at x + and x (or in general, it must satisfy appro- → ∞ → −∞ priate boundary conditions), therefore not any value of E is allowed. Instead, the boundary conditions usually impose restrictions on the energy eigenval- ues, denoted quantization conditions. In what follows we will present some properties which are consequences of Schrödinger’s wave equation. 3.3. SCHRÖDINGER’S WAVE EQUATION 55 Lemma 3.3.2. If the potential V (x) is even, then the solutions of the time- independent Schrödinger equation have a definite parity Proof. Consider the time-independent Schrödinger equation, ~2 ∂2 ϕ(x)+ V (x)ϕ(x)= Eϕ(x), −2m ∂x2 we can see that applying a parity transformation on it, we have ~2 ∂2 ϕ( x)+ V ( x)ϕ( x)= Eϕ( x). −2m ∂x2 − − − − If the potential is invariant under parity, then V ( x)= V (x),ϕ( x) satisfies − − the same equation as ψ(x), so both ϕ(x) and ϕ( x) are admissible solutions − to the same problem, up to boundary conditions. Thus, the solutions of a parity-symmetric problem can be expressed in terms of symmetric and anti- symmetric combinations, ϕ(x) ϕ( x) ϕ (x)= ± − , ± √2 leading to solutions with a definite parity: either even or odd. Now, we would like to show that the structure of the Schrödinger equation implies the existence of an important conservation law, that associated to the probabilistic interpretation of the quantum theory. In fact, there would be an unconciliable problem with Max Born’s probabilisitic interpreation, the so called Copenhagen interpretation, if it was not compatible with Schródinger’s evolution equation (see Ref. [10]). Proposition 3.2. If the potential is a real function and ψ D(H), then ∈ 2 2 2 ~ ∂ ψ ∂ ψ∗ ∂ ψ∗ ψ = +i~ (ψ∗ψ) . (3.3.5) −2m ∂x2 − ∂x2 ∂t Proof. Taking the fundamental equations for ψ(x, t) and ψ†(x, t) ~ ∂ H ψ = +i ∂t ψ, ∂ ψ†H† = i~ ψ†, − ∂t 56 CHAPTER 3. QUANTUM DYNAMICS We see that, for ψ D(H), ∈ ∂ (ψ,Hψ) (Hψ, ψ)=+i~ (ψ, ψ), − ∂t so the total probability (ψ, ψ) is preserved if the Hamiltonian is Hermitian. On the other hand, apply a complex conjugation transformation to Schrödinger’s time-dependent equation and we obtain the equation satisfied by ψ∗, 2 2 ~ ∂ ψ∗ ∂ψ∗ + V (x)∗ψ∗ = i~ . −2m ∂x2 − ∂t Then multiplying them from the left and right by ψ∗ and ψ, respectively, and subtracting one from the other we get 2 2 2 ~ ∂ ψ ∂ ψ∗ ∂ψ ∂ψ∗ ψ∗ ψ + ψ∗ (V (x) V (x)∗) ψ =+i~ ψ∗ + ψ −2m ∂x2 − ∂x2 − ∂t ∂t and if the potential is a real function, V (x)= V (x)∗, it simplifies to 2 2 2 ~ ∂ ψ ∂ ψ∗ ∂ ψ∗ ψ =+i~ (ψ∗ψ) . −2m ∂x2 − ∂x2 ∂t Notice that the expression above corresponds to a conservation law, since it is of the form of a continuity equation, ∂ρ ∂J + =0, (3.3.6) ∂t ∂x associated to the conservation of the total charge, the total quantum proba- bility, Q = ρ dx = ψ∗ψ dx =(ψ, ψ) (3.3.7) Z Z since its variation vanishes, ∂Q + + J ∞ =0, (3.3.8) ∂t −∞ if the associated current, i~ ∂ ∂ J = ψ ψ∗ ψ∗ ψ , (3.3.9) 2m ∂x − ∂x 3.4. TIME DEPENDENCE OF EXPECTED VALUES 57 is localized, vanishing at the faraway boundaries. The continuity equation above shows that there is a flow of the object J (inwards or outwards) if the density varies (increases or decreases). Notice that if the wave function is real, as is the case of a decaying exponential solution, there is no flux of probability, meaning the probability distribution remains unchanged with time. Note 3.f. In problems where an incident particles scatters, due to the pres- ence of a potential, for instance, one can define the fraction of the wave which is transmitted and that which is reflected. The quantities are given by the so called transmission and reflection coefficients, T = Jt , R = Jr . Ji Ji 3.4 Time dependence of Expected Values It is time now to use our knowledge about the Schrödinger equation to extract information about the evolution of the expectation values associated to a certain observable. Proposition 3.3. Let to be an operator corresponding to a real valued- O observable, then d 1 ∂ = [ ,H]+ O , dtO i~ O ∂t where [A, B]= AB BA. − Proof. Given a certain operator , its expected value on a certain state may O in principle be time-dependent if the reference state is evolving, (t)=(ψ, ψ)= ψ∗(x, t) ψ(x, t) dx (3.4.10) O O O Z Thus we can write d ∂ψ∗ ∂ ∂ψ = dx ψ + dx ψ∗ O ψ + dx ψ∗ + dtO ∂t O ∂t O ∂t Z Z Z 1 ∂ 1 = dx ψ∗ H ψ + dx ψ∗ O ψ + dx ψ∗ Hψ −i~ O ∂t O i~ Z Z Z 1 ∂ = dx ψ∗ ( H H ) ψ + dx ψ∗ O ψ. i~ O − O ∂t Z Z 58 CHAPTER 3. QUANTUM DYNAMICS It can then be cast in the more compact form, d 1 ∂ = [ ,H]+ O . dtO i~ O ∂t Note 3.g. This is the quantum analogue of the the Poisson brackets in clas- sical mechanics, suggesting one to interpret the quantum expectation value as the classical function itself in the classical limit, when ~ 0 and 1 [ , ] → i~ · · → , . {· ·}P B It becomes evident that if the observables do not depend explicitly on time, we have the Corolary 3.4.1. If does not depend explicitly on time O d 1 = [ ,H]. (3.4.11) dtO i~ O So if the observable is not depende explicitly on time the expectation values of the observables are preserved in time if and only if [ ,H]=0. These O objects are considered constants of motion if this happens. An important message to take from this point is that the Hamiltonian determines not only the evolution of the quatum states, as prescribed by Schrödinger’s equation, but it also dictates the time evolution of the expectation values of observables, as can be seen from the relation above. 3.4.1 Newton’s 2nd Law and Quantum Mechanics A remarkable consequence of the general relation between the expectation of any quantum mechanical operator and the expectation of the commutator of that operator with the Hamiltonian of the system, as seen in the Proposi- tion 3.3, is the so called Ehrenfest Theorem, which we now discuss. 3.4. TIME DEPENDENCE OF EXPECTED VALUES 59 The Proposition 3.3 is particularly useful when we evaluate it for simple P 2 Hamiltonian of the form H = 2m + V (X). It can be shown that 1 2 i~ [X,H] = 2m [X,P ]= m P, [P,H] = [P,V (X)] = i~ ∂V , − ∂X and since nor X or P depends explicitly on time, we have dX 1 1 dt = i~ [X,H]= m P, (3.4.12) dP = 1 [P,H]= ∂V . dt i~ − ∂X Note 3.h. The relations expressed in Eq. 3.4.12 are equivalent to Hamilton’s equations of motion (for a conservative potential) [11], dX 1 ∂H dt = m P = ∂P , dP = ∂V = ∂H , dt − ∂X − ∂X A simple way to interpret Eq. 3.4.12 is given by Newton’s second law for mechanics, dX¯ ¯ M dt = P, dP¯ = ∂V , dt − ∂X leading to d2 ∂V M X = = F. dt2 −∂X The content of the last expression is that Newton’s second law, F = ma, is after all, still valid at the quantum level in some sense: it holds true if one is interested only in expectation values. The same results can also be constructed from a slightly different ap- proach, starting from the definition of the expectation value of the position observable, d d x˙ = dt x = dt ψ(x)∗ x ψ(x) dx ∗ ∂ψ(x) x ψ x ψ x x ∂ψ(x) dx. = ∂t R ( )+ ( )∗ ∂t R 60 CHAPTER 3. QUANTUM DYNAMICS From Schrödinger’s wave equation ~ ∂2ψ(x)∗ ∂2ψ(x) x˙ = 2 x ψ(x) ψ(x)∗ x 2 dx 2mi ∂x − ∂x ~ ψ x ∂ ψ x dx p , = mi R ( )∗ ∂x ( ) = m as expected. Similar calculationsR provide us with the corresponding equation for p˙. In the end, Newton’s second law can again be stated: F¯ = mx¨. 3.5 Quantum Pictures We have seen that according to Schrödinger’s description the observ- ables are described by constant operators acting vectors which vary in time according to ΨS(t) = U(t)ΨS(0) = U(t)ΨH whereas in Heisenberg’s frame- work the states are fixed. We can choose Heisenberg states to coincide with Schrödinger’s at the initial time, ΨH =ΨS(0) = U(t)†ΨS(t). Not surprisingly we have that in either picture the time dependent ex- pectation value of an observable in a given state coincide, = (Ψ (t), Ψ (t)) = (Ψ , (t)Ψ ). O S OS S H OH H This requirement imposes restrictions also on the time dependence of the operators and we can relate them via (t) = U(t)† U(t), OH OS so that the operator has a constant form according to Schrödinger, = U(t) (t)U(t)†, OS OH but has a dynamical nature in Heisenberg’s picture. The time evolution given by Schrödinger’s equation has the following counterpart in Heisenberg’s Picture, d (t) i~ OH = [H, (t)], (3.5.13) dt OH dictating the dynamics of Heisenberg’s operators. 3.5. QUANTUM PICTURES 61 There is, however, a further useful formulation, due to Dirac. To explore it one should be able to decompose the Hamiltonian in a time-independent term H0 and an interacting time-dependent part V (t), H = H0 + V (t). We then introduce the simpler evolution operator, i ~ H0t U0(t) = e− , together with a transformed quantum state, ΨI (t)= U0(t)†ΨS(t), or ΨS(t) = U0(t)ΨI (t). This new vector evolves in time according to the interaction potential ∂ i~ Ψ (t)= V Ψ (t), ∂t I I I where we have defined VI (t) = U0(t)†V (t)U0(t). Moreover, also the trans- formed observable, (t)= U (t)† (t)U (t), OI 0 OS 0 evolves in time, but governed by the noninteracting part of the Hamiltonian H0, d (t) i~ OI = [H , (t)]. dt 0 OI The latter formalism is also called the interaction picture for a quantum description, since the states evolution rely essentially on the interaction po- tentical VI . In the following table we summarize how states and observables evolve in time under the different pictures we have presented. 62 CHAPTER 3. QUANTUM DYNAMICS Heisenberg Dirac Schrödinger State Vector Invariant Evolves with VI Evolves with H Observable Operator Evolves with H Evolves with H0 Invariant Table 3.1: Quantum pictures 3.6 Exercises 1. Show that, for a discrete spectra En, the general solution of the time- dependent Schrödinger equation is the linear superposition of the sta- tionary states ∞ i ~ En t ψ(x, t)= anψn(x)e− , n=1 X where ψn(x) are eigenfunctions of H with eigenvalues En. 2. Use the spectral theorem to obtain a continuous version of exercise 1. 3. Show that when an particle is confined to a box, its energy levels are quantized. 4. Show that when an electron is confined to a square well, even though there is a nonvanishing probability of finding it outside the box, there is no flux of probability to the outer region. 5. Show (and interprete) that when a Gaussian wave packet is left to evolve according to Schrödinger’s equation, one has the following rela- tion ~2 ~2t2 (∆x)2(∆p)2 = 1+ . 4 m2σ2 Chapter 4 Approximation Methods So far we have seen some examples of quantum problems which can be completely solved in closed form. Overall we can mention the problems of a free particle, of a particle in a box, of particles scattering on simple one- dimensional barriers, the harmonic oscillator, the problem of central forces, like the Hydrogen atom, just to name a few. In order to overcome the difficulties in “diagonalizing” a problem we present in this chapter some useful techniques to help us investigate real problems. The methods can be divided in two categories: the perturbative and the non-perturbative approaches. 4.1 Non-perturbative methods Here we start the description of problems for which a closed form solution is either very difficult to find and unknown or - more commonly - not possible. In what follows we present some convenient methods which can be applied to various problems, but whose validity is restricted. 4.1.1 Variational Methods A useful method and of simple implementation is the one that permits us to determine the ground state energy of a quantum problem, that is the minimal value of the energy E0, given a specific dynamics, characterized by 63 64 CHAPTER 4. APPROXIMATION METHODS a Hamiltonian H. Proposition 4.1. The energy expectation value of a system is never less than its ground state energy, H E . ≥ 0 Proof. We start by denoting ϕn the eigenstates of the system, Hϕn(x)= En ϕn(x), so that any state can be written as an expansion in this eigenbasis, ϕ(x)= ϕn(x) dxϕn∗ (x)ϕ(x)= ϕn(x)(ϕn,ϕ). n n X Z X This general state can be used to evaluate the expectation value of a certain observable. In the case of the energy operator, we have (ϕ,Hϕ) H = (ϕ,ϕ) which can be rewritten in the following form (E E ) (ϕ ,ϕ) 2 H = E + n=1 n − 0 | n | . 0 (ϕ ,ϕ) 2 P n=0 | n | Since we are considering E0 to be associatedP to the ground state, all other en- ergy levels are higher and all terms in the summation above are non negative [12]. Note 4.a. If one uses the theorem stating that the ground state of a quantum problem is nondegenerate, then all terms in the summation can be show to be positive. You can find a proof in the book of Simon/Reed, "Methods of Mathematical Physics", vol.4 "Analysis of Operators". Therefore, we have a general result that the expectation value for an energy measurement is greater than, or equal, the ground state energy, H ≥ E0. 4.1. NON-PERTURBATIVE METHODS 65 Alternatively, we can expand the generic state as a combination of eigen- states but with unknown coefficients, cn(α), ϕα(x)= cn(α) ϕn(x). n X In this case, we can show that the expression (ϕ ,Hϕ ) H(α) = α α (ϕα,ϕα) reduces to, using c (α) 2 =1, n | n | P H(α)= c (α) 2E . | n | n n X Once again we can argue that its minimum value is indeed E0, occurring when the expectation value is evaluated on the ground state ϕ0(x), H(α) E . ≥ 0 These results seem dull at first sight but they serve to a useful purpose, namely of determming the ground state of a certain system about which we do not know much. Complicated system tend to occur in physical situations more frequently than the idealized settings commonly found in textbooks. In general one is a priori not convinced about the best tools to solve a quantum problem and here we present a fairly universal technique which has also a simple implementation. The knowledge of a single energy eigenvalue in the whole spectrum might sound insignificant in principle. However, in many physical applications, es- pecially when one is concerned about statitic properties of quantum systems, the ground state energy contains sufficient information to allow a character- ization of the system. As we minimize the expectation value of a Hamiltonian operator whose spectrum is bounded from bellow, we are capable of finding the value of the 66 CHAPTER 4. APPROXIMATION METHODS newly introduced parameter α = α0 which satisfies the following minimum condition, ∂H(α) =0. ∂α α0 We can then estimate the ground state energy of a system as E0 . H(α0). Although very general, this approach presents the disadvantadge that it relies on one’s ability to construct the most effective form for the coefficients cn(α), or ultimately for ϕα(x) itself, as some choices will lead to better results than others. In order to optimize the method it is advisable to make use of symmetriy properties and physical reasoning to start with a convenient ansatz for these objects. In some cases it is possible to find the exact ground state ϕ0(x)= ϕα0 (x). Next we investigate an alternative non-perturbative method which is more compatible with the other end of the spectrum, that is it more suited for high eigenvalues, being a complementaray approach. 4.1.2 Extension to Excited States Proposition 4.2. If you have an enumerable spectrum which can be ordered, E < E E then the upper bound on the n-th energy level is given 0 1 ≤ 2 ≤··· by E ψ,Hψ n ≤ if the state ψ used is orthogonal to all previous n 1 states. − The proof is left as an exercise. 4.1. NON-PERTURBATIVE METHODS 67 4.1.3 A 2 2 example × 1 √2 H = √2 0 1 1 φ(α) = | i √1+ α2 α 1+2√2α E(α)= φ(α) H φ(α) = h | | i 1+ α2 dE =0 dα α0 α = √2 0 − 1 1 φ(α0) = | i √3 √2 − E E(α )= 1 0 ≤ 0 − Suppose we know the ground state of the system 1 1 1 = | i √3 √2 − The orthogonal state is 1 √2 φ = | i √ 1 3 E φ H φ =2 1 ≤h | | i 68 CHAPTER 4. APPROXIMATION METHODS 4.1.4 Method os Successive Powers Proposition 4.3. Assuming the spectrum is bounded, E E E E , 0 ≤ 1 ≤ 2 ≤···≤ M the largest eigenvalue reads φ Hm+1 ψ EM = lim h | | i, m φ Hm ψ →∞ h | | i where the state φ may be any, as long as it is orthogonal to all eigenfucntions of H, φ, ψ =0. n 6 Note 4.b. If this projection φ, ψn vanishes we have to take another φ state. Proof. Projecting on a test function φ the m-th action of the Hamiltonian operator on a (unknown) combination of its eigenfucntions m m φ,H ψ = cnEn φ, ψn . n X Thus, in the limit of large m the largest eigenvalue E contributions → ∞ M will dominate and m+1 m+1 φ,H ψ cM EM φ, ψM lim = m m m →∞ φ,H ψ cM EM φ, ψM is a well defined ratio if φ, ψ =0, as assumed, and provides the estimative M 6 for EM above. 4.1.5 WKB - Semiclassical approximation The method developed by Wentzel, Kramers and Brillouin - from which the acronym stems - considers the Schrödinger equation, ~2 ψ′′(x)+(V (x) E) ψ(x)=0, −2m − in a semiclassical approach. 4.1. NON-PERTURBATIVE METHODS 69 Proposition 4.4. A good approximation for the eigenfunctions of a Schrödinger problem in the semicassical limit is given by plane waves of the form i ~ R √2m(E V (x))dx e± − ψ(x) 1 . ∼ 4 2m(E V (x)) − Proof. For that it considers a representation similar to the one found in Hamilton-Jacobi’s formulation of classical mechanics, reexpressing the wave- function in terms of Hamilton’s principal (action) funcion, i ψ(x)= e ~ S(x), so the equation of motion becomes 2 2 i~ S′′(x)+ S′(x) =2m(E V (x)) = p(x) , − − where we have introduced a position-dependent momentum function p(x) as an analogy with classical mechanics. Here we are more interested in the excited states so that the quantum systems behaves more similarly to classical mechanics. In this so called semiclassical limit the wave properties are less important and -similarly to what is done in the geometric limit of physical optics- we can restrain to the optical approximation, 2 ~ S′′(x) << S′(x) . Note 4.c. Also in the semiclassical limit, ~ 0. → This allows us to express the principal function in terms of the external potential, or the equivalent classical momentum and corresponding wave- length, 2π~ S(x) dx 2m(E V (x)) = dxp(x)= dx . ≃ ± − ± ± λ(x) Z Z Z One is then able top show that a solution for the wave function is a combination of incoming and outgoing plane waves, c1 1 R p(x)dx c2 + 1 R p(x)dx ψ(x)= e− i~ + e i~ , p(x) p(x) whose form is valid forp a general potential.p 70 CHAPTER 4. APPROXIMATION METHODS It follows from this result a quantization condition which was used in the early days of the quantum theory, Corolary 4.1.1. The action associated to a quantum problem, over a period, must be quantized in multiples of π, 1 p(x)dx = n + π. 2 I Proof. Care must be taken, though, if the potential allows for the existence of turning points, which occur when the energy E given makes the parti- cle confined in the potential V (x). The turning points xTP appear when V (xTP ) = E, so that p(xTP )=0 and the movement stops. Denoting the turning points on the left and on the right by a and b respectively, it is left as an exercise to show that near these points one has [13] c 1 x π ψ(x) cos ~ p(x)dx , a The simple analysis thatp in the region between the turning points the solu- tion must be well behaved permits us to recover to famous Bohr-Sommerfeld quantization condition, 1 x π π 1 b p(x)dx p(x)dx = mπ, ~ − 4 − 4 − ~ Za Zx or more simply, b 1 p(x)dx = m + π. 2 Za It is tricky to state which method is best for a particular problem. Each method has its advantadges and its drawbacks. So far we have presented for- mulations which are more useful in opposite regimes of low and high energy 4.2. TIME-INDEPENDENT PERTURBATION THEORY 71 levels. In the next section we will show a very geneal way of solving some Schrödinger problems with arbitrary precision, a method based on perturba- tion theory. 4.2 Time-independent Perturbation Theory We know return to the fundamental time-independent Schrödinger equa- tion. In the event of us having a simple Hamiltonian operator H0 for which 0 0 we know the eigenvalues En and the associated eigenfunctions ϕn(x), we can write 0 0 0 H0 ϕn(x)= En ϕn(x), as it happens for a variety of problems: particle in a box, harmonic oscillator, hydrogen atom, to name a few. If, on the other hand, we have a Hamiltonian operator we do not how to diagonalize exactly we might still be able to solve the problem if we are able to recognize the challenging operator as a small deviation from a known one. That means we can express the new Hamiltonian as the old one with the simple introduction of a perturbation, H = H0 + λH1. Then we ask ourselves what the eigenstates ϕn(x) and eigenvalues En for this deformed operator are, that is we would like to write down the following eiganvalue equation, Hϕn(x)= En ϕn(x). Before proceeding, we must reminf ourselves that if the small perturbation hypothesis is true, in the end we should obtain eigenvalues En which are not 0 too different from the reference ones En, or equivalently, 0 λ (ϕn,H1 ϕn) << En. 72 CHAPTER 4. APPROXIMATION METHODS Because the eigenstates of H form a complete basis ϕ0 (x) , we can use 0 { n } those states to expand the eigenvectors of H, 0 ϕk(x)= ckm ϕm(x), m=n X6 with the normalization (ϕm,ϕn) = δmn. In a similar way as the Hamiltonian operator is a deviation from an undeformed one, we can expand the deformed wavefunction as a deviation from the original wavefunction by including corrections, order by order on λ, ϕ (x)= ϕ0 (x)+ λϕ1 (x)+ λ2ϕ2 (x)+ . n n n n ··· The normalization condition of the new eigenfunctions leads to the orthonor- mality of the correction terms with respect to the original eigenstate, 0 j (ϕn,ϕn)= δj0, and 0 k 0 0 (ϕn,ϕn)= ckm (ϕn,ϕm)=0. m=n X6 At this point, we need to establish a distinction between systems which contain and systems which do not contain degenerate states. We begin with the simplest scenario, of no eigenstates sharing the same eigenvalue. 4.2.1 Time-independent perturbation: Non-degenerate Proposition 4.5. The first order correction on the energy levels of a per- turbed Hamiltonian is given by 1 0 0 En =(ϕn,H1 ϕn). 4.2. TIME-INDEPENDENT PERTURBATION THEORY 73 Proof. By starting with the perturbed Hamiltonian, H = H0 + λH1 we can make use of the perturbation expansion of the eigenfunctions and eigenvalues to obtain (H + λH ) ϕ0 (x)+ λϕ1 (x)+ = E0 + λE1 + ϕ0 (x)+ λϕ1 (x)+ . 0 1 n n ··· n n ··· n n ··· On both sides of the equation one has now a power series on the perturbation parameter λ and in order for the equation to be valid for any value of λ we must equate each coefficient of the expansion, giving rise a set of simpler equations, 0 0 0 H0 ϕn(x) = En ϕn(x), 1 0 0 1 1 0 H0 ϕn(x)+ H1 ϕn(x) = En ϕn(x)+ En ϕn(x), 2 1 0 2 1 1 2 0 H0 ϕn(x)+ H1 ϕn(x) = En ϕn(x)+ En ϕn(x), +En ϕn(x), . . The terms independent of λ, associated to the coefficients of power λ0, are the first to appear in the expansion. They correspond simply to the reference eigenproblem, for which we know the spectrum, 0 0 0 H0 ϕn(x) = En ϕn(x). In the sequence we have the first order correction, corresponding to the coefficients of the power λ1, 1 0 0 1 1 0 H0 ϕn(x)+ H1 ϕn(x) = En ϕn(x)+ En ϕn(x). Notice that in the above equation we have two unknowns, the first pertur- 1 bation on eigenvalues, En, and the first perturbation on the eigenfunctions, 74 CHAPTER 4. APPROXIMATION METHODS 1 ϕn(x). As we have argued, the latter can be expanded in the basis of the eigenstates of H0, 1 0 ϕn(x)= c1m ϕm(x). m=n X6 Inserting them in the equation above, we are left with E0 E0 c ϕ0 (x)+ H ϕ0 (x) = E1 ϕ0 (x). (4.2.1) m − n 1m m 1 n n n m=n X6 1 In order to compute the first eigenvalue correction En one needs to take 0 the scalar product of this equation with the unperturbed eigenfunction ϕn(x). This produces 1 0 0 En =(ϕn,H1 ϕn), so it is enough to compute the expectation value of the perturbation operator on the original ground state. On the other hand, if we want to determine the first order correction to the eigenvectors we have the following proposition Proposition 4.6. The first nontrivial term in the eigenfucntion perturbation expansion is (ϕ0 ,H ϕ0 ) ϕ1 (x)= ϕ0 (x) m 1 n . n − m E0 E0 m=n m n X6 − 0 Proof. Projecting the same equation (4.2.1) on a state ϕk(x), allows us to write (ϕ0 ,H ϕ0 ) c = m 1 n , 1m − E0 E0 m − n so the expansion in the original basis becomes is the one presented above. Following a similar path we can compute the second order corrections as well, based on the equation 2 1 0 2 1 1 2 0 H0 ϕn(x)+ H1 ϕn(x)= En ϕn(x)+ En ϕn(x), +En ϕn(x) (4.2.2) 4.2. TIME-INDEPENDENT PERTURBATION THEORY 75 1 1 where En and ϕn(x) have been proviously determined, leaving us to compute 2 2 the unknowns En and ϕn(x). Proposition 4.7. The second order corrections to the eigenvalues are always negative. Besides, close levels contribute more and tend repel each other. Proof. Making use of (4.2.2) we write down n 0 2 0 1 2 E0 (ϕn,ϕn)+(ϕn,H1 ϕn)= En, leading to (ϕ0 ,H ϕ0 ) 2 E2 = | m 1 n | , n − E0 E0 m=n m n X6 − indicating the validity of the proposition. For states in the continuum the sums must be replaced by integrals. 4.2.2 Time-independent perturbation: Degenerate For degenerate states, a more careful analysis is required. Suppose we an unperturbed system with degenerate eigenstates, H ϕ0 (x)= E0 ϕ0 (x), i =1, ,g 0 n,i n n,i ··· where g denotes de degeneracy associated to each eigenvalue, that is it rep- resents the number of eigenfunctions sharing a same eigenvalue [14]. As before, we can still propose a perturbation expansion for the charac- teristic energies, E = E0 + λE1 + λ2E2 + , n n n n ··· and for the characteristic functions, ϕ (x)= ϕ0 (x)+ λϕ1 (x)+ λ2 ϕ2 (x)+ , n n n n ··· 76 CHAPTER 4. APPROXIMATION METHODS so that at zero-th order we have the familiar relation 0 0 0 H0 ϕn(x)= En ϕn(x). Considering now the undeformed states to be degenerate one has, more generally, g g 0 0 0 H0 ai ϕn,i(x)= En aiϕn,i(x), i=1 i=1 X X which is automatically satisfied. Proposition 4.8. In the presence of degeneracies the first order correction 1 to the energy eigenvalues En are determined by the following equation g 0 0 1 (ϕn,j,H1ϕn,i) ai = En aj. i=1 X Proof. In order to determine the first corrections we consider 1 0 0 0 1 0 H0 ϕn,i(x)+ H1 ϕn,i(x) = En ϕn,i(x)+ En ϕn,i(x), which leads to g g 0 0 0 0 1 0 H0 c1mϕm(x)+ H1 aiϕn,i(x) = En c1mϕm(x)+ En aiϕn,i(x), m=n i=1 m=n i=1 X6 X X6 X or g g c E0 E0 ϕ0 (x)+ a H ϕ0 (x) = a E1 ϕ0 (x). 1m m − n n i 1 n i n n,i m=n i=1 i=1 X6 X X 0 Projecting this equation on a function ϕn,j(x), one can show that the propo- sition is true. With the introduction of the abbreviation V (ϕ0 ,H ϕ0 ) ij ≡ n,j 1 n,i 4.3. THE ANHARMONIC OSCILLATOR 77 the equation which determines the first eigenvalue corrections can be written in a more convenient form, V11 V12 a1 a1 ··· 1 V21 V22 a2 = E a2 . . . ···. . n . ...... Therefore, the message we get from these computations is that one must choose a basis in the degenerate space for which the perturbation is diagonal. We turn our attention now to an explicit simple example, of system pre- senting a double degeneracy. The associated degenerate eigenspace is such that we can write V V a a 11 12 1 = E1 1 V V a n a 21 22 2 2 1 so if we wish to determine the value of the energy correction En, we must diagonalize the interaction operator, V E1 V 11 n 12 =0 V− V E1 21 22 − n which gives rise to the characteristic equation 2 E1 (V + V ) E1 + V V V V =0. n − 11 22 n 11 22 − 12 21 This can be put in a favorable form 2 E1 Tr V E1 + det V =0, n − n from which we see that the first energy correction represents a level splitting of the spectrum, 1 E1 = Tr V √Tr V 4 det V. n 2 ± − 4.3 The Anharmonic Oscillator As an application of the perturbation theory, consider now a small devi- ation of the harmonic oscillator, with a quartic interaction, ~2 d2 mω2 H = + x2 + αx4, −2m dx2 2 78 CHAPTER 4. APPROXIMATION METHODS for which there is not a closed form set of eigenstates and eigenvalues. Our first attempt, based on what we have studied so far, is to assume the Hamil- tonian of our interest to be decomposed as H = H0 + αH1 where ~2 d2 mω2 H = + x2, 0 −2m dx2 2 and 4 H1 = x , with a perturbation parameter α, considered to be relatively small. Since this operator is not explicitly time-dependent we can apply the time-independent perturbation formalism, with the unperturbed eigenvalues 1 E0 = n + ~ω, n 2 and unperturbed eigenfunctions written in terms of Hermite polynomials Hen(x), 1 4 2 0 π x ϕ (x)= e− 2 Hen(x). n √2nn! The first correction to the characteristic energies is given by, in units of 1 α− , 3h2 E1 =(ϕ0 , x4 ϕ0 )= 2n(n +1)+1 n n n 4m2ω2 and for the eigenfunctions we have to compute the more laborious quantity, [15] (ϕ0 , x4 ϕ0 ) (ϕ0 , x4 ϕ0 ) ϕ1 (x)= ϕ0 (x) m n = ϕ0 (x) m n . n − m E0 E0 − m ~ω(m n) m=n m n m=n X6 − X6 − 4.4. TIME-DEPENDENT PERTURBATION 79 4.4 Time-dependent Perturbation So far we have treated perturbatively the time-independent Schrödinger eigenvalue equation. It is time to employ the same principle of expanding the states as a perturbative expansion to address the more general problem of the time-dependent Schrödinger equation, which must be considered if, for example, the external potential varies with time [14]. For that matter, our starting point is, not surprisingly, the fundamental equation, ∂ i~ ψ(x, t)= H(x, t)ψ(x, t). ∂t Proposition 4.9. The general solution of Schrödinger’s equation can be ex- pressed as a combination of eigenfunctions of the form ψ(x, t)= cn φn(x) χn(t). n X Proof. We shall assume that the Hamiltonian has such a structure that allows us to separate the operator in the following fashion H(x, t)= H0(x)+ H1(t), an additive separability condition of the Hamiltonian. If that is the case we have a separability of wavefunction, but in the product form, ψ(x, t)= φ(x) χ(t). Substituting it in the time-dependent Schrödinger equation we can write χ (t) 1 1 i~ ′ H (t)χ(t)= H (x)φ(x)= E, χ(t) − χ(t) 1 φ(x) 0 where E is a constant to guarantee that both the χ(t) and φ(x) vary inde- pendently. This procedure generates two separate equations, H0(x) φ(x)= Eφ(x), 80 CHAPTER 4. APPROXIMATION METHODS and d i~ χ(t)= H (t)+ E χ(t). dt 1 Here φ(x) is, in principle known and χ(t) can be solved similarly to the stationary Schrödinger equation and the general solution can be expressed as in the proposition. In what follows we present how the solution can be constructed in the spirit of perturbation theory. 4.4.1 Time-dependent Perturbation Theory Considering a Hamiltonian whose time dependence may be, at least par- tially, separated in the following way H(x, t)= H0(x)+ V (x, t), we can make use of the perturbation approach if at every instant we have that the deformation from the original Hamiltinonian is indeed small, V (x, t) << H0(x). Proposition 4.10. The first time-dependent correction coefficients must sat- isfy the following equation 1 c˙1 (t)= c0 (t) eiωmnt V (t). m i~ n mn n X Proof. In the absence of perturbation we know that the Schrödinger equation 0 0 0 H0(x, t) ψn(x, t)= En ψn(x, t) has the general solution i 0 0 0 ~ Ent ψn(x, t)= φn(x)e− . 4.4. TIME-DEPENDENT PERTURBATION 81 We can use this complete set of eigenfunctions to expand the perturbed state i 0 0 0 ~ Ent ψn(x, t)= cn(t)ψn(x, t), = cn(t)φn(x)e− , n n X X and replacing it in ∂ i ψ(x, t)= H(x, t) ψ(x, t) ∂t gives ∂ 0 i E0 t 0 i E0 t i~ c (t)φ (x)e− ~ n = H (x)+ V (x, t) c (t)φ (x)e− ~ n . ∂t n n 0 n n n ! n ! X X The above expression simplifies to i i ~ ~ Emt ~ Ent 0 0 i c˙m(t) e− = cn(t) e− (ϕm,V (x, t) ϕn) n X and can be written in the compact form, 1 c˙ (t)= c (t) eiωmnt V (t), m i~ n mn n X where the first correction coefficient depends crucially on the unperturbed 0 eigenstates through cn(t), culminating with the proposition. Bellow we determine the explicit form of the time dependence of the coefficients cn(t). We can choose the initial state to obey 0 cn(t0)= cn = δns so we have ′ 1 1 iωmst c (t)= dt′ V (t′)e . m i~ ms Z This procedure can be repeated to give rise to higher and higher contributions forming the corrected eigenfunction coefficients c (t) = c0 (t)+ c1 (t)+ m m m ··· ′ 1 iωmst = δ + dt′ V (t′) e + . ms i~ ms ··· Z 82 CHAPTER 4. APPROXIMATION METHODS Proposition 4.11. The second order correction reads 2 ′ ′′ 2 1 iωnst iωmnt c (t) = dt′′ dt′ V (t′) V (t′′) e e . m i~ ns mn n X Z Z Proof. Using 1 c˙2 (t) = c1 (t) eiωmnt V (t) m i~ n mn n X with the information we have already gathered produces the equation 2 ′ 2 1 iωnst iωmnt c˙ (t) = dt′ V (t′) V (t) e e m i~ ns mn n X Z so the second correction solution behaves as proposed. We can keep going producing higher terms and constructing more and more precise eigenfunctions. The simplest example to consider as a time-dependent two-level system is a problem of the familiar form H = H0 + V (t), for which the equation of motion reads ∂ c (t) V V e iωt c (t) i~ 1 = 11 12 − 1 . ∂t c (t) V eiωt V c (t) 2 21 22 2 +i ωt i ωt Multiplying both sides by a phase, e 2 and e− 2 , respectively, with the introduction of new variables, +i ωt i ωt ϕ1(t)= c1(t) e 2 and ϕ2(t)= c2(t) e− 2 , furnishes us with ~ω ∂ ϕ1(t) V11 + V12 ϕ1(t) i~ = 2 ~ . ∂t ϕ (t) V V ω ϕ (t) 2 21 22 − 2 2 4.5. DIRAC’S INTERACTION PICTURE 83 With this transformation we were able to circumvent the problem of hav- ing a time-dependent Hamiltonian since we have eliminated the time de- pendence. By diagonalizing the problem we can express it in the following form ∂ ψ (t) E 0 ψ (t) i~ 1 = 1 1 , ∂t ψ (t) 0 E ψ (t) 2 2 2 where the eigenvalues are given by V + V 1 E = 11 22 (V V + ~ω)2 +4V 2 . 1,2 2 ± 2 11 − 22 12 q 4.5 Dirac’s interaction picture Previously we have discussed about Dirac’s interaction picture as opposed to Heisenberg’s and Schrödinger’s formulations. Here we go back to this point and show how it can be a useful tool to tackle time-dependent problems of the form H = H0 + V (t). We define, as we have already seen, the interaction state, i ~ H0t ψI (t) = e ψ(t), and the interaction observable, i i H0t H0t = e ~ e− ~ , OI O so that initially both states coincide, ψI (0) = ψ(0). If we also introduce the interaction potential, i i ~ H0t ~ H0t VI = e V e− , 84 CHAPTER 4. APPROXIMATION METHODS we can use to write the fundamental Schrödinger equation according to ∂ ∂ i i~ ψ (t)= i~ e ~ H0tψ(t) = V ψ (t). ∂t I ∂t I I Therefore, the time evolution of the interaction state is solely governed by the interaction potential. Projecting it on a given eigenstate of the original problem ∂ i~ (ψ , ψ ) = (ψ ,V ψ )= (ψ ,V ψ )(ψ , ψ ) ∂t n I n I I n I m m I m X gives rise to ∂ i~ c (t)= eiωnmtV c (t). ∂t n nm m m X This result, not surprisingly, coincides with the expression we have just ob- tained with a different approach. In the sequence we explore some consequences of the formalism con- structed applied to special cases. 4.6 Furhter Applications and Fermi’s Golden Rule Step Function / Constant Another problem which can be easily addressed with this framework is that of a potential which has a simple time dependence: initially the potential is not present but at some point in time it is turned on, by switching a button, for example, and it remains constant until it is switched off. Mathematically a potential of this type can be represented by a Heaviside step function, 1, t