Math 215/255 Lecture Notes 2019
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Ordinary Differential Equation 2
Unit - II Linear Systems- Let us consider a system of first order differential equations of the form dx = F(,)x y dt (1) dy = G(,)x y dt Where t is an independent variable. And x & y are dependent variables. The system (1) is called a linear system if both F(x, y) and G(x, y) are linear in x and y. dx = a ()t x + b ()t y + f() t dt 1 1 1 Also system (1) can be written as (2) dy = a ()t x + b ()t y + f() t dt 2 2 2 ∀ = [ ] Where ai t)( , bi t)( and fi () t i 2,1 are continuous functions on a, b . Homogeneous and Non-Homogeneous Linear Systems- The system (2) is called a homogeneous linear system, if both f1 ( t )and f2 ( t ) are identically zero and if both f1 ( t )and f2 t)( are not equal to zero, then the system (2) is called a non-homogeneous linear system. x = x() t Solution- A pair of functions defined on [a, b]is said to be a solution of (2) if it satisfies y = y() t (2). dx = 4x − y ........ A dt Example- (3) dy = 2x + y ........ B dt dx d 2 x dx From A, y = 4x − putting in B we obtain − 5 + 6x = 0is a 2 nd order differential dt dt 2 dt x = e2t equation. The auxiliary equation is m2 − 5m + 6 = 0 ⇒ m = 3,2 so putting x = e2t in A, we x = e3t obtain y = 2e2t again putting x = e3t in A, we obtain y = e3t . -
MATH 2370, Practice Problems
MATH 2370, Practice Problems Kiumars Kaveh Problem: Prove that an n × n complex matrix A is diagonalizable if and only if there is a basis consisting of eigenvectors of A. Problem: Let A : V ! W be a one-to-one linear map between two finite dimensional vector spaces V and W . Show that the dual map A0 : W 0 ! V 0 is surjective. Problem: Determine if the curve 2 2 2 f(x; y) 2 R j x + y + xy = 10g is an ellipse or hyperbola or union of two lines. Problem: Show that if a nilpotent matrix is diagonalizable then it is the zero matrix. Problem: Let P be a permutation matrix. Show that P is diagonalizable. Show that if λ is an eigenvalue of P then for some integer m > 0 we have λm = 1 (i.e. λ is an m-th root of unity). Hint: Note that P m = I for some integer m > 0. Problem: Show that if λ is an eigenvector of an orthogonal matrix A then jλj = 1. n Problem: Take a vector v 2 R and let H be the hyperplane orthogonal n n to v. Let R : R ! R be the reflection with respect to a hyperplane H. Prove that R is a diagonalizable linear map. Problem: Prove that if λ1; λ2 are distinct eigenvalues of a complex matrix A then the intersection of the generalized eigenspaces Eλ1 and Eλ2 is zero (this is part of the Spectral Theorem). 1 Problem: Let H = (hij) be a 2 × 2 Hermitian matrix. Use the Min- imax Principle to show that if λ1 ≤ λ2 are the eigenvalues of H then λ1 ≤ h11 ≤ λ2. -
Phase Plane Diagrams of Difference Equations
PHASE PLANE DIAGRAMS OF DIFFERENCE EQUATIONS TANYA DEWLAND, JEROME WESTON, AND RACHEL WEYRENS Abstract. We will be determining qualitative features of a dis- crete dynamical system of homogeneous difference equations with constant coefficients. By creating phase plane diagrams of our system we can visualize these features, such as convergence, equi- librium points, and stability. 1. Introduction Continuous systems are often approximated as discrete processes, mean- ing that we look only at the solutions for positive integer inputs. Dif- ference equations are recurrence relations, and first order difference equations only depend on the previous value. Using difference equa- tions, we can model discrete dynamical systems. The observations we can determine, by analyzing phase plane diagrams of difference equa- tions, are if we are modeling decay or growth, convergence, stability, and equilibrium points. For this paper, we will only consider two dimensional systems. Suppose x(k) and y(k) are two functions of the positive integers that satisfy the following system of difference equations: x(k + 1) = ax(k) + by(k) y(k + 1) = cx(k) + dy(k): This system is more simply written in matrix from z(k + 1) = Az(k); x(k) a b where z(k) = is a column vector and A = . If z(0) = z y(k) c d 0 is the initial condition then the solution to the initial value prob- lem z(k + 1) = Az(k); z(0) = z0; 1 2 TANYA DEWLAND, JEROME WESTON, AND RACHEL WEYRENS is k z(k) = A z0: Each z(k) is represented as a point (x(k); y(k)) in the Euclidean plane. -
Perturbation Theory and Exact Solutions
PERTURBATION THEORY AND EXACT SOLUTIONS by J J. LODDER R|nhtdnn Report 76~96 DISSIPATIVE MOTION PERTURBATION THEORY AND EXACT SOLUTIONS J J. LODOER ASSOCIATIE EURATOM-FOM Jun»»76 FOM-INST1TUUT VOOR PLASMAFYSICA RUNHUIZEN - JUTPHAAS - NEDERLAND DISSIPATIVE MOTION PERTURBATION THEORY AND EXACT SOLUTIONS by JJ LODDER R^nhuizen Report 76-95 Thisworkwat performed at part of th«r«Mvchprogmmncof thcHMCiattofiafrccmentof EnratoniOTd th« Stichting voor FundtmenteelOiutereoek der Matctk" (FOM) wtihnnmcWMppoft from the Nederhmdie Organiutic voor Zuiver Wetemchap- pcigk Onderzoek (ZWO) and Evntom It it abo pabHtfMd w a the* of Ac Univenrty of Utrecht CONTENTS page SUMMARY iii I. INTRODUCTION 1 II. GENERALIZED FUNCTIONS DEFINED ON DISCONTINUOUS TEST FUNC TIONS AND THEIR FOURIER, LAPLACE, AND HILBERT TRANSFORMS 1. Introduction 4 2. Discontinuous test functions 5 3. Differentiation 7 4. Powers of x. The partie finie 10 5. Fourier transforms 16 6. Laplace transforms 20 7. Hubert transforms 20 8. Dispersion relations 21 III. PERTURBATION THEORY 1. Introduction 24 2. Arbitrary potential, momentum coupling 24 3. Dissipative equation of motion 31 4. Expectation values 32 5. Matrix elements, transition probabilities 33 6. Harmonic oscillator 36 7. Classical mechanics and quantum corrections 36 8. Discussion of the Pu strength function 38 IV. EXACTLY SOLVABLE MODELS FOR DISSIPATIVE MOTION 1. Introduction 40 2. General quadratic Kami1tonians 41 3. Differential equations 46 4. Classical mechanics and quantum corrections 49 5. Equation of motion for observables 51 V. SPECIAL QUADRATIC HAMILTONIANS 1. Introduction 53 2. Hamiltcnians with coordinate coupling 53 3. Double coordinate coupled Hamiltonians 62 4. Symmetric Hamiltonians 63 i page VI. DISCUSSION 1. Introduction 66 ?. -
Chapter 8 Stability Theory
Chapter 8 Stability theory We discuss properties of solutions of a first order two dimensional system, and stability theory for a special class of linear systems. We denote the independent variable by ‘t’ in place of ‘x’, and x,y denote dependent variables. Let I ⊆ R be an interval, and Ω ⊆ R2 be a domain. Let us consider the system dx = F (t, x, y), dt (8.1) dy = G(t, x, y), dt where the functions are defined on I × Ω, and are locally Lipschitz w.r.t. variable (x, y) ∈ Ω. Definition 8.1 (Autonomous system) A system of ODE having the form (8.1) is called an autonomous system if the functions F (t, x, y) and G(t, x, y) are constant w.r.t. variable t. That is, dx = F (x, y), dt (8.2) dy = G(x, y), dt Definition 8.2 A point (x0, y0) ∈ Ω is said to be a critical point of the autonomous system (8.2) if F (x0, y0) = G(x0, y0) = 0. (8.3) A critical point is also called an equilibrium point, a rest point. Definition 8.3 Let (x(t), y(t)) be a solution of a two-dimensional (planar) autonomous system (8.2). The trace of (x(t), y(t)) as t varies is a curve in the plane. This curve is called trajectory. Remark 8.4 (On solutions of autonomous systems) (i) Two different solutions may represent the same trajectory. For, (1) If (x1(t), y1(t)) defined on an interval J is a solution of the autonomous system (8.2), then the pair of functions (x2(t), y2(t)) defined by (x2(t), y2(t)) := (x1(t − s), y1(t − s)), for t ∈ s + J (8.4) is a solution on interval s + J, for every arbitrary but fixed s ∈ R. -
Lecture 2: Spectral Theorems
Lecture 2: Spectral Theorems This lecture introduces normal matrices. The spectral theorem will inform us that normal matrices are exactly the unitarily diagonalizable matrices. As a consequence, we will deduce the classical spectral theorem for Hermitian matrices. The case of commuting families of matrices will also be studied. All of this corresponds to section 2.5 of the textbook. 1 Normal matrices Definition 1. A matrix A 2 Mn is called a normal matrix if AA∗ = A∗A: Observation: The set of normal matrices includes all the Hermitian matrices (A∗ = A), the skew-Hermitian matrices (A∗ = −A), and the unitary matrices (AA∗ = A∗A = I). It also " # " # 1 −1 1 1 contains other matrices, e.g. , but not all matrices, e.g. 1 1 0 1 Here is an alternate characterization of normal matrices. Theorem 2. A matrix A 2 Mn is normal iff ∗ n kAxk2 = kA xk2 for all x 2 C : n Proof. If A is normal, then for any x 2 C , 2 ∗ ∗ ∗ ∗ ∗ 2 kAxk2 = hAx; Axi = hx; A Axi = hx; AA xi = hA x; A xi = kA xk2: ∗ n n Conversely, suppose that kAxk = kA xk for all x 2 C . For any x; y 2 C and for λ 2 C with jλj = 1 chosen so that <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij, we expand both sides of 2 ∗ 2 kA(λx + y)k2 = kA (λx + y)k2 to obtain 2 2 ∗ 2 ∗ 2 ∗ ∗ kAxk2 + kAyk2 + 2<(λhAx; Ayi) = kA xk2 + kA yk2 + 2<(λhA x; A yi): 2 ∗ 2 2 ∗ 2 Using the facts that kAxk2 = kA xk2 and kAyk2 = kA yk2, we derive 0 = <(λhAx; Ayi − λhA∗x; A∗yi) = <(λhx; A∗Ayi − λhx; AA∗yi) = <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij: n ∗ ∗ n Since this is true for any x 2 C , we deduce (A A − AA )y = 0, which holds for any y 2 C , meaning that A∗A − AA∗ = 0, as desired. -
ATTRACTORS: STRANGE and OTHERWISE Attractor - in Mathematics, an Attractor Is a Region of Phase Space That "Attracts" All Nearby Points As Time Passes
ATTRACTORS: STRANGE AND OTHERWISE Attractor - In mathematics, an attractor is a region of phase space that "attracts" all nearby points as time passes. That is, the changing values have a trajectory which moves across the phase space toward the attractor, like a ball rolling down a hilly landscape toward the valley it is attracted to. PHASE SPACE - imagine a standard graph with an x-y axis; it is a phase space. We plot the position of an x-y variable on the graph as a point. That single point summarizes all the information about x and y. If the values of x and/or y change systematically a series of points will plot as a curve or trajectory moving across the phase space. Phase space turns numbers into pictures. There are as many phase space dimensions as there are variables. The strange attractor below has 3 dimensions. LIMIT CYCLE (OR PERIODIC) ATTRACTOR STRANGE (OR COMPLEX) ATTRACTOR A system which repeats itself exactly, continuously, like A strange (or chaotic) attractor is one in which the - + a clock pendulum (left) Position (right) trajectory of the points circle around a region of phase space, but never exactly repeat their path. That is, they do have a predictable overall form, but the form is made up of unpredictable details. Velocity More important, the trajectory of nearby points diverge 0 rapidly reflecting sensitive dependence. Many different strange attractors exist, including the Lorenz, Julian, and Henon, each generated by a Velocity Y different equation. 0 Z The attractors + (right) exhibit fractal X geometry. Velocity 0 ATTRACTORS IN GENERAL We can generalize an attractor as any state toward which a Velocity system naturally evolves. -
Polarization Fields and Phase Space Densities in Storage Rings: Stroboscopic Averaging and the Ergodic Theorem
Physica D 234 (2007) 131–149 www.elsevier.com/locate/physd Polarization fields and phase space densities in storage rings: Stroboscopic averaging and the ergodic theorem✩ James A. Ellison∗, Klaus Heinemann Department of Mathematics and Statistics, University of New Mexico, Albuquerque, NM 87131, United States Received 1 May 2007; received in revised form 6 July 2007; accepted 9 July 2007 Available online 14 July 2007 Communicated by C.K.R.T. Jones Abstract A class of orbital motions with volume preserving flows and with vector fields periodic in the “time” parameter θ is defined. Spin motion coupled to the orbital dynamics is then defined, resulting in a class of spin–orbit motions which are important for storage rings. Phase space densities and polarization fields are introduced. It is important, in the context of storage rings, to understand the behavior of periodic polarization fields and phase space densities. Due to the 2π time periodicity of the spin–orbit equations of motion the polarization field, taken at a sequence of increasing time values θ,θ 2π,θ 4π,... , gives a sequence of polarization fields, called the stroboscopic sequence. We show, by using the + + Birkhoff ergodic theorem, that under very general conditions the Cesaro` averages of that sequence converge almost everywhere on phase space to a polarization field which is 2π-periodic in time. This fulfills the main aim of this paper in that it demonstrates that the tracking algorithm for stroboscopic averaging, encoded in the program SPRINT and used in the study of spin motion in storage rings, is mathematically well-founded. -
Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation -
Phase Plane Methods
Chapter 10 Phase Plane Methods Contents 10.1 Planar Autonomous Systems . 680 10.2 Planar Constant Linear Systems . 694 10.3 Planar Almost Linear Systems . 705 10.4 Biological Models . 715 10.5 Mechanical Models . 730 Studied here are planar autonomous systems of differential equations. The topics: Planar Autonomous Systems: Phase Portraits, Stability. Planar Constant Linear Systems: Classification of isolated equilib- ria, Phase portraits. Planar Almost Linear Systems: Phase portraits, Nonlinear classi- fications of equilibria. Biological Models: Predator-prey models, Competition models, Survival of one species, Co-existence, Alligators, doomsday and extinction. Mechanical Models: Nonlinear spring-mass system, Soft and hard springs, Energy conservation, Phase plane and scenes. 680 Phase Plane Methods 10.1 Planar Autonomous Systems A set of two scalar differential equations of the form x0(t) = f(x(t); y(t)); (1) y0(t) = g(x(t); y(t)): is called a planar autonomous system. The term autonomous means self-governing, justified by the absence of the time variable t in the functions f(x; y), g(x; y). ! ! x(t) f(x; y) To obtain the vector form, let ~u(t) = , F~ (x; y) = y(t) g(x; y) and write (1) as the first order vector-matrix system d (2) ~u(t) = F~ (~u(t)): dt It is assumed that f, g are continuously differentiable in some region D in the xy-plane. This assumption makes F~ continuously differentiable in D and guarantees that Picard's existence-uniqueness theorem for initial d ~ value problems applies to the initial value problem dt ~u(t) = F (~u(t)), ~u(0) = ~u0. -
Calculus and Differential Equations II
Calculus and Differential Equations II MATH 250 B Linear systems of differential equations Linear systems of differential equations Calculus and Differential Equations II Second order autonomous linear systems We are mostly interested with2 × 2 first order autonomous systems of the form x0 = a x + b y y 0 = c x + d y where x and y are functions of t and a, b, c, and d are real constants. Such a system may be re-written in matrix form as d x x a b = M ; M = : dt y y c d The purpose of this section is to classify the dynamics of the solutions of the above system, in terms of the properties of the matrix M. Linear systems of differential equations Calculus and Differential Equations II Existence and uniqueness (general statement) Consider a linear system of the form dY = M(t)Y + F (t); dt where Y and F (t) are n × 1 column vectors, and M(t) is an n × n matrix whose entries may depend on t. Existence and uniqueness theorem: If the entries of the matrix M(t) and of the vector F (t) are continuous on some open interval I containing t0, then the initial value problem dY = M(t)Y + F (t); Y (t ) = Y dt 0 0 has a unique solution on I . In particular, this means that trajectories in the phase space do not cross. Linear systems of differential equations Calculus and Differential Equations II General solution The general solution to Y 0 = M(t)Y + F (t) reads Y (t) = C1 Y1(t) + C2 Y2(t) + ··· + Cn Yn(t) + Yp(t); = U(t) C + Yp(t); where 0 Yp(t) is a particular solution to Y = M(t)Y + F (t). -
Visualizing Quantum Mechanics in Phase Space
Visualizing quantum mechanics in phase space Heiko Baukea) and Noya Ruth Itzhak Max-Planck-Institut für Kernphysik, Saupfercheckweg 1, 69117 Heidelberg, Germany (Dated: 17 January 2011) We examine the visualization of quantum mechanics in phase space by means of the Wigner function and the Wigner function flow as a complementary approach to illustrating quantum mechanics in configuration space by wave func- tions. The Wigner function formalism resembles the mathematical language of classical mechanics of non-interacting particles. Thus, it allows a more direct comparison between classical and quantum dynamical features. PACS numbers: 03.65.-w, 03.65.Ca Keywords: 1. Introduction 2. Visualizing quantum dynamics in position space or momentum Quantum mechanics is a corner stone of modern physics and technology. Virtually all processes that take place on an space atomar scale require a quantum mechanical description. For example, it is not possible to understand chemical reactionsor A (one-dimensional) quantum mechanical position space the characteristics of solid states without a sound knowledge wave function maps each point x on the position axis to a of quantum mechanics. Quantum mechanical effects are uti- time dependent complex number Ψ(x, t). Equivalently, one lized in many technical devices ranging from transistors and may consider the wave function Ψ˜ (p, t) in momentum space, Flash memory to tunneling microscopes and quantum cryp- which is given by a Fourier transform of Ψ(x, t), viz. tography, just to mention a few applications. Quantum effects, however, are not directly accessible to hu- 1 ixp/~ Ψ˜ (p, t) = Ψ(x, t)e− dx . (1) man senses, the mathematical formulations of quantum me- (2π~)1/2 chanics are abstract and its implications are often unintuitive in terms of classical physics.