Intro to Differential Equations: a Subtle Subject

Intro to Differential Equations: a Subtle Subject

Intro to Differential Equations: A Subtle Subject Big Question: Given a differential equation, does it have any solutions? (Some have many solutions. Others don't have any at all.) Which have solutions and which don't? This is a very difficult question. Q: If our differential equation has solutions, can we find all of the solutions explicitly? (i.e.: Can we find actual formulas for the solutions?) Usually no. Only in rare cases can we actually find explicit solutions. Best Case Scenario: Would be finding all explicit solutions. But to reiterate: This is usually not possible! Next Best Thing: If we are lucky, we might be able to: (1) Determine qualitative behavior of solutions. (2) Prove an abstract existence-uniqueness theorem. (3) Construct approximations to solutions by a numerical method. Note: Even if a differential equation has solutions, different solution functions might have different domains! The domain need not be all of R. For instance: Example: The differential equation y0 = y2 has infinitely many solutions. ◦ One solution has its domain as all of R. ◦ All other solutions have domains like (−∞; a) [ (a; 1) for some a 2 R. Types of Differential Equations There are four ways to distinguish differential equations: ◦ ODE vs PDE ◦ Single Equations vs Systems of Equations ◦ 1st-order vs 2nd-order vs 3rd-order etc. ◦ Linear vs Nonlinear This Class: Only ODEs.* Mostly linear ones (but some nonlinear). ◦ 1st-Order ODEs. (Week 1) ◦ 1st-Order systems of ODEs. (Weeks 3-6) ◦ 2nd-Order ODEs. (Weeks 6-9) (* At the end of the course, we'll examine a PDE (Laplace Equation).) 1st-Order ODEs General Form: A first-order ODE looks like: dy = F (t; y): dt There is no general method for finding explicit solutions to all first-order ODEs. (Life is unfair.) However, there are a few things we can say about 1st-Order ODEs. (0) Special Cases: Some special kinds of first-order ODEs can be solved explicitly (more or less). A couple of these are discussed below. (1) Geometric Insight: Direction fields. Direction fields are a geometric technique for gaining insight into the be- havior of solutions to first-order ODEs. (2) Abstract Existence-Uniqueness Theorem: Q: Do all first-order ODE's have solutions? This is subtle. A first answer in this direction is given by an abstract existence-uniqueness theorem. We will discuss this much later in the course. (3) Approximations to Solutions: There are numerical methods (e.g.: Euler's Method) for approximating solutions. We will discuss this much later in the course. Special Kinds of 1st-Order ODEs (A) Linear. These have the form dy = p(t)y + g(t): dt They are called homogeneous if g(t) ≡ 0. Else, they are non-homogeneous. To solve: Method of Integrating Factors. (B) Separable. These have the form dy dy f(t) = f(t)g(y) or = : dt dt g(y) To solve: Method of Separation of Variables. Method of Integrating Factors Goal: Explicitly solve first-order linear ODE. That is: dy + p(t)y = g(t): dt Method: 0. Start with y0 + p(t)y = g(t). 1. Multiply both sides by the integrating factor, Z µ(t) = exp p(t) dt : Result is: µ(t)y0 + µ(t)p(t)y = µ(t)g(t): 2. Write the result as (µ(t)y)0 = µ(t)g(t): 3. Integrate and solve for y(t). Method of Separation of Variables Goal: Explicitly solve first-order separable ODE. That is: dy = f(t)g(y): dt Method: dy 0. Start with = f(t)g(y): dt 1. Move all functions of y to one side, and all functions of t to the other: dy = f(t) dt: g(y) 2. Integrate and solve for y. dy f(t) Note: Separable equations can also be written in the form = . dt g(y) The text describes them in the form f(t) dt + g(y) dy = 0, which is also the same thing. Eigenvalues and Eigenvectors: Motivation The simplest matrices are the diagonal matrices. These are of the form 0 1 d1 0 ··· 0 B 0 d ··· 0 C B 2 C D = B . .. C: @ . 0 A 0 0 ··· dn Geometrically, D acts as follows: ◦ Vectors along the x1-axis are stretched/contracted/reflected by d1. ◦ Vectors along the xn-axis are stretched/contracted/reflected by dn. Note: Stretching along the xi-axis happens if jdij > 1. Contracting along the xi-axis happens if jdij < 1. Reflection happens if di is negative. Hopes and Dreams: ◦ Maybe every matrix is really a diagonal matrix after a \change-of-basis." ◦ Maybe every matrix A is similar to a diagonal matrix, meaning that A = T DT −1 for a diagonal matrix D. ◦ Maybe every matrix geometrically acts like a diagonal matrix, stretching some line L1 by a factor d1, some other line L2 by a factor d2, and so on. Reality Check: None of these hopes are true as stated. (Life is hard.) But there is hope in the following two theorems. Def: A matrix A is diagonalizable iff A is similar to a diagonal matrix. That is: A = T DT −1 for some diagonal D and some invertible T . Diagonalization Criterion: Let A be an n × n matrix. Then: A is diagonalizable () A has n linearly independent eigenvectors. Spectral Theorem: If A is a symmetric matrix, then A is diagonalizable. Eigenvalues and Eigenvectors: Definitions Def: Let A be an n × n matrix. An eigenvector of A is a non-zero vector v 2 Cn, v 6= 0 such that Av = λv for some scalar λ 2 C. The scale factor λ 2 C is the eigenvalue associated to the eigenvector v. Note: If v is an eigenvector for A, then so too is every scalar multiple cv. ) Every eigenvector of A determines an entire line of eigenvectors of A. Q: How to find eigenvectors and their eigenvalues? A: Observe that: Av = λv for some v 6= 0 () (A − λI)v = 0 for some v 6= 0 () det(A − λI) = 0: Def: The characteristic polynomial of A is p(λ) = det(A − λI): So: Eigenvalues are the zeros of the characteristic polynomial! By the Fundamental Theorem of Algebra, we can factor p(λ): n m1 mk p(λ) = (−1) (λ − λ1) ··· (λ − λk) : The algebraic multiplicity of λj is the corresponding exponent mj. Def: Let A be an n × n matrix. The eigenspace of λ 2 C is n E(λ) = fv 2 C j Av = λvg = N(A − λI): The geometric multiplicity of λ is: GeoMult(λ) = dim(E(λ)) = # of linearly independent eigenvectors with eigenvalue λ. Fact: For every eigenvalue λ, we have: GeoMult(λ) ≤ AlgMult(λ): Eigenvalues and Eigenvectors: More Facts To Find Eigenvectors of A: (1) First find the eigenvalues. They will be the zeros of the characteristic polynomial p(λ) = det(A − λI): (2) For each eigenvalue λ that you found in part (1), solve the equation (A − λI)v = 0 for vectors v. These vectors are the eigenvectors associated to λ. (In fact, these vectors are the elements of E(λ) = N(A − λI).) Fact: Let A be an n × n matrix with eigenvalues λ1; : : : ; λn (each repeated according to its algebraic multiplicity). (a) tr(A) = λ1 + ··· + λn. (b) det(A) = λ1 ··· λn. Fact: Let A be an n × n matrix (with real entries). If λ = a + ib is an eigenvalue of A, then λ = a − ib is an eigenvalue of A. Fact: If v1;:::; vk are eigenvectors associated to distinct eigenvalues, then fv1;:::; vkg is linearly independent. Diagonalization Criterion: Let A be an n × n matrix. Then: A is diagonalizable () A has n linearly independent eigenvectors () Every eigenvalue λ of A has GeoMult(λ) = AlgMult(λ): Review: Dot Products Recall: The dot product (or inner product) of u; v 2 Rn is u · v = hu; vi = u1v1 + ··· + unvn: 2 2 2 Also: kvk = v1 + ··· + vn = hv; vi. Def: A set of vectors fv1;:::; vkg is an orthogonal set iff every pair of (distinct) vectors is orthogonal: hvi; vji = 0 for all i 6= j. Fact: If fv1;:::; vkg is an orthogonal set of non-zero vectors, then fv1;:::; vkg is linearly independent. Symmetric Matrices Def: An n × n matrix is symmetric iff AT = A. Theorem: Let A be a symmetric matrix. Then: (a) All eigenvalues of A are real. (b) If v1; v2 are eigenvectors of A corresponding to different eigenvalues, then hv1; v2i = 0. (c) (Spectral Theorem) There exists a set of n linearly independent eigen- vectors for A. Thus, A is diagonalizable. N.B.: Parts (a), (b), (c) are stated under the assumption that A is symmet- ric. If A is not symmetric, then (a), (b), (c) may or may not be true! Jordan Form Diagonalization Criterion: Let A be an n × n matrix. Then: A is diagonalizable () A has n linearly independent eigenvectors () Every eigenvalue λ of A has GeoMult(λ) = AlgMult(λ): Q: If A is not diagonalizable, is A still similar to a \simple" matrix? A: Yes: If we can't diagonalize, the next best thing is Jordan form. Def: A Jordan block is a matrix of the form 0 1 λ 1 B λ 1 C λ 1 B C λ or or ::: or B .. .. C 0 λ @ . .A λ A matrix is in Jordan form if it is of the form 0 1 J1 B J C B 2 C J = B .. C; @ . A Jk where each Ji is a Jordan block (possibly of different sizes and λ's). Example: The following two matrices are both in Jordan form: 04 1 04 1 B 7 1 C B 7 C B C B C B 7 C B 7 C J = B C and J = B C B 4 1 C B 4 C B C B C @ 4 1A @ 4 A 4 4 Notice: Diagonal matrices are exactly the Jordan form matrices in which every Jordan block has size 1.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    19 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us