Sect. 15.7 (Min/Max/Saddle) —MATH 2421

Total Page:16

File Type:pdf, Size:1020Kb

Sect. 15.7 (Min/Max/Saddle) —MATH 2421 Sect. 15.7 (Min/Max/Saddle) —MATH 2421 Kawai Sect. 15.7 (Min/Max/Saddle) 1. Single-variable functions: (a) In principle, a constant function (horizontal line piece) has many local minimum and maximum points, but this is not interesting. We REALLY want: If f has a local maximum at x = a, then f (x) < f (a) for all x in some (small) interval around x = a. Often, we write the open interval this way: (a , a + ) , where could be a really small positive number. Thus, a local maximum should really be at the top of a hill, or, theoretically, it could be an isolated point: 1, x = 0 f (x) = 0, x = 0 6 We note that there is no mention of differentiability or even continuity in the definition of local maximum. The definition is similar for a local minimum. (b) If f is differentiable (and therefore continuous), then all local extrema must have f 0 (a) = 0, a horizontal tangent line. The converse is not true. If f 0 (a) = 0, then x = a could be an inflectionpoint. [Example: f (x) = x3 at x = 0.] If x = a is a critical number of f, then either f 0 (a) = 0 or f 0 (a) is undefined. This takes care of sharp turns (cusps) and jump discontinuities. (c) If f 00 (a) exists and is not equal to zero, then the curve is concave up when f 00 (a) > 0 and concave down when f 00 (a) < 0. It is inconclusive when f 00 (a) = 0. If the concavity changes at x = a, then x = a is definitely an inflection point. However, we see that f (x) = x4 must have a local minimum at x = 0 since the function is nonnegative, but f 00 (0) = 0 because the bottom of the bowl is too flat. (d) Second Derivative Test: If x = a is a critical number via f 0 (a) = 0, then the second derivative (if it exists) may classify the points as an extremum or an inflection point. As in the previous bullet, if the concavity changes at x = a, then we must have f 00 (a) = 0, and this is an inflection point. If the curve is concave up in the neighborhood around x = a, then x = a must be a local minimum. Thus, if f 00 (a) > 0, then we know with certainty that x = a is a local minimum. However, this definition also tells us what to examine when f 00 (a) = 0 and the curve is very flat x = a. Similarly, if the curve is concave down in that neighborhood, then x = a must be a local maximum. (e) If f 00 (a) = 0 then it an inflection point if and only if the concavity changes! As we have seen above, if the concavity does not change and f 0 (a) = 0, then it is a local extremum. 1 2. For a 2-var. function, z = f (x, y) , we say that a point (x0, y0) is a critical point (not a critical number) if: fx (x0, y0) = 0 AND fy (x0, y0) = 0 OR either partial derivative is undefined. In the latter case, this could be due to a sharp turn in the surface or a vertical tangent line/plane, or a jump discontinuity. 3. The Second Partials Test is similar to the Calc. 1 Second Derivative Test. Recall that in Calc. 3, for a 2-var. function, there are FOUR 2nd partial derivatives: fxx, fxy, fyx, fyy. From Clairaut, we know that fxy = fyx, so there are only three partial derivatives that we need to calculate for the test. From the Calc. 2 Maclaurin series, we recall that if we have an infinitely differentiable function at x = 0, then we can find the “best”quadratic (parabola) approximation to y = f (x) using this polynomial: f (0) f (0) p (x) = f (0) + 0 x + 00 x2. 1! 2! This assumes that f 00 (0) = 0 and we want the best parabola around the point (0, f (0)) . When we move away from x =6 0, then the approximation fits poorly, in general. Amazingly, this idea can be extended to 2nd degree polynomials in x and y. The Calc. 3 “best”quadric approximation to z = f (x, y) near (x, y) = (0, 0) is: f (0, 0) f (0, 0) f (0, 0) f (0, 0) f (0, 0) z = f (x, y) f (0, 0)+ x x+ y y+ xx x2+ xy xy+ yy y2 1! 1! 2! 1! 1! 2! If we assume that fx (0, 0) = fy (0, 0) = 0, then (0, 0) is a critical point and we have: f (0, 0) f (0, 0) f (0, 0) z = f (0, 0) + xx x2 + xy xy + yy y2 2! 1! 1! 2! and this can only be one of two quadric surfaces. It is either an elliptic paraboloid, in which case, (0, 0) is either a local min. or max. Else, it must be a hyperbold paraboloid, in which case, (0, 0) is a saddle point. [A saddle point has at least one direction that goes up, and at least one direction that goes down. Thus, it cannot be a local extremum.] 4. Example 1: Suppose we have z = x2 + xy + y2. We are sure that we have a critical point at (0, 0): fx = 2x + y = 0 fy = x + 2y = 0 and that we have a unique solution by substitution. 2 If we can get a graph of the level curve at k = 1, then we have: x2 + xy + y2 = 1 2 y It’san ellipse! 1 Since this is a quadric surface, it must be an elliptic paraboloid and (0, 0) is either a local min. or local max. •2 •1 1 2 x •1 •2 What if we could not produce this contour? (Sadness.) We can look at the 2nd partial derivatives and find: fxx = 2, fxy = 1. and fyy = 2. 5. The proof of the 2nd Partials Test in Appendix B is probably not very convincing, so I present another way to look at this. If you don’t know how to multiply matrices, then don’t worry. This idea is presented to students who have some advanced background in linear algebra. If we have an ellipse or hyperbola of the form: Ax2 + Bxy + Cy2 = 1, then let’sassume that both A and C are positive. It turns out that if A and C have opposite signs, then it must be a hyperbola (different algebra for that). If A and C are positive, then the relative size of B will determine which conic section this is. Matrix stuff: This is called a quadratic form. A B/2 x x y = Ax2 + Bxy + Cy2 = 1 B/2 C y By multiplying the vectors on the sides of the 2 2 matrix, we get a 1 1 scalar polynomial. So the trick is to DIAGONALIZE/rotate the 2 2 matrix so that it ends up looking like this (a diagonal matrix): P 0 0 Q For our matrix above, we want to diagonalize: 1 1/2 1/2 1 and the following rotation matrix does the trick: p2/2 p2/2 rotates points in the = +45 direction. p2/2 p2/2 3 Here is the diagonalization: p2/2 p2/2 1 1/2 p2/2 p2/2 3/2 0 = p2/2 p2/2 1/2 1 p2/2 p2/2 0 1/2 In a new set of xy-coordinates, the equation of the ellipse becomes: 3 1 x2 + y2 = 1 and this must be an ellipse. 2 2 6. The Second Partials Test simplifies this decision greatly. It turns out that we can determine the signs of the diagonal elements. Notice in the above example, the SUM of the diagonal elements in our original matrix was 2, and the sum of the diagonal elements in our rotated matrix is also 2. This is called the trace, and this value is invariant (does not change) under rotation. Thus, after our rotation, if both numbers are positive, then the critical point must be a local min. because it is a concave up elliptic paraboloid. If both numbers are negative, then the critical point must be a local max. If the numbers have opposite sign, then the critical point must be a saddle point, since the quadric surface is a hyperbolic paraboloid. 7. Here is the actual test: (a) We assume that (x0, y0) is a critical point with fx = fy = 0. (b) Evaluate fxx (x0, y0) , fyy (x0, y0) , and fxy (x0, y0) . (c) Let the discriminant: 2 d = fxxfyy (fxy) for that critical point. If d > 0, and fxx & fyy are positive, then it must be local min. If d > 0, and fxx & fyy are negative, then it must be local min. If d < 0, then it must be a saddle point. If d = 0, then the test is inconclusive (in a similar way that the Calc. 1 Second Derivative Test can be inconclusive). 8. Example 2: Suppose we have the function from p. 942: f (x, y) = x (x 2) y (y + 3) (a) Critical points. Again, we assume that we do not have any critical points where either fx or fy are undefined. The first condition gives us: fx = 0 = 2y (y + 3) (x 1) x = 1 OR y = 0 OR y = 3. ) Similarly, we have fy = 0 = x (x 2) (2y + 3) x = 0 OR x = 2 OR y = 3/2. ) 4 Now we must pair these conditions (there are 9 possible pairs taking one from fx = 0 and one from fy = 0).
Recommended publications
  • Math 2374: Multivariable Calculus and Vector Analysis
    Math 2374: Multivariable Calculus and Vector Analysis Part 25 Fall 2012 Extreme points Definition If f : U ⊂ Rn ! R is a scalar function, • a point x0 2 U is called a local minimum of f if there is a neighborhood V of x0 such that for all points x 2 V, f (x) ≥ f (x0), • a point x0 2 U is called a local maximum of f if there is a neighborhood V of x0 such that for all points x 2 V, f (x) ≤ f (x0), • a point x0 2 U is called a local, or relative, extremum of f if it is either a local minimum or a local maximum, • a point x0 2 U is called a critical point of f if either f is not differentiable at x0 or if rf (x0) = Df (x0) = 0, • a critical point that is not a local extremum is called a saddle point. First Derivative Test for Local Extremum Theorem If U ⊂ Rn is open, the function f : U ⊂ Rn ! R is differentiable, and x0 2 U is a local extremum, then Df (x0) = 0; that is, x0 is a critical point. Proof. Suppose that f achieves a local maximum at x0, then for all n h 2 R , the function g(t) = f (x0 + th) has a local maximum at t = 0. Thus from one-variable calculus g0(0) = 0. By chain rule 0 g (0) = [Df (x0)] h = 0 8h So Df (x0) = 0. Same proof for a local minimum. Examples Ex-1 Find the maxima and minima of the function f (x; y) = x2 + y 2.
    [Show full text]
  • A Stability Boundary Based Method for Finding Saddle Points on Potential Energy Surfaces
    JOURNAL OF COMPUTATIONAL BIOLOGY Volume 13, Number 3, 2006 © Mary Ann Liebert, Inc. Pp. 745–766 A Stability Boundary Based Method for Finding Saddle Points on Potential Energy Surfaces CHANDAN K. REDDY and HSIAO-DONG CHIANG ABSTRACT The task of finding saddle points on potential energy surfaces plays a crucial role in under- standing the dynamics of a micromolecule as well as in studying the folding pathways of macromolecules like proteins. The problem of finding the saddle points on a high dimen- sional potential energy surface is transformed into the problem of finding decomposition points of its corresponding nonlinear dynamical system. This paper introduces a new method based on TRUST-TECH (TRansformation Under STability reTained Equilibria CHaracter- ization) to compute saddle points on potential energy surfaces using stability boundaries. Our method explores the dynamic and geometric characteristics of stability boundaries of a nonlinear dynamical system. A novel trajectory adjustment procedure is used to trace the stability boundary. Our method was successful in finding the saddle points on different potential energy surfaces of various dimensions. A simplified version of the algorithm has also been used to find the saddle points of symmetric systems with the help of some analyt- ical knowledge. The main advantages and effectiveness of the method are clearly illustrated with some examples. Promising results of our method are shown on various problems with varied degrees of freedom. Key words: potential energy surfaces, saddle points, stability boundary, minimum gradient point, computational chemistry. I. INTRODUCTION ecently, there has been a lot of interest across various disciplines to understand a wide Rvariety of problems related to bioinformatics.
    [Show full text]
  • 3.2 Sources, Sinks, Saddles, and Spirals
    3.2. Sources, Sinks, Saddles, and Spirals 161 3.2 Sources, Sinks, Saddles, and Spirals The pictures in this section show solutions to Ay00 By0 Cy 0. These are linear equations with constant coefficients A; B; and C .C The graphsC showD solutions y on the horizontal axis and their slopes y0 dy=dt on the vertical axis. These pairs .y.t/;y0.t// depend on time, but time is not inD the pictures. The paths show where the solution goes, but they don’t show when. Each specific solution starts at a particular point .y.0/;y0.0// given by the initial conditions. The point moves along its path as the time t moves forward from t 0. D We know that the solutions to Ay00 By0 Cy 0 depend on the two solutions to 2 C C D As Bs C 0 (an ordinary quadratic equation for s). When we find the roots s1 and C C D s2, we have found all possible solutions : s1t s2t s1t s2t y c1e c2e y c1s1e c2s2e (1) D C 0 D C The numbers s1 and s2 tell us which picture we are in. Then the numbers c1 and c2 tell us which path we are on. Since s1 and s2 determine the picture for each equation, it is essential to see the six possibilities. We write all six here in one place, to compare them. Later they will appear in six different places, one with each figure. The first three have real solutions s1 and s2. The last three have complex pairs s a i!.
    [Show full text]
  • Lecture 8: Maxima and Minima
    Maxima and minima Lecture 8: Maxima and Minima Rafikul Alam Department of Mathematics IIT Guwahati Rafikul Alam IITG: MA-102 (2013) Maxima and minima Local extremum of f : Rn ! R Let f : U ⊂ Rn ! R be continuous, where U is open. Then • f has a local maximum at p if there exists r > 0 such that f (x) ≤ f (p) for x 2 B(p; r): • f has a local minimum at a if there exists > 0 such that f (x) ≥ f (p) for x 2 B(p; ): A local maximum or a local minimum is called a local extremum. Rafikul Alam IITG: MA-102 (2013) Maxima and minima Figure: Local extremum of z = f (x; y) Rafikul Alam IITG: MA-102 (2013) Maxima and minima Necessary condition for extremum of Rn ! R Critical point: A point p 2 U is a critical point of f if rf (p) = 0: Thus, when f is differentiable, the tangent plane to z = f (x) at (p; f (p)) is horizontal. Theorem: Suppose that f has a local extremum at p and that rf (p) exists. Then p is a critical point of f , i.e, rf (p) = 0: 2 2 Example: Consider f (x; y) = x − y : Then fx = 2x = 0 and fy = −2y = 0 show that (0; 0) is the only critical point of f : But (0; 0) is not a local extremum of f : Rafikul Alam IITG: MA-102 (2013) Maxima and minima Saddle point Saddle point: A critical point of f that is not a local extremum is called a saddle point of f : Examples: 2 2 • The point (0; 0) is a saddle point of f (x; y) = x − y : 2 2 2 • Consider f (x; y) = x y + y x: Then fx = 2xy + y = 0 2 and fy = 2xy + x = 0 show that (0; 0) is the only critical point of f : But (0; 0) is a saddle point.
    [Show full text]
  • Calculus and Differential Equations II
    Calculus and Differential Equations II MATH 250 B Linear systems of differential equations Linear systems of differential equations Calculus and Differential Equations II Second order autonomous linear systems We are mostly interested with2 × 2 first order autonomous systems of the form x0 = a x + b y y 0 = c x + d y where x and y are functions of t and a, b, c, and d are real constants. Such a system may be re-written in matrix form as d x x a b = M ; M = : dt y y c d The purpose of this section is to classify the dynamics of the solutions of the above system, in terms of the properties of the matrix M. Linear systems of differential equations Calculus and Differential Equations II Existence and uniqueness (general statement) Consider a linear system of the form dY = M(t)Y + F (t); dt where Y and F (t) are n × 1 column vectors, and M(t) is an n × n matrix whose entries may depend on t. Existence and uniqueness theorem: If the entries of the matrix M(t) and of the vector F (t) are continuous on some open interval I containing t0, then the initial value problem dY = M(t)Y + F (t); Y (t ) = Y dt 0 0 has a unique solution on I . In particular, this means that trajectories in the phase space do not cross. Linear systems of differential equations Calculus and Differential Equations II General solution The general solution to Y 0 = M(t)Y + F (t) reads Y (t) = C1 Y1(t) + C2 Y2(t) + ··· + Cn Yn(t) + Yp(t); = U(t) C + Yp(t); where 0 Yp(t) is a particular solution to Y = M(t)Y + F (t).
    [Show full text]
  • Maxima, Minima, and Saddle Points
    Real Analysis III (MAT312β) Department of Mathematics University of Ruhuna A.W.L. Pubudu Thilan Department of Mathematics University of Ruhuna | Real Analysis III(MAT312β) 1/80 Chapter 12 Maxima, Minima, and Saddle Points Department of Mathematics University of Ruhuna | Real Analysis III(MAT312β) 2/80 Introduction A scientist or engineer will be interested in the ups and downs of a function, its maximum and minimum values, its turning points. For instance, locating extreme values is the basic objective of optimization. In the simplest case, an optimization problem consists of maximizing or minimizing a real function by systematically choosing input values from within an allowed set and computing the value of the function. Department of Mathematics University of Ruhuna | Real Analysis III(MAT312β) 3/80 Introduction Cont... Drawing a graph of a function using a computer graph plotting package will reveal behavior of the function. But if we want to know the precise location of maximum and minimum points, we need to turn to algebra and differential calculus. In this Chapter we look at how we can find maximum and minimum points in this way. Department of Mathematics University of Ruhuna | Real Analysis III(MAT312β) 4/80 Chapter 12 Section 12.1 Single Variable Functions Department of Mathematics University of Ruhuna | Real Analysis III(MAT312β) 5/80 Local maximum and local minimum The local maximum and local minimum (plural: maxima and minima) of a function, are the largest and smallest value that the function takes at a point within a given interval. It may not be the minimum or maximum for the whole function, but locally it is.
    [Show full text]
  • Analysis of Asymptotic Escape of Strict Saddle Sets in Manifold Optimization
    Analysis of Asymptotic Escape of Strict Saddle Sets in Manifold Optimization Thomas Y. Hou∗ , Zhenzhen Li y , and Ziyun Zhang z Abstract. In this paper, we provide some analysis on the asymptotic escape of strict saddles in manifold optimization using the projected gradient descent (PGD) algorithm. One of our main contributions is that we extend the current analysis to include non-isolated and possibly continuous saddle sets with complicated geometry. We prove that the PGD is able to escape strict critical submanifolds under certain conditions on the geometry and the distribution of the saddle point sets. We also show that the PGD may fail to escape strict saddles under weaker assumptions even if the saddle point set has zero measure and there is a uniform escape direction. We provide a counterexample to illustrate this important point. We apply this saddle analysis to the phase retrieval problem on the low-rank matrix manifold, prove that there are only a finite number of saddles, and they are strict saddles with high probability. We also show the potential application of our analysis for a broader range of manifold optimization problems. Key words. Manifold optimization, projected gradient descent, strict saddles, stable manifold theorem, phase retrieval AMS subject classifications. 58D17, 37D10, 65F10, 90C26 1. Introduction. Manifold optimization has long been studied in mathematics. From signal and imaging science [38][7][36], to computer vision [47][48] and quantum informa- tion [40][16][29], it finds applications in various disciplines.
    [Show full text]
  • Instability Analysis of Saddle Points by a Local Minimax Method
    INSTABILITY ANALYSIS OF SADDLE POINTS BY A LOCAL MINIMAX METHOD JIANXIN ZHOU Abstract. The objective of this work is to develop some tools for local insta- bility analysis of multiple critical points, which can be computationally carried out. The Morse index can be used to measure local instability of a nondegen- erate saddle point. However, it is very expensive to compute numerically and is ineffective for degenerate critical points. A local (weak) linking index can also be defined to measure local instability of a (degenerate) saddle point. But it is still too difficult to compute. In this paper, a local instability index, called a local minimax index, is defined by using a local minimax method. This new instability index is known beforehand and can help in finding a saddle point numerically. Relations between the local minimax index and other local insta- bility indices are established. Those relations also provide ways to numerically compute the Morse, local linking indices. In particular, the local minimax index can be used to define a local instability index of a saddle point relative to a reference (trivial) critical point even in a Banach space while others failed to do so. 1. Introduction Multiple solutions with different performance, maneuverability and instability in- dices exist in many nonlinear problems in natural or social sciences [1,7, 9,24,29,31, 33,35]. When cases are variational, the problems reduce to solving the Euler- Lagrange equation (1.1) J 0(u) = 0; where J is a C1 generic energy functional on a Banach space H and J 0 its Frechet derivative.
    [Show full text]
  • Gradient Descent Can Take Exponential Time to Escape Saddle Points
    Gradient Descent Can Take Exponential Time to Escape Saddle Points Simon S. Du Chi Jin Carnegie Mellon University University of California, Berkeley [email protected] [email protected] Jason D. Lee Michael I. Jordan University of Southern California University of California, Berkeley [email protected] [email protected] Barnabás Póczos Aarti Singh Carnegie Mellon University Carnegie Mellon University [email protected] [email protected] Abstract Although gradient descent (GD) almost always escapes saddle points asymptot- ically [Lee et al., 2016], this paper shows that even with fairly natural random initialization schemes and non-pathological functions, GD can be significantly slowed down by saddle points, taking exponential time to escape. On the other hand, gradient descent with perturbations [Ge et al., 2015, Jin et al., 2017] is not slowed down by saddle points—it canfind an approximate local minimizer in polynomial time. This result implies that GD is inherently slower than perturbed GD, and justifies the importance of adding perturbations for efficient non-convex optimization. While our focus is theoretical, we also present experiments that illustrate our theoreticalfindings. 1 Introduction Gradient Descent (GD) and its myriad variants provide the core optimization methodology in machine learning problems. Given a functionf(x), the basic GD method can be written as: x(t+1) x (t) η f x(t) , (1) ← − ∇ where η is a step size, assumedfixed in the current paper.� While� precise characterizations of the rate of convergence GD are available for convex problems, there is far less understanding of GD for non-convex problems. Indeed, for general non-convex problems, GD is only known tofind a stationary point (i.e., a point where the gradient equals zero) in polynomial time [Nesterov, 2013].
    [Show full text]
  • Escaping from Saddle Points on Riemannian Manifolds
    Escaping from saddle points on Riemannian manifolds Yue Suny, Nicolas Flammarionz, Maryam Fazely y Department of Electrical and Computer Engineering, University of Washington, Seattle z School of Computer and Communication Sciences, EPFL Abstract Taylor series and smoothness assumptions Example – Burer-Monteiro factorization. Notation: Expx(y) denotes the exponential map, gradf(x) and H(x) d×d Let A 2 S , the problem 6 We consider minimizing a nonconvex, smooth function f(x) on a are Riemannian gradient and Hessian of f(x); x is a saddle point, and smooth manifold x 2 M. We show that a perturbed Riemannian −1 max trace(AX); 5 u = Exp (u). 4 x X2 d×d gradient algorithm converges to a second-order stationary point in a S 3 • Riemannian gradient descent. Let f have a β-Lipschitz gradient. Function value number of iterations that is polynomial in appropriate smoothness s:t: diag(X) = 1;X 0; rank(X) ≤ r: 2 q There exists η = Θ(1/β) such that Riemannian gradient descent step 1 parameters of f and M, and polylog in dimension. This matches the can be factorized as + + 0 0 5 10 15 20 25 30 35 40 45 best known rate for unconstrained smooth minimization. u = Exp (−ηgradf(u)) (c.f. Euclidean case: u = u − ηrf(u)) Iterations u max trace(AY Y T ); s:t: diag(YY T ) = 1: η Iteration versus function value. The iterations 2 Y 2 d×p monotonically decreases f by 2kgradf(u)k . R start from a saddle point, is perturbed by the Background and motivation ^ when r(r + 1)=2 ≤ d, p(p + 1)=2 ≥ d.
    [Show full text]
  • Escaping from Saddle Points on Riemannian Manifolds
    Escaping from saddle points on Riemannian manifolds Yue Suny, Nicolas Flammarionz, Maryam Fazely y Department of Electrical and Computer Engineering, University of Washington, Seattle z School of Computer and Communication Sciences, EPFL, Lausanne, Switzerland November 8, 2019 1 / 24 Manifold constrained optimization We consider the problem minimize f (x); subject to x 2 M x Same as Euclidean space, generally we cannot find global optimum in polynomial time, so we want to find an approximate local minimum. Plot of saddle point in Euclidean Contour of function value on space. sphere. 2 / 24 Example of manifolds d Pd 2 2 1. Sphere. fx 2 R : i=1 xi = r g. m×n T 2. Stiefel manifold. fX 2 R : X X = I g. 3. Grassmannian manifold. Grass(p; n) is set of p dimensional n subspaces in R . m×n T 4. Bruer-Monteiro relaxation. fX 2 R : diag(X X ) = 1g. 3 / 24 Curve A curve in a continuous map γ : t !M. Usually t 2 [0; 1], where γ(0) and γ(1) are start and end points of the curve. 4 / 24 Tangent vector and tangent space We use x 2 M can be start point of many γ(t + τ) − γ(t) γ_ (t) = lim curves, and a tangent space Tx M is τ!0 τ the set of tangent vectors at x. as the velocity of the curve, γ_ (t) is a Tangent space is a metric space. tangent vector at γ(t) 2 M. 5 / 24 Gradient of a function Let f : M! R be a function defined on M, and γ be a curve on M.
    [Show full text]
  • Qualitative Analysis of Differential Equations
    QUALITATIVE ANALYSIS OF DIFFERENTIAL EQUATIONS Alexander Panfilov arXiv:1803.05291v1 [math.GM] 10 Mar 2018 det A D=0 stable spiral non−stable spiral 6 3 5 4 2 stable center non−stable node node tr A 1 saddle 2 QUALITATIVE ANALYSIS OF DIFFERENTIAL EQUATIONS Alexander Panfilov Theoretical Biology, Utrecht University, Utrecht c 2010 2010 2 Contents 1 Preliminaries 5 1.1 Basic algebra . 5 1.1.1 Algebraic expressions . 5 1.1.2 Limits . 6 1.1.3 Equations . 7 1.1.4 Systems of equations . 8 1.2 Functions of one variable . 9 1.3 Graphs of functions of one variable . 11 1.4 Implicit function graphs . 18 1.5 Exercises . 19 2 Selected topics of calculus 23 2.1 Complex numbers . 23 2.2 Matrices . 25 2.3 Eigenvalues and eigenvectors . 27 2.4 Functions of two variables . 32 2.5 Exercises . 35 3 Differential equations of one variable 37 3.1 Differential equations of one variable and their solutions . 37 3.1.1 Definitions . 37 3.1.2 Solution of a differential equation . 39 3.2 Qualitative methods of analysis of differential equations of one variable . 42 3.2.1 Phase portrait . 42 3.2.2 Equilibria, stability, global plan . 43 3.3 Systems with parameters. Bifurcations. 46 3.4 Exercises . 48 4 System of two linear differential equations 51 4.1 Phase portraits and equilibria . 51 4.2 General solution of linear system . 53 4.3 Real eigen values. Saddle, node. 56 4.3.1 Saddle; l1 < 0;l2 > 0, or l1 > 0;l2 < 0 ....................
    [Show full text]