Nonlinear Systems Theory
Matthew M. Peet Arizona State University
Lecture 02: Nonlinear Systems Theory
Overview
Our next goal is to extend LMI’s and optimization to nonlinear systems analysis.
Today we will discuss 1. Nonlinear Systems Theory 1.1 Existence and Uniqueness 1.2 Contractions and Iterations 1.3 Gronwall-Bellman Inequality 2. Stability Theory 2.1 Lyapunov Stability 2.2 Lyapunov’s Direct Method 2.3 A Collection of Converse Lyapunov Results The purpose of this lecture is to show that Lyapunov stability can be solved Exactly via optimization of polynomials.
M. Peet Lecture 02: 1 / 56 Ordinary Nonlinear Differential Equations Computing Stability and Domain of Attraction
Consider: A System of Nonlinear Ordinary Differential Equations
x˙(t) = f(x(t))
Problem: Stability Given a specific polynomial f : Rn → Rn, find the largest X ⊂ Rn such that for any x(0) ∈ X, limt→∞ x(t) = 0.
M. Peet Lecture 02: 2 / 56 Nonlinear Dynamical Systems Long-Range Weather Forecasting and the Lorentz Attractor
A model of atmospheric convection analyzed by E.N. Lorenz, Journal of Atmospheric Sciences, 1963. x˙ = σ(y − x)y ˙ = rx − y − xz z˙ = xy − bz
M. Peet Lecture 02: 3 / 56 Stability and Periodic Orbits The Poincar´e-BendixsonTheorem and van der Pol Oscillator
3
2 An oscillating circuit model: 1 y˙ = −x − (x2 − 1)y
y 0 x˙ = y −1
−2 Domain−of−attraction −3 −3 −2 −1 0 1 2 3 x Figure : The van der Pol oscillator in reverse
Theorem 1 (Poincar´e-Bendixson).
Invariant sets in R2 always contain a limit cycle or fixed point.
M. Peet Lecture 02: 4 / 56 Stability of Ordinary Differential Equations
Consider x˙(t) = f(x(t))
n with x(0) ∈ R .
Theorem 2 (Lyapunov Stability). Suppose there exists a continuous V and α, β, γ > 0 where
βkxk2 ≤ V (x) ≤ αkxk2 −∇V (x)T f(x) ≥ γkxk2
for all x ∈ X. Then any sub-level set of V in X is a Domain of Attraction.
M. Peet Lecture 02: 5 / 56 Mathematical Preliminaries Cauchy Problem
The first question people ask is the Cauchy problem: Autonomous System: Definition 3. The system x˙(t) = f(x(t)) is said to satisfy the Cauchy problem if there exists a n continuous function x : [0, tf ] → R such that x˙ is defined and x˙(t) = f(x(t)) for all t ∈ [0, tf ]
If f is continuous, the solution must be continuously differentiable. Controlled Systems: • For a controlled system, we have x˙(t) = f(x(t), u(t)). • At this point u is undefined, so for the Cauchy problem, we take x˙(t) = f(t, x(t)) • In this lecture, we consider the autonomous system.
I Including t complicates the analysis.
I However, results are almost all the same.
M. Peet Lecture 02: 6 / 56 Lecture 1++, 2006
• Nonlinear Phenomena and Stability theory
◮ Nonlinear phenomena [Khalil Ch 3.1] ◮ existence and uniqueness Nonlinear Control Theory 2006 ◮ finite escape time ◮ peaking
◮ Linear system theory revisited ◮ Second order systems [Khalil Ch 2.4, 2.6] ◮ periodic solutions / limit cycles ◮ Stability theory [Khalil Ch. 4] ◮ Lyapunov Theory revisited ◮ exponential stability Ordinary Differential Equations◮ quadratic stability ◮ time-varying systems Existence of Solutions ◮ invariant sets ◮ center manifold theorem There exist many systems for which no solution exists or for which a solution Existence problemsonly of exists solutions over a finite time interval. Finite Escape Time
Even for something as simple as 2 Finite escape time of dx/dt = x Example: The differential equation 5 2 x˙(t) = x(t) x(0) = x 4.5 dx 2 0 x , x 0 x0 dt = ( )= 4 has the solution 3.5 has the solution 3 x0 x0 1 x(t) = 2.5
x t , 0 t x(t) ( )= 1 x t ≤ < x 1 − x0t − 0 0 2 1.5 Finite escape time which clearly has escape time 1 1 t f = 1 0.5 x0 t = e 0 x0 0 1 2 3 4 5 Time t
Figure : Simulation of x˙ = x2 for several x(0) Uniqueness Problems Previous problem is like the water-tank problem in backward M. Peet time Lecture 02: 7 / 56 Example: The equation x˙ √x, x 0 0 has many solutions: = ( )= (Substitute τ t in differential equation). =− t C 2 4 t C x t ( − ) / > ( ) = 0 t C ≤
2
1.5
1
0.5 x(t)
0
−0.5
−1 0 1 2 3 4 5 Time t dh dt a√h, h : height water level / =− ( ) Compare with water tank: Change to backward-time: “If I see it empty, when was it full?”)
Existence and Uniqueness The peaking phenomenon
Theorem
Let Ω R denote the ball Example: Controlled linear system with right-half plane zero Feedback can change location of poles but not location of zero Ω z; z a R R = { q − q≤ } (unstable pole-zero cancellation not allowed). If f is Lipschitz-continuous: s 1 ω 2 o (1) Gcl s 2(− + ) 2 f z f y K z y , for all z, y Ω ( )= s 2ω os ω o q ( )− ( )q ≤ q − q ∈ + + then x˙ t f x t , x 0 a has a unique solution in ( )= ( ( )) ( )= 0 t R C , A step response will reveal a transient which grows in amplitude ≤ < / R for faster closed loop poles s ω o, see Figure on next slide. where C max f x =− R = ΩR q ( )q see [Khalil Ch. 3]
1 Ordinary Differential Equations Non-Uniqueness A classical example of a system without a unique solution is x˙(t) = x(t)1/3 x(0) = 0 For the given initial condition, it is easy to verify that 2t 3/2 x(t) = 0 and x(t) = 3 both satisfy the differential equation.
1.6 1.6
1.4 1.4
1.2 1.2
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
Figure : Matlab simulation of Figure : Matlab simulation of x˙(t) = x(t)1/3 with x(0) = 0 x˙(t) = x(t)1/3 with x(0) = .000001 M. Peet Lecture 02: 8 / 56 Lecture 1++, 2006
• Nonlinear Phenomena and Stability theory
◮ Nonlinear phenomena [Khalil Ch 3.1] ◮ existence and uniqueness Nonlinear Control Theory 2006 ◮ finite escape time ◮ peaking
◮ Linear system theory revisited ◮ Second order systems [Khalil Ch 2.4, 2.6] ◮ periodic solutions / limit cycles ◮ Stability theory [Khalil Ch. 4] ◮ Lyapunov Theory revisited ◮ exponential stability ◮ quadratic stability ◮ time-varying systems ◮ invariant sets ◮ center manifold theorem
Existence problems of solutions Finite Escape Time
Finite escape time of dx/dt = x2 Example: The differential equation 5
4.5 dx 2 x , x 0 x0 dt = ( )= 4 3.5 has the solution 3
x0 1 2.5
x t , 0 t x(t) ( )= 1 x t ≤ < x − 0 0 2 1.5 Finite escape time 1 1 t f = x0 0.5 0 0 1 2 3 4 5 Time t
Ordinary Differential Equations Uniqueness Problems Previous problem is like the water-tank problem in backward Non-Uniqueness time Example: The equation x˙ √x, x 0 0 has many solutions: = ( )= (Substitute τ t in differential equation). Another Example of a system with several solutions is given by =− t C 2 4 t C x˙(t) = x(t) x t x(0) = 0 ( − ) / > ( ) = 0 t C p ≤
For the given initial condition, it is easy 2 to verify that for any C, 1.5
(t−C)2 t > C 1 x(t) = 4 0.5 (0 t ≤ C x(t)
0 satisfies the differential equation. −0.5
−1 0 1 2 3 4 5 Time t √ √ dh dt a h, h : height water level Figure : Several solutions of x˙ = x / =− ( ) Compare with water tank: Change to backward-time: “If I see it empty, when was it full?”)
M. Peet Lecture 02: 9 / 56 Existence and Uniqueness The peaking phenomenon
Theorem
Let Ω R denote the ball Example: Controlled linear system with right-half plane zero Feedback can change location of poles but not location of zero Ω z; z a R R = { q − q≤ } (unstable pole-zero cancellation not allowed). If f is Lipschitz-continuous: s 1 ω 2 o (1) Gcl s 2(− + ) 2 f z f y K z y , for all z, y Ω ( )= s 2ω os ω o q ( )− ( )q ≤ q − q ∈ + + then x˙ t f x t , x 0 a has a unique solution in ( )= ( ( )) ( )= 0 t R C , A step response will reveal a transient which grows in amplitude ≤ < / R for faster closed loop poles s ω o, see Figure on next slide. where C max f x =− R = ΩR q ( )q
see [Khalil Ch. 3]
1 Ordinary Differential Equations Customary Notions of Continuity
Definition 4. For normed linear spaces X,Y , a function f : X → Y is said to be continuous at the point x0 ⊂ X if for any > 0, there exists a δ > 0 such that kx − x0k < δ implies kf(x) − f(x0)k < .
M. Peet Lecture 02: 10 / 56 Ordinary Differential Equations Customary Notions of Continuity
Definition 5. For normed linear spaces X,Y , a function f : A ⊂ X → Y is said to be continuous on B ⊂ A if it is continuous for any point x0 ∈ B. A function is said to be simply continuous if B = A.
Definition 6. For normed linear spaces X,Y , a function f : A ⊂ X → Y is said to be uniformly continuous on B ⊂ A if for any > 0, there exists a δ > 0 such that for any points x, y ∈ B, kx − yk < δ implies kf(x) − f(y)k < .A function is said to be simply uniformly continuous if B = A.
M. Peet Lecture 02: 11 / 56 Lipschitz Continuity A Quantitative Notion of Continuity Definition 7. We say the function f is Lipschitz continuous on X if there exists some L > 0 such that kf(x) − f(y)k ≤ Lkx − yk for any x, y ∈ X. The constant L is referred to as the Lipschitz constant for f on X. Definition 8. We say the function f is Locally Lipschitz continuous on X if for every x ∈ X, there exists a neighborhood, B of x such that f is Lipschitz continuous on B. Definition 9. We say the function f is globally Lipschitz if it is Lipschitz continuous on its entire domain. It turns out that smoothness of the vector field is the critical factor. • Not a Necessary condition, however. • The Lipschitz constant, L, allows us to quantify the roughness of the vector field. M. Peet Lecture 02: 12 / 56 Ordinary Differential Equations Existence and Uniqueness
Theorem 10 (Simple).
n n n Suppose x0 ∈ R , f : R → R and there exist L, r such that for any x, y ∈ B(x0, r), kf(x) − f(y)k ≤ Lkx − yk 1 r and kf(x)k ≤ c. Let b < min{ L , c }. Then there exists a unique differentiable map x ∈ C[0, b], such that x(0) = x0, x(t) ∈ B(x0, r) and x˙(t) = f(x(t)). Because the approach to its proof is so powerful, it is worth presenting the proof of the existence theorem.
M. Peet Lecture 02: 13 / 56 Ordinary Differential Equations Contraction Mapping Theorem
Theorem 11 (Contraction Mapping Principle). Let (X, k·k) be a complete normed space and let P : X → X. Suppose there exists a ρ < 1 such that
kP x − P yk ≤ ρkx − yk for all x, y ∈ X.
Then there is a unique x∗ ∈ X such that P x∗ = x∗. Furthermore for y ∈ X, define the sequence {xi} as x1 = y and xi = P xi−1 for i > 2. Then ∗ limi→∞ xi = x . Some Observations: • Proof: Show that P ky is a Cauchy sequence for any y ∈ X. • For a differentiable function P , P is a contraction if and only if kP˙ k < 1. • In our case, X is the space of solutions. The contraction is
t (P x)(t) = x0 + f(x(s))ds Z0
M. Peet Lecture 02: 14 / 56 Ordinary Differential Equations Contraction Mapping Theorem
This contraction derives from the fundamental theorem of calculus. Theorem 12 (Fundamental Theorem of Calculus).
n n Suppose x ∈ C and f : M × R → R is continuous and M = [0, tf ] ⊂ R. Then the following are equivalent. • x is differentiable at any t ∈ M and
x˙(t) = f(x(t)) at all t ∈ M (1)
x(0) = x0 (2)
• t x(t) = x0 + f(x(s))ds for all t ∈ M Z0
M. Peet Lecture 02: 15 / 56 Ordinary Differential Equations Existence and Uniqueness
First recall what we are trying to prove: Theorem 13 (Simple).
n n n Suppose x0 ∈ R , f : R → R and there exist L, r such that for any x, y ∈ B(x0, r), kf(x) − f(y)k ≤ Lkx − yk 1 r and kf(x)k ≤ c. Let b < min{ L , c }. Then there exists a unique differentiable map x ∈ C[0, b], such that x(0) = x0, x(t) ∈ B(x0, r) and x˙(t) = f(x(t)). We will show that t (P x)(t) = x0 + f(x(s))ds Z0 is a contraction.
M. Peet Lecture 02: 16 / 56 Proof of Existence Theorem Proof.
For given x0, define the space B = {x ∈ C[0, b]: x(0) = x0, x(t) ∈ B(x0, r)} with norm supt∈[0,b]kx(t)k which is complete. Define the map P as t P x(t) = x0 + f(x(s))ds Z0 We first show that P maps B to B. Suppose x ∈ B. To show P x ∈ B, we first show that P x a continuous function of t. t2 t2 kP x(t2) − P x(t1)k = k f(x(s))dsk ≤ kf(x(s))kds ≤ c(t2 − t1) Zt1 Zt1
Thus P x is continuous. Clearly P x(0) = x0. Now we show x(t) ∈ B(x0, r). t sup kP x(t) − x0k = sup k f(x(s))dsk t∈[0,b] t∈[0,b] Z0 b ≤ kf(x(s))kds Z0 ≤ bc < r
M. Peet Lecture 02: 17 / 56 Proof of Existence Theorem Proof. Now we have shown that P : B → B. To prove existence and uniqueness, we show that Φ is a contraction. t kP x − P yk = sup k f(x(s)) − f(y(s))dsk t∈[0,b] Z0 t b ≤ sup ( kf(x(s)) − f(y(s))kds) ≤ kf(x(s)) − f(y(s))kds t∈[0,b] Z0 Z0 b ≤ L kx(s) − y(s)kds ≤ Lbkx − yk Z0 Thus, since Lb < 1, the map is a contraction with a unique fixed point x ∈ B such that t x(t) = x0 + f(x(s))ds 0 By the fundamental theorem of calculus,Z this means that x is a differentiable function such that for t ∈ [0, b] x˙ = f(x(t))
M. Peet Lecture 02: 18 / 56 Picard Iteration Make it so
This proof is particularly important because it provides a way of actually constructing the solution. Picard-Lindel¨ofIteration: • From the proof, unique solution of P x∗ = x∗ is a solution of x˙ ∗ = f(x∗), where t (P x)(t) = x0 + f(x(s))ds Z0 • From the contraction mapping theorem, the solution P x∗ = x∗ can be found as x∗ = lim P kz for any z ∈ B k→∞
M. Peet Lecture 02: 19 / 56 Extension of Existence Theorem
Note that this existence theorem only guarantees existence on the interval
1 r t ∈ 0, or t ∈ 0, L c Where h i
• r is the size of the neighborhood near x0
• L is a Lipschitz constant for f in the neighborhood of x0
• c is a bound for f in the neighborhood of x0 Note further that this theorem only gives a solution for a particular initial condition x0 • It does not imply existence of the Solution Map However, convergence of the solution map can also be proven.
M. Peet Lecture 02: 20 / 56 Illustration of Picard Iteration
This is a plot of Picard iterations for the solution map of x˙ = −x3. z(t, x) = 0; P z(t, x) = x; P 2z(t, x) = x − tx3; P 3z(t, x) = x − tx3 + 3t2x5 − 3t3x7 + t4x9
1.5
1.0
0.5
0.2 0.4 0.6 0.8 1.0
Figure : The solution for x0 = 1
Convergence is only guaranteed on interval t ∈ [0,.333].
M. Peet Lecture 02: 21 / 56 Extension of Existence Theorem
Theorem 14 (Extension Theorem).
For a given set W and r, define the set Wr := {x : kx − yk ≤ r, y ∈ W }. Suppose that there exists a domain D and K > 0 such that kf(t, x) − f(t, y)k ≤ Kkx − yk for all x, y ∈ D ⊂ Rn and t > 0. Suppose there exists a compact set W and r > 0 such that Wr ⊂ D. Furthermore suppose that it has been proven that for x0 ∈ W , any solution to
x˙(t) = f(t, x), x(0) = x0
must lie entirely in W . Then, for x0 ∈ W , there exists a unique solution x with x(0) = x0 such that x lies entirely in W .
M. Peet Lecture 02: 22 / 56 Illustration of Extended Picard Iteration
Picard iteration can also be used with the extension theorem • Final time of previous Picard iterate is used to seed next Picard iterate. Definition 15. Suppose that the solution map φ exists on t ∈ [0, ∞] and kφ(t, x)k ≤ Kkxk for any x ∈ Br. Suppose that f has Lipschitz factor L on B4Kr and is bounded on 2Kr 1 B4Kr with bound Q. Given T < min{ Q , L }, let z = 0 and define
k k G0 (t, x) := (P z)(t, x)
and for i > 0, define the functions Gi recursively as
k k k Gi+1(t, x) := (P z)(t, Gi (T, x)).
k Define the concatenation of the Gi as
k k G (t, x) := Gi (t − iT, x) ∀ t ∈ [iT, iT + T ] and i = 1, ··· , ∞.
M. Peet Lecture 02: 23 / 56 Illustration of Extended Picard Iteration
We take the previous approximation to the solution map and extend it. xHtL 1.0
0.9
0.8
0.7
0.6
0.5
t 0.2 0.4 0.6 0.8 1.0
k Figure : The Solution map φ and the functions Gi for k = 1, 2, 3, 4, 5 and i = 1, 2, 3 for the system x˙(t) = −x(t)3. The interval of convergence of the Picard Iteration is 1 T = 3 .
M. Peet Lecture 02: 24 / 56 Stability Definitions
Whenever you are trying to prove stability, Please define your notion of stability!
We denote the set of bounded continuous functions by ¯ C := {x ∈ C : kx(t)k ≤ r, r ≥ 0} with norm kxk = suptkx(t)k. Definition 16. The system is locally Lyapunov stable on D where D contains an open neighborhood of the origin if it defines a unique map Φ: D → C¯ which is continuous at the origin.
The system is locally Lyapunov stable on D if for any > 0, there exists a δ() such that for kx(0)k ≤ δ(), x(0) ⊂ D we have kx(t)k ≤ for all t ≥ 0
M. Peet Lecture 02: 25 / 56 Stability Definitions
Definition 17. The system is globally Lyapunov stable if it defines a unique map Φ: Rn → C¯ which is continuous at the origin. We define the subspace of bounded continuous functions which tend to the ¯ origin by G := {x ∈ C : limt→∞ x(t) = 0} with norm kxk = suptkx(t)k. Definition 18. The system is locally asymptotically stable on D where D contains an open neighborhood of the origin if it defines a map Φ: D → G which is continuous at the origin.
M. Peet Lecture 02: 26 / 56 Stability Definitions
Definition 19. The system is globally asymptotically stable if it defines a map Φ: Rn → G which is continuous at the origin.
Definition 20. The system is locally exponentially stable on D if it defines a map Φ: D → G where k(Φx)(t)k ≤ Ke−γtkxk for some positive constants K, γ > 0 and any x ∈ D.
Definition 21. The system is globally exponentially stable if it defines a map Φ: Rn → G where k(Φx)(t)k ≤ Ke−γtkxk
for some positive constants K, γ > 0 and any x ∈ Rn.
M. Peet Lecture 02: 27 / 56 Lyapunov Theorem
x˙ = f(x), f(0) = 0
Theorem 22. Let V : D → R be a continuously differentiable function such that
V (0) = 0 V (x) > 0 for x ∈ D, x 6= 0 ∇V (x)T f(x) ≤ 0 for x ∈ D.
• Then x˙ = f(x) is well-posed and locally Lyapunov stable on the largest sublevel set of V contained in D. • Furthermore, if ∇V (x) < 0 for x ∈ D, x 6= 0, then x˙ = f(x) is locally asymptotically stable on the largest sublevel set of V contained in D.
M. Peet Lecture 02: 28 / 56 Lyapunov Theorem
Sublevel Set: For a given Lyapunov function V and positive constant γ, we denote the set Vγ = {x : V (x) ≤ γ}. Proof. Existence: Denote the largest bounded sublevel set of V contained in the T interior of D by Vγ∗ . Because V˙ (x(t)) = ∇V (x(t)) f(x(t)) ≤ 0 is continuous, if x(0) ∈ Vγ∗ , then x(t) ∈ Vγ∗ for all t ≥ 0. Therefore since f is continuous and Vγ∗ is compact, by the extension theorem, there is a unique solution for any initial condition x(0) ∈ Vγ∗ . 0 0 Lyapunov Stability: Given any > 0, choose e < e with B() ⊂ Vγ∗ , choose
γi such that Vγi ⊂ B(). Now, choose δ > 0 such that B(δ) ⊂ Vγi . Then
B(δ) ⊂ Vγi ⊂ B() and hence if x(0) ∈ B(δ), we have 0 x(0) ∈ Vγi ⊂ B() ⊂ B( ). Asymptotic Stability:
• V monotone decreasing implies limt→ V (x(t)) = 0. • V (x) = 0 implies x = 0. • Proof omitted.
M. Peet Lecture 02: 29 / 56 Lyapunov Theorem Exponential Stability
Theorem 23. Suppose there exists a continuously differentiable function V and constants c1, c2, c3 > and radius r > 0 such that the following holds for all x ∈ B(r).
p p c1kxk ≤ V (x) ≤ c2kxk T p ∇V (x) f(x) ≤ −c3kxk
Then x˙ = f(x) is exponentially stable on any ball contained in the largest sublevel set contained in B(r).
Exponential Stability allows a quantitative prediction of system behavior.
M. Peet Lecture 02: 30 / 56 The Gronwall-Bellman Inequality Exponential Stability
Lemma 24 (Gronwall-Bellman). Let λ be continuous and µ be continuous and nonnegative. Let y be continuous and satisfy for t ≤ b,
t y(t) ≤ λ(t) + µ(s)y(s)ds. Za Then t t y(t) ≤ λ(t) + λ(s)µ(s) exp µ(τ)dτ ds Za Zs If λ and µ are constants, then
y(t) ≤ λeµt.
For λ(t) = y(0), the condition can be differentiated to obtain y˙(t) ≤ µ(t)y(t).
M. Peet Lecture 02: 31 / 56 Lyapunov Theorem Exponential Stability
Proof. We begin by noting that we already satisfy the conditions for existence, uniqueness and asymptotic stability and that x(t) ∈ B(r). For simplicity, we take p = 2. Now, observe that
2 c3 V˙ (x(t)) ≤ −c3kx(t)k ≤ − V (x(t)) c2
Which implies by the Gronwall-Bellman inequality (µ = −c3 , λ = V (x(0))) c2 that − c3 t V (x(t)) ≤ V (x(0))e c2 . Hence
c c 2 1 1 − 3 t c2 − 3 t 2 kx(t)k ≤ V (x(t)) ≤ e c2 V (x(0)) ≤ e c2 kx(0)k . c1 c1 c1
M. Peet Lecture 02: 32 / 56 Lyapunov Theorem Invariance Sometimes, we want to prove convergence to a set. Recall
Vγ = {x , V (x) ≤ γ} Definition 25. A set, X, is Positively Invariant if x(0) ∈ X implies x(t) ∈ X for all t ≥ 0. Theorem 26. Suppose that there exists some continuously differentiable function V such that V (x) > 0 for x ∈ D, x 6= 0 ∇V (x)T f(x) ≤ 0 for x ∈ D.
for all x ∈ D. Then for any γ such that the level set X = {x : V (x) = γ} ⊂ D, we have that Vγ is positively invariant. Furthermore, if ∇V (x)T f(x) ≤ 0 for x ∈ D, then for any δ such that X ⊂ Vδ ⊂ D, we have that any trajectory starting in Vδ will approach the sublevel set Vγ .
M. Peet Lecture 02: 33 / 56 Converse Lyapunov Theory
In fact, stable systems always have Lyapunov functions. Suppose that there exists a continuously differentiable function function, called the solution map, g(x, s) such that
∂ g(x, s) = f(g(x, s)) and g(x, 0) = x ∂s is satisfied. Converse Form 1: δ V (x) = g(s, x)T g(s, x)ds Z0
M. Peet Lecture 02: 34 / 56 Converse Lyapunov Theory
Converse Form 1:
δ V (x) = g(s, x)T g(s, x)ds Z0 For a linear system, g(s, x) = eAsx. • This recalls the proof of feasibility of the Lyapunov inequality
AT P + P A < 0
• The solution was given by
∞ ∞ T xT P x = xT eA seAsxds = g(s, x)T g(s, x)ds Z0 Z0
M. Peet Lecture 02: 35 / 56 Converse Lyapunov Theory
Theorem 27. Suppose that there exist K and λ such that g satisfies
kg(x, t)k ≤ Kkg(x, 0)ke−λt
Then there exists a function V and constants c1, c2, and c3 such that V satisfies
2 2 c1kxk ≤ V (x) ≤ c2kxk T 2 ∇V (x) f(x) ≤ −c3kxk
M. Peet Lecture 02: 36 / 56 Converse Lyapunov Theory
Proof. There are 3 parts to the proof, of which 2 are relatively minor. But part 3 is tricky. The main hurdle is to choose δ > 0 sufficiently large 2 Part 1: Show that V (x) ≤ c2kxk . Then
δ V (x) = kg(s, x)k2ds 0 Z δ ≤ K2kg(x, 0)k2 e−2λsds Z0 K2 = kxk2 (1 − e−2λδ) = c kxk2 2λ 2
K2 −2λδ where c2 = 2λ (1 − e ). This part holds for any δ > 0.
M. Peet Lecture 02: 37 / 56 Converse Lyapunov Theory
Proof. 2 Part 2: Show that V (x) ≥ c1kxk . Lipschitz continuity of f implies kf(x)k ≤ Lkxk. By the fundamental identity
t t kx(t)k ≤ kx(0)k + kf(x(s))kds ≤ kx(0)k + Lkx(s)kds Z0 Z0 Hence by the Gronwall-Bellman inequality kx(0)ke−Lt ≤ kx(t)k ≤ kx(0)keLt. Thus we have that kg(x, t)k2 ≥ kxk2e−Lt. This implies
δ δ V (x) = kg(s, x)k2ds ≥ kxk2 e−2Lsds Z0 Z0 1 = kxk2 (1 − e−2Lδ) = c kxk2 2L 1
1 −2Lδ where c1 = 2L (1 − e ). This part also holds for any δ > 0.
M. Peet Lecture 02: 38 / 56 Converse Lyapunov Theory Proof, Part 3. T 2 Part 3: Show that ∇V (x) f(x) ≤ −c3kxk . This requires differentiating the solution map with respect to initial conditions. We first prove the identity
gt(t, x) = −gx(t, x)f(x) We start with a modified version of the fundamental identity t 0 g(t, x) = g(0, x) + f(g(s, x))ds = g(0, x) + f(g(s + t, x))ds Z0 Z−t By the Leibnitz rule for the differentiation of integrals, we find 0 T gt(t, x) = f(g(0, x)) + ∇f(g(s + t, x)) gs(s + t, x)ds Z−t t T = f(x) + ∇f(g(s, x)) gs(s, x)ds Z0 t Also, we have T gx(t, x) = I + ∇f(g(s, x)) gx(s, x)ds Z0
M. Peet Lecture 02: 39 / 56 Converse Lyapunov Theory
Proof, Part 3. Now
gt(t, x) − gx(t, x)f(x) t t T T = x + ∇f(g(s, x)) gs(s, x)ds + f(x) + ∇f(g(s, x)) gx(s, x)f(x)ds 0 0 t Z Z T = ∇f(g(s, x)) (gs(s, x) − gx(s, x)f(x)) ds Z0 By, e.g., Gronwall-Bellman, this implies
gt(t, x) − gx(t, x)f(x) = 0.
We conclude that gt(t, x) = gx(t, x)f(x) Which is interesting.
M. Peet Lecture 02: 40 / 56 Converse Lyapunov Theory Proof, Part 3. With this identity in hand, we proceed: T δ T T ∇V (x) f(x) = ∇x g(s, x) g(s, x)ds f(x) Z0 ! δ T = 2 g(s, x) gx(s, x)f(x)ds Z0 δ δ d = 2 g(s, x)T g (s, x)ds = kg(s, x)k2ds s ds Z0 Z0 = kg(δ, x)k2 − kg(0, x)k2 ≤ K2kxk2e−2λδ − kxk2 = − 1 − K2e−2λδ kxk2