Theory of Ordinary Differential Equations
Existence, Uniqueness and Stability
Jishan Hu and Wei-Ping Li
Department of Mathematics The Hong Kong University of Science and Technology ii
Copyright c 2004 by Jishan Hu and Wei-Ping Li Contents
1 Existence and Uniqueness 1 1.1 Some Basics 1 1.2 Uniqueness Theorem 6 1.3 Continuity 8 1.4 Existence Theorem 13 1.5 Local Existence Theorem and The Peano Theorem 19 1.5.1 Local Existence Theorem 19 1.5.2 The Peano Theorem 21 1.6 Linear Systems 23 1.7 Continuation of Solutions 27 1.8 Miscellaneous Problems 29
2 Plane Autonomous Systems 33 2.1 Plane Autonomous Systems 33 2.2 Linear Autonomous Systems 39 2.3 Complete Classification for Linear Autonomous Systems 44 2.4 Liapunov Direct Method 52 2.5 A Test For Instability 61 2.6 Nonlinear Oscillations 65 2.6.1 Undamped Oscillations 65 iii iv CONTENTS
2.6.2 Damped Oscillations 66 2.7 Miscellaneous Problems 67 1 Existence and Uniqueness
1.1 SOME BASICS
A normal system of first order ordinary differential equations (ODEs) is
dx1 = X1(x1, . . . , xn; t), dt . (1.1) . dx n = X (x , . . . , x ; t). dt n 1 n
Many varieties of ODEs can be reduced to this form. For example, consider an n-th order ODE
dnx dx dn−1x = F t, x, ,..., . dtn dt dtn−1
dx dn−1x Let x = x, x = , . . . , x = . Then the ODE can be changed to 1 2 dt n dtn−1
dx1 = x2, dt . . dx n−1 = x , dt n dx n = F (t, x , . . . , x ). dt 1 n 1 2 1 EXISTENCE AND UNIQUENESS
Let us review some notations and facts on vectors and vector-valued functions. The normal sys- tem (1.1) can be written as its vector form:
dx = X(x, t). dt
p 2 2 For a vector x = (x1, . . . , xn), define kxk = x1 + ··· + xn. The inner product is defined to be x · y = x1y1 + ··· + xnyn where y = (y1, . . . , yn). We have the triangle inequality
kx + yk ≤ kxk + kyk, and the Schwarz inequality
|x · y| ≤ kxk · kyk.
For a vector valued function x(t) = (x1(t), . . . , xn(t)), we have
0 0 0 x (t) = (x1(t), . . . , xn(t)) , and
Z b Z b Z b ! x(t)dt = x1(t)dt, . . . , xn(t)dt . a a a
We have the following useful inequality
Z b Z b
x(t)dt ≤ kx(t)k dt. a a A vector field X(x) = X1(x1, . . . , xn),...,Xn(x1, . . . , xn) is aid to be continuous if each Xi is a continuous function of x1, . . . , xn. Since only a few simple types of differential equations can be solved explicitly in terms of known elementary function, in this chapter, we are going to explore the conditions on the function X such that the differential system has a solution. We also study whether the solution is unique, subject some additional initial conditions.
Example 1.1.1 Show that the differential equation dy 1 = − y−1 dt 2 does not have solution satisfying y(0) = 0 for t > 0.
Solution Acturally, the general solution of this differential equation is
y2 = −t + C,
where C is an arbitrary constant. The initial condition implies C = 0. Thus, we have y2 = −t, that shows there exists no solution for t > 0. 2 1.1 SOME BASICS 3
Example 1.1.2 Show that the differential equation x0 = x2/3 has infinitely many solutions satisfying x(0) = 0 on every interval [0, b]. Solution Define 0, if 0 ≤ t < c; xc(t) = (t − c)3 , if c ≤ t ≤ b. 27
It is easy to check for any c, the function xc satisfies the differential equation and xc(0) = 0. 2 Definition 1.1.1 A vector-valued function X(x, t) is said to satisfy a Lipschitz condition in a region R in (x, t)-space if, for some constant L (called the Lipschitz constant), we have kX(x, t) − X(y, t)k ≤ Lkx − yk, (1.2) whenever (x, t) ∈ R and (y, t) ∈ R.1 ∂X Lemma 1.1.2 If X(x, t) has continuous partial derivatives i on a bounded closed convex domain ∂xj R, then it satisfies a Lipschitz condition in R.2 Proof Denote M = sup |∂Xi/∂xj | . x∈R 1≤i,j≤n Since X has continuous partial derivatives on the bounded closed region R, we know that M is finite. For x, y ∈ R and 0 ≤ s ≤ 1, (1 − s)x + sy is a line segment connecting x and y. Thus it is in the domain R since R is convex. For each component Xi, we have, for fixed x, y, t, regarding Xi((1 − s)x + sy) as a function of the variable s, n d X ∂Xi Xi((1 − s)x + sy, t) = ((1 − s)x + sy, t)(yk − xk). ds ∂xk k=1 Using the mean value theorem, we have
n X ∂Xi Xi(y, t) − Xi(x, t) = ((1 − σi)x + σiy, t)(yk − xk), ∂xk k=1
1It is worth to point out the following: (i) The left hand side of the inequality (1.2) is evaluated at (x, t) and (y, t), with the same t for these two points. (ii) The constant L is independent of x, y and t. However, it depends on R. In other words, for a given function X(x, t), its Lipschitz constant may change if the domain R is different. In fact, for the same function X(x, t), it can be a Lipschitz function in some regions, but not a Lipschitz function in some other regions.
2Here are some explanation of some concepts and terminologies in the Lemma. (i) A bounded closed domain is also called compact. It has the following property that any continuous function on a compact set has an absolute maximum and absolute minimum. ∂Xi (ii) That X(x, t) = X1(x1, . . . , xn, t),...,Xn(x1, . . . , xn, t) has continuous partial derivatives means that ∂xj ∂X and i are continuous for all i, j. Sometimes we use X(x, t) ∈ C1(D) for this. ∂t (iii) A convex domain D means that for any two points x and y in D, (1 − t)x + ty ∈ D for 0 ≤ t ≤ 1. 4 1 EXISTENCE AND UNIQUENESS
for some σi between 0 and 1. The Schwarz inequality gives
n n 2!1/2 n !1/2 X ∂Xi X ∂Xi X 2 ((1 − σi)x + σiy, t)(yk − xk) ≤ ((1 − σi)x + σiy, t) |yk − xk| ∂xk ∂xk k=1 k=1 k=1 n !1/2 X 2 √ ≤ M · ky − xk = nM · ky − xk. k=1 Thus, n !1/2 X 2 kX(y, t) − X(x, t)k = |Xi(y, t) − Xi(x, t)| i=1 n !1/2 X √ 2 ≤ nM · ky − xk = nM · ky − xk. i=1 The Lipschitz condition follows, with the Lipschitz constant nM. 2
Example 1.1.3 Determine whether the function
x2 + 1 X(x, t) = · t x satisfies a Lipschitz condition in the domains: (1) R1 = [1, 2] × [0, 1]; (2) R2 = (1, 2) × [0, 1]; (3) R3 = [1, 2] × [0, +∞); (4) R4 = [1, +∞) × [0,T ]; (5) R5 = (0, 1) × [0, 1].
Solution (1) Since the function X(x, t) is continuously differentiable in the bounded closed convex domain R = [1, 2] × [0, 1], by Lemma 1.1.2, we know that the function satisfies a Lipschitz conditio in this region.
(2) Since the function X(x, t) satisfis a Lipschitz condition in R1, and R2 ⊂ R1, we know that the function satisfis the same Lipschitz inequality in R2. Hence the function X(x, t) satisfis a Lipschitz condition in R2.
(3) Since R3 = [1, 2] × [0, +∞) is not a bounded region, we cannot apply Lemma 1.1.2 in this case. Since
|X(x, t) − X(y, t)| xy − 1 = · |t| > |t| → ∞, |x − y| xy as t → +∞, there exists no contant L, independent of x, y and t, such that
|X(x, t) − X(y, t)| ≤ L |x − y| .
Hence, the function X is not a Lipschitz function in R3.
(4) Again, R4 = [1, +∞) × [0,T ] is not bounded and Lemma 1.1.2 does not apply in this case. For (x, t), (y, t) ∈ R4,
xy − 1 |X(x, t) − X(y, t)| = · |t| · |x − y| xy 1 ≤ 1 − · T · |x − y| xy ≤ T |x − y|.
Thus, the function X is a Lipschitz function in R4, with L = T . 1.1 SOME BASICS 5
(5) Since R5 = (0, 1) × [0, 1] is not a closed region, Lemma 1.1.2 does not apply in this case. In fact, the function X is continuously differentiable in R, with
∂X(x, t) x2 − 1 = t. ∂x x2 ∂X(x, t) But the derivative is not bounded when t 6= 0, since → ∞ as x → 0+ when t 6= 0. ∂x Actually, we can show that the function X does not satisfy any Lipschitz condition in R5. To see that, 1 2 we take x = , y = and t = 1, and have n n n2 |X(x, t) − X(y, t)| = − 1, 2 n2 for n ≥ 2. Obviously, there exists no constant L such that − 1 ≤ L for all n. Therefore, the function X 2 does not satisfy any Lipschitz conditions in the domain R5. 2 d3x Example 1.1.4 Reduce the ODE + x2 = 1 to an equivalent first-order system and determine dt3 in which domain or domains the resulting system satisfies a Lipschitz condition.
0 00 Solution Let x1 = x, x2 = x , x3 = x , we get the equivalent first order system
dx1 = x2, dt dx 2 = x , dt 3 dx3 = 1 − x2. dt 1
2 Denote x = (x1, x2, x3), y = (y1, y2, y3) and X(x, t) = (x2, x3, 1 − x1). Then, X(x, t) − X(y, t) = (x2 − 2 2 y2, x3 − y3, −x1 + y1 ), which implies
p 2 2 2 2 kX(x, t) − X(y, t)k = (x2 − y2) + (x3 − y3) + (x1 − y1) (x1 + y1) .
We can see that whenever x1 and y1 are bounded, to be precise, x1 + y1 is bounded, the Lipschitz condition holds. In fact, consider the domain n o
R = (x1, x2, x3, t) |x1| ≤ M for some constant M. Take L = pmax{4M 2, 1}, then whenever (x, t), (y, t) ∈ R, we have
p 2 2 2 2 kX(x, t) − X(y, t)k ≤ (x2 − y2) + (x3 − y3) + 4M (x1 − y1)
p 2 2 2 2 2 2 ≤ L (x2 − y2) + L (x3 − y3) + L (x1 − y1) = L kx − yk .
Therefore X(x, t) satisfies the Lipschitz condition with the Lipschitz constant L = pmax{4M 2, 1} in the domain R. 2
Exercise 1.1 ∂X 1. Show that the function X(x, t) = |x| satisfies a Lipschitz condition everywhere, but does not exist at x = 0. ∂x 6 1 EXISTENCE AND UNIQUENESS
1 2. For f(x, t) = x 3 t, does it satisfy a Lipschitz condition over the interval [1, 2]×[0,T ]? Does it satisfy a Lipschitz condition over the interval [−1, 1] × [0,T ]? Here T is a constant.
3. Does the function f(x, t) = x3 satisfy a Lipschitz condition over [0, +∞) × [0, +∞)? xt 4. Does the function f(x, t) = satisfy a Lipschitz condition over [0, +∞)×[0,T ]? Here T is a constant. x2 + t2 + 1
1.2 UNIQUENESS THEOREM
In this section, we prove the following uniqueness theorem. Theorem 1.2.1 (Uniqueness Theorem) If the vector field X(x, t) satisfies a Lipschitz condition in a domain R, then there is at most one solution x(t) of the differential system dx = X(x, t), (1.3) dt that satisfies a given initial condition x(a) = c in R. Proof Assume that X(x, t) satisfies the Lipschitz condition
kX(x, t) − X(y, t)k ≤ L kx − yk , for any (x, t), (y, t) ∈ R. Let x(t) = (x1(t), . . . , xn(t)) and y(t) = (y1(t), . . . , yn(t)) be two solutions satisfying the same initial condition x(a) = c = y(a). n 2 P 2 Denote σ(t) = kx(t) − y(t)k = [xk(t) − yk(t)] ≥ 0. Then, k=1
n 0 X 0 0 σ (t) = 2 xk(t) − yk(t) · [xk(t) − yk(t)] k=1 n X = 2 [Xk(x(t), t) − Xk(y(t), t)] · [xk(t) − yk(t)] k=1 = 2 [X(x(t), t) − X(y(t), t)] · [x(t) − y(t)] .
By the Schwarz inequality, we have
0 0 σ (t) ≤ σ (t) = 2 |(X(x, t) − X(y, t)) · (x − y)| ≤ 2 kX(x, t) − X(y, t)k · kx − yk ≤ 2L kx − yk2 = 2Lσ(t).
The last inequality implies d σ(t)e−2Lt = σ0(t) − 2Lσ(t) e−2Lt ≤ 0. dt Hence σ(t)e−2Lt is a decreasing function. Therefore for t > a, σ(t)e−2Lt ≤ σ(a)e−2La = 0. Since σ(t) ≥ 0, we have σ(t) = 0 for t ≥ a, i.e., x(t) = y(t) for t ≥ a. To argue for t < a, as above we again have
−σ0(t) ≤ |σ0(t)| ≤ 2Lσ(t), 1.2 UNIQUENESS THEOREM 7 which implies d σ(t)e+2Lt = σ0(t) + 2Lσ(t) e+2Lt ≥ 0. dt So σ(t)e+2Lt is an increasing function. Therefore for t < a, σ(t)e+2Lt ≤ σ(a)e+2La = 0. Again we have σ(t) = 0 for t ≤ a, i.e., x(t) = y(t) for t ≤ a. 2
Example 1.2.1 Find a region where the differential equation x0 = x + 3x1/3 has a unique solution, i.e., find (x0, t0) such that the solution x(t) of the differential equation with x(t0) = x0 is unique in a neighborhood of (x0, t0).
Solution We will show the following
(1) the differential equation has a unique solution with x(t0) = x0, x0 6= 0;
(2) there are more than one solution satisfying x(t0) = 0.
(1) For any given (x0, t0), with x0 6= 0, we choose a small δ > 0 such that 0 6∈ [x0 − δ, x0 + δ]. By Lemma 1.1.2, we know that the function X(x, t) = x + 3x1/3 satisfies a Lipschitz condition in the region n o
R = (x, t) |x − x0| ≤ δ, |t − t0| ≤ T = [x0 − δ, x0 + δ] × [t0 − T, t0 + T ], where T > 0 is any fixed constant.3 By Theorem 1.2.1, we conclude that the differential equation has a unique solution with x(t0) = x0, x0 6= 0. (2) It is easy to see that x(t) ≡ 0 is one solution of the differential equation with x(t0) = 0. We only need to show that there exists another solution which also satisfies x(t0) = 0. Consider the improper integral Z x du 1/3 . 0 u + 3u For any c > 0, we know that Z c Z c du du 1 2/3 0 < 1/3 < 1/3 = c . 0 u + 3u 0 3u 2 Hence the improper integral converges for c > 0. This allows us to define an implicit function x(t) by
Z x du 1/3 = t − t0. 0 u + 3u
We should have x(t0) = 0, since the last equation becomes an identity when setting t = t0. Obviously, this function x(t) 6≡ 0, otherwise we will have t ≡ t0, a contradiction. This function x(t) certainly satisfies the differential equation, which can be seen easily by differentiating both sides of the last equation. 2
3 We can make the following argument. Let (x, t), (y, t) ∈ R = [x0 − δ, x0 + δ] × [t0 − T, t0 + T ]. Then, by the mean value theorem, we have
∂X −2/3 |X(x, t) − X(y, t)| = (ξ, t) · |x − y| = 1 − ξ · |x − y| , ∂x
∂X −2/3 where ξ is between x and y. This implies ξ ∈ [x0 − δ, x0 + δ]. The function (x, t) = 1 + x is continuous in x ∂x in the compact set [x0 − δ, x0 + δ], since 0 6∈ [x0 − δ, x0 + δ]. This indicates that there is a number L (might depend on x0 and δ) such that
−2/3 1 − ξ ≤ L. Thus, the function X(x, t) satisfies a Lipschitz condition in [x0 − δ, x0 + δ] × R. 8 1 EXISTENCE AND UNIQUENESS
Exercise 1.2
1. Show that for 0 < α < 1, the initial value problem dx = xα, dt x(0) = 0, has at least two solutions.
2. Show that the initial value problem dx = sin t · | sin x|, dt x(0) = c, has a unique solution.
3. Show that the initial value problem: dx x2 = 1 + , dt 1 + t2 x(0) = 1, has at most one solution.
4. Show that the initial value problem dx = t sin x, dt x(0) = 1, has at most one solution.
5. Show that the initial value problem du = tu + sin u, dt u(0) = 1, has a unique solution.
6. State and prove a uniqueness theorem for the differential equation y00 = F (t, y, y0).
1.3 CONTINUITY
Theorem 1.3.1 (Continuity Theorem) Let x(t) and y(t) be any two solutions of the differential equation (1.3) in T1 ≤ t ≤ T2, where X(x, t) is continuous and satisfies the Lipschitz condition (1.2) in some region R that contains the region where x(t) and y(t) are defined. Then
kx(t) − y(t)k ≤ eL|t−a| kx(a) − y(a)k ,
for any a, t ∈ [T1,T2].
Proof Let us first assume that t ≥ a. Then, for the function σ(t) = kx(t) − y(t)k2, as in the proof of Uniqueness Theorem 1.2.1, we have σ0(t) ≤ 2Lσ(t), which implies d σ(t)e−2Lt ≤ 0. dt 1.3 CONTINUITY 9
Integrating the last inequality from a to t gives
σ(t)e−2Lt ≤ σ(a)e−2La, which yields the desired inequality. In the case t ≤ a, again as in the proof of Uniqueness Theorem 1.2.1, we obtain
−σ0(t) ≤ 2Lσ(t), which implies d σ(t)e+2Lt ≥ 0. dt Now we integrate the last inequality from t to a to have
σ(t)e2Lt ≤ σ(a)e2La.
This implies the desired inequality in the case t ≤ a. 2
Corollary 1.3.2 Let x(t) be the solution of the differential equation (1.3) satisfying the initial condition x(a, c) = c. Let the hypotheses of Continuity Theorem 1.3.3 be satisfied, and let the function x(t, c) be defined for kc − c0k ≤ K and |t − a| ≤ T . Then
(1) x(t, c) is a continuous function of both variables;
(2) if c → c0, then x(t, c) → x(t, c0) uniformly for |t − a| ≤ T .
Proof For (a), it is obvious that x(t, c) is continuous in t since x(t, c) is a solution of the differential equation (1.3), which is differentiable. To see that x(t, c) is continuous in c, we take c1 and c2, with 1 0 2 0 1 2 c − c ≤ K and c − c ≤ K. Since x(t, c ) and x(t, c ) are solutions of (1.3), by Continuity Theo- rem 1.3.3, we get
1 2 L|t−a| 1 2 L|t−a| 1 2 x(t, c ) − x(t, c ) ≤ e x(a, c ) − x(a, c ) = e c − c .
1 2 Let t be fixed. Hence, for any given > 0, take δ = , whenever c − c < δ, we have eL|t−a|
1 2 L|t−a| x(t, c ) − x(t, c ) < e δ = .
Therefore, x(t, c) is continuous in c. For (b), in above, we can take 1 δ = min · . t∈[a−T,a+T ] eL|t−a| 1 Since min > 0 is a finite number, hence, δ > 0 is independent of t. Therefore, x(t, c) is t∈[a−T,a+T ] eL|t−a| continuous in c uniformly for |t − a| ≤ T . 2
Theorem 1.3.3 (Strong Continuity Theorem) Assume that
(1) x(t) and y(t) satisfy the differential equations
dx dy = X(x, t) and = Y(y, t), dt dt 10 1 EXISTENCE AND UNIQUENESS
respectively, on T1 ≤ t ≤ T2. (2) The functions X and Y be defined and continuous in a common domain R, and satisfy
kX(z, t) − Y(z, t)k ≤ ,
for any (z, t) ∈ R, with T1 ≤ t ≤ T2. (3) X(x, t) satisfies the Lipschitz condition (1.2) in R.4 Then kx(t) − y(t)k ≤ kx(a) − y(a)k eL|t−a| + eL|t−a| − 1 , L for any a, t ∈ [T1,T2].
2 Proof Assume that a, t ∈ [T1,T2]. Let us first consider the case t ≥ a. Define σ(t) = kx(t) − y(t)k = n P 2 [xk(t) − yk(t)] ≥ 0. We have k=1
σ0(t) = 2 x0(t) − y0(t) · [x(t) − y(t)] = 2 [X(x(t), t) − Y(y(t), t)] · [x(t) − y(t)] = 2 [X(x(t), t) − X(y(t), t)] · [x(t) − y(t)] + 2 [X(y(t), t) − Y(y(t), t)] · [x(t) − y(t)] .
The Schwarz inequality implies
0 0 σ (t) ≤ σ (t) ≤ 2 kX(x(t), t) − X(y(t), t)k · kx(t) − y(t)k + 2 kX(y(t), t) − Y(y(t), t)k · kx(t) − y(t)k ≤ 2L kx(t) − y(t)k2 + 2 kx(t) − y(t)k = 2Lσ(t) + 2pσ(t).
(1) Let us assume that σ(a) > 0. Consider the initial value problem
du √ = 2Lu + 2 u, t ≥ a, dt u(a) = σ(a).
Formally, we can introduce the substitution v(t) = pu(t). This gives the equivalent differential equation
dv 2v = 2Lv2 + 2v. dt v(t) = 0 for all t is not a solution since v(a) = pu(a) 6= 0. Thus we can divide both sides of the equation by v(t) to obtain a linear differential equation for v. This leads to the following initial value problem dv − Lv = , t ≥ a, dt v(a) = pu(a).
4The function Y(y, t) may not satisfy a Lipschitz condition. 1.3 CONTINUITY 11
The solution of the initial value problem is pu(t) = v(t) = pu(a)eL(t−a) + eL(t−a) − 1 . L Claim: σ(t) ≤ u(t). Suppose not, there exists t1 ≥ a such that σ(t1) > u(t1). Let t0 = sup{t|a ≤ t < t1, σ(t) ≤ u(t)}. Take a sequence tn ≤ t0, tn → t0 such that σ(tn) ≤ u(tn). By the continuity of σ(t) and u(t), we must 0 have σ(t0) = lim σ(tn) ≤ lim u(tn) = u(t0). Take any sequence tn → t0 with tn > t0, we must have n→∞ n→∞ 0 0 σ(tn) < u(tn). Taking limits, we have σ(t0) ≥ u(t0). Therefore we σ(t0) = u(t0). Since u(t0) > 0, we have, for t0 ≤ t ≤ t1, [σ(t) − u(t)]0 ≤ 2L[σ(t) − u(t)] + 2[pσ(t) − pu(t)] σ(t) − u(t) = 2L[σ(t) − u(t)] + 2 · pσ(t) + pu(t) σ(t) − u(t) ≤ 2L[σ(t) − u(t)] + 2 · p , u(t0) which implies h n p oi0 [σ(t) − u(t)] · exp − 2L + 2/ u(t0) t ≤ 0.
Integrating the last inequality from t0 to t, we have n p o n p o [σ(t) − u(t)] · exp − 2L + 2/ u(t0) t ≤ [σ(t0) − u(t0)] · exp − 2L + 2/ u(a) a = 0.
Therefore σ(t) ≤ u(t) for t0 ≤ t ≤ t1 which is a contradiction to σ(t1) > u(t1). The Claim is proved. Thus, σ(t) ≤ u(t), which implies
kx(t) − y(t)k = pσ(t) ≤ pu(t) = pu(a)eL(t−a) + eL(t−a) − 1 L = pσ(a)eL(t−a) + eL(t−a) − 1 L = kx(a) − y(a)k eL(t−a) + eL(t−a) − 1 . L (2) For σ(a) = 0, we have to modify the discussion above. For each positive integer n, we consider the following initial value problem du √ = 2Lu + 2 u, t ≥ a, dt u(a) = 1/n. The discussion above can be applied to this problem, and we have the solution h i2 u (t) = n−1/2eL|t−a| + eL|t−a| − 1 . n L
If we can show that σ(t) ≤ un(t) for t ≥ a, then, after taking n → ∞, we obtain the desired inequality in the case σ(a) = 0. The last inequality can be proved by contradiction. In fact, if
σ(t1) > un(t1), 12 1 EXISTENCE AND UNIQUENESS
for some t1 > a, then there exists t0 to be the largest t in the interval a < t ≤ t1 such that σ(t0) ≤ un(t0). Obviously, σ(t0) = un(t0) > 0 and σ(t) > un(t) for t0 < t ≤ t1. But this is impossible according to the discussion in the case σ(a) > 0 by replacing a by t0. We can use the inequality −σ0(t) ≤ |σ0(t)| ≤ 2Lσ(t) + 2pσ(t) and a similar argument to prove the following inequality kx(t) − y(t)k ≤ kx(a) − y(a)k eL(a−t) + eL(a−t) − 1 . L 2
Exercise 1.3
1. Assume that ϕ(t), ψ(t), w(t) are continuous on [a, b], with w(t) > 0. If the inequality Z t ϕ(t) ≤ ψ(t) + w(s)ϕ(s)ds a holds on [a, b], then Z t Z t ϕ(t) ≤ ψ(t) + w(s)ψ(s) exp w(u)du ds. a s Further, if ψ(t) is non-negative and monotonically increasing, then Z t ϕ(t) ≤ ψ(t) exp w(s)ds . a
2. Prove Strong Continuity Theorem in detail for t ≤ a.
3. Let x(t) and y(t) be the solutions of the initial value problems x00 + (1 + t2)x = 0, 0 < t < 1, x(0) = 1, x0(0) = 0 and y00 + (1 + t2)y = δ, 0 < t < 1, δ = const, y(0) = 1, y0(0) = 0, respectively. Estimate max |x(t) − y(t)|. t∈[0,1]
4. Let X(x, t, s) be continuous for kx − ck ≤ K, |t − a| ≤ T , and |s − s0| ≤ S, and let it satisfy kX(x, t, s) − X(y, t, s)k ≤ L kx − yk . Show that the solution x(t, s) of dx = X(x, t, s) dt satisfying x(a, s) = c is a continuous function of s. 1.4 EXISTENCE THEOREM 13
1.4 EXISTENCE THEOREM
In this section, we study existence of the differential equation (1.3). The idea is to establish an equivalent integral equation for any given initial value problem. Then we show that the iteration of the integral operator converges to a solution.
Theorem 1.4.1 Let X(x, t) be a continuous function. Then a function x(t) is a solution of the initial value problem dx = X(x, t), dt (1.4) x(a) = c if and only if it is a solution of the integral equation
Z t x(t) = c + X(x(s), s)ds. (1.5) a Proof Let us assume that x(t) is a solution of the initial value problem (1.4). The Fundamental Theorem of Calculus implies that Z t 0 xk(t) = xk(a) + xk(s)ds. a Using (1.4), we have the integral equation (1.5). Conversely, if x(t) is a solution of the integral equation (1.5), then x(a) = c and, by the Fundamental Theorem of Calculus, we have 0 xk(t) = Xk(x(t), t), k = 1, . . . , n. dx These imply that x(t) satisfies = X(x, t). 2 dt For a given X(x, t), if it is defined for all x in |t − a| ≤ T , and is continuous, then we can define an operator U by Z t U(x) = c + X(x(s), s)ds. (1.6) a For this operator, its domain is x(t) x(t) is continuous in the interval |t − a| ≤ T and its range is y(t) y(t) is continuously differentiable in the interval |t − a| ≤ T and y(a) = c .
Thus, a solution of the integral equation (1.5) is a fixed point of the operator U:
x = U(x).
Theorem 1.4.1 can be re-stated as the following:
Theorem 1.4.2 Let X(x, t) be a continuous function. Then a function x(t) is a solution of the initial value problem (1.4) if and only if the operator U defined in (1.6) has a fixed point in C[a − T, a + T ]. 14 1 EXISTENCE AND UNIQUENESS
Now we can use the operator U to generate a sequence of functions {xn} from the given initial data by the successive iteration:
x0(t) ≡ c, xn = U(xn−1) = U n(x0), n = 1, 2,.... (1.7)
This is call a Picard iteration.5 Now we show that, under some conditions, a Picard iteration converges to a solution of the initial value problem (1.4).
Lemma 1.4.3 Assume that X(x, t) is continuous and satisfies the Lipschitz condition (1.2) on the interval |t − a| ≤ T for all x, y. Then the Picard iteration (1.7) converges uniformly for |t − a| ≤ T .
Proof Let M = sup ||X(c, t)|| < +∞. Without loss of generality we can assume that a = 0 and t ≥ a. |t−a|≤T In other words, we prove the lemma on the interval 0 ≤ t ≤ T . The proof for the general a and t < a can be deduced from this case by the substitutions t → t + a and t → a − t. We first prove by induction that
n n n−1 (M/L)(Lt) x (t) − x (t) ≤ , n = 1, 2,.... (1.8) n! In fact, for n = 1, Z t 1 0 0 x (t) − x (t) = X(x (s), s) ds 0 Z t 0 ≤ X(x (s), s) ds 0 Z t (M/L)(Lt)1 ≤ M ds = Mt = . 0 1! For n = 2, Z t 2 1 1 0 x (t) − x (t) = X(x (s), s) − X(x (s), s) ds 0 Z t 1 0 ≤ X(x (s), s) − X(x (s), s) ds 0 Z t 1 0 ≤ L x (s) − x (s) ds 0 Z t LMt2 (M/L)(Lt)2 ≤ L Ms ds = = . 0 2 2! Assume that the desired inequality holds for n = k:
(M/L)(Lt)k xk(t) − xk−1(t) ≤ . k!
5The Picard iteration is started from x0 = c, given by the initial data. 1.4 EXISTENCE THEOREM 15
Then, Z t h i k+1 k k k−1 x (t) − x (t) = X(x (s), s) − X(x (s), s) ds 0 Z t k k−1 ≤ X(x (s), s) − X(x (s), s) ds 0 Z t k k−1 ≤ L x (s) − x (s) ds 0 Z t (M/L)(Ls)k (M/L)(Lt)k+1 ≤ L ds = . 0 k! (k + 1)! Hence, the estimates (1.8) hold. Next, we show that the sequence {xn(t)} converges uniformly for 0 ≤ t ≤ T . Indeed, for 0 ≤ t ≤ T , since
(M/L)(Lt)k (M/L)(LT )k ≤ , k! k! and the positive series ∞ X (M/L)(LT )k k! k=1 is convergent to (M/L) eLT − 1, by the Comparison Test, the series
∞ 0 X h k k−1 i x (t) + x (t) − x (t) k=1
(M/L)(LT )k is also convergent. Actually, the convergence is uniform for 0 ≤ t ≤ T , since is independent of k! t. The n-th partial sum, n 0 X h k k−1 i n x (t) + x (t) − x (t) = x (t). k=1 Therefore, the sequence of functions {xn(t)} converges uniformly. 2 Theorem 1.4.4 (Existence Theorem) Assume that X(x, t) is continuous and satisfies the Lip- schitz condition (1.2) on the interval |t − a| ≤ T for all x, y. Then the initial value problem (1.4) has a unique solution on the interval |t − a| ≤ T . Proof The uniqueness is a direct consequence of Uniqueness Theorem 1.2.1. We only need to prove the existence. By Lemma 1.4.3, the sequence {xn(t)} defined by the Picard iteration with x0(t) ≡ c is uniformly convergent. Denote x∞(t) the limit function. We show that x∞(t) is a solution of the integral equation (1.5). By the definition, Z t xn+1(t) = c + X(xn(s), s) ds. 0 The left hand side is uniformly convergent to x∞(t). By the Lipschitz condition,
kX(xm(s), s) − X(xn(s), s)k ≤ L kxm(s) − xn(s)k , and so the integral on the right hand side is also uniformly convergent. Since X(x, s) is continuous, we know that X(xn(s), s) → X(x∞(s), s). 16 1 EXISTENCE AND UNIQUENESS
Hence, we obtain Z t x∞(t) = c + X(x∞(s), s) ds. 0 Finally, by Theorem 1.4.1, we conclude that the function x∞(t) is a solution of the initial value prob- lem (1.4). 2 Example 1.4.1 Solve the initial value problem dx = x, x(0) = 1. dt by the Picard iteration. Solution For this initial value problem, the integral operator U is defined as
Z t U(x(t)) = 1 + x(s) ds 0 Thus, the Picard iteration gives
x0(t) = 1, Z t x1(t) = U(x0(t)) = 1 + 1 ds = 1 + t, 0 Z t x2(t) = U(x1(t)) = 1 + (1 + s) ds 0 t2 = 1 + t + , 2 . . t2 tn x (t) = 1 + t + + ... + , n 2 n!
xn+1(t) = U(xn(t)) Z t s2 sn = 1 + 1 + s + + ... + ds 0 2 n! t2 tn tn+1 = 1 + t + + ... + + . 2 n! (n + 1)!
t t We see that the sequence {xn(t)} converges uniformly to the function e . Hence we get the solution x(t) = e by the Picard iteration. 2 Example 1.4.2 Verify the Taylor series for sin t and cos t by applying the Picard iteration to the first order system corresponding to the second order initial value problem x00 = −x, x(0) = 0, x0(0) = 1. Solution The associated first order system is
dx1 = x2, dt dx 2 = −x , dt 1 1.4 EXISTENCE THEOREM 17 with the initial condition x(0) = (x1(0), x2(0)) = (0, 1).
The corresponding X(x, t) = (x2, −x1), and the initial value c = (0, 1). The Picard iteration yields
x0(t) = (0, 1), Z t x1(t) = (0, 1) + (1, 0) ds 0 = (0, 1) + (t, 0) = (t, 1), Z t x2(t) = (0, 1) + (1, −s) ds 0 t2 t2 = (0, 1) + t, − = t, 1 − , 2 2
Z t s2 x3(t) = (0, 1) + 1 − , −s ds 0 2 t3 t2 = (0, 1) + t − , − 3! 2 t3 t2 = t − , 1 − , 3! 2 Z t s2 s3 x4(t) = (0, 1) + 1 − , −s + ds 0 2 3! t3 t2 t4 = (0, 1) + t − , − + 3! 2 4! t3 t2 t4 = t − , 1 − + . 3! 2! 4!
We claim that for n = 0, 1, 2,...,
t2n+1 t2n x2n+1(t) = t + ··· + (−1)n , 1 + ··· + (−1)n , (2n + 1)! (2n)! t2n+1 t2n+2 x2n+2(t) = t + ··· + (−1)n , 1 + ··· + (−1)n+1 . (2n + 1)! (2n + 2)!
We prove this Claim by mathematical induction. When n = 0, they hold due to the calculations above. Suppose they hold when n = k. Then we have
Z t s2k+2 s2k+1 x2k+3(t) = (0, 1) + 1 + ··· + (−1)k+1 , −s − · · · − (−1)k ds 0 (2n + 2)! (2k + 1)! t2k+3 t2k+2 = t + ··· + (−1)k+1 , 1 + ··· + (−1)k+1 , (2k + 3)! (2k + 2)! Z t s2k+2 s2k+3 x2k+4(t) = (0, 1) + 1 + ··· + (−1)k+1 , −s − · · · − (−1)k+1 ds 0 (2k + 2)! (2k + 3)! t2k+3 t2k+4 = t + ··· + (−1)k+1 , 1 + ··· + (−1)k+2 . (2k + 3)! (2k + 4)! 18 1 EXISTENCE AND UNIQUENESS
That is, the claim is also true for n = k + 1. It can be easily shown that the associated initial value problem has a unique solution
x(t) = (sin t, cos t).
The uniqueness can be verified by checking that the conditions of Uniqueness Theorem 1.2.1 hold for the associated problem. Since {xn(t)} converges to the unique solution by Existence Theorem 1.4.4, we have
t3 t2n−1 sin t = t − + ··· + (−1)n−1 + ··· , 3! (2n − 1)! t2 t2n cos t = 1 − + ··· + (−1)n + ··· . 2! (2n)! 2
Example 1.4.3 Let f(x, t) be a function satisfying the Lipschitz condition |f(x, t) − f(y, t)| ≤ L|x − y| for all 0 ≤ t ≤ T and all x and y. Suppose f(x, t) is continuous and bounded. Let 0 M = sup |f(x, t)|. Let x(t) be a solution of x = f(x, t) with initial value x(0) = c and xk(t) −∞ Solution: Use mathematical induction. When k = 0, we have 0 |x(t) − x0(t)| = |x(t) − x(0)| = |x (ξ)| for some 0 < ξ < t by the Mean Value Theorem. Since |x0(ξ)| = |f(ξ, t)| ≤ M, the desired inequality holds for k = 0. Suppose the inequality holds for k = n. Since Z t Z t |x(t) − xn+1(t)| = f(x(s), s)ds − f(xn(s), s)ds 0 0 Z t ≤ |f(x(s), s) − f(xn(s), s)| ds 0 Z t ≤ L|x(s) − xn(s)|ds 0 Z t MLn ≤ L sn+1ds 0 (n + 1)! tn+2 = MLn+1 , (n + 2)! we know that the inequality holds when k = n + 1. By mathematical induction, the inequality holds for all k ≥ 0. 2 1.5 LOCAL EXISTENCE THEOREM AND THE PEANO THEOREM 19 Exercise 1.4 1. For the initial value problem x0 = tx, x(0) = 1, obtain the n-th approximation of the Picard iteration. Use mathematical induction to justify the formula. 0 2. For the initial value problem x = t + x, x(0) = 0, obtain x0(t), x1(t), x2(t) and the n-th term of the sequence of the Picard approximation. Use mathematical induction to justify the formula. 3. Assume that X(x, t) is continuous and satisfies the Lipschitz condition (1.2) on the interval |t − a| ≤ T for all x, y. For any function f(t), continuous on |t − a| ≤ T , define the sequence {xn(t)} by Z t x0(t) = f(t), xn(t) = c + X(xn−1(s), s)ds, n = 1, 2,.... a Show that {xn(t)} is uniformaly convergent to the unique solution of the initial value problem (1.4). 4. Show that the initial value problem y00 + y2 − 1 = 0, y(0) = y0, 0 y (0) = y1 is equivalent to the integral equation Z t 2 y(t) = y0 + y1t − (t − s) y (s) − 1 ds. 0 5. Solve the integral equation Z t y(t) = cos t + sin(τ − t) · y(τ)dτ. 0 6. (1) Show that the initial value problem (y00 + (1 + t2)y = 0, t > 0, y(0) = 1, y0(0) = 0 is equivalent to the integral equation Z t y(t) = cos t + sin(τ − t) · τ 2y(τ)dτ. 0 (2) Show that the successive approximations {φn(t)} defined by φ0(t) = 0, Z t 2 φn(t) = cos t + sin(τ − t) · τ φn−1(τ)dτ, n = 1, 2,... 0 converges uniformly to the solution in (1) for 0 ≤ t < T . Here T is any fixed constant. 1.5 LOCAL EXISTENCE THEOREM AND THE PEANO THEOREM 1.5.1 Local Existence Theorem In the Existence Theorem 1.4.4, the function X(x, t) satisfies a Lipschitz condition (1.2) for all x. This condition is quite strong and many functions may fail this condition. 20 1 EXISTENCE AND UNIQUENESS Example 1.5.1 Show that the conclusion of Theorem 1.4.4 fails for the initial value problem dx = ex, dt x(0) = c. Solution The solution of the initial value problem is given implicitly by e−c − e−x = t. It is obvious that the function is defined only in −∞ < t < e−c. Hence, there is no > 0 such that the differential equation has a solution defined on all of |t| < for every initial value, since we can always take a sufficient large c > such that e−c < . Thus, the conclusion of Theorem 1.4.4 fails for this initial value problem. The cause of this failure is that the function X(x) = ex does not satisfy a Lipschitz condition for all x. In fact, |X(x, t) − X(0, t)| ex − 1 = |x − 0| x is unbounded for large values of x. 2 However, if the function X(x, t) satisfies a Lipschitz condition in a bounded domain, then a solution exists in a limited region. Theorem 1.5.1 (Local Existence Theorem) Assume that X(x, t) is continuous and satisfies the Lipschitz condition (1.2) in the closed domain kx − ck ≤ K, |t − a| ≤ T . Then the initial value problem (1.4) has a unique solution in the interval |t − a| ≤ min{T, K/M}, where M = sup kX(x, t)k . kx−ck≤K |t−a|≤T Proof The existence can be proved as in the proof of Theorem 1.4.4, except that we have to modify the estimates for the sequence {xn(t)} by showing that if x(t) is defined and continuous on |t−a| ≤ min{T, K/M} satisfying (a) x(t) is defined and continuous on |t − a| ≤ min{T, K/M}; (b) x(a) = c; (c) kx(t) − ck ≤ K on |t − a| ≤ min{T, K/M}, then one iteration y = U(x) still satisfies these three conditions. Indeed, (a) and (b) are obvious. To show (c), we have the following Z t ky(t) − ck = X(x(s), s) ds a Z t ≤ kX(x(s), s)k ds a K ≤ M · |t − a| ≤ M · = K. M 2 1.5 LOCAL EXISTENCE THEOREM AND THE PEANO THEOREM 21 1.5.2 The Peano Theorem The Lipschitz condition plays an important role in the proof of the Local Existence Theorem. However, if we drop it, we are still able to prove the existence, by a more sophisticated argument. Theorem 1.5.2 (Peano Existence Theorem) Assume that X(x, t) is continuous in the closed domain kx − ck ≤ K, |t − a| ≤ T . Then the initial value problem (1.4) has at least one solution in the interval |t − a| ≤ min{T, K/M}, where M = sup kX(x, t)k . kx−ck≤K |t−a|≤T To prove this theorem, we need the following definition and the famous Arzel´a-AscoliTheorem. Definition 1.5.3 A family of functions F is said to be equicontinuous on [a, b] if for any given > 0, there exists a number δ > 0 such that kx(t) − x(s)k < whenever |t − s| < δ for every function x ∈ F and t, s ∈ [a, b]. Arzel´a-AscoliTheorem Assume that the sequence {xn(t)} is bounded and equicontinuous on [a, b]. Then there exists a subsequence {xni (t)} that is uniformly convergent on [a, b]. Now we are ready to prove Peano Existence Theorem 1.5.2. Proof Denote T1 = min{T, K/M}. As argued at the beginning of the proof of Lemma 1.4.3, we only need to prove the theorem on 0 ≤ t ≤ T1, with a = 0. n We first construct a sequence of bounded equicontinuous functions {x (t)} on [0,T1]. For each n, define c, for 0 ≤ t ≤ T1/n, n x (t) = Z t−T1/n n c + X (x (s), s) ds, for T1/n < t ≤ T1. 0 The above formula defines the value of xn(t) recursively in terms of the previous values of xn(t). We can use mathematical induction to show that kxn(t) − ck ≤ K n on [0,T1]. Indeed, on [0,T1/n], it is trivial since x (t) = c. If we assume that the inequality holds on [0, k · T1/n] (0 ≤ k < n), then on [k · T1/n, (k + 1) · T1/n], Z t−T1/n n n kx (t) − ck = X (x (s), s) ds 0 ≤ M · |t − T1/n| ≤ M · T1 ≤ K. n Hence, the sequence {x (t)} is uniformly bounded on [0,T1]: kxn(t)k ≤ kck + K. 22 1 EXISTENCE AND UNIQUENESS n The equicontinuity of the sequence {x (t)} on [0,T1] can be proven by the following estimates: for any t1, t2 ∈ [0,T1], 0, if t , t ∈ [0,T /n], 1 2 1 Z t2−T 1/n n X(x (s), s)ds , if t1 ∈ [0,T1/n] and t2 ∈ (T1/n, T1], 0 n n Z t1−T 1/n kx (t1) − x (t2)k = n X(x (s), s)ds , if t2 ∈ [0,T1/n] and t1 ∈ (T1/n, T1], 0 Z t2−T 1/n n X(x (s), s)ds , if t1, t2 ∈ (T1/n, T1] t1−T 1/n ≤ M|t1 − t2|. By Arzel´a-AscoliTheorem, we know that there exists a uniformly convergent subsequence {xni (t)} that ∞ ∞ converges to a continuous function x (t) on [0,T1] as ni → ∞. We can show that the function x (t) is actually a solution of the initial value problem (1.4). Indeed, for any fixed t ∈ (0,T1], we take ni sufficiently n large such that T1/ni < t. Thus, by the definition of {x (t)}, we have Z t Z t xni (t) = c + X (xni (s), s) ds − X (xni (s), s) ds. 0 t−T1/ni Z t Z t ni ∞ As ni → ∞, since X (x, t) is uniformly continuous, we have X (x (s), s) ds → X (x (s), s) ds; the 0 0 second integral of the last equation tends to zero, since Z t Z t T X (xni (s), s) ds ≤ Mds = M · 1 → 0. ni t−T1/ni t−T1/ni Hence, we know that the function x∞(t) satisfies the integral equation Z t x∞(t) = c + X (x∞(s), s) ds. 0 2 Without the Lipschitz condition, it is known that solution of initial value problem is not necessarily unique. For instance, it can be earily see that the initial value problem dx 3 = x1/3, t ≥ 0, dt 2 x(0) = 0, is not unique, with the following two solutions 3/2 x1(t) = 0, x2(t) = t . Exercise 1.5 1.6 LINEAR SYSTEMS 23 1. Determine the existence region of the solution to the initial value problem dx = t2 + x2, dt x(0) = 0, where the equation is defined in the region R = {(x, t)||x| ≤ 1, |t| ≤ 1}. Find the first four approximations x0(t), x1(t), x2(t), x3(t) of the Picard iteration. 2. Determine the existence region of the solution to the initial value problem dx = t2 − x2, dt x(0) = 0, where the equation is defined in the region R = {(x, t)||x| ≤ 1, |t| ≤ 1}. Find the first four approximations x0(t), x1(t), x2(t), x3(t) of the Picard iteration. 3. Assume that the functions F (x1, x2, . . . , xn, t) is continuous and satisfies a Lipschitz condition in |t − a| ≤ T , kx − ck ≤ K. Then the initial value problem ( u(n) = F (u, u0, u0, . . . , u(n−1), t), (i) u (a) = ci, 0 ≤ i ≤ n − 1, has a unique solution on |t − a| ≤ min{T, K/M}, where