Matrix Differential Operator Method of Finding a Particular Solution to A

Total Page:16

File Type:pdf, Size:1020Kb

Matrix Differential Operator Method of Finding a Particular Solution to A MATRIX DIFFERENTIAL OPERATOR METHOD OF FINDING A PARTICULAR SOLUTION TO A NONHOMOGENEOUS LINEAR ORDINARY DIFFERENTIAL EQUATION WITH CONSTANT COEFICIENTS JOZEF FECENKO Abstract. The article presents a matrix differential operator and a pseu- doinverse matrix differential operator for finding a particular solution to nonhomogeneous linear ordinary differential equations (ODE) with con- stant coeficients with special types of the right-hand side. Calculation requires the determination of an inverse or pseudoinverse matrix. If the matrix is singular, the Moore-Penrose pseudoinverse matrix is used for the calculation, which is simply calculated as the inverse submatrix of the considered matrix. It is shown that block matrices are effectively used to calculate a particular solution. 1. Introduction The idea of representing the processes of calculus, differentiation and in- tegration, as operators has a long history that goes back to the prominent German polymath Leibniz, G.W. The French mathematician L. F. A. Arbo- gast was the first mathematician to separate the symbols of operation from those of quantity, introducing systematically the operator notation DF for the derivative of the function. This approach was further developed by F. J. Servois who developed convenient notations. He was followed by a school of British and Irish mathematicians. R. B. Carmichael and G. Boole describing the application of operator methods to ordinary and partial differential equa- tions. Other prominent personalities who contributed to the development of operational calculus were, for example, physicist O. Heaviside, mathematicians arXiv:2101.02037v1 [math.GM] 31 Dec 2020 T.J. Bromwich, J.R. Carson, J. Mikusi´nski, N. Wiener, and others [13]. 2010 Mathematics Subject Classification. 34A30, 15A09. Key words and phrases. operator method of solving differential equations, solving differ- ential equations using a matrix differential operator, More-Penrose pseudoinverse matrix, block matrices. This paper was supported by the Slovak Grant Agency VEGA No. 1/0647/19. I owe thanks to J´an Buˇsa and Graham Luffrum for their valuable comments and advice that helped improve the text in many places. 1 2 JOZEF FECENKO Although the problem of solving ordinary nonhomogeneous linear differen- tial equations with constant coefficients is generally known, we will deal with the method of finding a particular solution of differential equation using a matrix differential operator method. This method, like the method of unde- termined coefficients, can be used only in special cases, if the right-hand side of the differential equation is typical, i.e. it is a constant, a polynomial function, exponential function eαx, sine or cosine functions sin βx or cos βx, or finite sums and products of these functions (α, β constants). In some cases, the method can be used as a support method for determining the particular solu- tion of a differential equation using the method of undetermined coefficients or the differential operator method or for evaluation of indefinite integrals. We introduce the differential operator notation. It is sometimes convenient to adopt the notation Dy, D2y, D3y, · · · , Dny dy d2y d3y dny to denote , , , · · · , . The symbols Dy, D2y,... are called differ- dx dx2 dx3 dxn ential operators [1] and have properties analogous to those of algebraic quan- tities [3]. Table 1: Operator Techniques [1], [10] A. n n φ(D) ckRk(x) ckφ(D)Rk(x) k=0 k=0 B. P P φ(D) · ψ(D) ψ(D) · φ(D) φ(D), ψ(D) polynomial operators in D C. φ(D)(xf(x)) xφ(D)f(x)+ φ′(D)f(x) D. Dneαx αneαx, n is non-negative integer E. Dn [eaxR(x)] eax (D + α)n R(x) n is non-negative integer F. n D2n sin βx −β2 sin βx n D2n cos βx −β2 cos βx Using the operator notation, we shall agree to write the differential equation (n) (n−1) ′ any + an−1y + . + a1y + a0y = f(x), an 6= 0, ai ∈ R (1.1) as n n−1 anD + an−1D + · · · + a1D + a0 y = f(x) (1.2) MATRIX DIFFERENTIAL OPERATOR METHOD FOR PARTICULAR SOLUTION ODE 3 or briefly φ(D)y = f(x), (1.3) where n n−1 φ(D)= anD + an−1D + · · · + a1D + a0 (1.4) is called an operator polynomial in D. If we want to emphasize the degree of the polynomial operator (1.4), we shall write it in the form φn(D). The use of the operator calculus for solving linear differential equations is well dealt with in several publications, e.g. [3], [1], [10]. 2. Matrix differential operator Definition 1. If S = {v1, v2,..., vn} is a set of vectors in a vector space V, then the set of all linear combination of v1, v2,..., vn is called the span of v1, v2,..., vn and is denoted by span(v1, v2,..., vn) or span(S). Let G be a vector space of all differentiable function. Consider the subspace V ⊂ G given by V = span (f1(x),f2(x), · · · ,fn(x)) , (2.1) where we assume that the functions f1(x),f2(x), · · · ,fn(x) are linearly inde- pendent. Since the set B = {f1(x),f2(x), · · · ,fn(x)} is linearly independent, it is a basis for V. The functions fi(x), i = 1, 2,...,n expressed in basis B using base vector coordinates are usually written 1 0 0 0 1 0 [f1(x)]B = . , [f2(x)]B = . , · · · , [fn(x)]B = . 0 0 1 The vector [fi(x)]B has in the i-th row 1 and 0 otherwise. Further, assume that the differential operator D maps V into itself Let n D(fi(x)) = cijfj(x), i = 1, 2, · · · , n, j X=1 where cij ∈ R, i, j = 1, 2,...,n are constants. Then ci1 c i2 [D (fi(x))]B = . , i, j = 1, 2,...,n . cin 4 JOZEF FECENKO and (see [9]) c11 c21 · · · cn1 c c · · · c . 12 22 n2 [D]B = [[D (f1(x))]B . [D (f2(x))]B . · · · . [D (fn(x))]B]= . .. c1n c2n · · · cnn (2.2) If n f(x)= αifi(x), αi ∈ R, i = 1, 2, · · · ,n, f(x) ∈ V i X=1 then α1 α 2 [f(x)]B = . αn We express what is the derivative of the function f(x) n n n n n Df(x)= αiDfi(x)= αi cijfj(x)= αicijfj(x) i i j j i X=1 X=1 X=1 X=1 X=1 respectively n ci1αi i=1 n c11 c21 · · · cn1 α1 P ci2αi c12 c22 · · · cn2 α2 [D(f(x))] = i=1 = . . B . . .. P . . . n c1n c2n · · · cnn αn cinαi i=1 P Let us next simply denote [D]B as DB. The matrix DB we will called a matrix differential operator corresponding to a vector space V with the considered basis B. Denote β1 β ′ 2 [D(f(x)]B = f B = . and [f(x)]B = f B . βn then ′ f B = DBf B MATRIX DIFFERENTIAL OPERATOR METHOD FOR PARTICULAR SOLUTION ODE 5 Note that the matrix transformation DB : V → V defined by ′ DBf B = f B is a linear transformation. As mentioned in (2.2), i-th, (i = 1, 2, · · · ,n) column of the matrix D ex- presses the derivative of the function fi(x). The following properties of the matrix differential operator are given with- out proof. Theorem 1. Let DB be a matrix differential operator of vector space V with a basis B. Then • DB (f B + gB)= DBf B + DBgB R • DB (cf B)= cDB (f B) , c ∈ ′′ 2 • f B = DB (DBf B)= DBf B (n) (n−1) (0) • f B = DBf B , f B = f B (n) n • f B = DBf B n • i-th (i = 1, 2, · · · ,n) column of the matrix DB expresses n-th derivative of function fi(x). Example 1. Let us consider V = span(xe2x sin 3x,xe2x cos 3x, e2x sin 3x, e2x cos 3x) (2.3) The Wronskian W (xe2x sin 3x,xe2x cos 3x, e2x sin 3x, e2x cos 3x) = 324 · e8x 6= 0. It follows that the functions in (2.3) are linearly independent. Then the set B = {xe2x sin 3x,xe2x cos 3x, e2x sin 3x, e2x cos 3x} is a basis for V. Applying differential operator D to a general element of V, we see that D axe2x sin 3x + bxe2x cos 3x + ce2x sin 3x + de2x cos 3x 2x 2x 2x = (2a − 3b)xe sin 3x + (3a + 2b)xe cos 3x + (a + 2c − 3d)e sin 3x + (b + 3c + 2d)e2x cos 3x, which is again in V. We found the corresponding matrix differential operator DB in the vector space V with the basis B. Here the construction of DB is shown schematically 6 JOZEF FECENKO x x x x cos3 sin 3 cos3 sin 3 x x 2 2 x x 2 2 xe e e xe D D D D xe2x sin 3x 2 −3 0 0 xe2x cos 3x 3 2 0 0 D = (2.4) B e2x sin 3x 1 0 2 −3 e2x cos 3x 0 1 3 2 Let us calculate the first and second derivative of the functi on f(x)= xe2x sin 3x + 2xe2x cos 3x − 2e2x cos 3x using the matrix differential operator (2.4). We have 2 −3 0 0 1 −4 3 2 0 0 2 7 [Df(x)] = D [f(x)] = = B B B 1 0 2 −3 0 7 0 1 3 2 −2 −2 So that Df(x)= −4xe2x sin 3x + 7xe2x cos 3x + 7e2x sin 3x − 2e2x cos 3x The second derivative 2 −3 0 0 −4 −29 ′′ ′ 3 2 0 0 7 2 f = D f = = B B B 1 0 2 −3 7 16 0 1 3 2 −2 24 D2f(x)= −29xe2x sin 3x + 2xe2x cos 3x + 16e2x sin 3x + 24e2x cos 3x For calculating higher powers of the matrix DB it is sometimes advanta- geous to express it as a block matrix (see example 11) or use Jordan matrix decomposition.
Recommended publications
  • Linear Differential Equations in N Th-Order ;
    Ordinary Differential Equations Second Course Babylon University 2016-2017 Linear Differential Equations in n th-order ; An nth-order linear differential equation has the form ; …. (1) Where the coefficients ;(j=0,1,2,…,n-1) ,and depends solely on the variable (not on or any derivative of ) If then Eq.(1) is homogenous ,if not then is nonhomogeneous . …… 2 A linear differential equation has constant coefficients if all the coefficients are constants ,if one or more is not constant then has variable coefficients . Examples on linear differential equations ; first order nonhomogeneous second order homogeneous third order nonhomogeneous fifth order homogeneous second order nonhomogeneous Theorem 1: Consider the initial value problem given by the linear differential equation(1)and the n initial Conditions; … …(3) Define the differential operator L(y) by …. (4) Where ;(j=0,1,2,…,n-1) is continuous on some interval of interest then L(y) = and a linear homogeneous differential equation written as; L(y) =0 Definition: Linearly Independent Solution; A set of functions … is linearly dependent on if there exists constants … not all zero ,such that …. (5) on 1 Ordinary Differential Equations Second Course Babylon University 2016-2017 A set of functions is linearly dependent if there exists another set of constants not all zero, that is satisfy eq.(5). If the only solution to eq.(5) is ,then the set of functions is linearly independent on The functions are linearly independent Theorem 2: If the functions y1 ,y2, …, yn are solutions for differential
    [Show full text]
  • The Wronskian
    Jim Lambers MAT 285 Spring Semester 2012-13 Lecture 16 Notes These notes correspond to Section 3.2 in the text. The Wronskian Now that we know how to solve a linear second-order homogeneous ODE y00 + p(t)y0 + q(t)y = 0 in certain cases, we establish some theory about general equations of this form. First, we introduce some notation. We define a second-order linear differential operator L by L[y] = y00 + p(t)y0 + q(t)y: Then, a initial value problem with a second-order homogeneous linear ODE can be stated as 0 L[y] = 0; y(t0) = y0; y (t0) = z0: We state a result concerning existence and uniqueness of solutions to such ODE, analogous to the Existence-Uniqueness Theorem for first-order ODE. Theorem (Existence-Uniqueness) The initial value problem 00 0 0 L[y] = y + p(t)y + q(t)y = g(t); y(t0) = y0; y (t0) = z0 has a unique solution on an open interval I containing the point t0 if p, q and g are continuous on I. The solution is twice differentiable on I. Now, suppose that we have obtained two solutions y1 and y2 of the equation L[y] = 0. Then 00 0 00 0 y1 + p(t)y1 + q(t)y1 = 0; y2 + p(t)y2 + q(t)y2 = 0: Let y = c1y1 + c2y2, where c1 and c2 are constants. Then L[y] = L[c1y1 + c2y2] 00 0 = (c1y1 + c2y2) + p(t)(c1y1 + c2y2) + q(t)(c1y1 + c2y2) 00 00 0 0 = c1y1 + c2y2 + c1p(t)y1 + c2p(t)y2 + c1q(t)y1 + c2q(t)y2 00 0 00 0 = c1(y1 + p(t)y1 + q(t)y1) + c2(y2 + p(t)y2 + q(t)y2) = 0: We have just established the following theorem.
    [Show full text]
  • An Introduction to Pseudo-Differential Operators
    An introduction to pseudo-differential operators Jean-Marc Bouclet1 Universit´ede Toulouse 3 Institut de Math´ematiquesde Toulouse [email protected] 2 Contents 1 Background on analysis on manifolds 7 2 The Weyl law: statement of the problem 13 3 Pseudodifferential calculus 19 3.1 The Fourier transform . 19 3.2 Definition of pseudo-differential operators . 21 3.3 Symbolic calculus . 24 3.4 Proofs . 27 4 Some tools of spectral theory 41 4.1 Hilbert-Schmidt operators . 41 4.2 Trace class operators . 44 4.3 Functional calculus via the Helffer-Sj¨ostrandformula . 50 5 L2 bounds for pseudo-differential operators 55 5.1 L2 estimates . 55 5.2 Hilbert-Schmidt estimates . 60 5.3 Trace class estimates . 61 6 Elliptic parametrix and applications 65 n 6.1 Parametrix on R ................................ 65 6.2 Localization of the parametrix . 71 7 Proof of the Weyl law 75 7.1 The resolvent of the Laplacian on a compact manifold . 75 7.2 Diagonalization of ∆g .............................. 78 7.3 Proof of the Weyl law . 81 A Proof of the Peetre Theorem 85 3 4 CONTENTS Introduction The spirit of these notes is to use the famous Weyl law (on the asymptotic distribution of eigenvalues of the Laplace operator on a compact manifold) as a case study to introduce and illustrate one of the many applications of the pseudo-differential calculus. The material presented here corresponds to a 24 hours course taught in Toulouse in 2012 and 2013. We introduce all tools required to give a complete proof of the Weyl law, mainly the semiclassical pseudo-differential calculus, and then of course prove it! The price to pay is that we avoid presenting many classical concepts or results which are not necessary for our purpose (such as Borel summations, principal symbols, invariance by diffeomorphism or the G˚ardinginequality).
    [Show full text]
  • Linear Differential Operators
    3 Linear Differential Operators Differential equations seem to be well suited as models for systems. Thus an understanding of differential equations is at least as important as an understanding of matrix equations. In Section 1.5 we inverted matrices and solved matrix equations. In this chapter we explore the analogous inversion and solution process for linear differential equations. Because of the presence of boundary conditions, the process of inverting a differential operator is somewhat more complex than the analogous matrix inversion. The notation ordinarily used for the study of differential equations is designed for easy handling of boundary conditions rather than for understanding of differential operators. As a consequence, the concept of the inverse of a differential operator is not widely understood among engineers. The approach we use in this chapter is one that draws a strong analogy between linear differential equations and matrix equations, thereby placing both these types of models in the same conceptual frame- work. The key concept is the Green’s function. It plays the same role for a linear differential equation as does the inverse matrix for a matrix equa- tion. There are both practical and theoretical reasons for examining the process of inverting differential operators. The inverse (or integral form) of a differential equation displays explicitly the input-output relationship of the system. Furthermore, integral operators are computationally and theoretically less troublesome than differential operators; for example, differentiation emphasizes data errors, whereas integration averages them. Consequently, the theoretical justification for applying many of the com- putational procedures of later chapters to differential systems is based on the inverse (or integral) description of the system.
    [Show full text]
  • First Integrals of Differential Operators from SL(2, R) Symmetries
    mathematics Article First Integrals of Differential Operators from SL(2, R) Symmetries Paola Morando 1 , Concepción Muriel 2,* and Adrián Ruiz 3 1 Dipartimento di Scienze Agrarie e Ambientali-Produzione, Territorio, Agroenergia, Università Degli Studi di Milano, 20133 Milano, Italy; [email protected] 2 Departamento de Matemáticas, Facultad de Ciencias, Universidad de Cádiz, 11510 Puerto Real, Spain 3 Departamento de Matemáticas, Escuela de Ingenierías Marina, Náutica y Radioelectrónica, Universidad de Cádiz, 11510 Puerto Real, Spain; [email protected] * Correspondence: [email protected] Received: 30 September 2020; Accepted: 27 November 2020; Published: 4 December 2020 Abstract: The construction of first integrals for SL(2, R)-invariant nth-order ordinary differential equations is a non-trivial problem due to the nonsolvability of the underlying symmetry algebra sl(2, R). Firstly, we provide for n = 2 an explicit expression for two non-constant first integrals through algebraic operations involving the symmetry generators of sl(2, R), and without any kind of integration. Moreover, although there are cases when the two first integrals are functionally independent, it is proved that a second functionally independent first integral arises by a single quadrature. This result is extended for n > 2, provided that a solvable structure for an integrable distribution generated by the differential operator associated to the equation and one of the prolonged symmetry generators of sl(2, R) is known. Several examples illustrate the procedures. Keywords: differential operator; first integral; solvable structure; integrable distribution 1. Introduction The study of nth-order ordinary differential equations admitting the unimodular Lie group SL(2, R) is a non-trivial problem due to the nonsolvability of the underlying symmetry algebra sl(2, R).
    [Show full text]
  • 1 Function Spaces, Linear Operators, and Green's Functions
    1 1 Function Spaces, Linear Operators, and Green’s Functions 1.1 Function Spaces Consider the set of all complex-valued functions of the real variable x, denoted by f (x), g(x), ..., and defined on the interval (a, b). We shall restrict ourselves to those functions which are square-integrable.Definetheinner product of any two of the latter functions by b ∗ (f , g) ≡ f (x) g (x) dx, (1.1.1) a in which f ∗(x) is the complex conjugate of f (x). The following properties of the inner product follow from definition (1.1.1): (f , g)∗ = (g, f ), ( f , g + h) = ( f , g) + ( f , h), (1.1.2) ( f , αg) = α( f , g), (αf , g) = α∗( f , g), with α a complex scalar. While the inner product of any two functions is in general a complex number, the inner product of a function with itself is a real number and is nonnegative. This prompts us to define the norm of a function by 1 b 2 ∗ f ≡ ( f , f ) = f (x) f (x)dx , (1.1.3) a provided that f is square-integrable , i.e., f < ∞. Equation (1.1.3) constitutes a proper definition for a norm since it satisfies the following conditions: scalar = | | · (i) αf α f , for all comlex α, multiplication (ii) positivity f > 0, for all f = 0, = = f 0, if and only if f 0, triangular (iii) f + g ≤ f + g . inequality (1.1.4) Applied Mathematical Methods in Theoretical Physics, Second Edition. Michio Masujima Copyright 2009 WILEY-VCH Verlag GmbH & Co.
    [Show full text]
  • Sturm Liouville Theory
    1 Sturm-Liouville System 1.1 Sturm-Liouville Dierential Equation Denition 1. A second ordered dierential equation of the form d d − p(x) y + q(x)y = λω(x)y; x 2 [a; b] (1) dx dx with p; q and ! specied such that p(x) > 0 and !(x) > 0 for x 2 (a; b), is called a Sturm- Liouville(SL) dierential equation. Note that SL dierential equation is essentially an eigenvalue problem since λ is not specied. While solving SL equation, both λ and y must be determined. Example 2. The Schrodinger equation 2 − ~ + V (x) = E 2m on an interval [a; b] is a SL dierential equation. 1.2 Boundary Conditions Boundary conditions for a solution y of a dierential equation on interval [a; b] are classied as follows: Mixed Boundary Conditions Boundary conditions of the form 0 cay(a) + day (a) = α 0 cby(b) + dby (b) = β (2) where, ca; da; cb; db; α and β are constants, are called mixed Dirichlet-Neumann boundary conditions. When both α = β = 0 the boundary conditions are said to be homogeneous. Special cases are Dirichlet BC (da = db = 0) and Neumann BC (ca = cb = 0) Periodic Boundary Conditions Boundary conditions of the form y(a) = y(b) y0(a) = y0(b) (3) are called periodic boundary conditions. Example 3. Some examples are 1. Stretched vibrating string clamped at two ends: Dirichlet BC 2. Electrostatic potential on the surface of a volume: Dirichlet BC 3. Electrostatic eld on the surface of a volume: Neumann BC 4. A heat-conducting rod with two ends in heat baths: Dirichlet BC 1 1.3 Sturm-Liouville System (Problem) Denition 4.
    [Show full text]
  • Step Functions, Delta Functions, and the Variation of Parameters Formula
    STEP FUNCTIONS, DELTA FUNCTIONS, AND THE VARIATION OF PARAMETERS FORMULA STEPHEN SCHECTER 1. The unit step function and piecewise continuous functions The Heaviside unit step function u(t) is given by 0 if t< 0, u(t)= (1 if t> 0. The function u(t) is not defined at t = 0. Often we will not worry about the value of a function at a point where it is discontinuous, since often it doesn’t matter. u(t) u(t−a) 1−u(t−b) u(t−a)−u(t−b) 1 1 1 1 a b a b Figure 1.1. Heaviside unit step function. The function u(t) turns on at t = 0. The function u(t − a) is just u(t) shifted, or dragged, so that it turns on at t = a. The function 1 − u(t − b) turns off at t = b. The function u(t − a) − u(t − b), with a < b, turns on a t = a and turns off at t = b. We can use the unit step function to crop and shift functions. Multiplying f(t) by u(t−a) crops f(t) so that it turns on at t = a: 0 if t < a, u(t − a)f(t)= (f(t) if t > a. Multiplying f(t) by u(t − a) − u(t − b), with a < b, crops f(t) so that it turns on at t = a and turns off at t = b: 0 if t < a, (u(t − a) − u(t − b))f(t)= f(t) if a<t<b, 0 if t > b.
    [Show full text]
  • Gradients and Directional Derivatives R Horan & M Lavelle
    Intermediate Mathematics Gradients and Directional Derivatives R Horan & M Lavelle The aim of this package is to provide a short self assessment programme for students who want to obtain an ability in vector calculus to calculate gradients and directional derivatives. Copyright c 2004 [email protected] , [email protected] Last Revision Date: August 24, 2004 Version 1.0 Table of Contents 1. Introduction (Vectors) 2. Gradient (Grad) 3. Directional Derivatives 4. Final Quiz Solutions to Exercises Solutions to Quizzes The full range of these packages and some instructions, should they be required, can be obtained from our web page Mathematics Support Materials. Section 1: Introduction (Vectors) 3 1. Introduction (Vectors) The base vectors in two dimensional Cartesian coordinates are the unit vector i in the positive direction of the x axis and the unit vector j in the y direction. See Diagram 1. (In three dimensions we also require k, the unit vector in the z direction.) The position vector of a point P (x, y) in two dimensions is xi + yj . We will often denote this important vector by r . See Diagram 2. (In three dimensions the position vector is r = xi + yj + zk .) y 6 Diagram 1 y 6 Diagram 2 P (x, y) 3 r 6 j yj 6- - - - 0 i x 0 xi x Section 1: Introduction (Vectors) 4 The vector differential operator ∇ , called “del” or “nabla”, is defined in three dimensions to be: ∂ ∂ ∂ ∇ = i + j + k . ∂x ∂y ∂z Note that these are partial derivatives! This vector operator may be applied to (differentiable) scalar func- tions (scalar fields) and the result is a special case of a vector field, called a gradient vector field.
    [Show full text]
  • Differential Operators Commuting with Convolution Integral Operators
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS 91, 80-93 (1983) Differential Operators Commuting with Convolution Integral Operators F. ALBERTOGRUNBAUM * Department of Mathematics and Lawrence Berkeley Laboratory, University of California, Berkeley, California 94720 Submitted by P. D. Lax Received March 13, 1981 1. INTRODUCTION The need to make a detailed study of the spectral structure of the convolution type integral operator acting on L*(B) arises in a number of applications, including optics [ 11, radio astronomy [2,3], electron microscopy [4], and x-ray tomography [5,6]. For instance, consider the problem of determining the values of a function g(x), x E R, which is known to be supported in the set B, from the knowledge of its Fourier transform g(L) for A E A (both A and B are compact and have nonempty interiors). While in principle g(x) is completely determined by analytic continuation, the practical solution of this problem is a different matter. One is led to the consideration of the operator (1), where k(r) stands for the Fourier transform of the characteristic function of the set A. K turns out to be a compact operator with spectrum in (0, 1), and a detailed spectral decom- position of K exhibits the directions along which the problem is ill posed [6, 71. In the case when A and B are intervals centered at the origin the detailed structure of the eigenfunctions and eigenvalues of (1) has been displayed in * Partly supported by NSF Grant MCS 78-06718 and by the Director, Oflice of Energy Research, Office of Basic Energy Sciences, Engineering, Mathematical, and Geosciences Division of the U.S.
    [Show full text]
  • Differential Operator Method of Finding a Particular Solution to an Ordinary Nonhomogeneous Linear Differential Equation with Constant Coefficients
    Differential Operator Method of Finding A Particular Solution to An Ordinary Nonhomogeneous Linear Differential Equation with Constant Coefficients Wenfeng Chen Department of Mathematics and Physics, College of Arts and Sciences, SUNY Polytechnic Institute, Utica, NY 13502, USA We systematically introduce the idea of applying differential operator method to find a par- ticular solution of an ordinary nonhomogeneous linear differential equation with constant coefficients when the nonhomogeneous term is a polynomial function, exponential function, sine function, cosine function or any possible product of these functions. In particular, differ- ent from the differential operator method introduced in literature, we propose and highlight utilizing the definition of the inverse of differential operator to determine a particular solu- tion. We suggest that this method should be introduced in textbooks and widely used for determining a particular solution of an ordinary nonhomogeneous linear differential equation with constant coefficients in parallel to the method of undetermined coefficients. arXiv:1802.09343v1 [math.GM] 16 Feb 2018 2 I. INTRODUCTION When we determine a particular solution of an ordinary nonhomogeneous linear differential equa- tion of constant coefficients with nonhomogeneous terms being a polynomial function, an exponen- tial function, a sine function, a cosine function or any possible products of these functions, usually only the method of undetermined coefficients is introduced in undergraduate textbooks [1–3]. The basic idea of this method is first assuming the general form of a particular solution with the co- efficients undetermined, based on the nonhomogeneous term, and then substituting the assumed solution to determine the coefficients. The calculation process of this method is very lengthy and time-consuming.
    [Show full text]
  • Sturm–Liouville Theory
    2 Sturm–Liouville Theory So far, we’ve examined the Fourier decomposition of functions defined on some interval (often scaled to be from π to π). We viewed this expansion as an infinite dimensional − analogue of expanding a finite dimensional vector into its components in an orthonormal basis. But this is just the tip of the iceberg. Recalling other games we play in linear algebra, you might well be wondering whether we couldn’t have found some other basis in which to expand our functions. You might also wonder whether there shouldn’t be some role for matrices in this story. If so, read on! 2.1 Self-adjoint matrices We’ll begin by reviewing some facts about matrices. Let V and W be finite dimensional vector spaces (defined, say, over the complex numbers) with dim V = n and dim W = m. Suppose we have a linear map M : V W . By linearity, we know what M does to any → vector v V if we know what it does to a complete set v , v ,..., v of basis vectors ∈ { 1 2 n} in V . Furthermore, given a basis w , w ,..., w of W we can represent the map M in { 1 2 m} terms of an m n matrix M whose components are × Mai =(wa,Mvi) for a =1, . , m and i =1, . , n , (2.1) where ( , ) is the inner product in W . We’ll be particularly interested in the case m = n, when the matrix M is square and the map M takes M : V W = V is an isomorphism of vector spaces.
    [Show full text]