Change of Basis Between Classical Orthogonal Polynomials

Total Page:16

File Type:pdf, Size:1020Kb

Change of Basis Between Classical Orthogonal Polynomials Change of Basis between Classical Orthogonal Polynomials D.A. Wolfram College of Engineering & Computer Science The Australian National University, Canberra, ACT 0200 [email protected] Abstract Classical orthogonal polynomials have widespread applications includ- ing in numerical integration, solving differential equations, and interpola- tion. Changing basis between classical orthogonal polynomials can affect the convergence, accuracy, and stability of solutions. We provide a general method for changing basis between any pair of classical orthogonal polynomials by using algebraic expressions called coefficient functions that evaluate to connection coefficients. The method builds directly on previous work on the change of basis groupoid. The scope has fifteen kinds of classical orthogonal polynomials including the classes of Jacobi, Gegenbauer and generalized Laguerre polynomials. The method involves the mappings to and from the monomials for these polynomial bases. Sixteen coefficient functions appear to be new for polynomials that do not have definite parity. We derive the remainder from seven sources in the literature. We give a complete summary of thirty coefficient functions. This enables change of basis to be defined algebraically and uniformly between any pair of classical orthogonal polynomial bases by using a vector dot product to compose two coefficient functions to give a third. A definition of Jacobi polynomials uses a basis of shifted monomials. We find a key new mapping between the monomials and Jacobi polynomi- arXiv:2108.13631v1 [math.CA] 31 Aug 2021 als by using a general mapping between shifted monomials and monomials. It yields new mappings for the Chebyshev polynomials of the third and fourth kinds and their shifted versions. 1 Introduction and Related Work Change of basis of a finite vector space has numerous significant applications in scientific computing and engineering. Applications of change of basis with clas- sical orthogonal polynomials include improving properties of spectral methods MSC: Primary 33C45; Secondary 15A03, 20N02 1 for solving differential equations numerically. These properties include better convergence [6, 9], lower computational complexity [11, 15], and better numeri- cal stability [15]. Previous related work gives recurrence relations for finding connection coef- ficients for orthogonal polynomials, e.g., [14], for Jacobi polynomials [13], or by other methods [5, 7] and for special cases [4, 6]. This work derives coefficient functions that evaluate to connection coeffi- cients for mappings between the classical orthogonal polynomials and the mono- mials. We write the coefficient functions as algebraic expressions. From the framework of the change of basis groupoid [20], a vector dot product equa- tion enables us to compose coefficient functions in order to derive a coefficient function for any pair of classical orthogonal polynomials. 2 Classifications of Orthogonal Polynomials Thirteen kinds of orthogonal polynomials are classified by Koornwinder et al. [10, §18.3] as classical orthogonal polynomials: (α,β) • Jacobi, Pn (x) (λ) • Gegenbauer, Cn (x) • Chebyshev of the first to fourth kinds: Tn(x),Un(x), Vn(x), Wn(x) ∗ ∗ • Shifted Chebyshev of the first and second kinds: Tn (x),Un(x) ∗ • Legendre and Shifted Legendre, Pn(x), Pn (x) (α) • Generalised Laguerre, Ln (x) • Physicists’ and Probabilists’ Hermite: Hn(x), and Hen(x). The Jacobi, Gegenbauer and generalized Laguerre polynomials depend on the respective parameters, α and β, λ, and α. Andrews and Askey [3] give a more general classification of the classical orthogonal polynomials and stated “there are a number of different places to put the boundary for the classical polynomials”. We also include the shifted Chebyshev polynomials of the third and fourth ∗ ∗ kinds here, Vn (x) and Wn (x), because the classification is not fixed. 3 Related Classical Orthogonal Polynomials We shall use some of the properties described here that relate different kinds of classical orthogonal polynomials. Specifically, Chebyshev polynomials of the third and fourth kinds and Jacobi polynomials, and Chebyshev polynomials of the second kind and Legendre polynomials as special cases of Gegenbauer polynomials. 2 The Jacobi and Gegenbauer polynomials are related to some of the other classical orthogonal polynomials. The Gegenbauer polynomials are a special 1 1 case of the Jacobi polynomials where α = β, α = λ 2 and α> 1 or λ> 2 . From Szeg¨o[16, §4.7], − − − Γ(α + 1) Γ(Γ(n +2α + 1) C(λ)(x)= P (α,α)(x) (1) n 2α +1 n + α +1 n 1 λ 1 1 Γ( + 2 ) Γ(n +2λ) (λ− 2 ,λ− 2 ) = 1 Pn (x) (2) Γ(2λ) Γ(n + λ + 2 ) Chebyshev polynomials of the first kind are related to the Gegenbauer poly- nomials, e.g., Arfken [2]: (α) T0(x)=C0 (x) (3) (α) n Cn (x) Tn(x)= lim where n> 0. (4) 2 α→0 α The Chebyshev polynomials of the second kind are a special case of the (1) Gegenbauer polynomials with Un(x)= Cn (x). Chebyshev polynomials of the third and fourth kinds, Vn(x) and Wn(x), are special cases of the Jacobi poly- nomials, e.g., [12, §1.2.3]. The Legendre polynomials are also a special case of Gegenbauer polynomials 1 ( 2 ) with Pn(x)= Cn (x). 4 Definitions and Background Equations We consider a method of defining algebraically a change of basis mapping be- tween finite bases of classical orthogonal polynomials or the monomials. The method uses the matrix dot product of change of basis matrices. From the change of basis groupoid [20], given two change of basis matrices MtsMsv, we have Mtv = MtsMsv. (5) The elements of Mtv are formed from the vector dot products of the rows of Mts and the columns of Msv. All classical orthogonal polynomials and the monomials satisfy the following condition. It leads to an optimization of the vector dot products. Definition 1. Let V be a vector space of finite dimension n > 0 and B1 and B2 be bases of V . The basis B2 is a triangular basis with respect to B1 if and only if there is a permutation v1,...,vn of the coordinate vectors of B2 with respect to B1 such that the n n{ matrix v}T ,...,vT is an upper triangular matrix. × 1 n 3 Definition 2. Suppose that s and t are triangular bases of a vector space V . The mapping : s t satisfies T → n sn = α(n, k)tn−k (6) k X=0 where α(n, k) R, i.e., each basis vector of s is a unique linear combination of the basis vectors∈ of t. The function α is called a coefficient function and it evaluates to a connection coefficient. If α(n, k) is the coefficient function in equation (6), then the elements of the change of basis matrix Mts are defined by α(j, j i) if j i m = − ≥ (7) ij 0 if j<i. where 0 i, j n. ≤ ≤ Suppose that s, t and v are triangular bases, α1(n, k) is the coefficient func- tion for a change of basis matrix Mts and α2(n, k) is a coefficient function for Msv following equation (7). Then k α (n, k)= α (n v, k v)α (n, v) (8) 3 1 − − 2 v=0 X where 0 k n, is the coefficient function for Mtv. The basis≤ ≤s is called the exchange basis [20]. We shall use the monomials as the exchange basis. 4.1 Parity and Bases Six of the classical orthogonal polynomials have definite parity, and nine do not: the Jacobi polynomials, the generalized Laguerre polynomials, Chebyshev polynomials of the third and fourth kinds, and shifted orthogonal polynomials: T ∗, U ∗, V ∗, W ∗ and P ∗. (α,β) For example, the Jacobi Polynomials Pn (x) can be expressed in terms x−1 x−1 2 x−1 n of the basis 1, ( 2 ), ( 2 ) ,..., ( 2 ) , e.g., [10, equation 18.5.7]. The gen- { (α) } eralised Laguerre Polynomials Ln (x) can be expressed in terms of the basis 1, x, x2,...,xn . { The Chebyshev} Polynomials of the second kind have definite parity and they can be defined by n ⌊ 2 ⌋ n k −2k n k n−2k Un(x)=2 ( 1) 2 − x . − k k X=0 where n 0. When n is even, Un(x) can be expressed in terms of the ba- 2 ≥ 4 n sis 1, x , x ,...,x . When n is odd, Un(x) can be expressed in terms of x, x{3, x5,...,xn . } { } 4 Equation (8) can be optimized when the bases t and v have definite parity and we are concerned with finding the coefficient function for Mtv with basis vectors that all have either even parity or odd parity. In the example above, we have a coefficient function β(n, k) for non-zero coefficients: n k β(n, k) = ( 1)k2n−2k − − k where 0 k n . ≤ ≤⌊ 2 ⌋ Given β1(n, k) and β2(n, k) corresponding to the change of basis matrices Mts and Msv respectively, we have k β (n, k)= β (n 2v, k v)β (n, v) (9) 3 1 − − 2 v=0 X where 0 k n . This optimization excludes terms that are zero [20, §5.1]. ≤ ≤⌊ 2 ⌋ From equation (5), Mtv = MtsMsv, parity is not relevant if at least one of t or v does not have definite parity. In this case, the following equation can be used when β is known, but α is not. k β(n, 2 ) if k is even α(n, k)= (10) 0 if k is odd. The upper triangle of the associated change of basis matrix has zero and non-zero alternating elements. 5 Classical Orthogonal Polynomials and the Monomials We summarize here results of mappings between the classical orthogonal poly- nomials and the monomials. With the monomials as the exchange basis, we can find mappings between any pair of classical orthogonal polynomials.
Recommended publications
  • Chebyshev Polynomials of the Second, Third and Fourth Kinds in Approximation, Indefinite Integration, and Integral Transforms *
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Journal of Computational and Applied Mathematics 49 (1993) 169-178 169 North-Holland CAM 1429 Chebyshev polynomials of the second, third and fourth kinds in approximation, indefinite integration, and integral transforms * J.C. Mason Applied and Computational Mathematics Group, Royal Military College of Science, Shrivenham, Swindon, Wiltshire, United Kingdom Dedicated to Dr. D.F. Mayers on the occasion of his 60th birthday Received 18 February 1992 Revised 31 March 1992 Abstract Mason, J.C., Chebyshev polynomials of the second, third and fourth kinds in approximation, indefinite integration, and integral transforms, Journal of Computational and Applied Mathematics 49 (1993) 169-178. Chebyshev polynomials of the third and fourth kinds, orthogonal with respect to (l+ x)‘/‘(l- x)-‘I* and (l- x)‘/*(l+ x)-‘/~, respectively, on [ - 1, 11, are less well known than traditional first- and second-kind polynomials. We therefore summarise basic properties of all four polynomials, and then show how some well-known properties of first-kind polynomials extend to cover second-, third- and fourth-kind polynomials. Specifically, we summarise a recent set of first-, second-, third- and fourth-kind results for near-minimax constrained approximation by series and interpolation criteria, then we give new uniform convergence results for the indefinite integration of functions weighted by (1 + x)-i/* or (1 - x)-l/* using third- or fourth-kind polynomial expansions, and finally we establish a set of logarithmically singular integral transforms for which weighted first-, second-, third- and fourth-kind polynomials are eigenfunctions.
    [Show full text]
  • Generalizations of Chebyshev Polynomials and Polynomial Mappings
    TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 359, Number 10, October 2007, Pages 4787–4828 S 0002-9947(07)04022-6 Article electronically published on May 17, 2007 GENERALIZATIONS OF CHEBYSHEV POLYNOMIALS AND POLYNOMIAL MAPPINGS YANG CHEN, JAMES GRIFFIN, AND MOURAD E.H. ISMAIL Abstract. In this paper we show how polynomial mappings of degree K from a union of disjoint intervals onto [−1, 1] generate a countable number of spe- cial cases of generalizations of Chebyshev polynomials. We also derive a new expression for these generalized Chebyshev polynomials for any genus g, from which the coefficients of xn can be found explicitly in terms of the branch points and the recurrence coefficients. We find that this representation is use- ful for specializing to polynomial mapping cases for small K where we will have explicit expressions for the recurrence coefficients in terms of the branch points. We study in detail certain special cases of the polynomials for small degree mappings and prove a theorem concerning the location of the zeroes of the polynomials. We also derive an explicit expression for the discrimi- nant for the genus 1 case of our Chebyshev polynomials that is valid for any configuration of the branch point. 1. Introduction and preliminaries Akhiezer [2], [1] and, Akhiezer and Tomˇcuk [3] introduced orthogonal polynomi- als on two intervals which generalize the Chebyshev polynomials. He observed that the study of properties of these polynomials requires the use of elliptic functions. In thecaseofmorethantwointervals,Tomˇcuk [17], investigated their Bernstein-Szeg˝o asymptotics, with the theory of Hyperelliptic integrals, and found expressions in terms of a certain Abelian integral of the third kind.
    [Show full text]
  • Chebyshev and Fourier Spectral Methods 2000
    Chebyshev and Fourier Spectral Methods Second Edition John P. Boyd University of Michigan Ann Arbor, Michigan 48109-2143 email: [email protected] http://www-personal.engin.umich.edu/jpboyd/ 2000 DOVER Publications, Inc. 31 East 2nd Street Mineola, New York 11501 1 Dedication To Marilyn, Ian, and Emma “A computation is a temptation that should be resisted as long as possible.” — J. P. Boyd, paraphrasing T. S. Eliot i Contents PREFACE x Acknowledgments xiv Errata and Extended-Bibliography xvi 1 Introduction 1 1.1 Series expansions .................................. 1 1.2 First Example .................................... 2 1.3 Comparison with finite element methods .................... 4 1.4 Comparisons with Finite Differences ....................... 6 1.5 Parallel Computers ................................. 9 1.6 Choice of basis functions .............................. 9 1.7 Boundary conditions ................................ 10 1.8 Non-Interpolating and Pseudospectral ...................... 12 1.9 Nonlinearity ..................................... 13 1.10 Time-dependent problems ............................. 15 1.11 FAQ: Frequently Asked Questions ........................ 16 1.12 The Chrysalis .................................... 17 2 Chebyshev & Fourier Series 19 2.1 Introduction ..................................... 19 2.2 Fourier series .................................... 20 2.3 Orders of Convergence ............................... 25 2.4 Convergence Order ................................. 27 2.5 Assumption of Equal Errors ...........................
    [Show full text]
  • Dymore User's Manual Chebyshev Polynomials
    Dymore User's Manual Chebyshev polynomials Olivier A. Bauchau August 27, 2019 Contents 1 Definition 1 1.1 Zeros and extrema....................................2 1.2 Orthogonality relationships................................3 1.3 Derivatives of Chebyshev polynomials..........................5 1.4 Integral of Chebyshev polynomials............................5 1.5 Products of Chebyshev polynomials...........................5 2 Chebyshev approximation of functions of a single variable6 2.1 Expansion of a function in Chebyshev polynomials...................6 2.2 Evaluation of Chebyshev expansions: Clenshaw's recurrence.............7 2.3 Derivatives and integrals of Chebyshev expansions...................7 2.4 Products of Chebyshev expansions...........................8 2.5 Examples.........................................9 2.6 Clenshaw-Curtis quadrature............................... 10 3 Chebyshev approximation of functions of two variables 12 3.1 Expansion of a function in Chebyshev polynomials................... 12 3.2 Evaluation of Chebyshev expansions: Clenshaw's recurrence............. 13 3.3 Derivatives of Chebyshev expansions.......................... 14 4 Chebychev polynomials 15 4.1 Examples......................................... 16 1 Definition Chebyshev polynomials [1,2] form a series of orthogonal polynomials, which play an important role in the theory of approximation. The lowest polynomials are 2 3 4 2 T0(x) = 1;T1(x) = x; T2(x) = 2x − 1;T3(x) = 4x − 3x; T4(x) = 8x − 8x + 1;::: (1) 1 and are depicted in fig.1. The polynomials can be generated from the following recurrence rela- tionship Tn+1 = 2xTn − Tn−1; n ≥ 1: (2) 1 0.8 0.6 0.4 0.2 0 −0.2 −0.4 CHEBYSHEV POLYNOMIALS −0.6 −0.8 −1 −1 −0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8 1 XX Figure 1: The seven lowest order Chebyshev polynomials It is possible to give an explicit expression of Chebyshev polynomials as Tn(x) = cos(n arccos x): (3) This equation can be verified by using elementary trigonometric identities.
    [Show full text]
  • Approximation Atkinson Chapter 4, Dahlquist & Bjork Section 4.5
    Approximation Atkinson Chapter 4, Dahlquist & Bjork Section 4.5, Trefethen's book Topics marked with ∗ are not on the exam 1 In approximation theory we want to find a function p(x) that is `close' to another function f(x). We can define closeness using any metric or norm, e.g. Z 2 2 kf(x) − p(x)k2 = (f(x) − p(x)) dx or kf(x) − p(x)k1 = sup jf(x) − p(x)j or Z kf(x) − p(x)k1 = jf(x) − p(x)jdx: In order for these norms to make sense we need to restrict the functions f and p to suitable function spaces. The polynomial approximation problem takes the form: Find a polynomial of degree at most n that minimizes the norm of the error. Naturally we will consider (i) whether a solution exists and is unique, (ii) whether the approximation converges as n ! 1. In our section on approximation (loosely following Atkinson, Chapter 4), we will first focus on approximation in the infinity norm, then in the 2 norm and related norms. 2 Existence for optimal polynomial approximation. Theorem (no reference): For every n ≥ 0 and f 2 C([a; b]) there is a polynomial of degree ≤ n that minimizes kf(x) − p(x)k where k · k is some norm on C([a; b]). Proof: To show that a minimum/minimizer exists, we want to find some compact subset of the set of polynomials of degree ≤ n (which is a finite-dimensional space) and show that the inf over this subset is less than the inf over everything else.
    [Show full text]
  • 8.3 - Chebyshev Polynomials
    8.3 - Chebyshev Polynomials 8.3 - Chebyshev Polynomials Chebyshev polynomials Definition Chebyshev polynomial of degree n ≥= 0 is defined as Tn(x) = cos (n arccos x) ; x 2 [−1; 1]; or, in a more instructive form, Tn(x) = cos nθ ; x = cos θ ; θ 2 [0; π] : Recursive relation of Chebyshev polynomials T0(x) = 1 ;T1(x) = x ; Tn+1(x) = 2xTn(x) − Tn−1(x) ; n ≥ 1 : Thus 2 3 4 2 T2(x) = 2x − 1 ;T3(x) = 4x − 3x ; T4(x) = 8x − 8x + 1 ··· n−1 Tn(x) is a polynomial of degree n with leading coefficient 2 for n ≥ 1. 8.3 - Chebyshev Polynomials Orthogonality Chebyshev polynomials are orthogonal w.r.t. weight function w(x) = p 1 . 1−x2 Namely, Z 1 T (x)T (x) 0 if m 6= n np m dx = (1) 2 π −1 1 − x 2 if n = m for each n ≥ 1 Theorem (Roots of Chebyshev polynomials) The roots of Tn(x) of degree n ≥ 1 has n simple zeros in [−1; 1] at 2k−1 x¯k = cos 2n π ; for each k = 1; 2 ··· n : Moreover, Tn(x) assumes its absolute extrema at 0 kπ 0 k x¯k = cos n ; with Tn(¯xk) = (−1) ; for each k = 0; 1; ··· n : 0 −1 k For k = 0; ··· n, Tn(¯xk) = cos n cos (cos(kπ=n)) = cos(kπ) = (−1) : 8.3 - Chebyshev Polynomials Definition A monic polynomial is a polynomial with leading coefficient 1. The monic Chebyshev polynomial T~n(x) is defined by dividing Tn(x) by 2n−1; n ≥ 1.
    [Show full text]
  • Algorithms for Classical Orthogonal Polynomials
    Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustr. 7, D-14195 Berlin - Dahlem Wolfram Ko epf Dieter Schmersau Algorithms for Classical Orthogonal Polynomials at Berlin Fachb ereich Mathematik und Informatik der Freien Universit Preprint SC Septemb er Algorithms for Classical Orthogonal Polynomials Wolfram Ko epf Dieter Schmersau koepfzibde Abstract In this article explicit formulas for the recurrence equation p x A x B p x C p x n+1 n n n n n1 and the derivative rules 0 x p x p x p x p x n n+1 n n n n1 n and 0 p x p x x p x x n n n n1 n n resp ectively which are valid for the orthogonal p olynomial solutions p x of the dierential n equation 00 0 x y x x y x y x n of hyp ergeometric typ e are develop ed that dep end only on the co ecients x and x which themselves are p olynomials wrt x of degrees not larger than and resp ectively Partial solutions of this problem had b een previously published by Tricomi and recently by Yanez Dehesa and Nikiforov Our formulas yield an algorithm with which it can b e decided whether a given holonomic recur rence equation ie one with p olynomial co ecients generates a family of classical orthogonal p olynomials and returns the corresp onding data density function interval including the stan dardization data in the armative case In a similar way explicit formulas for the co ecients of the recurrence equation and the dierence rule x rp x p x p x p x n n n+1 n n n n1 of the classical orthogonal p olynomials of a discrete variable are given that dep end only
    [Show full text]
  • Polynomial Interpolation and Chebyshev Polynomials
    Lecture 7: Polynomial interpolation and Chebyshev polynomials Michael S. Floater September 30, 2018 These notes are based on Section 3.1 of the book. 1 Lagrange interpolation Interpolation describes the problem of finding a function that passes through a given set of values f0,f1,...,fn at data points x0,x1,...,xn. Such a function is called an interpolant. If p is an interpolant then p(xi) = fi, i = 0, 1,...,n. Let us suppose that the points and values are in R. If the values fi are samples f(xi) from some real function f then p approximates f. One choice of p is a polynomial of degree at most n. Using the Lagrange basis functions n x − xj Lk(x)= , k =0, 1,...,n, xk − xj Yj=0 j=6 k we can express p explicitly as n p(x)= fkLk(x). Xk=0 Because Lk(xi) = 0 when i =6 k and Lk(xk) = 1, it is easy to check that p(xi) = fi for i = 0, 1,...,n. The interpolant p is also unique among all polynomial interpolants of degree ≤ n. This is because if q is another such interpolant, the difference r = p − q would be a polynomial of degree ≤ n 1 such that r(xi)=0for i =0, 1,...,n. Thus r would have at least n +1 zeros and by a result from algebra this would imply that r = 0. For this kind of interpolation there is an error formula. n+1 Theorem 1 Suppose f ∈ C [a,b] and let fi = f(xi), i =0, 1,...,n, where x0,x1,...,xn are distinct points in [a,b].
    [Show full text]
  • Asymptotics of the Chebyshev Polynomials of General Sets
    Joint Work with Jacob Christiansen and Maxim Zinchenko Part 1: Inv. Math, 208 (2017), 217-245 Part 2: Submitted (also with Peter Yuditskii) Part 3: Pavlov Memorial Volume, to appear. Asymptotics of the Chebyshev Polynomials of Chebyshev Polynomials General Sets Alternation Theorem Potential Theory Barry Simon Lower Bounds and Special Sets IBM Professor of Mathematics and Theoretical Physics, Emeritus Faber Fekete California Institute of Technology Szegő Theorem Pasadena, CA, U.S.A. Widom’s Work Totik Widom Bounds Main Results Totik Widom Bounds Szegő–Widom Asymptotics Part 2: Submitted (also with Peter Yuditskii) Part 3: Pavlov Memorial Volume, to appear. Asymptotics of the Chebyshev Polynomials of Chebyshev Polynomials General Sets Alternation Theorem Potential Theory Barry Simon Lower Bounds and Special Sets IBM Professor of Mathematics and Theoretical Physics, Emeritus Faber Fekete California Institute of Technology Szegő Theorem Pasadena, CA, U.S.A. Widom’s Work Totik Widom Bounds Main Results Totik Widom Joint Work with Jacob Christiansen and Maxim Zinchenko Bounds Part 1: Inv. Math, 208 (2017), 217-245 Szegő–Widom Asymptotics Part 3: Pavlov Memorial Volume, to appear. Asymptotics of the Chebyshev Polynomials of Chebyshev Polynomials General Sets Alternation Theorem Potential Theory Barry Simon Lower Bounds and Special Sets IBM Professor of Mathematics and Theoretical Physics, Emeritus Faber Fekete California Institute of Technology Szegő Theorem Pasadena, CA, U.S.A. Widom’s Work Totik Widom Bounds Main Results Totik Widom Joint Work with Jacob Christiansen and Maxim Zinchenko Bounds Part 1: Inv. Math, 208 (2017), 217-245 Szegő–Widom Part 2: Submitted (also with Peter Yuditskii) Asymptotics Asymptotics of the Chebyshev Polynomials of Chebyshev Polynomials General Sets Alternation Theorem Potential Theory Barry Simon Lower Bounds and Special Sets IBM Professor of Mathematics and Theoretical Physics, Emeritus Faber Fekete California Institute of Technology Szegő Theorem Pasadena, CA, U.S.A.
    [Show full text]
  • Orthogonal Polynomial Expansions for the Riemann Xi Function Dan Romik
    Orthogonal polynomial expansions for the Riemann xi function Dan Romik Author address: Department of Mathematics, University of California, Davis, One Shields Ave, Davis CA 95616, USA E-mail address: [email protected] Contents Chapter 1. Introduction 1 1.1. Background 1 1.2. Our new results: Tur´an'sprogram revisited and extended; expansion of Ξ(t) in new orthogonal polynomial bases 3 1.3. Previous work involving the polynomials fn 5 1.4. How to read this paper 6 1.5. Acknowledgements 6 Chapter 2. The Hermite expansion of Ξ(t) 7 2.1. The basic convergence result for the Hermite expansion 7 2.2. Preliminaries 8 2.3. Proof of Theorem 2.1 8 2.4. An asymptotic formula for the coefficients b2n 12 2.5. The Poisson flow, P´olya-De Bruijn flow and the De Bruijn-Newman constant 17 Chapter 3. Expansion of Ξ(t) in the polynomials fn 20 3.1. Main results 20 3.2. Proof of Theorem 3.1 21 3.3. Proof of Theorem 3.2 23 3.4. The Poisson flow associated with the fn-expansion 26 3.5. Evolution of the zeros under the Poisson flow 27 Chapter 4. Radial Fourier self-transforms 30 4.1. Radial Fourier self-transforms on Rd and their construction from balanced functions 30 4.2. The radial function A(r) associated to !(x) 31 4.3. An orthonormal basis for radial self-transforms 33 4.4. Constructing new balanced functions from old 35 4.5. The functions ν(x) and B(r) 35 4.6.
    [Show full text]
  • Orthogonal Polynomials and Classical Orthogonal Polynomials
    International Journal of Mechanical Engineering and Technology (IJMET) Volume 9, Issue 10, October 2018, pp. 1613–1630, Article ID: IJMET_09_10_164 Available online at http://iaeme.com/Home/issue/IJMET?Volume=9&Issue=10 ISSN Print: 0976-6340 and ISSN Online: 0976-6359 © IAEME Publication Scopus Indexed ORTHOGONAL POLYNOMIALS AND CLASSICAL ORTHOGONAL POLYNOMIALS DUNIA ALAWAI JARWAN Education for Girls College, Al-Anbar University, Ministry of Higher Education and Scientific Research, Iraq ABSTRACT The focus of this project is to clarify the concept of orthogonal polynomials in the case of continuous internal and discrete points on R and the Gram – Schmidt orthogonalization process of conversion to many orthogonal limits and the characteristics of this method. We have highlighted the classical orthogonal polynomials as an example of orthogonal polynomials because of they are great importance in physical practical applications. In this project, we present 3 types (Hermite – Laguerre – Jacobi) of classical orthogonal polynomials by clarifying the different formulas of each type and how to reach some formulas, especially the form of the orthogonality relation of each. Keywords: Polynomials, Classical Orthogonal, Monic Polynomial, Gram – Schmidt Cite this Article Dunia Alawai Jarwan, Orthogonal Polynomials and Classical Orthogonal Polynomials, International Journal of Mechanical Engineering and Technology, 9(10), 2018, pp. 1613–1630. http://iaeme.com/Home/issue/IJMET?Volume=9&Issue=10 1. INTRODUCTION The mathematics is the branch where the lots of concepts are included. An orthogonality is the one of the concept among them. Here we focuse on the orthogonal polynomial sequence. The orthogonal polynomial are divided in two classes i.e. classical orthogonal polynomials, Discrete orthogonal polynomials and Sieved orthogonal polynomials .There are different types of classical orthogonal polynomials such that Jacobi polynomials, Associated Laguerre polynomials and Hermite polynomials.
    [Show full text]
  • Orthogonal Polynomials: an Illustrated Guide
    Orthogonal Polynomials: An Illustrated Guide Avaneesh Narla December 10, 2018 Contents 1 Definitions 1 2 Example 1 2 3 Three-term Recurrence Relation 3 4 Christoffel-Darboux Formula 5 5 Zeros 6 6 Gauss Quadrature 8 6.1 Lagrange Interpolation . .8 6.2 Gauss quadrature formula . .8 7 Classical Orthogonal Polynomials 11 7.1 Hermite Polynomials . 11 7.2 Laguerre Polynomials . 12 7.3 Legendre Polynomials . 14 7.4 Jacobi Polynomials . 16 7.5 Chebyshev Polynomials of the First Kind . 17 7.6 Chebyshev Polynomials of the Second Kind . 19 7.7 Gegenbauer polynomials . 20 1 Definitions Orthogonal polynomials are orthogonal with respect to a certain function, known as the weight function w(x), and a defined interval. The weight function must be continuous and positive such that its moments (µn) exist. Z b n µn := w(x)x dx; n = 0; 1; 2; ::: a The interval may be infinite. We now define the inner product of two polynomials as follows Z 1 hf; giw(x) := w(x)f(x)g(x) dx −∞ 1 We will drop the subscript indicating the weight function in future cases. Thus, as always, a 1 sequence of polynomials fpn(x)gn=0 with deg(pn(x)) = n are called orthogonal polynomials for a weight function w if hpm; pni = hnδmn Above, the delta function is the Kronecker Delta Function There are two possible normalisations: If hn = 1 8n 2 f0; 1; 2:::g, the sequence is orthonormal. If the coefficient of highest degree term is 1 for all elements in the sequence, the sequence is monic.
    [Show full text]