Orthogonal Polynomials: an Illustrated Guide

Total Page:16

File Type:pdf, Size:1020Kb

Orthogonal Polynomials: an Illustrated Guide Orthogonal Polynomials: An Illustrated Guide Avaneesh Narla December 10, 2018 Contents 1 Definitions 1 2 Example 1 2 3 Three-term Recurrence Relation 3 4 Christoffel-Darboux Formula 5 5 Zeros 6 6 Gauss Quadrature 8 6.1 Lagrange Interpolation . .8 6.2 Gauss quadrature formula . .8 7 Classical Orthogonal Polynomials 11 7.1 Hermite Polynomials . 11 7.2 Laguerre Polynomials . 12 7.3 Legendre Polynomials . 14 7.4 Jacobi Polynomials . 16 7.5 Chebyshev Polynomials of the First Kind . 17 7.6 Chebyshev Polynomials of the Second Kind . 19 7.7 Gegenbauer polynomials . 20 1 Definitions Orthogonal polynomials are orthogonal with respect to a certain function, known as the weight function w(x), and a defined interval. The weight function must be continuous and positive such that its moments (µn) exist. Z b n µn := w(x)x dx; n = 0; 1; 2; ::: a The interval may be infinite. We now define the inner product of two polynomials as follows Z 1 hf; giw(x) := w(x)f(x)g(x) dx −∞ 1 We will drop the subscript indicating the weight function in future cases. Thus, as always, a 1 sequence of polynomials fpn(x)gn=0 with deg(pn(x)) = n are called orthogonal polynomials for a weight function w if hpm; pni = hnδmn Above, the delta function is the Kronecker Delta Function There are two possible normalisations: If hn = 1 8n 2 f0; 1; 2:::g, the sequence is orthonormal. If the coefficient of highest degree term is 1 for all elements in the sequence, the sequence is monic. We note that orthogonal polynomials are thus linearly independent and generate bases for all poly- nomials of any arbitrary degree. Important Immediate Properties: 1. hQ; pniw(x) = ckpn and hQ; pniw(x) 6= 0 =) Q has a term of degree n. Question 1. Can you have two distinct sets of orthogonal polynomials corresponding to the same weight function but different intervals? What about the same interval? Is there any relation be- tween the two distinct sets? Also, do all weight functions have corresponding sequence of orthogonal polynomials? If not, what are the defining properties of the weight functions that do? What additional restrictions do you need on w(x) to make the inner product space complete? Answers: For two sets of orthogonal polynomials corresponding to the same weight function but different intervals, look at the example set and the Legendre Polynomials. For two sets of orthogonal polynomials corresponding to the same weight function but the same interval, look at the two kinds of Chebyshev Polynomials. 2 Example 1 We choose the weight function to be a constant (say c), and choose the interval (0,a). We can construct the sequence through the Gram-Schmidt process (by moving from the basis f1; x; x2; x3:::g 2 to the orthogonal basis). We start with the sequence f1; x; x ; :::g. We choose p0(x) = 1 for now, we can later normalise the sequence to be orthonormal. Thus by the Gram-Schmidt Process, Proof. hx; p0(x)i hx; 1i a p1(x) = x − projp0(x)x = x − p0(x) = x − = x − hp0(x); p0(x)i h1; 1i 2 R a ca2 R a Since, hx; 1i = 0 c · x dx = 2 and h1; 1i = 0 cdx = ca 2 2 2 2 2 2 hx ; p0(x)i hx ; p1(x)i =) p2(x) = x − projp0(x)x − projp1(x)x = x − p0(x) − p1(x) = hp0(x); p0(x)i hp1(x); p1(x)i 2 2 a 3 4 2 2 2 2 hx ; 1i hx ; x − 2 i a 2 ca 1 ca 12 a 2 a a 2 a x − − a a x − = x − · − · 3 x − = x − −ax+ = x −ax+ h1; 1i hx − 2 ; x − 2 i 2 3 ca 12 ca 2 3 2 6 as, Z a ca3 hx2; 1i = c · x2 dx = 0 3 2 Z a 4 3 a 4 4 4 2 2 cx cax ca ca ca hx ; (x − a=2)i = cx · (x − a=2) dx = − = − = 0 4 6 0 4 6 12 Z a 3 2 2 a 3 3 3 3 2 cx cax ca x ca ca ca ca h(x − a=2); (x − a=2)i = c(x − a=2) dx = − + = − + = 0 3 2 4 0 3 2 4 12 And similarly by continuing the Gram-Schmidt process, we get that 3a 3a2 a3 9a2 2a3 a4 p (x) = x3 − x2 + x − ; p (x) = x4 − 2ax3 + x2 − x + 3 2 5 20 4 7 7 70 5a 20a2 5a3 5a4 a5 p (x) = x5 − x4 + x3 − x2 + x − 5 2 9 6 42 252 p R a 2 The orthonormal polynomials would be qn(x) = pn(x)= hn, where hn = 0 cpn(x) dx Question 2. What is the general expression for pn(x) in the above example? What about hn(x)? Is this a complete inner product space? 3 Three-term Recurrence Relation 1 Theorem 1. A sequence of orthogonal polynomials fpn(x)gn=0 satisfies pn+1(x) = (Anx + Bn)pn(x) + Cnpn−1(x); n = 1; 2; 3::: 3 where, kn+1 An hn An = ; n = 0; 1; 2; 3::: and Cn = − · ; n = 1; 2; 3::: kn An−1 hn−1 where kn is the leading coefficient of pn(x) Proof. We first note that the degree of pn+1(x) is the same as the degree of xpn(x). Thus, by defining the coefficients of the leading terms of each polynomial, we note that the leading term of p (x) is k xn+1, which is equal to the leading term of kn+1 xp := A xp . n+1 n+1 kn n n n =) hpn+1; xpn(x)i = An (1) Now we consider the lower order terms. pn+1 − Anxpn is a polynomial of degree ≤ n and n X pn+1 − Anxpn = ckpk(x) (2) k=0 This is because orthogonal polynomials are the basis for all polynomials. The orthogonality prop- erty now gives us that, n X hpn+1(x) − Anxpn(x); pk(x)i = cmhpm(x); pk(x)i = ckhpk(x); pk(x)i = hkck (3) m=0 = hpn+1(x); pk(x)i − hAnxpn(x); pk(x)i = −Anhpn(x); xpk(x)i (4) −An =) ck = hpn(x); xpk(x)i hk But, hpn(x); xpk(x)i = 0 for k < n − 1. Thus, 8k < n − 1; ck = 0 Therefore, from (2) 1 1 =) pn+1 − Anxpn = −An hpn(x); xpn(x)i · pn(x) + hpn(x); xpn−1(x)i · pn−1(x) hn hn−1 Or, equivalently pn+1(x) − Anxpn(x) = cnpn(x) + cn−1pn−1(x); n = 1; 2; 3::: Further, by using the same argument as (1), we have that kn−1 An hn hn−1cn−1 = −Anhpn(x); xpn−1(x)i = −An hn =) cn−1 = − · kn An−1 hn−1 The term B in Th. 1 can be represented as −An hp (x); xp (x)i but the inner product cannot n hn n n be put in terms of coefficients as simply as the other terms. Also, we note that if we represent the polynomials as monic polynomials, we get the simpler representation, hn pn+1(x) = xpn(x) + Bnpn(x) + Cnpn−1(x) where Cn = − ; n = 1; 2; 3::: hn−1 If they are orthogonal polynomials, we get the representation, hn+1 An pn+1(x) = ((Anx + Bn)pn(x) + Cnpn−1(x)) where Cn = − ; n = 1; 2; 3::: hn An−1 Question 3. Which is more computationally viable, the Gram-Schmidt Process or the Three-Term Recurrence Relation? Is there any other algorithm to do it better (like FFT)? 4 4 Christoffel-Darboux Formula As a consequence of the Three-Term Recurrence Relation, we have that 1 Theorem 2. A sequence of orthogonal polynomials fpn(x)gn=0 satisfies n X pk(x)pk(y) kn pn+1(x)pn(y) − pn+1(y)pn(x) = · ; n = 0; 1; 2::: hk hnkn+1 x − y k=0 and as y ! x n 2 X pk(x) kn 0 0 = · (pn+1(x)pn(x) − pn+1(x)pn(x)); n = 0; 1; 2::: hk hnkn+1 k=0 Proof. The three term recurrence relation implies that pn+1(x)pn(y) = (Anx + Bn)pn(x)pn(y) + Cnpn−1(x)pn(y); n = 1; 2; 3::: and pn+1(y)pn(x) = (Anx + Bn)pn(y)pn(x) + Cnpn−1(y)pn(x); n = 1; 2; 3::: Subtracting the two equations above, we have pn+1(x)pn(y) − pn+1(y)pn(x) = An(x − y)pn(x)pn(y) + Cn[pn−1(x)pn(y) − pn−1(y)pn(x)] =) An(x − y)pn(x)pn(y) = Cnpn−1(x)pn(y) − Cnpn−1(y)pn(x) − pn+1(x)pn(y) + pn+1(y)pn(x) Dividing all through by Anhn, we get (x − y)p (x)p (y) −p (y)p (x) + p (x)p (y) −p (y)p (x) + p (x)p (y) =) n n = n−1 n n−1 n + n+1 n n+1 n hn An−1hn−1 Anhn n n n X (x − y)pk(x)pk(y) X −pk−1(y)pk(x) + pk−1(x)pk(y) X −pk+1(y)pk(x) + pk+1(x)pk(y) =) = + hk Ak−1hk−1 Akhk k=1 k=1 k=1 By reordering, n n n X pk(x)pk(y) X pk+1(x)pk(y) − pk(x)pk+1(y) X pk(x)pk−1(y) − pk−1(x)pk(y) =) (x − y) = − hk Akhk Ak−1hk−1 k=1 k=1 k=1 By eliminating all the common terms in the two sums on the RHS, n X pk(x)pk(y) pn+1(x)pn(y) − pn(x)pn+1(y) p1(x)p0(y) − p0(x)p1(y) =) (x − y) = − hk Anhn A0h0 k=1 p (x)p (y) − p (x)p (y) k (p (x)k − k p (y)) = n+1 n n n+1 − 0 1 0 0 1 Anhn k1h0 as A = k1 and p (x) = p (y) = k 0 k0 0 0 0 k2(x−y) Further, p (x) − p (y) = k (x − y) =) k0(p1(x)k0−k0p1(y)) = 0 = (x − y) p0(x)p0(y) 1 1 1 k1h0 h0 h0 n X pk(x)pk(y) pn+1(x)pn(y) − pn(x)pn+1(y) =) (x − y) = hk Anhn k=0 5 n X pk(x)pk(y) kn =) = (pn+1(x)pn(y) − pn(x)pn+1(y)) hk (x − y)kn+1hn k=0 Thus, we prove the Christoffel-Darboux formula.
Recommended publications
  • The Matroid Theorem We First Review Our Definitions: a Subset System Is A
    CMPSCI611: The Matroid Theorem Lecture 5 We first review our definitions: A subset system is a set E together with a set of subsets of E, called I, such that I is closed under inclusion. This means that if X ⊆ Y and Y ∈ I, then X ∈ I. The optimization problem for a subset system (E, I) has as input a positive weight for each element of E. Its output is a set X ∈ I such that X has at least as much total weight as any other set in I. A subset system is a matroid if it satisfies the exchange property: If i and i0 are sets in I and i has fewer elements than i0, then there exists an element e ∈ i0 \ i such that i ∪ {e} ∈ I. 1 The Generic Greedy Algorithm Given any finite subset system (E, I), we find a set in I as follows: • Set X to ∅. • Sort the elements of E by weight, heaviest first. • For each element of E in this order, add it to X iff the result is in I. • Return X. Today we prove: Theorem: For any subset system (E, I), the greedy al- gorithm solves the optimization problem for (E, I) if and only if (E, I) is a matroid. 2 Theorem: For any subset system (E, I), the greedy al- gorithm solves the optimization problem for (E, I) if and only if (E, I) is a matroid. Proof: We will show first that if (E, I) is a matroid, then the greedy algorithm is correct. Assume that (E, I) satisfies the exchange property.
    [Show full text]
  • Chebyshev Polynomials of the Second, Third and Fourth Kinds in Approximation, Indefinite Integration, and Integral Transforms *
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Journal of Computational and Applied Mathematics 49 (1993) 169-178 169 North-Holland CAM 1429 Chebyshev polynomials of the second, third and fourth kinds in approximation, indefinite integration, and integral transforms * J.C. Mason Applied and Computational Mathematics Group, Royal Military College of Science, Shrivenham, Swindon, Wiltshire, United Kingdom Dedicated to Dr. D.F. Mayers on the occasion of his 60th birthday Received 18 February 1992 Revised 31 March 1992 Abstract Mason, J.C., Chebyshev polynomials of the second, third and fourth kinds in approximation, indefinite integration, and integral transforms, Journal of Computational and Applied Mathematics 49 (1993) 169-178. Chebyshev polynomials of the third and fourth kinds, orthogonal with respect to (l+ x)‘/‘(l- x)-‘I* and (l- x)‘/*(l+ x)-‘/~, respectively, on [ - 1, 11, are less well known than traditional first- and second-kind polynomials. We therefore summarise basic properties of all four polynomials, and then show how some well-known properties of first-kind polynomials extend to cover second-, third- and fourth-kind polynomials. Specifically, we summarise a recent set of first-, second-, third- and fourth-kind results for near-minimax constrained approximation by series and interpolation criteria, then we give new uniform convergence results for the indefinite integration of functions weighted by (1 + x)-i/* or (1 - x)-l/* using third- or fourth-kind polynomial expansions, and finally we establish a set of logarithmically singular integral transforms for which weighted first-, second-, third- and fourth-kind polynomials are eigenfunctions.
    [Show full text]
  • CONTINUITY in the ALEXIEWICZ NORM Dedicated to Prof. J
    131 (2006) MATHEMATICA BOHEMICA No. 2, 189{196 CONTINUITY IN THE ALEXIEWICZ NORM Erik Talvila, Abbotsford (Received October 19, 2005) Dedicated to Prof. J. Kurzweil on the occasion of his 80th birthday Abstract. If f is a Henstock-Kurzweil integrable function on the real line, the Alexiewicz norm of f is kfk = sup j I fj where the supremum is taken over all intervals I ⊂ . Define I the translation τx by τxfR(y) = f(y − x). Then kτxf − fk tends to 0 as x tends to 0, i.e., f is continuous in the Alexiewicz norm. For particular functions, kτxf − fk can tend to 0 arbitrarily slowly. In general, kτxf − fk > osc fjxj as x ! 0, where osc f is the oscillation of f. It is shown that if F is a primitive of f then kτxF − F k kfkjxj. An example 1 6 1 shows that the function y 7! τxF (y) − F (y) need not be in L . However, if f 2 L then kτxF − F k1 6 kfk1jxj. For a positive weight function w on the real line, necessary and sufficient conditions on w are given so that k(τxf − f)wk ! 0 as x ! 0 whenever fw is Henstock-Kurzweil integrable. Applications are made to the Poisson integral on the disc and half-plane. All of the results also hold with the distributional Denjoy integral, which arises from the completion of the space of Henstock-Kurzweil integrable functions as a subspace of Schwartz distributions. Keywords: Henstock-Kurzweil integral, Alexiewicz norm, distributional Denjoy integral, Poisson integral MSC 2000 : 26A39, 46Bxx 1.
    [Show full text]
  • Cardinality Constrained Combinatorial Optimization: Complexity and Polyhedra
    Takustraße 7 Konrad-Zuse-Zentrum D-14195 Berlin-Dahlem fur¨ Informationstechnik Berlin Germany RUDIGER¨ STEPHAN1 Cardinality Constrained Combinatorial Optimization: Complexity and Polyhedra 1Email: [email protected] ZIB-Report 08-48 (December 2008) Cardinality Constrained Combinatorial Optimization: Complexity and Polyhedra R¨udigerStephan Abstract Given a combinatorial optimization problem and a subset N of natural numbers, we obtain a cardinality constrained version of this problem by permitting only those feasible solutions whose cardinalities are elements of N. In this paper we briefly touch on questions that addresses common grounds and differences of the complexity of a combinatorial optimization problem and its cardinality constrained version. Afterwards we focus on polytopes associated with cardinality constrained combinatorial optimiza- tion problems. Given an integer programming formulation for a combina- torial optimization problem, by essentially adding Gr¨otschel’s cardinality forcing inequalities [11], we obtain an integer programming formulation for its cardinality restricted version. Since the cardinality forcing inequal- ities in their original form are mostly not facet defining for the associated polyhedra, we discuss possibilities to strengthen them. In [13] a variation of the cardinality forcing inequalities were successfully integrated in the system of linear inequalities for the matroid polytope to provide a com- plete linear description of the cardinality constrained matroid polytope. We identify this polytope as a master polytope for our class of problems, since many combinatorial optimization problems can be formulated over the intersection of matroids. 1 Introduction, Basics, and Complexity Given a combinatorial optimization problem and a subset N of natural numbers, we obtain a cardinality constrained version of this problem by permitting only those feasible solutions whose cardinalities are elements of N.
    [Show full text]
  • Generalizations of Chebyshev Polynomials and Polynomial Mappings
    TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 359, Number 10, October 2007, Pages 4787–4828 S 0002-9947(07)04022-6 Article electronically published on May 17, 2007 GENERALIZATIONS OF CHEBYSHEV POLYNOMIALS AND POLYNOMIAL MAPPINGS YANG CHEN, JAMES GRIFFIN, AND MOURAD E.H. ISMAIL Abstract. In this paper we show how polynomial mappings of degree K from a union of disjoint intervals onto [−1, 1] generate a countable number of spe- cial cases of generalizations of Chebyshev polynomials. We also derive a new expression for these generalized Chebyshev polynomials for any genus g, from which the coefficients of xn can be found explicitly in terms of the branch points and the recurrence coefficients. We find that this representation is use- ful for specializing to polynomial mapping cases for small K where we will have explicit expressions for the recurrence coefficients in terms of the branch points. We study in detail certain special cases of the polynomials for small degree mappings and prove a theorem concerning the location of the zeroes of the polynomials. We also derive an explicit expression for the discrimi- nant for the genus 1 case of our Chebyshev polynomials that is valid for any configuration of the branch point. 1. Introduction and preliminaries Akhiezer [2], [1] and, Akhiezer and Tomˇcuk [3] introduced orthogonal polynomi- als on two intervals which generalize the Chebyshev polynomials. He observed that the study of properties of these polynomials requires the use of elliptic functions. In thecaseofmorethantwointervals,Tomˇcuk [17], investigated their Bernstein-Szeg˝o asymptotics, with the theory of Hyperelliptic integrals, and found expressions in terms of a certain Abelian integral of the third kind.
    [Show full text]
  • Some Properties of AP Weight Function
    Journal of the Institute of Engineering, 2016, 12(1): 210-213 210 TUTA/IOE/PCU © TUTA/IOE/PCU Printed in Nepal Some Properties of AP Weight Function Santosh Ghimire Department of Engineering Science and Humanities, Institute of Engineering Pulchowk Campus, Tribhuvan University, Kathmandu, Nepal Corresponding author: [email protected] Received: June 20, 2016 Revised: July 25, 2016 Accepted: July 28, 2016 Abstract: In this paper, we briefly discuss the theory of weights and then define A1 and Ap weight functions. Finally we prove some of the properties of AP weight function. Key words: A1 weight function, Maximal functions, Ap weight function. 1. Introduction The theory of weights play an important role in various fields such as extrapolation theory, vector-valued inequalities and estimates for certain class of non linear differential equation. Moreover, they are very useful in the study of boundary value problems for Laplace's equation in Lipschitz domains. In 1970, Muckenhoupt characterized positive functions w for which the Hardy-Littlewood maximal operator M maps Lp(Rn, w(x)dx) to itself. Muckenhoupt's characterization actually gave the better understanding of theory of weighted inequalities which then led to the introduction of Ap class and consequently the development of weighted inequalities. 2. Definitions n Definition: A locally integrable function on R that takes values in the interval (0,∞) almost everywhere is called a weight. So by definition a weight function can be zero or infinity only on a set whose Lebesgue measure is zero. We use the notation to denote the w-measure of the set E and we reserve the notation Lp(Rn,w) or Lp(w) for the weighted L p spaces.
    [Show full text]
  • Chebyshev and Fourier Spectral Methods 2000
    Chebyshev and Fourier Spectral Methods Second Edition John P. Boyd University of Michigan Ann Arbor, Michigan 48109-2143 email: [email protected] http://www-personal.engin.umich.edu/jpboyd/ 2000 DOVER Publications, Inc. 31 East 2nd Street Mineola, New York 11501 1 Dedication To Marilyn, Ian, and Emma “A computation is a temptation that should be resisted as long as possible.” — J. P. Boyd, paraphrasing T. S. Eliot i Contents PREFACE x Acknowledgments xiv Errata and Extended-Bibliography xvi 1 Introduction 1 1.1 Series expansions .................................. 1 1.2 First Example .................................... 2 1.3 Comparison with finite element methods .................... 4 1.4 Comparisons with Finite Differences ....................... 6 1.5 Parallel Computers ................................. 9 1.6 Choice of basis functions .............................. 9 1.7 Boundary conditions ................................ 10 1.8 Non-Interpolating and Pseudospectral ...................... 12 1.9 Nonlinearity ..................................... 13 1.10 Time-dependent problems ............................. 15 1.11 FAQ: Frequently Asked Questions ........................ 16 1.12 The Chrysalis .................................... 17 2 Chebyshev & Fourier Series 19 2.1 Introduction ..................................... 19 2.2 Fourier series .................................... 20 2.3 Orders of Convergence ............................... 25 2.4 Convergence Order ................................. 27 2.5 Assumption of Equal Errors ...........................
    [Show full text]
  • Dymore User's Manual Chebyshev Polynomials
    Dymore User's Manual Chebyshev polynomials Olivier A. Bauchau August 27, 2019 Contents 1 Definition 1 1.1 Zeros and extrema....................................2 1.2 Orthogonality relationships................................3 1.3 Derivatives of Chebyshev polynomials..........................5 1.4 Integral of Chebyshev polynomials............................5 1.5 Products of Chebyshev polynomials...........................5 2 Chebyshev approximation of functions of a single variable6 2.1 Expansion of a function in Chebyshev polynomials...................6 2.2 Evaluation of Chebyshev expansions: Clenshaw's recurrence.............7 2.3 Derivatives and integrals of Chebyshev expansions...................7 2.4 Products of Chebyshev expansions...........................8 2.5 Examples.........................................9 2.6 Clenshaw-Curtis quadrature............................... 10 3 Chebyshev approximation of functions of two variables 12 3.1 Expansion of a function in Chebyshev polynomials................... 12 3.2 Evaluation of Chebyshev expansions: Clenshaw's recurrence............. 13 3.3 Derivatives of Chebyshev expansions.......................... 14 4 Chebychev polynomials 15 4.1 Examples......................................... 16 1 Definition Chebyshev polynomials [1,2] form a series of orthogonal polynomials, which play an important role in the theory of approximation. The lowest polynomials are 2 3 4 2 T0(x) = 1;T1(x) = x; T2(x) = 2x − 1;T3(x) = 4x − 3x; T4(x) = 8x − 8x + 1;::: (1) 1 and are depicted in fig.1. The polynomials can be generated from the following recurrence rela- tionship Tn+1 = 2xTn − Tn−1; n ≥ 1: (2) 1 0.8 0.6 0.4 0.2 0 −0.2 −0.4 CHEBYSHEV POLYNOMIALS −0.6 −0.8 −1 −1 −0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8 1 XX Figure 1: The seven lowest order Chebyshev polynomials It is possible to give an explicit expression of Chebyshev polynomials as Tn(x) = cos(n arccos x): (3) This equation can be verified by using elementary trigonometric identities.
    [Show full text]
  • Approximation Atkinson Chapter 4, Dahlquist & Bjork Section 4.5
    Approximation Atkinson Chapter 4, Dahlquist & Bjork Section 4.5, Trefethen's book Topics marked with ∗ are not on the exam 1 In approximation theory we want to find a function p(x) that is `close' to another function f(x). We can define closeness using any metric or norm, e.g. Z 2 2 kf(x) − p(x)k2 = (f(x) − p(x)) dx or kf(x) − p(x)k1 = sup jf(x) − p(x)j or Z kf(x) − p(x)k1 = jf(x) − p(x)jdx: In order for these norms to make sense we need to restrict the functions f and p to suitable function spaces. The polynomial approximation problem takes the form: Find a polynomial of degree at most n that minimizes the norm of the error. Naturally we will consider (i) whether a solution exists and is unique, (ii) whether the approximation converges as n ! 1. In our section on approximation (loosely following Atkinson, Chapter 4), we will first focus on approximation in the infinity norm, then in the 2 norm and related norms. 2 Existence for optimal polynomial approximation. Theorem (no reference): For every n ≥ 0 and f 2 C([a; b]) there is a polynomial of degree ≤ n that minimizes kf(x) − p(x)k where k · k is some norm on C([a; b]). Proof: To show that a minimum/minimizer exists, we want to find some compact subset of the set of polynomials of degree ≤ n (which is a finite-dimensional space) and show that the inf over this subset is less than the inf over everything else.
    [Show full text]
  • 32 FA15 Abstracts
    32 FA15 Abstracts IP1 metric simple exclusion process and the KPZ equation. In Vector-Valued Nonsymmetric and Symmetric Jack addition, the experiments of Takeuchi and Sano will be and Macdonald Polynomials briefly discussed. For each partition τ of N there are irreducible representa- Craig A. Tracy tions of the symmetric group SN and the associated Hecke University of California, Davis algebra HN (q) on a real vector space Vτ whose basis is [email protected] indexed by the set of reverse standard Young tableaux of shape τ. The talk concerns orthogonal bases of Vτ -valued polynomials of x ∈ RN . The bases consist of polyno- IP6 mials which are simultaneous eigenfunctions of commuta- Limits of Orthogonal Polynomials and Contrac- tive algebras of differential-difference operators, which are tions of Lie Algebras parametrized by κ and (q, t) respectively. These polynomi- als reduce to the ordinary Jack and Macdonald polynomials In this talk, I will discuss the connection between superin- when the partition has just one part (N). The polynomi- tegrable systems and classical systems of orthogonal poly- als are constructed by means of the Yang-Baxter graph. nomials in particular in the expansion coefficients between There is a natural bilinear form, which is positive-definite separable coordinate systems, related to representations of for certain ranges of parameter values depending on τ,and the (quadratic) symmetry algebras. This connection al- there are integral kernels related to the bilinear form for lows us to extend the Askey scheme of classical orthogonal the group case, of Gaussian and of torus type. The mate- polynomials and the limiting processes within the scheme.
    [Show full text]
  • Orthogonal Functions: the Legendre, Laguerre, and Hermite Polynomials
    ORTHOGONAL FUNCTIONS: THE LEGENDRE, LAGUERRE, AND HERMITE POLYNOMIALS THOMAS COVERSON, SAVARNIK DIXIT, ALYSHA HARBOUR, AND TYLER OTTO Abstract. The Legendre, Laguerre, and Hermite equations are all homogeneous second order Sturm-Liouville equations. Using the Sturm-Liouville Theory we will be able to show that polynomial solutions to these equations are orthogonal. In a more general context, finding that these solutions are orthogonal allows us to write a function as a Fourier series with respect to these solutions. 1. Introduction The Legendre, Laguerre, and Hermite equations have many real world practical uses which we will not discuss here. We will only focus on the methods of solution and use in a mathematical sense. In solving these equations explicit solutions cannot be found. That is solutions in in terms of elementary functions cannot be found. In many cases it is easier to find a numerical or series solution. There is a generalized Fourier series theory which allows one to write a function f(x) as a linear combination of an orthogonal system of functions φ1(x),φ2(x),...,φn(x),... on [a; b]. The series produced is called the Fourier series with respect to the orthogonal system. While the R b a f(x)φn(x)dx coefficients ,which can be determined by the formula cn = R b 2 , a φn(x)dx are called the Fourier coefficients with respect to the orthogonal system. We are concerned only with showing that the Legendre, Laguerre, and Hermite polynomial solutions are orthogonal and can thus be used to form a Fourier series. In order to proceed we must define an inner product and define what it means for a linear operator to be self- adjoint.
    [Show full text]
  • Orthogonal Polynomials on the Unit Circle Associated with the Laguerre Polynomials
    PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 129, Number 3, Pages 873{879 S 0002-9939(00)05821-4 Article electronically published on October 11, 2000 ORTHOGONAL POLYNOMIALS ON THE UNIT CIRCLE ASSOCIATED WITH THE LAGUERRE POLYNOMIALS LI-CHIEN SHEN (Communicated by Hal L. Smith) Abstract. Using the well-known fact that the Fourier transform is unitary, we obtain a class of orthogonal polynomials on the unit circle from the Fourier transform of the Laguerre polynomials (with suitable weights attached). Some related extremal problems which arise naturally in this setting are investigated. 1. Introduction This paper deals with a class of orthogonal polynomials which arise from an application of the Fourier transform on the Laguerre polynomials. We shall briefly describe the essence of our method. Let Π+ denote the upper half plane fz : z = x + iy; y > 0g and let Z 1 H(Π+)=ff : f is analytic in Π+ and sup jf(x + yi)j2 dx < 1g: 0<y<1 −∞ It is well known that, from the Paley-Wiener Theorem [4, p. 368], the Fourier transform provides a unitary isometry between the spaces L2(0; 1)andH(Π+): Since the Laguerre polynomials form a complete orthogonal basis for L2([0; 1);xαe−x dx); the application of Fourier transform to the Laguerre polynomials (with suitable weight attached) generates a class of orthogonal rational functions which are com- plete in H(Π+); and by composition of which with the fractional linear transfor- mation (which maps Π+ conformally to the unit disc) z =(2t − i)=(2t + i); we obtain a family of polynomials which are orthogonal with respect to the weight α t j j sin 2 dt on the boundary z = 1 of the unit disc.
    [Show full text]