Projection and Its Importance in Scientific Computing ______

Total Page:16

File Type:pdf, Size:1020Kb

Projection and Its Importance in Scientific Computing ______ Projection and its Importance in Scientific Computing ________________________________________________ Stan Tomov EECS Department The University of Tennessee, Knoxville March 22, 2017 COCS 594 Lecture Notes Slide 1 / 41 03/22/2017 Contact information office : Claxton 317 phone : (865) 974-6317 email : [email protected] Additional reference materials: [1] R.Barrett, M.Berry, T.F.Chan, J.Demmel, J.Donato, J. Dongarra, V. Eijkhout, R.Pozo, C.Romine, and H.Van der Vorst, Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods (2nd edition) http://netlib2.cs.utk.edu/linalg/html_templates/Templates.html [2] Yousef Saad, Iterative methods for sparse linear systems (1st edition) http://www-users.cs.umn.edu/~saad/books.html Slide 2 / 41 Topics as related to high-performance scientific computing‏ Projection in Scientific Computing Sparse matrices, PDEs, Numerical parallel implementations solution, Tools, etc. Iterative Methods Slide 3 / 41 Topics on new architectures – multicore, GPUs (CUDA & OpenCL), MIC Projection in Scientific Computing Sparse matrices, PDEs, Numerical parallel implementations solution, Tools, etc. Iterative Methods Slide 4 / 41 Outline Part I – Fundamentals Part II – Projection in Linear Algebra Part III – Projection in Functional Analysis (e.g. PDEs) HPC with Multicore and GPUs Slide 5 / 41 Part I Fundamentals Slide 6 / 41 Projection in Scientific Computing [ an example – in solvers for PDE discretizations ] A model leading to self-consistent iteration with need for high-performance diagonalization and orthogonalization routines Slide 7 / 41 What is Projection? Here are two examples (from linear algebra) (from functional analysis) u u = f(x) The error (u – Pu) to be Pu orthogonalThe error to(u vector– Pu) toe be orthogonal to vector e 1 e = 1 Pu e 0 1 P : orthogonal projection of P : best approximation (projection) of f(x) vector u on e in span{ e } ⊂ C[0,1] Slide 8 / 41 Definition Projection is a linear transformation P from a linear space V to itself such that P2 = P equivalently Let V is direct sum of subspaces V1 and V2 V = V1 V2 i.e. for u ∈ V there are unique u1∈ V1 and u2∈V2 s.t. u = u1+u2 Then P: VV1 is defined for u ∈ V as Pu ≡ u1 Slide 9 / 41 Importance in Scientific Computing To compute approximations Pu u where dim V1 << dim V V = V1 V2 When computation directly in V is not feasible or even possible. A few examples: – Interpolation (via polynomial interpolation/projection) – Image compression – Sparse iterative linear solvers and eigensolvers – Finite Element/Volume Method approximations – Least-squares approximations, etc. Slide 10 / 41 Projection in R2 In R2 with Euclidean inner-product, i.e. for x, y ∈ R2 T (x, y) = x1 y1 + x2 y2 ( = y x = x⋅y ) and || x || = (x, x)1/2 u (u, e) Pu = e (Exercise) || e ||2 e Pu i.e. for || e || = 1 P : orthogonal projection of Pu = (u, e) e vector u on e Slide 11 / 41 Projection in Rn / Cn Similarly to R2 P : Orthogonal projection of u into span{e1, ... , em}, m ≤ n. Let ei, i = 1 ... m is orthonormal basis, i.e. (ei, ej) = 0 for i≠j and (ei, ej) = 1 for i=j P u = (u, e1) e1 + ... + (u, em) em ( Exercise ) Orthogonal projection of u on e1 Slide 12 / 41 How to get an orthonormal basis? Can get one from every subspace by Gram-Schmidt orthogonalization: Input : m linearly independent vectors x1, ..., xm Output : m orthonormal vectors x1, ..., xm 1. x1 = x1 / || x1 || 2. do i = 2, m 3. xi = xi - (xi, x1) x1 - ... - (xi, xi-1) xi-1 (Exercise: xi x1 , ... , xi-1) 4. xi = xi / || xi || Orthogonal projection of x on x 5. enddo i 1 Known as Classical Gram-Schmidt (CGM) orthogonalization Slide 13 / 41 How to get an orthonormal basis? What if we replace line 3 with the following (3')? 3. xi = xi - (xi, x1) x1 - ... - (xi, xi-1) xi-1 3'. do j = 1, i-1 xi xi = xi – (xi, xj) xj enddo Equivalent in exact arithmetic (Exercise) but not with round-off errors (next) ! Known as Modified Gram-Schmidt (MGS) orthogonalization Slide 14 / 41 CGS vs MGS [ Results from Julien Langou: ] Scalability of MGS and CGS on two different clusters for matrices of various size m=[500 1000 2000 4000] per processor, n = 100 0.35 3.5 0.3 3 MGS 0500 0.25 2.5 MGS 1000 MGS 0500 ) MGS 2000 ) c s e ( 0.2 s 2 MGS 1000 ( MGS 4000 e e m i 0.15 CGS 0500 1.5 CGS 0500 m t i 0.1 CGS 1000 t 1 CGS 1000 CGS 2000 0.05 0.5 CGS 4000 0 0 1 2 4 8 16 32 64 1 2 4 8 16 32 # procs # procs Slide 15 / 41 || I - Q^TQ || [ ResultsJulienLangou:] [ from 1.00E+01 1.00E+03 1.00E-15 1.00E-13 1.00E-11 1.00E-09 1.00E-07 1.00E-05 1.00E-03 1.00E-01 Accuracy ofMGS Accuracy 1.00E+01 1.00E+02 1.00E+03 1.00E+04 vs K(A) 1.00E+05 CGS on matrices ofincreasingconditionnumber CGSonmatrices 1.00E+06 1.00E+07 1.00E+08 CGS 1.00E+09 1.00E+10 MGS CGS vs ||A - QR||/||A|| MGS 1.00E-19 1.00E-18 1.00E-17 1.00E-16 1.00E-15 1.00E+01 1.00E+02 1.00E+03 1.00E+04 K(A) 1.00E+05 1.00E+06 1.00E+07 Slide Slide 1.00E+08 16 1.00E+09 / 41 / 1.00E+10 QR factorization Let A = [x1, ... , xm] be the input for CGS/MGS and Q = [q1, ... , qm] the output; R : an upper m m triangular matrix defined from the CGR/MGS. Then A = Q R Slide 17 / 41 Other QR factorizations What about the following? [known as Cholesky QR] 1. G = ATA 2. G = L LT (Cholesky factorization) 3. Q = A (LT)-1 Does Q have orthonormal columns (i.e. QTQ=I), i.e. A = QLT to be a QR factorization (Exercise) When is this feasible and how compares to CGS and MGS? Slide 18 / 41 Other QR factorizations Feasible when n >> m Allows efficient parallel implementation: blocking both computation and communication AT A G = Investigate numerically accuracy P1 P2 . and scalability (compare to CGS and MGS) Exercise Slide 19 / 41 How is done in LAPACK? Using Householder reflectors x w H = I – 2 w wT w = ? so that H x1 = e1 Hx w = ... (compute or look at the reference books) Allows us to construct Xk Hk-1 ... H1 X = Q (Exercise( LAPACK implementation : “delayed update” of the trailing matrix + “accumulate transformation” to apply it as BLAS 3 Slide 20 / 41 Part II Projection in Linear Algebra Slide 21 / 41 Projection into general basis How to define projection without orthogonalization of a basis? – Sometimes is not feasible to orthogonalize – Often the case in functional analysis (e.g. Finite Element Method, Finite Volume Method, etc.) where the basis is “linearly independent” functions (more later, and Lecture 2) We saw if X = [x1, ... , xm] is an orthonormal basis (*) P u = (u, x1) x1 + ... + (u, xm) xm How does (*) change if X are just linearly independent ? P u = ? Slide 22 / 41 Projection into a general basis The problem: T Find the coefficients C = (c1 ... cm) in P u = c1 x1 + c2 x2 + . + cm xm = X C so that u – Pu ⊥ span{x1, ..., xm} or ⇔ so that the error e in (1) u = P u + e is ⊥ span{x1, ..., xm} Slide 23 / 41 Projection into a general basis (1) u = P u + e = c1 x1 + c2 x2 + . + cm xm + e Multiply (1) on both sides by “test” vector/function xj (terminology from functional analysis) for j = 1,..., m 0 (u , xj) = c1 (x1 , xj)+ c2 (x2 , xj)+ . + cm (xm , xj) + (e, xj) i.e., m equations for m unknowns In matrix notations (XT X) C = XT u (Exercise) XT Pu = XT u XTX is the so called Gram matrix (nonsingular; why?) => there exists a unique solution C Slide 24 / 41 Normal equations System (XT X) C = XT u is known also as Normal Equations The Method of Normal Equations: Finding the projection (approximation) Pu ≡ XC (approximation) of u in X by solving the Normal Equations system Slide 25 / 41 Least Squares (LS) Equivalently, system P u (XT X) C = XT u gives also the solution of the LS problem min || X C – u || C Rm 2 ∈ 2 2 2 since || v1 – u || = || (v1 – Pu) – e || = || v1 – Pu || + || e || 2 2 ≥ || e || = || Pu – u || for ∀ v1 ∈ V1 Slide 26 / 41 LS nm n Note that the usual notations for LS is: For A∈R , b∈R find min || A x – b || Solving LS with QR xfactorization∈Rm T T Let A = Q R, Q A = R = R1 , Q b = c m 0 d n - m Then 2 T T 2 2 2 || Ax – b || = || Q Ax – Q b || = || R1x – c || + || d || i.e. we get minimum if x is such that R1 x = c Slide 27 / 41 Projection and iterative solvers The problem : Solve Ax = b in Rn Iterative solution: at iteration i extract an approximate n xi from just a subspace V = span[v1, ..., vm] of R How? As on slide 22, impose constraints: n b – Ax subspace W = span[w1,...,wm] of R , i.e. (*) (Ax, wi) = (b, wi) for ∀ wi ∈W= span[w1,...,wm] Conditions (*) known also as Petrov-Galerkin conditions Projection is orthogonal: V and W are the same (Galerkin conditions) or oblique : V and W are different Slide 28 / 41 Matrix representation Let V = [v1, ..., vm], W = [w1,...,wm] m solves i.e.
Recommended publications
  • Orthogonal Projection, Low Rank Approximation, and Orthogonal Bases
    Week 11 Orthogonal Projection, Low Rank Approximation, and Orthogonal Bases 11.1 Opening Remarks 11.1.1 Low Rank Approximation * View at edX 463 Week 11. Orthogonal Projection, Low Rank Approximation, and Orthogonal Bases 464 11.1.2 Outline 11.1. Opening Remarks..................................... 463 11.1.1. Low Rank Approximation............................. 463 11.1.2. Outline....................................... 464 11.1.3. What You Will Learn................................ 465 11.2. Projecting a Vector onto a Subspace........................... 466 11.2.1. Component in the Direction of ............................ 466 11.2.2. An Application: Rank-1 Approximation...................... 470 11.2.3. Projection onto a Subspace............................. 474 11.2.4. An Application: Rank-2 Approximation...................... 476 11.2.5. An Application: Rank-k Approximation...................... 478 11.3. Orthonormal Bases.................................... 481 11.3.1. The Unit Basis Vectors, Again........................... 481 11.3.2. Orthonormal Vectors................................ 482 11.3.3. Orthogonal Bases.................................. 485 11.3.4. Orthogonal Bases (Alternative Explanation).................... 488 11.3.5. The QR Factorization................................ 492 11.3.6. Solving the Linear Least-Squares Problem via QR Factorization......... 493 11.3.7. The QR Factorization (Again)........................... 494 11.4. Change of Basis...................................... 498 11.4.1. The Unit Basis Vectors,
    [Show full text]
  • On Multivariate Interpolation
    On Multivariate Interpolation Peter J. Olver† School of Mathematics University of Minnesota Minneapolis, MN 55455 U.S.A. [email protected] http://www.math.umn.edu/∼olver Abstract. A new approach to interpolation theory for functions of several variables is proposed. We develop a multivariate divided difference calculus based on the theory of non-commutative quasi-determinants. In addition, intriguing explicit formulae that connect the classical finite difference interpolation coefficients for univariate curves with multivariate interpolation coefficients for higher dimensional submanifolds are established. † Supported in part by NSF Grant DMS 11–08894. April 6, 2016 1 1. Introduction. Interpolation theory for functions of a single variable has a long and distinguished his- tory, dating back to Newton’s fundamental interpolation formula and the classical calculus of finite differences, [7, 47, 58, 64]. Standard numerical approximations to derivatives and many numerical integration methods for differential equations are based on the finite dif- ference calculus. However, historically, no comparable calculus was developed for functions of more than one variable. If one looks up multivariate interpolation in the classical books, one is essentially restricted to rectangular, or, slightly more generally, separable grids, over which the formulae are a simple adaptation of the univariate divided difference calculus. See [19] for historical details. Starting with G. Birkhoff, [2] (who was, coincidentally, my thesis advisor), recent years have seen a renewed level of interest in multivariate interpolation among both pure and applied researchers; see [18] for a fairly recent survey containing an extensive bibli- ography. De Boor and Ron, [8, 12, 13], and Sauer and Xu, [61, 10, 65], have systemati- cally studied the polynomial case.
    [Show full text]
  • Inner Product Spaces Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 2, 2007)
    MAT067 University of California, Davis Winter 2007 Inner Product Spaces Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 2, 2007) The abstract definition of vector spaces only takes into account algebraic properties for the addition and scalar multiplication of vectors. For vectors in Rn, for example, we also have geometric intuition which involves the length of vectors or angles between vectors. In this section we discuss inner product spaces, which are vector spaces with an inner product defined on them, which allow us to introduce the notion of length (or norm) of vectors and concepts such as orthogonality. 1 Inner product In this section V is a finite-dimensional, nonzero vector space over F. Definition 1. An inner product on V is a map ·, · : V × V → F (u, v) →u, v with the following properties: 1. Linearity in first slot: u + v, w = u, w + v, w for all u, v, w ∈ V and au, v = au, v; 2. Positivity: v, v≥0 for all v ∈ V ; 3. Positive definiteness: v, v =0ifandonlyifv =0; 4. Conjugate symmetry: u, v = v, u for all u, v ∈ V . Remark 1. Recall that every real number x ∈ R equals its complex conjugate. Hence for real vector spaces the condition about conjugate symmetry becomes symmetry. Definition 2. An inner product space is a vector space over F together with an inner product ·, ·. Copyright c 2007 by the authors. These lecture notes may be reproduced in their entirety for non- commercial purposes. 2NORMS 2 Example 1. V = Fn n u =(u1,...,un),v =(v1,...,vn) ∈ F Then n u, v = uivi.
    [Show full text]
  • Polynomial Interpolation
    Numerical Methods 401-0654 6 Polynomial Interpolation Throughout this chapter n N is an element from the set N if not otherwise specified. ∈ 0 0 6.1 Polynomials D-ITET, ✬ ✩D-MATL Definition 6.1.1. We denote by n n 1 n := R t αnt + αn 1t − + ... + α1t + α0 R: α0, α1,...,αn R . P ∋ 7→ − ∈ ∈ the R-vectorn space of polynomials with degree n. o ≤ ✫ ✪ ✬ ✩ Theorem 6.1.2 (Dimension of space of polynomials). It holds that dim n = n +1 and n P P ⊂ 6.1 C∞(R). p. 192 ✫ ✪ n n 1 ➙ Remark 6.1.1 (Polynomials in Matlab). MATLAB: αnt + αn 1t − + ... + α0 Vector Numerical − Methods (αn, αn 1,...,α0) (ordered!). 401-0654 − △ Remark 6.1.2 (Horner scheme). Evaluation of a polynomial in monomial representation: Horner scheme p(t)= t t t t αnt + αn 1 + αn 2 + + α1 + α0 . ··· − − ··· Code 6.1: Horner scheme, polynomial in MATLAB format 1 function y = polyval (p,x) 2 y = p(1); for i =2: length (p), y = x y+p( i ); end ∗ D-ITET, D-MATL Asymptotic complexity: O(n) Use: MATLAB “built-in”-function polyval(p,x); △ 6.2 p. 193 6.2 Polynomial Interpolation: Theory Numerical Methods 401-0654 Goal: (re-)constructionofapolynomial(function)frompairs of values (fit). ✬ Lagrange polynomial interpolation problem ✩ Given the nodes < t < t < < tn < and the values y ,...,yn R compute −∞ 0 1 ··· ∞ 0 ∈ p n such that ∈P j 0, 1,...,n : p(t )= y . ∀ ∈ { } j j ✫ ✪ D-ITET, D-MATL 6.2 p. 194 6.2.1 Lagrange polynomials Numerical Methods 401-0654 ✬ ✩ Definition 6.2.1.
    [Show full text]
  • Sparse Polynomial Interpolation and Testing
    Sparse Polynomial Interpolation and Testing by Andrew Arnold A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philsophy in Computer Science Waterloo, Ontario, Canada, 2016 © Andrew Arnold 2016 I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. ii Abstract Interpolation is the process of learning an unknown polynomial f from some set of its evalua- tions. We consider the interpolation of a sparse polynomial, i.e., where f is comprised of a small, bounded number of terms. Sparse interpolation dates back to work in the late 18th century by the French mathematician Gaspard de Prony, and was revitalized in the 1980s due to advancements by Ben-Or and Tiwari, Blahut, and Zippel, amongst others. Sparse interpolation has applications to learning theory, signal processing, error-correcting codes, and symbolic computation. Closely related to sparse interpolation are two decision problems. Sparse polynomial identity testing is the problem of testing whether a sparse polynomial f is zero from its evaluations. Sparsity testing is the problem of testing whether f is in fact sparse. We present effective probabilistic algebraic algorithms for the interpolation and testing of sparse polynomials. These algorithms assume black-box evaluation access, whereby the algorithm may specify the evaluation points. We measure algorithmic costs with respect to the number and types of queries to a black-box oracle.
    [Show full text]
  • Basics of Euclidean Geometry
    This is page 162 Printer: Opaque this 6 Basics of Euclidean Geometry Rien n'est beau que le vrai. |Hermann Minkowski 6.1 Inner Products, Euclidean Spaces In a±ne geometry it is possible to deal with ratios of vectors and barycen- ters of points, but there is no way to express the notion of length of a line segment or to talk about orthogonality of vectors. A Euclidean structure allows us to deal with metric notions such as orthogonality and length (or distance). This chapter and the next two cover the bare bones of Euclidean ge- ometry. One of our main goals is to give the basic properties of the transformations that preserve the Euclidean structure, rotations and re- ections, since they play an important role in practice. As a±ne geometry is the study of properties invariant under bijective a±ne maps and projec- tive geometry is the study of properties invariant under bijective projective maps, Euclidean geometry is the study of properties invariant under certain a±ne maps called rigid motions. Rigid motions are the maps that preserve the distance between points. Such maps are, in fact, a±ne and bijective (at least in the ¯nite{dimensional case; see Lemma 7.4.3). They form a group Is(n) of a±ne maps whose corresponding linear maps form the group O(n) of orthogonal transformations. The subgroup SE(n) of Is(n) corresponds to the orientation{preserving rigid motions, and there is a corresponding 6.1. Inner Products, Euclidean Spaces 163 subgroup SO(n) of O(n), the group of rotations.
    [Show full text]
  • Sparse Polynomial Interpolation Codes and Their Decoding Beyond Half the Minimum Distance *
    Sparse Polynomial Interpolation Codes and their Decoding Beyond Half the Minimum Distance * Erich L. Kaltofen Clément Pernet Dept. of Mathematics, NCSU U. J. Fourier, LIP-AriC, CNRS, Inria, UCBL, Raleigh, NC 27695, USA ÉNS de Lyon [email protected] 46 Allée d’Italie, 69364 Lyon Cedex 7, France www4.ncsu.edu/~kaltofen [email protected] http://membres-liglab.imag.fr/pernet/ ABSTRACT General Terms: Algorithms, Reliability We present algorithms performing sparse univariate pol- Keywords: sparse polynomial interpolation, Blahut's ynomial interpolation with errors in the evaluations of algorithm, Prony's algorithm, exact polynomial fitting the polynomial. Based on the initial work by Comer, with errors. Kaltofen and Pernet [Proc. ISSAC 2012], we define the sparse polynomial interpolation codes and state that their minimal distance is precisely the code-word length 1. INTRODUCTION divided by twice the sparsity. At ISSAC 2012, we have Evaluation-interpolation schemes are a key ingredient given a decoding algorithm for as much as half the min- in many of today's computations. Model fitting for em- imal distance and a list decoding algorithm up to the pirical data sets is a well-known one, where additional minimal distance. information on the model helps improving the fit. In Our new polynomial-time list decoding algorithm uses particular, models of natural phenomena often happen sub-sequences of the received evaluations indexed by to be sparse, which has motivated a wide range of re- an arithmetic progression, allowing the decoding for a search including compressive sensing [4], and sparse in- larger radius, that is, more errors in the evaluations terpolation of polynomials [23,1, 15, 13,9, 11].
    [Show full text]
  • Chinese Remainder Theorem
    THE CHINESE REMAINDER THEOREM KEITH CONRAD We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens 1. Introduction The Chinese remainder theorem says we can uniquely solve every pair of congruences having relatively prime moduli. Theorem 1.1. Let m and n be relatively prime positive integers. For all integers a and b, the pair of congruences x ≡ a mod m; x ≡ b mod n has a solution, and this solution is uniquely determined modulo mn. What is important here is that m and n are relatively prime. There are no constraints at all on a and b. Example 1.2. The congruences x ≡ 6 mod 9 and x ≡ 4 mod 11 hold when x = 15, and more generally when x ≡ 15 mod 99, and they do not hold for other x. The modulus 99 is 9 · 11. We will prove the Chinese remainder theorem, including a version for more than two moduli, and see some ways it is applied to study congruences. 2. A proof of the Chinese remainder theorem Proof. First we show there is always a solution. Then we will show it is unique modulo mn. Existence of Solution. To show that the simultaneous congruences x ≡ a mod m; x ≡ b mod n have a common solution in Z, we give two proofs. First proof: Write the first congruence as an equation in Z, say x = a + my for some y 2 Z. Then the second congruence is the same as a + my ≡ b mod n: Subtracting a from both sides, we need to solve for y in (2.1) my ≡ b − a mod n: Since (m; n) = 1, we know m mod n is invertible.
    [Show full text]
  • Linear Algebra Chapter 5: Norms, Inner Products and Orthogonality
    MATH 532: Linear Algebra Chapter 5: Norms, Inner Products and Orthogonality Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2015 [email protected] MATH 532 1 Outline 1 Vector Norms 2 Matrix Norms 3 Inner Product Spaces 4 Orthogonal Vectors 5 Gram–Schmidt Orthogonalization & QR Factorization 6 Unitary and Orthogonal Matrices 7 Orthogonal Reduction 8 Complementary Subspaces 9 Orthogonal Decomposition 10 Singular Value Decomposition 11 Orthogonal Projections [email protected] MATH 532 2 Vector Norms Vector[0] Norms 1 Vector Norms 2 Matrix Norms Definition 3 Inner Product Spaces Let x; y 2 Rn (Cn). Then 4 Orthogonal Vectors n T X 5 Gram–Schmidt Orthogonalization & QRx Factorizationy = xi yi 2 R i=1 6 Unitary and Orthogonal Matrices n X ∗ = ¯ 2 7 Orthogonal Reduction x y xi yi C i=1 8 Complementary Subspaces is called the standard inner product for Rn (Cn). 9 Orthogonal Decomposition 10 Singular Value Decomposition 11 Orthogonal Projections [email protected] MATH 532 4 Vector Norms Definition Let V be a vector space. A function k · k : V! R≥0 is called a norm provided for any x; y 2 V and α 2 R 1 kxk ≥ 0 and kxk = 0 if and only if x = 0, 2 kαxk = jαj kxk, 3 kx + yk ≤ kxk + kyk. Remark The inequality in (3) is known as the triangle inequality. [email protected] MATH 532 5 Vector Norms Remark Any inner product h·; ·i induces a norm via (more later) p kxk = hx; xi: We will show that the standard inner product induces the Euclidean norm (cf.
    [Show full text]
  • Anl-7826 Chinese Remainder Ind Interpolation
    ANL-7826 CHINESE REMAINDER IND INTERPOLATION ALGORITHMS John D. Lipson UolC-AUA-USAEC- ARGONNE NATIONAL LABORATORY, ARGONNE, ILLINOIS Prepared for the U.S. ATOMIC ENERGY COMMISSION under contract W-31-109-Eng-38 The facilities of Argonne National Laboratory are owned by the United States Govern­ ment. Under the terms of a contract (W-31 -109-Eng-38) between the U. S. Atomic Energy 1 Commission. Argonne Universities Association and The University of Chicago, the University employs the staff and operates the Laboratory in accordance with policies and programs formu­ lated, approved and reviewed by the Association. MEMBERS OF ARGONNE UNIVERSITIES ASSOCIATION The University of Arizona Kansas State University The Ohio State University Carnegie-Mellon University The University of Kansas Ohio University Case Western Reserve University Loyola University The Pennsylvania State University The University of Chicago Marquette University Purdue University University of Cincinnati Michigan State University Saint Louis University Illinois Institute of Technology The University of Michigan Southern Illinois University University of Illinois University of Minnesota The University of Texas at Austin Indiana University University of Missouri Washington University Iowa State University Northwestern University Wayne State University The University of Iowa University of Notre Dame The University of Wisconsin NOTICE This report was prepared as an account of work sponsored by the United States Government. Neither the United States nor the United States Atomic Energy Commission, nor any of their employees, nor any of their contractors, subcontrac­ tors, or their employees, makes any warranty, express or implied, or assumes any legal liability or responsibility for the accuracy, completeness or usefulness of any information, apparatus, product or process disclosed, or represents that its use would not infringe privately-owned rights.
    [Show full text]
  • Polynomial Interpolation: Summary ∏
    Polynomial Interpolation 1 Polynomial Interpolation: Summary • Statement of problem. We are given a table of n + 1 data points (xi,yi): x x0 x1 x2 ... xn y y0 y1 y2 ... yn and seek a polynomial p of lowest degree such that p(xi) = yi for 0 ≤ i ≤ n Such a polynomial is said to interpolate the data. • Finding the interpolating polynomial using the Vandermonde matrix. 2 n Here, pn(x) = ao +a1x +a2x +...+anx where the coefficients are found by imposing p(xi) = yi for each i = 0 to n. The resulting system is 2 n 1 x0 x0 ... x0 a0 y0 1 x x2 ... xn a y 1 1 1 1 1 1 x x2 ... xn a y 2 2 2 2 = 2 . . .. . . . . . 2 n 1 xn xn ... xn an yn The matrix above is called the Vandermonde matrix. If this was singular it would imply that for some nonzero set of coefficients the associated polynomial of degree ≤ n would have n + 1 zeros. This can’t be so this matrix equation can be solved for the unknown coefficients of the polynomial. • The Lagrange interpolation polynomial. n x − x p (x) = y ` (x) + y ` (x) + ... + y ` (x) where ` (x) = ∏ j n 0 0 1 1 n n i x − x j = 0 i j j 6= i • The Newton interpolation polynomial (Divided Differences) pn(x) = c0 + c1(x − x0) + c2(x − x0)(x − x1) + ... + cn(x − x0)(x − x1)···(x − xn−1) where these coefficients will be found using divided differences. It is fairly clear however that c0 = y0 and c = y1−y0 .
    [Show full text]
  • Polynomial Interpolation: Standard Lagrange Interpolation, Barycentric Lagrange Interpolation and Chebyshev Interpolation
    Polynomial 9 Interpolation Lab Objective: Learn and compare three methods of polynomial interpolation: standard Lagrange interpolation, Barycentric Lagrange interpolation and Chebyshev interpolation. Explore Runge's phe- nomenon and how the choice of interpolating points aect the results. Use polynomial interpolation to study air polution by approximating graphs of particulates in air. Polynomial Interpolation Polynomial interpolation is the method of nding a polynomial that matches a function at specic points in its range. More precisely, if f(x) is a function on the interval [a; b] and p(x) is a poly- nomial then p(x) interpolates the function f(x) at the points x0; x1; : : : ; xn if p(xj) = f(xj) for all j = 0; 1; : : : ; n. In this lab most of the discussion is focused on using interpolation as a means of approximating functions or data, however, polynomial interpolation is useful in a much wider array of applications. Given a function and a set of unique points n , it can be shown that there exists f(x) fxigi=0 a unique interpolating polynomial p(x). That is, there is one and only one polynomial of degree n that interpolates f(x) through those points. This uniqueness property is why, for the remainder of this lab, an interpolating polynomial is referred to as the interpolating polynomial. One approach to nding the unique interpolating polynomial of degree n is Lagrange interpolation. Lagrange interpolation Given a set n of points to interpolate, a family of basis functions with the following property fxigi=1 n n is constructed: ( 0 if i 6= j Lj(xi) = : 1 if i = j The Lagrange form of this family of basis functions is n Y (x − xk) k=1;k6=j (9.1) Lj(x) = n Y (xj − xk) k=1;k6=j 1 2 Lab 9.
    [Show full text]