Approximation in High-Dimensional and Nonparametric Statistics: a Unified Approach

Total Page:16

File Type:pdf, Size:1020Kb

Approximation in High-Dimensional and Nonparametric Statistics: a Unified Approach Approximation in high-dimensional and nonparametric statistics: a unified approach Yanjun Han Department of Electrical Engineering, Stanford University Joint work with: Jiantao Jiao Stanford EE Rajarshi Mukherjee Stanford Stats Tsachy Weissman Stanford EE July 4, 2016 Outline 1 Problem Setup 2 High-dimensional Parametric Setting 3 Infinite Dimensional Nonparametric Setting Upper bound Lower bound General Lr norm 2 / 49 1 Problem Setup 2 High-dimensional Parametric Setting 3 Infinite Dimensional Nonparametric Setting Upper bound Lower bound General Lr norm 3 / 49 Problem: estimation of functionals Given i.i.d. samples X1; ··· ; Xn ∼ P, we would like to estimate a one-dimensional functional F (P) 2 R: Parametric case: P = (p1; ··· ; pS ) is discrete, and S X F (P) = I (pi ) i=1 High dimensional: S & n Nonparametric case: P is continuous with density f , and Z F (P) = I (f (x))dx 4 / 49 Parametric case: when the functional is smooth... When I (·) is everywhere differentiable... H´ajek{Le Cam Theory The plug-in approach F (Pn) is asymptotically efficient, where Pn is the empirical distribution 5 / 49 Nonparametric case: when the functional is smooth... When I (·) is four times differentiable with bounded I (4), Taylor expansion yields Z Z 1 I (f (x))dx = I (f^) + I (1)(f^)(f − f^) + I (2)(f^)(f − f^)2 2 1 + I (3)(f^)(f − f^)3 + O((f − f^)4) dx 6 where f^ is a \good" estimator of f (e.g., a kernel estimate) Key observation: suffice to deal with linear (see, e.g., Nemirovski'00), quadratic (Bickel and Ritov'88, Birge and Massart'95) and cubic terms (Kerkyacharian and Picard'96) separately. Require bias reduction 6 / 49 What if I (·) is non-smooth? Bias dominates when I (·) is non-smooth: Theorem (`1 norm of Gaussian mean, Cai{Low'11) 2 −1 Pn For yi ∼ N (θi ; σ ); i = 1; ··· ; n and F (θ) = n i=1 jθi j, the plug-in estimator satisfies 2 2 2 σ sup Eθ (F (y) − F (θ)) σ + θ2 n |{z} n R squared bias |{z} variance Theorem (Discrete entropy, Jiao{Venkat{H.{Weissman'15) PS For X1; ··· ; Xn ∼ P = (p1; ··· ; pS ) and F (P) = i=1 −pi ln pi , the plug-in estimator satisfies 2 2 2 S (ln S) sup EP (F (Pn) − F (P)) 2 + P2M n n S |{z} | {z } squared bias variance 7 / 49 The optimal estimator Theorem (`1 norm of Gaussian mean, Cai{Low'11) 2 −1 Pn For yi ∼ N (θi ; σ ); i = 1; ··· ; n and F (θ) = n i=1 jθi j, 2 2 ^ σ inf sup Eθ F − F (θ) F^ θ2 n ln n R |{z} squared bias Theorem (Discrete entropy, Jiao{Venkat{H.{Weissman'15) PS For X1; ··· ; Xn ∼ P = (p1; ··· ; pS ) and F (P) = i=1 −pi ln pi , 2 2 ^ 2 S (ln S) inf sup EP (F − F (P)) + ^ 2 F P2MS (n ln n) n | {z } | {z } squared bias variance Effective sample size enlargement: n samples −! n ln n samples 8 / 49 Optimal estimator for `1 norm unbiased estimate plug-in of best polynomial approximation of order c2 ln n jxj \nonsmooth" \smooth" p p x 0 −c1 ln n c1 ln n 9 / 49 Optimal estimator for entropy f (pi ) unbiased estimate of best polynomial approximation of order c2 ln n 00 f (^pi )^pi (1−p^i ) f (^pi ) − 2n \nonsmooth" \smooth" 0 pi c1 ln n 1 n 10 / 49 The general recipe For a statistical model (Pθ : θ 2 Θ), consider estimating the functional F (θ) which is non-analytic at Θ0 ⊂ Θ, and θ^n is a natural estimator for θ. 1 Classify the Regime: Compute θ^n, and declare that we are in the \non-smooth" regime if θ^n is \close" enough to Θ0. Otherwise declare we are in the \smooth" regime; 2 Estimate: If θ^n falls in the \smooth" regime, use an estimator \similar" to F (θ^n) to estimate F (θ); If θ^n falls in the \non-smooth" regime, replace the functional F (θ) in the \non-smooth" regime by an approximation Fappr(θ) (another functional) which can be estimated without bias, then apply an unbiased estimator for the functional Fappr(θ). 11 / 49 Questions How to determine the \non-smooth" regime? In the \smooth" regime, what does \ `similar' to F (θ^n)" mean precisely? In the \non-smooth" regime, what approximation (including which kind, which degree, and on which region) should be employed? What if the domain of θ^n is different from (usually larger than) that of θ? 12 / 49 1 Problem Setup 2 High-dimensional Parametric Setting 3 Infinite Dimensional Nonparametric Setting Upper bound Lower bound General Lr norm 13 / 49 Estimation of information divergence Given joint independent samples X1; ··· ; Xm ∼ P = (p1; ··· ; pS ) and Y1; ··· ; Yn ∼ Q = (q1; ··· ; qS ), we would like to estimate the L1 distance and the Kullback{Leibler (KL) divergence: S X kP − Qk1 = jpi − qi j i=1 (PS pi pi ln if P Q; D(PkQ) = i=1 qi +1 otherwise. In the latter case, we assume a bounded likelihood ratio: pi =qi ≤ u(S) for any i. 14 / 49 Localization Definition (Localization) Consider a statistical model (Pθ)θ2Θ and an estimator θ^ 2 Θ^ of θ, where Θ ⊂ Θ.^ A localization of level r 2 [0; 1], or an r-localization, is a ^ collection of sets fU(x)gx2Θ^ , where U(x) ⊂ Θ for any x 2 Θ, and ^ sup Pθ(θ2 = U(θ)) ≤ r: θ2Θ Naturally induce a reverse localization V (θ) = fθ^ : U(θ^) 3 θg Localization always exists, but we seek for a small one Different from confidence set: usually r n−A 15 / 49 p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A Θ^ = Θ = R θ^ ∼ N (θ; σ2) 16 / 49 p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A θ^ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 16 / 49 p ∼ σ ln n V (θ) θ 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A p ∼ σ ln n U(θ^) θ^ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 16 / 49 p ∼ σ ln n V (θ) 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A p ∼ σ ln n U(θ^) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 16 / 49 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 16 / 49 V (θ) θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n Localization in Gaussian model: r n−A p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 2 Θ^ = Θ = R 2 θ^ ∼ N (θ; σ I2) θ^ 16 / 49 V (θ) θ p r2 ∼ σ ln n Localization in Gaussian model: r n−A p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 2 Θ^ = Θ = R 2 θ^ ∼ N (θ; σ I2) U(θ^) θ^ p r1 ∼ σ ln n 16 / 49 V (θ) θ p r2 ∼ σ ln n Localization in Gaussian model: r n−A p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 2 Θ^ = Θ = R θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r1 ∼ σ ln n 16 / 49 Localization in Gaussian model: r n−A p p ∼ σ ln n ∼ σ ln n U(θ^) V (θ) θ^ θ Θ^ = Θ = R θ^ ∼ N (θ; σ2) 2 Θ^ = Θ = R V (θ) θ^ ∼ N (θ; σ2I ) 2 θ U(θ^) θ^ p r2 ∼ σ ln n p r1 ∼ σ ln n 16 / 49 q ln n ∼ p ln n ∼ n n ln n U(^p) n V (p) ln n ln n p^ < n p > n 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A 0 1 Θ^ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 q ln n ∼ p ln n ∼ n n U(^p) V (p) ln n ln n p^ < n p > n 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A ln n n 0 1 Θ^ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 q ln n ∼ p ln n ∼ n n U(^p) V (p) ln n p > n 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A ln n n 0 1 ln n ^ p^ < n Θ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 q p ln n ∼ n V (p) ln n p > n 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A ln n ∼ n ln n U(^p) n 0 1 ln n ^ p^ < n Θ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 q p ln n ∼ n V (p) 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A ln n ∼ n ln n U(^p) n 0 1 ln n ln n ^ p^ < n p > n Θ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) V (θ) n ln n 2 Θ p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A q ln n ∼ p ln n ∼ n n ln n U(^p) n V (p) 0 1 ln n ln n ^ p^ < n p > n Θ = Θ = [0; 1] np^ ∼ B(n; p) 17 / 49 V (θ) n ln n 2 p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w Localization in Binomial model: r minfm; ng−A q ln n ∼ p ln n ∼ n n ln n U(^p) n V (p) 0 1 ln n ln n ^ p^ < n p > n Θ = Θ = [0; 1] np^ ∼ B(n; p) 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) Θ 17 / 49 V (θ) n ln n 2 p q θ = (p1; p2) ln m ∼ l ∼ q 2 1 m p1 ln m l2 ∼ m w ^ n U(θ) n ln ∼ 1 w Localization in Binomial model: r minfm; ng−A q ln n ∼ p ln n ∼ n n ln n U(^p) n V (p) 0 1 ln n ln n ^ p^ < n p > n Θ = Θ = [0; 1] np^ ∼ B(n; p) 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) Θ θ^ = (^p1; p^2) 17 / 49 V (θ) n ln n 2 p q θ = (p1; p2) ∼ q 2 p1 ln m l2 ∼ m w Localization in Binomial model: r minfm; ng−A q ln n ∼ p ln n ∼ n n ln n U(^p) n V (p) 0 1 ln n ln n ^ p^ < n p > n Θ = Θ = [0; 1] np^ ∼ B(n; p) 2 Θ^ = [0; 1] :(mp^1; np^2) ∼ B(m; p1) × B(n; p2) Θ ln m l1 ∼ m ^ n U(θ) n ln θ^ = (^p1; p^2) ∼ 1 w 17 / 49 V (θ) n ln n 2 p q ∼ q 2 p1 ln m l2 ∼ m w Localization
Recommended publications
  • Fen Fakültesi Matematik Bölümü Y
    İZMİR INSTITUTE OF TECHNOLOGY GRADUATE SCHOOL OF ENGINEERING AND SCIENCES DEPARTMENT OF MATHEMATICS CURRICULUM OF THE GRADUATE PROGRAMS M.S. in Mathematics (Thesis) Core Courses MATH 596 Graduate Seminar (0-2) NC AKTS:9 MATH 599 Scientific Research Techniques and Publication Ethics (0-2) NC AKTS:8 MATH 500 M.S. Thesis (0-1) NC AKTS:26 MATH 8XX Special Studies (8-0) NC AKTS:4 In addition, the following courses must be taken. MATH 515 Real Analysis (3-0)3 AKTS:8 MATH 527 Basic Abstract Algebra (3-0)3 AKTS:8 *All M.S. students must register Graduate Seminar course until the beginning of their 4th semester. Total credit (min.) :21 Number of courses with credit (min.) : 7 M.S. in Mathematics (Non-thesis) Core Courses MATH 516 Complex Analysis (3-0)3 AKTS:8 MATH 527 Basic Abstract Algebra (3-0)3 AKTS:8 MATH 533 Ordinary Differential Equations (3-0)3 AKTS:8 MATH 534 Partial Differential Equations (3-0)3 AKTS:8 MATH 573 Modern Geometry I (3-0)3 AKTS:8 MATH 595 Graduation Project (0-2) NC AKTS:5 MATH 596 Graduate Seminar (0-2) NC AKTS:9 MATH 599 Scientific Research Techniques and Publication Ethics (0-2) NC AKTS:8 Total credit (min.) :30 Number of courses with credit (min.) :10 Ph.D. in Mathematics Core Courses MATH 597 Comprehensive Studies (0-2) NC AKTS:9 MATH 598 Graduate Seminar in PhD (0-2) NC AKTS:9 MATH 599 Scientific Research Techniques and Publication Ethics* (0-2) NC AKTS:8 MATH 600 Ph.D.
    [Show full text]
  • RESOURCES in NUMERICAL ANALYSIS Kendall E
    RESOURCES IN NUMERICAL ANALYSIS Kendall E. Atkinson University of Iowa Introduction I. General Numerical Analysis A. Introductory Sources B. Advanced Introductory Texts with Broad Coverage C. Books With a Sampling of Introductory Topics D. Major Journals and Serial Publications 1. General Surveys 2. Leading journals with a general coverage in numerical analysis. 3. Other journals with a general coverage in numerical analysis. E. Other Printed Resources F. Online Resources II. Numerical Linear Algebra, Nonlinear Algebra, and Optimization A. Numerical Linear Algebra 1. General references 2. Eigenvalue problems 3. Iterative methods 4. Applications on parallel and vector computers 5. Over-determined linear systems. B. Numerical Solution of Nonlinear Systems 1. Single equations 2. Multivariate problems C. Optimization III. Approximation Theory A. Approximation of Functions 1. General references 2. Algorithms and software 3. Special topics 4. Multivariate approximation theory 5. Wavelets B. Interpolation Theory 1. Multivariable interpolation 2. Spline functions C. Numerical Integration and Differentiation 1. General references 2. Multivariate numerical integration IV. Solving Differential and Integral Equations A. Ordinary Differential Equations B. Partial Differential Equations C. Integral Equations V. Miscellaneous Important References VI. History of Numerical Analysis INTRODUCTION Numerical analysis is the area of mathematics and computer science that creates, analyzes, and implements algorithms for solving numerically the problems of continuous mathematics. Such problems originate generally from real-world applications of algebra, geometry, and calculus, and they involve variables that vary continuously; these problems occur throughout the natural sciences, social sciences, engineering, medicine, and business. During the second half of the twentieth century and continuing up to the present day, digital computers have grown in power and availability.
    [Show full text]
  • A Short Course on Approximation Theory
    A Short Course on Approximation Theory N. L. Carothers Department of Mathematics and Statistics Bowling Green State University ii Preface These are notes for a topics course offered at Bowling Green State University on a variety of occasions. The course is typically offered during a somewhat abbreviated six week summer session and, consequently, there is a bit less material here than might be associated with a full semester course offered during the academic year. On the other hand, I have tried to make the notes self-contained by adding a number of short appendices and these might well be used to augment the course. The course title, approximation theory, covers a great deal of mathematical territory. In the present context, the focus is primarily on the approximation of real-valued continuous functions by some simpler class of functions, such as algebraic or trigonometric polynomials. Such issues have attracted the attention of thousands of mathematicians for at least two centuries now. We will have occasion to discuss both venerable and contemporary results, whose origins range anywhere from the dawn of time to the day before yesterday. This easily explains my interest in the subject. For me, reading these notes is like leafing through the family photo album: There are old friends, fondly remembered, fresh new faces, not yet familiar, and enough easily recognizable faces to make me feel right at home. The problems we will encounter are easy to state and easy to understand, and yet their solutions should prove intriguing to virtually anyone interested in mathematics. The techniques involved in these solutions entail nearly every topic covered in the standard undergraduate curriculum.
    [Show full text]
  • Zonotopal Algebra and Forward Exchange Matroids
    Zonotopal algebra and forward exchange matroids Matthias Lenz1,2 Technische Universit¨at Berlin Sekretariat MA 4-2 Straße des 17. Juni 136 10623 Berlin GERMANY Abstract Zonotopal algebra is the study of a family of pairs of dual vector spaces of multivariate polynomials that can be associated with a list of vectors X. It connects objects from combinatorics, geometry, and approximation theory. The origin of zonotopal algebra is the pair (D(X), P(X)), where D(X) denotes the Dahmen–Micchelli space that is spanned by the local pieces of the box spline and P(X) is a space spanned by products of linear forms. The first main result of this paper is the construction of a canonical basis for D(X). We show that it is dual to the canonical basis for P(X) that is already known. The second main result of this paper is the construction of a new family of zonotopal spaces that is far more general than the ones that were recently studied by Ardila–Postnikov, Holtz–Ron, Holtz–Ron–Xu, Li–Ron, and others. We call the underlying combinatorial structure of those spaces forward exchange matroid. A forward exchange matroid is an ordered matroid together with a subset of its set of bases that satisfies a weak version of the basis exchange axiom. Keywords: zonotopal algebra, matroid, Tutte polynomial, hyperplane arrangement, multivariate spline, graded vector space, kernel of differential operators, Hilbert series 2010 MSC: Primary 05A15, 05B35, 13B25, 16S32, 41A15. Secondary: 05C31, arXiv:1204.3869v4 [math.CO] 31 Mar 2016 41A63, 47F05. Email address: [email protected] (Matthias Lenz) 1Present address: D´epartement de math´ematiques, Universit´ede Fribourg, Chemin du Mus´ee 23, 1700 Fribourg, Switzerland 2The author was supported by a Sofia Kovalevskaya Research Prize of Alexander von Humboldt Foundation awarded to Olga Holtz and by a PhD scholarship from the Berlin Mathematical School (BMS) 22nd November 2018 1.
    [Show full text]
  • Math 541 - Numerical Analysis Approximation Theory Discrete Least Squares Approximation
    Math 541 - Numerical Analysis Approximation Theory Discrete Least Squares Approximation Joseph M. Mahaffy, [email protected] Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research Center San Diego State University San Diego, CA 92182-7720 http://jmahaffy.sdsu.edu Spring 2018 Discrete Least Squares Approximation — Joseph M. Mahaffy, [email protected] (1/97) Outline 1 Approximation Theory: Discrete Least Squares Introduction Discrete Least Squares A Simple, Powerful Approach 2 Application: Cricket Thermometer Polynomial fits Model Selection - BIC and AIC Different Norms Weighted Least Squares 3 Application: U. S. Population Polynomial fits Exponential Models 4 Application: Pharmokinetics Compartment Models Dog Study with Opioid Exponential Peeling Nonlinear least squares fits Discrete Least Squares Approximation — Joseph M. Mahaffy, [email protected] (2/97) Introduction Approximation Theory: Discrete Least Squares Discrete Least Squares A Simple, Powerful Approach Introduction: Matching a Few Parameters to a Lot of Data. Sometimes we get a lot of data, many observations, and want to fit it to a simple model. 8 6 4 2 0 0 1 2 3 4 5 Underlying function f(x) = 1 + x + x^2/25 Measured Data Average Linear Best Fit Quadratic Best Fit PDF-link: code. Discrete Least Squares Approximation — Joseph M. Mahaffy, [email protected] (3/97) Introduction Approximation Theory: Discrete Least Squares Discrete Least Squares A Simple, Powerful Approach Why a Low Dimensional Model? Low dimensional models (e.g. low degree polynomials) are easy to work with, and are quite well behaved (high degree polynomials can be quite oscillatory.) All measurements are noisy, to some degree.
    [Show full text]
  • Arxiv:1106.4415V1 [Math.DG] 22 Jun 2011 R,Rno Udai Form
    JORDAN STRUCTURES IN MATHEMATICS AND PHYSICS Radu IORDANESCU˘ 1 Institute of Mathematics of the Romanian Academy P.O.Box 1-764 014700 Bucharest, Romania E-mail: [email protected] FOREWORD The aim of this paper is to offer an overview of the most important applications of Jordan structures inside mathematics and also to physics, up- dated references being included. For a more detailed treatment of this topic see - especially - the recent book Iord˘anescu [364w], where sugestions for further developments are given through many open problems, comments and remarks pointed out throughout the text. Nowadays, mathematics becomes more and more nonassociative (see 1 § below), and my prediction is that in few years nonassociativity will govern mathematics and applied sciences. MSC 2010: 16T25, 17B60, 17C40, 17C50, 17C65, 17C90, 17D92, 35Q51, 35Q53, 44A12, 51A35, 51C05, 53C35, 81T05, 81T30, 92D10. Keywords: Jordan algebra, Jordan triple system, Jordan pair, JB-, ∗ ∗ ∗ arXiv:1106.4415v1 [math.DG] 22 Jun 2011 JB -, JBW-, JBW -, JH -algebra, Ricatti equation, Riemann space, symmet- ric space, R-space, octonion plane, projective plane, Barbilian space, Tzitzeica equation, quantum group, B¨acklund-Darboux transformation, Hopf algebra, Yang-Baxter equation, KP equation, Sato Grassmann manifold, genetic alge- bra, random quadratic form. 1The author was partially supported from the contract PN-II-ID-PCE 1188 517/2009. 2 CONTENTS 1. Jordan structures ................................. ....................2 § 2. Algebraic varieties (or manifolds) defined by Jordan pairs ............11 § 3. Jordan structures in analysis ....................... ..................19 § 4. Jordan structures in differential geometry . ...............39 § 5. Jordan algebras in ring geometries . ................59 § 6. Jordan algebras in mathematical biology and mathematical statistics .66 § 7.
    [Show full text]
  • Feynman Quantization
    3 FEYNMAN QUANTIZATION An introduction to path-integral techniques Introduction. By Richard Feynman (–), who—after a distinguished undergraduate career at MIT—had come in as a graduate student to Princeton, was deeply involved in a collaborative effort with John Wheeler (his thesis advisor) to shake the foundations of field theory. Though motivated by problems fundamental to quantum field theory, as it was then conceived, their work was entirely classical,1 and it advanced ideas so radicalas to resist all then-existing quantization techniques:2 new insight into the quantization process itself appeared to be called for. So it was that (at a beer party) Feynman asked Herbert Jehle (formerly a student of Schr¨odinger in Berlin, now a visitor at Princeton) whether he had ever encountered a quantum mechanical application of the “Principle of Least Action.” Jehle directed Feynman’s attention to an obscure paper by P. A. M. Dirac3 and to a brief passage in §32 of Dirac’s Principles of Quantum Mechanics 1 John Archibald Wheeler & Richard Phillips Feynman, “Interaction with the absorber as the mechanism of radiation,” Reviews of Modern Physics 17, 157 (1945); “Classical electrodynamics in terms of direct interparticle action,” Reviews of Modern Physics 21, 425 (1949). Those were (respectively) Part III and Part II of a projected series of papers, the other parts of which were never published. 2 See page 128 in J. Gleick, Genius: The Life & Science of Richard Feynman () for a popular account of the historical circumstances. 3 “The Lagrangian in quantum mechanics,” Physicalische Zeitschrift der Sowjetunion 3, 64 (1933). The paper is reprinted in J.
    [Show full text]
  • Faster Algorithms Via Approximation Theory
    Foundations and Trends R in Theoretical Computer Science Vol. 9, No. 2 (2013) 125–210 c 2014 S. Sachdeva and N. K. Vishnoi DOI: 10.1561/0400000065 Faster Algorithms via Approximation Theory Sushant Sachdeva Nisheeth K. Vishnoi Yale University Microsoft Research [email protected] [email protected] Contents Introduction 126 I APPROXIMATION THEORY 135 1 Uniform Approximations 136 2 Chebyshev Polynomials 140 3 Approximating Monomials 145 4 Approximating the Exponential 149 5 Lower Bounds for Polynomial Approximations 153 6 Approximating the Exponential using Rational Functions 158 7 Approximating the Exponential using Rational Functions with Negative Poles 163 ii iii II APPLICATIONS 171 8 Simulating Random Walks 174 9 Solving Equations via the Conjugate Gradient Method 179 10 Computing Eigenvalues via the Lanczos Method 185 11 Computing the Matrix Exponential 191 12 Matrix Inversion via Exponentiation 196 References 205 Abstract This monograph presents ideas and techniques from approximation theory for approximating functions such as xs; x−1 and e−x, and demonstrates how these results play a crucial role in the design of fast algorithms for problems which are increasingly relevant. The key lies in the fact that such results imply faster ways to compute primitives such as Asv; A−1v; exp(−A)v; eigenvalues, and eigenvectors, which are fundamental to many spectral algorithms. Indeed, many fast algorithms reduce to the computation of such primitives, which have proved useful for speeding up several fundamental computations such as random walk simulation, graph partitioning, and solving systems of linear equations. S. Sachdeva and N. K. Vishnoi. Faster Algorithms via Approximation Theory.
    [Show full text]
  • Publications by Lloyd N. Trefethen
    Publications by Lloyd N. Trefethen I. Books I.1. Numerical Conformal Mapping, editor, 269 pages. Elsevier, 1986. I.2. Finite Difference and Spectral Methods for Ordinary and Partial Differential Equations, x+315 pages. Graduate textbook, privately published at http://comlab.ox.ac.uk/nick.trefethen/, 1996. I.3. Numerical Linear Algebra, with David Bau III, xii+361 pages. SIAM, 1997. I.4. Spectral Methods in MATLAB, xviii+165 pages. SIAM, 2000. I.5. Schwarz-Christoffel Mapping, with Tobin A. Driscoll, xvi+132 pages. Cambridge U. Press, 2002. I.6. Spectra and Pseudospectra: The Behavior of Nonnormal Matrices and Operators, with Mark Embree, xviii+606 pages. Princeton U. Press, 2005. I.7. Trefethen’s Index Cards: Forty years of Notes about People, Words, and Mathematics, xv + 368 pages. World Scientific Publishing, 2011. I.8. Approximation Theory and Approximation Practice. SIAM, to appear in 2013. In addition to these a book has been published by other authors about the international problem solving challenge I organised in 2002 (item VIII.8 below): F. Bornemann, D. Laurie, S. Wagon and D. Waldvogel, The SIAM 100-Digit Challenge, SIAM, 2004, to appear in German translation as Abenteuer Numerik: Eine Führung entlang der Herausforderung von Trefethen, Springer, 2006. II. Finite difference and spectral methods for partial differential equations II.1. Group velocity in finite difference schemes. SIAM Review 24 (1982), 113-136. II.2. Group velocity interpretation of the stability theory of Gustafsson, Kreiss, and Sundström. J. Comp. Phys. 49 (1983), 199-217. II.3. On Lp-instability and dispersion at discontinuities in finite difference schemes.
    [Show full text]
  • Topics in High-Dimensional Approximation Theory
    TOPICS IN HIGH-DIMENSIONAL APPROXIMATION THEORY DISSERTATION Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of the Ohio State University By Yeonjong Shin Graduate Program in Mathematics The Ohio State University 2018 Dissertation Committee: Dongbin Xiu, Advisor Ching-Shan Chou Chuan Xue c Copyright by Yeonjong Shin 2018 ABSTRACT Several topics in high-dimensional approximation theory are discussed. The fundamental problem in approximation theory is to approximate an unknown function, called the target function by using its data/samples/observations. Depending on different scenarios on the data collection, different approximation techniques need to be sought in order to utilize the data properly. This dissertation is concerned with four different approximation methods in four different scenarios in data. First, suppose the data collecting procedure is resource intensive, which may require expensive numerical simulations or experiments. As the performance of the approximation highly depends on the data set, one needs to carefully decide where to collect data. We thus developed a method of designing a quasi-optimal point set which guides us where to collect the data, before the actual data collection procedure. We showed that the resulting quasi-optimal set notably outperforms than other standard choices. Second, it is not always the case that we obtain the exact data. Suppose the data is corrupted by unexpected external errors. These unexpected corruption errors can have big magnitude and in most cases are non-random. We proved that a well-known classical method, the least absolute deviation (LAD), could effectively eliminate the corruption er- rors.
    [Show full text]
  • Function Approximation and the Remez Algorithm 11
    FUNCTION APPROXIMATION AND THE REMEZ ALGORITHM ABIY TASISSA Abstract. The Remez exchange algorithm is explained. An implementation in Python is tested on different test functions. We make a comparison with SLSQP(Sequential Least Squares Programming) optimizer. Key words. Remez exchange, Minimax polynomial, polynomial interpolation 1. Overview. Given a set of data points and values measured at these points, it is of interest to to determine the function that does a `good' approximation. This problem belongs to the realm of function approximation or interpolation theory and finds its use in scientific computations[7]. In the past, these problems were handled using complicated tables and the evaluations were done using calculators or just by hand[4]. To increase the number of significant digits, it meant more manual work to be done. With the coming of computers, the computations were very fast and there was no need to store values as things could be computed on the fly. Approximating functions by rational functions is one class of problems in approx- imation theory. We are interested in the approximation of a function by a rational function in the L1 norm. That is we want to minimize the maximum vertical dis- tance between the function in consideration and a rational function. This is called the Minimax Approximation[7]. The theoretical framework for this was originally done by Chebyshev[16]. He stated what is called equioscillation theorem that gave a necessary and sufficient condition for a minimax polynomial/rational function[7]. By 1915, all the theoretical results regarding Minimax approximations had been es- tablished [19]. However, as we will see in later sections, the theoretical tools are useful but don't lead to any feasible way to find the Minimax rational function.
    [Show full text]
  • Approximate Approximations Mathematical I Surveys I and I Monographs I
    http://dx.doi.org/10.1090/surv/141 Approximate Approximations Mathematical I Surveys I and I Monographs I Volume 141 I Approximate Approximations I Vladimir Maz'ya I Gunther Schmidt EDITORIAL COMMITTEE Jerry L. Bona Michael G. Eastwood Ralph L. Cohen Michael P. Loss J. T. Stafford, Chair 2000 Mathematics Subject Classification. Primary 41A30. For additional information and updates on this book, visit www.ams.org/bookpages/surv-141 Library of Congress Cataloging-in-Publication Data Maz'ia, V. G. Approximate approximations / Vladimir Maz'ya, Gunther Schmidt. p. cm. — (Mathematical surveys and monographs ; v. 141) Includes bibliographical references and index. ISBN 978-0-8218-4203-4 (alk. paper) 1. Approximation theory. I. Schmidt, Gunther, 1952- II. Title. QA221.M37 2007 511'.4—dc22 2007060769 Copying and reprinting. Individual readers of this publication, and nonprofit libraries acting for them, are permitted to make fair use of the material, such as to copy a chapter for use in teaching or research. Permission is granted to quote brief passages from this publication in reviews, provided the customary acknowledgment of the source is given. Republication, systematic copying, or multiple reproduction of any material in this publication is permitted only under license from the American Mathematical Society. Requests for such permission should be addressed to the Acquisitions Department, American Mathematical Society, 201 Charles Street, Providence, Rhode Island 02904-2294, USA. Requests can also be made by e-mail to [email protected]. © 2007 by the American Mathematical Society. All rights reserved. The American Mathematical Society retains all rights except those granted to the United States Government.
    [Show full text]