Signals and Systems Lecture (S2) Orthogonal Functions and Fourier
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Orthogonal Functions, Radial Basis Functions (RBF) and Cross-Validation (CV)
Exercise 6: Orthogonal Functions, Radial Basis Functions (RBF) and Cross-Validation (CV) CSElab Computational Science & Engineering Laboratory Outline 1. Goals 2. Theory/ Examples 3. Questions Goals ⚫ Orthonormal Basis Functions ⚫ How to construct an orthonormal basis ? ⚫ Benefit of using an orthonormal basis? ⚫ Radial Basis Functions (RBF) ⚫ Cross Validation (CV) Motivation: Orthonormal Basis Functions ⚫ Computation of interpolation coefficients can be compute intensive ⚫ Adding or removing basis functions normally requires a re-computation of the coefficients ⚫ Overcome by orthonormal basis functions Gram-Schmidt Orthonormalization 푛 ⚫ Given: A set of vectors 푣1, … , 푣푘 ⊂ 푅 푛 ⚫ Goal: Generate a set of vectors 푢1, … , 푢푘 ⊂ 푅 such that 푢1, … , 푢푘 is orthonormal and spans the same subspace as 푣1, … , 푣푘 ⚫ Click here if the animation is not playing Recap: Projection of Vectors ⚫ Notation: Denote dot product as 푎Ԧ ⋅ 푏 = 푎Ԧ, 푏 1 (bilinear form). This implies the norm 푢 = 푎Ԧ, 푎Ԧ 2 ⚫ Define the scalar projection of v onto u as 푃푢 푣 = 푢,푣 | 푢 | 푢,푣 푢 ⚫ Hence, is the part of v pointing in direction of u | 푢 | | 푢 | 푢 푢 and 푣∗ = 푣 − , 푣 is orthogonal to 푢 푢 | 푢 | Gram-Schmidt for Vectors 푛 ⚫ Given: A set of vectors 푣1, … , 푣푘 ⊂ 푅 푣1 푢1 = | 푣1 | 푣2,푢1 푢1 푢2 = 푣2 − = 푣2 − 푣2, 푢1 푢1 as 푢1 normalized 푢1 푢1 푢3 = 푣3 − 푣3, 푢1 푢1 What’s missing? Gram-Schmidt for Vectors 푛 ⚫ Given: A set of vectors 푣1, … , 푣푘 ⊂ 푅 푣1 푢1 = | 푣1 | 푣2,푢1 푢1 푢2 = 푣2 − = 푣2 − 푣2, 푢1 푢1 as 푢1 normalized 푢1 푢1 푢3 = 푣3 − 푣3, 푢1 푢1 What’s missing? 푢3is orthogonal to 푢1, but not yet to 푢2 푢3 = 푢3 − 푢3, 푢2 푢2 If the animation is not playing, please click here Gram-Schmidt for Functions ⚫ Given: A set of functions 푔1, … , 푔푘 .We want 휙1, … , 휙푘 ⚫ Bridge to Gram-Schmidt for vectors: ∞ ⚫ Define 푔푖, 푔푗 = −∞ 푔푖 푥 푔푗 푥 푑푥 ⚫ What’s the corresponding norm? Gram-Schmidt for Functions ⚫ Given: A set of functions 푔1, … , 푔푘 .We want 휙1, … , 휙푘 ⚫ Bridge to Gram-Schmidt for vectors: ∞ ⚫ 푔 , 푔 = 푔 푥 푔 푥 푑푥 . -
Anatomically Informed Basis Functions — Anatomisch Informierte Basisfunktionen
Anatomically Informed Basis Functions | Anatomisch Informierte Basisfunktionen Dissertation zur Erlangung des akademischen Grades doctor rerum naturalium (Dr.rer.nat.), genehmigt durch die Fakult¨atf¨urNaturwissenschaften der Otto{von{Guericke{Universit¨atMagdeburg von Diplom-Informatiker Stefan Kiebel geb. am 10. Juni 1968 in Wadern/Saar Gutachter: Prof. Dr. Lutz J¨ancke Prof. Dr. Hermann Hinrichs Prof. Dr. Cornelius Weiller Eingereicht am: 27. Februar 2001 Verteidigung am: 9. August 2001 Abstract In this thesis, a method is presented that incorporates anatomical information into the statistical analysis of functional neuroimaging data. Available anatomical informa- tion is used to explicitly specify spatial components within a functional volume that are assumed to carry evidence of functional activation. After estimating the activity by fitting the same spatial model to each functional volume and projecting the estimates back into voxel-space, one can proceed with a conventional time-series analysis such as statistical parametric mapping (SPM). The anatomical information used in this work comprised the reconstructed grey matter surface, derived from high-resolution T1-weighted magnetic resonance images (MRI). The spatial components specified in the model were of low spatial frequency and confined to the grey matter surface. By explaining the observed activity in terms of these components, one efficiently captures spatially smooth response components induced by underlying neuronal activations lo- calised close to or within the grey matter sheet. Effectively, the method implements a spatially variable anatomically informed deconvolution and consequently the method was named anatomically informed basis functions (AIBF). AIBF can be used for the analysis of any functional imaging modality. In this thesis it was applied to simu- lated and real functional MRI (fMRI) and positron emission tomography (PET) data. -
A Local Radial Basis Function Method for the Numerical Solution of Partial Differential Equations Maggie Elizabeth Chenoweth [email protected]
Marshall University Marshall Digital Scholar Theses, Dissertations and Capstones 1-1-2012 A Local Radial Basis Function Method for the Numerical Solution of Partial Differential Equations Maggie Elizabeth Chenoweth [email protected] Follow this and additional works at: http://mds.marshall.edu/etd Part of the Numerical Analysis and Computation Commons Recommended Citation Chenoweth, Maggie Elizabeth, "A Local Radial Basis Function Method for the Numerical Solution of Partial Differential Equations" (2012). Theses, Dissertations and Capstones. Paper 243. This Thesis is brought to you for free and open access by Marshall Digital Scholar. It has been accepted for inclusion in Theses, Dissertations and Capstones by an authorized administrator of Marshall Digital Scholar. For more information, please contact [email protected]. A LOCAL RADIAL BASIS FUNCTION METHOD FOR THE NUMERICAL SOLUTION OF PARTIAL DIFFERENTIAL EQUATIONS Athesissubmittedto the Graduate College of Marshall University In partial fulfillment of the requirements for the degree of Master of Arts in Mathematics by Maggie Elizabeth Chenoweth Approved by Dr. Scott Sarra, Committee Chairperson Dr. Anna Mummert Dr. Carl Mummert Marshall University May 2012 Copyright by Maggie Elizabeth Chenoweth 2012 ii ACKNOWLEDGMENTS I would like to begin by expressing my sincerest appreciation to my thesis advisor, Dr. Scott Sarra. His knowledge and expertise have guided me during my research endeavors and the process of writing this thesis. Dr. Sarra has also served as my teaching mentor, and I am grateful for all of his encouragement and advice. It has been an honor to work with him. I would also like to thank the other members of my thesis committee, Dr. -
EOF/PC Analysis
ATM 552 Notes: Matrix Methods: EOF, SVD, ETC. D.L.Hartmann Page 68 4. Matrix Methods for Analysis of Structure in Data Sets: Empirical Orthogonal Functions, Principal Component Analysis, Singular Value Decomposition, Maximum Covariance Analysis, Canonical Correlation Analysis, Etc. In this chapter we discuss the use of matrix methods from linear algebra, primarily as a means of searching for structure in data sets. Empirical Orthogonal Function (EOF) analysis seeks structures that explain the maximum amount of variance in a two-dimensional data set. One dimension in the data set represents the dimension in which we are seeking to find structure, and the other dimension represents the dimension in which realizations of this structure are sampled. In seeking characteristic spatial structures that vary with time, for example, we would use space as the structure dimension and time as the sampling dimension. The analysis produces a set of structures in the first dimension, which we call the EOF’s, and which we can think of as being the structures in the spatial dimension. The complementary set of structures in the sampling dimension (e.g. time) we can call the Principal Components (PC’s), and they are related one-to-one to the EOF’s. Both sets of structures are orthogonal in their own dimension. Sometimes it is helpful to sacrifice one or both of these orthogonalities to produce more compact or physically appealing structures, a process called rotation of EOF’s. Singular Value Decomposition (SVD) is a general decomposition of a matrix. It can be used on data matrices to find both the EOF’s and PC’s simultaneously. -
Complete Set of Orthogonal Basis Mathematical Physics
R. I. Badran Complete set of orthogonal basis Mathematical Physics Complete set of orthogonal functions Discrete set of vectors: ˆ ˆ ˆ ˆ ˆ ˆ The two vectors A AX i A y j Az k and B BX i B y j Bz k are 3 orthogonal if A B 0 or Ai Bi 0 . i1 Continuous set of functions on an interval (a, b): a) The two continuous functions A(x) and B (x) are orthogonal b on the interval (a, b) if A(x)B(x)dx 0. a b) The two complex functions A (x) and B (x) are orthogonal on b the interval (a, b) if A (x)B(x)dx 0 , where A*(x) is the a complex conjugate of A (x). c) For a whole set of functions An (x) (where n= 1, 2, 3,) and on the interval (a, b) b 0 if m n An (x)Am (x)dx a , const.t 0 if m n An (x) is called a set of orthogonal functions. Examples: 0 m n sin nx sin mxdx i) , m n 0 where sin nx is a set of orthogonal functions on the interval (-, ). Similarly 0 m n cos nx cos mxdx if if m n 0 R. I. Badran Complete set of orthogonal basis Mathematical Physics ii) sin nx cos mxdx 0 for any n and m 0 (einx ) eimxdx iii) 2 1 vi) P (x)Pm (x)dx 0 unless. m 1 [Try to prove this; also solve problems (2, 5) of section 6]. -
Linear Basis Function Models
Linear basis function models Linear basis function models Linear models for regression Francesco Corona FC - Fortaleza Linear basis function models Linear models for regression Linear basis function models Linear models for regression The focus so far on unsupervised learning, we turn now to supervised learning ◮ Regression The goal of regression is to predict the value of one or more continuous target variables t, given the value of a D-dimensional vector x of input variables ◮ e.g., polynomial curve fitting FC - Fortaleza Linear basis function models Linear basis function models Linear models for regression (cont.) Training data of N = 10 points, blue circles ◮ each comprising an observation of the input variable x along with the 1 corresponding target variable t t 0 The unknown function sin(2πx) is used to generate the data, green curve ◮ −1 Goal: Predict the value of t for some new value of x ◮ 0 x 1 without knowledge of the green curve The input training data x was generated by choosing values of xn, for n = 1,..., N, that are spaced uniformly in the range [0, 1] The target training data t was obtained by computing values sin(2πxn) of the function and adding a small level of Gaussian noise FC - Fortaleza Linear basis function models Linear basis function models Linear models for regression (cont.) ◮ We shall fit the data using a polynomial function of the form M 2 M j y(x, w)= w0 + w1x + w2x + · · · + wM x = wj x (1) Xj=0 ◮ M is the polynomial order, x j is x raised to the power of j ◮ Polynomial coefficients w0,..., wM are collected -
Introduction to Radial Basis Function Networks
Intro duction to Radial Basis Function Networks Mark J L Orr Centre for Cognitive Science University of Edinburgh Buccleuch Place Edinburgh EH LW Scotland April Abstract This do cumentisanintro duction to radial basis function RBF networks a typ e of articial neural network for application to problems of sup ervised learning eg regression classication and time series prediction It is now only available in PostScript an older and now unsupp orted hyp ertext ver sion maybeavailable for a while longer The do cumentwas rst published in along with a package of Matlab functions implementing the metho ds describ ed In a new do cument Recent Advances in Radial Basis Function Networks b ecame available with a second and improved version of the Matlab package mjoancedacuk wwwancedacukmjopapersintrops wwwancedacukmjointrointrohtml wwwancedacukmjosoftwarerbfzip wwwancedacukmjopapersrecadps wwwancedacukmjosoftwarerbfzip Contents Intro duction Sup ervised Learning Nonparametric Regression Classication and Time Series Prediction Linear Mo dels Radial Functions Radial Basis Function Networks Least Squares The Optimal WeightVector The Pro jection Matrix Incremental Op erations The Eective NumberofParameters Example Mo del Selection Criteria CrossValidation Generalised -
Linear Models for Regression
Machine Learning Srihari Linear Models for Regression Sargur Srihari [email protected] 1 Machine Learning Srihari Topics in Linear Regression • What is regression? – Polynomial Curve Fitting with Scalar input – Linear Basis Function Models • Maximum Likelihood and Least Squares • Stochastic Gradient Descent • Regularized Least Squares 2 Machine Learning Srihari The regression task • It is a supervised learning task • Goal of regression: – predict value of one or more target variables t – given d-dimensional vector x of input variables – With dataset of known inputs and outputs • (x1,t1), ..(xN,tN) • Where xi is an input (possibly a vector) known as the predictor • ti is the target output (or response) for case i which is real-valued – Goal is to predict t from x for some future test case • We are not trying to model the distribution of x – We dont expect predictor to be a linear function of x • So ordinary linear regression of inputs will not work • We need to allow for a nonlinear function of x • We don’t have a theory of what form this function to take 3 Machine Learning Srihari An example problem • Fifty points generated (one-dimensional problem) – With x uniform from (0,1) – y generated from formula y=sin(1+x2)+noise • Where noise has N(0,0.032) distribution • Noise-free true function and data points are as shown 4 Machine Learning Srihari Applications of Regression 1. Expected claim amount an insured person will make (used to set insurance premiums) or prediction of future prices of securities 2. Also used for algorithmic -
A Complete System of Orthogonal Step Functions
PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 132, Number 12, Pages 3491{3502 S 0002-9939(04)07511-2 Article electronically published on July 22, 2004 A COMPLETE SYSTEM OF ORTHOGONAL STEP FUNCTIONS HUAIEN LI AND DAVID C. TORNEY (Communicated by David Sharp) Abstract. We educe an orthonormal system of step functions for the interval [0; 1]. This system contains the Rademacher functions, and it is distinct from the Paley-Walsh system: its step functions use the M¨obius function in their definition. Functions have almost-everywhere convergent Fourier-series expan- sions if and only if they have almost-everywhere convergent step-function-series expansions (in terms of the members of the new orthonormal system). Thus, for instance, the new system and the Fourier system are both complete for Lp(0; 1); 1 <p2 R: 1. Introduction Analytical desiderata and applications, ever and anon, motivate the elaboration of systems of orthogonal step functions|as exemplified by the Haar system, the Paley-Walsh system and the Rademacher system. Our motivation is the example- based classification of digitized images, represented by rational points of the real interval [0,1], the domain of interest in the sequel. It is imperative to establish the \completeness" of any orthogonal system. Much is known, in this regard, for the classical systems, as this has been the subject of numerous investigations, and we use the latter results to establish analogous properties for the orthogonal system developed herein. h i RDefinition 1. Let the inner product f(x);g(x) denote the Lebesgue integral 1 0 f(x)g(x)dx: Definition 2. -
Radial Basis Functions
Acta Numerica (2000), pp. 1–38 c Cambridge University Press, 2000 Radial basis functions M. D. Buhmann Mathematical Institute, Justus Liebig University, 35392 Giessen, Germany E-mail: [email protected] Radial basis function methods are modern ways to approximate multivariate functions, especially in the absence of grid data. They have been known, tested and analysed for several years now and many positive properties have been identified. This paper gives a selective but up-to-date survey of several recent developments that explains their usefulness from the theoretical point of view and contributes useful new classes of radial basis function. We consider particularly the new results on convergence rates of interpolation with radial basis functions, as well as some of the various achievements on approximation on spheres, and the efficient numerical computation of interpolants for very large sets of data. Several examples of useful applications are stated at the end of the paper. CONTENTS 1 Introduction 1 2 Convergence rates 5 3 Compact support 11 4 Iterative methods for implementation 16 5 Interpolation on spheres 25 6 Applications 28 References 34 1. Introduction There is a multitude of ways to approximate a function of many variables: multivariate polynomials, splines, tensor product methods, local methods and global methods. All of these approaches have many advantages and some disadvantages, but if the dimensionality of the problem (the number of variables) is large, which is often the case in many applications from stat- istics to neural networks, our choice of methods is greatly reduced, unless 2 M. D. Buhmann we resort solely to tensor product methods. -
Basis Functions
Basis Functions Volker Tresp Summer 2017 1 Nonlinear Mappings and Nonlinear Classifiers • Regression: { Linearity is often a good assumption when many inputs influence the output { Some natural laws are (approximately) linear F = ma { But in general, it is rather unlikely that a true function is linear • Classification: { Linear classifiers also often work well when many inputs influence the output { But also for classifiers, it is often not reasonable to assume that the classification boundaries are linear hyper planes 2 Trick • We simply transform the input into a high-dimensional space where the regressi- on/classification is again linear! • Other view: let's define appropriate features • Other view: let's define appropriate basis functions • XOR-type problem with patterns 0 0 ! +1 1 0 ! −1 0 1 ! −1 1 1 ! +1 3 XOR is not linearly separable 4 Trick: Let's Add Basis Functions • Linear Model: input vector: 1; x1; x2 • Let's consider x1x2 in addition • The interaction term x1x2 couples two inputs nonlinearly 5 With a Third Input z3 = x1x2 the XOR Becomes Linearly Separable f(x) = 1 − 2x1 − 2x2 + 4x1x2 = φ1(x) − 2φ2(x) − 2φ3(x) + 4φ4(x) with φ1(x) = 1; φ2(x) = x1; φ3(x) = x2; φ4(x) = x1x2 6 f(x) = 1 − 2x1 − 2x2 + 4x1x2 7 Separating Planes 8 A Nonlinear Function 9 f(x) = x − 0:3x3 2 3 Basis functions φ1(x) = 1; φ2(x) = x; φ3(x) = x ; φ4(x) = x und w = (0; 1; 0; −0:3) 10 Basic Idea • The simple idea: in addition to the original inputs, we add inputs that are calculated as deterministic functions of the existing inputs, and treat them as additional inputs -
FOURIER ANALYSIS 1. the Best Approximation Onto Trigonometric
FOURIER ANALYSIS ERIK LØW AND RAGNAR WINTHER 1. The best approximation onto trigonometric polynomials Before we start the discussion of Fourier series we will review some basic results on inner–product spaces and orthogonal projections mostly presented in Section 4.6 of [1]. 1.1. Inner–product spaces. Let V be an inner–product space. As usual we let u, v denote the inner–product of u and v. The corre- sponding normh isi given by v = v, v . k k h i A basic relation between the inner–productp and the norm in an inner– product space is the Cauchy–Scwarz inequality. It simply states that the absolute value of the inner–product of u and v is bounded by the product of the corresponding norms, i.e. (1.1) u, v u v . |h i|≤k kk k An outline of a proof of this fundamental inequality, when V = Rn and is the standard Eucledian norm, is given in Exercise 24 of Section 2.7k·k of [1]. We will give a proof in the general case at the end of this section. Let W be an n dimensional subspace of V and let P : V W be the corresponding projection operator, i.e. if v V then w∗ =7→P v W is the element in W which is closest to v. In other∈ words, ∈ v w∗ v w for all w W. k − k≤k − k ∈ It follows from Theorem 12 of Chapter 4 of [1] that w∗ is characterized by the conditions (1.2) v P v, w = v w∗, w =0 forall w W.