Topics in Mathematics of Data Science Lecture Notes

Total Page:16

File Type:pdf, Size:1020Kb

Topics in Mathematics of Data Science Lecture Notes Ten Lectures and Forty-Two Open Problems in the Mathematics of Data Science Afonso S. Bandeira December, 2015 Preface These are notes from a course I gave at MIT on the Fall of 2015 entitled: \18.S096: Topics in Mathematics of Data Science". These notes are not in final form and will be continuously edited and/or corrected (as I am sure they contain many typos). Please use at your own risk and do let me know if you find any typo/mistake. Part of the content of this course is greatly inspired by a course I took from Amit Singer while a graduate student at Princeton. Amit's course was inspiring and influential on my research interests. I can only hope that these notes may one day inspire someone's research in the same way that Amit's course inspired mine. These notes also include a total of forty-two open problems (now 41, as in meanwhile Open Problem 1.3 has been solved [MS15]!). This list of problems does not necessarily contain the most important problems in the field (al- though some will be rather important). I have tried to select a mix of important, perhaps approachable, and fun problems. Hopefully you will enjoy thinking about these problems as much as I do! I would like to thank all the students who took my course, it was a great and interactive audience! I would also like to thank Nicolas Boumal, Ludwig Schmidt, and Jonathan Weed for letting me know of several typos. Thank you also to Nicolas Boumal, Dustin G. Mixon, Bernat Guillen Pegueroles, Philippe Rigollet, and Francisco Unda for suggesting open problems. Contents 0.1 List of open problems . 4 0.2 A couple of Open Problems . 6 0.2.1 Koml´osConjecture . 6 0.2.2 Matrix AM-GM inequality . 7 0.3 Brief Review of some linear algebra tools . 7 0.3.1 Singular Value Decomposition . 7 0.3.2 Spectral Decomposition . 8 1 0.3.3 Trace and norm . 8 0.4 Quadratic Forms . 9 1 Principal Component Analysis in High Dimensions and the Spike Model 10 1.1 Dimension Reduction and PCA . 10 1.1.1 PCA as best d-dimensional affine fit . 10 1.1.2 PCA as d-dimensional projection that preserves the most variance . 12 1.1.3 Finding the Principal Components . 13 1.1.4 Which d should we pick? . 13 1.1.5 A related open problem . 14 1.2 PCA in high dimensions and Marcenko-Pastur . 15 1.2.1 A related open problem . 17 1.3 Spike Models and BBP transition . 18 1.3.1 A brief mention of Wigner matrices . 22 1.3.2 An open problem about spike models . 23 2 Graphs, Diffusion Maps, and Semi-supervised Learning 24 2.1 Graphs . 24 2.1.1 Cliques and Ramsey numbers . 25 2.2 Diffusion Maps . 29 2.2.1 A couple of examples . 32 2.2.2 Diffusion Maps of point clouds . 33 2.2.3 A simple example . 34 2.2.4 Similar non-linear dimensional reduction techniques . 34 2.3 Semi-supervised learning . 35 2.3.1 An interesting experience and the Sobolev Embedding Theorem . 38 3 Spectral Clustering and Cheeger's Inequality 41 3.1 Clustering . 41 3.1.1 k-means Clustering . 41 3.2 Spectral Clustering . 43 3.3 Two clusters . 45 3.3.1 Normalized Cut . 46 3.3.2 Normalized Cut as a spectral relaxation . 48 3.4 Small Clusters and the Small Set Expansion Hypothesis . 53 3.5 Computing Eigenvectors . 53 3.6 Multiple Clusters . 54 4 Concentration Inequalities, Scalar and Matrix Versions 55 4.1 Large Deviation Inequalities . 55 4.1.1 Sums of independent random variables . 55 4.2 Gaussian Concentration . 60 4.2.1 Spectral norm of a Wigner Matrix . 62 4.2.2 Talagrand's concentration inequality . 62 2 4.3 Other useful large deviation inequalities . 63 4.3.1 Additive Chernoff Bound . 63 4.3.2 Multiplicative Chernoff Bound . 63 4.3.3 Deviation bounds on χ2 variables . 63 4.4 Matrix Concentration . 64 4.5 Optimality of matrix concentration result for gaussian series . 66 4.5.1 An interesting observation regarding random matrices with independent matrices 68 4.6 A matrix concentration inequality for Rademacher Series . 69 4.6.1 A small detour on discrepancy theory . 69 4.6.2 Back to matrix concentration . 70 4.7 Other Open Problems . 75 4.7.1 Oblivious Sparse Norm-Approximating Projections . 75 4.7.2 k-lifts of graphs . 76 4.8 Another open problem . 77 5 Johnson-Lindenstrauss Lemma and Gordons Theorem 78 5.1 The Johnson-Lindenstrauss Lemma . 78 5.1.1 Optimality of the Johnson-Lindenstrauss Lemma . 80 5.1.2 Fast Johnson-Lindenstrauss . 80 5.2 Gordon's Theorem . 81 5.2.1 Gordon's Escape Through a Mesh Theorem . 83 5.2.2 Proof of Gordon's Theorem . 83 5.3 Sparse vectors and Low-rank matrices . 85 5.3.1 Gaussian width of k-sparse vectors . 85 5.3.2 The Restricted Isometry Property and a couple of open problems . 86 5.3.3 Gaussian width of rank-r matrices . 87 6 Compressed Sensing and Sparse Recovery 89 6.1 Duality and exact recovery . 91 6.2 Finding a dual certificate . 92 6.3 A different approach . 93 6.4 Partial Fourier matrices satisfying the Restricted Isometry Property . 94 6.5 Coherence and Gershgorin Circle Theorem . 94 6.5.1 Mutually Unbiased Bases . 95 6.5.2 Equiangular Tight Frames . 96 6.5.3 The Paley ETF . 97 6.6 The Kadison-Singer problem . 97 7 Group Testing and Error-Correcting Codes 98 7.1 Group Testing . 98 7.2 Some Coding Theory and the proof of Theorem 7.3 . 102 7.2.1 Boolean Classification . 103 7.2.2 The proof of Theorem 7.3 . 104 7.3 In terms of linear Bernoulli algebra . 105 3 7.3.1 Shannon Capacity . 105 7.3.2 The deletion channel . 106 8 Approximation Algorithms and Max-Cut 108 8.1 The Max-Cut problem . 108 8.2 Can αGW be improved? . 110 8.3 A Sums-of-Squares interpretation . 111 8.4 The Grothendieck Constant . 114 8.5 The Paley Graph . 115 8.6 An interesting conjecture regarding cuts and bisections . 115 9 Community detection and the Stochastic Block Model 117 9.1 Community Detection . 117 9.2 Stochastic Block Model . 117 9.3 What does the spike model suggest? . 117 9.3.1 Three of more communities . 119 9.4 Exact recovery . ..
Recommended publications
  • On the Linear Stability of Plane Couette Flow for an Oldroyd-B Fluid and Its
    J. Non-Newtonian Fluid Mech. 127 (2005) 169–190 On the linear stability of plane Couette flow for an Oldroyd-B fluid and its numerical approximation Raz Kupferman Institute of Mathematics, The Hebrew University, Jerusalem, 91904, Israel Received 14 December 2004; received in revised form 10 February 2005; accepted 7 March 2005 Abstract It is well known that plane Couette flow for an Oldroyd-B fluid is linearly stable, yet, most numerical methods predict spurious instabilities at sufficiently high Weissenberg number. In this paper we examine the reasons which cause this qualitative discrepancy. We identify a family of distribution-valued eigenfunctions, which have been overlooked by previous analyses. These singular eigenfunctions span a family of non- modal stress perturbations which are divergence-free, and therefore do not couple back into the velocity field. Although these perturbations decay eventually, they exhibit transient amplification during which their “passive" transport by shearing streamlines generates large cross- stream gradients. This filamentation process produces numerical under-resolution, accompanied with a growth of truncation errors. We believe that the unphysical behavior has to be addressed by fine-scale modelling, such as artificial stress diffusivity, or other non-local couplings. © 2005 Elsevier B.V. All rights reserved. Keywords: Couette flow; Oldroyd-B model; Linear stability; Generalized functions; Non-normal operators; Stress diffusion 1. Introduction divided into two main groups: (i) numerical solutions of the boundary value problem defined by the linearized system The linear stability of Couette flow for viscoelastic fluids (e.g. [2,3]) and (ii) stability analysis of numerical methods is a classical problem whose study originated with the pi- for time-dependent flows, with the objective of understand- oneering work of Gorodtsov and Leonov (GL) back in the ing how spurious solutions emerge in computations.
    [Show full text]
  • A Highly Efficient Spectral-Galerkin Method Based on Tensor Product for Fourth-Order Steklov Equation with Boundary Eigenvalue
    An et al. Journal of Inequalities and Applications (2016)2016:211 DOI 10.1186/s13660-016-1158-1 R E S E A R C H Open Access A highly efficient spectral-Galerkin method based on tensor product for fourth-order Steklov equation with boundary eigenvalue Jing An1,HaiBi1 and Zhendong Luo2* *Correspondence: [email protected] Abstract 2School of Mathematics and Physics, North China Electric Power In this study, a highly efficient spectral-Galerkin method is posed for the fourth-order University, No. 2, Bei Nong Road, Steklov equation with boundary eigenvalue. By making use of the spectral theory of Changping District, Beijing, 102206, compact operators and the error formulas of projective operators, we first obtain the China Full list of author information is error estimates of approximative eigenvalues and eigenfunctions. Then we build a 1 ∩ 2 available at the end of the article suitable set of basis functions included in H0() H () and establish the matrix model for the discrete spectral-Galerkin scheme by adopting the tensor product. Finally, we use some numerical experiments to verify the correctness of the theoretical results. MSC: 65N35; 65N30 Keywords: fourth-order Steklov equation with boundary eigenvalue; spectral-Galerkin method; error estimates; tensor product 1 Introduction Increasing attention has recently been paid to numerical approximations for Steklov equa- tions with boundary eigenvalue, arising in fluid mechanics, electromagnetism, etc. (see, e.g.,[–]). However, most existing work usually treated the second-order Steklov equa- tions with boundary eigenvalue and there are relatively few articles treating fourth-order ones. The fourth-order Steklov equations with boundary eigenvalue also have been used in both mathematics and physics, for example, the main eigenvalues play a very key role in the positivity-preserving properties for the biharmonic-operator under the border conditions w = w – χwν =on∂ (see []).
    [Show full text]
  • Chebyshev and Fourier Spectral Methods 2000
    Chebyshev and Fourier Spectral Methods Second Edition John P. Boyd University of Michigan Ann Arbor, Michigan 48109-2143 email: [email protected] http://www-personal.engin.umich.edu/jpboyd/ 2000 DOVER Publications, Inc. 31 East 2nd Street Mineola, New York 11501 1 Dedication To Marilyn, Ian, and Emma “A computation is a temptation that should be resisted as long as possible.” — J. P. Boyd, paraphrasing T. S. Eliot i Contents PREFACE x Acknowledgments xiv Errata and Extended-Bibliography xvi 1 Introduction 1 1.1 Series expansions .................................. 1 1.2 First Example .................................... 2 1.3 Comparison with finite element methods .................... 4 1.4 Comparisons with Finite Differences ....................... 6 1.5 Parallel Computers ................................. 9 1.6 Choice of basis functions .............................. 9 1.7 Boundary conditions ................................ 10 1.8 Non-Interpolating and Pseudospectral ...................... 12 1.9 Nonlinearity ..................................... 13 1.10 Time-dependent problems ............................. 15 1.11 FAQ: Frequently Asked Questions ........................ 16 1.12 The Chrysalis .................................... 17 2 Chebyshev & Fourier Series 19 2.1 Introduction ..................................... 19 2.2 Fourier series .................................... 20 2.3 Orders of Convergence ............................... 25 2.4 Convergence Order ................................. 27 2.5 Assumption of Equal Errors ...........................
    [Show full text]
  • Isospectral Families of High-Order Systems∗
    Isospectral Families of High-order Systems∗ Peter Lancaster† Department of Mathematics and Statistics University of Calgary Calgary T2N 1N4, Alberta, Canada and Uwe Prells School of Mechanical, Materials, Manufacturing Engineering & Management University of Nottingham Nottingham NG7 2RD United Kingdom November 28, 2006 Keywords: Matrix Polynomials, Isospectral Systems, Structure Preserving Transformations, Palindromic Symmetries. Abstract Earlier work of the authors concerning the generation of isospectral families of second or- der (vibrating) systems is generalized to higher-order systems (with no spectrum at infinity). Results and techniques are developed first for systems without symmetries, then with Hermi- tian symmetry and, finally, with palindromic symmetry. The construction of linearizations which retain such symmetries is discussed. In both cases, the notion of strictly isospectral families of systems is introduced - implying that properties of both the spectrum and the sign-characteristic are preserved. Open questions remain in the case of strictly isospectral families of palindromic systems. Intimate connections between Hermitian and unitary sys- tems are discussed in an Appendix. ∗The authors are grateful to an anonymous reviewer for perceptive comments which led to significant improve- ments in exposition. †The research of Peter Lancaster was partly supported by a grant from the Natural Sciences and Engineering Research Council of Canada 1 1 Introduction ` By a high-order system we mean an n × n matrix polynomial L(λ) = A`λ + ··· + A1λ + A0 n×n with coefficients in C . Such a polynomial with A` 6= 0 is said to have degree `, (sometimes referred to as “order” `) and “high-order” implies that ` ≥ 2. It will be assumed throughout that the leading coefficient A` is nonsingular.
    [Show full text]
  • Introduction to Spectral Methods
    Introduction to spectral methods Eric Gourgoulhon Laboratoire de l’Univers et de ses Th´eories (LUTH) CNRS / Observatoire de Paris Meudon, France Based on a collaboration with Silvano Bonazzola, Philippe Grandcl´ement,Jean-Alain Marck & J´eromeNovak [email protected] http://www.luth.obspm.fr 4th EU Network Meeting, Palma de Mallorca, Sept. 2002 1 Plan 1. Basic principles 2. Legendre and Chebyshev expansions 3. An illustrative example 4. Spectral methods in numerical relativity 2 1 Basic principles 3 Solving a partial differential equation Consider the PDE with boundary condition Lu(x) = s(x); x 2 U ½ IRd (1) Bu(y) = 0; y 2 @U; (2) where L and B are linear differential operators. Question: What is a numerical solution of (1)-(2)? Answer: It is a function u¯ which satisfies (2) and makes the residual R := Lu¯ ¡ s small. 4 What do you mean by “small” ? Answer in the framework of Method of Weighted Residuals (MWR): Search for solutions u¯ in a finite-dimensional sub-space PN of some Hilbert space W (typically a L2 space). Expansion functions = trial functions : basis of PN : (Á0;:::;ÁN ) XN u¯ is expanded in terms of the trial functions: u¯(x) = u˜n Án(x) n=0 Test functions : family of functions (Â0;:::;ÂN ) to define the smallness of the residual R, by means of the Hilbert space scalar product: 8 n 2 f0;:::;Ng; (Ân;R) = 0 5 Various numerical methods Classification according to the trial functions Án: Finite difference: trial functions = overlapping local polynomials of low order Finite element: trial functions = local smooth functions (polynomial
    [Show full text]
  • Math 660-Lecture 15: Spectral Methods: Fourier Pseduo-Spectral Method
    Math 660-Lecture 15: Spectral methods: Fourier pseduo-spectral method The topics here will be in the book `Spectral methods in Matlab' (Chap- ter 3); Another reference book is `Spectral methods in chemistry and physics.' 1 Spectral methods v.s. pseudo-spectral methods 1.1 Spectral method A spectral method is to represent the solution of a differential equation in terms of a basis of some vector space and then reduce the differential equations to an ODE system for the coefficients. In this sense, the (contin- uous) Fourier Transform and the Finite Element methods are all spectral methods. In particular, suppose we have some equation T (u) = f(x) where T is some temporal-spatial operator for the differential equation. Then, under the basis we have the error: X X R(x) = T ( ci(t)φi(x)) − f~iφi i i Pick test functions fχjg, which is the dual basis of fφi(x)g (if the space is a Hilbert space like in FEM, these two are identical) and we require hχj;Ri = 0: This then gives a system of equations for cj. • Advantages: Reduce differential operators to multiplications, and solv- ing ODEs might be easier. • Disadvatange: The multiplication in the spatial variable becomes con- volution. Then, we may have a big matrix-vector multiplication. 1.2 Pseudo-spectral method Here, represent the solution in terms of the basis but impose the equa- tion only at discrete points. As we can see, the pseudo-spectral method in some sense is a spectral method in discrete space. There are some typical pseudo-spectral methods 1 • Fourier Pseudo-spectral method with trigonometric basis • Chebyshev pseudo-spectral method with Chebyshev polynomial basis • Collocation methods with polynomial basis (especially for ODEs solvers); (Some people also call the pseudo-spectral method the collocation method.) For discrete points, we often choose χj = δ(x − xj), so that we require R(xj) = 0: The pseudo-spectral method may have efficient algorithms, which makes solving PDEs fast.
    [Show full text]
  • Validated and Numerically Efficient Chebyshev Spectral Methods For
    Validated and numerically efficient Chebyshev spectral methods for linear ordinary differential equations Florent Bréhard, Nicolas Brisebarre, Mioara Joldes To cite this version: Florent Bréhard, Nicolas Brisebarre, Mioara Joldes. Validated and numerically efficient Chebyshev spectral methods for linear ordinary differential equations. ACM Transactions on Mathematical Soft- ware, Association for Computing Machinery, 2018, 44 (4), pp.44:1-44:42. 10.1145/3208103. hal- 01526272v3 HAL Id: hal-01526272 https://hal.archives-ouvertes.fr/hal-01526272v3 Submitted on 5 Jun 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Validated and numerically efficient Chebyshev spectral methods for linear ordinary differential equations Florent Bréhard∗ Nicolas Brisebarrey Mioara Joldeşz Abstract In this work we develop a validated numerics method for the solu- tion of linear ordinary differential equations (LODEs). A wide range of algorithms (i.e., Runge-Kutta, collocation, spectral methods) exist for nu- merically computing approximations of the solutions. Most of these come with proofs of asymptotic convergence, but usually, provided error bounds are non-constructive. However, in some domains like critical systems and computer-aided mathematical proofs, one needs validated effective error bounds. We focus on both the theoretical and practical complexity anal- ysis of a so-called a posteriori quasi-Newton validation method, which mainly relies on a fixed-point argument of a contracting map.
    [Show full text]
  • Applying an Iterative Method Numerically to Solve N × N Matrix
    Applying an iterative method numerically to solve n × n rsta.royalsocietypublishing.org matrix Wiener–Hopf equations with exponential factors Research Matthew J. Priddin1, Anastasia V. Kisil2 Article submitted to journal and Lorna J. Ayton1 1DAMTP, University of Cambridge, Cambridge, CB3 Subject Areas: 0WA, UK 2Department of Mathematics, The University wave motion, applied mathematics, of Manchester, Manchester, M13 9PL, UK differential equations Keywords: This paper presents a generalisation of a recent iterative approach to solving a class of 2 2 Wiener–Hopf equations, × matrix Wiener–Hopf equations involving exponential Riemann-Hilbert problem, iterative factors. We extend the method to square matrices of methods, scattering, n-part arbitrary dimension n, as arise in mixed boundary boundaries value problems with n junctions. To demonstrate the method we consider the classical problem of Author for correspondence: scattering a plane wave by a set of collinear plates. The results are compared to other known Matthew Priddin methods. We describe an effective implementation e-mail: [email protected] using a spectral method to compute the required Cauchy transforms. The approach is ideally suited to obtaining far-field directivity patterns of utility to applications. Convergence in iteration is fastest for large wavenumbers, but remains practical at modest wavenumbers to achieve a high degree of accuracy. 1. Introduction The Wiener–Hopf technique [1] provides a powerful tool to approach varied problems such as the solution of integral equations and random processes. Applications may so be found in such diverse fields as aeroacoustics [2], metamaterials [3], geophysics [4], crack propagation [5] and financial mathematics [6]. The mathematical problem underlying the Wiener– Hopf technique may be simply stated: find functions that are analytic in adjacent (or overlapping) regions of a complex manifold that satisfy a prescribed jump condition at their common boundary.
    [Show full text]
  • Msc in Applied Mathematics
    MSc in Applied Mathematics Title: Mathematical and Technological Aspects of Quantum Information Processing Author: Jordi Tura i Brugués Advisor: Sebastià Xambó i Descamps Co-Advisor: Antonio Acín dal Maschio Department: Departament de Matemàtica Aplicada II Academic year: 2010-2011 Universitat Polit`ecnicade Catalunya Facultat de Matem`atiquesi Estad´ıstica Master's Degree Thesis Mathematical and Technological Aspects of Quantum Information Processing Jordi Tura i Brugu´es Advisor: Sebasti`aXamb´oi Descamps Co-Advisor: Antonio Ac´ındal Maschio Departament de Matem`aticaAplicada II Abstract Keywords: Quantum, information, processing, computation, coding, cryptography, entanglement, field theory, entropy, key distribution, protocols. Since its genesis, quantum mechanics has proved to be a very accurate model for predicting the behavior of the world below the nanoscale. However, crucial breakthroughs in technology were needed in order to be able to effectively access and manipulate such small magnitudes. During the last twenty years, the field of quantum information processing has experienced a growing interest, in its many variants, both theoretically and practically. Despite being still at a very basic stage, expectations are high. The uniqueness of quantum phenomena (superposition of states, creation of entanglement, etc.) have no classical analogue and allow novelties such as another paradigm of computation, more secure com- munications, quantum teleportation, quantum dense coding, etc. which are presented and analyzed here. The aim of this Thesis is to present in a unified way the main mathematical methods used in quantum information processing, as well as the state of the art of their corresponding technological implementa- tions. Our contributions are based in making a self-contained presentation; seeking completeness, that is, treating the most relevant fields of research involved, focusing on their relations; and picking the most relevant, insightful references, given the quantity of literature produced in this field.
    [Show full text]
  • Computing Spectral Measures and Spectral Types
    Commun. Math. Phys. 384, 433–501 (2021) Communications in Digital Object Identifier (DOI) https://doi.org/10.1007/s00220-021-04072-4 Mathematical Physics Computing Spectral Measures and Spectral Types Matthew J. Colbrook Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, UK. E-mail: [email protected] Received: 26 February 2020 / Accepted: 16 March 2021 Published online: 11 April 2021 – © The Author(s) 2021 Abstract: Spectral measures arise in numerous applications such as quantum mechan- ics, signal processing, resonance phenomena, and fluid stability analysis. Similarly, spectral decompositions (into pure point, absolutely continuous and singular contin- uous parts) often characterise relevant physical properties such as the long-time dy- namics of quantum systems. Despite new results on computing spectra, there remains no general method able to compute spectral measures or spectral decompositions of infinite-dimensional normal operators. Previous efforts have focused on specific exam- ples where analytical formulae are available (or perturbations thereof) or on classes of operators that carry a lot of structure. Hence the general computational problem is predominantly open. We solve this problem by providing the first set of general algo- rithms that compute spectral measures and decompositions of a wide class of operators. Given a matrix representation of a self-adjoint or unitary operator, such that each col- umn decays at infinity at a known asymptotic rate, we show how to compute spectral measures and decompositions. We discuss how these methods allow the computation of objects such as the functional calculus, and how they generalise to a large class of partial differential operators, allowing, for example, solutions to evolution PDEs such as the linear Schrödinger equation on L2(Rd ).
    [Show full text]
  • Analysis of Parametric Models — Linear Methods and Approximations
    Analysis of parametric models — linear methods and approximations∗ Hermann G. Matthies* and Roger Ohayono *Institute of Scientific Computing Technische Universität Braunschweig 38092 Braunschweig, Germany e-mail: [email protected] oLaboratoire de Mécanique des Structures et des Systèmes Couplés Conservatoire National des Arts et Métiers (CNAM) 75141 Paris Cedex 03, France Abstract Parametric models in vector spaces are shown to possess an associated linear map. This linear operator leads directly to reproducing kernel Hilbert spaces and affine- / linear- representations in terms of tensor products. From the associated linear map analogues of covariance or rather correlation operators can be formed. The associated linear map in fact provides a factorisation of the correlation. Its spectral decomposition, and the associated Karhunen-Loève- or proper orthogonal decomposition in a tensor product follow directly. It is shown that all factorisations of a certain class are unitarily equivalent, as well as that every factorisation induces a different representation, and vice versa. A completely equivalent spectral and factorisation analysis can be carried out in kernel space. The relevance of these abstract constructions is shown on a number of mostly familiar examples, thus unifying many such constructions under one the- oretical umbrella. From the factorisation one obtains tensor representations, which may be cascaded, leading to tensors of higher degree. When carried over to a dis- arXiv:1806.01101v2 [math.NA] 17 Jun 2018 cretised level in the form of a model order reduction, such factorisations allow very sparse low-rank approximations which lead to very efficient computations especially in high dimensions. Key words: parametric models, reproducing kernel Hilbert space, correlation, factorisation, spectral decomposition, representation MSC subject classifications: 35B30, 37M99, 41A05, 41A45, 41A63, 60G20, 60G60, 65J99, 93A30 ∗Partly supported by the Deutsche Forschungsgemeinschaft (DFG) through SFB 880 and SPP 1886.
    [Show full text]
  • Spectral Algorithms
    Spectral Algorithms Ravindran Kannan Santosh Vempala August, 2009 ii Summary. Spectral methods refer to the use of eigenvalues, eigenvectors, sin- gular values and singular vectors. They are widely used in Engineering, Ap- plied Mathematics and Statistics. More recently, spectral methods have found numerous applications in Computer Science to \discrete" as well \continuous" problems. This book describes modern applications of spectral methods, and novel algorithms for estimating spectral parameters. In the first part of the book, we present applications of spectral methods to problems from a variety of topics including combinatorial optimization, learning and clustering. The second part of the book is motivated by efficiency considerations. A fea- ture of many modern applications is the massive amount of input data. While sophisticated algorithms for matrix computations have been developed over a century, a more recent development is algorithms based on \sampling on the fly” from massive matrices. Good estimates of singular values and low rank ap- proximations of the whole matrix can be provably derived from a sample. Our main emphasis in the second part of the book is to present these sampling meth- ods with rigorous error bounds. We also present recent extensions of spectral methods from matrices to tensors and their applications to some combinatorial optimization problems. Contents I Applications 1 1 The Best-Fit Subspace 3 1.1 Singular Value Decomposition . .4 1.2 Algorithms for computing the SVD . .8 1.3 The k-variance problem . .8 1.4 Discussion . 11 2 Mixture Models 13 2.1 Probabilistic separation . 14 2.2 Geometric separation . 14 2.3 Spectral Projection .
    [Show full text]