Functional and Structured Tensor Analysis for Engineers

Total Page:16

File Type:pdf, Size:1020Kb

Functional and Structured Tensor Analysis for Engineers UNM BOOK DRAFT September 4, 2003 5:21 pm Functional and Structured Tensor Analysis for Engineers A casual (intuition-based) introduction to vector and tensor analysis with reviews of popular notations used in contemporary materials modeling R. M. Brannon University of New Mexico, Albuquerque Copyright is reserved. Individual copies may be made for personal use. No part of this document may be reproduced for profit. Contact author at [email protected] NOTE: When using Adobe’s “acrobat reader” to view this document, the page numbers in acrobat will not coincide with the page numbers shown at the bottom of each page of this document. Note to draft readers: The most useful textbooks are the ones with fantastic indexes. The book’s index is rather new and still under construction. It would really help if you all could send me a note whenever you discover that an important entry is miss- ing from this index. I’ll be sure to add it. This work is a community effort. Let’s try to make this document helpful to others. FUNCTIONAL AND STRUCTURED TENSOR ANALYSIS FOR ENGINEERS A casual (intuition-based) introduction to vector and tensor analysis with reviews of popular notations used in contemporary materials modeling Rebecca M. Brannon† †University of New Mexico Adjunct professor [email protected] Abstract Elementary vector and tensor analysis concepts are reviewed in a manner that proves useful for higher-order tensor analysis of anisotropic media. In addition to reviewing basic matrix and vector analysis, the concept of a tensor is cov- ered by reviewing and contrasting numerous different definition one might see in the literature for the term “tensor.” Basic vector and tensor operations are provided, as well as some lesser-known operations that are useful in materials modeling. Considerable space is devoted to “philosophical” discussions about relative merits of the many (often conflicting) tensor notation systems in popu- lar use. ii Acknowledgments An indeterminately large (but, of course, countable) set of people who have offered advice, encouragement, and fantastic suggestions throughout the years that I’ve spent writing this document. I say years because the seeds for this document were sown back in 1986, when I was a co-op student at Los Alamos National Laboratories, and I made the mistake of asking my supervisor, Norm Johnson, “what’s a tensor?” His reply? “read the appendix of R.B. “Bob” Bird’s book, Dynamics of Polymeric Liquids. I did — and got hooked. Bird’s appendix (which has nothing to do with polymers) is an outstanding and succinct summary of vector and tensor analysis. Reading it motivated me, as an under- graduate, to take my first graduate level continuum mechanics class from Dr. H.L. “Buck” Schreyer at the University of New Mexico. Buck Schreyer used multiple underlines beneath symbols as a teaching aid to help his students keep track of the different kinds of strange new objects (tensors) appearing in his lectures, and I have adopted his notation in this document. Later taking Buck’s beginning and advanced finite element classes further improved my command of matrix analysis and partial differential equations. Buck’s teach- ing pace was fast, so we all struggled to keep up. Buck was careful to explain that he would often cover esoteric subjects principally to enable us to effectively read the litera- ture, though sometimes merely to give us a different perspective on what we had already learned. Buck armed us with a slew of neat tricks or fascinating insights that were rarely seen in any publications. I often found myself “secretly” using Buck’s tips in my own work, and then struggling to figure out how to explain how I was able to come up with these “miracle instant answers” — the effort to reproduce my results using conventional (better known) techniques helped me learn better how to communicate difficult concepts to a broader audience. While taking Buck’s continuum mechanics course, I simulta- neously learned variational mechanics from Fred Ju (also at UNM), which was fortunate timing because Dr. Ju’s refreshing and careful teaching style forced me to make enlighten- ing connections between his class and Schreyer’s class. Taking thermodynamics from A. Razanni (UNM) helped me improve my understanding of partial derivatives and their applications (furthermore, my interactions with Buck Schreyer helped me figure out how gas thermodynamics equations generalized to the solid mechanics arena). Following my undergraduate experiences at UNM, I was fortunate to learn advanced applications of con- tinuum mechanics from my Ph.D advisor, Prof. Walt Drugan (U. Wisconsin), who intro- duced me to even more (often completely new) viewpoints to add to my tensor analysis toolbelt. While at Wisconsin, I took an elasticity course from Prof. Chen, who was enam- oured of doing all proofs entirely in curvilinear notation, so I was forced to improve my abilities in this area (curvilinear analysis is not covered in this book, but it may be found in a separate publication, Ref. [6]. A slightly different spin on curvilinear analysis came when I took Arthur Lodge’s “Elastic Liquids” class. My third continuum mechanics course, this time taught by Millard Johnson (U. Wisc), introduced me to the usefulness of “Rossetta stone” type derivations of classic theorems, done using multiple notations to make them clear to every reader. It was here where I conceded that no single notation is superior, and I had better become darn good at them all. At Wisconsin, I took a class on Greens functions and boundary value problems from the noted mathematician R. Dickey, who really drove home the importance of projection operations in physical applications, and instilled in me the irresistible habit of examining operators for their properties and iii classifying them as outlined in our class textbook [12]; it was Dickey who finally got me into the habit of looking for analogies between seemingly unrelated operators and sets so that my strong knowledge. Dickey himself got sideswiped by this habit when I solved one of his exam questions by doing it using a technique that I had learned in Buck Schreyer’s continuum mechanics class and which I realized would also work on the exam question by merely re-interpreting the vector dot product as the inner product that applies for continu- ous functions. As I walked into my Ph.D. defense, I warned Dickey (who was on my com- mittee) that my thesis was really just a giant application of the projection theorem, and he replied “most are, but you are distinguished by recognizing the fact!” Even though neither this book nor very many of my other publications (aside from Ref. [6], of course) employ curvilinear notation, my exposure to it has been invaluable to lend insight to the relation- ship between so-called “convected coordinates” and “unconvected reference spaces” often used in materials modeling. Having gotten my first exposure to tensor analysis from read- ing Bird’s polymer book, I naturally felt compelled to take his macromolecular fluid dynamics course at U. Wisc, which solidified several concepts further. Bird’s course was immediately followed by an applied analysis course, taught by ____, where more correct “mathematician’s” viewpoints on tensor analysis were drilled into me (the textbook for this course [17] is outstanding, and don’t be swayed by the fact that “chemical engineer- ing” is part of its title — the book applies to any field of physics). These and numerous other academic mentors I’ve had throughout my career have given me a wonderfully bal- anced set of analysis tools, and I wish I could thank them enough. For the longest time, this “Acknowledgement” section said only “Acknowledgements to be added. Stay tuned...” Assigning such low priority to the acknowledgements section was a gross tactical error on my part. When my colleagues offered assistance and sugges- tions in the earliest days of error-ridden rough drafts of this book, I thought to myself “I should thank them in my acknowledgements section.” A few years later, I sit here trying to recall the droves of early reviewers. I remember contributions from Glenn Randers-Pher- son because his advice for one of my other publications proved to be incredibly helpful, and he did the same for this more elementary document as well. A few folks (Mark Chris- ten, Allen Robinson, Stewart Silling, Paul Taylor, Tim Trucano) in my former department at Sandia National Labs also came forward with suggestions or helpful discussions that were incorporated into this book. While in my new department at Sandia National Labora- tories, I continued to gain new insight, especially from Dan Segalman and Bill Scherz- inger. Part of what has driven me to continue to improve this document has been the numer- ous encouraging remarks (approximately one per week) that I have received from researchers and students all over the world who have stumbled upon the pdf draft version of this document that I originally wrote as a student’s guide when I taught Continuum Mechanics at UNM. I don’t recall the names of people who sent me encouraging words in the early days, but some recent folks are Ricardo Colorado, Vince Owens, Dave Dooli- nand Mr. Jan Cox. Jan was especially inspiring because he was so enthusiastic about this work that he spent an entire afternoon disscussing it with me after a business trip I made to his home city, Oakland CA. Even some professors [such as Lynn Bennethum (U. Colo- rado), Ron Smelser (U. Idaho), Tom Scarpas (TU Delft), Sanjay Arwad (JHU), Kaspar William (U. Colorado), Walt Gerstle (U. New Mexico)] have told me that they have iv Copyright is reserved.
Recommended publications
  • FUNCTIONAL ANALYSIS 1. Banach and Hilbert Spaces in What
    FUNCTIONAL ANALYSIS PIOTR HAJLASZ 1. Banach and Hilbert spaces In what follows K will denote R of C. Definition. A normed space is a pair (X, k · k), where X is a linear space over K and k · k : X → [0, ∞) is a function, called a norm, such that (1) kx + yk ≤ kxk + kyk for all x, y ∈ X; (2) kαxk = |α|kxk for all x ∈ X and α ∈ K; (3) kxk = 0 if and only if x = 0. Since kx − yk ≤ kx − zk + kz − yk for all x, y, z ∈ X, d(x, y) = kx − yk defines a metric in a normed space. In what follows normed paces will always be regarded as metric spaces with respect to the metric d. A normed space is called a Banach space if it is complete with respect to the metric d. Definition. Let X be a linear space over K (=R or C). The inner product (scalar product) is a function h·, ·i : X × X → K such that (1) hx, xi ≥ 0; (2) hx, xi = 0 if and only if x = 0; (3) hαx, yi = αhx, yi; (4) hx1 + x2, yi = hx1, yi + hx2, yi; (5) hx, yi = hy, xi, for all x, x1, x2, y ∈ X and all α ∈ K. As an obvious corollary we obtain hx, y1 + y2i = hx, y1i + hx, y2i, hx, αyi = αhx, yi , Date: February 12, 2009. 1 2 PIOTR HAJLASZ for all x, y1, y2 ∈ X and α ∈ K. For a space with an inner product we define kxk = phx, xi . Lemma 1.1 (Schwarz inequality).
    [Show full text]
  • Linear Spaces
    Chapter 2 Linear Spaces Contents FieldofScalars ........................................ ............. 2.2 VectorSpaces ........................................ .............. 2.3 Subspaces .......................................... .............. 2.5 Sumofsubsets........................................ .............. 2.5 Linearcombinations..................................... .............. 2.6 Linearindependence................................... ................ 2.7 BasisandDimension ..................................... ............. 2.7 Convexity ............................................ ............ 2.8 Normedlinearspaces ................................... ............... 2.9 The `p and Lp spaces ............................................. 2.10 Topologicalconcepts ................................... ............... 2.12 Opensets ............................................ ............ 2.13 Closedsets........................................... ............. 2.14 Boundedsets......................................... .............. 2.15 Convergence of sequences . ................... 2.16 Series .............................................. ............ 2.17 Cauchysequences .................................... ................ 2.18 Banachspaces....................................... ............... 2.19 Completesubsets ....................................... ............. 2.19 Transformations ...................................... ............... 2.21 Lineartransformations.................................. ................ 2.21
    [Show full text]
  • General Inner Product & Fourier Series
    General Inner Products 1 General Inner Product & Fourier Series Advanced Topics in Linear Algebra, Spring 2014 Cameron Braithwaite 1 General Inner Product The inner product is an algebraic operation that takes two vectors of equal length and com- putes a single number, a scalar. It introduces a geometric intuition for length and angles of vectors. The inner product is a generalization of the dot product which is the more familiar operation that's specific to the field of real numbers only. Euclidean space which is limited to 2 and 3 dimensions uses the dot product. The inner product is a structure that generalizes to vector spaces of any dimension. The utility of the inner product is very significant when proving theorems and runing computations. An important aspect that can be derived is the notion of convergence. Building on convergence we can move to represenations of functions, specifally periodic functions which show up frequently. The Fourier series is an expansion of periodic functions specific to sines and cosines. By the end of this paper we will be able to use a Fourier series to represent a wave function. To build toward this result many theorems are stated and only a few have proofs while some proofs are trivial and left for the reader to save time and space. Definition 1.11 Real Inner Product Let V be a real vector space and a; b 2 V . An inner product on V is a function h ; i : V x V ! R satisfying the following conditions: (a) hαa + α0b; ci = αha; ci + α0hb; ci (b) hc; αa + α0bi = αhc; ai + α0hc; bi (c) ha; bi = hb; ai (d) ha; aiis a positive real number for any a 6= 0 Definition 1.12 Complex Inner Product Let V be a vector space.
    [Show full text]
  • Function Space Tensor Decomposition and Its Application in Sports Analytics
    East Tennessee State University Digital Commons @ East Tennessee State University Electronic Theses and Dissertations Student Works 12-2019 Function Space Tensor Decomposition and its Application in Sports Analytics Justin Reising East Tennessee State University Follow this and additional works at: https://dc.etsu.edu/etd Part of the Applied Statistics Commons, Multivariate Analysis Commons, and the Other Applied Mathematics Commons Recommended Citation Reising, Justin, "Function Space Tensor Decomposition and its Application in Sports Analytics" (2019). Electronic Theses and Dissertations. Paper 3676. https://dc.etsu.edu/etd/3676 This Thesis - Open Access is brought to you for free and open access by the Student Works at Digital Commons @ East Tennessee State University. It has been accepted for inclusion in Electronic Theses and Dissertations by an authorized administrator of Digital Commons @ East Tennessee State University. For more information, please contact [email protected]. Function Space Tensor Decomposition and its Application in Sports Analytics A thesis presented to the faculty of the Department of Mathematics East Tennessee State University In partial fulfillment of the requirements for the degree Master of Science in Mathematical Sciences by Justin Reising December 2019 Jeff Knisley, Ph.D. Nicole Lewis, Ph.D. Michele Joyner, Ph.D. Keywords: Sports Analytics, PCA, Tensor Decomposition, Functional Analysis. ABSTRACT Function Space Tensor Decomposition and its Application in Sports Analytics by Justin Reising Recent advancements in sports information and technology systems have ushered in a new age of applications of both supervised and unsupervised analytical techniques in the sports domain. These automated systems capture large volumes of data points about competitors during live competition.
    [Show full text]
  • Using Functional Distance Measures When Calibrating Journey-To-Crime Distance Decay Algorithms
    USING FUNCTIONAL DISTANCE MEASURES WHEN CALIBRATING JOURNEY-TO-CRIME DISTANCE DECAY ALGORITHMS A Thesis Submitted to the Graduate Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements for the degree of Master of Natural Sciences in The Interdepartmental Program of Natural Sciences by Joshua David Kent B.S., Louisiana State University, 1994 December 2003 ACKNOWLEDGMENTS The work reported in this research was partially supported by the efforts of Dr. James Mitchell of the Louisiana Department of Transportation and Development - Information Technology section for donating portions of the Geographic Data Technology (GDT) Dynamap®-Transportation data set. Additional thanks are paid for the thirty-five residence of East Baton Rouge Parish who graciously supplied the travel data necessary for the successful completion of this study. The author also wishes to acknowledge the support expressed by Dr. Michael Leitner, Dr. Andrew Curtis, Mr. DeWitt Braud, and Dr. Frank Cartledge - their efforts helped to make this thesis possible. Finally, the author thanks his wonderful wife and supportive family for their encouragement and tolerance. ii TABLE OF CONTENTS ACKNOWLEDGMENTS .............................................................................................................. ii LIST OF TABLES.......................................................................................................................... v LIST OF FIGURES ......................................................................................................................
    [Show full text]
  • Calculus of Variations
    MIT OpenCourseWare http://ocw.mit.edu 16.323 Principles of Optimal Control Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.323 Lecture 5 Calculus of Variations • Calculus of Variations • Most books cover this material well, but Kirk Chapter 4 does a particularly nice job. • See here for online reference. x(t) x*+ αδx(1) x*- αδx(1) x* αδx(1) −αδx(1) t t0 tf Figure by MIT OpenCourseWare. Spr 2008 16.323 5–1 Calculus of Variations • Goal: Develop alternative approach to solve general optimization problems for continuous systems – variational calculus – Formal approach will provide new insights for constrained solutions, and a more direct path to the solution for other problems. • Main issue – General control problem, the cost is a function of functions x(t) and u(t). � tf min J = h(x(tf )) + g(x(t), u(t), t)) dt t0 subject to x˙ = f(x, u, t) x(t0), t0 given m(x(tf ), tf ) = 0 – Call J(x(t), u(t)) a functional. • Need to investigate how to find the optimal values of a functional. – For a function, we found the gradient, and set it to zero to find the stationary points, and then investigated the higher order derivatives to determine if it is a maximum or minimum. – Will investigate something similar for functionals. June 18, 2008 Spr 2008 16.323 5–2 • Maximum and Minimum of a Function – A function f(x) has a local minimum at x� if f(x) ≥ f(x �) for all admissible x in �x − x�� ≤ � – Minimum can occur at (i) stationary point, (ii) at a boundary, or (iii) a point of discontinuous derivative.
    [Show full text]
  • Fact Sheet Functional Analysis
    Fact Sheet Functional Analysis Literature: Hackbusch, W.: Theorie und Numerik elliptischer Differentialgleichungen. Teubner, 1986. Knabner, P., Angermann, L.: Numerik partieller Differentialgleichungen. Springer, 2000. Triebel, H.: H¨ohere Analysis. Harri Deutsch, 1980. Dobrowolski, M.: Angewandte Funktionalanalysis, Springer, 2010. 1. Banach- and Hilbert spaces Let V be a real vector space. Normed space: A norm is a mapping k · k : V ! [0; 1), such that: kuk = 0 , u = 0; (definiteness) kαuk = jαj · kuk; α 2 R; u 2 V; (positive scalability) ku + vk ≤ kuk + kvk; u; v 2 V: (triangle inequality) The pairing (V; k · k) is called a normed space. Seminorm: In contrast to a norm there may be elements u 6= 0 such that kuk = 0. It still holds kuk = 0 if u = 0. Comparison of two norms: Two norms k · k1, k · k2 are called equivalent if there is a constant C such that: −1 C kuk1 ≤ kuk2 ≤ Ckuk1; u 2 V: If only one of these inequalities can be fulfilled, e.g. kuk2 ≤ Ckuk1; u 2 V; the norm k · k1 is called stronger than the norm k · k2. k · k2 is called weaker than k · k1. Topology: In every normed space a canonical topology can be defined. A subset U ⊂ V is called open if for every u 2 U there exists a " > 0 such that B"(u) = fv 2 V : ku − vk < "g ⊂ U: Convergence: A sequence vn converges to v w.r.t. the norm k · k if lim kvn − vk = 0: n!1 1 A sequence vn ⊂ V is called Cauchy sequence, if supfkvn − vmk : n; m ≥ kg ! 0 for k ! 1.
    [Show full text]
  • Inner Product Spaces
    CHAPTER 6 Woman teaching geometry, from a fourteenth-century edition of Euclid’s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition and scalar multiplication) of R2 and R3. We ignored other important features, such as the notions of length and angle. These ideas are embedded in the concept we now investigate, inner products. Our standing assumptions are as follows: 6.1 Notation F, V F denotes R or C. V denotes a vector space over F. LEARNING OBJECTIVES FOR THIS CHAPTER Cauchy–Schwarz Inequality Gram–Schmidt Procedure linear functionals on inner product spaces calculating minimum distance to a subspace Linear Algebra Done Right, third edition, by Sheldon Axler 164 CHAPTER 6 Inner Product Spaces 6.A Inner Products and Norms Inner Products To motivate the concept of inner prod- 2 3 x1 , x 2 uct, think of vectors in R and R as x arrows with initial point at the origin. x R2 R3 H L The length of a vector in or is called the norm of x, denoted x . 2 k k Thus for x .x1; x2/ R , we have The length of this vector x is p D2 2 2 x x1 x2 . p 2 2 x1 x2 . k k D C 3 C Similarly, if x .x1; x2; x3/ R , p 2D 2 2 2 then x x1 x2 x3 . k k D C C Even though we cannot draw pictures in higher dimensions, the gener- n n alization to R is obvious: we define the norm of x .x1; : : : ; xn/ R D 2 by p 2 2 x x1 xn : k k D C C The norm is not linear on Rn.
    [Show full text]
  • Calculus of Variations Raju K George, IIST
    Calculus of Variations Raju K George, IIST Lecture-1 In Calculus of Variations, we will study maximum and minimum of a certain class of functions. We first recall some maxima/minima results from the classical calculus. Maxima and Minima Let X and Y be two arbitrary sets and f : X Y be a well-defined function having domain → X and range Y . The function values f(x) become comparable if Y is IR or a subset of IR. Thus, optimization problem is valid for real valued functions. Let f : X IR be a real valued → function having X as its domain. Now x X is said to be maximum point for the function f if 0 ∈ f(x ) f(x) x X. The value f(x ) is called the maximum value of f. Similarly, x X is 0 ≥ ∀ ∈ 0 0 ∈ said to be a minimum point for the function f if f(x ) f(x) x X and in this case f(x ) is 0 ≤ ∀ ∈ 0 the minimum value of f. Sufficient condition for having maximum and minimum: Theorem (Weierstrass Theorem) Let S IR and f : S IR be a well defined function. Then f will have a maximum/minimum ⊆ → under the following sufficient conditions. 1. f : S IR is a continuous function. → 2. S IR is a bound and closed (compact) subset of IR. ⊂ Note that the above conditions are just sufficient conditions but not necessary. Example 1: Let f : [ 1, 1] IR defined by − → 1 x = 0 f(x)= − x x = 0 | | 6 1 −1 +1 −1 Obviously f(x) is not continuous at x = 0.
    [Show full text]
  • Collaborative Fashion Recommendation: a Functional Tensor Factorization Approach
    Collaborative Fashion Recommendation: A Functional Tensor Factorization Approach Yang Hu†, Xi Yi‡, Larry S. Davis§ Institute for Advanced Computer Studies, University of Maryland, College Park, MD 20742 [email protected]†, [email protected]‡, [email protected]§ ABSTRACT With the rapid expansion of online shopping for fashion products, effective fashion recommendation has become an increasingly important problem. In this work, we study the problem of personalized outfit recommendation, i.e. auto- matically suggesting outfits to users that fit their personal (a) Users 1 fashion preferences. Unlike existing recommendation sys- tems that usually recommend individual items, we suggest sets of items, which interact with each other, to users. We propose a functional tensor factorization method to model the interactions between user and fashion items. To effec- tively utilize the multi-modal features of the fashion items, we use a gradient boosting based method to learn nonlinear (b) Users 2 functions to map the feature vectors from the feature space into some low dimensional latent space. The effectiveness of the proposed algorithm is validated through extensive ex- periments on real world user data from a popular fashion- focused social network. Categories and Subject Descriptors (c) Users 3 H.3.3 [Information Search and Retrieval]: Retrieval models; I.2.6 [Learning]: Knowledge acquisition Figure 1: Examples of fashion sets created by three Keywords Polyvore users. Different users have different style preferences. Our task is to automatically recom- Recommendation systems; Collaborative filtering; Tensor mend outfits to users that fit their personal taste. factorization; Learning to rank; Gradient boosting 1. INTRODUCTION book3, where people showcase their personal styles and con- With the proliferation of social networks, people share al- nect to others that share similar fashion taste.
    [Show full text]
  • An Introduction to Some Aspects of Functional Analysis
    An introduction to some aspects of functional analysis Stephen Semmes Rice University Abstract These informal notes deal with some very basic objects in functional analysis, including norms and seminorms on vector spaces, bounded linear operators, and dual spaces of bounded linear functionals in particular. Contents 1 Norms on vector spaces 3 2 Convexity 4 3 Inner product spaces 5 4 A few simple estimates 7 5 Summable functions 8 6 p-Summable functions 9 7 p =2 10 8 Bounded continuous functions 11 9 Integral norms 12 10 Completeness 13 11 Bounded linear mappings 14 12 Continuous extensions 15 13 Bounded linear mappings, 2 15 14 Bounded linear functionals 16 1 15 ℓ1(E)∗ 18 ∗ 16 c0(E) 18 17 H¨older’s inequality 20 18 ℓp(E)∗ 20 19 Separability 22 20 An extension lemma 22 21 The Hahn–Banach theorem 23 22 Complex vector spaces 24 23 The weak∗ topology 24 24 Seminorms 25 25 The weak∗ topology, 2 26 26 The Banach–Alaoglu theorem 27 27 The weak topology 28 28 Seminorms, 2 28 29 Convergent sequences 30 30 Uniform boundedness 31 31 Examples 32 32 An embedding 33 33 Induced mappings 33 34 Continuity of norms 34 35 Infinite series 35 36 Some examples 36 37 Quotient spaces 37 38 Quotients and duality 38 39 Dual mappings 39 2 40 Second duals 39 41 Continuous functions 41 42 Bounded sets 42 43 Hahn–Banach, revisited 43 References 44 1 Norms on vector spaces Let V be a vector space over the real numbers R or the complex numbers C.
    [Show full text]
  • The Sparse Manifold Transform
    The Sparse Manifold Transform Yubei Chen1,2 Dylan M Paiton1,3 Bruno A Olshausen1,3,4 1Redwood Center for Theoretical Neuroscience 2Department of Electrical Engineering and Computer Science 3Vision Science Graduate Group 4Helen Wills Neuroscience Institute & School of Optometry University of California, Berkeley Berkeley, CA 94720 [email protected] Abstract We present a signal representation framework called the sparse manifold transform that combines key ideas from sparse coding, manifold learning, and slow feature analysis. It turns non-linear transformations in the primary sensory signal space into linear interpolations in a representational embedding space while maintaining approximate invertibility. The sparse manifold transform is an unsupervised and generative framework that explicitly and simultaneously models the sparse dis- creteness and low-dimensional manifold structure found in natural scenes. When stacked, it also models hierarchical composition. We provide a theoretical descrip- tion of the transform and demonstrate properties of the learned representation on both synthetic data and natural videos. 1 Introduction Inspired by Pattern Theory [40], we attempt to model three important and pervasive patterns in natural signals: sparse discreteness, low dimensional manifold structure and hierarchical composition. Each of these concepts have been individually explored in previous studies. For example, sparse coding [43, 44] and ICA [5, 28] can learn sparse and discrete elements that make up natural signals. Manifold learning [56, 48, 38, 4] was proposed to model and visualize low-dimensional continuous transforms such as smooth 3D rotations or translations of a single discrete element. Deformable, compositional models [60, 18] allow for a hierarchical composition of components into a more abstract representation.
    [Show full text]