Multilinear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

Multilinear Algebra Appendix A Multilinear Algebra This chapter presents concepts from multilinear algebra based on the basic properties of finite dimensional vector spaces and linear maps. The primary aim of the chapter is to give a concise introduction to alternating tensors which are necessary to define differential forms on manifolds. Many of the stated definitions and propositions can be found in Lee [1], Chaps. 11, 12 and 14. Some definitions and propositions are complemented by short and simple examples. First, in Sect. A.1 dual and bidual vector spaces are discussed. Subsequently, in Sects. A.2–A.4, tensors and alternating tensors together with operations such as the tensor and wedge product are introduced. Lastly, in Sect. A.5, the concepts which are necessary to introduce the wedge product are summarized in eight steps. A.1 The Dual Space Let V be a real vector space of finite dimension dim V = n.Let(e1,...,en) be a basis of V . Then every v ∈ V can be uniquely represented as a linear combination i v = v ei , (A.1) where summation convention over repeated indices is applied. The coefficients vi ∈ R arereferredtoascomponents of the vector v. Throughout the whole chapter, only finite dimensional real vector spaces, typically denoted by V , are treated. When not stated differently, summation convention is applied. Definition A.1 (Dual Space)Thedual space of V is the set of real-valued linear functionals ∗ V := {ω : V → R : ω linear} . (A.2) The elements of the dual space V ∗ are called linear forms on V . © Springer International Publishing Switzerland 2015 123 S.R. Eugster, Geometric Continuum Mechanics and Induced Beam Theories, Lecture Notes in Applied and Computational Mechanics 75, DOI 10.1007/978-3-319-16495-3 124 Appendix A: Multilinear Algebra The dual space, equipped with pointwise addition and scalar multiplication is again a real vector space. 1 n ∗ Proposition A.2 Given any basis (e1,...,en) for V , let ε ,...,ε ∈ V be the linear forms defined by εi ( ) := δi , e j j (A.3) δi where j is the Kronecker delta symbol defined by 1ifi = j δi := (A.4) j 0ifi = j . 1 n ∗ Then (ε ,...,ε ) is a basis for V , called the dual basis to (e1,...,en) and conse- quently dim V ∗ = dim V . Any linear form ω can be uniquely represented as a linear combination i ω = ωi ε , (A.5) with components ωi = ω(ei ). Proof We need to show that (ε1,...,εn)(i) spans the dual space V ∗ and (ii) is linearly independent. Let v ∈ V and ω ∈ V ∗. (i) Due to the linearity of ω we have (A.1) i (A.2) i ω(v) = ω(v ei ) = v ω(ei ). (A.6) Writing the linear form as a linear combination (A.5) and applying it to v, it follows by linearity of the linear form and the definition of the dual basis (A.3) that ω( ) = ω εi (v j ) (A=.2) ω v j εi ( ) (A=.3) ω v j δi (A=.4) ω vi . v i e j i e j i j i (A.7) 1 n Let ωi = ω(ei ). Then comparison of (A.6) and (A.7) proves that (ε ,...,ε ) spans V ∗. (ii) To show that (ε1,...,εn) is linearly independent, suppose some linear com- i bination equals zero, i.e. ω = ωi ε = 0. Applying to both sides an arbitrary vector v ∈ V , it follows by the computation of (A.7) that i i ∀v ∈ V : ω(v) = 0 ⇒∀v ∈ R: v ωi = 0 ⇒ ωi = 0 . Thus, the only linear combination of elements of (ε1,...,εn) that sums to zero is the trivial one. This proves the linear independency of (ε1,...,εn). The application of a linear form ω ∈ V ∗ on a vector v ∈ V is called the duality pairing and is expressed in components as in (A.7). Appendix A: Multilinear Algebra 125 Since the dual space V ∗ is also a vector space, we may consider the dual space of V ∗, called the bidual space V ∗∗ := (V ∗)∗. For each vector space V , there exists a linear isomorphism ∗∗ ∗ φ : V → V , v → φv :∀ω ∈ V ,φv(ω) = ω(v). Hence, we identify the vectors of the bidual space V ∗∗ naturally with the vectors of the vector space V . For convenience, we suppress the function φ in our notation and write φv(ω) =: v(ω). ∗∗ The duality pairing between the identified bidual basis ei ∈ V and the dual basis ε j ∈ V ∗ is evaluated as (ε j ) = δ j . ei i A.2 Multilinear Forms and Tensors Definition A.3 (Multilinear Form, Tensor) Suppose V1,...,Vk are vector spaces. AmapF: V1 ×···×Vk → R is said to be multilinear, if it is linear in each argument, i.e. for any i ∈ {1,...,k}, v j ∈ Vj , wi ∈ Vi , a, b ∈ R F(v1,...,avi +bwi ,...,vk) = aF(v1,...,vi ,...,vk)+bF(v1,...,wi ,...,vk). We refer to multilinear F as multilinear form or tensor of rank k.1 The set of such multilinear forms is denoted by L(V1,...,Vk; R). Example A.4 Let V and U be vector spaces, then the multilinear form F: V ∗ × V × U → R,(ω, v, u) → F(ω, v, u) is a tensor of rank 3. 1 n Example A.5 Let (e1,...,en) be the basis vectors of V and (ε ...,ε ) the corre- sponding dual basis. Then the basis vectors ε j : V → R , v → ε j (v) = v j , ∗ ei : V → R , ω → ei (ω) = ωi , 1 There exists also an abstract definition of tensor spaces as quotient spaces of free vector spaces, cf. [1], Chap.12. Since we are assuming the vector spaces to be finite dimensional, there exists a canonical isomorphism between the abstract tensor space and the space of multilinear forms. Accordingly, the two terms multilinear form and tensor are used synonymously. 126 Appendix A: Multilinear Algebra are tensors of rank 1, projecting the vector v and the covector ω to their j-th and i-th component, respectively. Definition A.6 (Tensor Product)LetV1,...,Vk, W1,...,Wl be vector spaces, and let F ∈ L(V1,...,Vk; R) and G ∈ L(W1,...,Wl ; R). Define a function F ⊗ G: V1 ×···×Vk × W1 ×···×Wl → R by (F ⊗ G)(v1,...,vk, w1, ··· , wk) := F(v1,...,vk)G(w1, ··· , wk). (A.8) From the multilinearity of F and G it follows that (F ⊗ G) is multi linear too. So, F ⊗ G ∈ L(V1,...,Vk, W1,...,Wl ; R) and is called the tensor product of F and G. Example A.7 Let U and V be vector spaces, F: V × U → R and g: V ∗ → R.The tensor product of F and g is given by the multilinear form F ⊗ g: V × U × V ∗ → R , defined by its application on the vectors (v, u, ω) as (v, u, ω) → (F ⊗ g)(v, u, ω) = F(v, u)g(ω). Applying the definition of the tensor product (A.8) several times, it follows straight forward, that the tensor product is bilinear and associative. Insofar it is allowed to write the tensor products between the tensors F, G and H without any brackets as F ⊗ (G ⊗ H) = (F ⊗ G) ⊗ H =: F ⊗ G ⊗ H . With the tensor product, it is possible to find the building blocks of tensors of arbi- trary rank. The following proposition is formulated and proved for L(U, V, V ∗; R), but extends on a basis of the space of multilinear forms. Proposition A.8 Let U, V be vector spaces of dimensions k and l with bases 1 k 1 l (b1,...,bk) and (e1,...,el ), respectively. Let (β ,...,β ) and (ε ,...,ε ) be the corresponding dual bases of U ∗ and V ∗, respectively. Then the set α i B = β ⊗ ε ⊗ e j : 1 ≤ α ≤ k, 1 ≤ i, j ≤ l is a basis for F ∈ L(U, V, V ∗; R), which therefore has dimension kl2. Summing repeated greek indices from 1 to k and repeated roman indices from 1 to l, any multilinear form F can be written as a linear combination Appendix A: Multilinear Algebra 127 = j βα ⊗ εi ⊗ , F Fαi e j (A.9) j = ( , ,εj ) with Fαi F bα ei . Proof The proof is very similar to the proof of Proposition A.2. We need to show that B is linearly independent and spans L(U, V, V ∗; R).Letu ∈ U, v ∈ V and ω ∈ V ∗. Due to the multilinearity of F it follows directly that α i j j α i F(u, v, ω) = F(u bα,v ei ,ωj ε ) = F(bα, ei ,ε )u v ω j . (A.10) Writing the multilinear form as a linear combination (A.9) and applying it to the same vectors u, v and ω, it follows by the definition of the tensor product that ( , , ω) (A=.9) ( j βα ⊗ εi ⊗ )( , , ω) F u v Fαi e j u v (A=.8) j βα( β )εi (vm ) (ω εn) Fαi u bβ em e j n (A.11) (A=.3) j β vmω δα δi δn (A=.4) j αvi ω . Fαi u n β m j Fαi u j j = ( , ,εj ) B Let Fαi F bα ei . Then comparison of (A.10) and (A.11) proves that spans L(U, V, V ∗). To show that B is linearly independent, suppose some linear combination equals zero: = j βα ⊗ εi ⊗ = F Fαi e j 0 Applying the multilinear form F to arbitrary vectors u ∈ U, v ∈ V and ω ∈ V ∗,it follows by the computation of (A.11) that ∗ ∀u ∈ U, v ∈ V, ω ∈ V : F(u, v, ω) = 0 ⇒∀ α,vi ,ω ∈ R: j αvi ω = u j Fαi u j 0 ⇒ j = .
Recommended publications
  • Tensors and Differential Forms on Vector Spaces
    APPENDIX A TENSORS AND DIFFERENTIAL FORMS ON VECTOR SPACES Since only so much of the vast and growing field of differential forms and differentiable manifolds will be actually used in this survey, we shall attempt to briefly review how the calculus of exterior differential forms on vector spaces can serve as a replacement for the more conventional vector calculus and then introduce only the most elementary notions regarding more topologically general differentiable manifolds, which will mostly be used as the basis for the discussion of Lie groups, in the following appendix. Since exterior differential forms are special kinds of tensor fields – namely, completely-antisymmetric covariant ones – and tensors are important to physics, in their own right, we shall first review the basic notions concerning tensors and multilinear algebra. Presumably, the reader is familiar with linear algebra as it is usually taught to physicists, but for the “basis-free” approach to linear and multilinear algebra (which we shall not always adhere to fanatically), it would also help to have some familiarity with the more “abstract-algebraic” approach to linear algebra, such as one might learn from Hoffman and Kunze [ 1], for instance. 1. Tensor algebra. – A tensor algebra is a type of algebra in which multiplication takes the form of the tensor product. a. Tensor product. – Although the tensor product of vector spaces can be given a rigorous definition in a more abstract-algebraic context (See Greub [ 2], for instance), for the purposes of actual calculations with tensors and tensor fields, it is usually sufficient to say that if V and W are vector spaces of dimensions n and m, respectively, then the tensor product V ⊗ W will be a vector space of dimension nm whose elements are finite linear combinations of elements of the form v ⊗ w, where v is a vector in V and w is a vector in W.
    [Show full text]
  • Tensors Notation • Contravariant Denoted by Superscript Ai Took a Vector and Gave for Vector Calculus Us a Vector
    Tensors Notation • contravariant denoted by superscript Ai took a vector and gave For vector calculus us a vector • covariant denoted by subscript Ai took a scaler and gave us a vector Review To avoid confusion in cartesian coordinates both types are the same so • Vectors we just opt for the subscript. Thus a vector x would be x1,x2,x3 R3 • Summation representation of an n by n array As it turns out in an cartesian space and other rectilinear coordi- nate systems there is no difference between contravariant and covariant • Gradient, Divergence and Curl vectors. This will not be the case for other coordinate systems such a • Spherical Harmonics (maybe) curvilinear coordinate systems or in 4 dimensions. These definitions are closely related to the Jacobian. Motivation If you tape a book shut and try to spin it in the air on each indepen- Definitions for Tensors of Rank 2 dent axis you will notice that it spins fine on two axes but not on the Rank 2 tensors can be written as a square array. They have con- third. That’s the inertia tensor in your hands. Similar are the polar- travariant, mixed, and covariant forms. As we might expect in cartesian izations tensor, index of refraction tensor and stress tensor. But tensors coordinates these are the same. also show up in all sorts of places that don’t connect to an anisotropic material property, in fact even spherical harmonics are tensors. What are the similarities and differences between such a plethora of tensors? Vector Calculus and Identifers The mathematics of tensors is particularly useful for de- Tensor analysis extends deep into coordinate transformations of all scribing properties of substances which vary in direction– kinds of spaces and coordinate systems.
    [Show full text]
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Geometrical Tools for Embedding Fields, Submanifolds, and Foliations Arxiv
    Geometrical tools for embedding fields, submanifolds, and foliations Antony J. Speranza∗ Perimeter Institute for Theoretical Physics, 31 Caroline St. N, Waterloo, ON N2L 2Y5, Canada Abstract Embedding fields provide a way of coupling a background structure to a theory while preserving diffeomorphism-invariance. Examples of such background structures include embedded submani- folds, such as branes; boundaries of local subregions, such as the Ryu-Takayanagi surface in holog- raphy; and foliations, which appear in fluid dynamics and force-free electrodynamics. This work presents a systematic framework for computing geometric properties of these background structures in the embedding field description. An overview of the local geometric quantities associated with a foliation is given, including a review of the Gauss, Codazzi, and Ricci-Voss equations, which relate the geometry of the foliation to the ambient curvature. Generalizations of these equations for cur- vature in the nonintegrable normal directions are derived. Particular care is given to the question of which objects are well-defined for single submanifolds, and which depend on the structure of the foliation away from a submanifold. Variational formulas are provided for the geometric quantities, which involve contributions both from the variation of the embedding map as well as variations of the ambient metric. As an application of these variational formulas, a derivation is given of the Jacobi equation, describing perturbations of extremal area surfaces of arbitrary codimension. The embedding field formalism is also applied to the problem of classifying boundary conditions for general relativity in a finite subregion that lead to integrable Hamiltonians. The framework developed in this paper will provide a useful set of tools for future analyses of brane dynamics, fluid arXiv:1904.08012v2 [gr-qc] 26 Apr 2019 mechanics, and edge modes for finite subregions of diffeomorphism-invariant theories.
    [Show full text]
  • Abstract Tensor Systems and Diagrammatic Representations
    Abstract tensor systems and diagrammatic representations J¯anisLazovskis September 28, 2012 Abstract The diagrammatic tensor calculus used by Roger Penrose (most notably in [7]) is introduced without a solid mathematical grounding. We will attempt to derive the tools of such a system, but in a broader setting. We show that Penrose's work comes from the diagrammisation of the symmetric algebra. Lie algebra representations and their extensions to knot theory are also discussed. Contents 1 Abstract tensors and derived structures 2 1.1 Abstract tensor notation . 2 1.2 Some basic operations . 3 1.3 Tensor diagrams . 3 2 A diagrammised abstract tensor system 6 2.1 Generation . 6 2.2 Tensor concepts . 9 3 Representations of algebras 11 3.1 The symmetric algebra . 12 3.2 Lie algebras . 13 3.3 The tensor algebra T(g)....................................... 16 3.4 The symmetric Lie algebra S(g)................................... 17 3.5 The universal enveloping algebra U(g) ............................... 18 3.6 The metrized Lie algebra . 20 3.6.1 Diagrammisation with a bilinear form . 20 3.6.2 Diagrammisation with a symmetric bilinear form . 24 3.6.3 Diagrammisation with a symmetric bilinear form and an orthonormal basis . 24 3.6.4 Diagrammisation under ad-invariance . 29 3.7 The universal enveloping algebra U(g) for a metrized Lie algebra g . 30 4 Ensuing connections 32 A Appendix 35 Note: This work relies heavily upon the text of Chapter 12 of a draft of \An Introduction to Quantum and Vassiliev Invariants of Knots," by David M.R. Jackson and Iain Moffatt, a yet-unpublished book at the time of writing.
    [Show full text]
  • On the Second Dual of the Space of Continuous Functions
    ON THE SECOND DUAL OF THE SPACE OF CONTINUOUS FUNCTIONS BY SAMUEL KAPLAN Introduction. By "the space of continuous functions" we mean the Banach space C of real continuous functions on a compact Hausdorff space X. C, its first dual, and its second dual are not only Banach spaces, but also vector lattices, and this aspect of them plays a central role in our treatment. In §§1-3, we collect the properties of vector lattices which we need in the paper. In §4 we add some properties of the first dual of C—the space of Radon measures on X—denoted by L in the present paper. In §5 we imbed C in its second dual, denoted by M, and then identify the space of all bounded real functions on X with a quotient space of M, in fact with a topological direct summand. Thus each bounded function on X represents an entire class of elements of M. The value of an element/ of M on an element p. of L is denoted as usual by/(p.)- If/lies in C then of course f(u) =p(f) (usually written J fdp), and thus the multiplication between Afand P is an extension of the multiplication between L and C. Now in integration theory, for each pEL, there is a stand- ard "extension" of ffdu to certain of the bounded functions on X (we confine ourselves to bounded functions in this paper), which are consequently called u-integrable. The question arises: how is this "extension" related to the multi- plication between M and L.
    [Show full text]
  • Bornologically Isomorphic Representations of Tensor Distributions
    Bornologically isomorphic representations of distributions on manifolds E. Nigsch Thursday 15th November, 2018 Abstract Distributional tensor fields can be regarded as multilinear mappings with distributional values or as (classical) tensor fields with distribu- tional coefficients. We show that the corresponding isomorphisms hold also in the bornological setting. 1 Introduction ′ ′ ′r s ′ Let D (M) := Γc(M, Vol(M)) and Ds (M) := Γc(M, Tr(M) ⊗ Vol(M)) be the strong duals of the space of compactly supported sections of the volume s bundle Vol(M) and of its tensor product with the tensor bundle Tr(M) over a manifold; these are the spaces of scalar and tensor distributions on M as defined in [?, ?]. A property of the space of tensor distributions which is fundamental in distributional geometry is given by the C∞(M)-module isomorphisms ′r ∼ s ′ ∼ r ′ Ds (M) = LC∞(M)(Tr (M), D (M)) = Ts (M) ⊗C∞(M) D (M) (1) (cf. [?, Theorem 3.1.12 and Corollary 3.1.15]) where C∞(M) is the space of smooth functions on M. In[?] a space of Colombeau-type nonlinear generalized tensor fields was constructed. This involved handling smooth functions (in the sense of convenient calculus as developed in [?]) in par- arXiv:1105.1642v1 [math.FA] 9 May 2011 ∞ r ′ ticular on the C (M)-module tensor products Ts (M) ⊗C∞(M) D (M) and Γ(E) ⊗C∞(M) Γ(F ), where Γ(E) denotes the space of smooth sections of a vector bundle E over M. In[?], however, only minor attention was paid to questions of topology on these tensor products.
    [Show full text]
  • Multilinear Algebra and Applications July 15, 2014
    Multilinear Algebra and Applications July 15, 2014. Contents Chapter 1. Introduction 1 Chapter 2. Review of Linear Algebra 5 2.1. Vector Spaces and Subspaces 5 2.2. Bases 7 2.3. The Einstein convention 10 2.3.1. Change of bases, revisited 12 2.3.2. The Kronecker delta symbol 13 2.4. Linear Transformations 14 2.4.1. Similar matrices 18 2.5. Eigenbases 19 Chapter 3. Multilinear Forms 23 3.1. Linear Forms 23 3.1.1. Definition, Examples, Dual and Dual Basis 23 3.1.2. Transformation of Linear Forms under a Change of Basis 26 3.2. Bilinear Forms 30 3.2.1. Definition, Examples and Basis 30 3.2.2. Tensor product of two linear forms on V 32 3.2.3. Transformation of Bilinear Forms under a Change of Basis 33 3.3. Multilinear forms 34 3.4. Examples 35 3.4.1. A Bilinear Form 35 3.4.2. A Trilinear Form 36 3.5. Basic Operation on Multilinear Forms 37 Chapter 4. Inner Products 39 4.1. Definitions and First Properties 39 4.1.1. Correspondence Between Inner Products and Symmetric Positive Definite Matrices 40 4.1.1.1. From Inner Products to Symmetric Positive Definite Matrices 42 4.1.1.2. From Symmetric Positive Definite Matrices to Inner Products 42 4.1.2. Orthonormal Basis 42 4.2. Reciprocal Basis 46 4.2.1. Properties of Reciprocal Bases 48 4.2.2. Change of basis from a basis to its reciprocal basis g 50 B B III IV CONTENTS 4.2.3.
    [Show full text]
  • Sisältö Part I: Topological Vector Spaces 4 1. General Topological
    Sisalt¨ o¨ Part I: Topological vector spaces 4 1. General topological vector spaces 4 1.1. Vector space topologies 4 1.2. Neighbourhoods and filters 4 1.3. Finite dimensional topological vector spaces 9 2. Locally convex spaces 10 2.1. Seminorms and semiballs 10 2.2. Continuous linear mappings in locally convex spaces 13 3. Generalizations of theorems known from normed spaces 15 3.1. Separation theorems 15 Hahn and Banach theorem 17 Important consequences 18 Hahn-Banach separation theorem 19 4. Metrizability and completeness (Fr`echet spaces) 19 4.1. Baire's category theorem 19 4.2. Complete topological vector spaces 21 4.3. Metrizable locally convex spaces 22 4.4. Fr´echet spaces 23 4.5. Corollaries of Baire 24 5. Constructions of spaces from each other 25 5.1. Introduction 25 5.2. Subspaces 25 5.3. Factor spaces 26 5.4. Product spaces 27 5.5. Direct sums 27 5.6. The completion 27 5.7. Projektive limits 28 5.8. Inductive locally convex limits 28 5.9. Direct inductive limits 29 6. Bounded sets 29 6.1. Bounded sets 29 6.2. Bounded linear mappings 31 7. Duals, dual pairs and dual topologies 31 7.1. Duals 31 7.2. Dual pairs 32 7.3. Weak topologies in dualities 33 7.4. Polars 36 7.5. Compatible topologies 39 7.6. Polar topologies 40 PART II Distributions 42 1. The idea of distributions 42 1.1. Schwartz test functions and distributions 42 2. Schwartz's test function space 43 1 2.1. The spaces C (Ω) and DK 44 1 2 2.2.
    [Show full text]
  • Causalx: Causal Explanations and Block Multilinear Factor Analysis
    To appear: Proc. of the 2020 25th International Conference on Pattern Recognition (ICPR 2020) Milan, Italy, Jan. 10-15, 2021. CausalX: Causal eXplanations and Block Multilinear Factor Analysis M. Alex O. Vasilescu1;2 Eric Kim2;1 Xiao S. Zeng2 [email protected] [email protected] [email protected] 1Tensor Vision Technologies, Los Angeles, California 2Department of Computer Science,University of California, Los Angeles Abstract—By adhering to the dictum, “No causation without I. INTRODUCTION:PROBLEM DEFINITION manipulation (treatment, intervention)”, cause and effect data Developing causal explanations for correct results or for failures analysis represents changes in observed data in terms of changes from mathematical equations and data is important in developing in the causal factors. When causal factors are not amenable for a trustworthy artificial intelligence, and retaining public trust. active manipulation in the real world due to current technological limitations or ethical considerations, a counterfactual approach Causal explanations are germane to the “right to an explanation” performs an intervention on the model of data formation. In the statute [15], [13] i.e., to data driven decisions, such as those case of object representation or activity (temporal object) rep- that rely on images. Computer graphics and computer vision resentation, varying object parts is generally unfeasible whether problems, also known as forward and inverse imaging problems, they be spatial and/or temporal. Multilinear algebra, the algebra have been cast as causal inference questions [40], [42] consistent of higher order tensors, is a suitable and transparent framework for disentangling the causal factors of data formation. Learning a with Donald Rubin’s quantitative definition of causality, where part-based intrinsic causal factor representations in a multilinear “A causes B” means “the effect of A is B”, a measurable framework requires applying a set of interventions on a part- and experimentally repeatable quantity [14], [17].
    [Show full text]
  • Solution: the First Element of the Dual Basis Is the Linear Function Α
    Homework assignment 7 pp. 105 3 Exercise 2. Let B = f®1; ®2; ®3g be the basis for C defined by ®1 = (1; 0; ¡1) ®2 = (1; 1; 1) ®3 = (2; 2; 0): Find the dual basis of B. Solution: ¤ The first element of the dual basis is the linear function ®1 such ¤ ¤ ¤ that ®1(®1) = 1; ®1(®2) = 0 and ®1(®3) = 0. To describe such a function more explicitly we need to find its values on the standard basis vectors e1, e2 and e3. To do this express e1; e2; e3 through ®1; ®2; ®3 (refer to the solution of Exercise 1 pp. 54-55 from Homework 6). For each i = 1; 2; 3 you will find the numbers ai; bi; ci such that e1 = ai®1 + bi®2 + ci®3 (i.e. the coordinates of ei relative to the ¤ ¤ ¤ basis ®1; ®2; ®3). Then by linearity of ®1 we get that ®1(ei) = ai. Then ®2(ei) = bi, ¤ and ®3(ei) = ci. This is the answer. It can also be reformulated as follows. If P is the transition matrix from the standard basis e1; e2; e3 to ®1; ®2; ®3, i.e. ¡1 t (®1; ®2; ®3) = (e1; e2; e3)P , then (P ) is the transition matrix from the dual basis ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¡1 t e1; e2; e3 to the dual basis ®1; ®2; a3, i.e. (®1; ®2; a3) = (e1; e2; e3)(P ) . Note that this problem is basically the change of coordinates problem: e.g. ¤ 3 the value of ®1 on the vector v 2 C is the first coordinate of v relative to the basis ®1; ®2; ®3.
    [Show full text]
  • Vectors and Dual Vectors
    Vectors By now you have a pretty good experience with \vectors". Usually, a vector is defined as a quantity that has a direction and a magnitude, such as a position vector, velocity vector, acceleration vector, etc. However, the notion of a vector has a considerably wider realm of applicability than these examples might suggest. The set of all real numbers forms a vector space, as does the set of all complex numbers. The set of functions on a set (e.g., functions of one variable, f(x)) form a vector space. Solutions of linear homogeneous equations form a vector space. We begin by giving the abstract rules for forming a space of vectors, also known as a vector space. A vector space V is a set equipped with an operation of \addition" and an additive identity. The elements of the set are called vectors, which we shall denote as ~u, ~v, ~w, etc. For now, you can think of them as position vectors in order to keep yourself sane. Addition, is an operation in which two vectors, say ~u and ~v, can be combined to make another vector, say, ~w. We denote this operation by the symbol \+": ~u + ~v = ~w: (1) Do not be fooled by this simple notation. The \addition" of vectors may be quite a different operation than ordinary arithmetic addition. For example, if we view position vectors in the x-y plane as \arrows" drawn from the origin, the addition of vectors is defined by the parallelogram rule. Clearly this rule is quite different than ordinary \addition".
    [Show full text]