Tensors on a Vector Space

Total Page:16

File Type:pdf, Size:1020Kb

Tensors on a Vector Space APPENDIX B Tensors on a vector space In this Appendix, we gather mathematical definitions and results pertaining to tensors. The purpose is mostly to introduce the “modern”, geometrical view on tensors, which defines them by their action on vectors or one-forms, i.e. in a coordinate-independent way (Sec. B.1), in contrast to the “old” definition based on their behavior under basis transformations (Sec. B.2). The reader is assumed to already possess enough knowledge on linear algebra to know what are vectors, linear (in)dependence, (multi)linearity, matrices. Similarly, the notions of group, field, application/function/mapping. are used without further mention. In the remainder of these lecture notes, we actually consider tensors on real vector spaces, i.e. for which the underlying base field K of scalars is the set R of real numbers; here we remain more general. Einstein’s summation convention is used throughout. B.1 Vectors, one-forms and tensors B.1.1 Vectors . are by definition the elements ~c of a vector space V , i.e. of a set with 1) a binary operation (“addition”) with which it is an Abelian group, and 2) a multiplication with “scalars”—elements of a base field K—which is associative, has an identity element, and is distributive with respect to both additions on V and on K. Introducing a basis B = f~eig, i.e. a family of linearly independent vectors that span the whole space V , one associates to each vector ~c its uniquely defined components fcig, elements of the base field K, such that i ~c = c ~ei: (B.1) If the number of vectors of a basis is finite—in which case this holds for all bases— and equal to some integer D—which is the same for all bases—, the space V is said to be finite-dimensional and D is its dimension (over K): D = dim V . We shall assume that this is the case in the remainder of this Section. B.1.2 One-forms . on a vector space V are the linear applications, hereafter denoted as h, from V into the base field of scalars K. e The set of 1-forms on V , equipped with the “natural” addition and scalar multiplication, is itself a vector space over the field K, denoted by V ∗ and said to be dual to V . ∗ ∗ If V is finite-dimensional, so is V , with dim V = dim V . Given a basis B = f~eig in V , one can then construct its dual basis B∗ = fjg in V ∗ such that e j j (~ei) = δi ; (B.2) j e where δi denotes the usual Kronecker delta symbol. The components of a 1-form h on a given basis will be denoted as fhjg: j e h = hj : (B.3) e e B.1 Vectors, one-forms and tensors 53 Remarks: ∗ The choice of notations, in particular the position of indices, is not innocent! Thus, if fjg denotes the dual base to f~eig, the reader can trivially check that e i i c = (~c) and hj = h(~ej): (B.4) e e ∗ In the “old” language, the vectors of V resp. the 1-forms of V ∗ were designated as “contravariant vectors” resp. “covariant vectors” or “covectors”, and their coordinates as “contravariant” resp. ”co- variant” coordinates. The latter two, applying to the components, remain useful short denominations, especially when applied to tensors (see below). Yet in truth they are not different components of a same mathemat- ical quantity, but components of different objects between which a “natural” correspondence was introduced, in particular by using a metric tensor as in § B.1.4. B.1.3 Tensors :::::::B.1.3 a :::::::::::::::::::::::::::Definition and first results Let V be a vector space with base field K, and m, n denote two nonnegative integers. The multilinear applications of m one-forms—elements of V ∗—and n vectors—elements of V —into m K are referred to as the tensors of type n on V , where linearity should hold with respect to every argument. The integer m + n is the order (or often, but improperly, rank) of the tensor. Already known objects arise as special cases of this definition when either m or n is zero: 0 K • the 0 -tensors are simply the scalars of the base field ; 1 (19) • the 0 -tensors coincide with vectors; 0 0 • the 1 -tensors are the one-forms. More generally, the n -tensors are also known as (multi- linear) n-forms. 2 • Eventually, 0 -tensors are sometimes called “bivectors” or “dyadics”. Tensors will generically be denoted as T, irrespective of their rank, unless the latter is 0 or 1. A tensor may be symmetric or antisymmetric under the exchange of two of its arguments, either both vectors or both 1-forms. Generalizing, it may be totally symmetric—as e.g. the metric tensor we shall encounter below—, or antisymmetric. An instance of the latter case is the determinant, which is the only (up to a multiplicative factor) totally antisymmetric D-form on a vector space of dimension D. m ∗ m ∗ n 0 0 Remark: Consider a n -tensor T :(V ) ×(V ) ! K, and let m ≤ m, n ≤ n be two nonnegative 0 0 integers. For every m -uplet of one-forms fhig and n -uplet of vectors f~cjg—and corresponding multiplets of argument positions, although heree we take for simplicity the first ones—the object T h1; : : : ; hm0 ; · ;:::; · ;~c1; : : : ;~cn0 ; · ;:::; · ; e e where the dots denote “empty” arguments, can be applied to m − m0 one-forms and n − n0 vectors 0 0 to yield a scalar. That is, the tensor T induces a multilinear application(20) from (V ∗)m × (V ∗)n m−m0 into the set of n−n0 -tensors. 1 For example, the 1 -tensors are in natural correspondence with the linear applications from V into V , i.e. in turn with the square matrices of order dim V . (19)More accurately, they are the elements of the double dual of V , which is always homomorphic to V . (20)Rather, the number of such applications is the number of independent—under consideration of possible symmetries—combinations of m0 resp. n0 one-form resp. vector arguments. 54 Tensors on a vector space :::::::B.1.3 b :::::::::::::::::::::::Operations on tensors The tensors of a given type, with the addition and scalar multiplication inherited from V ; form a vector space on K. Besides these natural addition and multiplication, one defines two further operations on tensors, the outer product or tensor product—which increases the rank—and the contraction, which decreases the rank. 0 m m0 0 Consider two tensors T and T , of respective types n and n0 . Their outer product T ⊗T is m+m0 0 0 a tensor of type n+n0 satisfying for every (m + m )-uplet (h1; : : : ; hm; : : : ; hm+m ) of 1-forms and 0 every (n + n )-uplet (~c1; : : : ;~cn; : : : ;~cn+n0 ) of vectors the identitye e e 0 T ⊗ T h1; : : : ; hm+m0 ;~c1; : : : ;~cn+n0 = e e 0 T h1; : : : ; hm;~c1; : : : ;~cn T hm+1; : : : ; hm+m0 ;~cn+1; : : : ;~cn+n0 : e e e e For instance, the outer product of two 1-forms h, h0 is a 2-form h ⊗ h0 such that for every pair of vectors (~c;~c 0), h ⊗ h0(~c;~c 0) = h(~c) h0(~c 0). In turn,e e the outer producte e of two vectors ~c, ~c 0 is a 2 0 0 0 0 0 0 0 -tensor ~c ⊗ ~c suche thate for everye paire of 1-forms (h; h ), ~c ⊗ ~c (h; h ) = h(~c) h (~c ). m Tensors of type n that can be written as outere productse ofeme vectorse ande n one-forms are sometimes called simple tensors. m Let T be a n -tensor, where both m and n are non-zero. To define the contraction over its j-th one-form and k-th vector arguments, the easiest—apart from introducing the tensor components—is to write T as a sum of simple tensors. By applying in each of the summand the k-th one-form to m−1 the j-th vector, which gives a number, one obtains a sum of simple tensors of type n−1 , which is the result of the contraction operation. Examples of contractions will be given after the metric tensor has been introduced. :::::::B.1.3 c ::::::::::::::::::::Tensor coordinates j ∗ Let f~eig resp. f g denote bases on a vector space V of dimension D resp. on its dual V —in principle, they neede not be dual to each other, although using dual bases is what is implicitly always done in practice—and m, n be two nonnegative integers. m+n j1 jn The D simple tensors f~ei1 ⊗ · · · ⊗~eim ⊗ ⊗ · · · ⊗ g, where each ik or jk runs from 1 to D, m form a basis of the tensors of type n . The componentse e of a tensor T on this basis will be denoted i1:::im as fT j1:::jn g: i1:::im j1 jn T = T j1:::jn ~ei1 ⊗ · · · ⊗~eim ⊗ ⊗ · · · ⊗ ; (B.5a) e e where i1:::im i1 im T j1:::jn = T( ; : : : ; ;~ej1 ; : : : ;~ejn ): (B.5b) e e The possible symmetry or antisymmetry of a tensor with respect to the exchange of two of its arguments translates into the corresponding symmetry or antisymmetry of the components when exchanging the respective indices. In turn, the contraction of T over its j-th one-form and k-th vector arguments yields the tensor :::ij−1;`;ij+1;::: with components T :::jk−1;`;jk+1;:::, with summation over the repeated index `.
Recommended publications
  • Topology and Physics 2019 - Lecture 2
    Topology and Physics 2019 - lecture 2 Marcel Vonk February 12, 2019 2.1 Maxwell theory in differential form notation Maxwell's theory of electrodynamics is a great example of the usefulness of differential forms. A nice reference on this topic, though somewhat outdated when it comes to notation, is [1]. For notational simplicity, we will work in units where the speed of light, the vacuum permittivity and the vacuum permeability are all equal to 1: c = 0 = µ0 = 1. 2.1.1 The dual field strength In three dimensional space, Maxwell's electrodynamics describes the physics of the electric and magnetic fields E~ and B~ . These are three-dimensional vector fields, but the beauty of the theory becomes much more obvious if we (a) use a four-dimensional relativistic formulation, and (b) write it in terms of differential forms. For example, let us look at Maxwells two source-free, homogeneous equations: r · B = 0;@tB + r × E = 0: (2.1) That these equations have a relativistic flavor becomes clear if we write them out in com- ponents and organize the terms somewhat suggestively: x y z 0 + @xB + @yB + @zB = 0 x z y −@tB + 0 − @yE + @zE = 0 (2.2) y z x −@tB + @xE + 0 − @zE = 0 z y x −@tB − @xE + @yE + 0 = 0 Note that we also multiplied the last three equations by −1 to clarify the structure. All in all, we see that we have four equations (one for each space-time coordinate) which each contain terms in which the four coordinate derivatives act. Therefore, we may be tempted to write our set of equations in more \relativistic" notation as ^µν @µF = 0 (2.3) 1 with F^µν the coordinates of an antisymmetric two-tensor (i.
    [Show full text]
  • S-96.510 Advanced Field Theory Course for Graduate Students Lecture Viewgraphs, Fall Term 2004
    S-96.510 Advanced Field Theory Course for graduate students Lecture viewgraphs, fall term 2004 I.V.Lindell Helsinki University of Technology Electromagnetics Laboratory Espoo, Finland I.V.Lindell: Advanced Field Theory, 2004 Helsinki University of Technology 00.01 1 Contents [01] Complex Vectors and Dyadics [02] Dyadic Algebra [03] Basic Electromagnetic Equations [04] Conditions for Fields and Media [05] Duality Transformation [06] Affine Transformation [07] Electromagnetic Field Solutions [08] Singularities and Complex Sources [09] Plane Waves [10] Source Equivalence [11] Huygens’ Principle [12] Field Decompositions Vector Formulas, Dyadic Identites as an appendix I.V.Lindell: Advanced Field Theory, 2004 Helsinki University of Technology 00.02 2 Foreword This lecture material contains all viewgraphs associated with the gradu- ate course S-96.510 Advanced Field Theory given at the Department of Electrical and Communications Engineering, fall 2004. The course is based on Chapters 1–6 of the book Methods for Electromagnetic Field Analysis (Oxford University Press 1992, 2nd edition IEEE Press, 1995, 3rd printing Wiley 2002) by this author. The figures drawn by hand on the blackboard could not be added to the present material. Otaniemi, September 13 2004 I.V. Lindell I.V.Lindell: Advanced Field Theory, 2004 Helsinki University of Technology 00.03 3 I.V.Lindell: Advanced Field Theory, 2004 Helsinki University of Technology 00.04 4 S-96.510 Advanced Field Theory 1. Complex Vectors and Dyadics I.V.Lindell I.V.Lindell: Advanced Field Theory, 2004
    [Show full text]
  • Tensor Manipulation in GPL Maxima
    Tensor Manipulation in GPL Maxima Viktor Toth http://www.vttoth.com/ February 1, 2008 Abstract GPL Maxima is an open-source computer algebra system based on DOE-MACSYMA. GPL Maxima included two tensor manipulation packages from DOE-MACSYMA, but these were in various states of disrepair. One of the two packages, CTENSOR, implemented component-based tensor manipulation; the other, ITENSOR, treated tensor symbols as opaque, manipulating them based on their index properties. The present paper describes the state in which these packages were found, the steps that were needed to make the packages fully functional again, and the new functionality that was implemented to make them more versatile. A third package, ATENSOR, was also implemented; fully compatible with the identically named package in the commercial version of MACSYMA, ATENSOR implements abstract tensor algebras. 1 Introduction GPL Maxima (GPL stands for the GNU Public License, the most widely used open source license construct) is the descendant of one of the world’s first comprehensive computer algebra systems (CAS), DOE-MACSYMA, developed by the United States Department of Energy in the 1960s and the 1970s. It is currently maintained by 18 volunteer developers, and can be obtained in source or object code form from http://maxima.sourceforge.net/. Like other computer algebra systems, Maxima has tensor manipulation capability. This capability was developed in the late 1970s. Documentation is scarce regarding these packages’ origins, but a select collection of e-mail messages by various authors survives, dating back to 1979-1982, when these packages were actively maintained at M.I.T. When this author first came across GPL Maxima, the tensor packages were effectively non-functional.
    [Show full text]
  • Bornologically Isomorphic Representations of Tensor Distributions
    Bornologically isomorphic representations of distributions on manifolds E. Nigsch Thursday 15th November, 2018 Abstract Distributional tensor fields can be regarded as multilinear mappings with distributional values or as (classical) tensor fields with distribu- tional coefficients. We show that the corresponding isomorphisms hold also in the bornological setting. 1 Introduction ′ ′ ′r s ′ Let D (M) := Γc(M, Vol(M)) and Ds (M) := Γc(M, Tr(M) ⊗ Vol(M)) be the strong duals of the space of compactly supported sections of the volume s bundle Vol(M) and of its tensor product with the tensor bundle Tr(M) over a manifold; these are the spaces of scalar and tensor distributions on M as defined in [?, ?]. A property of the space of tensor distributions which is fundamental in distributional geometry is given by the C∞(M)-module isomorphisms ′r ∼ s ′ ∼ r ′ Ds (M) = LC∞(M)(Tr (M), D (M)) = Ts (M) ⊗C∞(M) D (M) (1) (cf. [?, Theorem 3.1.12 and Corollary 3.1.15]) where C∞(M) is the space of smooth functions on M. In[?] a space of Colombeau-type nonlinear generalized tensor fields was constructed. This involved handling smooth functions (in the sense of convenient calculus as developed in [?]) in par- arXiv:1105.1642v1 [math.FA] 9 May 2011 ∞ r ′ ticular on the C (M)-module tensor products Ts (M) ⊗C∞(M) D (M) and Γ(E) ⊗C∞(M) Γ(F ), where Γ(E) denotes the space of smooth sections of a vector bundle E over M. In[?], however, only minor attention was paid to questions of topology on these tensor products.
    [Show full text]
  • On Manifolds of Tensors of Fixed Tt-Rank
    ON MANIFOLDS OF TENSORS OF FIXED TT-RANK SEBASTIAN HOLTZ, THORSTEN ROHWEDDER, AND REINHOLD SCHNEIDER Abstract. Recently, the format of TT tensors [19, 38, 34, 39] has turned out to be a promising new format for the approximation of solutions of high dimensional problems. In this paper, we prove some new results for the TT representation of a tensor U ∈ Rn1×...×nd and for the manifold of tensors of TT-rank r. As a first result, we prove that the TT (or compression) ranks ri of a tensor U are unique and equal to the respective seperation ranks of U if the components of the TT decomposition are required to fulfil a certain maximal rank condition. We then show that d the set T of TT tensors of fixed rank r forms an embedded manifold in Rn , therefore preserving the essential theoretical properties of the Tucker format, but often showing an improved scaling behaviour. Extending a similar approach for matrices [7], we introduce certain gauge conditions to obtain a unique representation of the tangent space TU T of T and deduce a local parametrization of the TT manifold. The parametrisation of TU T is often crucial for an algorithmic treatment of high-dimensional time-dependent PDEs and minimisation problems [33]. We conclude with remarks on those applications and present some numerical examples. 1. Introduction The treatment of high-dimensional problems, typically of problems involving quantities from Rd for larger dimensions d, is still a challenging task for numerical approxima- tion. This is owed to the principal problem that classical approaches for their treatment normally scale exponentially in the dimension d in both needed storage and computa- tional time and thus quickly become computationally infeasable for sensible discretiza- tions of problems of interest.
    [Show full text]
  • Solution: the First Element of the Dual Basis Is the Linear Function Α
    Homework assignment 7 pp. 105 3 Exercise 2. Let B = f®1; ®2; ®3g be the basis for C defined by ®1 = (1; 0; ¡1) ®2 = (1; 1; 1) ®3 = (2; 2; 0): Find the dual basis of B. Solution: ¤ The first element of the dual basis is the linear function ®1 such ¤ ¤ ¤ that ®1(®1) = 1; ®1(®2) = 0 and ®1(®3) = 0. To describe such a function more explicitly we need to find its values on the standard basis vectors e1, e2 and e3. To do this express e1; e2; e3 through ®1; ®2; ®3 (refer to the solution of Exercise 1 pp. 54-55 from Homework 6). For each i = 1; 2; 3 you will find the numbers ai; bi; ci such that e1 = ai®1 + bi®2 + ci®3 (i.e. the coordinates of ei relative to the ¤ ¤ ¤ basis ®1; ®2; ®3). Then by linearity of ®1 we get that ®1(ei) = ai. Then ®2(ei) = bi, ¤ and ®3(ei) = ci. This is the answer. It can also be reformulated as follows. If P is the transition matrix from the standard basis e1; e2; e3 to ®1; ®2; ®3, i.e. ¡1 t (®1; ®2; ®3) = (e1; e2; e3)P , then (P ) is the transition matrix from the dual basis ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¤ ¡1 t e1; e2; e3 to the dual basis ®1; ®2; a3, i.e. (®1; ®2; a3) = (e1; e2; e3)(P ) . Note that this problem is basically the change of coordinates problem: e.g. ¤ 3 the value of ®1 on the vector v 2 C is the first coordinate of v relative to the basis ®1; ®2; ®3.
    [Show full text]
  • Vectors and Dual Vectors
    Vectors By now you have a pretty good experience with \vectors". Usually, a vector is defined as a quantity that has a direction and a magnitude, such as a position vector, velocity vector, acceleration vector, etc. However, the notion of a vector has a considerably wider realm of applicability than these examples might suggest. The set of all real numbers forms a vector space, as does the set of all complex numbers. The set of functions on a set (e.g., functions of one variable, f(x)) form a vector space. Solutions of linear homogeneous equations form a vector space. We begin by giving the abstract rules for forming a space of vectors, also known as a vector space. A vector space V is a set equipped with an operation of \addition" and an additive identity. The elements of the set are called vectors, which we shall denote as ~u, ~v, ~w, etc. For now, you can think of them as position vectors in order to keep yourself sane. Addition, is an operation in which two vectors, say ~u and ~v, can be combined to make another vector, say, ~w. We denote this operation by the symbol \+": ~u + ~v = ~w: (1) Do not be fooled by this simple notation. The \addition" of vectors may be quite a different operation than ordinary arithmetic addition. For example, if we view position vectors in the x-y plane as \arrows" drawn from the origin, the addition of vectors is defined by the parallelogram rule. Clearly this rule is quite different than ordinary \addition".
    [Show full text]
  • Matrices and Tensors
    APPENDIX MATRICES AND TENSORS A.1. INTRODUCTION AND RATIONALE The purpose of this appendix is to present the notation and most of the mathematical tech- niques that are used in the body of the text. The audience is assumed to have been through sev- eral years of college-level mathematics, which included the differential and integral calculus, differential equations, functions of several variables, partial derivatives, and an introduction to linear algebra. Matrices are reviewed briefly, and determinants, vectors, and tensors of order two are described. The application of this linear algebra to material that appears in under- graduate engineering courses on mechanics is illustrated by discussions of concepts like the area and mass moments of inertia, Mohr’s circles, and the vector cross and triple scalar prod- ucts. The notation, as far as possible, will be a matrix notation that is easily entered into exist- ing symbolic computational programs like Maple, Mathematica, Matlab, and Mathcad. The desire to represent the components of three-dimensional fourth-order tensors that appear in anisotropic elasticity as the components of six-dimensional second-order tensors and thus rep- resent these components in matrices of tensor components in six dimensions leads to the non- traditional part of this appendix. This is also one of the nontraditional aspects in the text of the book, but a minor one. This is described in §A.11, along with the rationale for this approach. A.2. DEFINITION OF SQUARE, COLUMN, AND ROW MATRICES An r-by-c matrix, M, is a rectangular array of numbers consisting of r rows and c columns: ¯MM..
    [Show full text]
  • Duality, Part 1: Dual Bases and Dual Maps Notation
    Duality, part 1: Dual Bases and Dual Maps Notation F denotes either R or C. V and W denote vector spaces over F. Define ': R3 ! R by '(x; y; z) = 4x − 5y + 2z. Then ' is a linear functional on R3. n n Fix (b1;:::; bn) 2 C . Define ': C ! C by '(z1;:::; zn) = b1z1 + ··· + bnzn: Then ' is a linear functional on Cn. Define ': P(R) ! R by '(p) = 3p00(5) + 7p(4). Then ' is a linear functional on P(R). R 1 Define ': P(R) ! R by '(p) = 0 p(x) dx. Then ' is a linear functional on P(R). Examples: Linear Functionals Definition: linear functional A linear functional on V is a linear map from V to F. In other words, a linear functional is an element of L(V; F). n n Fix (b1;:::; bn) 2 C . Define ': C ! C by '(z1;:::; zn) = b1z1 + ··· + bnzn: Then ' is a linear functional on Cn. Define ': P(R) ! R by '(p) = 3p00(5) + 7p(4). Then ' is a linear functional on P(R). R 1 Define ': P(R) ! R by '(p) = 0 p(x) dx. Then ' is a linear functional on P(R). Linear Functionals Definition: linear functional A linear functional on V is a linear map from V to F. In other words, a linear functional is an element of L(V; F). Examples: Define ': R3 ! R by '(x; y; z) = 4x − 5y + 2z. Then ' is a linear functional on R3. Define ': P(R) ! R by '(p) = 3p00(5) + 7p(4). Then ' is a linear functional on P(R). R 1 Define ': P(R) ! R by '(p) = 0 p(x) dx.
    [Show full text]
  • SPINORS and SPACE–TIME ANISOTROPY
    Sergiu Vacaru and Panayiotis Stavrinos SPINORS and SPACE{TIME ANISOTROPY University of Athens ————————————————— c Sergiu Vacaru and Panyiotis Stavrinos ii - i ABOUT THE BOOK This is the first monograph on the geometry of anisotropic spinor spaces and its applications in modern physics. The main subjects are the theory of grav- ity and matter fields in spaces provided with off–diagonal metrics and asso- ciated anholonomic frames and nonlinear connection structures, the algebra and geometry of distinguished anisotropic Clifford and spinor spaces, their extension to spaces of higher order anisotropy and the geometry of gravity and gauge theories with anisotropic spinor variables. The book summarizes the authors’ results and can be also considered as a pedagogical survey on the mentioned subjects. ii - iii ABOUT THE AUTHORS Sergiu Ion Vacaru was born in 1958 in the Republic of Moldova. He was educated at the Universities of the former URSS (in Tomsk, Moscow, Dubna and Kiev) and reveived his PhD in theoretical physics in 1994 at ”Al. I. Cuza” University, Ia¸si, Romania. He was employed as principal senior researcher, as- sociate and full professor and obtained a number of NATO/UNESCO grants and fellowships at various academic institutions in R. Moldova, Romania, Germany, United Kingdom, Italy, Portugal and USA. He has published in English two scientific monographs, a university text–book and more than hundred scientific works (in English, Russian and Romanian) on (super) gravity and string theories, extra–dimension and brane gravity, black hole physics and cosmolgy, exact solutions of Einstein equations, spinors and twistors, anistoropic stochastic and kinetic processes and thermodynamics in curved spaces, generalized Finsler (super) geometry and gauge gravity, quantum field and geometric methods in condensed matter physics.
    [Show full text]
  • General Relativity Fall 2019 Lecture 11: the Riemann Tensor
    General Relativity Fall 2019 Lecture 11: The Riemann tensor Yacine Ali-Ha¨ımoud October 8th 2019 The Riemann tensor quantifies the curvature of spacetime, as we will see in this lecture and the next. RIEMANN TENSOR: BASIC PROPERTIES α γ Definition { Given any vector field V , r[αrβ]V is a tensor field. Let us compute its components in some coordinate system: σ σ λ σ σ λ r[µrν]V = @[µ(rν]V ) − Γ[µν]rλV + Γλ[µrν]V σ σ λ σ λ λ ρ = @[µ(@ν]V + Γν]λV ) + Γλ[µ @ν]V + Γν]ρV 1 = @ Γσ + Γσ Γρ V λ ≡ Rσ V λ; (1) [µ ν]λ ρ[µ ν]λ 2 λµν where all partial derivatives of V µ cancel out after antisymmetrization. σ Since the left-hand side is a tensor field and V is a vector field, we conclude that R λµν is a tensor field as well { this is the tensor division theorem, which I encourage you to think about on your own. You can also check that explicitly from the transformation law of Christoffel symbols. This is the Riemann tensor, which measures the non-commutation of second derivatives of vector fields { remember that second derivatives of scalar fields do commute, by assumption. It is completely determined by the metric, and is linear in its second derivatives. Expression in LICS { In a LICS the Christoffel symbols vanish but not their derivatives. Let us compute the latter: 1 1 @ Γσ = @ gσδ (@ g + @ g − @ g ) = ησδ (@ @ g + @ @ g − @ @ g ) ; (2) µ νλ 2 µ ν λδ λ νδ δ νλ 2 µ ν λδ µ λ νδ µ δ νλ since the first derivatives of the metric components (thus of its inverse as well) vanish in a LICS.
    [Show full text]
  • A Some Basic Rules of Tensor Calculus
    A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e.
    [Show full text]