Einstein Summation Notation
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Package 'Einsum'
Package ‘einsum’ May 15, 2021 Type Package Title Einstein Summation Version 0.1.0 Description The summation notation suggested by Einstein (1916) <doi:10.1002/andp.19163540702> is a concise mathematical notation that implicitly sums over repeated indices of n- dimensional arrays. Many ordinary matrix operations (e.g. transpose, matrix multiplication, scalar product, 'diag()', trace etc.) can be written using Einstein notation. The notation is particularly convenient for expressing operations on arrays with more than two dimensions because the respective operators ('tensor products') might not have a standardized name. License MIT + file LICENSE Encoding UTF-8 SystemRequirements C++11 Suggests testthat, covr RdMacros mathjaxr RoxygenNote 7.1.1 LinkingTo Rcpp Imports Rcpp, glue, mathjaxr R topics documented: einsum . .2 einsum_package . .3 Index 5 1 2 einsum einsum Einstein Summation Description Einstein summation is a convenient and concise notation for operations on n-dimensional arrays. Usage einsum(equation_string, ...) einsum_generator(equation_string, compile_function = TRUE) Arguments equation_string a string in Einstein notation where arrays are separated by ’,’ and the result is separated by ’->’. For example "ij,jk->ik" corresponds to a standard matrix multiplication. Whitespace inside the equation_string is ignored. Unlike the equivalent functions in Python, einsum() only supports the explicit mode. This means that the equation_string must contain ’->’. ... the arrays that are combined. All arguments are converted to arrays with -
Linear Algebra I
Linear Algebra I Martin Otto Winter Term 2013/14 Contents 1 Introduction7 1.1 Motivating Examples.......................7 1.1.1 The two-dimensional real plane.............7 1.1.2 Three-dimensional real space............... 14 1.1.3 Systems of linear equations over Rn ........... 15 1.1.4 Linear spaces over Z2 ................... 21 1.2 Basics, Notation and Conventions................ 27 1.2.1 Sets............................ 27 1.2.2 Functions......................... 29 1.2.3 Relations......................... 34 1.2.4 Summations........................ 36 1.2.5 Propositional logic.................... 36 1.2.6 Some common proof patterns.............. 37 1.3 Algebraic Structures....................... 39 1.3.1 Binary operations on a set................ 39 1.3.2 Groups........................... 40 1.3.3 Rings and fields...................... 42 1.3.4 Aside: isomorphisms of algebraic structures...... 44 2 Vector Spaces 47 2.1 Vector spaces over arbitrary fields................ 47 2.1.1 The axioms........................ 48 2.1.2 Examples old and new.................. 50 2.2 Subspaces............................. 53 2.2.1 Linear subspaces..................... 53 2.2.2 Affine subspaces...................... 56 2.3 Aside: affine and linear spaces.................. 58 2.4 Linear dependence and independence.............. 60 3 4 Linear Algebra I | Martin Otto 2013 2.4.1 Linear combinations and spans............. 60 2.4.2 Linear (in)dependence.................. 62 2.5 Bases and dimension....................... 65 2.5.1 Bases............................ 65 2.5.2 Finite-dimensional vector spaces............. 66 2.5.3 Dimensions of linear and affine subspaces........ 71 2.5.4 Existence of bases..................... 72 2.6 Products, sums and quotients of spaces............. 73 2.6.1 Direct products...................... 73 2.6.2 Direct sums of subspaces................ -
Notes on Manifolds
Notes on Manifolds Justin H. Le Department of Electrical & Computer Engineering University of Nevada, Las Vegas [email protected] August 3, 2016 1 Multilinear maps A tensor T of order r can be expressed as the tensor product of r vectors: T = u1 ⊗ u2 ⊗ ::: ⊗ ur (1) We herein fix r = 3 whenever it eases exposition. Recall that a vector u 2 U can be expressed as the combination of the basis vectors of U. Transform these basis vectors with a matrix A, and if the resulting vector u0 is equivalent to uA, then the components of u are said to be covariant. If u0 = A−1u, i.e., the vector changes inversely with the change of basis, then the components of u are contravariant. By Einstein notation, we index the covariant components of a tensor in subscript and the contravariant components in superscript. Just as the components of a vector u can be indexed by an integer i (as in ui), tensor components can be indexed as Tijk. Additionally, as we can view a matrix to be a linear map M : U ! V from one finite-dimensional vector space to another, we can consider a tensor to be multilinear map T : V ∗r × V s ! R, where V s denotes the s-th-order Cartesian product of vector space V with itself and likewise for its algebraic dual space V ∗. In this sense, a tensor maps an ordered sequence of vectors to one of its (scalar) components. Just as a linear map satisfies M(a1u1 + a2u2) = a1M(u1) + a2M(u2), we call an r-th-order tensor multilinear if it satisfies T (u1; : : : ; a1v1 + a2v2; : : : ; ur) = a1T (u1; : : : ; v1; : : : ; ur) + a2T (u1; : : : ; v2; : : : ; ur); (2) for scalars a1 and a2. -
Vector Spaces in Physics
San Francisco State University Department of Physics and Astronomy August 6, 2015 Vector Spaces in Physics Notes for Ph 385: Introduction to Theoretical Physics I R. Bland TABLE OF CONTENTS Chapter I. Vectors A. The displacement vector. B. Vector addition. C. Vector products. 1. The scalar product. 2. The vector product. D. Vectors in terms of components. E. Algebraic properties of vectors. 1. Equality. 2. Vector Addition. 3. Multiplication of a vector by a scalar. 4. The zero vector. 5. The negative of a vector. 6. Subtraction of vectors. 7. Algebraic properties of vector addition. F. Properties of a vector space. G. Metric spaces and the scalar product. 1. The scalar product. 2. Definition of a metric space. H. The vector product. I. Dimensionality of a vector space and linear independence. J. Components in a rotated coordinate system. K. Other vector quantities. Chapter 2. The special symbols ij and ijk, the Einstein summation convention, and some group theory. A. The Kronecker delta symbol, ij B. The Einstein summation convention. C. The Levi-Civita totally antisymmetric tensor. Groups. The permutation group. The Levi-Civita symbol. D. The cross Product. E. The triple scalar product. F. The triple vector product. The epsilon killer. Chapter 3. Linear equations and matrices. A. Linear independence of vectors. B. Definition of a matrix. C. The transpose of a matrix. D. The trace of a matrix. E. Addition of matrices and multiplication of a matrix by a scalar. F. Matrix multiplication. G. Properties of matrix multiplication. H. The unit matrix I. Square matrices as members of a group. -
The Mechanics of the Fermionic and Bosonic Fields: an Introduction to the Standard Model and Particle Physics
The Mechanics of the Fermionic and Bosonic Fields: An Introduction to the Standard Model and Particle Physics Evan McCarthy Phys. 460: Seminar in Physics, Spring 2014 Aug. 27,! 2014 1.Introduction 2.The Standard Model of Particle Physics 2.1.The Standard Model Lagrangian 2.2.Gauge Invariance 3.Mechanics of the Fermionic Field 3.1.Fermi-Dirac Statistics 3.2.Fermion Spinor Field 4.Mechanics of the Bosonic Field 4.1.Spin-Statistics Theorem 4.2.Bose Einstein Statistics !5.Conclusion ! 1. Introduction While Quantum Field Theory (QFT) is a remarkably successful tool of quantum particle physics, it is not used as a strictly predictive model. Rather, it is used as a framework within which predictive models - such as the Standard Model of particle physics (SM) - may operate. The overarching success of QFT lends it the ability to mathematically unify three of the four forces of nature, namely, the strong and weak nuclear forces, and electromagnetism. Recently substantiated further by the prediction and discovery of the Higgs boson, the SM has proven to be an extraordinarily proficient predictive model for all the subatomic particles and forces. The question remains, what is to be done with gravity - the fourth force of nature? Within the framework of QFT theoreticians have predicted the existence of yet another boson called the graviton. For this reason QFT has a very attractive allure, despite its limitations. According to !1 QFT the gravitational force is attributed to the interaction between two gravitons, however when applying the equations of General Relativity (GR) the force between two gravitons becomes infinite! Results like this are nonsensical and must be resolved for the theory to stand. -
Multilinear Algebra and Applications July 15, 2014
Multilinear Algebra and Applications July 15, 2014. Contents Chapter 1. Introduction 1 Chapter 2. Review of Linear Algebra 5 2.1. Vector Spaces and Subspaces 5 2.2. Bases 7 2.3. The Einstein convention 10 2.3.1. Change of bases, revisited 12 2.3.2. The Kronecker delta symbol 13 2.4. Linear Transformations 14 2.4.1. Similar matrices 18 2.5. Eigenbases 19 Chapter 3. Multilinear Forms 23 3.1. Linear Forms 23 3.1.1. Definition, Examples, Dual and Dual Basis 23 3.1.2. Transformation of Linear Forms under a Change of Basis 26 3.2. Bilinear Forms 30 3.2.1. Definition, Examples and Basis 30 3.2.2. Tensor product of two linear forms on V 32 3.2.3. Transformation of Bilinear Forms under a Change of Basis 33 3.3. Multilinear forms 34 3.4. Examples 35 3.4.1. A Bilinear Form 35 3.4.2. A Trilinear Form 36 3.5. Basic Operation on Multilinear Forms 37 Chapter 4. Inner Products 39 4.1. Definitions and First Properties 39 4.1.1. Correspondence Between Inner Products and Symmetric Positive Definite Matrices 40 4.1.1.1. From Inner Products to Symmetric Positive Definite Matrices 42 4.1.1.2. From Symmetric Positive Definite Matrices to Inner Products 42 4.1.2. Orthonormal Basis 42 4.2. Reciprocal Basis 46 4.2.1. Properties of Reciprocal Bases 48 4.2.2. Change of basis from a basis to its reciprocal basis g 50 B B III IV CONTENTS 4.2.3. -
An Attempt to Intuitively Introduce the Dot, Wedge, Cross, and Geometric Products
An attempt to intuitively introduce the dot, wedge, cross, and geometric products Peeter Joot March 21, 2008 1 Motivation. Both the NFCM and GAFP books have axiomatic introductions of the gener- alized (vector, blade) dot and wedge products, but there are elements of both that I was unsatisfied with. Perhaps the biggest issue with both is that they aren’t presented in a dumb enough fashion. NFCM presents but does not prove the generalized dot and wedge product operations in terms of symmetric and antisymmetric sums, but it is really the grade operation that is fundamental. You need that to define the dot product of two bivectors for example. GAFP axiomatic presentation is much clearer, but the definition of general- ized wedge product as the totally antisymmetric sum is a bit strange when all the differential forms book give such a different definition. Here I collect some of my notes on how one starts with the geometric prod- uct action on colinear and perpendicular vectors and gets the familiar results for two and three vector products. I may not try to generalize this, but just want to see things presented in a fashion that makes sense to me. 2 Introduction. The aim of this document is to introduce a “new” powerful vector multiplica- tion operation, the geometric product, to a student with some traditional vector algebra background. The geometric product, also called the Clifford product 1, has remained a relatively obscure mathematical subject. This operation actually makes a great deal of vector manipulation simpler than possible with the traditional methods, and provides a way to naturally expresses many geometric concepts. -
Dirac Notation 1 Vectors
Physics 324, Fall 2001 Dirac Notation 1 Vectors 1.1 Inner product T Recall from linear algebra: we can represent a vector V as a column vector; then V y = (V )∗ is a row vector, and the inner product (another name for dot product) between two vectors is written as AyB = A∗B1 + A∗B2 + : (1) 1 2 ··· In conventional vector notation, the above is just A~∗ B~ . Note that the inner product of a vector with itself is positive definite; we can define the· norm of a vector to be V = pV V; (2) j j y which is a non-negative real number. (In conventional vector notation, this is V~ , which j j is the length of V~ ). 1.2 Basis vectors We can expand a vector in a set of basis vectors e^i , provided the set is complete, which means that the basis vectors span the whole vectorf space.g The basis is called orthonormal if they satisfy e^iye^j = δij (orthonormality); (3) and an orthonormal basis is complete if they satisfy e^ e^y = I (completeness); (4) X i i i where I is the unit matrix. (note that a column vectore ^i times a row vectore ^iy is a square matrix, following the usual definition of matrix multiplication). Assuming we have a complete orthonormal basis, we can write V = IV = e^ e^yV V e^ ;V (^eyV ) : (5) X i i X i i i i i ≡ i ≡ The Vi are complex numbers; we say that Vi are the components of V in the e^i basis. -
Abstract Tensor Systems As Monoidal Categories
Abstract Tensor Systems as Monoidal Categories Aleks Kissinger Dedicated to Joachim Lambek on the occasion of his 90th birthday October 31, 2018 Abstract The primary contribution of this paper is to give a formal, categorical treatment to Penrose’s abstract tensor notation, in the context of traced symmetric monoidal categories. To do so, we introduce a typed, sum-free version of an abstract tensor system and demonstrate the construction of its associated category. We then show that the associated category of the free abstract tensor system is in fact the free traced symmetric monoidal category on a monoidal signature. A notable consequence of this result is a simple proof for the soundness and completeness of the diagrammatic language for traced symmetric monoidal categories. 1 Introduction This paper formalises the connection between monoidal categories and the ab- stract index notation developed by Penrose in the 1970s, which has been used by physicists directly, and category theorists implicitly, via the diagrammatic languages for traced symmetric monoidal and compact closed categories. This connection is given as a representation theorem for the free traced symmet- ric monoidal category as a syntactically-defined strict monoidal category whose morphisms are equivalence classes of certain kinds of terms called Einstein ex- pressions. Representation theorems of this kind form a rich history of coherence results for monoidal categories originating in the 1960s [17, 6]. Lambek’s con- arXiv:1308.3586v1 [math.CT] 16 Aug 2013 tribution [15, 16] plays an essential role in this history, providing some of the earliest examples of syntactically-constructed free categories and most of the key ingredients in Kelly and Mac Lane’s proof of the coherence theorem for closed monoidal categories [11]. -
Quantum Chromodynamics (QCD) QCD Is the Theory That Describes the Action of the Strong Force
Quantum chromodynamics (QCD) QCD is the theory that describes the action of the strong force. QCD was constructed in analogy to quantum electrodynamics (QED), the quantum field theory of the electromagnetic force. In QED the electromagnetic interactions of charged particles are described through the emission and subsequent absorption of massless photons (force carriers of QED); such interactions are not possible between uncharged, electrically neutral particles. By analogy with QED, quantum chromodynamics predicts the existence of gluons, which transmit the strong force between particles of matter that carry color, a strong charge. The color charge was introduced in 1964 by Greenberg to resolve spin-statistics contradictions in hadron spectroscopy. In 1965 Nambu and Han introduced the octet of gluons. In 1972, Gell-Mann and Fritzsch, coined the term quantum chromodynamics as the gauge theory of the strong interaction. In particular, they employed the general field theory developed in the 1950s by Yang and Mills, in which the carrier particles of a force can themselves radiate further carrier particles. (This is different from QED, where the photons that carry the electromagnetic force do not radiate further photons.) First, QED Lagrangian… µ ! # 1 µν LQED = ψeiγ "∂µ +ieAµ $ψe − meψeψe − Fµν F 4 µν µ ν ν µ Einstein notation: • F =∂ A −∂ A EM field tensor when an index variable µ • A four potential of the photon field appears twice in a single term, it implies summation µ •γ Dirac 4x4 matrices of that term over all the values of the index -
About Matrices, Tensors and Various Abbrevia- Tions
About matrices, tensors and various abbrevia- tions The objects we deal with in this course are rather complicated. We therefore use a simplifying notation where at every step as much as possible of this complexity is hidden. Basically this means that objects that are made out of many elements or components. are written without indices as much as possible. We also have several different types of indices present. The ones that show up in this course are: • Dirac-Indices a; b; c • Lorentz indices, both upper and lower µ, ν; ρ • SU(2)L indices i; j; k • SU(3)c indices α; β; γ Towards the right I have written the type of symbols used in this note to denote a particular type of index. The course contains even more, three-vector indices or various others denoting sums over types of quarks and/or leptons. An object is called scalar or singlet if it has no index of a particular type of index, a vector if it has one, a matrix if it has two and a tensor if it has two or more. A vector can also be called a column or row matrix. Examples are b with elements bi: 0 1 b1 B C B b2 C b = (b1; b2; : : : ; bn) b = (b1 b2 ··· bn) b = B . C (1) B . C @ . A bn In the first case is a vector, the second a row matrix and the last a column vector. Objects with two indices are called a matrix or sometimes a tensor and de- noted by 0 1 c11 c12 ··· c1n B C B c21 c22 ··· c2n C c = (cij) = B . -
Tensor Algebra
TENSOR ALGEBRA Continuum Mechanics Course (MMC) - ETSECCPB - UPC Introduction to Tensors Tensor Algebra 2 Introduction SCALAR , , ... v VECTOR vf, , ... MATRIX σε,,... ? C,... 3 Concept of Tensor A TENSOR is an algebraic entity with various components which generalizes the concepts of scalar, vector and matrix. Many physical quantities are mathematically represented as tensors. Tensors are independent of any reference system but, by need, are commonly represented in one by means of their “component matrices”. The components of a tensor will depend on the reference system chosen and will vary with it. 4 Order of a Tensor The order of a tensor is given by the number of indexes needed to specify without ambiguity a component of a tensor. a Scalar: zero dimension 3.14 1.2 v 0.3 a , a Vector: 1 dimension i 0.8 0.1 0 1.3 2nd order: 2 dimensions A, A E 02.40.5 ij rd A , A 3 order: 3 dimensions 1.3 0.5 5.8 A , A 4th order … 5 Cartesian Coordinate System Given an orthonormal basis formed by three mutually perpendicular unit vectors: eeˆˆ12,, ee ˆˆ 23 ee ˆˆ 31 Where: eeeˆˆˆ1231, 1, 1 Note that 1 if ij eeˆˆi j ij 0 if ij 6 Cylindrical Coordinate System x3 xr1 cos x(,rz , ) xr2 sin xz3 eeeˆˆˆr cosθθ 12 sin eeeˆˆˆsinθθ cos x2 12 eeˆˆz 3 x1 7 Spherical Coordinate System x3 xr1 sin cos xrxr, , 2 sin sin xr3 cos ˆˆˆˆ x2 eeeer sinθφ sin 123sin θ cos φ cos θ eeeˆˆˆ cosφφ 12sin x1 eeeeˆˆˆˆφ cosθφ sin 123cos θ cos φ sin θ 8 Indicial or (Index) Notation Tensor Algebra 9 Tensor Bases – VECTOR A vector v can be written as a unique linear combination of the three vector basis eˆ for i 1, 2, 3 .