Tensors Ocontravariant and Covariant Tensors Oeinstein Notation Ointerpretation What Is a Tensor?

Total Page:16

File Type:pdf, Size:1020Kb

Tensors Ocontravariant and Covariant Tensors Oeinstein Notation Ointerpretation What Is a Tensor? Tensors oContravariant and covariant tensors oEinstein notation oInterpretation What is a Tensor? ! It is the generalization of a: � – Matrix operator for >2 dimensions output input – Vector data to >1 dimension �� �� or ! Why do we need it? – Came out of differential geometry. – Was used by Albert Einstein to formulate the theory of General Relativity. – It’s needed for many problems, including Deep NNs. Contravariant and Covariant Vectors � = �� ! Contravariant vectors: – “Column Vectors” that represent data – Vectors that describe the position of something ! contravariant – � for 0 ≤ � < � and � ∈ ℜ covariant ! Covariant vector: � = � is 1×� 1 – “Row vectors” that operate on data × � – Gradient vectors is � – �! for 0 ≤ � < � Picture of multiplication ! Einstein notation $%& ! ! � = �! � = - �! � !"# – Leave out the sum to make notation cleaner – Always sum over any two indices that appear twice Vector-Matrix Products as Tensors � = �� covariant rows contravariant ! 1D Contravariant vectors: ! = – � for 0 ≤ � < � � ' � is ! is � – � for 0 ≤ � < �( � is �×� � × × 1 1 ! 2D tensor (i.e., matrix): ) Picture of multiplication – � ! for 0 ≤ � < �( and 0 ≤ � < �' – There is a gradient for each component �) ! Einstein notation $!%& ) ) ! ) ! � = � ! � = - � ! � !"# – Leave out the sum to make notation cleaner – Always sum over any two indices that appear twice Tensor Products Sum over pairs ! Einstein notation of indices #(,#) #(,#) %(,%) � = � %(,%) � ! 2-D Contravariant Tensors: % ,% – � ( ) for 0 ≤ �&, �' < �( #(,#) – � for 0 ≤ �&, �' < �) ! 4-D Tensor: #(,#) – � %(,%) for 0 ≤ �&, �' < �) and 0 ≤ �&, �' < �( – � is known as a tensor • 2D covariant input • 2D contravariant output Picture of a Tensor Product ! For example, if we have ! ! "!,"" � = � "!,"" � 3D tensor contravariant 2D tensor � = � � � � – Tensor 3-D tensor � has Input: 2D image indexed by �&, �* Output: 1D vector indexed by � – General idea: � ∈ ℜ**×*+ Some Useful Definitions ! Delta functions: 1 if � = � �# = 1 % 0 if � ≠ � – So then we have that , , # � % = � # � % !Gradient w.r.t. vector ∇( �� = � – In tensor notation, we have that # ∇ �� = �# ( % % !Gradient w.r.t. matrix ∇- �� =? ? – First, it must have 1 output dimension and 2 input dimensions. So we have that ∇ �� # = �# �%) - %(,%) %( GD for Single Layer NN oStructure of the Gradient oGradients for NN parameters oUpdates for NN parameters Gradient Direction for Single Layer NN ! Single layer NN: �. � = �� �� + � � �� + � ⋅ � �= � � � – We will need gradient w.r.t the parameters � = �, �, � : ∇.�. � = ∇-� -,/,0 � , ∇/� -,/,0 � , ∇0� -,/,0 � – Later, we will also need: ∇(�. � Gradient Structure for Single Layer NN �� ∇'� + + + ��= ∇$� ∇%� ∇&� �� �� �� ! Single layer NN: – Parameters are � = �, �, � – We will need the parameter gradients: ∇.�. � = ∇-� -,/,0 � , ∇/� -,/,0 � , ∇0� -,/,0 � – And the input gradient: ∇(�. � Gradient w.r.t. � )* " = (& !" + % " !" + & ⋅ ( #$ ! % ( ! For this case, ∇1� = ∇1�2 � = ∇1� � �� + � = � ∇� �� + � � – Using Einstein notation 3 ∇ � = �3 ∇� 3! �3" 1 4 3! 3" 4 This is a matrix or 2-D tensor Gradient w.r.t. � )* " = (& !" + % " !" + & ⋅ ( #$ ! For this case, ! % ( ∇5� = � ∇� �� + � ∇5 �� – Using Einstein notation, since This is a tensor! # # %) ∇- �� % ,% = � %( � – Then ( ) ∇ � 3! = �3! ∇� 3" ∇ �� 3# 5 4!,4" 3" 3# 5 4!,4" = �3! ∇� 3" �3# �4" 3" 3# 4! �� ∇ � 3! = �3! ∇� 3" �4" 5 4!,4" 3" 4! Gradient w.r.t. � )* " = (& !" + % " !" + & ⋅ ( #$ ! % ( ! For this case, ∇6� = � ∇� �� + � ∇6 �� + � = � ∇� � = � ∇� – Using Einstein notation, ∇ � 3! = �3! ∇� 3" 6 4! 3" 4! Gradient w.r.t. � )* " = (& !" + % " !" + & ⋅ ( #$ ! % ( ! For this case, ∇7� = ∇7 �� This is a tensor! – Using Einstein notation, since ∇ �� # = �# �%) 0 %(,%) %( – We have that ∇ � 3! = �3! �4" 7 4!,4" 4! Update Direction for � Gradient step: � ← � + ��1 where � is given by ;<= 2 3! 3" 4" �4 ,4 = −∇5� � = 2 �8,3 � 3 ∇�8 �8 ! " � ! " 4! 89: $%& 2 �, 1×�! �"×�" , �! � �!×�" ∇�, = !"# �, For efficiency, computation goes this way! Update Direction for � Gradient step: � ← � + ��1 where � is given by ;<= 2 3! 3" �4 = −∇6� � = 2 �8,3 � 3 ∇�8 ! � ! " 4! 89: $%& 2 �, 1×�! = 1×�" , �"×�" � �!×�" ∇�, !"# �, For efficiency, computation goes this way! Update Direction for � Gradient step: � ← � + ��1 where � is given by ;<= 2 � = −∇ � � = 2 � � 4" 4!,4" 7 � 8,4! 8 89: $%& � 1×� 2 , ! = , �, � �" !"# For efficiency, computation goes this way! Local and Global Minima oOpen and Closed Sets oConvex Sets and Functions oProperties of Convex Functions oLocal Minimum, Saddle Points, and Global Minima oOptimization Theorems Open and Closed Sets ! Define: – � ⊂ ℜ$ $ – Open ball of diameter � is � �, � = � ∈ ℜ : � − �+ < � . ! A set � is open if – At every point, there is an open ball contained in �. – ∀� ∈ �, ∃� > 0 s.t. � �, � ⊂ �. ! A set � is closed if �[ = ℜ\ − � is open. ! A set � is compact if it is closed and bounded. !Facts: – ℜ$ is both open and closed, but it is not compact. – If � is compact, than every sequence in � has a limit point in �. Convexity Sets ! A set � is convex if ∀� ∈ 0,1 , ∀�, � ∈ �, then �� + 1 − � � ∈ � Line connecting Line connecting points is points is always in set sometimes outside set Convex Set Nonconvex Set ! Properties: – The intersection of convex sets is convex Convexity Functions ! Let �: � → ℜ where � is a convex set. Then we say that � is a convex function if ∀� ∈ 0,1 , ∀�, � ∈ �, then � �� + 1 − � � ≤ �� � + 1 − � � � Line is always Line is sometimes above function below function � � � � � � Convex Function Nonconvex Function ! Properties: – The sums of convex functions is convex – The maximum of a set of convex functions is convex – If � � is convex, then � �� is also convex. – � � is concave if −� � is convex. – � � = �� + � is both convex and concave Properties of Convex Functions ! Valuable properties of convex functions • The sums of convex functions is convex. – Let � � = ∑# �# � . If �# � are convex, then � � is convex. • The maximum of a set of convex functions is convex. – Let � � = max �# � . If �# � are convex, then � � is convex. # • The second derivative of a convex function is positive. '!( – Let � � have two continuous derivatives, and let �$,& � = be the ')"')# Hessian of � at �. Then � � is convex if and only if � � is non-negative definite for all �. • A convex function of a linear transform is convex. – If � � is convex, then � �� is also convex. • An affine transform is both convex and concave. – � � = �� + � is both convex and concave. • A function is concave if its negative is convex. – � � is concave if −� � is convex. Local Minimum ! Let �: � → ℜ where � ⊂ ℜ\. – We say that �+ ∈ � is a local minimum of � if there ∃� > 0 s.t. ∀ � ∈ � ∩ � �, � , � � ≥ � �+ . local minimum �+ • Necessary condition for local minima: – Let � be continuously differentiable and let �+ be a local minimum, then ∇� �+ = 0. �J a local minimum ∇� �J = 0 Saddle Point ! Let �: � → ℜ where � ⊂ ℜ\ be a continuously differentiable function. – We say that �+ ∈ � is a saddle point of � if ∇� �+ = 0 and �+ is not a local minimum. saddle point in 2D* saddle point in 1D � + * * Graph of � � = �& − �* – Saddle points can cause more problems than you might think. *Shared from Wikipedia Global Minimum ! Let �: � → ℜ where � ⊂ ℜ\. – We say that �+ ∈ � is a global minimum of � if ∀ � ∈ �, � � ≥ � �+ . GD steps global minimum Local minimum �+ ! Comments: – In general, finding global minimum is difficult. – Gradient descent optimization typically becomes trapped in local minima. Optimization Theorems ! Let �: � → ℜ where � ⊂ ℜ\. – If � is continuous and � is compact, then � takes on a global minimum in �. – If � is convex on �, then any local minimum is a global minimum. – If � is continuously differentiable and convex on �, then ∇� �* = 0 implies that �* ∈ � is a global minimum of �. ! Important facts: – Global minimum may not be unique. – If � is closed but not bounded, then it may not take on a global minimum. – Generally speaking, gradient descent algorithms converge to the global minimum of continuously differentiable convex functions. – Most interesting functions in ML are not convex! Convex Function GD steps global minimum .
Recommended publications
  • Package 'Einsum'
    Package ‘einsum’ May 15, 2021 Type Package Title Einstein Summation Version 0.1.0 Description The summation notation suggested by Einstein (1916) <doi:10.1002/andp.19163540702> is a concise mathematical notation that implicitly sums over repeated indices of n- dimensional arrays. Many ordinary matrix operations (e.g. transpose, matrix multiplication, scalar product, 'diag()', trace etc.) can be written using Einstein notation. The notation is particularly convenient for expressing operations on arrays with more than two dimensions because the respective operators ('tensor products') might not have a standardized name. License MIT + file LICENSE Encoding UTF-8 SystemRequirements C++11 Suggests testthat, covr RdMacros mathjaxr RoxygenNote 7.1.1 LinkingTo Rcpp Imports Rcpp, glue, mathjaxr R topics documented: einsum . .2 einsum_package . .3 Index 5 1 2 einsum einsum Einstein Summation Description Einstein summation is a convenient and concise notation for operations on n-dimensional arrays. Usage einsum(equation_string, ...) einsum_generator(equation_string, compile_function = TRUE) Arguments equation_string a string in Einstein notation where arrays are separated by ’,’ and the result is separated by ’->’. For example "ij,jk->ik" corresponds to a standard matrix multiplication. Whitespace inside the equation_string is ignored. Unlike the equivalent functions in Python, einsum() only supports the explicit mode. This means that the equation_string must contain ’->’. ... the arrays that are combined. All arguments are converted to arrays with
    [Show full text]
  • Notes on Manifolds
    Notes on Manifolds Justin H. Le Department of Electrical & Computer Engineering University of Nevada, Las Vegas [email protected] August 3, 2016 1 Multilinear maps A tensor T of order r can be expressed as the tensor product of r vectors: T = u1 ⊗ u2 ⊗ ::: ⊗ ur (1) We herein fix r = 3 whenever it eases exposition. Recall that a vector u 2 U can be expressed as the combination of the basis vectors of U. Transform these basis vectors with a matrix A, and if the resulting vector u0 is equivalent to uA, then the components of u are said to be covariant. If u0 = A−1u, i.e., the vector changes inversely with the change of basis, then the components of u are contravariant. By Einstein notation, we index the covariant components of a tensor in subscript and the contravariant components in superscript. Just as the components of a vector u can be indexed by an integer i (as in ui), tensor components can be indexed as Tijk. Additionally, as we can view a matrix to be a linear map M : U ! V from one finite-dimensional vector space to another, we can consider a tensor to be multilinear map T : V ∗r × V s ! R, where V s denotes the s-th-order Cartesian product of vector space V with itself and likewise for its algebraic dual space V ∗. In this sense, a tensor maps an ordered sequence of vectors to one of its (scalar) components. Just as a linear map satisfies M(a1u1 + a2u2) = a1M(u1) + a2M(u2), we call an r-th-order tensor multilinear if it satisfies T (u1; : : : ; a1v1 + a2v2; : : : ; ur) = a1T (u1; : : : ; v1; : : : ; ur) + a2T (u1; : : : ; v2; : : : ; ur); (2) for scalars a1 and a2.
    [Show full text]
  • The Mechanics of the Fermionic and Bosonic Fields: an Introduction to the Standard Model and Particle Physics
    The Mechanics of the Fermionic and Bosonic Fields: An Introduction to the Standard Model and Particle Physics Evan McCarthy Phys. 460: Seminar in Physics, Spring 2014 Aug. 27,! 2014 1.Introduction 2.The Standard Model of Particle Physics 2.1.The Standard Model Lagrangian 2.2.Gauge Invariance 3.Mechanics of the Fermionic Field 3.1.Fermi-Dirac Statistics 3.2.Fermion Spinor Field 4.Mechanics of the Bosonic Field 4.1.Spin-Statistics Theorem 4.2.Bose Einstein Statistics !5.Conclusion ! 1. Introduction While Quantum Field Theory (QFT) is a remarkably successful tool of quantum particle physics, it is not used as a strictly predictive model. Rather, it is used as a framework within which predictive models - such as the Standard Model of particle physics (SM) - may operate. The overarching success of QFT lends it the ability to mathematically unify three of the four forces of nature, namely, the strong and weak nuclear forces, and electromagnetism. Recently substantiated further by the prediction and discovery of the Higgs boson, the SM has proven to be an extraordinarily proficient predictive model for all the subatomic particles and forces. The question remains, what is to be done with gravity - the fourth force of nature? Within the framework of QFT theoreticians have predicted the existence of yet another boson called the graviton. For this reason QFT has a very attractive allure, despite its limitations. According to !1 QFT the gravitational force is attributed to the interaction between two gravitons, however when applying the equations of General Relativity (GR) the force between two gravitons becomes infinite! Results like this are nonsensical and must be resolved for the theory to stand.
    [Show full text]
  • Multilinear Algebra and Applications July 15, 2014
    Multilinear Algebra and Applications July 15, 2014. Contents Chapter 1. Introduction 1 Chapter 2. Review of Linear Algebra 5 2.1. Vector Spaces and Subspaces 5 2.2. Bases 7 2.3. The Einstein convention 10 2.3.1. Change of bases, revisited 12 2.3.2. The Kronecker delta symbol 13 2.4. Linear Transformations 14 2.4.1. Similar matrices 18 2.5. Eigenbases 19 Chapter 3. Multilinear Forms 23 3.1. Linear Forms 23 3.1.1. Definition, Examples, Dual and Dual Basis 23 3.1.2. Transformation of Linear Forms under a Change of Basis 26 3.2. Bilinear Forms 30 3.2.1. Definition, Examples and Basis 30 3.2.2. Tensor product of two linear forms on V 32 3.2.3. Transformation of Bilinear Forms under a Change of Basis 33 3.3. Multilinear forms 34 3.4. Examples 35 3.4.1. A Bilinear Form 35 3.4.2. A Trilinear Form 36 3.5. Basic Operation on Multilinear Forms 37 Chapter 4. Inner Products 39 4.1. Definitions and First Properties 39 4.1.1. Correspondence Between Inner Products and Symmetric Positive Definite Matrices 40 4.1.1.1. From Inner Products to Symmetric Positive Definite Matrices 42 4.1.1.2. From Symmetric Positive Definite Matrices to Inner Products 42 4.1.2. Orthonormal Basis 42 4.2. Reciprocal Basis 46 4.2.1. Properties of Reciprocal Bases 48 4.2.2. Change of basis from a basis to its reciprocal basis g 50 B B III IV CONTENTS 4.2.3.
    [Show full text]
  • Abstract Tensor Systems As Monoidal Categories
    Abstract Tensor Systems as Monoidal Categories Aleks Kissinger Dedicated to Joachim Lambek on the occasion of his 90th birthday October 31, 2018 Abstract The primary contribution of this paper is to give a formal, categorical treatment to Penrose’s abstract tensor notation, in the context of traced symmetric monoidal categories. To do so, we introduce a typed, sum-free version of an abstract tensor system and demonstrate the construction of its associated category. We then show that the associated category of the free abstract tensor system is in fact the free traced symmetric monoidal category on a monoidal signature. A notable consequence of this result is a simple proof for the soundness and completeness of the diagrammatic language for traced symmetric monoidal categories. 1 Introduction This paper formalises the connection between monoidal categories and the ab- stract index notation developed by Penrose in the 1970s, which has been used by physicists directly, and category theorists implicitly, via the diagrammatic languages for traced symmetric monoidal and compact closed categories. This connection is given as a representation theorem for the free traced symmet- ric monoidal category as a syntactically-defined strict monoidal category whose morphisms are equivalence classes of certain kinds of terms called Einstein ex- pressions. Representation theorems of this kind form a rich history of coherence results for monoidal categories originating in the 1960s [17, 6]. Lambek’s con- arXiv:1308.3586v1 [math.CT] 16 Aug 2013 tribution [15, 16] plays an essential role in this history, providing some of the earliest examples of syntactically-constructed free categories and most of the key ingredients in Kelly and Mac Lane’s proof of the coherence theorem for closed monoidal categories [11].
    [Show full text]
  • Quantum Chromodynamics (QCD) QCD Is the Theory That Describes the Action of the Strong Force
    Quantum chromodynamics (QCD) QCD is the theory that describes the action of the strong force. QCD was constructed in analogy to quantum electrodynamics (QED), the quantum field theory of the electromagnetic force. In QED the electromagnetic interactions of charged particles are described through the emission and subsequent absorption of massless photons (force carriers of QED); such interactions are not possible between uncharged, electrically neutral particles. By analogy with QED, quantum chromodynamics predicts the existence of gluons, which transmit the strong force between particles of matter that carry color, a strong charge. The color charge was introduced in 1964 by Greenberg to resolve spin-statistics contradictions in hadron spectroscopy. In 1965 Nambu and Han introduced the octet of gluons. In 1972, Gell-Mann and Fritzsch, coined the term quantum chromodynamics as the gauge theory of the strong interaction. In particular, they employed the general field theory developed in the 1950s by Yang and Mills, in which the carrier particles of a force can themselves radiate further carrier particles. (This is different from QED, where the photons that carry the electromagnetic force do not radiate further photons.) First, QED Lagrangian… µ ! # 1 µν LQED = ψeiγ "∂µ +ieAµ $ψe − meψeψe − Fµν F 4 µν µ ν ν µ Einstein notation: • F =∂ A −∂ A EM field tensor when an index variable µ • A four potential of the photon field appears twice in a single term, it implies summation µ •γ Dirac 4x4 matrices of that term over all the values of the index
    [Show full text]
  • About Matrices, Tensors and Various Abbrevia- Tions
    About matrices, tensors and various abbrevia- tions The objects we deal with in this course are rather complicated. We therefore use a simplifying notation where at every step as much as possible of this complexity is hidden. Basically this means that objects that are made out of many elements or components. are written without indices as much as possible. We also have several different types of indices present. The ones that show up in this course are: • Dirac-Indices a; b; c • Lorentz indices, both upper and lower µ, ν; ρ • SU(2)L indices i; j; k • SU(3)c indices α; β; γ Towards the right I have written the type of symbols used in this note to denote a particular type of index. The course contains even more, three-vector indices or various others denoting sums over types of quarks and/or leptons. An object is called scalar or singlet if it has no index of a particular type of index, a vector if it has one, a matrix if it has two and a tensor if it has two or more. A vector can also be called a column or row matrix. Examples are b with elements bi: 0 1 b1 B C B b2 C b = (b1; b2; : : : ; bn) b = (b1 b2 ··· bn) b = B . C (1) B . C @ . A bn In the first case is a vector, the second a row matrix and the last a column vector. Objects with two indices are called a matrix or sometimes a tensor and de- noted by 0 1 c11 c12 ··· c1n B C B c21 c22 ··· c2n C c = (cij) = B .
    [Show full text]
  • Tensor Algebra
    TENSOR ALGEBRA Continuum Mechanics Course (MMC) - ETSECCPB - UPC Introduction to Tensors Tensor Algebra 2 Introduction SCALAR , , ... v VECTOR vf, , ... MATRIX σε,,... ? C,... 3 Concept of Tensor A TENSOR is an algebraic entity with various components which generalizes the concepts of scalar, vector and matrix. Many physical quantities are mathematically represented as tensors. Tensors are independent of any reference system but, by need, are commonly represented in one by means of their “component matrices”. The components of a tensor will depend on the reference system chosen and will vary with it. 4 Order of a Tensor The order of a tensor is given by the number of indexes needed to specify without ambiguity a component of a tensor. a Scalar: zero dimension 3.14 1.2 v 0.3 a , a Vector: 1 dimension i 0.8 0.1 0 1.3 2nd order: 2 dimensions A, A E 02.40.5 ij rd A , A 3 order: 3 dimensions 1.3 0.5 5.8 A , A 4th order … 5 Cartesian Coordinate System Given an orthonormal basis formed by three mutually perpendicular unit vectors: eeˆˆ12,, ee ˆˆ 23 ee ˆˆ 31 Where: eeeˆˆˆ1231, 1, 1 Note that 1 if ij eeˆˆi j ij 0 if ij 6 Cylindrical Coordinate System x3 xr1 cos x(,rz , ) xr2 sin xz3 eeeˆˆˆr cosθθ 12 sin eeeˆˆˆsinθθ cos x2 12 eeˆˆz 3 x1 7 Spherical Coordinate System x3 xr1 sin cos xrxr, , 2 sin sin xr3 cos ˆˆˆˆ x2 eeeer sinθφ sin 123sin θ cos φ cos θ eeeˆˆˆ cosφφ 12sin x1 eeeeˆˆˆˆφ cosθφ sin 123cos θ cos φ sin θ 8 Indicial or (Index) Notation Tensor Algebra 9 Tensor Bases – VECTOR A vector v can be written as a unique linear combination of the three vector basis eˆ for i 1, 2, 3 .
    [Show full text]
  • Introduction to Quantum Mechanics Unit 0. Review of Linear Algebra
    Introduction to Quantum Mechanics Unit 0. Review of Linear Algebra A. Linear Space 1. Definition: A linear space is a collection S of vectors |a>, |b>, |c>, …. (usually infinite number of them) on which vector addition + is defined so that (i) (S,+) forms an Abelien group, i,e,, S is closed under + + is associate: (|a>+|b>)+|c> = |a>+(|b>+|c>) + is commutative: |a>+|b>=|b>+|a> Existence of identity: |a> + |0> = |a> for all |a> Existence of inverse: |a> + (|-a>) = |0> and also scalar (usually complex number) * is defined so that (ii) S is closed under * (iii) * and complex number multiplication is associative: α(β*|a>)=(αβ)*|a> (iv) 1|a>=|a> (and hence 0|a>=|0> and |a>=-|a>) (v) * and + is distributive: (α+β)|a>=α|a>+β|a> and α(|a>+|b>)=α|a>+α|b> 2. Basis and dimension (i) Definition of a basis: A basis is a collection of vectors B={|a1>, |a2>,|a3>,….|aN>} such that any vector in S can be written as a linear combination of a1>, |a2>,|a3>,….|aN>: a = α1 a1 + α 2 a 2 + α3 a 3 +Lα N a N where α1 ,α 2 ,Kα N are complex numbers Furthermore, ={|a1>, |a2>,|a3>,….|aN> are independent of each other, i.e. 0 = α1 a1 + α 2 a 2 + α 3 a 3 +Lα N a N if and only if α1 = α 2 = K = α N = 0 (ii) Choice of a basis is not unique. (iii) However, all basises have the same number of elements. This number (say, N) depends only on the linear space S, and it is known as the dimension of the linear space S.
    [Show full text]
  • Lost in the Tensors: Einstein's Struggles with Covariance Principles 1912-1916"
    JOHN EARMAN and CLARK GL YMOUR LOST IN THE TENSORS: EINSTEIN'S STRUGGLES WITH COVARIANCE PRINCIPLES 1912-1916" Introduction IN 1912 Einstein began to devote a major portion of his time and energy to an attempt to construct a relativistic theory of gravitation. A strong intimation of the struggle that lay ahead is contained in a letter to Arnold Sommerfeld dated October 29, 1912: At the moment I am working solely on the problem of gravitation and believe 1 will be able to overcome all difficulties with the help of a local, friendly mathemat- ician. But one thing is certain, that I have never worked so hard in my life, and that I have been injected with a great awe of mathematics, which in my naivet~ until now I only viewed as a pure luxury in its subtler forms! Compared to this problem the original theory of relativity is mere child's play.' Einstein's letter contained only a perfunctory reply to a query from Sommerfeld about the Debye-Born theory of specific heats. Obviously disappointed, Som- merfeld wrote to Hilbert: 'My letter to Einstein was in vain . Einstein is evidently so deeply mired in gravitation that he is deaf to everything else? Sommerfeld's words were more prophetic than he could possibly have known; the next three years were to see Einstein deeply mired in gravitation, sometimes seemingly hopelessly so. In large measure, Einstein's struggle resulted from his use and his misuse, his understanding and his misunderstanding of the nature and implications of covariance principles. In brief, considerations of general covariance were bound up with Einstein's motive for seeking a 'generalized' theory of relativity; mis- understandings about the meaning and implementation of this motivation threatened to wreck the search; and in the end, the desire for general covariance helped to bring Einstein back onto the track which led to what we now recognize *Present address c/o Department of Philosophy, University of Minnesota, Minneapolis, Minn, U.S.A.
    [Show full text]
  • Indicial Notation
    Appendix A Indicial Notation A.1 Indicial Notation for Vector and Matrix Operations In tensor analysis, an extensive use of indicial notation is made. Operations using Cartesian components of vectors and matrices can be expressed efficiently using this notation. A.1.1 Elements of a Matrix A set of n variables x1, x2,...,xn is denoted using indicial notation such as as xi , i = 1, 2,...,n. Let us consider now the equation that describes a plane in a three dimensional space x1, x2, x3, a1x1 + a2x2 + a3x3 = p, where ai and p are constants. This equation can be expressed alternatively as 3 ai xi = p. i=1 At this point, we can use the convention usually referred to as Einstein notation or Einstein summation convention, which implies summation over a set of indexed terms in a formula when an index appears twice in a single term along the variational rank of the indexes (1 to 3 in the above equation). Therefore, ai xi = p In the same spirit, © Springer International Publishing AG, part of Springer Nature 2018 227 E. Cueto and D. González, An Introduction to Structural Mechanics for Architects, Structural Integrity 4, https://doi.org/10.1007/978-3-319-72935-0 228 Appendix A: Indicial Notation • Vectorial scalar product: u · v = ui vi • Norm of a vector: √ √ ||u|| = u · u = ui ui • Derivative of a function: n ∂ f ∂ f df = dx = dx = f, dx ∂x i ∂x i i i i=1 i i 1 Let us define the Kronecker Delta δij as a symbol whose values can be, = δ = 1ifi j ij 0ifi = j In the same way, let us define the permutation symbol εijk as, ⎧ ⎨ 0ifi = jj= ki= k ε = , , ijk ⎩ 1ifi j k for an even number permutation −1ifi, j, k form an odd permutation A.1.
    [Show full text]
  • 1 Einstein Notation
    PH 108 : Electricity & Magnetism Einstein Notation Rwitaban Goswami 1 Einstein Notation 2.6. The gradient operator r behaves like a vector in \ some sense". For example, divergence of a curl (r:r × A~ = 0) for any A~, may suggest that it is just like A:~ B~ × C~ being zero if any two vectors are equal. Prove that r × r × F~ = r(r:F~ ) − r2F~ . To what extent does this look like the well known expansion of A~ × B~ × C~ ? Solution: We will be using Einstein summation convention. According to this convention: 1. The whole term is summed over any index which appears twice in that term 2. Any index appearing only once in the term in not summed over and must be present in both LHS and RHS Don't be confused by this. The convention only is to simply omit the P symbol. You can mentally insert the P symbol to make sense of what the summation will look like. Now let us go over some common symbols that are used extensively in Einstein's notation: • δij is the Kronecker-Delta symbol. It is defined as ( 1 if i = j δ = ij 0 if i 6= j • ijk is the Levi-Civita symbol. It is defined as 81 if ijk is a cyclic permutation of xyz, eg xyz, yzx, zxy <> ijk = −1 if ijk is an anticyclic permuation of xyz, eg xzy, zyx, yxz :>0 if in ijk any index is repeating, eg xxy, xyy etc Now, you can verify: ~ ~ A · B = AiBi ~ ~ A × B = ijkAjBk e^i wheree ^i is the unit vector in i direction In the cross product formula, when the P symbol is reinserted, this simply means X X X ijkAjBke^i i=x;y;z j=x;y;z k=x;y;z A few things to note: P • δii = 3.
    [Show full text]