Appendix a Spinors in Four Dimensions

Total Page:16

File Type:pdf, Size:1020Kb

Appendix a Spinors in Four Dimensions Appendix A Spinors in Four Dimensions In this appendix we collect the conventions used for spinors in both Minkowski and Euclidean spaces. In Minkowski space the flat metric has the 0 1 2 3 form ηµν = diag(−1, 1, 1, 1), and the coordinates are labelled (x ,x , x , x ). The analytic continuation into Euclidean space is madethrough the replace- ment x0 = ix4 (and in momentum space, p0 = −ip4) the coordinates in this case being labelled (x1,x2, x3, x4). The Lorentz group in four dimensions, SO(3, 1), is not simply connected and therefore, strictly speaking, has no spinorial representations. To deal with these types of representations one must consider its double covering, the spin group Spin(3, 1), which is isomorphic to SL(2, C). The group SL(2, C) pos- sesses a natural complex two-dimensional representation. Let us denote this representation by S andlet us consider an element ψ ∈ S with components ψα =(ψ1,ψ2) relative to some basis. The action of an element M ∈ SL(2, C) is β (Mψ)α = Mα ψβ. (A.1) This is not the only action of SL(2, C) which one could choose. Instead of M we could have used its complex conjugate M, its inverse transpose (M T)−1,or its inverse adjoint (M †)−1. All of them satisfy the same group multiplication law. These choices would correspond to the complex conjugate representation S, the dual representation S,and the dual complex conjugate representation S. We will use the following conventions for elements of these representations: α α˙ ψα ∈ S, ψα˙ ∈ S, ψ ∈ S, ψ ∈ S. (A.2) These representations are called spinorial representations and are not inde- pendent. The SL(2, C) invariant tensor αβ (and α˙ β˙ ) allows one to relate a representation to its dual. We will use the following conventions for this 204 Spinors in four dimensions 205 tensor: 12 21 21 = = −12 = − = 1, 1˙2˙ 2˙1˙ (A.3) 2˙1˙ = = −1˙2˙ = − = 1. It allows one to raise and lower spinor indices: α αβ β ψ = ψβ, ψα = αβψ , (A.4) α˙ α˙ β˙ β˙ ψ = ψβ˙ , ψα˙ = α˙ β˙ ψ . One can easily verify that these relations are consistent with the assignment ˙ βδ β β˙δ˙ β (A.2). Notice also that in our conventions αβ = δα, α˙ β˙ = δα˙ . The real form of the complex Lie algebras sl(2, C) and su(2)+ × su(2)− are the same. This allows one to use the notation (j1, j2) for the represen- tationsofSL(2, C) where j1 and j2 are the spins of the two su(2)s. The representation S corresponds to (2, 0) whilst S to (0, 2). Notice that the two su(2)’s are not independent but related by complex conjugation. Under complex conjugation the labels of the representations must be interchanged. This allows one to restrict to representations which are fixed by complex con- jugation in the case in which the two labels are the same. The resulting representation is real, and the simplest case, (2, 2), corresponds to the defin- ing representation of the Lorentz group SO(3, 1) or vector representation. The Clebsch–Gordan coefficients which intertwine between the vector representation and the (2, 2) of SL(2, C) involve the Pauli matrices: 0 1 0 −i 10 σ1 = ,σ2 = ,σ3 = , (A.5) 10 i 0 0 −1 and have theform µ 1 2 3 σ = (1,σ ,σ ,σ ) ˙ , (A.6) αβ˙ αβ allowing one to transform a four-vector pµ = (p0,p1,p2, p3) into a bi-spinor: µ p0 + p3 p1 − ip2 σ pµ = . (A.7) p1 + ip2 p0 − p3 The indices for the matrices σµ show how SL(2, C) acts in bi-spinor space. µ µ † If M ∈ SL(2, C), the action of M is given by σ pµ → M(σ pµ)M , which preserves hermiticity and the determinant. The determinant is related to the µ µ 2 2 2 2 Lorentz invariant quantity p pµ = −det(σ pµ) = −p0 +p1 +p2 +p3 (the minus sign in front of the determinant is owed to our choice of metric). Notice that 206 Topological Quantum Field Theory and Four-Manifolds M and −M act in the same way on the vector representation, showing the fact that SL(2, C) is a double cover of SO(3, 1). The dual Clebsch–Gordan coefficients are defined after raising the indices of σµ, ˙ (σµ)αβ˙ = α˙ ββασµ −→ (σµ) = (1, −σ1, −σ2, −σ3)(A.8) αβ˙ and allow to transform a vector into a bi-spinor corresponding to the dual representations. The matrices σµ and σµ satisfy the following set of useful identities: σµσν + σν σµ = − 2ηµν , (A.9) µ αβ˙ − β α˙ (σ ) (σµ)γρ˙ = 2δγ δρ˙ , which can be easily derived after using the following property of the Pauli matrices σi, i = 1, 2, 3: σiσj = δij + iijkσk (A.10) where ijk is the totally antisymmetric tensor with 123 =1. The generators of the Lorentz transformations can be constructed with the help of the matrices σµ and σµ. They take the form: 1 (σµν ) β = σµσν − σν σµ β, α 4 α ˙ (A.11) µν β˙ 1 µ ν ν µ β (σ ) ˙ = σ σ − σ σ ˙ , α 4 α which leads to the simple form: 1 i σ0i = − σi,σij = − ijkσk, 2 2 (A.12) 1 i σ0i = σi, σij = − ijkσk, 2 2 and satisfy the properties: i i σµν = µνρσσ , σµν = − µνρσσ , (A.13) 2 ρσ 2 ρσ 0123 being 0123 =1, = −1. The matrices (A.11) provide spinorial representations of the Lorentz group: M µν → iσµν ,Mµν =→ iσµν , (A.14) satisfying the Lorentz algebra: [M µν ,Mρτ ] = iηµρM ντ − iηµτ M νρ − iηνρM µτ + iηντ M µρ. (A.15) Spinors in four dimensions 207 Under a Lorentz transformation parametrized by a real antisymmetric pa- rameter ωµν spinors transform as: µν β µν β δψα = iωµν (M )α ψβ = −ωµν (σ )α ψβ, (A.16) µν β˙ − µν β˙ δψα˙ = iωµν (M )α˙ ψβ˙ = ωµν (σ )α˙ ψβ˙ . For the scalar product of spinors we will use the following conventions: α α ψχ = ψ χα = −ψαχ = χψ ˙ (A.17) α − α˙ ψχ = ψ χα˙ = ψα˙ χ = χψ. In Euclidean space the relevant group is the orthonormal group SO(4). Inthis case the covering group is Spin(4), which is isomorphic to SU(2)+ ⊗ SU(2)−. The representations are labelled by the spins of the two SU(2)s, which are independent. These spinorial representations are denoted by S+ and S−.The SU(2) invariant tensor can be used to raise and lower in- dices following the same conventions (equations (A.3) and (A.4)) as in the Minkowskian case. The Clebsh-Gordan coefficients which intertwine between the bi-spinorial representations and the defining representation of SO(4), or vector represent- ation, are σµ = (σ1, σ2, σ3,i), σµ =(−σ1, −σ2, −σ3, i). (A.18) Given a four vector (p1,p2, p3,p4) the negative of the determinant of its corre- µ 2 2 2 2 sponding bi-spinor now becomes −det(σ pµ) = p1 +p2 +p3 +p4,asexpected. The matrices σµ and σµ satisfy the identities σµσν + σν σµ = − 2δµν , (A.19) µ αβ˙ − β α˙ (σ ) (σµ)γρ˙ = 2δγ δρ˙ . In order to construct the generators of the rotation group SO(4) one first defines the following matrices: 1 (σµν ) β = σµσν − σν σµ β, α 4 α ˙ (A.20) µν β˙ 1 µ ν ν µ β (σ ) ˙ = σ σ − σ σ ˙ , α 4 α One then easily finds out that the generators M µν = iσµν (or M µν = iσµν ) satisfy the SO(4) algebra: [M µν ,Mρτ ] = iδµρM ντ − iδµτ M νρ − iδνρM µτ + iδντ M µρ. (A.21) 208 Topological Quantum Field Theory and Four-Manifolds The explicit forms of the matrices σµν and σµν are ⎛ ⎞ i 3 i 2 i 1 0 − 2 σ 2 σ 2 σ ⎜ i 3 − i 1 i 2 ⎟ µν ⎜ 2 σ 0 2 σ 2 σ ⎟ σ = ⎝ i 2 i 1 i 3 ⎠ , − 2 σ 2 σ 0 2 σ i 1 i 2 i 3 − 2 σ − 2 σ − 2 σ 0 (A.22) ⎛ ⎞ i 3 i 2 i 1 0 − 2 σ 2 σ − 2 σ ⎜ i 3 − i 1 − i 2 ⎟ µν ⎜ 2 σ 0 2 σ 2 σ ⎟ σ = ⎝ i 2 i 1 i 3 ⎠ . − 2 σ 2 σ 0 − 2 σ i 1 i 2 i 3 2 σ 2 σ 2 σ 0 These matrices act as projectors of antisymmetric tensors and two-forms into their self-dual (SD) and anti-self-dual (ASD) parts. Given an antisymmetric + − tensor Vµν one defines its SD part, Vµν , and ASD part, Vµν , as follows: 1 1 V ± = (V ± V ρσ), (A.23) µν 2 µν 2 µνρσ where µνρσ isthe totally antisymmetric tensor corresponding to thevolume µν µν form (1234 =1). The matrices σ and σ satisfy the relations: 1 1 σµν = − σρσ, σµν = σρσ (A.24) 2 µνρσ 2 µνρσ and therefore they act as projectors into SD and ASD parts, providing the corresponding bi-spinor form: µν γ − µν γ Vαβ =βγ(σ )α Vµν = βγ(σ )α Vµν , (A.25) µν γ˙ + µν γ˙ Vα˙ β˙ =β˙γ˙ (σ ) α˙ Vµν = α˙ γ˙ (σ ) α˙ Vµν . Thetensors Vαβ and Vα˙ β˙ are symmetric and transform as the (3, 0) and the (0, 3) representations of SO(4). The antisymmetric tensor Vµν can bealso written in spinorial form as µ ν Vαα,β˙ β˙ = (σ )αα˙ (σ )ββ˙ Vµν . (A.26) It turns out that the bi-spinor version of the decomposition of the antisym- metric tensor into SD and ASD parts takes the form: + − −→ Vµν = Vµν + Vµν Vαα,β˙ β˙ = α˙ β˙ Vαβ + αβVα˙ β˙ .
Recommended publications
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Handout 9 More Matrix Properties; the Transpose
    Handout 9 More matrix properties; the transpose Square matrix properties These properties only apply to a square matrix, i.e. n £ n. ² The leading diagonal is the diagonal line consisting of the entries a11, a22, a33, . ann. ² A diagonal matrix has zeros everywhere except the leading diagonal. ² The identity matrix I has zeros o® the leading diagonal, and 1 for each entry on the diagonal. It is a special case of a diagonal matrix, and A I = I A = A for any n £ n matrix A. ² An upper triangular matrix has all its non-zero entries on or above the leading diagonal. ² A lower triangular matrix has all its non-zero entries on or below the leading diagonal. ² A symmetric matrix has the same entries below and above the diagonal: aij = aji for any values of i and j between 1 and n. ² An antisymmetric or skew-symmetric matrix has the opposite entries below and above the diagonal: aij = ¡aji for any values of i and j between 1 and n. This automatically means the digaonal entries must all be zero. Transpose To transpose a matrix, we reect it across the line given by the leading diagonal a11, a22 etc. In general the result is a di®erent shape to the original matrix: a11 a21 a11 a12 a13 > > A = A = 0 a12 a22 1 [A ]ij = A : µ a21 a22 a23 ¶ ji a13 a23 @ A > ² If A is m £ n then A is n £ m. > ² The transpose of a symmetric matrix is itself: A = A (recalling that only square matrices can be symmetric).
    [Show full text]
  • Geometric-Algebra Adaptive Filters Wilder B
    1 Geometric-Algebra Adaptive Filters Wilder B. Lopes∗, Member, IEEE, Cassio G. Lopesy, Senior Member, IEEE Abstract—This paper presents a new class of adaptive filters, namely Geometric-Algebra Adaptive Filters (GAAFs). They are Faces generated by formulating the underlying minimization problem (a deterministic cost function) from the perspective of Geometric Algebra (GA), a comprehensive mathematical language well- Edges suited for the description of geometric transformations. Also, (directed lines) differently from standard adaptive-filtering theory, Geometric Calculus (the extension of GA to differential calculus) allows Fig. 1. A polyhedron (3-dimensional polytope) can be completely described for applying the same derivation techniques regardless of the by the geometric multiplication of its edges (oriented lines, vectors), which type (subalgebra) of the data, i.e., real, complex numbers, generate the faces and hypersurfaces (in the case of a general n-dimensional quaternions, etc. Relying on those characteristics (among others), polytope). a deterministic quadratic cost function is posed, from which the GAAFs are devised, providing a generalization of regular adaptive filters to subalgebras of GA. From the obtained update rule, it is shown how to recover the following least-mean squares perform calculus with hypercomplex quantities, i.e., elements (LMS) adaptive filter variants: real-entries LMS, complex LMS, that generalize complex numbers for higher dimensions [2]– and quaternions LMS. Mean-square analysis and simulations in [10]. a system identification scenario are provided, showing very good agreement for different levels of measurement noise. GA-based AFs were first introduced in [11], [12], where they were successfully employed to estimate the geometric Index Terms—Adaptive filtering, geometric algebra, quater- transformation (rotation and translation) that aligns a pair of nions.
    [Show full text]
  • Matrices and Tensors
    APPENDIX MATRICES AND TENSORS A.1. INTRODUCTION AND RATIONALE The purpose of this appendix is to present the notation and most of the mathematical tech- niques that are used in the body of the text. The audience is assumed to have been through sev- eral years of college-level mathematics, which included the differential and integral calculus, differential equations, functions of several variables, partial derivatives, and an introduction to linear algebra. Matrices are reviewed briefly, and determinants, vectors, and tensors of order two are described. The application of this linear algebra to material that appears in under- graduate engineering courses on mechanics is illustrated by discussions of concepts like the area and mass moments of inertia, Mohr’s circles, and the vector cross and triple scalar prod- ucts. The notation, as far as possible, will be a matrix notation that is easily entered into exist- ing symbolic computational programs like Maple, Mathematica, Matlab, and Mathcad. The desire to represent the components of three-dimensional fourth-order tensors that appear in anisotropic elasticity as the components of six-dimensional second-order tensors and thus rep- resent these components in matrices of tensor components in six dimensions leads to the non- traditional part of this appendix. This is also one of the nontraditional aspects in the text of the book, but a minor one. This is described in §A.11, along with the rationale for this approach. A.2. DEFINITION OF SQUARE, COLUMN, AND ROW MATRICES An r-by-c matrix, M, is a rectangular array of numbers consisting of r rows and c columns: ¯MM..
    [Show full text]
  • 5 the Dirac Equation and Spinors
    5 The Dirac Equation and Spinors In this section we develop the appropriate wavefunctions for fundamental fermions and bosons. 5.1 Notation Review The three dimension differential operator is : ∂ ∂ ∂ = , , (5.1) ∂x ∂y ∂z We can generalise this to four dimensions ∂µ: 1 ∂ ∂ ∂ ∂ ∂ = , , , (5.2) µ c ∂t ∂x ∂y ∂z 5.2 The Schr¨odinger Equation First consider a classical non-relativistic particle of mass m in a potential U. The energy-momentum relationship is: p2 E = + U (5.3) 2m we can substitute the differential operators: ∂ Eˆ i pˆ i (5.4) → ∂t →− to obtain the non-relativistic Schr¨odinger Equation (with = 1): ∂ψ 1 i = 2 + U ψ (5.5) ∂t −2m For U = 0, the free particle solutions are: iEt ψ(x, t) e− ψ(x) (5.6) ∝ and the probability density ρ and current j are given by: 2 i ρ = ψ(x) j = ψ∗ ψ ψ ψ∗ (5.7) | | −2m − with conservation of probability giving the continuity equation: ∂ρ + j =0, (5.8) ∂t · Or in Covariant notation: µ µ ∂µj = 0 with j =(ρ,j) (5.9) The Schr¨odinger equation is 1st order in ∂/∂t but second order in ∂/∂x. However, as we are going to be dealing with relativistic particles, space and time should be treated equally. 25 5.3 The Klein-Gordon Equation For a relativistic particle the energy-momentum relationship is: p p = p pµ = E2 p 2 = m2 (5.10) · µ − | | Substituting the equation (5.4), leads to the relativistic Klein-Gordon equation: ∂2 + 2 ψ = m2ψ (5.11) −∂t2 The free particle solutions are plane waves: ip x i(Et p x) ψ e− · = e− − · (5.12) ∝ The Klein-Gordon equation successfully describes spin 0 particles in relativistic quan- tum field theory.
    [Show full text]
  • A Some Basic Rules of Tensor Calculus
    A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e.
    [Show full text]
  • 13 the Dirac Equation
    13 The Dirac Equation A two-component spinor a χ = b transforms under rotations as iθn J χ e− · χ; ! with the angular momentum operators, Ji given by: 1 Ji = σi; 2 where σ are the Pauli matrices, n is the unit vector along the axis of rotation and θ is the angle of rotation. For a relativistic description we must also describe Lorentz boosts generated by the operators Ki. Together Ji and Ki form the algebra (set of commutation relations) Ki;Kj = iεi jkJk − ε Ji;Kj = i i jkKk ε Ji;Jj = i i jkJk 1 For a spin- 2 particle Ki are represented as i Ki = σi; 2 giving us two inequivalent representations. 1 χ Starting with a spin- 2 particle at rest, described by a spinor (0), we can boost to give two possible spinors α=2n σ χR(p) = e · χ(0) = (cosh(α=2) + n σsinh(α=2))χ(0) · or α=2n σ χL(p) = e− · χ(0) = (cosh(α=2) n σsinh(α=2))χ(0) − · where p sinh(α) = j j m and Ep cosh(α) = m so that (Ep + m + σ p) χR(p) = · χ(0) 2m(Ep + m) σ (pEp + m p) χL(p) = − · χ(0) 2m(Ep + m) p 57 Under the parity operator the three-moment is reversed p p so that χL χR. Therefore if we 1 $ − $ require a Lorentz description of a spin- 2 particles to be a proper representation of parity, we must include both χL and χR in one spinor (note that for massive particles the transformation p p $ − can be achieved by a Lorentz boost).
    [Show full text]
  • 1 Vectors & Tensors
    1 Vectors & Tensors The mathematical modeling of the physical world requires knowledge of quite a few different mathematics subjects, such as Calculus, Differential Equations and Linear Algebra. These topics are usually encountered in fundamental mathematics courses. However, in a more thorough and in-depth treatment of mechanics, it is essential to describe the physical world using the concept of the tensor, and so we begin this book with a comprehensive chapter on the tensor. The chapter is divided into three parts. The first part covers vectors (§1.1-1.7). The second part is concerned with second, and higher-order, tensors (§1.8-1.15). The second part covers much of the same ground as done in the first part, mainly generalizing the vector concepts and expressions to tensors. The final part (§1.16-1.19) (not required in the vast majority of applications) is concerned with generalizing the earlier work to curvilinear coordinate systems. The first part comprises basic vector algebra, such as the dot product and the cross product; the mathematics of how the components of a vector transform between different coordinate systems; the symbolic, index and matrix notations for vectors; the differentiation of vectors, including the gradient, the divergence and the curl; the integration of vectors, including line, double, surface and volume integrals, and the integral theorems. The second part comprises the definition of the tensor (and a re-definition of the vector); dyads and dyadics; the manipulation of tensors; properties of tensors, such as the trace, transpose, norm, determinant and principal values; special tensors, such as the spherical, identity and orthogonal tensors; the transformation of tensor components between different coordinate systems; the calculus of tensors, including the gradient of vectors and higher order tensors and the divergence of higher order tensors and special fourth order tensors.
    [Show full text]
  • Some Key Facts About Transpose
    Some key facts about transpose Let A be an m × n matrix. Then AT is the matrix which switches the rows and columns of A. For example 0 1 T 1 2 1 01 5 3 41 5 7 3 2 7 0 9 = B C @ A B3 0 2C 1 3 2 6 @ A 4 9 6 We have the following useful identities: (AT )T = A (A + B)T = AT + BT (kA)T = kAT Transpose Facts 1 (AB)T = BT AT (AT )−1 = (A−1)T ~v · ~w = ~vT ~w A deeper fact is that Rank(A) = Rank(AT ): Transpose Fact 2 Remember that Rank(B) is dim(Im(B)), and we compute Rank as the number of leading ones in the row reduced form. Recall that ~u and ~v are perpendicular if and only if ~u · ~v = 0. The word orthogonal is a synonym for perpendicular. n ? n If V is a subspace of R , then V is the set of those vectors in R which are perpendicular to every vector in V . V ? is called the orthogonal complement to V ; I'll often pronounce it \Vee perp" for short. ? n You can (and should!) check that V is a subspace of R . It is geometrically intuitive that dim V ? = n − dim V Transpose Fact 3 and that (V ?)? = V: Transpose Fact 4 We will prove both of these facts later in this note. In this note we will also show that Ker(A) = Im(AT )? Im(A) = Ker(AT )? Transpose Fact 5 As is often the case in math, the best order to state results is not the best order to prove them.
    [Show full text]
  • Properties of Transpose
    3.2, 3.3 Inverting Matrices P. Danziger Properties of Transpose Transpose has higher precedence than multiplica- tion and addition, so T T T T AB = A B and A + B = A + B As opposed to the bracketed expressions (AB)T and (A + B)T Example 1 1 2 1 1 0 1 Let A = ! and B = !. 2 5 2 1 1 0 Find ABT , and (AB)T . T 1 1 1 2 1 1 0 1 1 2 1 0 1 ABT = ! ! = ! 0 1 2 5 2 1 1 0 2 5 2 B C @ 1 0 A 2 3 = ! 4 7 Whereas (AB)T is undefined. 1 3.2, 3.3 Inverting Matrices P. Danziger Theorem 2 (Properties of Transpose) Given ma- trices A and B so that the operations can be pre- formed 1. (AT )T = A 2. (A + B)T = AT + BT and (A B)T = AT BT − − 3. (kA)T = kAT 4. (AB)T = BT AT 2 3.2, 3.3 Inverting Matrices P. Danziger Matrix Algebra Theorem 3 (Algebraic Properties of Matrix Multiplication) 1. (k + `)A = kA + `A (Distributivity of scalar multiplication I) 2. k(A + B) = kA + kB (Distributivity of scalar multiplication II) 3. A(B + C) = AB + AC (Distributivity of matrix multiplication) 4. A(BC) = (AB)C (Associativity of matrix mul- tiplication) 5. A + B = B + A (Commutativity of matrix ad- dition) 6. (A + B) + C = A + (B + C) (Associativity of matrix addition) 7. k(AB) = A(kB) (Commutativity of Scalar Mul- tiplication) 3 3.2, 3.3 Inverting Matrices P. Danziger The matrix 0 is the identity of matrix addition.
    [Show full text]
  • 16 the Transpose
    The Transpose Learning Goals: to expose some properties of the transpose We would like to see what happens with our factorization if we get zeroes in the pivot positions and have to employ row exchanges. To do this we need to look at permutation matrices. But to study these effectively, we need to know something about the transpose. So the transpose of a matrix is what you get by swapping rows for columns. In symbols, T ⎡2 −1⎤ T ⎡ 2 0 4⎤ ⎢ ⎥ A ij = Aji. By example, ⎢ ⎥ = 0 3 . ⎣−1 3 5⎦ ⎢ ⎥ ⎣⎢4 5 ⎦⎥ Let’s take a look at how the transpose interacts with algebra: T T T T • (A + B) = A + B . This is simple in symbols because (A + B) ij = (A + B)ji = Aji + Bji = Y T T T A ij + B ij = (A + B )ij. In practical terms, it just doesn’t matter if we add first and then flip or flip first and then add • (AB)T = BTAT. We can do this in symbols, but it really doesn’t help explain why this is n true: (AB)T = (AB) = T T = (BTAT) . To get a better idea of what is ij ji ∑a jkbki = ∑b ika kj ij k =1 really going on, look at what happens if B is a single column b. Then Ab is a combination of the columns of A. If we transpose this, we get a row which is a combination of the rows of AT. In other words, (Ab)T = bTAT. Then we get the whole picture of (AB)T by placing several columns side-by-side, which is putting a whole bunch of rows on top of each other in BT.
    [Show full text]
  • Differential Geometry: Discrete Exterior Calculus
    Differential Geometry: Discrete Exterior Calculus [Discrete Exterior Calculus (Thesis). Hirani, 2003] [Discrete Differential Forms for Computational Modeling. Desbrun et al., 2005] [Build Your Own DEC at Home. Elcott et al., 2006] Chains Recall: A k-chain of a simplicial complex Κ is linear combination of the k-simplices in Κ: c = ∑c(σ )⋅σ σ∈Κ k where c is a real-valued function. The dual of a k-chain is a k-cochain which is a linear map taking a k-chain to a real value. Chains and cochains related through evaluation: –Cochains: What we integrate –Chains: What we integrate over Boundary Operator Recall: The boundary ∂σ of a k-simplex σ is the signed union of all (k-1)-faces. ∂ ∂ ∂ ∂ +1 -1 The boundary operator extends linearly to act on chains: ∂c = ∂ ∑c(σ )⋅σ = ∑c(σ )⋅∂σ σ∈Κ k σ∈Κ k Discrete Exterior Derivative Recall: The exterior derivative d:Ωk→ Ωk+1 is the operator on cochains that is the complement of the boundary operator, defined to satisfy Stokes’ Theorem. -1 -2 -1 -3.5 -2 0.5 0.5 2 0 0.5 2 0 0.5 2 1.5 1 1 Since the boundary of the boundary is empty: ∂∂ = 0 dd = 0 The Laplacian Definition: The Laplacian of a function f is a function defined as the divergence of the gradient of f: ∆f = div(∇f ) = ∇ ⋅∇f The Laplacian Definition: The Laplacian of a function f is a function defined as the divergence of the gradient of f: ∆f = div(∇f ) = ∇ ⋅∇f Q.
    [Show full text]