Introduction to Geometric Algebra Lecture IV

Total Page:16

File Type:pdf, Size:1020Kb

Introduction to Geometric Algebra Lecture IV Introduction to Geometric Algebra Lecture IV Leandro A. F. Fernandes Manuel M. Oliveira [email protected] [email protected] Visgraf - Summer School in Computer Graphics - 2010 CG UFRGS Lecture IV Checkpoint Visgraf - Summer School in Computer Graphics - 2010 2 Checkpoint, Lecture I Multivector space Non-metric products . The outer product . The regressive product c Visgraf - Summer School in Computer Graphics - 2010 3 Checkpoint, Lecture II Metric spaces . Bilinear form defines a metric on the vector space, e.g., Euclidean metric . Metric matrix Some inner products The scalar product is a particular . Inner product of vectors case of the left and right contractions . Scalar product . Left contraction These metric products are . Right contraction backward compatible for 1-blades Visgraf - Summer School in Computer Graphics - 2010 4 Checkpoint, Lecture II Dualization Undualization By taking the undual, the dual representation of a blade can be correctly mapped back to its direct representation Venn Diagrams Visgraf - Summer School in Computer Graphics - 2010 5 Checkpoint, Lecture III Duality relationships between products . Dual of the outer product . Dual of the left contraction Visgraf - Summer School in Computer Graphics - 2010 6 Checkpoint, Lecture III Some non-linear products . Meet of blades . Join of blades . Delta product of blades Visgraf - Summer School in ComputerVenn GraphicsDiagrams - 2010 7 Today Lecture IV – Mon, January 18 . Geometric product . Versors . Rotors Visgraf - Summer School in Computer Graphics - 2010 8 Lecture IV Geometric Product Visgraf - Summer School in Computer Graphics - 2010 9 Geometric product of vectors Denoted by a white space, like Inner Product Outer Product standard multiplication Unique Feature An invertible product for vectors! Visgraf - Summer School in Computer Graphics - 2010 10 Geometric product of vectors The Inner Product is not Invertible Unique Feature An invertible product for vectors! γ ** Euclidean Metric Visgraf - Summer School in Computer Graphics - 2010 11 Geometric product of vectors The Outer Product is not Invertible Unique Feature An invertible product for vectors! Visgraf - Summer School in Computer Graphics - 2010 12 Geometric product of vectors Unique Feature An invertible product for vectors! Inverse geometric product, denoted by a slash, like standard division Visgraf - Summer School in Computer Graphics - 2010 13 Intuitive solutions for simple problems q r ? p t ** Euclidean Metric Visgraf - Summer School in Computer Graphics - 2010 14 Geometric product and multivector space The geometric product of two vectors is an element of mixed dimensionality Scalars Vector Space Bivector Space Trivector Space Visgraf - Summer School in Computer Graphics - 2010 15 Geometric product and multivector space The geometric product of two vectors is an element of mixed dimensionality Scalars Vector Space Bivector Space Trivector Space Geometric Meaning The interpretation of the resulting element depends on the operands. Visgraf - Summer School in Computer Graphics - 2010 16 Properties of the geometric product Scalars commute Distributivity Associativity Neither fully symmetric nor fully antisymmetric Visgraf - Summer School in Computer Graphics - 2010 17 Lecture IV Geometric product of basis blades Visgraf - Summer School in Computer Graphics - 2010 18 Geometric product of basis blades Let’s assume an orthogonal metric, i.e., Kronecker delta function With an orthogonal metric, there are two cases to be handled Metric factor Visgraf - Summer School in Computer Graphics - 2010 19 Geometric product of basis blades Let’s assume an orthogonal metric, i.e., . Case 1: blades consisting of different orthogonal factors The geometric product is equivalent to the outer product . Case 2: blades with some common factors The dependent-basis factors are replaced Visgraf - Summer School in Computer Graphics - 2010 by metric factors 20 Geometric product of basis blades Let’s assume a non-orthogonal metric, e.g., Apply the spectral theorem from ** The spectral theorem states that a linear algebra and reduce the problem matrix is orthogonally diagonalizable to the orthogonal metric case if and only if it is symmetric. Visgraf - Summer School in Computer Graphics - 2010 21 Geometric product of basis blades For non-orthogonal metrics 1. Compute the eigenvectors and eigenvalues of the metric matrix 2. Represent the input with respect to the eigenbasis • Apply a change of basis using the inverse of the eigenvector matrix 3. Compute the geometric product on this new orthogonal basis • The eigenvalues specify the new orthogonal metric 4. Get back to the original basis • Apply a change of basis using the original eigenvector matrix Visgraf - Summer School in Computer Graphics - 2010 22 Geometric product of basis blades For non-orthogonal metrics 1. Compute the eigenvectors and eigenvalues of the metric matrix 2. Represent the input with respect to the eigenbasis • Apply a change of basis using the inverse of the eigenvector matrix 3. Compute the geometric product on this new orthogonal basis See the Supplementary Material A of the • The eigenvalues specifyTutorial the atnew Sibgrapi metric 2009 for a purest 4. Get back to the originaltreatment basis of the geometric product • Apply a change of basis using the original eigenvector matrix Visgraf - Summer School in Computer Graphics - 2010 23 Lecture IV Subspace Products from Geometric Product Visgraf - Summer School in Computer Graphics - 2010 24 The most fundamental product of GA The subspace products can be derived from the geometric product The “grade extraction” operation extracts grade parts from multivector A general multivector variable in Visgraf - Summer School in Computer Graphics - 2010 25 The most fundamental product of GA The subspace products can be derived from the geometric product Outer product Scalar product Left contraction Right contraction Delta product The largest grade such that the result is not zero Visgraf - Summer School in Computer Graphics - 2010 26 Lecture IV Orthogonal Transformations as Versors Visgraf - Summer School in Computer Graphics - 2010 27 Reflection of vectors Mirror Input vector Vector a was reflected in vector v, resulting in vector a´ Visgraf - Summer School in Computer Graphics - 2010 28 k-Versor V is a k-versor. It is computed as the geometric product of k invertible vectors. Visgraf - Summer School in Computer Graphics - 2010 29 Rotation of subspaces How to rotate vector a in the plane by radians. Visgraf - Summer School in Computer Graphics - 2010 30 Rotation of subspaces How to rotate vector a in the plane by radians. Visgraf - Summer School in Computer Graphics - 2010 31 Rotation of subspaces Rotors How to rotate vector a in Unit versors encoding rotations. the plane by radians. They are build as the geometric product of an even number of unit invertible vectors. Visgraf - Summer School in Computer Graphics - 2010 32 Versor product for general multivectors Grade Involution The sign change under the grade involution exhibits a + - + - + - … pattern over the value of t. Visgraf - Summer School in Computer Graphics - 2010 33 The structure preservation property The structure preservation of versors holds for the geometric product, and hence to all other products in geometric algebra. The ○ symbol represents any product of geometricof geometric algebra, algebra and, as a consequence, any operation defined from the products Visgraf - Summer School in Computer Graphics - 2010 34 Inverse of versors The inverse of versors is computed as for the inverse of invertible blades Squared (reverse) norm Inverse Reverse (+ + – – + + – – … pattern over k) The norm of rotors is equal to one, so the inverse of a rotor is its reverse Visgraf - Summer School in Computer Graphics - 2010 35 Multivector classification It can be used for blades and versors . Use Euclidean metric for blades . Use the actual metric for versors Test if 푀 푀 푀 is truly the inverse of the multivector −1 −1 −1 grade 푀 푀 = 0 푀 푀 = 푀 푀 Test the grade preservation property grade 푀 퐞푖 푀 = 1 If the multivector is of a single grade then it is a blade; otherwise it is a versor Visgraf - Summer School in Computer Graphics - 2010 36 Credits William R. Hamilton Hermann G. Grassmann William K. Clifford (1805-1865) (1809-1877) (1845-1879) W.Grassmann, R.Clifford, Hamilton W. H. K. (1844) G. (1878) (1877) On Applications aVerwendung new species of derGrassmann's of imaginaryAusdehnungslehre extensivequantities fur algebraconnected die allgemeine. Am. with J. theMath. Theorie theory, der Polarenof quaternionsund denWalter Zusammenhang. Inde Proc. Gruyter of theUnd algebraischer Royal Co., vol. Irish 1, Gebilde Acad.n. 4, 350, .vol. J.-358 Reine 2, 424 Angew.-434 Math. (Crelle's J.), Walter de Gruyter Und Co., 84, 273-283 Visgraf - Summer School in Computer Graphics - 2010 37 Differences between algebras Clifford algebra . Developed in nongeometric directions . Permits us to construct elements by a universal addition . Arbitrary multivectors may be important Geometric algebra . The geometrically significant part of Clifford algebra . Only permits exclusively multiplicative constructions • The only elements that can be added are scalars, vectors, pseudovectors, and pseudoscalars . Only blades and versors are important Visgraf - Summer School in Computer Graphics - 2010 38 .
Recommended publications
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Exercise 1 Advanced Sdps Matrices and Vectors
    Exercise 1 Advanced SDPs Matrices and vectors: All these are very important facts that we will use repeatedly, and should be internalized. • Inner product and outer products. Let u = (u1; : : : ; un) and v = (v1; : : : ; vn) be vectors in n T Pn R . Then u v = hu; vi = i=1 uivi is the inner product of u and v. T The outer product uv is an n × n rank 1 matrix B with entries Bij = uivj. The matrix B is a very useful operator. Suppose v is a unit vector. Then, B sends v to u i.e. Bv = uvT v = u, but Bw = 0 for all w 2 v?. m×n n×p • Matrix Product. For any two matrices A 2 R and B 2 R , the standard matrix Pn product C = AB is the m × p matrix with entries cij = k=1 aikbkj. Here are two very useful ways to view this. T Inner products: Let ri be the i-th row of A, or equivalently the i-th column of A , the T transpose of A. Let bj denote the j-th column of B. Then cij = ri cj is the dot product of i-column of AT and j-th column of B. Sums of outer products: C can also be expressed as outer products of columns of A and rows of B. Pn T Exercise: Show that C = k=1 akbk where ak is the k-th column of A and bk is the k-th row of B (or equivalently the k-column of BT ).
    [Show full text]
  • Geometric-Algebra Adaptive Filters Wilder B
    1 Geometric-Algebra Adaptive Filters Wilder B. Lopes∗, Member, IEEE, Cassio G. Lopesy, Senior Member, IEEE Abstract—This paper presents a new class of adaptive filters, namely Geometric-Algebra Adaptive Filters (GAAFs). They are Faces generated by formulating the underlying minimization problem (a deterministic cost function) from the perspective of Geometric Algebra (GA), a comprehensive mathematical language well- Edges suited for the description of geometric transformations. Also, (directed lines) differently from standard adaptive-filtering theory, Geometric Calculus (the extension of GA to differential calculus) allows Fig. 1. A polyhedron (3-dimensional polytope) can be completely described for applying the same derivation techniques regardless of the by the geometric multiplication of its edges (oriented lines, vectors), which type (subalgebra) of the data, i.e., real, complex numbers, generate the faces and hypersurfaces (in the case of a general n-dimensional quaternions, etc. Relying on those characteristics (among others), polytope). a deterministic quadratic cost function is posed, from which the GAAFs are devised, providing a generalization of regular adaptive filters to subalgebras of GA. From the obtained update rule, it is shown how to recover the following least-mean squares perform calculus with hypercomplex quantities, i.e., elements (LMS) adaptive filter variants: real-entries LMS, complex LMS, that generalize complex numbers for higher dimensions [2]– and quaternions LMS. Mean-square analysis and simulations in [10]. a system identification scenario are provided, showing very good agreement for different levels of measurement noise. GA-based AFs were first introduced in [11], [12], where they were successfully employed to estimate the geometric Index Terms—Adaptive filtering, geometric algebra, quater- transformation (rotation and translation) that aligns a pair of nions.
    [Show full text]
  • New Foundations for Geometric Algebra1
    Text published in the electronic journal Clifford Analysis, Clifford Algebras and their Applications vol. 2, No. 3 (2013) pp. 193-211 New foundations for geometric algebra1 Ramon González Calvet Institut Pere Calders, Campus Universitat Autònoma de Barcelona, 08193 Cerdanyola del Vallès, Spain E-mail : [email protected] Abstract. New foundations for geometric algebra are proposed based upon the existing isomorphisms between geometric and matrix algebras. Each geometric algebra always has a faithful real matrix representation with a periodicity of 8. On the other hand, each matrix algebra is always embedded in a geometric algebra of a convenient dimension. The geometric product is also isomorphic to the matrix product, and many vector transformations such as rotations, axial symmetries and Lorentz transformations can be written in a form isomorphic to a similarity transformation of matrices. We collect the idea Dirac applied to develop the relativistic electron equation when he took a basis of matrices for the geometric algebra instead of a basis of geometric vectors. Of course, this way of understanding the geometric algebra requires new definitions: the geometric vector space is defined as the algebraic subspace that generates the rest of the matrix algebra by addition and multiplication; isometries are simply defined as the similarity transformations of matrices as shown above, and finally the norm of any element of the geometric algebra is defined as the nth root of the determinant of its representative matrix of order n. The main idea of this proposal is an arithmetic point of view consisting of reversing the roles of matrix and geometric algebras in the sense that geometric algebra is a way of accessing, working and understanding the most fundamental conception of matrix algebra as the algebra of transformations of multiple quantities.
    [Show full text]
  • Università Degli Studi Di Trieste a Gentle Introduction to Clifford Algebra
    Università degli Studi di Trieste Dipartimento di Fisica Corso di Studi in Fisica Tesi di Laurea Triennale A Gentle Introduction to Clifford Algebra Laureando: Relatore: Daniele Ceravolo prof. Marco Budinich ANNO ACCADEMICO 2015–2016 Contents 1 Introduction 3 1.1 Brief Historical Sketch . 4 2 Heuristic Development of Clifford Algebra 9 2.1 Geometric Product . 9 2.2 Bivectors . 10 2.3 Grading and Blade . 11 2.4 Multivector Algebra . 13 2.4.1 Pseudoscalar and Hodge Duality . 14 2.4.2 Basis and Reciprocal Frames . 14 2.5 Clifford Algebra of the Plane . 15 2.5.1 Relation with Complex Numbers . 16 2.6 Clifford Algebra of Space . 17 2.6.1 Pauli algebra . 18 2.6.2 Relation with Quaternions . 19 2.7 Reflections . 19 2.7.1 Cross Product . 21 2.8 Rotations . 21 2.9 Differentiation . 23 2.9.1 Multivectorial Derivative . 24 2.9.2 Spacetime Derivative . 25 3 Spacetime Algebra 27 3.1 Spacetime Bivectors and Pseudoscalar . 28 3.2 Spacetime Frames . 28 3.3 Relative Vectors . 29 3.4 Even Subalgebra . 29 3.5 Relative Velocity . 30 3.6 Momentum and Wave Vectors . 31 3.7 Lorentz Transformations . 32 3.7.1 Addition of Velocities . 34 1 2 CONTENTS 3.7.2 The Lorentz Group . 34 3.8 Relativistic Visualization . 36 4 Electromagnetism in Clifford Algebra 39 4.1 The Vector Potential . 40 4.2 Electromagnetic Field Strength . 41 4.3 Free Fields . 44 5 Conclusions 47 5.1 Acknowledgements . 48 Chapter 1 Introduction The aim of this thesis is to show how an approach to classical and relativistic physics based on Clifford algebras can shed light on some hidden geometric meanings in our models.
    [Show full text]
  • Coordinatization
    MATH 355 Supplemental Notes Coordinatization Coordinatization In R3, we have the standard basis i, j and k. When we write a vector in coordinate form, say 3 v 2 , (1) “ »´ fi 5 — ffi – fl it is understood as v 3i 2j 5k. “ ´ ` The numbers 3, 2 and 5 are the coordinates of v relative to the standard basis ⇠ i, j, k . It has p´ q “p q always been understood that a coordinate representation such as that in (1) is with respect to the ordered basis ⇠. A little thought reveals that it need not be so. One could have chosen the same basis elements in a di↵erent order, as in the basis ⇠ i, k, j . We employ notation indicating the 1 “p q coordinates are with respect to the di↵erent basis ⇠1: 3 v 5 , to mean that v 3i 5k 2j, r s⇠1 “ » fi “ ` ´ 2 —´ ffi – fl reflecting the order in which the basis elements fall in ⇠1. Of course, one could employ similar notation even when the coordinates are expressed in terms of the standard basis, writing v for r s⇠ (1), but whenever we have coordinatization with respect to the standard basis of Rn in mind, we will consider the wrapper to be optional. r¨s⇠ Of course, there are many non-standard bases of Rn. In fact, any linearly independent collection of n vectors in Rn provides a basis. Say we take 1 1 1 4 ´ ´ » 0fi » 1fi » 1fi » 1fi u , u u u ´ . 1 “ 2 “ 3 “ 4 “ — 3ffi — 1ffi — 0ffi — 2ffi — ffi —´ ffi — ffi — ffi — 0ffi — 4ffi — 2ffi — 1ffi — ffi — ffi — ffi —´ ffi – fl – fl – fl – fl As per the discussion above, these vectors are being expressed relative to the standard basis of R4.
    [Show full text]
  • MATH 304 Linear Algebra Lecture 14: Basis and Coordinates. Change of Basis
    MATH 304 Linear Algebra Lecture 14: Basis and coordinates. Change of basis. Linear transformations. Basis and dimension Definition. Let V be a vector space. A linearly independent spanning set for V is called a basis. Theorem Any vector space V has a basis. If V has a finite basis, then all bases for V are finite and have the same number of elements (called the dimension of V ). Example. Vectors e1 = (1, 0, 0,..., 0, 0), e2 = (0, 1, 0,..., 0, 0),. , en = (0, 0, 0,..., 0, 1) form a basis for Rn (called standard) since (x1, x2,..., xn) = x1e1 + x2e2 + ··· + xnen. Basis and coordinates If {v1, v2,..., vn} is a basis for a vector space V , then any vector v ∈ V has a unique representation v = x1v1 + x2v2 + ··· + xnvn, where xi ∈ R. The coefficients x1, x2,..., xn are called the coordinates of v with respect to the ordered basis v1, v2,..., vn. The mapping vector v 7→ its coordinates (x1, x2,..., xn) is a one-to-one correspondence between V and Rn. This correspondence respects linear operations in V and in Rn. Examples. • Coordinates of a vector n v = (x1, x2,..., xn) ∈ R relative to the standard basis e1 = (1, 0,..., 0, 0), e2 = (0, 1,..., 0, 0),. , en = (0, 0,..., 0, 1) are (x1, x2,..., xn). a b • Coordinates of a matrix ∈ M2,2(R) c d 1 0 0 0 0 1 relative to the basis , , , 0 0 1 0 0 0 0 0 are (a, c, b, d). 0 1 • Coordinates of a polynomial n−1 p(x) = a0 + a1x + ··· + an−1x ∈Pn relative to 2 n−1 the basis 1, x, x ,..., x are (a0, a1,..., an−1).
    [Show full text]
  • A Guided Tour to the Plane-Based Geometric Algebra PGA
    A Guided Tour to the Plane-Based Geometric Algebra PGA Leo Dorst University of Amsterdam Version 1.15{ July 6, 2020 Planes are the primitive elements for the constructions of objects and oper- ators in Euclidean geometry. Triangulated meshes are built from them, and reflections in multiple planes are a mathematically pure way to construct Euclidean motions. A geometric algebra based on planes is therefore a natural choice to unify objects and operators for Euclidean geometry. The usual claims of `com- pleteness' of the GA approach leads us to hope that it might contain, in a single framework, all representations ever designed for Euclidean geometry - including normal vectors, directions as points at infinity, Pl¨ucker coordinates for lines, quaternions as 3D rotations around the origin, and dual quaternions for rigid body motions; and even spinors. This text provides a guided tour to this algebra of planes PGA. It indeed shows how all such computationally efficient methods are incorporated and related. We will see how the PGA elements naturally group into blocks of four coordinates in an implementation, and how this more complete under- standing of the embedding suggests some handy choices to avoid extraneous computations. In the unified PGA framework, one never switches between efficient representations for subtasks, and this obviously saves any time spent on data conversions. Relative to other treatments of PGA, this text is rather light on the mathematics. Where you see careful derivations, they involve the aspects of orientation and magnitude. These features have been neglected by authors focussing on the mathematical beauty of the projective nature of the algebra.
    [Show full text]
  • Vectors, Change of Basis and Matrix Representation: Onto-Semiotic Approach in the Analysis of Creating Meaning
    International Journal of Mathematical Education in Science and Technology ISSN: 0020-739X (Print) 1464-5211 (Online) Journal homepage: http://www.tandfonline.com/loi/tmes20 Vectors, change of basis and matrix representation: onto-semiotic approach in the analysis of creating meaning Mariana Montiel , Miguel R. Wilhelmi , Draga Vidakovic & Iwan Elstak To cite this article: Mariana Montiel , Miguel R. Wilhelmi , Draga Vidakovic & Iwan Elstak (2012) Vectors, change of basis and matrix representation: onto-semiotic approach in the analysis of creating meaning, International Journal of Mathematical Education in Science and Technology, 43:1, 11-32, DOI: 10.1080/0020739X.2011.582173 To link to this article: http://dx.doi.org/10.1080/0020739X.2011.582173 Published online: 01 Aug 2011. Submit your article to this journal Article views: 174 View related articles Full Terms & Conditions of access and use can be found at http://www.tandfonline.com/action/journalInformation?journalCode=tmes20 Download by: [Universidad Pública de Navarra - Biblioteca] Date: 22 January 2017, At: 10:06 International Journal of Mathematical Education in Science and Technology, Vol. 43, No. 1, 15 January 2012, 11–32 Vectors, change of basis and matrix representation: onto-semiotic approach in the analysis of creating meaning Mariana Montiela*, Miguel R. Wilhelmib, Draga Vidakovica and Iwan Elstaka aDepartment of Mathematics and Statistics, Georgia State University, Atlanta, GA, USA; bDepartment of Mathematics, Public University of Navarra, Pamplona 31006, Spain (Received 26 July 2010) In a previous study, the onto-semiotic approach was employed to analyse the mathematical notion of different coordinate systems, as well as some situations and university students’ actions related to these coordinate systems in the context of multivariate calculus.
    [Show full text]
  • Orthonormal Bases
    Math 108B Professor: Padraic Bartlett Lecture 3: Orthonormal Bases Week 3 UCSB 2014 In our last class, we introduced the concept of \changing bases," and talked about writ- ing vectors and linear transformations in other bases. In the homework and in class, we saw that in several situations this idea of \changing basis" could make a linear transformation much easier to work with; in several cases, we saw that linear transformations under a certain basis would become diagonal, which made tasks like raising them to large powers far easier than these problems would be in the standard basis. But how do we find these \nice" bases? What does it mean for a basis to be\nice?" In this set of lectures, we will study one potential answer to this question: the concept of an orthonormal basis. 1 Orthogonality To start, we should define the notion of orthogonality. First, recall/remember the defini- tion of the dot product: n Definition. Take two vectors (x1; : : : xn); (y1; : : : yn) 2 R . Their dot product is simply the sum x1y1 + x2y2 + : : : xnyn: Many of you have seen an alternate, geometric definition of the dot product: n Definition. Take two vectors (x1; : : : xn); (y1; : : : yn) 2 R . Their dot product is the product jj~xjj · jj~yjj cos(θ); where θ is the angle between ~x and ~y, and jj~xjj denotes the length of the vector ~x, i.e. the distance from (x1; : : : xn) to (0;::: 0). These two definitions are equivalent: 3 Theorem. Let ~x = (x1; x2; x3); ~y = (y1; y2; y3) be a pair of vectors in R .
    [Show full text]
  • Geometric Algebra 4
    Geometric Algebra 4. Algebraic Foundations and 4D Dr Chris Doran ARM Research L4 S2 Axioms Elements of a geometric algebra are Multivectors can be classified by grade called multivectors Grade-0 terms are real scalars Grading is a projection operation Space is linear over the scalars. All simple and natural L4 S3 Axioms The grade-1 elements of a geometric The antisymmetric produce of r vectors algebra are called vectors results in a grade-r blade Call this the outer product So we define Sum over all permutations with epsilon +1 for even and -1 for odd L4 S4 Simplifying result Given a set of linearly-independent vectors We can find a set of anti-commuting vectors such that These vectors all anti-commute Symmetric matrix Define The magnitude of the product is also correct L4 S5 Decomposing products Make repeated use of Define the inner product of a vector and a bivector L4 S6 General result Grade r-1 Over-check means this term is missing Define the inner product of a vector Remaining term is the outer product and a grade-r term Can prove this is the same as earlier definition of the outer product L4 S7 General product Extend dot and wedge symbols for homogenous multivectors The definition of the outer product is consistent with the earlier definition (requires some proof). This version allows a quick proof of associativity: L4 S8 Reverse, scalar product and commutator The reverse, sometimes written with a dagger Useful sequence Write the scalar product as Occasionally use the commutator product Useful property is that the commutator Scalar product is symmetric with a bivector B preserves grade L4 S9 Rotations Combination of rotations Suppose we now rotate a blade So the product rotor is So the blade rotates as Rotors form a group Fermions? Take a rotated vector through a further rotation The rotor transformation law is Now take the rotor on an excursion through 360 degrees.
    [Show full text]
  • Notes on Change of Bases Northwestern University, Summer 2014
    Notes on Change of Bases Northwestern University, Summer 2014 Let V be a finite-dimensional vector space over a field F, and let T be a linear operator on V . Given a basis (v1; : : : ; vn) of V , we've seen how we can define a matrix which encodes all the information about T as follows. For each i, we can write T vi = a1iv1 + ··· + anivn 2 for a unique choice of scalars a1i; : : : ; ani 2 F. In total, we then have n scalars aij which we put into an n × n matrix called the matrix of T relative to (v1; : : : ; vn): 0 1 a11 ··· a1n B . .. C M(T )v := @ . A 2 Mn;n(F): an1 ··· ann In the notation M(T )v, the v showing up in the subscript emphasizes that we're taking this matrix relative to the specific bases consisting of v's. Given any vector u 2 V , we can also write u = b1v1 + ··· + bnvn for a unique choice of scalars b1; : : : ; bn 2 F, and we define the coordinate vector of u relative to (v1; : : : ; vn) as 0 1 b1 B . C n M(u)v := @ . A 2 F : bn In particular, the columns of M(T )v are the coordinates vectors of the T vi. Then the point of the matrix M(T )v is that the coordinate vector of T u is given by M(T u)v = M(T )vM(u)v; so that from the matrix of T and the coordinate vectors of elements of V , we can in fact reconstruct T itself.
    [Show full text]