Geometric-Algebra Adaptive Filters Wilder B

Total Page:16

File Type:pdf, Size:1020Kb

Geometric-Algebra Adaptive Filters Wilder B 1 Geometric-Algebra Adaptive Filters Wilder B. Lopes∗, Member, IEEE, Cassio G. Lopesy, Senior Member, IEEE Abstract—This paper presents a new class of adaptive filters, namely Geometric-Algebra Adaptive Filters (GAAFs). They are Faces generated by formulating the underlying minimization problem (a deterministic cost function) from the perspective of Geometric Algebra (GA), a comprehensive mathematical language well- Edges suited for the description of geometric transformations. Also, (directed lines) differently from standard adaptive-filtering theory, Geometric Calculus (the extension of GA to differential calculus) allows Fig. 1. A polyhedron (3-dimensional polytope) can be completely described for applying the same derivation techniques regardless of the by the geometric multiplication of its edges (oriented lines, vectors), which type (subalgebra) of the data, i.e., real, complex numbers, generate the faces and hypersurfaces (in the case of a general n-dimensional quaternions, etc. Relying on those characteristics (among others), polytope). a deterministic quadratic cost function is posed, from which the GAAFs are devised, providing a generalization of regular adaptive filters to subalgebras of GA. From the obtained update rule, it is shown how to recover the following least-mean squares perform calculus with hypercomplex quantities, i.e., elements (LMS) adaptive filter variants: real-entries LMS, complex LMS, that generalize complex numbers for higher dimensions [2]– and quaternions LMS. Mean-square analysis and simulations in [10]. a system identification scenario are provided, showing very good agreement for different levels of measurement noise. GA-based AFs were first introduced in [11], [12], where they were successfully employed to estimate the geometric Index Terms—Adaptive filtering, geometric algebra, quater- transformation (rotation and translation) that aligns a pair of nions. 3D point clouds (a recurrent problem in computer vision), while keeping a low computational footprint. This work takes a I. INTRODUCTION step further: it uses GA and GC to generate a new class of AFs DAPTIVE filters (AFs), usually described via linear able to naturally estimate hypersurfaces, hypervolumes, and A algebra (LA) and standard vector calculus [1], can be elements of greater dimensions (multivectors), generalizing interpreted as instances for geometric estimation, since the regular AFs from the literature. Filters like the regular least- weight vector to be estimated represents a directed line in an mean squares (LMS) – real entries, the Complex LMS (CLMS) underlying vector space, i.e, a n-dimensional vector encodes – complex entries [13], and the Quaternion LMS (QLMS) – the length and direction of an edge in a n-dimensional polytope quaternion entries [14]–[16], are recovered as special cases of (see Fig. 1). However, to estimate the polytope as a whole the more comprehensive GA-LMS introduced herein. (edges, faces, and so on), a regular AF designed in light of The text is organized as follows. Section II covers the LA might not be very helpful. transition from linear to geometric algebra, presenting the Indeed, LA has limitations regarding the representation of fundamentals of GA and GC. In Section III, the standard geometric structures [2, p. 20]. Take for instance the matrix quadratic cost function, usually adopted in adaptive filtering, product between two vectors: it always results in either a scalar is recovered as a particular case of a more comprehensive or a matrix (2-dimensional array of numbers). Thus, one may quadratic cost function that can only be written using the wonder if it is possible to construct a new kind of product geometric product. In Section IV the gradient of that cost that takes two vectors (directed lines or edges) and returns function is calculated via GC techniques and the GA-LMS arXiv:1608.03450v2 [math.OC] 21 Aug 2018 a hypersurface (the face of an n-dimensional polytope); or is formulated. Section V provides a mean-square analysis takes a vector and a hypersurface and returns another polytope. (steady state) with the support of the energy conservation Similar ideas have been present since the advent of algebra, in relations [1]. Experiments are shown in Section VI to assess an attempt to establish a deep connection with geometry [3], the performance of the GAAFs against the theoretical analysis [4]. in a system-identification scenario. Finally, conclusion remarks The geometric product, which is the product operation of are presented in Section VII. GA [5], [6], captures the aforementioned idea. It allows one Remarks about notation: Table I summarizes the notation to map a set of vectors not only onto scalars, but also onto convention. While those are deterministic quantities, their hypersurfaces and n-dimensional polytopes. The use of GA in- boldface versions represent random quantities. The name array creases the portfolio of geometric shapes and transformations of multivectors was chosen to avoid confusion with vectors in one can represent. Also, its extension to calculus, namely geo- Rn, which in this text have the usual meaning of collection metric calculus (GC), allows for a clear and compact way to of real numbers (row or column). In this sense, an array of multivectors can be interpreted as a “vector” that allows ∗R&D Dept. at UCit (ucit.fr) and OpenGA.org, [email protected]. yDept. of Electronic Systems Engineering, University of Sao Paulo, Brazil, cas- for hypercomplex entries (numbers constructed by adding [email protected]. “imaginary units” to real numbers [17]). 2 TABLE I that two nonparallel directed segments determine a parallel- SUMMARY OF NOTATION. ogram, a notion which can not be described by the inner Symbol Definition product. The multiplication a^b results in an oriented surface a Arrays of multivectors, vectors and scalars. or bivector (see Fig. 2a). Such a surface can be interpreted A General multivector or matrix. as the parallelogram (hyperplane) generated when vector a is r Rotor (type of multivector). Time-varying scalar, vector (or array of multivectors), swept on the direction determined by vector b. Alternatively, a(i); a ;A(i) i and general multivector (or matrix), respectively. the outer product can be defined as a function of the angle θ between a and b a ^ b = Ia;bjajjbjsinθ; (3) II. FROM LINEAR TO GEOMETRIC ALGEBRA 1 where Ia;b is the unit bivector that defines the orientation of To start transitioning from LA to GA, one needs to recall the the hyperplane a ^ b [2, p.66]. Note that in the particular case definition of an algebra over the reals [6], [9], [18]: a vector of 3D Euclidean space, (3) is reduced to the cross product space V over the field R, equipped with a bilinear map V × a × b = pjajjbjsinθ, where p is the unit vector normal to the V ! V denoted by ◦ (the product operation of the algebra), plane containing fa; bg. From Fig. 2a it can be concluded that is said to be an algebra over R if the following relations hold the outer product is noncommutative, i.e., a ^ b = −b ^ a: 8fa; b; cg 2 V and fα; βg 2 R, the orientation of the surface generated by sweeping a along (a + b) ◦ c = a ◦ c + b ◦ c (Left distributivity); b is opposite to the orientation of the surface generated by c ◦ (a + b) = c ◦ a + c ◦ b (Right distributivity); sweeping b along a. 2 (αa) ◦ (βb) = (αβ)(a ◦ b) (Compatibility with scalars). Finally, the geometric product of vectors a and b is denoted (1) by their juxtaposition ab and defined as Linear (matrix) algebra, utilized to describe adaptive fil- ab , a · b + a ^ b; (4) tering theory, is constructed from the definition above. The in terms of the inner (·) and outer (^) products [6, Sec. 2:2]. elements of this algebra are matrices and vectors, which Note that due to a ^ b = −(b ^ a) the geometric product is multiplied among themselves via the matrix product generate noncommutative, resulting in ab = −ba, and it is associative, new matrices and vectors. Additionally, to express the notion a(bc) = (ab)c, fa; b; cg 2 Rn. of vector length and angle between vectors, LA adopts the In linear algebra the fundamental elements are matri- bilinear form V × V ! R, i.e., inner product, which returns ces/vectors. In a similar way, in GA the basic elements are a real number as a result of the multiplication between two the so-called multivectors (Clifford numbers). The structure vectors in V (one says that V is a normed vector space) [18, of a multivector can be seen as a generalization of complex p. 180]. numbers and quaternions for higher dimensions. For instance, Geometric (Clifford) Algebra derives from (1), however a complex number α+jβ has a scalar part α combined with an with a different product operation. Such product, called geo- imaginary part jβ; quaternions like α+iβ1 +jβ2 +kβ3 expand metric product, is what allows for GA to be a mathematical that by adding two extra imaginary-valued parts. Multivectors language that unifies different algebraic systems trying to generalize this structure, which contains one scalar part and express geometric relations/transformations, e.g., rotation and several other parts named grades. Thus, a general multivector translation. The following systems are examples of algebras A has the form integrated into GA theory: vector/matrix algebra, complex X A = hAi0 + hAi1 + hAi2 + ··· = hAig, (5) numbers, and quaternions [2], [5], [6], [9]. Such unifying g quality is put into use in this work to expand the capabilities which is comprised of its g-grades (or g-vectors) h·ig, e.g., g = of AFs. 0 (scalars), g = 1 (vectors), g = 2 (bivectors, generated via The fundamentals of GA are provided in the sequel. For an the geometric multiplication of two vectors), g = 3 (trivectors, in-depth discussion, the reader is referred to [2]–[10], [19]– generated via the geometric multiplication of three vectors), [23].
Recommended publications
  • Vectors and Beyond: Geometric Algebra and Its Philosophical
    dialectica Vol. 63, N° 4 (2010), pp. 381–395 DOI: 10.1111/j.1746-8361.2009.01214.x Vectors and Beyond: Geometric Algebra and its Philosophical Significancedltc_1214 381..396 Peter Simons† In mathematics, like in everything else, it is the Darwinian struggle for life of ideas that leads to the survival of the concepts which we actually use, often believed to have come to us fully armed with goodness from some mysterious Platonic repository of truths. Simon Altmann 1. Introduction The purpose of this paper is to draw the attention of philosophers and others interested in the applicability of mathematics to a quiet revolution that is taking place around the theory of vectors and their application. It is not that new math- ematics is being invented – on the contrary, the basic concepts are over a century old – but rather that this old theory, having languished for many decades as a quaint backwater, is being rediscovered and properly applied for the first time. The philosophical importance of this quiet revolution is not that new applications for old mathematics are being found. That presumably happens much of the time. Rather it is that this new range of applications affords us a novel insight into the reasons why vectors and their mathematical kin find application at all in the real world. Indirectly, it tells us something general but interesting about the nature of the spatiotemporal world we all inhabit, and that is of philosophical significance. Quite what this significance amounts to is not yet clear. I should stress that nothing here is original: the history is quite accessible from several sources, and the mathematics is commonplace to those who know it.
    [Show full text]
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Abstract Tensor Systems and Diagrammatic Representations
    Abstract tensor systems and diagrammatic representations J¯anisLazovskis September 28, 2012 Abstract The diagrammatic tensor calculus used by Roger Penrose (most notably in [7]) is introduced without a solid mathematical grounding. We will attempt to derive the tools of such a system, but in a broader setting. We show that Penrose's work comes from the diagrammisation of the symmetric algebra. Lie algebra representations and their extensions to knot theory are also discussed. Contents 1 Abstract tensors and derived structures 2 1.1 Abstract tensor notation . 2 1.2 Some basic operations . 3 1.3 Tensor diagrams . 3 2 A diagrammised abstract tensor system 6 2.1 Generation . 6 2.2 Tensor concepts . 9 3 Representations of algebras 11 3.1 The symmetric algebra . 12 3.2 Lie algebras . 13 3.3 The tensor algebra T(g)....................................... 16 3.4 The symmetric Lie algebra S(g)................................... 17 3.5 The universal enveloping algebra U(g) ............................... 18 3.6 The metrized Lie algebra . 20 3.6.1 Diagrammisation with a bilinear form . 20 3.6.2 Diagrammisation with a symmetric bilinear form . 24 3.6.3 Diagrammisation with a symmetric bilinear form and an orthonormal basis . 24 3.6.4 Diagrammisation under ad-invariance . 29 3.7 The universal enveloping algebra U(g) for a metrized Lie algebra g . 30 4 Ensuing connections 32 A Appendix 35 Note: This work relies heavily upon the text of Chapter 12 of a draft of \An Introduction to Quantum and Vassiliev Invariants of Knots," by David M.R. Jackson and Iain Moffatt, a yet-unpublished book at the time of writing.
    [Show full text]
  • What's in a Name? the Matrix As an Introduction to Mathematics
    St. John Fisher College Fisher Digital Publications Mathematical and Computing Sciences Faculty/Staff Publications Mathematical and Computing Sciences 9-2008 What's in a Name? The Matrix as an Introduction to Mathematics Kris H. Green St. John Fisher College, [email protected] Follow this and additional works at: https://fisherpub.sjfc.edu/math_facpub Part of the Mathematics Commons How has open access to Fisher Digital Publications benefited ou?y Publication Information Green, Kris H. (2008). "What's in a Name? The Matrix as an Introduction to Mathematics." Math Horizons 16.1, 18-21. Please note that the Publication Information provides general citation information and may not be appropriate for your discipline. To receive help in creating a citation based on your discipline, please visit http://libguides.sjfc.edu/citations. This document is posted at https://fisherpub.sjfc.edu/math_facpub/12 and is brought to you for free and open access by Fisher Digital Publications at St. John Fisher College. For more information, please contact [email protected]. What's in a Name? The Matrix as an Introduction to Mathematics Abstract In lieu of an abstract, here is the article's first paragraph: In my classes on the nature of scientific thought, I have often used the movie The Matrix to illustrate the nature of evidence and how it shapes the reality we perceive (or think we perceive). As a mathematician, I usually field questions elatedr to the movie whenever the subject of linear algebra arises, since this field is the study of matrices and their properties. So it is natural to ask, why does the movie title reference a mathematical object? Disciplines Mathematics Comments Article copyright 2008 by Math Horizons.
    [Show full text]
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Recent Books on Vector Analysis. 463
    1921.] RECENT BOOKS ON VECTOR ANALYSIS. 463 RECENT BOOKS ON VECTOR ANALYSIS. Einführung in die Vehtoranalysis, mit Anwendungen auf die mathematische Physik. By R. Gans. 4th edition. Leip­ zig and Berlin, Teubner, 1921. 118 pp. Elements of Vector Algebra. By L. Silberstein. New York, Longmans, Green, and Company, 1919. 42 pp. Vehtor analysis. By C. Runge. Vol. I. Die Vektoranalysis des dreidimensionalen Raumes. Leipzig, Hirzel, 1919. viii + 195 pp. Précis de Calcul géométrique. By R. Leveugle. Preface by H. Fehr. Paris, Gauthier-Villars, 1920. lvi + 400 pp. The first of these books is for the most part a reprint of the third edition, which has been reviewed in this BULLETIN (vol. 21 (1915), p. 360). Some small condensation of results has been accomplished by making deductions by purely vector methods. The author still clings to the rather common idea that a vector is a certain triple on a certain set of coordinate axes, so that his development is rather that of a shorthand than of a study of expressions which are not dependent upon axes. The second book is a quite brief exposition of the bare fundamentals of vector algebra, the notation being that of Heaviside, which Silberstein has used before. It is designed primarily to satisfy the needs of those studying geometric optics. However the author goes so far as to mention a linear vector operator and the notion of dyad. The exposition is very simple and could be followed by high school students who had had trigonometry. The third of the books mentioned is an elementary exposition of vectors from the standpoint of Grassmann.
    [Show full text]
  • Appendix a Spinors in Four Dimensions
    Appendix A Spinors in Four Dimensions In this appendix we collect the conventions used for spinors in both Minkowski and Euclidean spaces. In Minkowski space the flat metric has the 0 1 2 3 form ηµν = diag(−1, 1, 1, 1), and the coordinates are labelled (x ,x , x , x ). The analytic continuation into Euclidean space is madethrough the replace- ment x0 = ix4 (and in momentum space, p0 = −ip4) the coordinates in this case being labelled (x1,x2, x3, x4). The Lorentz group in four dimensions, SO(3, 1), is not simply connected and therefore, strictly speaking, has no spinorial representations. To deal with these types of representations one must consider its double covering, the spin group Spin(3, 1), which is isomorphic to SL(2, C). The group SL(2, C) pos- sesses a natural complex two-dimensional representation. Let us denote this representation by S andlet us consider an element ψ ∈ S with components ψα =(ψ1,ψ2) relative to some basis. The action of an element M ∈ SL(2, C) is β (Mψ)α = Mα ψβ. (A.1) This is not the only action of SL(2, C) which one could choose. Instead of M we could have used its complex conjugate M, its inverse transpose (M T)−1,or its inverse adjoint (M †)−1. All of them satisfy the same group multiplication law. These choices would correspond to the complex conjugate representation S, the dual representation S,and the dual complex conjugate representation S. We will use the following conventions for elements of these representations: α α˙ ψα ∈ S, ψα˙ ∈ S, ψ ∈ S, ψ ∈ S.
    [Show full text]
  • Handout 9 More Matrix Properties; the Transpose
    Handout 9 More matrix properties; the transpose Square matrix properties These properties only apply to a square matrix, i.e. n £ n. ² The leading diagonal is the diagonal line consisting of the entries a11, a22, a33, . ann. ² A diagonal matrix has zeros everywhere except the leading diagonal. ² The identity matrix I has zeros o® the leading diagonal, and 1 for each entry on the diagonal. It is a special case of a diagonal matrix, and A I = I A = A for any n £ n matrix A. ² An upper triangular matrix has all its non-zero entries on or above the leading diagonal. ² A lower triangular matrix has all its non-zero entries on or below the leading diagonal. ² A symmetric matrix has the same entries below and above the diagonal: aij = aji for any values of i and j between 1 and n. ² An antisymmetric or skew-symmetric matrix has the opposite entries below and above the diagonal: aij = ¡aji for any values of i and j between 1 and n. This automatically means the digaonal entries must all be zero. Transpose To transpose a matrix, we reect it across the line given by the leading diagonal a11, a22 etc. In general the result is a di®erent shape to the original matrix: a11 a21 a11 a12 a13 > > A = A = 0 a12 a22 1 [A ]ij = A : µ a21 a22 a23 ¶ ji a13 a23 @ A > ² If A is m £ n then A is n £ m. > ² The transpose of a symmetric matrix is itself: A = A (recalling that only square matrices can be symmetric).
    [Show full text]
  • Multilinear Algebra and Applications July 15, 2014
    Multilinear Algebra and Applications July 15, 2014. Contents Chapter 1. Introduction 1 Chapter 2. Review of Linear Algebra 5 2.1. Vector Spaces and Subspaces 5 2.2. Bases 7 2.3. The Einstein convention 10 2.3.1. Change of bases, revisited 12 2.3.2. The Kronecker delta symbol 13 2.4. Linear Transformations 14 2.4.1. Similar matrices 18 2.5. Eigenbases 19 Chapter 3. Multilinear Forms 23 3.1. Linear Forms 23 3.1.1. Definition, Examples, Dual and Dual Basis 23 3.1.2. Transformation of Linear Forms under a Change of Basis 26 3.2. Bilinear Forms 30 3.2.1. Definition, Examples and Basis 30 3.2.2. Tensor product of two linear forms on V 32 3.2.3. Transformation of Bilinear Forms under a Change of Basis 33 3.3. Multilinear forms 34 3.4. Examples 35 3.4.1. A Bilinear Form 35 3.4.2. A Trilinear Form 36 3.5. Basic Operation on Multilinear Forms 37 Chapter 4. Inner Products 39 4.1. Definitions and First Properties 39 4.1.1. Correspondence Between Inner Products and Symmetric Positive Definite Matrices 40 4.1.1.1. From Inner Products to Symmetric Positive Definite Matrices 42 4.1.1.2. From Symmetric Positive Definite Matrices to Inner Products 42 4.1.2. Orthonormal Basis 42 4.2. Reciprocal Basis 46 4.2.1. Properties of Reciprocal Bases 48 4.2.2. Change of basis from a basis to its reciprocal basis g 50 B B III IV CONTENTS 4.2.3.
    [Show full text]
  • A Clifford Dyadic Superfield from Bilateral Interactions of Geometric Multispin Dirac Theory
    A CLIFFORD DYADIC SUPERFIELD FROM BILATERAL INTERACTIONS OF GEOMETRIC MULTISPIN DIRAC THEORY WILLIAM M. PEZZAGLIA JR. Department of Physia, Santa Clam University Santa Clam, CA 95053, U.S.A., [email protected] and ALFRED W. DIFFER Department of Phyaia, American River College Sacramento, CA 958i1, U.S.A. (Received: November 5, 1993) Abstract. Multivector quantum mechanics utilizes wavefunctions which a.re Clifford ag­ gregates (e.g. sum of scalar, vector, bivector). This is equivalent to multispinors con­ structed of Dirac matrices, with the representation independent form of the generators geometrically interpreted as the basis vectors of spacetime. Multiple generations of par­ ticles appear as left ideals of the algebra, coupled only by now-allowed right-side applied (dextral) operations. A generalized bilateral (two-sided operation) coupling is propoeed which includes the above mentioned dextrad field, and the spin-gauge interaction as partic­ ular cases. This leads to a new principle of poly-dimensional covariance, in which physical laws are invariant under the reshuffling of coordinate geometry. Such a multigeometric su­ perfield equation is proposed, whi~h is sourced by a bilateral current. In order to express the superfield in representation and coordinate free form, we introduce Eddington E-F double-frame numbers. Symmetric tensors can now be represented as 4D "dyads", which actually are elements of a global SD Clifford algebra.. As a restricted example, the dyadic field created by the Greider-Ross multivector current (of a Dirac electron) describes both electromagnetic and Morris-Greider gravitational interactions. Key words: spin-gauge, multivector, clifford, dyadic 1. Introduction Multi vector physics is a grand scheme in which we attempt to describe all ba­ sic physical structure and phenomena by a single geometrically interpretable Algebra.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • Matrices and Tensors
    APPENDIX MATRICES AND TENSORS A.1. INTRODUCTION AND RATIONALE The purpose of this appendix is to present the notation and most of the mathematical tech- niques that are used in the body of the text. The audience is assumed to have been through sev- eral years of college-level mathematics, which included the differential and integral calculus, differential equations, functions of several variables, partial derivatives, and an introduction to linear algebra. Matrices are reviewed briefly, and determinants, vectors, and tensors of order two are described. The application of this linear algebra to material that appears in under- graduate engineering courses on mechanics is illustrated by discussions of concepts like the area and mass moments of inertia, Mohr’s circles, and the vector cross and triple scalar prod- ucts. The notation, as far as possible, will be a matrix notation that is easily entered into exist- ing symbolic computational programs like Maple, Mathematica, Matlab, and Mathcad. The desire to represent the components of three-dimensional fourth-order tensors that appear in anisotropic elasticity as the components of six-dimensional second-order tensors and thus rep- resent these components in matrices of tensor components in six dimensions leads to the non- traditional part of this appendix. This is also one of the nontraditional aspects in the text of the book, but a minor one. This is described in §A.11, along with the rationale for this approach. A.2. DEFINITION OF SQUARE, COLUMN, AND ROW MATRICES An r-by-c matrix, M, is a rectangular array of numbers consisting of r rows and c columns: ¯MM..
    [Show full text]