Transpose & Dot Product Extended Example

Total Page:16

File Type:pdf, Size:1020Kb

Transpose & Dot Product Extended Example Transpose & Dot Product Def: The transpose of an m × n matrix A is the n × m matrix AT whose columns are the rows of A. So: The columns of AT are the rows of A. The rows of AT are the columns of A. 21 43 1 2 3 Example: If A = , then AT = 2 5 : 4 5 6 4 5 3 6 Convention: From now on, vectors v 2 Rn will be regarded as \columns" (i.e.: n × 1 matrices). Therefore, vT is a \row vector" (a 1 × n matrix). Observation: Let v; w 2 Rn. Then vT w = v · w. This is because: 2 3 w1 T . v w = v1 ··· vn 4 . 5 = v1w1 + ··· + vnwn = v · w: wn Where theory is concerned, the key property of transposes is the following: Prop 18.2: Let A be an m × n matrix. Then for x 2 Rn and y 2 Rm: (Ax) · y = x · (AT y): Here, · is the dot product of vectors. Extended Example Let A be a 5 × 3 matrix, so A: R3 ! R5. ◦ N(A) is a subspace of ◦ C(A) is a subspace of The transpose AT is a matrix, so AT : ! ◦ C(AT ) is a subspace of ◦ N(AT ) is a subspace of Observation: Both C(AT ) and N(A) are subspaces of . Might there be a geometric relationship between the two? (No, they're not equal.) Hm... Also: Both N(AT ) and C(A) are subspaces of . Might there be a geometric relationship between the two? (Again, they're not equal.) Hm... Orthogonal Complements Def: Let V ⊂ Rn be a subspace. The orthogonal complement of V is the set ? n V = fx 2 R j x · v = 0 for every v 2 V g: So, V ? consists of the vectors which are orthogonal to every vector in V . Fact: If V ⊂ Rn is a subspace, then V ? ⊂ Rn is a subspace. Examples in R3: ◦ The orthogonal complement of V = f0g is V ? = R3 ◦ The orthogonal complement of V = fz-axisg is V ? = fxy-planeg ◦ The orthogonal complement of V = fxy-planeg is V ? = fz-axisg ◦ The orthogonal complement of V = R3 is V ? = f0g Examples in R4: ◦ The orthogonal complement of V = f0g is V ? = R4 ◦ The orthogonal complement of V = fw-axisg is V ? = fxyz-spaceg ◦ The orthogonal complement of V = fzw-planeg is V ? = fxy-planeg ◦ The orthogonal complement of V = fxyz-spaceg is V ? = fw-axisg ◦ The orthogonal complement of V = R4 is V ? = f0g Prop 19.3-19.4-19.5: Let V ⊂ Rn be a subspace. Then: (a) dim(V ) + dim(V ?) = n (b) (V ?)? = V (c) V \ V ? = f0g (d) V + V ? = Rn. Part (d) means: \Every vector x 2 Rn can be written as a sum x = v + w where v 2 V and w 2 V ?." Also, it turns out that the expression x = v + w is unique: that is, there is only one way to write x as a sum of a vector in V and a vector in V ?. Meaning of C(AT ) and N(AT ) Q: What does C(AT ) mean? Well, the columns of AT are the rows of A. So: C(AT ) = column space of AT = span of columns of AT = span of rows of A: For this reason: We call C(AT ) the row space of A. Q: What does N(AT ) mean? Well: x 2 N(AT ) () AT x = 0 () (AT x)T = 0T () xT A = 0T : So, for an m × n matrix A, we see that: N(AT ) = fx 2 Rm j xT A = 0T g. For this reason: We call N(AT ) the left null space of A. Relationships among the Subspaces Theorem: Let A be an m × n matrix. Then: ◦ C(AT ) = N(A)? ◦ N(AT ) = C(A)? Corollary: Let A be an m × n matrix. Then: ◦ C(A) = N(AT )? ◦ N(A) = C(AT )? Prop 18.3: Let A be an m × n matrix. Then rank(A) = rank(AT ). Motivating Questions for Reading Problem 1: Let b 2 C(A). So, the system of equations Ax = b does have solutions, possibly infinitely many. Q: What is the solution x of Ax = b with kxk the smallest? Problem 2: Let b 2= C(A). So, the system of equations Ax = b does not have any solutions. In other words, Ax − b 6= 0. Q: What is the vector x that minimizes the error kAx−bk? That is, what is the vector x that comes closest to being a solution to Ax = b? Orthogonal Projection Def: Let V ⊂ Rn be a subspace. Then every vector x 2 Rn can be written uniquely as x = v + w; where v 2 V and w 2 V ?: n n The orthogonal projection onto V is the function ProjV : R ! R given by: ProjV (x) = v. (Note that ProjV ? (x) = w.) Prop 20.1: Let V ⊂ Rn be a subspace. Then: ProjV + ProjV ? = In: Of course, we already knew this: We have x = v+w = ProjV (x)+ProjV ? (x). n Formula: Let fv1;:::; vkg be a basis of V ⊂ R . Let A be the n × k matrix 2 3 A = 4v1 ··· vk5: Then: T −1 T ProjV = A(A A) A : (∗) Geometry Observations: Let V ⊂ Rn be a subspace, and x 2 Rn a vector. (1) The distance from x to V is: kProjV ? (x)k = kx − ProjV (x)k. (2) The vector in V that is closest to x is: ProjV (x). Derivation of (∗): Notice ProjV (x) is a vector in V = span(v1;:::; vk) = C(A) = Range(A), and k therefore ProjV (x) = Ay for some vector y 2 R . ? ? T Now notice that x − ProjV (x) = x − Ay is a vector in V = C(A) = N(A ), which means that AT (x − Ay) = 0, which means AT x = AT Ay. Now, it turns out that our matrix AT A is invertible (proof in L20), so we get y = (AT A)−1AT x. T −1 T Thus, ProjV (x) = Ay = A(A A) A x. ♦ Minimum Magnitude Solution Prop 19.6: Let b 2 C(A) (so Ax = b has solutions). Then there exists T exactly one vector x0 2 C(A ) with Ax0 = b. And: Among all solutions of Ax = b, the vector x0 has the smallest length. In other words: There is exactly one vector x0 in the row space of A which solves Ax = b { and this vector is the solution of smallest length. To Find x0: Start with any solution x of Ax = b. Then x0 = ProjC(AT )(x): Least Squares Approximation Idea: Suppose b 2= C(A). So, Ax = b has no solutions, so Ax − b 6= 0. We want to find the vector x∗ which minimizes the error kAx∗ − bk. That is, we want the vector x∗ for which Ax∗ is the closest vector in C(A) to b. In other words, we want the vector x∗ for which Ax∗ − b is orthogonal to C(A). So, Ax∗ − b 2 C(A)? = N(AT ), meaning that AT (Ax∗ − b) = 0, i.e.: AT Ax∗ = AT b: Quadratic Forms (Intro) Given an m × n matrix A, we can regard it as a linear transformation T : Rn ! Rm. In the special case where the matrix A is a symmetric matrix, we can also regard A as defining a \quadratic form": Def: Let A be a symmetric n × n matrix. The quadratic form associated n to A is the function QA : R ! R given by: QA(x) = x · Ax (· is the dot product) 2 3 x1 T . = x Ax = x1 ··· xn A 4 . 5 xn Notice that quadratic forms are not linear transformations! Orthonormal Bases Def: A basis fw1;:::; wkg for a subspace V is an orthonormal basis if: (1) The basis vectors are mutually orthogonal: wi · wj = 0 (for i 6= j); (2) The basis vectors are unit vectors: wi · wi = 1. (i.e.: kwik = 1) Orthonormal bases are nice for (at least) two reasons: (a) It is much easier to find the B-coordinates [v]B of a vector when the basis B is orthonormal; (b) It is much easier to find the projection matrix onto a subspace V when we have an orthonormal basis for V . n Prop: Let fw1;:::; wkg be an orthonormal basis for a subspace V ⊂ R . (a) Every vector v 2 V can be written v = (v · w1)w1 + ··· + (v · wk)wk: (b) For all x 2 Rn: ProjV (x) = (x · w1)w1 + ··· + (x · wk)wk: T (c) Let A be the matrix with columns fw1;:::; wkg. Then A A = Ik, so: T −1 T T ProjV = A(A A) A = AA : Orthogonal Matrices Def: An orthogonal matrix is an invertible matrix C such that C−1 = CT : n Example: Let fv1;:::; vng be an orthonormal basis for R . Then the matrix 2 3 C = 4v1 ··· vn5 is an orthogonal matrix. In fact, every orthogonal matrix C looks like this: the columns of any orthogonal matrix form an orthonormal basis of Rn. Where theory is concerned, the key property of orthogonal matrices is: Prop 22.4: Let C be an orthogonal matrix. Then for v; w 2 Rn: Cv · Cw = v · w: Gram-Schmidt Process Since orthonormal bases have so many nice properties, it would be great if we had a way of actually manufacturing orthonormal bases. That is: n Goal: We are given a basis fv1;:::; vkg for a subspace V ⊂ R . We would like an orthonormal basis fw1;:::; wkg for our subspace V . Notation: We will let V1 = span(v1) V2 = span(v1; v2) . Vk = span(v1;:::; vk) = V: Idea: Build an orthonormal basis for V1, then for V2, :::; up to Vk = V .
Recommended publications
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Determinant Notes
    68 UNIT FIVE DETERMINANTS 5.1 INTRODUCTION In unit one the determinant of a 2× 2 matrix was introduced and used in the evaluation of a cross product. In this chapter we extend the definition of a determinant to any size square matrix. The determinant has a variety of applications. The value of the determinant of a square matrix A can be used to determine whether A is invertible or noninvertible. An explicit formula for A–1 exists that involves the determinant of A. Some systems of linear equations have solutions that can be expressed in terms of determinants. 5.2 DEFINITION OF THE DETERMINANT a11 a12 Recall that in chapter one the determinant of the 2× 2 matrix A = was a21 a22 defined to be the number a11a22 − a12 a21 and that the notation det (A) or A was used to represent the determinant of A. For any given n × n matrix A = a , the notation A [ ij ]n×n ij will be used to denote the (n −1)× (n −1) submatrix obtained from A by deleting the ith row and the jth column of A. The determinant of any size square matrix A = a is [ ij ]n×n defined recursively as follows. Definition of the Determinant Let A = a be an n × n matrix. [ ij ]n×n (1) If n = 1, that is A = [a11], then we define det (A) = a11 . n 1k+ (2) If na>=1, we define det(A) ∑(-1)11kk det(A ) k=1 Example If A = []5 , then by part (1) of the definition of the determinant, det (A) = 5.
    [Show full text]
  • What's in a Name? the Matrix As an Introduction to Mathematics
    St. John Fisher College Fisher Digital Publications Mathematical and Computing Sciences Faculty/Staff Publications Mathematical and Computing Sciences 9-2008 What's in a Name? The Matrix as an Introduction to Mathematics Kris H. Green St. John Fisher College, [email protected] Follow this and additional works at: https://fisherpub.sjfc.edu/math_facpub Part of the Mathematics Commons How has open access to Fisher Digital Publications benefited ou?y Publication Information Green, Kris H. (2008). "What's in a Name? The Matrix as an Introduction to Mathematics." Math Horizons 16.1, 18-21. Please note that the Publication Information provides general citation information and may not be appropriate for your discipline. To receive help in creating a citation based on your discipline, please visit http://libguides.sjfc.edu/citations. This document is posted at https://fisherpub.sjfc.edu/math_facpub/12 and is brought to you for free and open access by Fisher Digital Publications at St. John Fisher College. For more information, please contact [email protected]. What's in a Name? The Matrix as an Introduction to Mathematics Abstract In lieu of an abstract, here is the article's first paragraph: In my classes on the nature of scientific thought, I have often used the movie The Matrix to illustrate the nature of evidence and how it shapes the reality we perceive (or think we perceive). As a mathematician, I usually field questions elatedr to the movie whenever the subject of linear algebra arises, since this field is the study of matrices and their properties. So it is natural to ask, why does the movie title reference a mathematical object? Disciplines Mathematics Comments Article copyright 2008 by Math Horizons.
    [Show full text]
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Appendix a Spinors in Four Dimensions
    Appendix A Spinors in Four Dimensions In this appendix we collect the conventions used for spinors in both Minkowski and Euclidean spaces. In Minkowski space the flat metric has the 0 1 2 3 form ηµν = diag(−1, 1, 1, 1), and the coordinates are labelled (x ,x , x , x ). The analytic continuation into Euclidean space is madethrough the replace- ment x0 = ix4 (and in momentum space, p0 = −ip4) the coordinates in this case being labelled (x1,x2, x3, x4). The Lorentz group in four dimensions, SO(3, 1), is not simply connected and therefore, strictly speaking, has no spinorial representations. To deal with these types of representations one must consider its double covering, the spin group Spin(3, 1), which is isomorphic to SL(2, C). The group SL(2, C) pos- sesses a natural complex two-dimensional representation. Let us denote this representation by S andlet us consider an element ψ ∈ S with components ψα =(ψ1,ψ2) relative to some basis. The action of an element M ∈ SL(2, C) is β (Mψ)α = Mα ψβ. (A.1) This is not the only action of SL(2, C) which one could choose. Instead of M we could have used its complex conjugate M, its inverse transpose (M T)−1,or its inverse adjoint (M †)−1. All of them satisfy the same group multiplication law. These choices would correspond to the complex conjugate representation S, the dual representation S,and the dual complex conjugate representation S. We will use the following conventions for elements of these representations: α α˙ ψα ∈ S, ψα˙ ∈ S, ψ ∈ S, ψ ∈ S.
    [Show full text]
  • Handout 9 More Matrix Properties; the Transpose
    Handout 9 More matrix properties; the transpose Square matrix properties These properties only apply to a square matrix, i.e. n £ n. ² The leading diagonal is the diagonal line consisting of the entries a11, a22, a33, . ann. ² A diagonal matrix has zeros everywhere except the leading diagonal. ² The identity matrix I has zeros o® the leading diagonal, and 1 for each entry on the diagonal. It is a special case of a diagonal matrix, and A I = I A = A for any n £ n matrix A. ² An upper triangular matrix has all its non-zero entries on or above the leading diagonal. ² A lower triangular matrix has all its non-zero entries on or below the leading diagonal. ² A symmetric matrix has the same entries below and above the diagonal: aij = aji for any values of i and j between 1 and n. ² An antisymmetric or skew-symmetric matrix has the opposite entries below and above the diagonal: aij = ¡aji for any values of i and j between 1 and n. This automatically means the digaonal entries must all be zero. Transpose To transpose a matrix, we reect it across the line given by the leading diagonal a11, a22 etc. In general the result is a di®erent shape to the original matrix: a11 a21 a11 a12 a13 > > A = A = 0 a12 a22 1 [A ]ij = A : µ a21 a22 a23 ¶ ji a13 a23 @ A > ² If A is m £ n then A is n £ m. > ² The transpose of a symmetric matrix is itself: A = A (recalling that only square matrices can be symmetric).
    [Show full text]
  • Geometric-Algebra Adaptive Filters Wilder B
    1 Geometric-Algebra Adaptive Filters Wilder B. Lopes∗, Member, IEEE, Cassio G. Lopesy, Senior Member, IEEE Abstract—This paper presents a new class of adaptive filters, namely Geometric-Algebra Adaptive Filters (GAAFs). They are Faces generated by formulating the underlying minimization problem (a deterministic cost function) from the perspective of Geometric Algebra (GA), a comprehensive mathematical language well- Edges suited for the description of geometric transformations. Also, (directed lines) differently from standard adaptive-filtering theory, Geometric Calculus (the extension of GA to differential calculus) allows Fig. 1. A polyhedron (3-dimensional polytope) can be completely described for applying the same derivation techniques regardless of the by the geometric multiplication of its edges (oriented lines, vectors), which type (subalgebra) of the data, i.e., real, complex numbers, generate the faces and hypersurfaces (in the case of a general n-dimensional quaternions, etc. Relying on those characteristics (among others), polytope). a deterministic quadratic cost function is posed, from which the GAAFs are devised, providing a generalization of regular adaptive filters to subalgebras of GA. From the obtained update rule, it is shown how to recover the following least-mean squares perform calculus with hypercomplex quantities, i.e., elements (LMS) adaptive filter variants: real-entries LMS, complex LMS, that generalize complex numbers for higher dimensions [2]– and quaternions LMS. Mean-square analysis and simulations in [10]. a system identification scenario are provided, showing very good agreement for different levels of measurement noise. GA-based AFs were first introduced in [11], [12], where they were successfully employed to estimate the geometric Index Terms—Adaptive filtering, geometric algebra, quater- transformation (rotation and translation) that aligns a pair of nions.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • Matrices and Tensors
    APPENDIX MATRICES AND TENSORS A.1. INTRODUCTION AND RATIONALE The purpose of this appendix is to present the notation and most of the mathematical tech- niques that are used in the body of the text. The audience is assumed to have been through sev- eral years of college-level mathematics, which included the differential and integral calculus, differential equations, functions of several variables, partial derivatives, and an introduction to linear algebra. Matrices are reviewed briefly, and determinants, vectors, and tensors of order two are described. The application of this linear algebra to material that appears in under- graduate engineering courses on mechanics is illustrated by discussions of concepts like the area and mass moments of inertia, Mohr’s circles, and the vector cross and triple scalar prod- ucts. The notation, as far as possible, will be a matrix notation that is easily entered into exist- ing symbolic computational programs like Maple, Mathematica, Matlab, and Mathcad. The desire to represent the components of three-dimensional fourth-order tensors that appear in anisotropic elasticity as the components of six-dimensional second-order tensors and thus rep- resent these components in matrices of tensor components in six dimensions leads to the non- traditional part of this appendix. This is also one of the nontraditional aspects in the text of the book, but a minor one. This is described in §A.11, along with the rationale for this approach. A.2. DEFINITION OF SQUARE, COLUMN, AND ROW MATRICES An r-by-c matrix, M, is a rectangular array of numbers consisting of r rows and c columns: ¯MM..
    [Show full text]
  • Glossary of Linear Algebra Terms
    INNER PRODUCT SPACES AND THE GRAM-SCHMIDT PROCESS A. HAVENS 1. The Dot Product and Orthogonality 1.1. Review of the Dot Product. We first recall the notion of the dot product, which gives us a familiar example of an inner product structure on the real vector spaces Rn. This product is connected to the Euclidean geometry of Rn, via lengths and angles measured in Rn. Later, we will introduce inner product spaces in general, and use their structure to define general notions of length and angle on other vector spaces. Definition 1.1. The dot product of real n-vectors in the Euclidean vector space Rn is the scalar product · : Rn × Rn ! R given by the rule n n ! n X X X (u; v) = uiei; viei 7! uivi : i=1 i=1 i n Here BS := (e1;:::; en) is the standard basis of R . With respect to our conventions on basis and matrix multiplication, we may also express the dot product as the matrix-vector product 2 3 v1 6 7 t î ó 6 . 7 u v = u1 : : : un 6 . 7 : 4 5 vn It is a good exercise to verify the following proposition. Proposition 1.1. Let u; v; w 2 Rn be any real n-vectors, and s; t 2 R be any scalars. The Euclidean dot product (u; v) 7! u · v satisfies the following properties. (i:) The dot product is symmetric: u · v = v · u. (ii:) The dot product is bilinear: • (su) · v = s(u · v) = u · (sv), • (u + v) · w = u · w + v · w.
    [Show full text]
  • 28. Exterior Powers
    28. Exterior powers 28.1 Desiderata 28.2 Definitions, uniqueness, existence 28.3 Some elementary facts 28.4 Exterior powers Vif of maps 28.5 Exterior powers of free modules 28.6 Determinants revisited 28.7 Minors of matrices 28.8 Uniqueness in the structure theorem 28.9 Cartan's lemma 28.10 Cayley-Hamilton Theorem 28.11 Worked examples While many of the arguments here have analogues for tensor products, it is worthwhile to repeat these arguments with the relevant variations, both for practice, and to be sensitive to the differences. 1. Desiderata Again, we review missing items in our development of linear algebra. We are missing a development of determinants of matrices whose entries may be in commutative rings, rather than fields. We would like an intrinsic definition of determinants of endomorphisms, rather than one that depends upon a choice of coordinates, even if we eventually prove that the determinant is independent of the coordinates. We anticipate that Artin's axiomatization of determinants of matrices should be mirrored in much of what we do here. We want a direct and natural proof of the Cayley-Hamilton theorem. Linear algebra over fields is insufficient, since the introduction of the indeterminate x in the definition of the characteristic polynomial takes us outside the class of vector spaces over fields. We want to give a conceptual proof for the uniqueness part of the structure theorem for finitely-generated modules over principal ideal domains. Multi-linear algebra over fields is surely insufficient for this. 417 418 Exterior powers 2. Definitions, uniqueness, existence Let R be a commutative ring with 1.
    [Show full text]
  • A Some Basic Rules of Tensor Calculus
    A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e.
    [Show full text]