A Brief Review of Matrices and Linear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

A Brief Review of Matrices and Linear Algebra A Brief Review of Matrices and Linear Algebra Dr. Robert L. Williams II Mechanical Engineering Ohio University © 2016 Dr. Bob Productions [email protected] people.ohio.edu/williar4 This document is intended as a reference guide to help students review matrices and linear algebra for use in kinematics, dynamics, controls, biomechanics, and robotics. The usefulness of this document extends well beyond these fields. However, it IS NOT intended to replace a textbook in this field of mathematics. 2 Table of Contents MATRIX DEFINITION ........................................................................................................................... 3 SPECIAL MATRICES ............................................................................................................................. 4 MATRIX OPERATIONS ........................................................................................................................ 5 MATRIX ADDITION .................................................................................................................................. 5 MATRIX MULTIPLICATION WITH A SCALAR ............................................................................................. 5 MATRIX MULTIPLICATION ....................................................................................................................... 6 MATRIX DETERMINANT ......................................................................................................................... 10 MATRIX INVERSION ............................................................................................................................... 12 SOLVING A SYSTEM OF LINEAR EQUATIONS ........................................................................... 15 MATRIX EXAMPLES IN MATLAB ................................................................................................... 17 3 Matrix Definition A matrix is an m x n array of numbers, where m is the number of rows and n is the number of columns. aa11 12 a 1n aa a A 21 22 2n aamm12 a mn Matrices may be used to simplify and standardize the solution of n linear equations in n unknowns (where m = n). Matrices are used in velocity, acceleration, and dynamics linear equations (matrices are not used in analytical position analysis, which requires a non-linear solution). 4 Special Matrices These are demonstrated for 3x3 matrices, but apply to all matrix sizes. aaa11 12 13 square matrix (m = n = 3) A aaa 21 22 23 aaa31 32 33 a11 00 diagonal matrix Aa 00 22 00a33 100 identity matrix I 010 3 001 aaa11 21 31 transpose matrix AaaaT (switch rows & columns) 12 22 32 aaa13 23 33 aaa11 12 13 symmetric matrix AAT aaa 12 22 23 aaa13 23 33 x1 column vector (3x1 matrix) X x2 x3 T row vector (1x3 matrix) Xxxx 123 5 Matrix Operations Matrix Addition add like terms and keep the results in place CAB cc11 12 aa 11 12 bb 11 12 abab 11 11 12 12 cc21 22 aa 21 22 bb 21 22 abab 21 21 22 22 Matrix Multiplication with a Scalar multiply each matrix term by the scalar k and keep the results in place aa11 12 kaka 11 12 kA k a21 a 22 ka 21 ka 22 6 Matrix Multiplication CAB In general, AB BA The row and column indices must line up as follows. CAB (x)(x)(x)m nm p p n That is, in a matrix multiplication product, the number of columns p in the left-hand matrix must equal the number of rows p in the right-hand matrix. If this condition is not met, the matrix multiplication is undefined and cannot be done. The size of the resulting matrix [C] is from the number of rows m of the left-hand matrix and the number of columns n of the right-hand matrix, m x n. Multiplication proceeds by multiplying like terms and adding them, along the rows of the left- hand matrix and down the columns of the right-hand matrix (use your index fingers from the left and right hands for the left-hand rows and right-hand columns, respectively). Examples: CAB aab11 12 11 aab21 22 21 ab11 11 ab 12 21 ab21 11 ab 22 21 (2x1) (2x2)(2x1 ) Note the inner indices (p = 2) must match, as stated above, and the dimension of the result is dictated by the outer indices, i.e. m x n = 2x1. For multiplying matrices of multiple columns, follow the same procedure, first using the right index finger down the first column to yield the first column of the result and then using the right index finger down the second column to yield the second column of the result; both use rows 1 and 2. 7 CAB aabb11 12 11 12 aabb21 22 21 22 ab11 11 ab 12 21 ab 11 12 ab 12 22 ab21 11 ab 22 21 ab 21 12 ab 22 22 (2x2) (2x2)(2x2 ) The inner indices (p = 2) must match and the dimension of the result is dictated by the outer indices, i.e. m x n = 2x2. Another analytical example shows non-square matrix multiplication is possible: CAB b aaa 11 11 12 13 b 21 aaa21 22 23 b31 ab11 11 ab 12 21 ab 13 31 ab21 11 ab 22 21 ab 23 31 (2x1) (2x3)(3x1 ) The inner indices (p = 3) must match and the dimension of the result is dictated by the outer indices, i.e. m x n = 2x1. 8 We conclude with two strange but legit analytical matrix multiplication examples. CAB b11 aaab 11 12 13 21 b31 ab11 11 ab 12 21 ab 13 31 (1x1) (1x3)(3x1 ) This example is the dot product between two column vectors, i.e. the dot product can be found by transposing the first vector and using matrix multiplication: T A BAB The last example is even stranger because the matrix multiplication here only involves one multiplication for each term in the result. CAB a11 abbb 21 11 12 13 a31 ab11 11 ab 11 12 ab 11 13 ab ab ab 21 11 21 12 21 13 ab31 11 ab 31 12 ab 31 13 (3x3) (3x1)(1x3 ) This happens whenever the number of columns in the left-hand matrix equals the number of rows in the right-hand matrix, and both are p = 1. 9 Numerical Matrix Multiplication Examples 78 123 A B 98 456 76 CAB 78 123 98 456 76 7 18 21 8 16 18 46 42 28 45 42 32 40 36 115 108 (2x2) (2x3)(3x2 ) DBA 78 123 98 456 76 73214402148 395469 93218402748 415875 72414302136 314457 (3x3) (3x2)(2x3 ) Note in this example that A BBA ; they are not even of the same size. 10 Matrix Determinant The determinant of a square n x n matrix is a scalar. The matrix determinant is undefined for a non-square matrix. The determinant of a square matrix A is denoted det(A) or A . The determinant notation should not be confused with the absolute-value symbol. The MATLAB function for matrix determinant is det(A). If a nonhomogeneous system of n linear equations in n unknowns is dependent, the coefficient matrix A is singular, and the determinant of matrix A is zero. In this case no unique solution exists to these equations. On the other hand, if the matrix determinant is non-zero, then the matrix is non- singular, the system of equations is independent, and a unique solution exists. The formula to calculate a 2 x 2 matrix determinant is straight-forward. aa11 12 A aa21 22 A aa11 22 aa 21 12 To calculate the determinant of 3 x 3 and larger square matrices, we can expand about any one row or column, utilizing sub-matrix determinants. Each sub-determinant is formed by crossing out the current row and its column and retaining the remaining terms as an n–1 x n–1 square matrix, each of whose determinants must also be evaluated in the process. The pivot term (the entry in the cross-out row and column) multiplies the sub-matrix determinants, and there is an alternating + / – / + / – etc. sign pattern. Here is an explicit example for a 3 x 3 matrix, expanding about the first row (all other options will yield identical results). aaa11 12 13 A aaa 21 22 23 aaa31 32 33 aa22 23 aa 21 23 aa 21 22 Aa11 a 12 a 13 aa32 33 aa 31 33 aa 31 32 aaa11()()() 22 33 aa 32 23 aaa 12 21 33 aa 31 23 aaa 13 21 32 aa 31 22 11 For a 3 x 3 matrix only, the determinant can alternatively be calculated as shown below, by copying columns 1 and 2 outside the matrix, multiplying the downward diagonals with + signs and multiplying the upward diagonals with – signs (the result is the same as in the above formula). abc ab Adefde ghk gh aek bfg cdh gec hfa kdb a()()() ek hf b kd fg c dh ge aaa11 12 13 aa 11 12 Aa 21 a 22 a 23 a 21 a 22 aaa31 32 33 aa 31 32 aaa11 22 33 aaa 12 23 31 aaa 13 21 32 aaa 31 22 13 aaa 32 23 11 aaa 33 21 12 aaa11()()() 22 33 aa 32 23 aaa 12 23 31 aa 33 21 aaa 13 21 32 aa 31 22 aaa11()( 22 33 aa 32 23 aaa 12 21 33 aa 31 231321323122)(aaa aa ) A common usage of the 3 x 3 matrix determinant is to calculate the cross product PP12 . ijkˆˆˆ pppp pppp pp 12yz 12 zy ˆˆ11yz11xzˆ 11 xy PP1211 px pyz p 1 i j k pppp12xz 12 zx pp pp 22yzpp22xz 22 xy p222x ppyz pppp12xy 12 yx 12 Matrix Inversion Since we cannot divide by a matrix, we multiply by the matrix inverse instead. Given CAB , solve for [B]. CAB A 11CAAB IB B 1 BAC Matrix [A] must be square (m = n) to invert. The following math facts must be true: 11 A AAAI where [I] is the identity matrix, the matrix 1 (ones on the diagonal and zeros everywhere else). To calculate the matrix inverse use the following expression. 1 adjoint(A ) A A where A is the determinant of [A]. T adjoint(A ) cofactor(A ) ij cofactor(A) aAij(1) ij minor minor Aij is the determinant of the submatrix of [A] with row i and column j removed. 13 For another example, given CAB , solve for [A] CAB CB11 ABB AI A 1 A CB In general the order of matrix multiplication and inversion is crucial and cannot be changed.
Recommended publications
  • Introduction to Linear Bialgebra
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of New Mexico University of New Mexico UNM Digital Repository Mathematics and Statistics Faculty and Staff Publications Academic Department Resources 2005 INTRODUCTION TO LINEAR BIALGEBRA Florentin Smarandache University of New Mexico, [email protected] W.B. Vasantha Kandasamy K. Ilanthenral Follow this and additional works at: https://digitalrepository.unm.edu/math_fsp Part of the Algebra Commons, Analysis Commons, Discrete Mathematics and Combinatorics Commons, and the Other Mathematics Commons Recommended Citation Smarandache, Florentin; W.B. Vasantha Kandasamy; and K. Ilanthenral. "INTRODUCTION TO LINEAR BIALGEBRA." (2005). https://digitalrepository.unm.edu/math_fsp/232 This Book is brought to you for free and open access by the Academic Department Resources at UNM Digital Repository. It has been accepted for inclusion in Mathematics and Statistics Faculty and Staff Publications by an authorized administrator of UNM Digital Repository. For more information, please contact [email protected], [email protected], [email protected]. INTRODUCTION TO LINEAR BIALGEBRA W. B. Vasantha Kandasamy Department of Mathematics Indian Institute of Technology, Madras Chennai – 600036, India e-mail: [email protected] web: http://mat.iitm.ac.in/~wbv Florentin Smarandache Department of Mathematics University of New Mexico Gallup, NM 87301, USA e-mail: [email protected] K. Ilanthenral Editor, Maths Tiger, Quarterly Journal Flat No.11, Mayura Park, 16, Kazhikundram Main Road, Tharamani, Chennai – 600 113, India e-mail: [email protected] HEXIS Phoenix, Arizona 2005 1 This book can be ordered in a paper bound reprint from: Books on Demand ProQuest Information & Learning (University of Microfilm International) 300 N.
    [Show full text]
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Eigenvalues and Eigenvectors
    Jim Lambers MAT 605 Fall Semester 2015-16 Lecture 14 and 15 Notes These notes correspond to Sections 4.4 and 4.5 in the text. Eigenvalues and Eigenvectors In order to compute the matrix exponential eAt for a given matrix A, it is helpful to know the eigenvalues and eigenvectors of A. Definitions and Properties Let A be an n × n matrix. A nonzero vector x is called an eigenvector of A if there exists a scalar λ such that Ax = λx: The scalar λ is called an eigenvalue of A, and we say that x is an eigenvector of A corresponding to λ. We see that an eigenvector of A is a vector for which matrix-vector multiplication with A is equivalent to scalar multiplication by λ. Because x is nonzero, it follows that if x is an eigenvector of A, then the matrix A − λI is singular, where λ is the corresponding eigenvalue. Therefore, λ satisfies the equation det(A − λI) = 0: The expression det(A−λI) is a polynomial of degree n in λ, and therefore is called the characteristic polynomial of A (eigenvalues are sometimes called characteristic values). It follows from the fact that the eigenvalues of A are the roots of the characteristic polynomial that A has n eigenvalues, which can repeat, and can also be complex, even if A is real. However, if A is real, any complex eigenvalues must occur in complex-conjugate pairs. The set of eigenvalues of A is called the spectrum of A, and denoted by λ(A). This terminology explains why the magnitude of the largest eigenvalues is called the spectral radius of A.
    [Show full text]
  • Algebra of Linear Transformations and Matrices Math 130 Linear Algebra
    Then the two compositions are 0 −1 1 0 0 1 BA = = 1 0 0 −1 1 0 Algebra of linear transformations and 1 0 0 −1 0 −1 AB = = matrices 0 −1 1 0 −1 0 Math 130 Linear Algebra D Joyce, Fall 2013 The products aren't the same. You can perform these on physical objects. Take We've looked at the operations of addition and a book. First rotate it 90◦ then flip it over. Start scalar multiplication on linear transformations and again but flip first then rotate 90◦. The book ends used them to define addition and scalar multipli- up in different orientations. cation on matrices. For a given basis β on V and another basis γ on W , we have an isomorphism Matrix multiplication is associative. Al- γ ' φβ : Hom(V; W ) ! Mm×n of vector spaces which though it's not commutative, it is associative. assigns to a linear transformation T : V ! W its That's because it corresponds to composition of γ standard matrix [T ]β. functions, and that's associative. Given any three We also have matrix multiplication which corre- functions f, g, and h, we'll show (f ◦ g) ◦ h = sponds to composition of linear transformations. If f ◦ (g ◦ h) by showing the two sides have the same A is the standard matrix for a transformation S, values for all x. and B is the standard matrix for a transformation T , then we defined multiplication of matrices so ((f ◦ g) ◦ h)(x) = (f ◦ g)(h(x)) = f(g(h(x))) that the product AB is be the standard matrix for S ◦ T .
    [Show full text]
  • Things You Need to Know About Linear Algebra Math 131 Multivariate
    the standard (x; y; z) coordinate three-space. Nota- tions for vectors. Geometric interpretation as vec- tors as displacements as well as vectors as points. Vector addition, the zero vector 0, vector subtrac- Things you need to know about tion, scalar multiplication, their properties, and linear algebra their geometric interpretation. Math 131 Multivariate Calculus We start using these concepts right away. In D Joyce, Spring 2014 chapter 2, we'll begin our study of vector-valued functions, and we'll use the coordinates, vector no- The relation between linear algebra and mul- tation, and the rest of the topics reviewed in this tivariate calculus. We'll spend the first few section. meetings reviewing linear algebra. We'll look at just about everything in chapter 1. Section 1.2. Vectors and equations in R3. Stan- Linear algebra is the study of linear trans- dard basis vectors i; j; k in R3. Parametric equa- formations (also called linear functions) from n- tions for lines. Symmetric form equations for a line dimensional space to m-dimensional space, where in R3. Parametric equations for curves x : R ! R2 m is usually equal to n, and that's often 2 or 3. in the plane, where x(t) = (x(t); y(t)). A linear transformation f : Rn ! Rm is one that We'll use standard basis vectors throughout the preserves addition and scalar multiplication, that course, beginning with chapter 2. We'll study tan- is, f(a + b) = f(a) + f(b), and f(ca) = cf(a). We'll gent lines of curves and tangent planes of surfaces generally use bold face for vectors and for vector- in that chapter and throughout the course.
    [Show full text]
  • Schaum's Outline of Linear Algebra (4Th Edition)
    SCHAUM’S SCHAUM’S outlines outlines Linear Algebra Fourth Edition Seymour Lipschutz, Ph.D. Temple University Marc Lars Lipson, Ph.D. University of Virginia Schaum’s Outline Series New York Chicago San Francisco Lisbon London Madrid Mexico City Milan New Delhi San Juan Seoul Singapore Sydney Toronto Copyright © 2009, 2001, 1991, 1968 by The McGraw-Hill Companies, Inc. All rights reserved. Except as permitted under the United States Copyright Act of 1976, no part of this publication may be reproduced or distributed in any form or by any means, or stored in a database or retrieval system, without the prior writ- ten permission of the publisher. ISBN: 978-0-07-154353-8 MHID: 0-07-154353-8 The material in this eBook also appears in the print version of this title: ISBN: 978-0-07-154352-1, MHID: 0-07-154352-X. All trademarks are trademarks of their respective owners. Rather than put a trademark symbol after every occurrence of a trademarked name, we use names in an editorial fashion only, and to the benefit of the trademark owner, with no intention of infringement of the trademark. Where such designations appear in this book, they have been printed with initial caps. McGraw-Hill eBooks are available at special quantity discounts to use as premiums and sales promotions, or for use in corporate training programs. To contact a representative please e-mail us at [email protected]. TERMS OF USE This is a copyrighted work and The McGraw-Hill Companies, Inc. (“McGraw-Hill”) and its licensors reserve all rights in and to the work.
    [Show full text]
  • Math 217: Multilinearity of Determinants Professor Karen Smith (C)2015 UM Math Dept Licensed Under a Creative Commons By-NC-SA 4.0 International License
    Math 217: Multilinearity of Determinants Professor Karen Smith (c)2015 UM Math Dept licensed under a Creative Commons By-NC-SA 4.0 International License. A. Let V −!T V be a linear transformation where V has dimension n. 1. What is meant by the determinant of T ? Why is this well-defined? Solution note: The determinant of T is the determinant of the B-matrix of T , for any basis B of V . Since all B-matrices of T are similar, and similar matrices have the same determinant, this is well-defined—it doesn't depend on which basis we pick. 2. Define the rank of T . Solution note: The rank of T is the dimension of the image. 3. Explain why T is an isomorphism if and only if det T is not zero. Solution note: T is an isomorphism if and only if [T ]B is invertible (for any choice of basis B), which happens if and only if det T 6= 0. 3 4. Now let V = R and let T be rotation around the axis L (a line through the origin) by an 21 0 0 3 3 angle θ. Find a basis for R in which the matrix of ρ is 40 cosθ −sinθ5 : Use this to 0 sinθ cosθ compute the determinant of T . Is T othogonal? Solution note: Let v be any vector spanning L and let u1; u2 be an orthonormal basis ? for V = L . Rotation fixes ~v, which means the B-matrix in the basis (v; u1; u2) has 213 first column 405.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • Basic Multivector Calculus, Proceedings of FAN 2008, Hiroshima, Japan, 23+24 October 2008
    E. Hitzer, Basic Multivector Calculus, Proceedings of FAN 2008, Hiroshima, Japan, 23+24 October 2008. Basic Multivector Calculus Eckhard Hitzer, Department of Applied Physics, Faculty of Engineering, University of Fukui Abstract: We begin with introducing the generalization of real, complex, and quaternion numbers to hypercomplex numbers, also known as Clifford numbers, or multivectors of geometric algebra. Multivectors encode everything from vectors, rotations, scaling transformations, improper transformations (reflections, inversions), geometric objects (like lines and spheres), spinors, and tensors, and the like. Multivector calculus allows to define functions mapping multivectors to multivectors, differentiation, integration, function norms, multivector Fourier transformations and wavelet transformations, filtering, windowing, etc. We give a basic introduction into this general mathematical language, which has fascinating applications in physics, engineering, and computer science. more didactic approach for newcomers to geometric algebra. 1. Introduction G(I) is the full geometric algebra over all vectors in the “Now faith is being sure of what we hope for and certain of what we do not see. This is what the ancients were n-dimensional unit pseudoscalar I e1 e2 ... en . commended for. By faith we understand that the universe was formed at God’s command, so that what is seen was not made 1 A G (I) is the n-dimensional vector sub-space of out of what was visible.” [7] n The German 19th century mathematician H. Grassmann had grade-1 elements in G(I) spanned by e1,e2 ,...,en . For the clear vision, that his “extension theory (now developed to geometric calculus) … forms the keystone of the entire 1 vectors a,b,c An G (I) and scalars , ,, ; structure of mathematics.”[6] The algebraic “grammar” of this universal form of calculus is geometric algebra (or Clifford G(I) has the fundamental properties of algebra).
    [Show full text]
  • Span, Linear Independence and Basis Rank and Nullity
    Remarks for Exam 2 in Linear Algebra Span, linear independence and basis The span of a set of vectors is the set of all linear combinations of the vectors. A set of vectors is linearly independent if the only solution to c1v1 + ::: + ckvk = 0 is ci = 0 for all i. Given a set of vectors, you can determine if they are linearly independent by writing the vectors as the columns of the matrix A, and solving Ax = 0. If there are any non-zero solutions, then the vectors are linearly dependent. If the only solution is x = 0, then they are linearly independent. A basis for a subspace S of Rn is a set of vectors that spans S and is linearly independent. There are many bases, but every basis must have exactly k = dim(S) vectors. A spanning set in S must contain at least k vectors, and a linearly independent set in S can contain at most k vectors. A spanning set in S with exactly k vectors is a basis. A linearly independent set in S with exactly k vectors is a basis. Rank and nullity The span of the rows of matrix A is the row space of A. The span of the columns of A is the column space C(A). The row and column spaces always have the same dimension, called the rank of A. Let r = rank(A). Then r is the maximal number of linearly independent row vectors, and the maximal number of linearly independent column vectors. So if r < n then the columns are linearly dependent; if r < m then the rows are linearly dependent.
    [Show full text]
  • MATH 304 Linear Algebra Lecture 24: Scalar Product. Vectors: Geometric Approach
    MATH 304 Linear Algebra Lecture 24: Scalar product. Vectors: geometric approach B A B′ A′ A vector is represented by a directed segment. • Directed segment is drawn as an arrow. • Different arrows represent the same vector if • they are of the same length and direction. Vectors: geometric approach v B A v B′ A′ −→AB denotes the vector represented by the arrow with tip at B and tail at A. −→AA is called the zero vector and denoted 0. Vectors: geometric approach v B − A v B′ A′ If v = −→AB then −→BA is called the negative vector of v and denoted v. − Vector addition Given vectors a and b, their sum a + b is defined by the rule −→AB + −→BC = −→AC. That is, choose points A, B, C so that −→AB = a and −→BC = b. Then a + b = −→AC. B b a C a b + B′ b A a C ′ a + b A′ The difference of the two vectors is defined as a b = a + ( b). − − b a b − a Properties of vector addition: a + b = b + a (commutative law) (a + b) + c = a + (b + c) (associative law) a + 0 = 0 + a = a a + ( a) = ( a) + a = 0 − − Let −→AB = a. Then a + 0 = −→AB + −→BB = −→AB = a, a + ( a) = −→AB + −→BA = −→AA = 0. − Let −→AB = a, −→BC = b, and −→CD = c. Then (a + b) + c = (−→AB + −→BC) + −→CD = −→AC + −→CD = −→AD, a + (b + c) = −→AB + (−→BC + −→CD) = −→AB + −→BD = −→AD. Parallelogram rule Let −→AB = a, −→BC = b, −−→AB′ = b, and −−→B′C ′ = a. Then a + b = −→AC, b + a = −−→AC ′.
    [Show full text]