Linear Independence, Basis and Dimension

Linear Independence, Basis and Dimension

Linear Independence, Basis and Dimension P. Sam Johnson December 29, 2019 P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 1 / 52 Overview We usually associate \dimension" with something geometrical. We turn now to the task of finding a suitable algebraic definition of the dimension of a vector space. This will be done through the concept of a basis for the space. We discuss linear independence, basis and dimension in two lectures. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 2 / 52 Linear Independence, Basis and Dimension Suppose we are given a generating set A for a subspace S. One question that arises naturally is: is there any redundancy in A in the sense that some proper subset of A generates S? To answer this, we start with Definition 1. Let V be a vector space over F . A subset S of V is said to be linearly dependent (or simply, dependent) if there exist distinct vectors x1; x2;:::; xn in S and scalars α1; α2; : : : ; αn in F , not all of which are 0, such that α1x1 + α2x2 + ··· + αnxn = 0: A set which is not linearly dependent is called linearly independent. If the set S contains only finitely many vectors x1; x2;:::; xn, we sometimes say that x1; x2;:::; xn are dependent (or independent) instead of saying S is dependent (or independent). An infinite set S is said to be linearly independent if each finite subset of it is linearly independent. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 3 / 52 Linear Independence The following are easy consequences of the definition. Any set which contains a linearly dependent set is linearly dependent. Any subset of a linear independent set is linearly independent. Any set which contains the 0 vector is linearly dependent. A set S of vectors is linearly independent if and only if each finite subset of S is linearly independent, i.e., if and only if for distinct vectors x1; x2;:::; xn of S, α1x1 + α2x2 + ··· + αnxn = 0 implies each αi = 0. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 4 / 52 Linear Independence If x1; x2;:::; xn are linearly independent, then xi1 ; xi2 ;:::; xin are also linearly independent, for any permutation (i1; i2;:::; in) of (1; 2;:::; n), so the linear indepedence of a set of vectors is well-defined. Exercise 2. Give an example for the fact that \not all zero" cannot be replaced by \all non-zero" in the definition of linear independence. Exercise 3. Let X = fa1; a2;:::g where a1; a2;::: form a sequence of distinct real X numbers. Consider the function fi belonging to the vector space R taking value 1 at x = ai and 0 elsewhere. Then prove that the infinite set ff1; f2;:::g is linearly independent. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 5 / 52 Linear Independence Example 4. p p p In the vector space R over the fieldpQ,p thep sets f1; 2g and f 2; 3g are linearly independent and the set f 2; 3; 12g is linearly independent. Example 5. 2 n−1 The vectors 1; t; t ;:::; t in the vector space Pn form a linearly independent set. The vectors 1; t; t2;::: in the vector space P form an infinite linearly independent set. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 6 / 52 Linear Independency Test To prove that a set fx1; x2;:::; xng is linearly independent, we should claim that the zero vector can only be expressed as zero (scalar) linear combination of xi 's, 1 ≤ i ≤ n. That is, we claim that the equation α1x1 + ··· + αnxn = 0 has only the trivial solution α1 = ··· = αn = 0: Suppose α1x1 + ··· + αnxn = 0 for some scalars α1; : : : ; αn. We should claim that each αi is zero, to prove fx1; x2;:::; xng is linearly independent. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 7 / 52 Linear Dependency Test To prove a set fx1; x2;:::; xng is linearly dependent, we should find scalars α1; : : : ; αn not all zero such that α1x1 + ··· + αnxn = 0: That is, the zero vector can be expressed as some non-zero (scalar) linear combination of xi 's 1 ≤ i ≤ n. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 8 / 52 Basis Definition 6. Let V be a vector space. A basis for V is a linearly independent set of vectors in V which spans the space V . The space is finite dimensional if it has a finite basis. Example 7. Let F be a field and in F n let S be the subset consisting of the vectors e1; e2;:::; en defined by e1 = (1; 0; 0;:::; 0) e2 = (0; 1; 0;:::; 0) . en = (0; 0; 0;:::; 1): n The set S = fe1; e2;:::; eng is a basis for F , called the standard basis of F n. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 9 / 52 Basis of the Row Space of A Let A be an m × n matrix and let R be a row reduced echelon matrix which is row-equivalent to A. For an echelon matrix like R, the row space is clear. It contains all combinations of the rows, as every row space does. To every echelon matirx, with r pivots and r nonzero rows: Its nonzero rows are independent, and its row space has dimension r. Fortunately, it is equally easy to deal with the original matrix A. The row space of A has the same dimension r as the row space of R, and it has the same bases, because the two row spaces are the same. The reason is that each elementary operation leaves the row space unchanged. The rows in R are combinations of the original rows in A. It is true that A and R have different rows, but the combinations of the rows are identical. It is those combinations that make up the row space. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 10 / 52 Basis of F n×1 Exercise 8. Let P be an invertible n × n matrices with entries in the field. Prove that the columns of P form a basis for the space of column matrices, F n×1. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 11 / 52 Basis for the Solution Space Let A be an m × n matrix and let S be the solution space for the homogeneous system AX = 0. The purpose of elimination is to simplify a system of linear equations without changing any of the solutions. Let R be a row reduced echelon matrix which is row-equivalent to A. Then the system AX = 0 is reduced to Rx = 0, and this process is reversible. Therefore S is also the solution space for the system RX = 0. A basis for the solution space can be constructed by reducing to RX = 0, which has n − r free variables { corresponding to the columns of R that do not contain pivots. Then, in turn, we give to each free variable the value 1, to the other free variables the value 0, and solve RX = 0 by back substitution for the remaining (basic) variables. The n − r vectors produced in this way are a basis for the solution space for AX = 0. Refer Linear Algebra by Kenneth Hoffman and Ray Kunze, page 42, for details. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 12 / 52 An Example of an Infinite Basis Exercise 9. Let F be a subfield of the complex numbers and let V be the space of polynomial functions over F . Note that the members of V are the functions from F into F which have a rule of the form n f (x) = c0 + c1x + ··· + cnx : k Let fk (x) = x ; k = 0; 1; 2;::: . Prove that the infinite set ff0; f1; f2;:::g is a basis for V . [Hints : A function f from F into F is a polynomial function iff there exists an integer n and scalars c0; c1;:::; cn such that f = c0f0 + c1f1 + ··· + cnfn. A polynomial of degree n with complex coefficients cannot have more than n distinct roots.] Refer Linear Algebra by Kenneth Hoffman and Ray Kunze, page 43, for details. P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 13 / 52 Basis An infinite basis has been found for the space V of polynomial functions over F . But V might have a finite basis as it is not immediate from the definition. We shall see in the next theorem that this possibility doesn't occur. Example 10. The space of polynomial functions over F cannot have a finite basis. Refer Linear Algebra by Kenneth Hoffman and Ray Kunze, page 43, for details. Caution : Infinite bases have nothing to do with “infinite linear combinations". \Linear combination" always refers to “finite linear combination." P. Sam Johnson (NIT Karnataka) Linear Independence, Basis and Dimension December 29, 2019 14 / 52 Basis Theorem 11. Let V be a vector space which is spanned by a finite set of vectors β1; β2; : : : ; βm. Then any independent set of vectors in V is finite and contains no more than m elements. LA-1(P-4)T-5 To prove the above theorem, we need the following facts: Equivalent systems of linear equations have exactly the same equations. LA-1(P-4)L-1 If A and B are row-equivalent m × n matrices, the homogeneous systems of linear equations AX = 0 and BX = 0 have exactly the same solutions.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    52 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us