4.5 Basis and Dimension of a Vector Space

Total Page:16

File Type:pdf, Size:1020Kb

4.5 Basis and Dimension of a Vector Space 4.5. BASIS AND DIMENSION OF A VECTOR SPACE 135 4.5 Basis and Dimension of a Vector Space In the section on spanning sets and linear independence, we were trying to understand what the elements of a vector space looked like by studying how they could be generated. We learned that some subsets of a vector space could generate the entire vector space. Such subsets were called spanning sets. Other subsets did not generate the entire space, but their span was still a subspace of the underlying vector space. In some cases, the number of vectors in such a set was redundant in the sense that one or more of the vectors could be removed, without changing the span of the set. In other cases, there was not a unique way to generate some vectors in the space. In this section, we want to make this process of generating all the elements of a vector space more reliable, more e¢ cient. 4.5.1 Basis of a Vector Space De…nition 297 Let V denote a vector space and S = u1; u2; :::; un a subset of V . S is called a basis for V if the following is true: f g 1. S spans V . 2. S is linearly independent. This de…nition tells us that a basis has to contain enough vectors to generate the entire vector space. But it does not contain too many. In other words, if we removed one of the vectors, it would no longer generate the space. A basis is the vector space generalization of a coordinate system in R2 or R3. Example 298 We have already seen that the set S = e1; e2 where e1 = (1; 0) 2 f g and e2 = (0; 1) was a spanning set of R . It is also linearly independent for the only solution of the vector equation c1e1 + c2e2 = 0 is the trivial solution. Therefore, S is a basis for R2. It is called the standard basis for R2. These vectors also have a special name. (1; 0) is i and (0; 1) is j. 3 Example 299 Similarly, the standard basis for R is the set e1; e2; e3 where f g e1 = (1; 0; 0), e2 = (0; 1; 0) and e3 = (0; 0; 1). These vectors also have a special name. They are i; j and k respectively. 2 Example 300 Prove that S = 1; x; x is a basis for P2, the set of polynomials of degree less than or equal to 2. We need to prove that S spans P2 and is linearly independent. S spans P2. We already did this in the section on spanning sets. A typical polynomial of degree less than or equal to 2 is ax2 + bx + c. S is linearly independent. Here, we need to show that the only solution to a (1) + bx + cx2 = 0 (where 0 is the zero polynomial) is a = b = c = 0. 136 CHAPTER 4. VECTOR SPACES From algebra, we remember that two polynomials are equal if and only if their corresponding coe¢ cients are equal. The zero polynomial has all its coe¢ cients equal to zero. So, a (1) + bx + cx2 = 0 if and only if a = 0, b = 0, c = 0. Which proves that S is linearly independent. We will see more examples shortly. The next theorem outlines an important di¤erence between a basis and a spanning set. Any vector in a vector space can be represented in a unique way as a linear combination of the vectors of a basis.. Theorem 301 Let V denote a vector space and S = u1; u2; :::; un a basis of V . Every vector in V can be written in a unique wayf as a linear combinationg of vectors in S. Proof. Since S is a basis, we know that it spans V . If v V , then there 2 exists scalars c1; c2; :::; cn such that v =c1u1 + c2u2 + ::: + cnun. Suppose there is another way to write v. That is, there exist scalars d1; d2; :::; dn such that v =d1u1 + d2u2 + ::: + dnun. Then, c1u1 + c2u2 + ::: + cnun = d1u1 + d2u2 + ::: + dnun. In other words, (c1 d1) u1 + (c2 d2) u2 + ::: + (cn dn) un = 0. Since S is a basis, it must be linearly independent. The unique solution to (c1 d1) u1 +(c2 d2) u2 +:::+(cn dn) un = 0 must be the trivial solution. It follows that ci di = 0 for i = 1; 2; :::; n in other words ci = di for i = 1; 2; :::; n. Therefore, the two representations of v are the same. Remark 302 We say that any vector v of V has a unique representation with respect to the basis S. The scalars used in the linear representation are called the coordinates of the vector. For example, the vector (x; y) can be represented in the basis (1; 0) ; (0; 1) by the linear combination (x; y) = x (1; 0) + y (0; 1). Thus, x andfy are the coordinatesg of this vector (we knew that!). De…nition 303 If V is a vector space, and B = u1; u2; :::; un is an ordered basis of V , then we know that every vector v of Vfcan be expressedg as a linear combination of the vectors in S in a unique way. In others words, there exists unique scalars c1; c2; :::; cn such that v = c1u1 + c2u2 + ::: + cnun. These scalars are called the coordinates of v relative to the ordered basis B. Remark 304 The term "ordered basis" simply means that the order in which we list the elements is important. Indeed it is since each coordinate is with respect to one of the vector in the basis. We know that in R2, (2; 3) is not the same as (3; 2). Example 305 What are the coordinates of (1; 2; 3) with respect to the basis (1; 1; 0) ; (0; 1; 1) ; (1; 1; 1) ? f g One can indeed verify that this set is a basis for R3. Finding the coordinates of (1; 2; 3) with respect to this new basis amounts to …nding the numbers (a; b; c) such that a (1; 1; 0) + b (0; 1; 1) + c (1; 1; 1) = (1; 2; 3). This amounts to solving 4.5. BASIS AND DIMENSION OF A VECTOR SPACE 137 1 0 1 a 1 the system 1 1 1 b = 2 . The solution is 2 0 1 1 3 2 c 3 2 3 3 4 5 4 5 4 5 1 a 1 0 1 1 b = 1 1 1 2 2 c 3 2 0 1 1 3 2 3 3 4 5 4 1 5 4 5 = 1 2 2 3 4 5 The next theorem, deals with the number of vectors the basis of a given vector space can have. We will state the theorem without proof. Theorem 306 Let V denote a vector space and S = u1; u2; :::; un a basis of V . f g 1. Any subset of V containing more than n vectors must be dependent. 2. Any subset of V containing less than n vectors cannot span V . Proof. We prove each part separately. 1. Consider W = v1; v2; :::; vr a subset of V where r > n. We must show f g that W in dependent. Since S is a basis, we can writ each vi in term of elements in S. More speci…cally, there exists constants cij with 1 i r and 1 j n such that vi = ci1u1 + ci2u2 + ::: + cinun. Consider the linear combination r r djvj = dj (cj1u1 + cj2u2 + ::: + cjnun) = 0 j=1 j=1 X X d1c11 + d1c12 + ::: + d1c1n = 0 d2c21 + d2c22 + ::: + d2c2n = 0 So, we must solve 8 . where the unknowns > . <> drcr1 + drcr2 + ::: + drcrn = 0 are d1; d2; :::; dr. Since> we have more unknowns than equations, we are guaranteed that this:> homogeneous system will have a nontrivial solution. Thus W is dependent. 2. Consider W = v1; v2; :::; vr a subset of V where r < n. We must show that W doesf not span Vg. We do it by contradiction. We assume it does span V and show this would imply that S is dependent. Suppose that there exists constants cij with 1 i n and 1 j r such that ui = ci1v1 + ci2v2 + ::: + cirvr. Consider the linear combination n r djuj = dj (cj1v1 + cj2v2 + ::: + cjrvr) = 0 j=1 j=1 X X 138 CHAPTER 4. VECTOR SPACES d1c11 + d1c12 + ::: + d1c1r = 0 d2c21 + d2c22 + ::: + d2c2r = 0 So, we must solve 8 . where the unknowns > . <> dncn1 + dncn2 + ::: + dncnr = 0 are d1; d2; :::; dn.> Since we have more unknowns than equations, we are guaranteed that this:> homogeneous system will have a nontrivial solution. Thus S would be dependent. But it can’t be since it is a basis. Corollary 307 Let V denote a vector space. If V has a basis with n elements, then all the bases of V will have n elements. Proof. Assume that S1 is a basis of V with n elements and S2 is another basis with m elements. We need to show that m = n. Since S1 is a basis, S2 being also a basis implies that m n. If we had m > n, by the theorem, S2 would be dependent, hence not a basis. Similarly, since S2 is a basis, S1 being also a basis implies that n m. The only way we can have m n and n m is if m = n. 4.5.2 Dimension of a Vector Space All the bases of a vector space must have the same number of elements. This common number of elements has a name. De…nition 308 Let V denote a vector space.
Recommended publications
  • Introduction to Linear Bialgebra
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of New Mexico University of New Mexico UNM Digital Repository Mathematics and Statistics Faculty and Staff Publications Academic Department Resources 2005 INTRODUCTION TO LINEAR BIALGEBRA Florentin Smarandache University of New Mexico, [email protected] W.B. Vasantha Kandasamy K. Ilanthenral Follow this and additional works at: https://digitalrepository.unm.edu/math_fsp Part of the Algebra Commons, Analysis Commons, Discrete Mathematics and Combinatorics Commons, and the Other Mathematics Commons Recommended Citation Smarandache, Florentin; W.B. Vasantha Kandasamy; and K. Ilanthenral. "INTRODUCTION TO LINEAR BIALGEBRA." (2005). https://digitalrepository.unm.edu/math_fsp/232 This Book is brought to you for free and open access by the Academic Department Resources at UNM Digital Repository. It has been accepted for inclusion in Mathematics and Statistics Faculty and Staff Publications by an authorized administrator of UNM Digital Repository. For more information, please contact [email protected], [email protected], [email protected]. INTRODUCTION TO LINEAR BIALGEBRA W. B. Vasantha Kandasamy Department of Mathematics Indian Institute of Technology, Madras Chennai – 600036, India e-mail: [email protected] web: http://mat.iitm.ac.in/~wbv Florentin Smarandache Department of Mathematics University of New Mexico Gallup, NM 87301, USA e-mail: [email protected] K. Ilanthenral Editor, Maths Tiger, Quarterly Journal Flat No.11, Mayura Park, 16, Kazhikundram Main Road, Tharamani, Chennai – 600 113, India e-mail: [email protected] HEXIS Phoenix, Arizona 2005 1 This book can be ordered in a paper bound reprint from: Books on Demand ProQuest Information & Learning (University of Microfilm International) 300 N.
    [Show full text]
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • 1 Lifts of Polytopes
    Lecture 5: Lifts of polytopes and non-negative rank CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Lifts of polytopes 1.1 Polytopes and inequalities Recall that the convex hull of a subset X n is defined by ⊆ conv X λx + 1 λ x0 : x; x0 X; λ 0; 1 : ( ) f ( − ) 2 2 [ ]g A d-dimensional convex polytope P d is the convex hull of a finite set of points in d: ⊆ P conv x1;:::; xk (f g) d for some x1;:::; xk . 2 Every polytope has a dual representation: It is a closed and bounded set defined by a family of linear inequalities P x d : Ax 6 b f 2 g for some matrix A m d. 2 × Let us define a measure of complexity for P: Define γ P to be the smallest number m such that for some C s d ; y s ; A m d ; b m, we have ( ) 2 × 2 2 × 2 P x d : Cx y and Ax 6 b : f 2 g In other words, this is the minimum number of inequalities needed to describe P. If P is full- dimensional, then this is precisely the number of facets of P (a facet is a maximal proper face of P). Thinking of γ P as a measure of complexity makes sense from the point of view of optimization: Interior point( methods) can efficiently optimize linear functions over P (to arbitrary accuracy) in time that is polynomial in γ P . ( ) 1.2 Lifts of polytopes Many simple polytopes require a large number of inequalities to describe.
    [Show full text]
  • Multivector Differentiation and Linear Algebra 0.5Cm 17Th Santaló
    Multivector differentiation and Linear Algebra 17th Santalo´ Summer School 2016, Santander Joan Lasenby Signal Processing Group, Engineering Department, Cambridge, UK and Trinity College Cambridge [email protected], www-sigproc.eng.cam.ac.uk/ s jl 23 August 2016 1 / 78 Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. 2 / 78 Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. 3 / 78 Functional Differentiation: very briefly... Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. 4 / 78 Summary Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... 5 / 78 Overview The Multivector Derivative. Examples of differentiation wrt multivectors. Linear Algebra: matrices and tensors as linear functions mapping between elements of the algebra. Functional Differentiation: very briefly... Summary 6 / 78 We now want to generalise this idea to enable us to find the derivative of F(X), in the A ‘direction’ – where X is a general mixed grade multivector (so F(X) is a general multivector valued function of X). Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi. Then, provided A has same grades as X, it makes sense to define: F(X + tA) − F(X) A ∗ ¶XF(X) = lim t!0 t The Multivector Derivative Recall our definition of the directional derivative in the a direction F(x + ea) − F(x) a·r F(x) = lim e!0 e 7 / 78 Let us use ∗ to denote taking the scalar part, ie P ∗ Q ≡ hPQi.
    [Show full text]
  • Simplicial Complexes
    46 III Complexes III.1 Simplicial Complexes There are many ways to represent a topological space, one being a collection of simplices that are glued to each other in a structured manner. Such a collection can easily grow large but all its elements are simple. This is not so convenient for hand-calculations but close to ideal for computer implementations. In this book, we use simplicial complexes as the primary representation of topology. Rd k Simplices. Let u0; u1; : : : ; uk be points in . A point x = i=0 λiui is an affine combination of the ui if the λi sum to 1. The affine hull is the set of affine combinations. It is a k-plane if the k + 1 points are affinely Pindependent by which we mean that any two affine combinations, x = λiui and y = µiui, are the same iff λi = µi for all i. The k + 1 points are affinely independent iff P d P the k vectors ui − u0, for 1 ≤ i ≤ k, are linearly independent. In R we can have at most d linearly independent vectors and therefore at most d+1 affinely independent points. An affine combination x = λiui is a convex combination if all λi are non- negative. The convex hull is the set of convex combinations. A k-simplex is the P convex hull of k + 1 affinely independent points, σ = conv fu0; u1; : : : ; ukg. We sometimes say the ui span σ. Its dimension is dim σ = k. We use special names of the first few dimensions, vertex for 0-simplex, edge for 1-simplex, triangle for 2-simplex, and tetrahedron for 3-simplex; see Figure III.1.
    [Show full text]
  • Degrees of Freedom in Quadratic Goodness of Fit
    Submitted to the Annals of Statistics DEGREES OF FREEDOM IN QUADRATIC GOODNESS OF FIT By Bruce G. Lindsay∗, Marianthi Markatouy and Surajit Ray Pennsylvania State University, Columbia University, Boston University We study the effect of degrees of freedom on the level and power of quadratic distance based tests. The concept of an eigendepth index is introduced and discussed in the context of selecting the optimal de- grees of freedom, where optimality refers to high power. We introduce the class of diffusion kernels by the properties we seek these kernels to have and give a method for constructing them by exponentiating the rate matrix of a Markov chain. Product kernels and their spectral decomposition are discussed and shown useful for high dimensional data problems. 1. Introduction. Lindsay et al. (2008) developed a general theory for good- ness of fit testing based on quadratic distances. This class of tests is enormous, encompassing many of the tests found in the literature. It includes tests based on characteristic functions, density estimation, and the chi-squared tests, as well as providing quadratic approximations to many other tests, such as those based on likelihood ratios. The flexibility of the methodology is particularly important for enabling statisticians to readily construct tests for model fit in higher dimensions and in more complex data. ∗Supported by NSF grant DMS-04-05637 ySupported by NSF grant DMS-05-04957 AMS 2000 subject classifications: Primary 62F99, 62F03; secondary 62H15, 62F05 Keywords and phrases: Degrees of freedom, eigendepth, high dimensional goodness of fit, Markov diffusion kernels, quadratic distance, spectral decomposition in high dimensions 1 2 LINDSAY ET AL.
    [Show full text]
  • Arxiv:1910.10745V1 [Cond-Mat.Str-El] 23 Oct 2019 2.2 Symmetry-Protected Time Crystals
    A Brief History of Time Crystals Vedika Khemania,b,∗, Roderich Moessnerc, S. L. Sondhid aDepartment of Physics, Harvard University, Cambridge, Massachusetts 02138, USA bDepartment of Physics, Stanford University, Stanford, California 94305, USA cMax-Planck-Institut f¨urPhysik komplexer Systeme, 01187 Dresden, Germany dDepartment of Physics, Princeton University, Princeton, New Jersey 08544, USA Abstract The idea of breaking time-translation symmetry has fascinated humanity at least since ancient proposals of the per- petuum mobile. Unlike the breaking of other symmetries, such as spatial translation in a crystal or spin rotation in a magnet, time translation symmetry breaking (TTSB) has been tantalisingly elusive. We review this history up to recent developments which have shown that discrete TTSB does takes place in periodically driven (Floquet) systems in the presence of many-body localization (MBL). Such Floquet time-crystals represent a new paradigm in quantum statistical mechanics — that of an intrinsically out-of-equilibrium many-body phase of matter with no equilibrium counterpart. We include a compendium of the necessary background on the statistical mechanics of phase structure in many- body systems, before specializing to a detailed discussion of the nature, and diagnostics, of TTSB. In particular, we provide precise definitions that formalize the notion of a time-crystal as a stable, macroscopic, conservative clock — explaining both the need for a many-body system in the infinite volume limit, and for a lack of net energy absorption or dissipation. Our discussion emphasizes that TTSB in a time-crystal is accompanied by the breaking of a spatial symmetry — so that time-crystals exhibit a novel form of spatiotemporal order.
    [Show full text]
  • THE DIMENSION of a VECTOR SPACE 1. Introduction This Handout
    THE DIMENSION OF A VECTOR SPACE KEITH CONRAD 1. Introduction This handout is a supplementary discussion leading up to the definition of dimension of a vector space and some of its properties. We start by defining the span of a finite set of vectors and linear independence of a finite set of vectors, which are combined to define the all-important concept of a basis. Definition 1.1. Let V be a vector space over a field F . For any finite subset fv1; : : : ; vng of V , its span is the set of all of its linear combinations: Span(v1; : : : ; vn) = fc1v1 + ··· + cnvn : ci 2 F g: Example 1.2. In F 3, Span((1; 0; 0); (0; 1; 0)) is the xy-plane in F 3. Example 1.3. If v is a single vector in V then Span(v) = fcv : c 2 F g = F v is the set of scalar multiples of v, which for nonzero v should be thought of geometrically as a line (through the origin, since it includes 0 · v = 0). Since sums of linear combinations are linear combinations and the scalar multiple of a linear combination is a linear combination, Span(v1; : : : ; vn) is a subspace of V . It may not be all of V , of course. Definition 1.4. If fv1; : : : ; vng satisfies Span(fv1; : : : ; vng) = V , that is, if every vector in V is a linear combination from fv1; : : : ; vng, then we say this set spans V or it is a spanning set for V . Example 1.5. In F 2, the set f(1; 0); (0; 1); (1; 1)g is a spanning set of F 2.
    [Show full text]
  • Bases for Infinite Dimensional Vector Spaces Math 513 Linear Algebra Supplement
    BASES FOR INFINITE DIMENSIONAL VECTOR SPACES MATH 513 LINEAR ALGEBRA SUPPLEMENT Professor Karen E. Smith We have proven that every finitely generated vector space has a basis. But what about vector spaces that are not finitely generated, such as the space of all continuous real valued functions on the interval [0; 1]? Does such a vector space have a basis? By definition, a basis for a vector space V is a linearly independent set which generates V . But we must be careful what we mean by linear combinations from an infinite set of vectors. The definition of a vector space gives us a rule for adding two vectors, but not for adding together infinitely many vectors. By successive additions, such as (v1 + v2) + v3, it makes sense to add any finite set of vectors, but in general, there is no way to ascribe meaning to an infinite sum of vectors in a vector space. Therefore, when we say that a vector space V is generated by or spanned by an infinite set of vectors fv1; v2;::: g, we mean that each vector v in V is a finite linear combination λi1 vi1 + ··· + λin vin of the vi's. Likewise, an infinite set of vectors fv1; v2;::: g is said to be linearly independent if the only finite linear combination of the vi's that is zero is the trivial linear combination. So a set fv1; v2; v3;:::; g is a basis for V if and only if every element of V can be be written in a unique way as a finite linear combination of elements from the set.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • 28. Exterior Powers
    28. Exterior powers 28.1 Desiderata 28.2 Definitions, uniqueness, existence 28.3 Some elementary facts 28.4 Exterior powers Vif of maps 28.5 Exterior powers of free modules 28.6 Determinants revisited 28.7 Minors of matrices 28.8 Uniqueness in the structure theorem 28.9 Cartan's lemma 28.10 Cayley-Hamilton Theorem 28.11 Worked examples While many of the arguments here have analogues for tensor products, it is worthwhile to repeat these arguments with the relevant variations, both for practice, and to be sensitive to the differences. 1. Desiderata Again, we review missing items in our development of linear algebra. We are missing a development of determinants of matrices whose entries may be in commutative rings, rather than fields. We would like an intrinsic definition of determinants of endomorphisms, rather than one that depends upon a choice of coordinates, even if we eventually prove that the determinant is independent of the coordinates. We anticipate that Artin's axiomatization of determinants of matrices should be mirrored in much of what we do here. We want a direct and natural proof of the Cayley-Hamilton theorem. Linear algebra over fields is insufficient, since the introduction of the indeterminate x in the definition of the characteristic polynomial takes us outside the class of vector spaces over fields. We want to give a conceptual proof for the uniqueness part of the structure theorem for finitely-generated modules over principal ideal domains. Multi-linear algebra over fields is surely insufficient for this. 417 418 Exterior powers 2. Definitions, uniqueness, existence Let R be a commutative ring with 1.
    [Show full text]
  • Linear Algebra Handout
    Artificial Intelligence: 6.034 Massachusetts Institute of Technology April 20, 2012 Spring 2012 Recitation 10 Linear Algebra Review • A vector is an ordered list of values. It is often denoted using angle brackets: ha; bi, and its variable name is often written in bold (z) or with an arrow (~z). We can refer to an individual element of a vector using its index: for example, the first element of z would be z1 (or z0, depending on how we're indexing). Each element of a vector generally corresponds to a particular dimension or feature, which could be discrete or continuous; often you can think of a vector as a point in Euclidean space. p 2 2 2 • The magnitude (also called norm) of a vector x = hx1; x2; :::; xni is x1 + x2 + ::: + xn, and is denoted jxj or kxk. • The sum of a set of vectors is their elementwise sum: for example, ha; bi + hc; di = ha + c; b + di (so vectors can only be added if they are the same length). The dot product (also called scalar product) of two vectors is the sum of their elementwise products: for example, ha; bi · hc; di = ac + bd. The dot product x · y is also equal to kxkkyk cos θ, where θ is the angle between x and y. • A matrix is a generalization of a vector: instead of having just one row or one column, it can have m rows and n columns. A square matrix is one that has the same number of rows as columns. A matrix's variable name is generally a capital letter, often written in bold.
    [Show full text]