Bases and Dimension 281
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Introduction to Linear Bialgebra
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of New Mexico University of New Mexico UNM Digital Repository Mathematics and Statistics Faculty and Staff Publications Academic Department Resources 2005 INTRODUCTION TO LINEAR BIALGEBRA Florentin Smarandache University of New Mexico, [email protected] W.B. Vasantha Kandasamy K. Ilanthenral Follow this and additional works at: https://digitalrepository.unm.edu/math_fsp Part of the Algebra Commons, Analysis Commons, Discrete Mathematics and Combinatorics Commons, and the Other Mathematics Commons Recommended Citation Smarandache, Florentin; W.B. Vasantha Kandasamy; and K. Ilanthenral. "INTRODUCTION TO LINEAR BIALGEBRA." (2005). https://digitalrepository.unm.edu/math_fsp/232 This Book is brought to you for free and open access by the Academic Department Resources at UNM Digital Repository. It has been accepted for inclusion in Mathematics and Statistics Faculty and Staff Publications by an authorized administrator of UNM Digital Repository. For more information, please contact [email protected], [email protected], [email protected]. INTRODUCTION TO LINEAR BIALGEBRA W. B. Vasantha Kandasamy Department of Mathematics Indian Institute of Technology, Madras Chennai – 600036, India e-mail: [email protected] web: http://mat.iitm.ac.in/~wbv Florentin Smarandache Department of Mathematics University of New Mexico Gallup, NM 87301, USA e-mail: [email protected] K. Ilanthenral Editor, Maths Tiger, Quarterly Journal Flat No.11, Mayura Park, 16, Kazhikundram Main Road, Tharamani, Chennai – 600 113, India e-mail: [email protected] HEXIS Phoenix, Arizona 2005 1 This book can be ordered in a paper bound reprint from: Books on Demand ProQuest Information & Learning (University of Microfilm International) 300 N. -
HILBERT SPACE GEOMETRY Definition: a Vector Space Over Is a Set V (Whose Elements Are Called Vectors) Together with a Binary Operation
HILBERT SPACE GEOMETRY Definition: A vector space over is a set V (whose elements are called vectors) together with a binary operation +:V×V→V, which is called vector addition, and an external binary operation ⋅: ×V→V, which is called scalar multiplication, such that (i) (V,+) is a commutative group (whose neutral element is called zero vector) and (ii) for all λ,µ∈ , x,y∈V: λ(µx)=(λµ)x, 1 x=x, λ(x+y)=(λx)+(λy), (λ+µ)x=(λx)+(µx), where the image of (x,y)∈V×V under + is written as x+y and the image of (λ,x)∈ ×V under ⋅ is written as λx or as λ⋅x. Exercise: Show that the set 2 together with vector addition and scalar multiplication defined by x y x + y 1 + 1 = 1 1 + x 2 y2 x 2 y2 x λx and λ 1 = 1 , λ x 2 x 2 respectively, is a vector space. 1 Remark: Usually we do not distinguish strictly between a vector space (V,+,⋅) and the set of its vectors V. For example, in the next definition V will first denote the vector space and then the set of its vectors. Definition: If V is a vector space and M⊆V, then the set of all linear combinations of elements of M is called linear hull or linear span of M. It is denoted by span(M). By convention, span(∅)={0}. Proposition: If V is a vector space, then the linear hull of any subset M of V (together with the restriction of the vector addition to M×M and the restriction of the scalar multiplication to ×M) is also a vector space. -
21. Orthonormal Bases
21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else. -
Bases for Infinite Dimensional Vector Spaces Math 513 Linear Algebra Supplement
BASES FOR INFINITE DIMENSIONAL VECTOR SPACES MATH 513 LINEAR ALGEBRA SUPPLEMENT Professor Karen E. Smith We have proven that every finitely generated vector space has a basis. But what about vector spaces that are not finitely generated, such as the space of all continuous real valued functions on the interval [0; 1]? Does such a vector space have a basis? By definition, a basis for a vector space V is a linearly independent set which generates V . But we must be careful what we mean by linear combinations from an infinite set of vectors. The definition of a vector space gives us a rule for adding two vectors, but not for adding together infinitely many vectors. By successive additions, such as (v1 + v2) + v3, it makes sense to add any finite set of vectors, but in general, there is no way to ascribe meaning to an infinite sum of vectors in a vector space. Therefore, when we say that a vector space V is generated by or spanned by an infinite set of vectors fv1; v2;::: g, we mean that each vector v in V is a finite linear combination λi1 vi1 + ··· + λin vin of the vi's. Likewise, an infinite set of vectors fv1; v2;::: g is said to be linearly independent if the only finite linear combination of the vi's that is zero is the trivial linear combination. So a set fv1; v2; v3;:::; g is a basis for V if and only if every element of V can be be written in a unique way as a finite linear combination of elements from the set. -
Does Geometric Algebra Provide a Loophole to Bell's Theorem?
Discussion Does Geometric Algebra provide a loophole to Bell’s Theorem? Richard David Gill 1 1 Leiden University, Faculty of Science, Mathematical Institute; [email protected] Version October 30, 2019 submitted to Entropy Abstract: Geometric Algebra, championed by David Hestenes as a universal language for physics, was used as a framework for the quantum mechanics of interacting qubits by Chris Doran, Anthony Lasenby and others. Independently of this, Joy Christian in 2007 claimed to have refuted Bell’s theorem with a local realistic model of the singlet correlations by taking account of the geometry of space as expressed through Geometric Algebra. A series of papers culminated in a book Christian (2014). The present paper first explores Geometric Algebra as a tool for quantum information and explains why it did not live up to its early promise. In summary, whereas the mapping between 3D geometry and the mathematics of one qubit is already familiar, Doran and Lasenby’s ingenious extension to a system of entangled qubits does not yield new insight but just reproduces standard QI computations in a clumsy way. The tensor product of two Clifford algebras is not a Clifford algebra. The dimension is too large, an ad hoc fix is needed, several are possible. I further analyse two of Christian’s earliest, shortest, least technical, and most accessible works (Christian 2007, 2011), exposing conceptual and algebraic errors. Since 2015, when the first version of this paper was posted to arXiv, Christian has published ambitious extensions of his theory in RSOS (Royal Society - Open Source), arXiv:1806.02392, and in IEEE Access, arXiv:1405.2355. -
Determinants Math 122 Calculus III D Joyce, Fall 2012
Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations. -
28. Exterior Powers
28. Exterior powers 28.1 Desiderata 28.2 Definitions, uniqueness, existence 28.3 Some elementary facts 28.4 Exterior powers Vif of maps 28.5 Exterior powers of free modules 28.6 Determinants revisited 28.7 Minors of matrices 28.8 Uniqueness in the structure theorem 28.9 Cartan's lemma 28.10 Cayley-Hamilton Theorem 28.11 Worked examples While many of the arguments here have analogues for tensor products, it is worthwhile to repeat these arguments with the relevant variations, both for practice, and to be sensitive to the differences. 1. Desiderata Again, we review missing items in our development of linear algebra. We are missing a development of determinants of matrices whose entries may be in commutative rings, rather than fields. We would like an intrinsic definition of determinants of endomorphisms, rather than one that depends upon a choice of coordinates, even if we eventually prove that the determinant is independent of the coordinates. We anticipate that Artin's axiomatization of determinants of matrices should be mirrored in much of what we do here. We want a direct and natural proof of the Cayley-Hamilton theorem. Linear algebra over fields is insufficient, since the introduction of the indeterminate x in the definition of the characteristic polynomial takes us outside the class of vector spaces over fields. We want to give a conceptual proof for the uniqueness part of the structure theorem for finitely-generated modules over principal ideal domains. Multi-linear algebra over fields is surely insufficient for this. 417 418 Exterior powers 2. Definitions, uniqueness, existence Let R be a commutative ring with 1. -
Spanning Sets the Only Algebraic Operations That Are Defined in a Vector Space V Are Those of Addition and Scalar Multiplication
i i “main” 2007/2/16 page 258 i i 258 CHAPTER 4 Vector Spaces 23. Show that the set of all solutions to the nonhomoge- and let neous differential equation S1 + S2 = {v ∈ V : y + a1y + a2y = F(x), v = x + y for some x ∈ S1 and y ∈ S2} . where F(x) is nonzero on an interval I, is not a sub- 2 space of C (I). (a) Show that, in general, S1 ∪ S2 is not a subspace of V . 24. Let S1 and S2 be subspaces of a vector space V . Let (b) Show that S1 ∩ S2 is a subspace of V . S ∪ S ={v ∈ V : v ∈ S or v ∈ S }, 1 2 1 2 (c) Show that S1 + S2 is a subspace of V . S1 ∩ S2 ={v ∈ V : v ∈ S1 and v ∈ S2}, 4.4 Spanning Sets The only algebraic operations that are defined in a vector space V are those of addition and scalar multiplication. Consequently, the most general way in which we can combine the vectors v1, v2,...,vk in V is c1v1 + c2v2 +···+ckvk, (4.4.1) where c1,c2,...,ck are scalars. An expression of the form (4.4.1) is called a linear combination of v1, v2,...,vk. Since V is closed under addition and scalar multiplica- tion, it follows that the foregoing linear combination is itself a vector in V . One of the questions we wish to answer is whether every vector in a vector space can be obtained by taking linear combinations of a finite set of vectors. The following terminology is used in the case when the answer to this question is affirmative: DEFINITION 4.4.1 If every vector in a vector space V can be written as a linear combination of v1, v2, ..., vk, we say that V is spanned or generated by v1, v2, ..., vk and call the set of vectors {v1, v2,...,vk} a spanning set for V . -
Coordinatization
MATH 355 Supplemental Notes Coordinatization Coordinatization In R3, we have the standard basis i, j and k. When we write a vector in coordinate form, say 3 v 2 , (1) “ »´ fi 5 — ffi – fl it is understood as v 3i 2j 5k. “ ´ ` The numbers 3, 2 and 5 are the coordinates of v relative to the standard basis ⇠ i, j, k . It has p´ q “p q always been understood that a coordinate representation such as that in (1) is with respect to the ordered basis ⇠. A little thought reveals that it need not be so. One could have chosen the same basis elements in a di↵erent order, as in the basis ⇠ i, k, j . We employ notation indicating the 1 “p q coordinates are with respect to the di↵erent basis ⇠1: 3 v 5 , to mean that v 3i 5k 2j, r s⇠1 “ » fi “ ` ´ 2 —´ ffi – fl reflecting the order in which the basis elements fall in ⇠1. Of course, one could employ similar notation even when the coordinates are expressed in terms of the standard basis, writing v for r s⇠ (1), but whenever we have coordinatization with respect to the standard basis of Rn in mind, we will consider the wrapper to be optional. r¨s⇠ Of course, there are many non-standard bases of Rn. In fact, any linearly independent collection of n vectors in Rn provides a basis. Say we take 1 1 1 4 ´ ´ » 0fi » 1fi » 1fi » 1fi u , u u u ´ . 1 “ 2 “ 3 “ 4 “ — 3ffi — 1ffi — 0ffi — 2ffi — ffi —´ ffi — ffi — ffi — 0ffi — 4ffi — 2ffi — 1ffi — ffi — ffi — ffi —´ ffi – fl – fl – fl – fl As per the discussion above, these vectors are being expressed relative to the standard basis of R4. -
Review a Basis of a Vector Space 1
Review • Vectors v1 , , v p are linearly dependent if x1 v1 + x2 v2 + + x pv p = 0, and not all the coefficients are zero. • The columns of A are linearly independent each column of A contains a pivot. 1 1 − 1 • Are the vectors 1 , 2 , 1 independent? 1 3 3 1 1 − 1 1 1 − 1 1 1 − 1 1 2 1 0 1 2 0 1 2 1 3 3 0 2 4 0 0 0 So: no, they are dependent! (Coeff’s x3 = 1 , x2 = − 2, x1 = 3) • Any set of 11 vectors in R10 is linearly dependent. A basis of a vector space Definition 1. A set of vectors { v1 , , v p } in V is a basis of V if • V = span{ v1 , , v p} , and • the vectors v1 , , v p are linearly independent. In other words, { v1 , , vp } in V is a basis of V if and only if every vector w in V can be uniquely expressed as w = c1 v1 + + cpvp. 1 0 0 Example 2. Let e = 0 , e = 1 , e = 0 . 1 2 3 0 0 1 3 Show that { e 1 , e 2 , e 3} is a basis of R . It is called the standard basis. Solution. 3 • Clearly, span{ e 1 , e 2 , e 3} = R . • { e 1 , e 2 , e 3} are independent, because 1 0 0 0 1 0 0 0 1 has a pivot in each column. Definition 3. V is said to have dimension p if it has a basis consisting of p vectors. Armin Straub 1 [email protected] This definition makes sense because if V has a basis of p vectors, then every basis of V has p vectors. -
MATH 304 Linear Algebra Lecture 14: Basis and Coordinates. Change of Basis
MATH 304 Linear Algebra Lecture 14: Basis and coordinates. Change of basis. Linear transformations. Basis and dimension Definition. Let V be a vector space. A linearly independent spanning set for V is called a basis. Theorem Any vector space V has a basis. If V has a finite basis, then all bases for V are finite and have the same number of elements (called the dimension of V ). Example. Vectors e1 = (1, 0, 0,..., 0, 0), e2 = (0, 1, 0,..., 0, 0),. , en = (0, 0, 0,..., 0, 1) form a basis for Rn (called standard) since (x1, x2,..., xn) = x1e1 + x2e2 + ··· + xnen. Basis and coordinates If {v1, v2,..., vn} is a basis for a vector space V , then any vector v ∈ V has a unique representation v = x1v1 + x2v2 + ··· + xnvn, where xi ∈ R. The coefficients x1, x2,..., xn are called the coordinates of v with respect to the ordered basis v1, v2,..., vn. The mapping vector v 7→ its coordinates (x1, x2,..., xn) is a one-to-one correspondence between V and Rn. This correspondence respects linear operations in V and in Rn. Examples. • Coordinates of a vector n v = (x1, x2,..., xn) ∈ R relative to the standard basis e1 = (1, 0,..., 0, 0), e2 = (0, 1,..., 0, 0),. , en = (0, 0,..., 0, 1) are (x1, x2,..., xn). a b • Coordinates of a matrix ∈ M2,2(R) c d 1 0 0 0 0 1 relative to the basis , , , 0 0 1 0 0 0 0 0 are (a, c, b, d). 0 1 • Coordinates of a polynomial n−1 p(x) = a0 + a1x + ··· + an−1x ∈Pn relative to 2 n−1 the basis 1, x, x ,..., x are (a0, a1,..., an−1). -
Worksheet 1, for the MATLAB Course by Hans G. Feichtinger, Edinburgh, Jan
Worksheet 1, for the MATLAB course by Hans G. Feichtinger, Edinburgh, Jan. 9th Start MATLAB via: Start > Programs > School applications > Sci.+Eng. > Eng.+Electronics > MATLAB > R2007 (preferably). Generally speaking I suggest that you start your session by opening up a diary with the command: diary mydiary1.m and concluding the session with the command diary off. If you want to save your workspace you my want to call save today1.m in order to save all the current variable (names + values). Moreover, using the HELP command from MATLAB you can get help on more or less every MATLAB command. Try simply help inv or help plot. 1. Define an arbitrary (random) 3 × 3 matrix A, and check whether it is invertible. Calculate the inverse matrix. Then define an arbitrary right hand side vector b and determine the (unique) solution to the equation A ∗ x = b. Find in two different ways the solution to this problem, by either using the inverse matrix, or alternatively by applying Gauss elimination (i.e. the RREF command) to the extended system matrix [A; b]. In addition look at the output of the command rref([A,eye(3)]). What does it tell you? 2. Produce an \arbitrary" 7 × 7 matrix of rank 5. There are at east two simple ways to do this. Either by factorization, i.e. by obtaining it as a product of some 7 × 5 matrix with another random 5 × 7 matrix, or by purposely making two of the rows or columns linear dependent from the remaining ones. Maybe the second method is interesting (if you have time also try the first one): 3.