Review of Linear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators. Vector spaces and linear operators are studied in linear algebra. This lecture reviews definitions from linear algebra that we will need in the rest of the course! Unless stated otherwise, all vector spaces are over C. Recall: Any z 2 C is of the form z = a + ib for some a; b 2 R. Matrices An n × m matrix is an array of numbers aij arranged as follows: 0 1 a11 a12 :::a 1m Ba21 a22 :::a 2m C A = B C B . .. C @ . A an1 an2 :::a nm For any A, we write Aij to denote the entry in the i-th row and the j-th column of A. In the matrix above we have Aij = aij. We denote the set of all n × m complex matrices by Mn;m(C). Basic matrix operations Matrices can be added (if they are of the same shape) and multiplied by a constant (scalar). Both operations are performed entry-wise. If A; B 2 Mn;m(C) and c 2 C then (A + B)ij = Aij + Bij (cA)ij = cAij Example: a a a b b b a + b a + b a + b 11 12 13 + 11 12 13 = 11 11 12 12 13 13 a21 a22 a23 b21 b22 b23 a21 + b21 a22 + b22 a23 + b23 a a a ca ca ca c 11 12 13 = 11 12 13 a21 a22 a23 ca21 ca22 ca23 Conjugate transpose For A 2 Mn;m(C) we define • complex conjugate: A∗ and A¯ ∗ ∗ ∗ ∗ ∗ ∗ a11 a12 a13 a11 a12 a13 (A )ij = (Aij) = ∗ ∗ ∗ a21 a22 a23 a21 a22 a23 • transpose: AT 0 1 T a11 a21 T a11 a12 a13 (A )ij = Aji = @a12 a22A a21 a22 a23 a13 a23 • conjugate transpose(adjoint): Ay = (A∗)T 0 ∗ ∗ 1 y a11 a21 y ∗ a11 a12 a13 ∗ ∗ (A )ij = Aji = @a12 a22A a21 a22 a23 ∗ ∗ a13 a23 Warning: mathematicians often use A∗ to denote Ay. Vectors and inner product Vector is a matrix that has only one row or one column: row vector column vector 0 1 b1 Bb2 C a a :::a B C 1 2 n B . C @ . A bn Just like matrices, vectors of the same shape can be added and multiplied by scalars. They form a vector space denoted Cn. A row vector and a column vector can be multiplied together if they have the same number of entries. The result is a number: 0 1 b1 n Bb2 C X a a :::a · B C = a b + a b + ··· + a b = a b 1 2 n B . C 1 1 2 2 n n i i @ . A i=1 bn This is a special case of matrix product... Matrix multiplication If A is n × m and B is m × l matrix then C = A · B is the n × l matrix with entries given by m X Cik = AijBjk j=1 for all i = 1; : : : ; n and k = 1; : : : ; l. Note that Cik is just the product of the i-th row of A and the k-th column of B: l Note: It is possible to multiply A and B only m B if the number of columns of A agrees with the number of rows of B: m k [n × m] · [m × l] = [n × l] i n A C Properties of matrix multiplication • Associativity: (A · B) · C = A · (B · C) = ABC • Distributivity: A(B + C) = AB + AC and (A + B)C = AB + BC • Scalar multiplication: λ(AB) = (λA)B = A(λB) = (AB)λ • Complex conjugate: (AB)∗ = A∗B∗ • Transpose: (AB)T = BTAT (reversed order!) • Conjugate transpose: (AB)y = ByAy (reversed order!) Warning: Matrix multiplication is not commutative: AB 6= BA. In fact, BA might not even make sense even if AB is well-defined (e.g., when A is 3 × 2 and B is 2 × 1). Dirac bra-ket notation Column vector (ket) Row vector (bra) 0 1 a1 Ba2 C j i = B C h j = a∗ a∗ :::a ∗ B . C 1 2 n @ . A an Note: Ket and bra are conjugate transposes of each other: j iy ≡ h j h jy ≡ j i n For j i; j'i 2 C and A 2 Mn;n(C) the following are valid products: • inner product: h j · j'i ≡ h j'i 2 C • outer product: j i · h'j ≡ j ih'j 2 Mn×n(C) • matrix-vector product: A · j i ≡ Aj i and h j · A ≡ h jA Quiz n Let j i; j'i 2 C and A 2 Mn;n(C). Which of these make sense? (a) j i + h'j (h) j ih'jA (b) j ih'j (i) j iAh'j (c) Ah j (j) h jAj'i (d) j iA (k) h jAj'i + h j'i (e) h jA (l) h j'ih j (f) h jA + h'j (m) h j'iA (g) j ij'i (n) j ih j'i = h j'ij i Inner product The complex number hujvi is called the inner product of jui; jvi 2 Cn. 0 1 0 1 a1 b1 B . C B . C If jui = @ . A and jvi = @ . A then their inner product is an bn 0 1 b1 n X a∗ ··· a∗ B . C ∗ hujvi = 1 n @ . A = ai bi i=1 bn Simple observations: y • hujvi = hvjui = hvjui Pn 2 • hujui = i=1jaij is a real number • hujui ≥ 0 and hujui = 0 iff jui = 0 (the all-zeroes vector) Norm and orthogonality The norm of a vector jui is the non-negative real number q p Pn 2 kjuik = hujui = i=1jaij ≥ 0 Pn 2 A unit vector is a vector with norm 1, i.e., i=1jaij = 1. Recall: qubit states αj0i + βj1i are unit vectors since jαj2 + jβj2 = 1. Cauchy{Schwarz inequality: jhujvij ≤ kjuik kjvik for any jui; jvi 2 Cn. If jui and jvi are unit vectors then jhujvij 2 [0; 1], so jhujvij = cos θ for some angle θ 2 [0; π=2]. Vectors jui and jvi are orthogonal if hujvi = 0 (i.e., θ = π=2). Basis n A basis of C is a minimal collection of vectors jv1i;:::; jvni such that every vector jvi 2 Cn can be expressed as a linear combination of these: jvi = α1jv1i + ··· + αnjvni for some coefficients αi 2 C. In particular, jv1i;:::; jvni are linearly independent, meaning that no jvii can be expressed as a linear combination of the rest. The number n (the size of the basis) is uniquely determined by the vector space and is called its dimension. The dimension of Cn is n. Given a basis, every vector jvi can be uniquely represented as an n-tuple of scalars α1; : : : ; αn, called the coordinates of jvi in this basis. An orthonormal basis is a basis made up of unit vectors that are pairwise orthogonal: ( 1 if i = j hvijvji = δij ≡ 0 if i 6= j Standard basis for Cn The following are all bases for C2 (the last two are orthonormal): 3 4 1 0 1 1 1 1 ; ; p ; p 2 −i 0 1 2 1 2 −1 The standard or computational basis for Cn is 011 001 001 B0C B1C B0C j1i ≡ B C j2i ≡ B C · · · jni ≡ B C B.C B.C B.C @.A @.A @.A 0 0 1 (Sometimes we will use j0i; j1i;:::; jn − 1i instead, e.g., when n = 2.) Any vector can be expanded in the standard basis: 0 1 a1 n B . C X @ . A = aijii i=1 an Outer product and projectors With every pair of vectors jui 2 Cn, jvi 2 Cm we can associate a matrix juihvj 2 Mn;m(C) known as the outer product of jui and jvi. For any jwi 2 Cm, it satisfies the property juihvjjwi = hvjwijui If jui is a unit vector then juihuj is the projector on the one-dimensional subspace generated by jui. This can be easily seen if jui and jvi are real and we set cos θ = hujvi. Then juihujvi = hujvijui. jvi θ jui hujvijui Expanding a matrix in the standard basis 011 001 ··· 01 B0C B00 ··· 0C j1ih2j = B C 01 ··· 0 = B C B.C B. .. .C @.A @. .A 0 00 ··· 0 Similarly, jiihjj has 1 at coordinates (i; j) and zeroes everywhere else. Any n × m matrix can be expressed as a linear combination of outer products: 0 1 a11 a12 :::a 1m n m Ba21 a22 :::a 2m C X X B C = a jiihjj B . .. C ij @ . A i=1 j=1 an1 an2 :::a nm We can recover the (i; j) entry of A as follows: hijAjji = aij Eigenvalues and eigenvectors n An eigenvector of A 2 Mn;n(C) is a non-zero vector jvi 2 C such that Ajvi = λjvi for some complex number λ (the eigenvalue corresponding to jvi). The eigenvalues of A are the roots of the characteristic polynomial: det(A − λI) = 0 where det denotes the determinant and I is the n × n identity matrix: n X I = jiihij i=1 Each n × n matrix has at least one eigenvalue. Diagonal representation Matrix A 2 Mn;n(C) is diagonalisable if n X A = λijviihvij i=1 where jv1i;:::; jvni is an orthonormal set of eigenvectors of A with corresponding eigenvalues λ1; : : : ; λn. This is called spectral decomposition of A. Equivalently, A can be written as a diagonal matrix 0 1 λ1 B .
Recommended publications
  • Common Course Outline MATH 257 Linear Algebra 4 Credits
    Common Course Outline MATH 257 Linear Algebra 4 Credits The Community College of Baltimore County Description MATH 257 – Linear Algebra is one of the suggested elective courses for students majoring in Mathematics, Computer Science or Engineering. Included are geometric vectors, matrices, systems of linear equations, vector spaces, linear transformations, determinants, eigenvectors and inner product spaces. 4 Credits: 5 lecture hours Prerequisite: MATH 251 with a grade of “C” or better Overall Course Objectives Upon successfully completing the course, students will be able to: 1. perform matrix operations; 2. use Gaussian Elimination, Cramer’s Rule, and the inverse of the coefficient matrix to solve systems of Linear Equations; 3. find the inverse of a matrix by Gaussian Elimination or using the adjoint matrix; 4. compute the determinant of a matrix using cofactor expansion or elementary row operations; 5. apply Gaussian Elimination to solve problems concerning Markov Chains; 6. verify that a structure is a vector space by checking the axioms; 7. verify that a subset is a subspace and that a set of vectors is a basis; 8. compute the dimensions of subspaces; 9. compute the matrix representation of a linear transformation; 10. apply notions of linear transformations to discuss rotations and reflections of two dimensional space; 11. compute eigenvalues and find their corresponding eigenvectors; 12. diagonalize a matrix using eigenvalues; 13. apply properties of vectors and dot product to prove results in geometry; 14. apply notions of vectors, dot product and matrices to construct a best fitting curve; 15. construct a solution to real world problems using problem methods individually and in groups; 16.
    [Show full text]
  • Introduction to Linear Bialgebra
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of New Mexico University of New Mexico UNM Digital Repository Mathematics and Statistics Faculty and Staff Publications Academic Department Resources 2005 INTRODUCTION TO LINEAR BIALGEBRA Florentin Smarandache University of New Mexico, [email protected] W.B. Vasantha Kandasamy K. Ilanthenral Follow this and additional works at: https://digitalrepository.unm.edu/math_fsp Part of the Algebra Commons, Analysis Commons, Discrete Mathematics and Combinatorics Commons, and the Other Mathematics Commons Recommended Citation Smarandache, Florentin; W.B. Vasantha Kandasamy; and K. Ilanthenral. "INTRODUCTION TO LINEAR BIALGEBRA." (2005). https://digitalrepository.unm.edu/math_fsp/232 This Book is brought to you for free and open access by the Academic Department Resources at UNM Digital Repository. It has been accepted for inclusion in Mathematics and Statistics Faculty and Staff Publications by an authorized administrator of UNM Digital Repository. For more information, please contact [email protected], [email protected], [email protected]. INTRODUCTION TO LINEAR BIALGEBRA W. B. Vasantha Kandasamy Department of Mathematics Indian Institute of Technology, Madras Chennai – 600036, India e-mail: [email protected] web: http://mat.iitm.ac.in/~wbv Florentin Smarandache Department of Mathematics University of New Mexico Gallup, NM 87301, USA e-mail: [email protected] K. Ilanthenral Editor, Maths Tiger, Quarterly Journal Flat No.11, Mayura Park, 16, Kazhikundram Main Road, Tharamani, Chennai – 600 113, India e-mail: [email protected] HEXIS Phoenix, Arizona 2005 1 This book can be ordered in a paper bound reprint from: Books on Demand ProQuest Information & Learning (University of Microfilm International) 300 N.
    [Show full text]
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Exercise 1 Advanced Sdps Matrices and Vectors
    Exercise 1 Advanced SDPs Matrices and vectors: All these are very important facts that we will use repeatedly, and should be internalized. • Inner product and outer products. Let u = (u1; : : : ; un) and v = (v1; : : : ; vn) be vectors in n T Pn R . Then u v = hu; vi = i=1 uivi is the inner product of u and v. T The outer product uv is an n × n rank 1 matrix B with entries Bij = uivj. The matrix B is a very useful operator. Suppose v is a unit vector. Then, B sends v to u i.e. Bv = uvT v = u, but Bw = 0 for all w 2 v?. m×n n×p • Matrix Product. For any two matrices A 2 R and B 2 R , the standard matrix Pn product C = AB is the m × p matrix with entries cij = k=1 aikbkj. Here are two very useful ways to view this. T Inner products: Let ri be the i-th row of A, or equivalently the i-th column of A , the T transpose of A. Let bj denote the j-th column of B. Then cij = ri cj is the dot product of i-column of AT and j-th column of B. Sums of outer products: C can also be expressed as outer products of columns of A and rows of B. Pn T Exercise: Show that C = k=1 akbk where ak is the k-th column of A and bk is the k-th row of B (or equivalently the k-column of BT ).
    [Show full text]
  • Tensor Products
    Tensor products Joel Kamnitzer April 5, 2011 1 The definition Let V, W, X be three vector spaces. A bilinear map from V × W to X is a function H : V × W → X such that H(av1 + v2, w) = aH(v1, w) + H(v2, w) for v1,v2 ∈ V, w ∈ W, a ∈ F H(v, aw1 + w2) = aH(v, w1) + H(v, w2) for v ∈ V, w1, w2 ∈ W, a ∈ F Let V and W be vector spaces. A tensor product of V and W is a vector space V ⊗ W along with a bilinear map φ : V × W → V ⊗ W , such that for every vector space X and every bilinear map H : V × W → X, there exists a unique linear map T : V ⊗ W → X such that H = T ◦ φ. In other words, giving a linear map from V ⊗ W to X is the same thing as giving a bilinear map from V × W to X. If V ⊗ W is a tensor product, then we write v ⊗ w := φ(v ⊗ w). Note that there are two pieces of data in a tensor product: a vector space V ⊗ W and a bilinear map φ : V × W → V ⊗ W . Here are the main results about tensor products summarized in one theorem. Theorem 1.1. (i) Any two tensor products of V, W are isomorphic. (ii) V, W has a tensor product. (iii) If v1,...,vn is a basis for V and w1,...,wm is a basis for W , then {vi ⊗ wj }1≤i≤n,1≤j≤m is a basis for V ⊗ W .
    [Show full text]
  • On Manifolds of Tensors of Fixed Tt-Rank
    ON MANIFOLDS OF TENSORS OF FIXED TT-RANK SEBASTIAN HOLTZ, THORSTEN ROHWEDDER, AND REINHOLD SCHNEIDER Abstract. Recently, the format of TT tensors [19, 38, 34, 39] has turned out to be a promising new format for the approximation of solutions of high dimensional problems. In this paper, we prove some new results for the TT representation of a tensor U ∈ Rn1×...×nd and for the manifold of tensors of TT-rank r. As a first result, we prove that the TT (or compression) ranks ri of a tensor U are unique and equal to the respective seperation ranks of U if the components of the TT decomposition are required to fulfil a certain maximal rank condition. We then show that d the set T of TT tensors of fixed rank r forms an embedded manifold in Rn , therefore preserving the essential theoretical properties of the Tucker format, but often showing an improved scaling behaviour. Extending a similar approach for matrices [7], we introduce certain gauge conditions to obtain a unique representation of the tangent space TU T of T and deduce a local parametrization of the TT manifold. The parametrisation of TU T is often crucial for an algorithmic treatment of high-dimensional time-dependent PDEs and minimisation problems [33]. We conclude with remarks on those applications and present some numerical examples. 1. Introduction The treatment of high-dimensional problems, typically of problems involving quantities from Rd for larger dimensions d, is still a challenging task for numerical approxima- tion. This is owed to the principal problem that classical approaches for their treatment normally scale exponentially in the dimension d in both needed storage and computa- tional time and thus quickly become computationally infeasable for sensible discretiza- tions of problems of interest.
    [Show full text]
  • Bases for Infinite Dimensional Vector Spaces Math 513 Linear Algebra Supplement
    BASES FOR INFINITE DIMENSIONAL VECTOR SPACES MATH 513 LINEAR ALGEBRA SUPPLEMENT Professor Karen E. Smith We have proven that every finitely generated vector space has a basis. But what about vector spaces that are not finitely generated, such as the space of all continuous real valued functions on the interval [0; 1]? Does such a vector space have a basis? By definition, a basis for a vector space V is a linearly independent set which generates V . But we must be careful what we mean by linear combinations from an infinite set of vectors. The definition of a vector space gives us a rule for adding two vectors, but not for adding together infinitely many vectors. By successive additions, such as (v1 + v2) + v3, it makes sense to add any finite set of vectors, but in general, there is no way to ascribe meaning to an infinite sum of vectors in a vector space. Therefore, when we say that a vector space V is generated by or spanned by an infinite set of vectors fv1; v2;::: g, we mean that each vector v in V is a finite linear combination λi1 vi1 + ··· + λin vin of the vi's. Likewise, an infinite set of vectors fv1; v2;::: g is said to be linearly independent if the only finite linear combination of the vi's that is zero is the trivial linear combination. So a set fv1; v2; v3;:::; g is a basis for V if and only if every element of V can be be written in a unique way as a finite linear combination of elements from the set.
    [Show full text]
  • Determinants Math 122 Calculus III D Joyce, Fall 2012
    Determinants Math 122 Calculus III D Joyce, Fall 2012 What they are. A determinant is a value associated to a square array of numbers, that square array being called a square matrix. For example, here are determinants of a general 2 × 2 matrix and a general 3 × 3 matrix. a b = ad − bc: c d a b c d e f = aei + bfg + cdh − ceg − afh − bdi: g h i The determinant of a matrix A is usually denoted jAj or det (A). You can think of the rows of the determinant as being vectors. For the 3×3 matrix above, the vectors are u = (a; b; c), v = (d; e; f), and w = (g; h; i). Then the determinant is a value associated to n vectors in Rn. There's a general definition for n×n determinants. It's a particular signed sum of products of n entries in the matrix where each product is of one entry in each row and column. The two ways you can choose one entry in each row and column of the 2 × 2 matrix give you the two products ad and bc. There are six ways of chosing one entry in each row and column in a 3 × 3 matrix, and generally, there are n! ways in an n × n matrix. Thus, the determinant of a 4 × 4 matrix is the signed sum of 24, which is 4!, terms. In this general definition, half the terms are taken positively and half negatively. In class, we briefly saw how the signs are determined by permutations.
    [Show full text]
  • 28. Exterior Powers
    28. Exterior powers 28.1 Desiderata 28.2 Definitions, uniqueness, existence 28.3 Some elementary facts 28.4 Exterior powers Vif of maps 28.5 Exterior powers of free modules 28.6 Determinants revisited 28.7 Minors of matrices 28.8 Uniqueness in the structure theorem 28.9 Cartan's lemma 28.10 Cayley-Hamilton Theorem 28.11 Worked examples While many of the arguments here have analogues for tensor products, it is worthwhile to repeat these arguments with the relevant variations, both for practice, and to be sensitive to the differences. 1. Desiderata Again, we review missing items in our development of linear algebra. We are missing a development of determinants of matrices whose entries may be in commutative rings, rather than fields. We would like an intrinsic definition of determinants of endomorphisms, rather than one that depends upon a choice of coordinates, even if we eventually prove that the determinant is independent of the coordinates. We anticipate that Artin's axiomatization of determinants of matrices should be mirrored in much of what we do here. We want a direct and natural proof of the Cayley-Hamilton theorem. Linear algebra over fields is insufficient, since the introduction of the indeterminate x in the definition of the characteristic polynomial takes us outside the class of vector spaces over fields. We want to give a conceptual proof for the uniqueness part of the structure theorem for finitely-generated modules over principal ideal domains. Multi-linear algebra over fields is surely insufficient for this. 417 418 Exterior powers 2. Definitions, uniqueness, existence Let R be a commutative ring with 1.
    [Show full text]
  • A Some Basic Rules of Tensor Calculus
    A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e.
    [Show full text]
  • The Tensor Product and Induced Modules
    The Tensor Product and Induced Modules Nayab Khalid The Tensor Product The Tensor Product A Construction (via the Universal Property) Properties Examples References Nayab Khalid PPS Talk University of St. Andrews October 16, 2015 The Tensor Product and Induced Modules Overview of the Presentation Nayab Khalid The Tensor Product A 1 The Tensor Product Construction Properties Examples 2 A Construction References 3 Properties 4 Examples 5 References The Tensor Product and Induced Modules The Tensor Product Nayab Khalid The Tensor Product A Construction Properties The tensor product of modules is a construction that allows Examples multilinear maps to be carried out in terms of linear maps. References The tensor product can be constructed in many ways, such as using the basis of free modules. However, the standard, more comprehensive, definition of the tensor product stems from category theory and the universal property. The Tensor Product and Induced Modules The Tensor Product (via the Nayab Khalid Universal Property) The Tensor Product A Construction Properties Examples Definition References Let R be a commutative ring. Let E1; E2 and F be modules over R. Consider bilinear maps of the form f : E1 × E2 ! F : The Tensor Product and Induced Modules The Tensor Product (via the Nayab Khalid Universal Property) The Tensor Product Definition A Construction The tensor product E1 ⊗ E2 is the module with a bilinear map Properties Examples λ: E1 × E2 ! E1 ⊗ E2 References such that there exists a unique homomorphism which makes the following diagram commute. The Tensor Product and Induced Modules A Construction of the Nayab Khalid Tensor Product The Tensor Product Here is a more constructive definition of the tensor product.
    [Show full text]
  • Università Degli Studi Di Trieste a Gentle Introduction to Clifford Algebra
    Università degli Studi di Trieste Dipartimento di Fisica Corso di Studi in Fisica Tesi di Laurea Triennale A Gentle Introduction to Clifford Algebra Laureando: Relatore: Daniele Ceravolo prof. Marco Budinich ANNO ACCADEMICO 2015–2016 Contents 1 Introduction 3 1.1 Brief Historical Sketch . 4 2 Heuristic Development of Clifford Algebra 9 2.1 Geometric Product . 9 2.2 Bivectors . 10 2.3 Grading and Blade . 11 2.4 Multivector Algebra . 13 2.4.1 Pseudoscalar and Hodge Duality . 14 2.4.2 Basis and Reciprocal Frames . 14 2.5 Clifford Algebra of the Plane . 15 2.5.1 Relation with Complex Numbers . 16 2.6 Clifford Algebra of Space . 17 2.6.1 Pauli algebra . 18 2.6.2 Relation with Quaternions . 19 2.7 Reflections . 19 2.7.1 Cross Product . 21 2.8 Rotations . 21 2.9 Differentiation . 23 2.9.1 Multivectorial Derivative . 24 2.9.2 Spacetime Derivative . 25 3 Spacetime Algebra 27 3.1 Spacetime Bivectors and Pseudoscalar . 28 3.2 Spacetime Frames . 28 3.3 Relative Vectors . 29 3.4 Even Subalgebra . 29 3.5 Relative Velocity . 30 3.6 Momentum and Wave Vectors . 31 3.7 Lorentz Transformations . 32 3.7.1 Addition of Velocities . 34 1 2 CONTENTS 3.7.2 The Lorentz Group . 34 3.8 Relativistic Visualization . 36 4 Electromagnetism in Clifford Algebra 39 4.1 The Vector Potential . 40 4.2 Electromagnetic Field Strength . 41 4.3 Free Fields . 44 5 Conclusions 47 5.1 Acknowledgements . 48 Chapter 1 Introduction The aim of this thesis is to show how an approach to classical and relativistic physics based on Clifford algebras can shed light on some hidden geometric meanings in our models.
    [Show full text]