James B. Carrell Groups, Matrices, and Vector Spaces a Group Theoretic Approach to Linear Algebra Groups, Matrices, and Vector Spaces James B
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Lecture Notes: Qubit Representations and Rotations
Phys 711 Topics in Particles & Fields | Spring 2013 | Lecture 1 | v0.3 Lecture notes: Qubit representations and rotations Jeffrey Yepez Department of Physics and Astronomy University of Hawai`i at Manoa Watanabe Hall, 2505 Correa Road Honolulu, Hawai`i 96822 E-mail: [email protected] www.phys.hawaii.edu/∼yepez (Dated: January 9, 2013) Contents mathematical object (an abstraction of a two-state quan- tum object) with a \one" state and a \zero" state: I. What is a qubit? 1 1 0 II. Time-dependent qubits states 2 jqi = αj0i + βj1i = α + β ; (1) 0 1 III. Qubit representations 2 A. Hilbert space representation 2 where α and β are complex numbers. These complex B. SU(2) and O(3) representations 2 numbers are called amplitudes. The basis states are or- IV. Rotation by similarity transformation 3 thonormal V. Rotation transformation in exponential form 5 h0j0i = h1j1i = 1 (2a) VI. Composition of qubit rotations 7 h0j1i = h1j0i = 0: (2b) A. Special case of equal angles 7 In general, the qubit jqi in (1) is said to be in a superpo- VII. Example composite rotation 7 sition state of the two logical basis states j0i and j1i. If References 9 α and β are complex, it would seem that a qubit should have four free real-valued parameters (two magnitudes and two phases): I. WHAT IS A QUBIT? iθ0 α φ0 e jqi = = iθ1 : (3) Let us begin by introducing some notation: β φ1 e 1 state (called \minus" on the Bloch sphere) Yet, for a qubit to contain only one classical bit of infor- 0 mation, the qubit need only be unimodular (normalized j1i = the alternate symbol is |−i 1 to unity) α∗α + β∗β = 1: (4) 0 state (called \plus" on the Bloch sphere) 1 Hence it lives on the complex unit circle, depicted on the j0i = the alternate symbol is j+i: 0 top of Figure 1. -
MATH 2370, Practice Problems
MATH 2370, Practice Problems Kiumars Kaveh Problem: Prove that an n × n complex matrix A is diagonalizable if and only if there is a basis consisting of eigenvectors of A. Problem: Let A : V ! W be a one-to-one linear map between two finite dimensional vector spaces V and W . Show that the dual map A0 : W 0 ! V 0 is surjective. Problem: Determine if the curve 2 2 2 f(x; y) 2 R j x + y + xy = 10g is an ellipse or hyperbola or union of two lines. Problem: Show that if a nilpotent matrix is diagonalizable then it is the zero matrix. Problem: Let P be a permutation matrix. Show that P is diagonalizable. Show that if λ is an eigenvalue of P then for some integer m > 0 we have λm = 1 (i.e. λ is an m-th root of unity). Hint: Note that P m = I for some integer m > 0. Problem: Show that if λ is an eigenvector of an orthogonal matrix A then jλj = 1. n Problem: Take a vector v 2 R and let H be the hyperplane orthogonal n n to v. Let R : R ! R be the reflection with respect to a hyperplane H. Prove that R is a diagonalizable linear map. Problem: Prove that if λ1; λ2 are distinct eigenvalues of a complex matrix A then the intersection of the generalized eigenspaces Eλ1 and Eλ2 is zero (this is part of the Spectral Theorem). 1 Problem: Let H = (hij) be a 2 × 2 Hermitian matrix. Use the Min- imax Principle to show that if λ1 ≤ λ2 are the eigenvalues of H then λ1 ≤ h11 ≤ λ2. -
Chapter 7 Block Designs
Chapter 7 Block Designs One simile that solitary shines In the dry desert of a thousand lines. Epilogue to the Satires, ALEXANDER POPE The collection of all subsets of cardinality k of a set with v elements (k < v) has the ³ ´ v¡t property that any subset of t elements, with 0 · t · k, is contained in precisely k¡t subsets of size k. The subsets of size k provide therefore a nice covering for the subsets of a lesser cardinality. Observe that the number of subsets of size k that contain a subset of size t depends only on v; k; and t and not on the specific subset of size t in question. This is the essential defining feature of the structures that we wish to study. The example we just described inspires general interest in producing similar coverings ³ ´ v without using all the k subsets of size k but rather as small a number of them as possi- 1 2 CHAPTER 7. BLOCK DESIGNS ble. The coverings that result are often elegant geometrical configurations, of which the projective and affine planes are examples. These latter configurations form nice coverings only for the subsets of cardinality 2, that is, any two elements are in the same number of these special subsets of size k which we call blocks (or, in certain instances, lines). A collection of subsets of cardinality k, called blocks, with the property that every subset of size t (t · k) is contained in the same number (say ¸) of blocks is called a t-design. We supply the reader with constructions for t-designs with t as high as 5. -
Theory of Angular Momentum and Spin
Chapter 5 Theory of Angular Momentum and Spin Rotational symmetry transformations, the group SO(3) of the associated rotation matrices and the 1 corresponding transformation matrices of spin{ 2 states forming the group SU(2) occupy a very important position in physics. The reason is that these transformations and groups are closely tied to the properties of elementary particles, the building blocks of matter, but also to the properties of composite systems. Examples of the latter with particularly simple transformation properties are closed shell atoms, e.g., helium, neon, argon, the magic number nuclei like carbon, or the proton and the neutron made up of three quarks, all composite systems which appear spherical as far as their charge distribution is concerned. In this section we want to investigate how elementary and composite systems are described. To develop a systematic description of rotational properties of composite quantum systems the consideration of rotational transformations is the best starting point. As an illustration we will consider first rotational transformations acting on vectors ~r in 3-dimensional space, i.e., ~r R3, 2 we will then consider transformations of wavefunctions (~r) of single particles in R3, and finally N transformations of products of wavefunctions like j(~rj) which represent a system of N (spin- Qj=1 zero) particles in R3. We will also review below the well-known fact that spin states under rotations behave essentially identical to angular momentum states, i.e., we will find that the algebraic properties of operators governing spatial and spin rotation are identical and that the results derived for products of angular momentum states can be applied to products of spin states or a combination of angular momentum and spin states. -
Lecture 2: Spectral Theorems
Lecture 2: Spectral Theorems This lecture introduces normal matrices. The spectral theorem will inform us that normal matrices are exactly the unitarily diagonalizable matrices. As a consequence, we will deduce the classical spectral theorem for Hermitian matrices. The case of commuting families of matrices will also be studied. All of this corresponds to section 2.5 of the textbook. 1 Normal matrices Definition 1. A matrix A 2 Mn is called a normal matrix if AA∗ = A∗A: Observation: The set of normal matrices includes all the Hermitian matrices (A∗ = A), the skew-Hermitian matrices (A∗ = −A), and the unitary matrices (AA∗ = A∗A = I). It also " # " # 1 −1 1 1 contains other matrices, e.g. , but not all matrices, e.g. 1 1 0 1 Here is an alternate characterization of normal matrices. Theorem 2. A matrix A 2 Mn is normal iff ∗ n kAxk2 = kA xk2 for all x 2 C : n Proof. If A is normal, then for any x 2 C , 2 ∗ ∗ ∗ ∗ ∗ 2 kAxk2 = hAx; Axi = hx; A Axi = hx; AA xi = hA x; A xi = kA xk2: ∗ n n Conversely, suppose that kAxk = kA xk for all x 2 C . For any x; y 2 C and for λ 2 C with jλj = 1 chosen so that <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij, we expand both sides of 2 ∗ 2 kA(λx + y)k2 = kA (λx + y)k2 to obtain 2 2 ∗ 2 ∗ 2 ∗ ∗ kAxk2 + kAyk2 + 2<(λhAx; Ayi) = kA xk2 + kA yk2 + 2<(λhA x; A yi): 2 ∗ 2 2 ∗ 2 Using the facts that kAxk2 = kA xk2 and kAyk2 = kA yk2, we derive 0 = <(λhAx; Ayi − λhA∗x; A∗yi) = <(λhx; A∗Ayi − λhx; AA∗yi) = <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij: n ∗ ∗ n Since this is true for any x 2 C , we deduce (A A − AA )y = 0, which holds for any y 2 C , meaning that A∗A − AA∗ = 0, as desired. -
New Foundations for Geometric Algebra1
Text published in the electronic journal Clifford Analysis, Clifford Algebras and their Applications vol. 2, No. 3 (2013) pp. 193-211 New foundations for geometric algebra1 Ramon González Calvet Institut Pere Calders, Campus Universitat Autònoma de Barcelona, 08193 Cerdanyola del Vallès, Spain E-mail : [email protected] Abstract. New foundations for geometric algebra are proposed based upon the existing isomorphisms between geometric and matrix algebras. Each geometric algebra always has a faithful real matrix representation with a periodicity of 8. On the other hand, each matrix algebra is always embedded in a geometric algebra of a convenient dimension. The geometric product is also isomorphic to the matrix product, and many vector transformations such as rotations, axial symmetries and Lorentz transformations can be written in a form isomorphic to a similarity transformation of matrices. We collect the idea Dirac applied to develop the relativistic electron equation when he took a basis of matrices for the geometric algebra instead of a basis of geometric vectors. Of course, this way of understanding the geometric algebra requires new definitions: the geometric vector space is defined as the algebraic subspace that generates the rest of the matrix algebra by addition and multiplication; isometries are simply defined as the similarity transformations of matrices as shown above, and finally the norm of any element of the geometric algebra is defined as the nth root of the determinant of its representative matrix of order n. The main idea of this proposal is an arithmetic point of view consisting of reversing the roles of matrix and geometric algebras in the sense that geometric algebra is a way of accessing, working and understanding the most fundamental conception of matrix algebra as the algebra of transformations of multiple quantities. -
Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation -
Understanding the Spreading Power of All Nodes in a Network: a Continuous-Time Perspective
i i “Lawyer-mainmanus” — 2014/6/12 — 9:39 — page 1 — #1 i i Understanding the spreading power of all nodes in a network: a continuous-time perspective Glenn Lawyer ∗ ∗Max Planck Institute for Informatics, Saarbr¨ucken, Germany Submitted to Proceedings of the National Academy of Sciences of the United States of America Centrality measures such as the degree, k-shell, or eigenvalue central- epidemic. When this ratio exceeds this critical range, the dy- ity can identify a network’s most influential nodes, but are rarely use- namics approach the SI system as a limiting case. fully accurate in quantifying the spreading power of the vast majority Several approaches to quantifying the spreading power of of nodes which are not highly influential. The spreading power of all all nodes have recently been proposed, including the accessibil- network nodes is better explained by considering, from a continuous- ity [16, 17], the dynamic influence [11], and the impact [7] (See time epidemiological perspective, the distribution of the force of in- fection each node generates. The resulting metric, the Expected supplementary Note S1). These extend earlier approaches to Force (ExF), accurately quantifies node spreading power under all measuring centrality by explicitly incorporating spreading dy- primary epidemiological models across a wide range of archetypi- namics, and have been shown to be both distinct from previous cal human contact networks. When node power is low, influence is centrality measures and more highly correlated with epidemic a function of neighbor degree. As power increases, a node’s own outcomes [7, 11, 18]. Yet they retain the common foundation degree becomes more important. -
Unipotent Flows and Applications
Clay Mathematics Proceedings Volume 10, 2010 Unipotent Flows and Applications Alex Eskin 1. General introduction 1.1. Values of indefinite quadratic forms at integral points. The Op- penheim Conjecture. Let X Q(x1; : : : ; xn) = aijxixj 1≤i≤j≤n be a quadratic form in n variables. We always assume that Q is indefinite so that (so that there exists p with 1 ≤ p < n so that after a linear change of variables, Q can be expresses as: Xp Xn ∗ 2 − 2 Qp(y1; : : : ; yn) = yi yi i=1 i=p+1 We should think of the coefficients aij of Q as real numbers (not necessarily rational or integer). One can still ask what will happen if one substitutes integers for the xi. It is easy to see that if Q is a multiple of a form with rational coefficients, then the set of values Q(Zn) is a discrete subset of R. Much deeper is the following conjecture: Conjecture 1.1 (Oppenheim, 1929). Suppose Q is not proportional to a ra- tional form and n ≥ 5. Then Q(Zn) is dense in the real line. This conjecture was extended by Davenport to n ≥ 3. Theorem 1.2 (Margulis, 1986). The Oppenheim Conjecture is true as long as n ≥ 3. Thus, if n ≥ 3 and Q is not proportional to a rational form, then Q(Zn) is dense in R. This theorem is a triumph of ergodic theory. Before Margulis, the Oppenheim Conjecture was attacked by analytic number theory methods. (In particular it was known for n ≥ 21, and for diagonal forms with n ≥ 5). -
Similarity Preserving Linear Maps on Upper Triangular Matrix Algebras ୋ
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Linear Algebra and its Applications 414 (2006) 278–287 www.elsevier.com/locate/laa Similarity preserving linear maps on upper triangular matrix algebras ୋ Guoxing Ji ∗, Baowei Wu College of Mathematics and Information Science, Shaanxi Normal University, Xian 710062, PR China Received 14 June 2005; accepted 5 October 2005 Available online 1 December 2005 Submitted by P. Šemrl Abstract In this note we consider similarity preserving linear maps on the algebra of all n × n complex upper triangular matrices Tn. We give characterizations of similarity invariant subspaces in Tn and similarity preserving linear injections on Tn. Furthermore, we considered linear injections on Tn preserving similarity in Mn as well. © 2005 Elsevier Inc. All rights reserved. AMS classification: 47B49; 15A04 Keywords: Matrix; Upper triangular matrix; Similarity invariant subspace; Similarity preserving linear map 1. Introduction In the last few decades, many researchers have considered linear preserver problems on matrix or operator spaces. For example, there are many research works on linear maps which preserve spectrum (cf. [4,5,11]), rank (cf. [3]), nilpotency (cf. [10]) and similarity (cf. [2,6–8]) and so on. Many useful techniques have been developed; see [1,9] for some general techniques and background. Hiai [2] gave a characterization of the similarity preserving linear map on the algebra of all complex n × n matrices. ୋ This research was supported by the National Natural Science Foundation of China (no. 10571114), the Natural Science Basic Research Plan in Shaanxi Province of China (program no. -
Hadamard and Conference Matrices
Hadamard and conference matrices Peter J. Cameron December 2011 with input from Dennis Lin, Will Orrick and Gordon Royle Now det(H) is equal to the volume of the n-dimensional parallelepiped spanned by the rows of H. By assumption, each row has Euclidean length at most n1/2, so that det(H) ≤ nn/2; equality holds if and only if I every entry of H is ±1; > I the rows of H are orthogonal, that is, HH = nI. A matrix attaining the bound is a Hadamard matrix. Hadamard's theorem Let H be an n × n matrix, all of whose entries are at most 1 in modulus. How large can det(H) be? A matrix attaining the bound is a Hadamard matrix. Hadamard's theorem Let H be an n × n matrix, all of whose entries are at most 1 in modulus. How large can det(H) be? Now det(H) is equal to the volume of the n-dimensional parallelepiped spanned by the rows of H. By assumption, each row has Euclidean length at most n1/2, so that det(H) ≤ nn/2; equality holds if and only if I every entry of H is ±1; > I the rows of H are orthogonal, that is, HH = nI. Hadamard's theorem Let H be an n × n matrix, all of whose entries are at most 1 in modulus. How large can det(H) be? Now det(H) is equal to the volume of the n-dimensional parallelepiped spanned by the rows of H. By assumption, each row has Euclidean length at most n1/2, so that det(H) ≤ nn/2; equality holds if and only if I every entry of H is ±1; > I the rows of H are orthogonal, that is, HH = nI. -
Linear Operators Hsiu-Hau Lin [email protected] (Mar 25, 2010)
Linear Operators Hsiu-Hau Lin [email protected] (Mar 25, 2010) The notes cover linear operators and discuss linear independence of func- tions (Boas 3.7-3.8). • Linear operators An operator maps one thing into another. For instance, the ordinary func- tions are operators mapping numbers to numbers. A linear operator satisfies the properties, O(A + B) = O(A) + O(B);O(kA) = kO(A); (1) where k is a number. As we learned before, a matrix maps one vector into another. One also notices that M(r1 + r2) = Mr1 + Mr2;M(kr) = kMr: Thus, matrices are linear operators. • Orthogonal matrix The length of a vector remains invariant under rotations, ! ! x0 x x0 y0 = x y M T M : y0 y The constraint can be elegantly written down as a matrix equation, M T M = MM T = 1: (2) In other words, M T = M −1. For matrices satisfy the above constraint, they are called orthogonal matrices. Note that, for orthogonal matrices, computing inverse is as simple as taking transpose { an extremely helpful property for calculations. From the product theorem for the determinant, we immediately come to the conclusion det M = ±1. In two dimensions, any 2 × 2 orthogonal matrix with determinant 1 corresponds to a rotation, while any 2 × 2 orthogonal HedgeHog's notes (March 24, 2010) 2 matrix with determinant −1 corresponds to a reflection about a line. Let's come back to our good old friend { the rotation matrix, cos θ − sin θ ! cos θ sin θ ! R(θ) = ;RT = : (3) sin θ cos θ − sin θ cos θ It is straightforward to check that RT R = RRT = 1.