Logarithms and Square Roots of Real Matrices Existence, Uniqueness, and Applications in Medical Imaging
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Regular Linear Systems on Cp1 and Their Monodromy Groups V.P
Astérisque V. P. KOSTOV Regular linear systems onCP1 and their monodromy groups Astérisque, tome 222 (1994), p. 259-283 <http://www.numdam.org/item?id=AST_1994__222__259_0> © Société mathématique de France, 1994, tous droits réservés. L’accès aux archives de la collection « Astérisque » (http://smf4.emath.fr/ Publications/Asterisque/) implique l’accord avec les conditions générales d’uti- lisation (http://www.numdam.org/conditions). Toute utilisation commerciale ou impression systématique est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit contenir la présente mention de copyright. Article numérisé dans le cadre du programme Numérisation de documents anciens mathématiques http://www.numdam.org/ REGULAR LINEAR SYSTEMS ON CP1 AND THEIR MONODROMY GROUPS V.P. KOSTOV 1. INTRODUCTION 1.1 A meromorphic linear system of differential equations on CP1 can be pre sented in the form X = A(t)X (1) where A{t) is a meromorphic on CP1 n x n matrix function, " • " = d/dt. Denote its poles ai,... ,ap+i, p > 1. We consider the dependent variable X to be also n x n-matrix. Definition. System (1) is called fuchsian if all the poles of the matrix- function A(t) axe of first order. Definition. System (1) is called regular at the pole a,j if in its neighbour hood the solutions of the system are of moderate growth rate, i.e. IW-a^l^Odt-a^), Ni E R, j = 1,...., p + 1 Here || • || denotes an arbitrary norm in gl(n, C) and we consider a restriction of the solution to a sector with vertex at ctj and of a sufficiently small radius, i.e. -
The Rational and Jordan Forms Linear Algebra Notes
The Rational and Jordan Forms Linear Algebra Notes Satya Mandal November 5, 2005 1 Cyclic Subspaces In a given context, a "cyclic thing" is an one generated "thing". For example, a cyclic groups is a one generated group. Likewise, a module M over a ring R is said to be a cyclic module if M is one generated or M = Rm for some m 2 M: We do not use the expression "cyclic vector spaces" because one generated vector spaces are zero or one dimensional vector spaces. 1.1 (De¯nition and Facts) Suppose V is a vector space over a ¯eld F; with ¯nite dim V = n: Fix a linear operator T 2 L(V; V ): 1. Write R = F[T ] = ff(T ) : f(X) 2 F[X]g L(V; V )g: Then R = F[T ]g is a commutative ring. (We did considered this ring in last chapter in the proof of Caley-Hamilton Theorem.) 2. Now V acquires R¡module structure with scalar multiplication as fol- lows: Define f(T )v = f(T )(v) 2 V 8 f(T ) 2 F[T ]; v 2 V: 3. For an element v 2 V de¯ne Z(v; T ) = F[T ]v = ff(T )v : f(T ) 2 Rg: 1 Note that Z(v; T ) is the cyclic R¡submodule generated by v: (I like the notation F[T ]v, the textbook uses the notation Z(v; T ).) We say, Z(v; T ) is the T ¡cyclic subspace generated by v: 4. If V = Z(v; T ) = F[T ]v; we say that that V is a T ¡cyclic space, and v is called the T ¡cyclic generator of V: (Here, I di®er a little from the textbook.) 5. -
Euler's Square Root Laws for Negative Numbers
Ursinus College Digital Commons @ Ursinus College Transforming Instruction in Undergraduate Complex Numbers Mathematics via Primary Historical Sources (TRIUMPHS) Winter 2020 Euler's Square Root Laws for Negative Numbers Dave Ruch Follow this and additional works at: https://digitalcommons.ursinus.edu/triumphs_complex Part of the Curriculum and Instruction Commons, Educational Methods Commons, Higher Education Commons, and the Science and Mathematics Education Commons Click here to let us know how access to this document benefits ou.y Euler’sSquare Root Laws for Negative Numbers David Ruch December 17, 2019 1 Introduction We learn in elementary algebra that the square root product law pa pb = pab (1) · is valid for any positive real numbers a, b. For example, p2 p3 = p6. An important question · for the study of complex variables is this: will this product law be valid when a and b are complex numbers? The great Leonard Euler discussed some aspects of this question in his 1770 book Elements of Algebra, which was written as a textbook [Euler, 1770]. However, some of his statements drew criticism [Martinez, 2007], as we shall see in the next section. 2 Euler’sIntroduction to Imaginary Numbers In the following passage with excerpts from Sections 139—148of Chapter XIII, entitled Of Impossible or Imaginary Quantities, Euler meant the quantity a to be a positive number. 1111111111111111111111111111111111111111 The squares of numbers, negative as well as positive, are always positive. ...To extract the root of a negative number, a great diffi culty arises; since there is no assignable number, the square of which would be a negative quantity. Suppose, for example, that we wished to extract the root of 4; we here require such as number as, when multiplied by itself, would produce 4; now, this number is neither +2 nor 2, because the square both of 2 and of 2 is +4, and not 4. -
MAT247 Algebra II Assignment 5 Solutions
MAT247 Algebra II Assignment 5 Solutions 1. Find the Jordan canonical form and a Jordan basis for the map or matrix given in each part below. (a) Let V be the real vector space spanned by the polynomials xiyj (in two variables) with i + j ≤ 3. Let T : V V be the map Dx + Dy, where Dx and Dy respectively denote 2 differentiation with respect to x and y. (Thus, Dx(xy) = y and Dy(xy ) = 2xy.) 0 1 1 1 1 1 ! B0 1 2 1 C (b) A = B C over Q @0 0 1 -1A 0 0 0 2 Solution: (a) The operator T is nilpotent so its characterictic polynomial splits and its only eigenvalue is zero and K0 = V. We have Im(T) = spanf1; x; y; x2; xy; y2g; Im(T 2) = spanf1; x; yg Im(T 3) = spanf1g T 4 = 0: Thus the longest cycle for eigenvalue zero has length 4. Moreover, since the number of cycles of length at least r is given by dim(Im(T r-1)) - dim(Im(T r)), the number of cycles of lengths at least 4, 3, 2, 1 is respectively 1, 2, 3, and 4. Thus the number of cycles of lengths 4,3,2,1 is respectively 1,1,1,1. Denoting the r × r Jordan block with λ on the diagonal by Jλ,r, the Jordan canonical form is thus 0 1 J0;4 B J0;3 C J = B C : @ J0;2 A J0;1 We now proceed to find a corresponding Jordan basis, i.e. -
Lecture Notes in One File
Phys 6124 zipped! World Wide Quest to Tame Math Methods Predrag Cvitanovic´ January 2, 2020 Overview The career of a young theoretical physicist consists of treating the harmonic oscillator in ever-increasing levels of abstraction. — Sidney Coleman I am leaving the course notes here, not so much for the notes themselves –they cannot be understood on their own, without the (unrecorded) live lectures– but for the hyperlinks to various source texts you might find useful later on in your research. We change the topics covered year to year, in hope that they reflect better what a graduate student needs to know. This year’s experiment are the two weeks dedicated to data analysis. Let me know whether you would have preferred the more traditional math methods fare, like Bessels and such. If you have taken this course live, you might have noticed a pattern: Every week we start with something obvious that you already know, let mathematics lead us on, and then suddenly end up someplace surprising and highly non-intuitive. And indeed, in the last lecture (that never took place), we turn Coleman on his head, and abandon harmonic potential for the inverted harmonic potential, “spatiotemporal cats” and chaos, arXiv:1912.02940. Contents 1 Linear algebra7 HomeworkHW1 ................................ 7 1.1 Literature ................................ 8 1.2 Matrix-valued functions ......................... 9 1.3 A linear diversion ............................ 12 1.4 Eigenvalues and eigenvectors ...................... 14 1.4.1 Yes, but how do you really do it? . 18 References ................................... 19 exercises 20 2 Eigenvalue problems 23 HomeworkHW2 ................................ 23 2.1 Normal modes .............................. 24 2.2 Stable/unstable manifolds ....................... -
The Geometry of René Descartes
BOOK FIRST The Geometry of René Descartes BOOK I PROBLEMS THE CONSTRUCTION OF WHICH REQUIRES ONLY STRAIGHT LINES AND CIRCLES ANY problem in geometry can easily be reduced to such terms that a knowledge of the lengths of certain straight lines is sufficient for its construction.1 Just as arithmetic consists of only four or five operations, namely, addition, subtraction, multiplication, division and the extraction of roots, which may be considered a kind of division, so in geometry, to find required lines it is merely necessary to add or subtract ther lines; or else, taking one line which I shall call unity in order to relate it as closely as possible to numbers,2 and which can in general be chosen arbitrarily, and having given two other lines, to find a fourth line which shall be to one of the given lines as the other is to unity (which is the same as multiplication) ; or, again, to find a fourth line which is to one of the given lines as unity is to the other (which is equivalent to division) ; or, finally, to find one, two, or several mean proportionals between unity and some other line (which is the same as extracting the square root, cube root, etc., of the given line.3 And I shall not hesitate to introduce these arithmetical terms into geometry, for the sake of greater clearness. For example, let AB be taken as unity, and let it be required to multiply BD by BC. I have only to join the points A and C, and draw DE parallel to CA ; then BE is the product of BD and BC. -
On Multivariate Interpolation
On Multivariate Interpolation Peter J. Olver† School of Mathematics University of Minnesota Minneapolis, MN 55455 U.S.A. [email protected] http://www.math.umn.edu/∼olver Abstract. A new approach to interpolation theory for functions of several variables is proposed. We develop a multivariate divided difference calculus based on the theory of non-commutative quasi-determinants. In addition, intriguing explicit formulae that connect the classical finite difference interpolation coefficients for univariate curves with multivariate interpolation coefficients for higher dimensional submanifolds are established. † Supported in part by NSF Grant DMS 11–08894. April 6, 2016 1 1. Introduction. Interpolation theory for functions of a single variable has a long and distinguished his- tory, dating back to Newton’s fundamental interpolation formula and the classical calculus of finite differences, [7, 47, 58, 64]. Standard numerical approximations to derivatives and many numerical integration methods for differential equations are based on the finite dif- ference calculus. However, historically, no comparable calculus was developed for functions of more than one variable. If one looks up multivariate interpolation in the classical books, one is essentially restricted to rectangular, or, slightly more generally, separable grids, over which the formulae are a simple adaptation of the univariate divided difference calculus. See [19] for historical details. Starting with G. Birkhoff, [2] (who was, coincidentally, my thesis advisor), recent years have seen a renewed level of interest in multivariate interpolation among both pure and applied researchers; see [18] for a fairly recent survey containing an extensive bibli- ography. De Boor and Ron, [8, 12, 13], and Sauer and Xu, [61, 10, 65], have systemati- cally studied the polynomial case. -
The Invertible Matrix Theorem
The Invertible Matrix Theorem Ryan C. Daileda Trinity University Linear Algebra Daileda The Invertible Matrix Theorem Introduction It is important to recognize when a square matrix is invertible. We can now characterize invertibility in terms of every one of the concepts we have now encountered. We will continue to develop criteria for invertibility, adding them to our list as we go. The invertibility of a matrix is also related to the invertibility of linear transformations, which we discuss below. Daileda The Invertible Matrix Theorem Theorem 1 (The Invertible Matrix Theorem) For a square (n × n) matrix A, TFAE: a. A is invertible. b. A has a pivot in each row/column. RREF c. A −−−→ I. d. The equation Ax = 0 only has the solution x = 0. e. The columns of A are linearly independent. f. Null A = {0}. g. A has a left inverse (BA = In for some B). h. The transformation x 7→ Ax is one to one. i. The equation Ax = b has a (unique) solution for any b. j. Col A = Rn. k. A has a right inverse (AC = In for some C). l. The transformation x 7→ Ax is onto. m. AT is invertible. Daileda The Invertible Matrix Theorem Inverse Transforms Definition A linear transformation T : Rn → Rn (also called an endomorphism of Rn) is called invertible iff it is both one-to-one and onto. If [T ] is the standard matrix for T , then we know T is given by x 7→ [T ]x. The Invertible Matrix Theorem tells us that this transformation is invertible iff [T ] is invertible. -
Introduction to Quantum Field Theory
Utrecht Lecture Notes Masters Program Theoretical Physics September 2006 INTRODUCTION TO QUANTUM FIELD THEORY by B. de Wit Institute for Theoretical Physics Utrecht University Contents 1 Introduction 4 2 Path integrals and quantum mechanics 5 3 The classical limit 10 4 Continuous systems 18 5 Field theory 23 6 Correlation functions 36 6.1 Harmonic oscillator correlation functions; operators . 37 6.2 Harmonic oscillator correlation functions; path integrals . 39 6.2.1 Evaluating G0 ............................... 41 6.2.2 The integral over qn ............................ 42 6.2.3 The integrals over q1 and q2 ....................... 43 6.3 Conclusion . 44 7 Euclidean Theory 46 8 Tunneling and instantons 55 8.1 The double-well potential . 57 8.2 The periodic potential . 66 9 Perturbation theory 71 10 More on Feynman diagrams 80 11 Fermionic harmonic oscillator states 87 12 Anticommuting c-numbers 90 13 Phase space with commuting and anticommuting coordinates and quanti- zation 96 14 Path integrals for fermions 108 15 Feynman diagrams for fermions 114 2 16 Regularization and renormalization 117 17 Further reading 127 3 1 Introduction Physical systems that involve an infinite number of degrees of freedom are usually described by some sort of field theory. Almost all systems in nature involve an extremely large number of degrees of freedom. A droplet of water contains of the order of 1026 molecules and while each water molecule can in many applications be described as a point particle, each molecule has itself a complicated structure which reveals itself at molecular length scales. To deal with this large number of degrees of freedom, which for all practical purposes is infinite, we often regard a system as continuous, in spite of the fact that, at certain distance scales, it is discrete. -
Determinants of Commuting-Block Matrices by Istvan Kovacs, Daniel S
Determinants of Commuting-Block Matrices by Istvan Kovacs, Daniel S. Silver*, and Susan G. Williams* Let R beacommutative ring, and Matn(R) the ring of n × n matrices over R.We (i,j) can regard a k × k matrix M =(A ) over Matn(R)asablock matrix,amatrix that has been partitioned into k2 submatrices (blocks)overR, each of size n × n. When M is regarded in this way, we denote its determinant by |M|.Wewill use the symbol D(M) for the determinant of M viewed as a k × k matrix over Matn(R). It is important to realize that D(M)isann × n matrix. Theorem 1. Let R be acommutative ring. Assume that M is a k × k block matrix of (i,j) blocks A ∈ Matn(R) that commute pairwise. Then | | | | (1,π(1)) (2,π(2)) ··· (k,π(k)) (1) M = D(M) = (sgn π)A A A . π∈Sk Here Sk is the symmetric group on k symbols; the summation is the usual one that appears in the definition of determinant. Theorem 1 is well known in the case k =2;the proof is often left as an exercise in linear algebra texts (see [4, page 164], for example). The general result is implicit in [3], but it is not widely known. We present a short, elementary proof using mathematical induction on k.Wesketch a second proof when the ring R has no zero divisors, a proof that is based on [3] and avoids induction by using the fact that commuting matrices over an algebraically closed field can be simultaneously triangularized. -
Matrix Lie Groups
Maths Seminar 2007 MATRIX LIE GROUPS Claudiu C Remsing Dept of Mathematics (Pure and Applied) Rhodes University Grahamstown 6140 26 September 2007 RhodesUniv CCR 0 Maths Seminar 2007 TALK OUTLINE 1. What is a matrix Lie group ? 2. Matrices revisited. 3. Examples of matrix Lie groups. 4. Matrix Lie algebras. 5. A glimpse at elementary Lie theory. 6. Life beyond elementary Lie theory. RhodesUniv CCR 1 Maths Seminar 2007 1. What is a matrix Lie group ? Matrix Lie groups are groups of invertible • matrices that have desirable geometric features. So matrix Lie groups are simultaneously algebraic and geometric objects. Matrix Lie groups naturally arise in • – geometry (classical, algebraic, differential) – complex analyis – differential equations – Fourier analysis – algebra (group theory, ring theory) – number theory – combinatorics. RhodesUniv CCR 2 Maths Seminar 2007 Matrix Lie groups are encountered in many • applications in – physics (geometric mechanics, quantum con- trol) – engineering (motion control, robotics) – computational chemistry (molecular mo- tion) – computer science (computer animation, computer vision, quantum computation). “It turns out that matrix [Lie] groups • pop up in virtually any investigation of objects with symmetries, such as molecules in chemistry, particles in physics, and projective spaces in geometry”. (K. Tapp, 2005) RhodesUniv CCR 3 Maths Seminar 2007 EXAMPLE 1 : The Euclidean group E (2). • E (2) = F : R2 R2 F is an isometry . → | n o The vector space R2 is equipped with the standard Euclidean structure (the “dot product”) x y = x y + x y (x, y R2), • 1 1 2 2 ∈ hence with the Euclidean distance d (x, y) = (y x) (y x) (x, y R2). -
Math 54. Selected Solutions for Week 8 Section 6.1 (Page 282) 22. Let U = U1 U2 U3 . Explain Why U · U ≥ 0. Wh
Math 54. Selected Solutions for Week 8 Section 6.1 (Page 282) 2 3 u1 22. Let ~u = 4 u2 5 . Explain why ~u · ~u ≥ 0 . When is ~u · ~u = 0 ? u3 2 2 2 We have ~u · ~u = u1 + u2 + u3 , which is ≥ 0 because it is a sum of squares (all of which are ≥ 0 ). It is zero if and only if ~u = ~0 . Indeed, if ~u = ~0 then ~u · ~u = 0 , as 2 can be seen directly from the formula. Conversely, if ~u · ~u = 0 then all the terms ui must be zero, so each ui must be zero. This implies ~u = ~0 . 2 5 3 26. Let ~u = 4 −6 5 , and let W be the set of all ~x in R3 such that ~u · ~x = 0 . What 7 theorem in Chapter 4 can be used to show that W is a subspace of R3 ? Describe W in geometric language. The condition ~u · ~x = 0 is equivalent to ~x 2 Nul ~uT , and this is a subspace of R3 by Theorem 2 on page 187. Geometrically, it is the plane perpendicular to ~u and passing through the origin. 30. Let W be a subspace of Rn , and let W ? be the set of all vectors orthogonal to W . Show that W ? is a subspace of Rn using the following steps. (a). Take ~z 2 W ? , and let ~u represent any element of W . Then ~z · ~u = 0 . Take any scalar c and show that c~z is orthogonal to ~u . (Since ~u was an arbitrary element of W , this will show that c~z is in W ? .) ? (b).