The Monodromy of Meromorphic Projective Structures

Total Page:16

File Type:pdf, Size:1020Kb

The Monodromy of Meromorphic Projective Structures THE MONODROMY OF MEROMORPHIC PROJECTIVE STRUCTURES DYLAN G.L. ALLEGRETTI AND TOM BRIDGELAND Abstract. We study projective structures on a surface having poles of prescribed orders. We obtain a monodromy map from a complex manifold parameterising such structures to the stack of framed PGL2(C) local systems on the associated marked bordered surface. We prove that the image of this map is contained in the union of the domains of the cluster charts. We discuss a number of open questions concerning this monodromy map. 1. Introduction This paper is concerned with the monodromy of projective structures on Riemann surfaces. A projective structure can be viewed as a global generalization of a differential equation of the form y00(z) − '(z) · y(z) = 0; where primes denote differentiation with respect to z. In another language it is an oper for the group PGL2(C). Our main focus will be on the case when the potential '(z) has poles, but we begin by recalling some of the classical results on holomorphic projective structures. For excellent surveys on this material we recommend [17, 36, 43]. 1.1. Holomorphic projective structures. A projective structure P on a Riemann surface S is an atlas of holomorphic charts 1 [ fi : Ui ! P ;S = Ui; i2I −1 with the property that each transition function gij = fi ◦ fj is the restriction of an element of 1 G = Aut(P ) = PGL2(C): Projective structures are closely related to quadratic differentials. If P is a projective structure on a Riemann surface S, with a local chart z : U ! P1, and φ(z) = '(z) dz⊗2 (1) is a quadratic differential on S, written in terms of the co-ordinate z, then one obtains a chart in a new projective structure P0 = P + φ on the surface S by considering the ratio of two independent 1 2 DYLAN G.L. ALLEGRETTI AND TOM BRIDGELAND solutions to the differential equation y00(z) − '(z) · y(z) = 0: (2) This construction gives the set of projective structures on a fixed compact Riemann surface S of genus g = g(S) the structure of an affine space for the vector space of holomorphic quadratic differentials 0 ⊗2 ∼ 3g−3 H (S; !S ) = C : (3) Let us fix a closed, oriented surface S of genus g = g(S). A marked projective structure is then defined to be a triple (S; P; θ), where S is a Riemann surface equipped with a projective structure P, and θ is a marking, that is, an isotopy class of orientation-preserving diffeomorphisms θ : S ! S. Two such triples (Si; Pi; θi) will be considered to be equivalent if there is a biholomorphism f : S1 ! S2 which preserves the projective structures and the markings in the obvious way. The set P(S) of equivalence classes of marked projective structures has the natural structure of a complex manifold of dimension 6g − 6. There is an obvious forgetful map p: P(S) ! T(S) (4) to the Teichm¨ullerspace T(S), viewed as the moduli space of Riemann surfaces S equipped with a marking θ : S ! S. A relative version of the construction described above shows that this map (4) is an affine bundle for the vector bundle q : Q(S) ! T(S); whose fibre over a marked Riemann surface (S; θ) is the vector space (3). 1.2. Monodromy of projective structures. A projective structure P on a Riemann surface S has an associated developing map: a holomorphic map 1 f : S~ ! P ; where π : S~ ! S is the universal covering surface, such that any injective locally-defined map of the form f ◦ π−1 is a chart in P. Such a developing map gives rise to a monodromy representation ρ: π1(S) ! G; (5) defined by the the condition f(γ · x) = ρ(γ) · f(x). The developing map f is unique up to post-composition with an element of the group G, and the monodromy representation is therefore well-defined up to overall conjugation by an element of G. More abstractly, we can think in terms of a G local system naturally associated to the projective structure. THE MONODROMY OF MEROMORPHIC PROJECTIVE STRUCTURES 3 Let us fix a closed, oriented surface S as above and consider the quotient stack X(S) = Hom(π1(S);G)=G (6) parameterising representations (5) up to overall conjugation, or equivalently, isomorphism classes of G local systems on S. The monodromy of a marked projective structure on S defines a point of this stack in the obvious way. Gunning [26] proved that when g = g(S) > 1, this point lies in the open substack X∗(S) consisting of representations with non-commutative image. This substack is a (possibly non-Hausdorff) complex manifold, and there is a holomorphic map ∗ F : P(S) ! X (S) (7) sending a marked projective structure to its monodromy representation. The map F has been studied for over a century. Let us briefly recall some of the better known results. A result of Poincar´e[39] shows that F is injective when restricted to each fibre of the forgetful map (4). Hejhal [27] proved that F is a local homeomorphism, and Earle [18] and Hubbard [28] showed that F is moreover a local biholomorphism. It is also known that F has infinite fibres and is not a covering map of its image, see e.g. [17]. Finally, a famous theorem of Gallo, Kapovich and Marden [25] characterises the image of F . 1.3. Meromorphic projective structures. In this paper we study a monodromy map analogous to (7) but for meromorphic projective structures. The notion of a meromorphic projective structure has meaning because of the above-mentioned fact that the set of projective structures on a fixed Riemann surface S is an affine space for the space of quadratic differentials. Concretely, if we fix a holomorphic projective structure on S as above, the local charts in a meromorphic projective structure are obtained by taking ratios of solutions to the equation (2), where the quadratic differential (1) is now allowed to have poles. Note that when '(z) has a pole p of order m > 2 the equation (2) has an irregular singularity, and one should expect generalised monodromy in the form of Stokes data to enter the picture. A solution y(z) to the equation (2) defined in a sectorial neighbourhood centered at p is called subdominant if y(z) ! 0 as z ! p. Standard results in the theory of differential equations show that there are m − 2 distinguished sectors centered at p, known as Stokes sectors, in which there exist unique-up-to-scale subdominant solutions to (2). The rays in the centres of these distinguished sectors are called Stokes directions and coincide with the asymptotic directions at p of the horizontal foliation defined by the quadratic differential φ(z). 4 DYLAN G.L. ALLEGRETTI AND TOM BRIDGELAND Our results depend on a choice of genus g ≥ 0 and a non-empty collection of positive integers giving the orders of poles of the projective structures. It is more convenient to represent this data in the form of a marked bordered surface (S; M). This is a compact, connected, oriented surface S with (possibly empty) boundary, equipped with a non-empty collection of marked points M ⊂ S, such that each boundary component of S contains at least one point of M. We denote by P ⊂ M the set of internal marked points, which we also refer to as punctures. A meromorphic projective structure P on a compact Riemann surface S naturally determines a marked bordered surface (S; M). The surface S is the real oriented blow-up of S at the poles of P of order > 2, which are precisely the irregular singularities of the differential equation (2), and the boundary marked points correspond to the Stokes directions. The internal marked points are the poles of P of order ≤ 2, which are precisely the regular singularities of the equation (2). Let us fix a marked bordered surface (S; M). If g(S) = 0 we assume that jMj ≥ 3. In Section8 we introduce a complex manifold P(S; M) parameterising marked meromorphic projective structures in much the same way as before. The points of P(S; M) are equivalence-classes of triples (S; P; θ), where S is a Riemann surface equipped with a meromorphic projective structure P, and θ is an isotopy class of orientation-preserving diffeomorphisms between the marked bordered surface canonically associated to (S; P) and the fixed surface (S; M). It will be convenient to introduce two modifications to the manifold P(S; M). Firstly, we define a dense open subset ◦ P (S; M) ⊂ P(S; M); whose complement consists of projective structures with apparent singularities: regular singularities for which the corresponding monodromy transformation is trivial as an element of G = PGL2(C). We do this because the monodromy map of Theorem 1.1 is not well-defined for such projective structures. Secondly, we introduce a branched cover ∗ ◦ π : P (S; M) ! P (S; M) (8) of degree 2jPj, whose points parameterise marked projective structures equipped with a signing: a choice of eigenline for the monodromy around each regular singularity. The main point of this is to obtain a monodromy map taking values in the stack of framed local systems, which as we explain below, is rational and carries interesting birational co-ordinate systems. An analogous cover also turns out to be very natural in the context of moduli spaces of meromorphic quadratic differentials [14, Section 6.2]. THE MONODROMY OF MEROMORPHIC PROJECTIVE STRUCTURES 5 1.4. Monodromy and framed local systems. Let us now turn to the analogue of the character stack (6) in the meromorphic situation.
Recommended publications
  • CONSTRUCTING INTEGER MATRICES with INTEGER EIGENVALUES CHRISTOPHER TOWSE,∗ Scripps College
    Applied Probability Trust (25 March 2016) CONSTRUCTING INTEGER MATRICES WITH INTEGER EIGENVALUES CHRISTOPHER TOWSE,∗ Scripps College ERIC CAMPBELL,∗∗ Pomona College Abstract In spite of the proveable rarity of integer matrices with integer eigenvalues, they are commonly used as examples in introductory courses. We present a quick method for constructing such matrices starting with a given set of eigenvectors. The main feature of the method is an added level of flexibility in the choice of allowable eigenvalues. The method is also applicable to non-diagonalizable matrices, when given a basis of generalized eigenvectors. We have produced an online web tool that implements these constructions. Keywords: Integer matrices, Integer eigenvalues 2010 Mathematics Subject Classification: Primary 15A36; 15A18 Secondary 11C20 In this paper we will look at the problem of constructing a good problem. Most linear algebra and introductory ordinary differential equations classes include the topic of diagonalizing matrices: given a square matrix, finding its eigenvalues and constructing a basis of eigenvectors. In the instructional setting of such classes, concrete “toy" examples are helpful and perhaps even necessary (at least for most students). The examples that are typically given to students are, of course, integer-entry matrices with integer eigenvalues. Sometimes the eigenvalues are repeated with multipicity, sometimes they are all distinct. Oftentimes, the number 0 is avoided as an eigenvalue due to the degenerate cases it produces, particularly when the matrix in question comes from a linear system of differential equations. Yet in [10], Martin and Wong show that “Almost all integer matrices have no integer eigenvalues," let alone all integer eigenvalues.
    [Show full text]
  • Solutions to Math 53 Practice Second Midterm
    Solutions to Math 53 Practice Second Midterm 1. (20 points) dx dy (a) (10 points) Write down the general solution to the system of equations dt = x+y; dt = −13x−3y in terms of real-valued functions. (b) (5 points) The trajectories of this equation in the (x; y)-plane rotate around the origin. Is this dx rotation clockwise or counterclockwise? If we changed the first equation to dt = −x + y, would it change the direction of rotation? (c) (5 points) Suppose (x1(t); y1(t)) and (x2(t); y2(t)) are two solutions to this equation. Define the Wronskian determinant W (t) of these two solutions. If W (0) = 1, what is W (10)? Note: the material for this part will be covered in the May 8 and May 10 lectures. 1 1 (a) The corresponding matrix is A = , with characteristic polynomial (1 − λ)(−3 − −13 −3 λ) + 13 = 0, so that λ2 + 2λ + 10 = 0, so that λ = −1 ± 3i. The corresponding eigenvector v to λ1 = −1 + 3i must be a solution to 2 − 3i 1 −13 −2 − 3i 1 so we can take v = Now we just need to compute the real and imaginary parts of 3i − 2 e3it cos(3t) + i sin(3t) veλ1t = e−t = e−t : (3i − 2)e3it −2 cos(3t) − 3 sin(3t) + i(3 cos(3t) − 2 sin(3t)) The general solution is thus expressed as cos(3t) sin(3t) c e−t + c e−t : 1 −2 cos(3t) − 3 sin(3t) 2 (3 cos(3t) − 2 sin(3t)) dx (b) We can examine the direction field.
    [Show full text]
  • Fast Computation of the Rank Profile Matrix and the Generalized Bruhat Decomposition Jean-Guillaume Dumas, Clement Pernet, Ziad Sultan
    Fast Computation of the Rank Profile Matrix and the Generalized Bruhat Decomposition Jean-Guillaume Dumas, Clement Pernet, Ziad Sultan To cite this version: Jean-Guillaume Dumas, Clement Pernet, Ziad Sultan. Fast Computation of the Rank Profile Matrix and the Generalized Bruhat Decomposition. Journal of Symbolic Computation, Elsevier, 2017, Special issue on ISSAC’15, 83, pp.187-210. 10.1016/j.jsc.2016.11.011. hal-01251223v1 HAL Id: hal-01251223 https://hal.archives-ouvertes.fr/hal-01251223v1 Submitted on 5 Jan 2016 (v1), last revised 14 May 2018 (v2) HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Fast Computation of the Rank Profile Matrix and the Generalized Bruhat Decomposition Jean-Guillaume Dumas Universit´eGrenoble Alpes, Laboratoire LJK, umr CNRS, BP53X, 51, av. des Math´ematiques, F38041 Grenoble, France Cl´ement Pernet Universit´eGrenoble Alpes, Laboratoire de l’Informatique du Parall´elisme, Universit´ede Lyon, France. Ziad Sultan Universit´eGrenoble Alpes, Laboratoire LJK and LIG, Inria, CNRS, Inovall´ee, 655, av. de l’Europe, F38334 St Ismier Cedex, France Abstract The row (resp. column) rank profile of a matrix describes the stair-case shape of its row (resp.
    [Show full text]
  • Matrix Lie Groups
    Maths Seminar 2007 MATRIX LIE GROUPS Claudiu C Remsing Dept of Mathematics (Pure and Applied) Rhodes University Grahamstown 6140 26 September 2007 RhodesUniv CCR 0 Maths Seminar 2007 TALK OUTLINE 1. What is a matrix Lie group ? 2. Matrices revisited. 3. Examples of matrix Lie groups. 4. Matrix Lie algebras. 5. A glimpse at elementary Lie theory. 6. Life beyond elementary Lie theory. RhodesUniv CCR 1 Maths Seminar 2007 1. What is a matrix Lie group ? Matrix Lie groups are groups of invertible • matrices that have desirable geometric features. So matrix Lie groups are simultaneously algebraic and geometric objects. Matrix Lie groups naturally arise in • – geometry (classical, algebraic, differential) – complex analyis – differential equations – Fourier analysis – algebra (group theory, ring theory) – number theory – combinatorics. RhodesUniv CCR 2 Maths Seminar 2007 Matrix Lie groups are encountered in many • applications in – physics (geometric mechanics, quantum con- trol) – engineering (motion control, robotics) – computational chemistry (molecular mo- tion) – computer science (computer animation, computer vision, quantum computation). “It turns out that matrix [Lie] groups • pop up in virtually any investigation of objects with symmetries, such as molecules in chemistry, particles in physics, and projective spaces in geometry”. (K. Tapp, 2005) RhodesUniv CCR 3 Maths Seminar 2007 EXAMPLE 1 : The Euclidean group E (2). • E (2) = F : R2 R2 F is an isometry . → | n o The vector space R2 is equipped with the standard Euclidean structure (the “dot product”) x y = x y + x y (x, y R2), • 1 1 2 2 ∈ hence with the Euclidean distance d (x, y) = (y x) (y x) (x, y R2).
    [Show full text]
  • Quantum Cluster Algebras and Quantum Nilpotent Algebras
    Quantum cluster algebras and quantum nilpotent algebras K. R. Goodearl ∗ and M. T. Yakimov † ∗Department of Mathematics, University of California, Santa Barbara, CA 93106, U.S.A., and †Department of Mathematics, Louisiana State University, Baton Rouge, LA 70803, U.S.A. Proceedings of the National Academy of Sciences of the United States of America 111 (2014) 9696–9703 A major direction in the theory of cluster algebras is to construct construction does not rely on any initial combinatorics of the (quantum) cluster algebra structures on the (quantized) coordinate algebras. On the contrary, the construction itself produces rings of various families of varieties arising in Lie theory. We prove intricate combinatorial data for prime elements in chains of that all algebras in a very large axiomatically defined class of noncom- subalgebras. When this is applied to special cases, we re- mutative algebras possess canonical quantum cluster algebra struc- cover the Weyl group combinatorics which played a key role tures. Furthermore, they coincide with the corresponding upper quantum cluster algebras. We also establish analogs of these results in categorification earlier [7, 9, 10]. Because of this, we expect for a large class of Poisson nilpotent algebras. Many important fam- that our construction will be helpful in building a unified cat- ilies of coordinate rings are subsumed in the class we are covering, egorification of quantum nilpotent algebras. Finally, we also which leads to a broad range of application of the general results prove similar results for (commutative) cluster algebras using to the above mentioned types of problems. As a consequence, we Poisson prime elements.
    [Show full text]
  • Ratner's Work on Unipotent Flows and Impact
    Ratner’s Work on Unipotent Flows and Its Impact Elon Lindenstrauss, Peter Sarnak, and Amie Wilkinson Dani above. As the name suggests, these theorems assert that the closures, as well as related features, of the orbits of such flows are very restricted (rigid). As such they provide a fundamental and powerful tool for problems connected with these flows. The brilliant techniques that Ratner in- troduced and developed in establishing this rigidity have been the blueprint for similar rigidity theorems that have been proved more recently in other contexts. We begin by describing the setup for the group of 푑×푑 matrices with real entries and determinant equal to 1 — that is, SL(푑, ℝ). An element 푔 ∈ SL(푑, ℝ) is unipotent if 푔−1 is a nilpotent matrix (we use 1 to denote the identity element in 퐺), and we will say a group 푈 < 퐺 is unipotent if every element of 푈 is unipotent. Connected unipotent subgroups of SL(푑, ℝ), in particular one-parameter unipo- Ratner presenting her rigidity theorems in a plenary tent subgroups, are basic objects in Ratner’s work. A unipo- address to the 1994 ICM, Zurich. tent group is said to be a one-parameter unipotent group if there is a surjective homomorphism defined by polyno- In this note we delve a bit more into Ratner’s rigidity theo- mials from the additive group of real numbers onto the rems for unipotent flows and highlight some of their strik- group; for instance ing applications, expanding on the outline presented by Elon Lindenstrauss is Alice Kusiel and Kurt Vorreuter professor of mathemat- 1 푡 푡2/2 1 푡 ics at The Hebrew University of Jerusalem.
    [Show full text]
  • Linear Independence, the Wronskian, and Variation of Parameters
    LINEAR INDEPENDENCE, THE WRONSKIAN, AND VARIATION OF PARAMETERS JAMES KEESLING In this post we determine when a set of solutions of a linear differential equation are linearly independent. We first discuss the linear space of solutions for a homogeneous differential equation. 1. Homogeneous Linear Differential Equations We start with homogeneous linear nth-order ordinary differential equations with general coefficients. The form for the nth-order type of equation is the following. dnx dn−1x (1) a (t) + a (t) + ··· + a (t)x = 0 n dtn n−1 dtn−1 0 It is straightforward to solve such an equation if the functions ai(t) are all constants. However, for general functions as above, it may not be so easy. However, we do have a principle that is useful. Because the equation is linear and homogeneous, if we have a set of solutions fx1(t); : : : ; xn(t)g, then any linear combination of the solutions is also a solution. That is (2) x(t) = C1x1(t) + C2x2(t) + ··· + Cnxn(t) is also a solution for any choice of constants fC1;C2;:::;Cng. Now if the solutions fx1(t); : : : ; xn(t)g are linearly independent, then (2) is the general solution of the differential equation. We will explain why later. What does it mean for the functions, fx1(t); : : : ; xn(t)g, to be linearly independent? The simple straightforward answer is that (3) C1x1(t) + C2x2(t) + ··· + Cnxn(t) = 0 implies that C1 = 0, C2 = 0, ::: , and Cn = 0 where the Ci's are arbitrary constants. This is the definition, but it is not so easy to determine from it just when the condition holds to show that a given set of functions, fx1(t); x2(t); : : : ; xng, is linearly independent.
    [Show full text]
  • Introduction to Cluster Algebras. Chapters
    Introduction to Cluster Algebras Chapters 1–3 (preliminary version) Sergey Fomin Lauren Williams Andrei Zelevinsky arXiv:1608.05735v4 [math.CO] 30 Aug 2021 Preface This is a preliminary draft of Chapters 1–3 of our forthcoming textbook Introduction to cluster algebras, joint with Andrei Zelevinsky (1953–2013). Other chapters have been posted as arXiv:1707.07190 (Chapters 4–5), • arXiv:2008.09189 (Chapter 6), and • arXiv:2106.02160 (Chapter 7). • We expect to post additional chapters in the not so distant future. This book grew from the ten lectures given by Andrei at the NSF CBMS conference on Cluster Algebras and Applications at North Carolina State University in June 2006. The material of his lectures is much expanded but we still follow the original plan aimed at giving an accessible introduction to the subject for a general mathematical audience. Since its inception in [23], the theory of cluster algebras has been actively developed in many directions. We do not attempt to give a comprehensive treatment of the many connections and applications of this young theory. Our choice of topics reflects our personal taste; much of the book is based on the work done by Andrei and ourselves. Comments and suggestions are welcome. Sergey Fomin Lauren Williams Partially supported by the NSF grants DMS-1049513, DMS-1361789, and DMS-1664722. 2020 Mathematics Subject Classification. Primary 13F60. © 2016–2021 by Sergey Fomin, Lauren Williams, and Andrei Zelevinsky Contents Chapter 1. Total positivity 1 §1.1. Totally positive matrices 1 §1.2. The Grassmannian of 2-planes in m-space 4 §1.3.
    [Show full text]
  • Integer Matrix Approximation and Data Mining
    Noname manuscript No. (will be inserted by the editor) Integer Matrix Approximation and Data Mining Bo Dong · Matthew M. Lin · Haesun Park Received: date / Accepted: date Abstract Integer datasets frequently appear in many applications in science and en- gineering. To analyze these datasets, we consider an integer matrix approximation technique that can preserve the original dataset characteristics. Because integers are discrete in nature, to the best of our knowledge, no previously proposed technique de- veloped for real numbers can be successfully applied. In this study, we first conduct a thorough review of current algorithms that can solve integer least squares problems, and then we develop an alternative least square method based on an integer least squares estimation to obtain the integer approximation of the integer matrices. We discuss numerical applications for the approximation of randomly generated integer matrices as well as studies of association rule mining, cluster analysis, and pattern extraction. Our computed results suggest that our proposed method can calculate a more accurate solution for discrete datasets than other existing methods. Keywords Data mining · Matrix factorization · Integer least squares problem · Clustering · Association rule · Pattern extraction The first author's research was supported in part by the National Natural Science Foundation of China under grant 11101067 and the Fundamental Research Funds for the Central Universities. The second author's research was supported in part by the National Center for Theoretical Sciences of Taiwan and by the Ministry of Science and Technology of Taiwan under grants 104-2115-M-006-017-MY3 and 105-2634-E-002-001. The third author's research was supported in part by the Defense Advanced Research Projects Agency (DARPA) XDATA program grant FA8750-12-2-0309 and NSF grants CCF-0808863, IIS-1242304, and IIS-1231742.
    [Show full text]
  • An Introduction to the HESSIAN
    An introduction to the HESSIAN 2nd Edition Photograph of Ludwig Otto Hesse (1811-1874), circa 1860 https://upload.wikimedia.org/wikipedia/commons/6/65/Ludwig_Otto_Hesse.jpg See page for author [Public domain], via Wikimedia Commons I. As the students of the first two years of mathematical analysis well know, the Wronskian, Jacobian and Hessian are the names of three determinants and matrixes, which were invented in the nineteenth century, to make life easy to the mathematicians and to provide nightmares to the students of mathematics. II. I don’t know how the Hessian came to the mind of Ludwig Otto Hesse (1811-1874), a quiet professor father of nine children, who was born in Koenigsberg (I think that Koenigsberg gave birth to a disproportionate 1 number of famous men). It is possible that he was studying the problem of finding maxima, minima and other anomalous points on a bi-dimensional surface. (An alternative hypothesis will be presented in section X. ) While pursuing such study, in one variable, one first looks for the points where the first derivative is zero (if it exists at all), and then examines the second derivative at each of those points, to find out its “quality”, whether it is a maximum, a minimum, or an inflection point. The variety of anomalies on a bi-dimensional surface is larger than for a one dimensional line, and one needs to employ more powerful mathematical instruments. Still, also in two dimensions one starts by looking for points where the two first partial derivatives, with respect to x and with respect to y respectively, exist and are both zero.
    [Show full text]
  • LIE GROUPS and ALGEBRAS NOTES Contents 1. Definitions 2
    LIE GROUPS AND ALGEBRAS NOTES STANISLAV ATANASOV Contents 1. Definitions 2 1.1. Root systems, Weyl groups and Weyl chambers3 1.2. Cartan matrices and Dynkin diagrams4 1.3. Weights 5 1.4. Lie group and Lie algebra correspondence5 2. Basic results about Lie algebras7 2.1. General 7 2.2. Root system 7 2.3. Classification of semisimple Lie algebras8 3. Highest weight modules9 3.1. Universal enveloping algebra9 3.2. Weights and maximal vectors9 4. Compact Lie groups 10 4.1. Peter-Weyl theorem 10 4.2. Maximal tori 11 4.3. Symmetric spaces 11 4.4. Compact Lie algebras 12 4.5. Weyl's theorem 12 5. Semisimple Lie groups 13 5.1. Semisimple Lie algebras 13 5.2. Parabolic subalgebras. 14 5.3. Semisimple Lie groups 14 6. Reductive Lie groups 16 6.1. Reductive Lie algebras 16 6.2. Definition of reductive Lie group 16 6.3. Decompositions 18 6.4. The structure of M = ZK (a0) 18 6.5. Parabolic Subgroups 19 7. Functional analysis on Lie groups 21 7.1. Decomposition of the Haar measure 21 7.2. Reductive groups and parabolic subgroups 21 7.3. Weyl integration formula 22 8. Linear algebraic groups and their representation theory 23 8.1. Linear algebraic groups 23 8.2. Reductive and semisimple groups 24 8.3. Parabolic and Borel subgroups 25 8.4. Decompositions 27 Date: October, 2018. These notes compile results from multiple sources, mostly [1,2]. All mistakes are mine. 1 2 STANISLAV ATANASOV 1. Definitions Let g be a Lie algebra over algebraically closed field F of characteristic 0.
    [Show full text]
  • Chapter 2 Solving Linear Equations
    Chapter 2 Solving Linear Equations Po-Ning Chen, Professor Department of Computer and Electrical Engineering National Chiao Tung University Hsin Chu, Taiwan 30010, R.O.C. 2.1 Vectors and linear equations 2-1 • What is this course Linear Algebra about? Algebra (代數) The part of mathematics in which letters and other general sym- bols are used to represent numbers and quantities in formulas and equations. Linear Algebra (線性代數) To combine these algebraic symbols (e.g., vectors) in a linear fashion. So, we will not combine these algebraic symbols in a nonlinear fashion in this course! x x1x2 =1 x 1 Example of nonlinear equations for = x : 2 x1/x2 =1 x x x x 1 1 + 2 =4 Example of linear equations for = x : 2 x1 − x2 =1 2.1 Vectors and linear equations 2-2 The linear equations can always be represented as matrix operation, i.e., Ax = b. Hence, the central problem of linear algorithm is to solve a system of linear equations. x − 2y =1 1 −2 x 1 Example of linear equations. ⇔ = 2 3x +2y =11 32 y 11 • A linear equation problem can also be viewed as a linear combination prob- lem for column vectors (as referred by the textbook as the column picture view). In contrast, the original linear equations (the red-color one above) is referred by the textbook as the row picture view. 1 −2 1 Column picture view: x + y = 3 2 11 1 −2 We want to find the scalar coefficients of column vectors and to form 3 2 1 another vector .
    [Show full text]