Permutation Learning Via Lehmer Codes

Total Page:16

File Type:pdf, Size:1020Kb

Permutation Learning Via Lehmer Codes 24th European Conference on Artificial Intelligence - ECAI 2020 Santiago de Compostela, Spain Permutation Learning via Lehmer Codes A¨ıssatou Diallo1 Markus Zopf1 Johannes Furnkranz¨ 2 Abstract. Many machine learning problems require to learn to Another line of work investigated convex surrogates of the permu- permute a set of objects. Notable applications include ranking or tations themselves. The classical workflow for tackling the discrete sorting. One of the difficulties of learning in such a combinatorial optimization problem over the set of permutations of n elements is space is the definition of meaningful and differentiable distances and the following: (1) use permutation matrices to represent permuta- loss functions. Lehmer codes are an elegant way for mapping per- tions, (2) as these matrices are discrete, gradient-based optimisation mutations into a vector space where the Euclidean distance between is not possible; hence use a relaxation in order to transform them two codes corresponds to the Kendall tau distance between the corre- into the nearest convex, surrogate, doubly stochastic matrix, i.e., sponding rankings. This transformation, therefore, allows the use of into a matrix where all rows and all columns sum up to one. In a surrogate loss converting the original permutation learning problem particular, permutation matrices may be viewed as special cases of into a surrogate prediction problem which is easier to optimize. To doubly stochastic matrices which contain only zero and ones. that end, learning to predict Lehmer codes allows a transformation of In this paper, we propose a different approach for learning permu- the inference problem from permutations matrices to row-stochastic tations, which differs from previous work [6] in that we transform matrices, thereby removing the constraints of row and columns to the permutation matrix from the combinatorial space to the Hilbert sum to one. This permits a straight-forward optimization of the space using Lehmer codes [9]. Unlike [7], where the use of Lehmer permutation prediction problem. We demonstrate the effectiveness codes was proposed for label ranking, we employ them for the task of this approach for object ranking problems by providing empirical of object ranking. Lehmer codes provide a bijective mapping from results and comparing it to competitive baselines on different tasks. permutations onto sequences of integers. Moreover, Lehmer codes enjoy algorithmic advantages such as that the L1 norm of the code represents the Kendall’s tau distance between the permutation that 1 INTRODUCTION generated the code and the lexicographic permutation, which is a Permutation learning is a long-standing problem in many scientific commonly used metric for assessing the ordinal association between fields. The need to deal with permutations arises in a variety of ranking data. Lehmer codes provide vector representations of permu- applications, such as ranking [13], information retrieval or multi- tations in which the coordinates are decoupled, i.e., the values of each object tracking [6] for computer vision. The fundamental underlying coordinate are not restricted by the other coordinates. This mapping problem is the correct identification of the permutation that allows allows to transform the problem from learning a doubly stochastic to reconstruct the original sequence. Ranking, sorting and matching matrix to an integer valued vector which enjoys great advantages over problems are classical examples of the necessity to perform inference the direct optimization of permutation matrices. over such a combinatorial space, and they can be both represented by The remainder of the paper is organized as follows: Section 2 permutations over the set of objects to sort. However, one of the main lays the necessary mathematical background and defines the notation difficulties in dealing with such combinatorial objects is the factorial used throughout the paper, whereas Section 3 introduces prior work growth of the number of permutations with growing number of ele- on permutation learning and embedding of permutations. Section 4 ments. Moreover, there are mutual exclusivity constraints associated presents our primary contribution, motivating our theoretical choices. with permutations. Finally, the space of all possible permutations Finally, Section 5 presents experiments which compare our proposed of a set of objects is not smooth. In fact, permutations are discrete approach against alternative methods, illustrating the benefits of objects and thus it is not possible to compute straight marginals with Lehmer codes for the task of permutation learning. gradient-based methods. Thus, exact solutions are intractable to find for most cases. 2 BACKGROUND One strategy for working with combinatorial objects is to embed In this section, we formally state the permutation learning problem as them into continuous spaces, in order to impose a structure and a well as the notation used throughout the paper, which closely follows topology so that conventional tools for inference can be used. Prior the literature [7, 11]. works have proposed different approaches to deal with this problem. For example, several authors [4, 6] derived an approximation of a general probability distribution with a restricted set of basis 2.1 Notation functions, performing inference in the Fourier domain, and using Given an ordered list of items indexed by f1; :::; Ng , let us accompanying transformations to project back to the combinatorial denote without loss of generality, πIn as the n-dimensional identity space. permutation vector. A permutation is a rearrangement of the elements 1 Technische Universitat¨ Darmstadt, Germany, email: [diallo@aiphes, of πIn , i.e a bijection from πIn to πn, a generic permutation vector n n zopf@ke].tu-darmstadt.de with πn 2 P . P denotes the set of all permutations vector of 2 Johannes Kepler Universitat¨ Linz, Austria, email: juffi@faw.jku.at length n in the symmetric group, and has the cardinality n!. 24th European Conference on Artificial Intelligence - ECAI 2020 Santiago de Compostela, Spain 2.2 Learning problem The permutation learning problem aims at learning a function s : X!Pn that maps a feature space X to the permutation space Pn. Loosely speaking, given a sequence of shuffled items, the goal is to recover their natural order as defined by some specific criterion c. Formally, the underlying problem is to predict the permutation matrix n×n Pi 2 f0; 1g , given a sequence of shuffled items X~i such that −1 Pi recovers the original ordered sequence Xi. In such a setting, the cardinality of P is n!, which often is orders of magnitude larger than the number of training samples. This is one of the difficulties in dealing with combinatorial spaces such as n P . Moreover, permutations are represented as discrete points in the Figure 1: The Birkhoff (left) and Lehmer (right) polytopes of size 3. Euclidean space. This discrete nature typically prohibits gradient- Permutations matrices are in the column representation. based optimization, because these discrete solvers work by iteratively moving towards an optimum via stochastic updates along gradient a finite number of permutation matrices. The set of n × n doubly directions. A large part of the dedicated literature proposed different stochastic matrices is the called the Birkhoff polytope (cf. Figure 1 relaxations of the permutation matrix, the best known of which (left)). A natural consequence that arises from this theorem is involve the use of the Sinkhorn operator [18]. This operator is a to consider doubly stochastic matrices as a convex relaxation of procedure that takes as an input a non-negative square matrix and permutation matrices. However, it is difficult to learn matrices with outputs a doubly-stochastic matrix by iteratively normalizing rows this particular structure. The dedicated literature proposed to use and columns[1]. Sinkhorn normalization [18, 5] to ensure that columns and rows sums Another approach consists of embedding the permutations into to one. This procedure works by iteratively normalising rows and continuous spaces. An example of such line of work is given by columns of the square matrix. [7], which proposes three different encodings suitable to act as embeddings for the task of label ranking. Among these encodings, 2.4 Lehmer codes Lehmer codes showed great algorithmic advantages, in particular, the fact that the encoding and decoding step have a cost of O(n). Several works in theoretical computer science and discrete mathe- matics deal with effective ways to representing permutations. The 2.3 Permutation matrices Lehmer code [9] is a particular way to encode each possible permu- tation of a sequence of n numbers. Concretely, it is a bijection from As previously stated, there are several ways to represent permu- every permutation N to a function φ : N ! f0; 1;:::;N − 1g. tations. In matrix theory, a permutation matrix is a square matrix The Lehmer code,J alsoK known as inversionJ K vector, is a word of the with the peculiarity of having exactly a single unit value for each form cπ 2 CN , f0g × f0; 1g × f0; 2g × · · · × f0;N − 1g, where row and column and zero elsewhere. They are a natural and concise for j = 1; :::; N, way to represent permutations. In fact, given an ordered sequence cπ(j) = #fi 2 N : i < j; π(i) > π(j)g (2) of elements represented as a vector πIn , it is possible to derive the n×n J K permutation πn from the permutation matrix Pn 2 f0; 1g , Thus, the coordinate cπ(j) represents the number of elements i with simply by performing a vector-matrix multiplication: index smaller than j that are ranked higher than j in the permutation π. Consider the following example, which shows the canonical set of πn = PnπIn (1) items e, a permutation π, and the corresponding Lehmer code cπ: where Pn is obtained by swapping the rows of the identity matrix e 1 2 3 4 5 6 7 8 9 according to the desired permutation πn. The set of the n × n π 2 1 4 5 7 3 6 9 8 n! permutation matrices has cardinality and is the subset of the non- cπ 0 1 0 0 0 3 1 0 1 n×n singular matrices in R .
Recommended publications
  • Math 511 Advanced Linear Algebra Spring 2006
    MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006 Sherod Eubanks HOMEWORK 2 x2:1 : 2; 5; 9; 12 x2:3 : 3; 6 x2:4 : 2; 4; 5; 9; 11 Section 2:1: Unitary Matrices Problem 2 If ¸ 2 σ(U) and U 2 Mn is unitary, show that j¸j = 1. Solution. If ¸ 2 σ(U), U 2 Mn is unitary, and Ux = ¸x for x 6= 0, then by Theorem 2:1:4(g), we have kxkCn = kUxkCn = k¸xkCn = j¸jkxkCn , hence j¸j = 1, as desired. Problem 5 Show that the permutation matrices in Mn are orthogonal and that the permutation matrices form a sub- group of the group of real orthogonal matrices. How many different permutation matrices are there in Mn? Solution. By definition, a matrix P 2 Mn is called a permutation matrix if exactly one entry in each row n and column is equal to 1, and all other entries are 0. That is, letting ei 2 C denote the standard basis n th element of C that has a 1 in the i row and zeros elsewhere, and Sn be the set of all permutations on n th elements, then P = [eσ(1) j ¢ ¢ ¢ j eσ(n)] = Pσ for some permutation σ 2 Sn such that σ(k) denotes the k member of σ. Observe that for any σ 2 Sn, and as ½ 1 if i = j eT e = σ(i) σ(j) 0 otherwise for 1 · i · j · n by the definition of ei, we have that 2 3 T T eσ(1)eσ(1) ¢ ¢ ¢ eσ(1)eσ(n) T 6 .
    [Show full text]
  • Chapter Four Determinants
    Chapter Four Determinants In the first chapter of this book we considered linear systems and we picked out the special case of systems with the same number of equations as unknowns, those of the form T~x = ~b where T is a square matrix. We noted a distinction between two classes of T ’s. While such systems may have a unique solution or no solutions or infinitely many solutions, if a particular T is associated with a unique solution in any system, such as the homogeneous system ~b = ~0, then T is associated with a unique solution for every ~b. We call such a matrix of coefficients ‘nonsingular’. The other kind of T , where every linear system for which it is the matrix of coefficients has either no solution or infinitely many solutions, we call ‘singular’. Through the second and third chapters the value of this distinction has been a theme. For instance, we now know that nonsingularity of an n£n matrix T is equivalent to each of these: ² a system T~x = ~b has a solution, and that solution is unique; ² Gauss-Jordan reduction of T yields an identity matrix; ² the rows of T form a linearly independent set; ² the columns of T form a basis for Rn; ² any map that T represents is an isomorphism; ² an inverse matrix T ¡1 exists. So when we look at a particular square matrix, the question of whether it is nonsingular is one of the first things that we ask. This chapter develops a formula to determine this. (Since we will restrict the discussion to square matrices, in this chapter we will usually simply say ‘matrix’ in place of ‘square matrix’.) More precisely, we will develop infinitely many formulas, one for 1£1 ma- trices, one for 2£2 matrices, etc.
    [Show full text]
  • GROUP ACTIONS 1. Introduction the Groups Sn, An, and (For N ≥ 3)
    GROUP ACTIONS KEITH CONRAD 1. Introduction The groups Sn, An, and (for n ≥ 3) Dn behave, by their definitions, as permutations on certain sets. The groups Sn and An both permute the set f1; 2; : : : ; ng and Dn can be considered as a group of permutations of a regular n-gon, or even just of its n vertices, since rigid motions of the vertices determine where the rest of the n-gon goes. If we label the vertices of the n-gon in a definite manner by the numbers from 1 to n then we can view Dn as a subgroup of Sn. For instance, the labeling of the square below lets us regard the 90 degree counterclockwise rotation r in D4 as (1234) and the reflection s across the horizontal line bisecting the square as (24). The rest of the elements of D4, as permutations of the vertices, are in the table below the square. 2 3 1 4 1 r r2 r3 s rs r2s r3s (1) (1234) (13)(24) (1432) (24) (12)(34) (13) (14)(23) If we label the vertices in a different way (e.g., swap the labels 1 and 2), we turn the elements of D4 into a different subgroup of S4. More abstractly, if we are given a set X (not necessarily the set of vertices of a square), then the set Sym(X) of all permutations of X is a group under composition, and the subgroup Alt(X) of even permutations of X is a group under composition. If we list the elements of X in a definite order, say as X = fx1; : : : ; xng, then we can think about Sym(X) as Sn and Alt(X) as An, but a listing in a different order leads to different identifications 1 of Sym(X) with Sn and Alt(X) with An.
    [Show full text]
  • Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
    Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation
    [Show full text]
  • Alternating Sign Matrices, Extensions and Related Cones
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/311671190 Alternating sign matrices, extensions and related cones Article in Advances in Applied Mathematics · May 2017 DOI: 10.1016/j.aam.2016.12.001 CITATIONS READS 0 29 2 authors: Richard A. Brualdi Geir Dahl University of Wisconsin–Madison University of Oslo 252 PUBLICATIONS 3,815 CITATIONS 102 PUBLICATIONS 1,032 CITATIONS SEE PROFILE SEE PROFILE Some of the authors of this publication are also working on these related projects: Combinatorial matrix theory; alternating sign matrices View project All content following this page was uploaded by Geir Dahl on 16 December 2016. The user has requested enhancement of the downloaded file. All in-text references underlined in blue are added to the original document and are linked to publications on ResearchGate, letting you access and read them immediately. Alternating sign matrices, extensions and related cones Richard A. Brualdi∗ Geir Dahly December 1, 2016 Abstract An alternating sign matrix, or ASM, is a (0; ±1)-matrix where the nonzero entries in each row and column alternate in sign, and where each row and column sum is 1. We study the convex cone generated by ASMs of order n, called the ASM cone, as well as several related cones and polytopes. Some decomposition results are shown, and we find a minimal Hilbert basis of the ASM cone. The notion of (±1)-doubly stochastic matrices and a generalization of ASMs are introduced and various properties are shown. For instance, we give a new short proof of the linear characterization of the ASM polytope, in fact for a more general polytope.
    [Show full text]
  • Group Properties and Group Isomorphism
    GROUP PROPERTIES AND GROUP ISOMORPHISM Evelyn. M. Manalo Mathematics Honors Thesis University of California, San Diego May 25, 2001 Faculty Mentor: Professor John Wavrik Department of Mathematics GROUP PROPERTIES AND GROUP ISOMORPHISM I n t r o d u c t i o n T H E I M P O R T A N C E O F G R O U P T H E O R Y is relevant to every branch of Mathematics where symmetry is studied. Every symmetrical object is associated with a group. It is in this association why groups arise in many different areas like in Quantum Mechanics, in Crystallography, in Biology, and even in Computer Science. There is no such easy definition of symmetry among mathematical objects without leading its way to the theory of groups. In this paper we present the first stages of constructing a systematic method for classifying groups of small orders. Classifying groups usually arise when trying to distinguish the number of non-isomorphic groups of order n. This paper arose from an attempt to find a formula or an algorithm for classifying groups given invariants that can be readily determined without any other known assumptions about the group. This formula is very useful if we want to know if two groups are isomorphic. Mathematical objects are considered to be essentially the same, from the point of view of their algebraic properties, when they are isomorphic. When two groups Γ and Γ’ have exactly the same group-theoretic structure then we say that Γ is isomorphic to Γ’ or vice versa.
    [Show full text]
  • More Gaussian Elimination and Matrix Inversion
    Week 7 More Gaussian Elimination and Matrix Inversion 7.1 Opening Remarks 7.1.1 Introduction * View at edX 289 Week 7. More Gaussian Elimination and Matrix Inversion 290 7.1.2 Outline 7.1. Opening Remarks..................................... 289 7.1.1. Introduction..................................... 289 7.1.2. Outline....................................... 290 7.1.3. What You Will Learn................................ 291 7.2. When Gaussian Elimination Breaks Down....................... 292 7.2.1. When Gaussian Elimination Works........................ 292 7.2.2. The Problem.................................... 297 7.2.3. Permutations.................................... 299 7.2.4. Gaussian Elimination with Row Swapping (LU Factorization with Partial Pivoting) 304 7.2.5. When Gaussian Elimination Fails Altogether................... 311 7.3. The Inverse Matrix.................................... 312 7.3.1. Inverse Functions in 1D.............................. 312 7.3.2. Back to Linear Transformations.......................... 312 7.3.3. Simple Examples.................................. 314 7.3.4. More Advanced (but Still Simple) Examples................... 320 7.3.5. Properties...................................... 324 7.4. Enrichment......................................... 325 7.4.1. Library Routines for LU with Partial Pivoting................... 325 7.5. Wrap Up.......................................... 327 7.5.1. Homework..................................... 327 7.5.2. Summary...................................... 327 7.1. Opening
    [Show full text]
  • Pattern Avoidance in Alternating Sign Matrices
    PATTERN AVOIDANCE IN ALTERNATING SIGN MATRICES ROBERT JOHANSSON AND SVANTE LINUSSON Abstract. We generalize the definition of a pattern from permu- tations to alternating sign matrices. The number of alternating sign matrices avoiding 132 is proved to be counted by the large Schr¨oder numbers, 1, 2, 6, 22, 90, 394 . .. We give a bijection be- tween 132-avoiding alternating sign matrices and Schr¨oder-paths, which gives a refined enumeration. We also show that the 132, 123- avoiding alternating sign matrices are counted by every second Fi- bonacci number. 1. Introduction For some time, we have emphasized the use of permutation matrices when studying pattern avoidance. This and the fact that alternating sign matrices are generalizations of permutation matrices, led to the idea of studying pattern avoidance in alternating sign matrices. The main results, concerning 132-avoidance, are presented in Section 3. In Section 5, we enumerate sets of alternating sign matrices avoiding two patterns of length three at once. In Section 6 we state some open problems. Let us start with the basic definitions. Given a permutation π of length n we will represent it with the n×n permutation matrix with 1 in positions (i, π(i)), 1 ≤ i ≤ n and 0 in all other positions. We will also suppress the 0s. 1 1 3 1 2 1 Figure 1. The permutation matrix of 132. In this paper, we will think of permutations in terms of permutation matrices. Definition 1.1. (Patterns in permutations) A permutation π is said to contain a permutation τ as a pattern if the permutation matrix of τ is a submatrix of π.
    [Show full text]
  • Mathematics for Humanists
    Mathematics for Humanists Mathematics for Humanists Herbert Gintis xxxxxxxxx xxxxxxxxxx Press xxxxxxxxx and xxxxxx Copyright c 2021 by ... Published by ... All Rights Reserved Library of Congress Cataloging-in-Publication Data Gintis, Herbert Mathematics for Humanists/ Herbert Gintis p. cm. Includes bibliographical references and index. ISBN ...(hardcover: alk. paper) HB... xxxxxxxxxx British Library Cataloging-in-Publication Data is available The publisher would like to acknowledge the author of this volume for providing the camera-ready copy from which this book was printed This book has been composed in Times and Mathtime by the author Printed on acid-free paper. Printed in the United States of America 10987654321 This book is dedicated to my mathematics teachers: Pincus Shub, Walter Gottschalk, Abram Besikovitch, and Oskar Zariski Contents Preface xii 1 ReadingMath 1 1.1 Reading Math 1 2 The LanguageofLogic 2 2.1 TheLanguageofLogic 2 2.2 FormalPropositionalLogic 4 2.3 Truth Tables 5 2.4 ExercisesinPropositionalLogic 7 2.5 Predicate Logic 8 2.6 ProvingPropositionsinPredicateLogic 9 2.7 ThePerilsofLogic 10 3 Sets 11 3.1 Set Theory 11 3.2 PropertiesandPredicates 12 3.3 OperationsonSets 14 3.4 Russell’s Paradox 15 3.5 Ordered Pairs 17 3.6 MathematicalInduction 18 3.7 SetProducts 19 3.8 RelationsandFunctions 20 3.9 PropertiesofRelations 21 3.10 Injections,Surjections,andBijections 22 3.11 CountingandCardinality 23 3.12 The Cantor-Bernstein Theorem 24 3.13 InequalityinCardinalNumbers 25 3.14 Power Sets 26 3.15 TheFoundationsofMathematics
    [Show full text]
  • Matrices That Commute with a Permutation Matrix
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Matrices That Commute With a Permutation Matrix Jeffrey L. Stuart Department of Mathematics University of Southern Mississippi Hattiesburg, Mississippi 39406-5045 and James R. Weaver Department of Mathematics and Statistics University of West Florida Pensacola. Florida 32514 Submitted by Donald W. Robinson ABSTRACT Let P be an n X n permutation matrix, and let p be the corresponding permuta- tion. Let A be a matrix such that AP = PA. It is well known that when p is an n-cycle, A is permutation similar to a circulant matrix. We present results for the band patterns in A and for the eigenstructure of A when p consists of several disjoint cycles. These results depend on the greatest common divisors of pairs of cycle lengths. 1. INTRODUCTION A central theme in matrix theory concerns what can be said about a matrix if it commutes with a given matrix of some special type. In this paper, we investigate the combinatorial structure and the eigenstructure of a matrix that commutes with a permutation matrix. In doing so, we follow a long tradition of work on classes of matrices that commute with a permutation matrix of a specified type. In particular, our work is related both to work on the circulant matrices, the results for which can be found in Davis’s compre- hensive book [5], and to work on the centrosymmetric matrices, which can be found in the book by Aitken [l] and in the papers by Andrew [2], Bamett, LZNEAR ALGEBRA AND ITS APPLICATIONS 150:255-265 (1991) 255 Q Elsevier Science Publishing Co., Inc., 1991 655 Avenue of the Americas, New York, NY 10010 0024-3795/91/$3.50 256 JEFFREY L.
    [Show full text]
  • The Waldspurger Transform of Permutations and Alternating Sign Matrices
    Séminaire Lotharingien de Combinatoire 78B (2017) Proceedings of the 29th Conference on Formal Power Article #41, 12 pp. Series and Algebraic Combinatorics (London) The Waldspurger Transform of Permutations and Alternating Sign Matrices Drew Armstrong∗ and James McKeown Department of Mathematics, University of Miami, Coral Gables, FL Abstract. In 2005 J. L. Waldspurger proved the following theorem: Given a finite real reflection group G, the closed positive root cone is tiled by the images of the open weight cone under the action of the linear transformations 1 g. Shortly after this − E. Meinrenken extended the result to affine Weyl groups and then P. V. Bibikov and V. S. Zhgoon gave a uniform proof for a discrete reflection group acting on a simply- connected space of constant curvature. In this paper we show that the Waldspurger and Meinrenken theorems of type A give an interesting new perspective on the combinatorics of the symmetric group. In particular, for each permutation matrix g S we define a non-negative integer 2 n matrix WT(g), called the Waldspurger transform of g. The definition of the matrix WT(g) is purely combinatorial but it turns out that its columns are the images of the fundamental weights under 1 g, expressed in simple root coordinates. The possible − columns of WT(g) (which we call UM vectors) biject to many interesting structures including: unimodal Motzkin paths, abelian ideals in the Lie algebra sln(C), Young diagrams with maximum hook length n, and integer points inside a certain polytope. We show that the sum of the entries of WT(g) is half the entropy of the corresponding permutation g, which is known to equal the rank of g in the MacNeille completion of the Bruhat order.
    [Show full text]
  • 18.704 Supplementary Notes March 23, 2005 the Subgroup Ω For
    18.704 Supplementary Notes March 23, 2005 The subgroup Ω for orthogonal groups In the case of the linear group, it is shown in the text that P SL(n; F ) (that is, the group SL(n) of determinant one matrices, divided by its center) is usually a simple group. In the case of symplectic group, P Sp(2n; F ) (the group of symplectic matrices divided by its center) is usually a simple group. In the case of the orthog- onal group (as Yelena will explain on March 28), what turns out to be simple is not P SO(V ) (the orthogonal group of V divided by its center). Instead there is a mysterious subgroup Ω(V ) of SO(V ), and what is usually simple is P Ω(V ). The purpose of these notes is first to explain why this complication arises, and then to give the general definition of Ω(V ) (along with some of its basic properties). So why should the complication arise? There are some hints of it already in the case of the linear group. We made a lot of use of GL(n; F ), the group of all invertible n × n matrices with entries in F . The most obvious normal subgroup of GL(n; F ) is its center, the group F × of (non-zero) scalar matrices. Dividing by the center gives (1) P GL(n; F ) = GL(n; F )=F ×; the projective general linear group. We saw that this group acts faithfully on the projective space Pn−1(F ), and generally it's a great group to work with.
    [Show full text]