Metric Learning for Diffeomorphic Image Registration

Total Page:16

File Type:pdf, Size:1020Kb

Metric Learning for Diffeomorphic Image Registration Metric learning for diffeomorphic image registration. Fran¸cois-Xavier Vialard Metric learning for diffeomorphic image registration. Fran¸cois-XavierVialard Universit´eParis-Est Marne-la-Vall´ee joint work with M. Niethammer and R. Kwitt. IHP, March 2019. Metric learning for diffeomorphic image Outline registration. Fran¸cois-Xavier Vialard 1 Introduction to diffeomorphisms group and Riemannian tools 2 Choice of the metric 3 Spatially dependent metrics 4 Metric learning 5 SVF metric learning Metric learning for diffeomorphic image Example of problems of interest registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Given two shapes, find a diffeomorphism of R3 that maps one shape onto the other Metric learning for diffeomorphic image Example of problems of interest registration. 3 Fran¸cois-Xavier Given two shapes, find a diffeomorphism of R that maps one Vialard shape onto the other Introduction to diffeomorphisms group and Riemannian tools Different data types and different way of representing them. Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Figure{ Two slices of 3D brain images of the same subject at different ages Metric learning for diffeomorphic image Example of problems of interest registration. 3 Given two shapes, find a diffeomorphism of R that maps one Fran¸cois-Xavier shape onto the other Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Deformation by a diffeomorphism Figure{ Diffeomorphic deformation of the image Metric learning for diffeomorphic image Variety of shapes registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Figure{ Different anatomical structures extracted from MRI data Metric learning for diffeomorphic image Variety of shapes registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Figure{ Different anatomical structures extracted from MRI data Metric learning for diffeomorphic image A Riemannian approach to diffeomorphic registration. Fran¸cois-Xavier registration Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent Several diffeomorphic registration methods are available: metrics Metric learning • Free-form deformations B-spline-based diffeomorphisms by D. SVF metric learning Rueckert • Log-demons (X.Pennec et al.) • Large Deformations by Diffeomorphisms (M. Miller,A. Trouv´e,L. Younes) • ANTS Only the two last ones provide a Riemannian framework. Metric learning for diffeomorphic image A Riemannian approach to diffeomorphic registration. Fran¸cois-Xavier registration Vialard Introduction to diffeomorphisms group and Riemannian tools n Choice of the metric • vt 2 V a time dependent vector field on R . Spatially dependent • 't 2 Diff , the flow defined by metrics Metric learning SVF metric learning @t 't = vt ('t ) : (1) Action of the group of diffeomorphism G0 (flow at time 1): Π: G0 × C ! C ; : Π('; X ) = ':X 2 1 R 1 2 Right-invariant metric on G0: d('0;1; Id) = 2 0 jvt jV dt. −! Strong metric needed on V (Mumford and Michor: Vanishing Geodesic Distance on...) Right invariant distance on G0 Z 1 2 2 d(Id;') = inf jvt j dt ; 2 V v2L ([0;1];V ) 0 −! geodesics on G0. V is a Reproducing kernel Hilbert Space (RKHS): (pointwise evaluation continuous) =) Existence of a matrix function kV (kernel) defined on U × U such that: hv(x); ai = hkV (:; x)a; viV : Metric learning for diffeomorphic image Matching problems in a diffeomorphic framework registration. Fran¸cois-Xavier Vialard 1 n Introduction to U a domain in R diffeomorphisms group 1 and Riemannian tools 2 V a Hilbert space of C vector fields such that: Choice of the metric Spatially dependent kvk1;1 6 CjvjV : metrics Metric learning SVF metric learning Right invariant distance on G0 Z 1 2 2 d(Id;') = inf jvt j dt ; 2 V v2L ([0;1];V ) 0 −! geodesics on G0. Metric learning for diffeomorphic image Matching problems in a diffeomorphic framework registration. Fran¸cois-Xavier Vialard 1 n Introduction to U a domain in R diffeomorphisms group 1 and Riemannian tools 2 V a Hilbert space of C vector fields such that: Choice of the metric Spatially dependent kvk1;1 6 CjvjV : metrics Metric learning V is a Reproducing kernel Hilbert Space (RKHS): (pointwise SVF metric learning evaluation continuous) =) Existence of a matrix function kV (kernel) defined on U × U such that: hv(x); ai = hkV (:; x)a; viV : Metric learning for diffeomorphic image Matching problems in a diffeomorphic framework registration. Fran¸cois-Xavier Vialard 1 n Introduction to U a domain in R diffeomorphisms group 1 and Riemannian tools 2 V a Hilbert space of C vector fields such that: Choice of the metric Spatially dependent kvk1;1 6 CjvjV : metrics Metric learning V is a Reproducing kernel Hilbert Space (RKHS): (pointwise SVF metric learning evaluation continuous) =) Existence of a matrix function kV (kernel) defined on U × U such that: hv(x); ai = hkV (:; x)a; viV : Right invariant distance on G0 Z 1 2 2 d(Id;') = inf jvt j dt ; 2 V v2L ([0;1];V ) 0 −! geodesics on G0. Metric learning for diffeomorphic image Variational formulation registration. Fran¸cois-Xavier Vialard Introduction to Find the best deformation, minimize diffeomorphisms group and Riemannian tools J (') = inf d(':A; B)2 (2) Choice of the metric Spatially dependent '2GV | {z } metrics similarity measure Metric learning SVF metric learning Metric learning for diffeomorphic image Variational formulation registration. Fran¸cois-Xavier Find the best deformation, minimize Vialard Introduction to diffeomorphisms group 2 J (') = inf d(':A; B) (2) and Riemannian tools '2G V | {z } Choice of the metric similarity measure Spatially dependent metrics Tychonov regularization: Metric learning SVF metric learning 1 2 J (') = R(') + 2 d(':A; B) : (3) | {z } 2σ Regularization | {z } similarity measure Riemannian metric on GV : Z 1 1 2 R(') = jvt jV dt (4) 2 0 is a right-invariant metric on GV . Main issues for practical applications: • choice of the metric (prior), • choice of the similarity measure. Metric learning for diffeomorphic image Optimization problem registration. Fran¸cois-Xavier Minimizing Vialard Introduction to 1 diffeomorphisms group 1 Z 1 and Riemannian tools J (v) = jv j2 dt + d(' :A; B)2 : t V 2 0;1 Choice of the metric 2 0 2σ Spatially dependent In the case of landmarks: metrics Metric learning k SVF metric learning 1 Z 1 1 X J (') = jv j2 dt + k'(x ) − y k2 ; 2 t V 2σ2 i i 0 i=1 In the case of images: Z 2 2 d('0;1:I0; Itarget ) = jI0 ◦ '1;0 − Itarget j dx : U Metric learning for diffeomorphic image Optimization problem registration. Fran¸cois-Xavier Minimizing Vialard Introduction to 1 diffeomorphisms group 1 Z 1 and Riemannian tools J (v) = jv j2 dt + d(' :A; B)2 : t V 2 0;1 Choice of the metric 2 0 2σ Spatially dependent In the case of landmarks: metrics Metric learning k SVF metric learning 1 Z 1 1 X J (') = jv j2 dt + k'(x ) − y k2 ; 2 t V 2σ2 i i 0 i=1 In the case of images: Z 2 2 d('0;1:I0; Itarget ) = jI0 ◦ '1;0 − Itarget j dx : U Main issues for practical applications: • choice of the metric (prior), • choice of the similarity measure. Metric learning for diffeomorphic image Why does the Riemannian framework matter? registration. Fran¸cois-Xavier Vialard Generalizations of statistical tools in Euclidean space: Introduction to diffeomorphisms group • Distance often given by a Riemannian metric. and Riemannian tools • Straight lines ! geodesic defined by Choice of the metric Spatially dependent Z 1 metrics 2 Metric learning Variational definition: arg min kc_kc(t) dt = 0 ; c(t) 0 SVF metric learning Equivalent (local) definition: rc_ c_ =c ¨ + Γ(c)(c _; c_) = 0 : • Average ! Fr´echet/Karcher mean. Variational definition: arg minfx ! E[d 2(x; y)]dµ(y)g 2 Critical point definition: E[rx d (x; y)]dµ(y)] = 0 : • PCA ! Tangent PCA or PGA. • Geodesic regression, cubic regression...(variational or algebraic) Metric learning for diffeomorphic image Karcher mean on 3D images registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group Init. guesses and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning 1 iteration SVF metric learning 2 iterations 3 iterations 1 2 3 4 Ai Ai Ai Ai m Figure{ Average image estimates Ai , m 2 f1; ··· ; 4g after i =0, 1, 2 and 3 iterations. Metric learning for diffeomorphic image Interpolation, Extrapolation registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Figure{ Geodesic regression (MICCAI 2011) Metric learning for diffeomorphic image Interpolation, Extrapolation registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning SVF metric learning Figure{ Extrapolation of happiness Metric learning for diffeomorphic image registration. Fran¸cois-Xavier Vialard Introduction to diffeomorphisms group and Riemannian tools Choice of the metric Spatially dependent metrics Metric learning What metric to choose? SVF metric learning The norm on V is simply Z Z 2 1=2 2 kvkV = hv(x); (Lv)(x)i dx = (L v) (x) dx : Ω Ω Scale parameter important! 2 − kx−yk n kσ(x; y) = e σ2 kernel/operator (Id −σ∆) (5) • σ small: good matching but non regular deformations and more local minima. • σ large: poor matching but regular deformations and more global minima. Metric learning for diffeomorphic image Choosing the right-invariant metric registration. Right-invariant metric: Eulerian fluid dynamic viewpoint on Fran¸cois-Xavier Vialard regularization.
Recommended publications
  • The General Linear Group
    18.704 Gabe Cunningham 2/18/05 [email protected] The General Linear Group Definition: Let F be a field. Then the general linear group GLn(F ) is the group of invert- ible n × n matrices with entries in F under matrix multiplication. It is easy to see that GLn(F ) is, in fact, a group: matrix multiplication is associative; the identity element is In, the n × n matrix with 1’s along the main diagonal and 0’s everywhere else; and the matrices are invertible by choice. It’s not immediately clear whether GLn(F ) has infinitely many elements when F does. However, such is the case. Let a ∈ F , a 6= 0. −1 Then a · In is an invertible n × n matrix with inverse a · In. In fact, the set of all such × matrices forms a subgroup of GLn(F ) that is isomorphic to F = F \{0}. It is clear that if F is a finite field, then GLn(F ) has only finitely many elements. An interesting question to ask is how many elements it has. Before addressing that question fully, let’s look at some examples. ∼ × Example 1: Let n = 1. Then GLn(Fq) = Fq , which has q − 1 elements. a b Example 2: Let n = 2; let M = ( c d ). Then for M to be invertible, it is necessary and sufficient that ad 6= bc. If a, b, c, and d are all nonzero, then we can fix a, b, and c arbitrarily, and d can be anything but a−1bc. This gives us (q − 1)3(q − 2) matrices.
    [Show full text]
  • Normal Subgroups of the General Linear Groups Over Von Neumann Regular Rings L
    PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 96, Number 2, February 1986 NORMAL SUBGROUPS OF THE GENERAL LINEAR GROUPS OVER VON NEUMANN REGULAR RINGS L. N. VASERSTEIN1 ABSTRACT. Let A be a von Neumann regular ring or, more generally, let A be an associative ring with 1 whose reduction modulo its Jacobson radical is von Neumann regular. We obtain a complete description of all subgroups of GLn A, n > 3, which are normalized by elementary matrices. 1. Introduction. For any associative ring A with 1 and any natural number n, let GLn A be the group of invertible n by n matrices over A and EnA the subgroup generated by all elementary matrices x1'3, where 1 < i / j < n and x E A. In this paper we describe all subgroups of GLn A normalized by EnA for any von Neumann regular A, provided n > 3. Our description is standard (see Bass [1] and Vaserstein [14, 16]): a subgroup H of GL„ A is normalized by EnA if and only if H is of level B for an ideal B of A, i.e. E„(A, B) C H C Gn(A, B). Here Gn(A, B) is the inverse image of the center of GL„(,4/S) (when n > 2, this center consists of scalar invertible matrices over the center of the ring A/B) under the canonical homomorphism GL„ A —►GLn(A/B) and En(A, B) is the normal subgroup of EnA generated by all elementary matrices in Gn(A, B) (when n > 3, the group En(A, B) is generated by matrices of the form (—y)J'lx1'Jy:i''1 with x € B,y £ A,l < i ^ j < n, see [14]).
    [Show full text]
  • Variational Problems on Flows of Diffeomorphisms for Image Matching
    QUARTERLY OF APPLIED MATHEMATICS VOLUME LVI, NUMBER 3 SEPTEMBER 1998, PAGES 587-600 VARIATIONAL PROBLEMS ON FLOWS OF DIFFEOMORPHISMS FOR IMAGE MATCHING By PAUL DUPUIS (LCDS, Division of Applied Mathematics, Brown University, Providence, RI), ULF GRENANDER (Division of Applied Mathematics, Brown University, Providence, Rl), AND MICHAEL I. MILLER (Dept. of Electrical Engineering, Washington University, St. Louis, MO) Abstract. This paper studies a variational formulation of the image matching prob- lem. We consider a scenario in which a canonical representative image T is to be carried via a smooth change of variable into an image that is intended to provide a good fit to the observed data. The images are all defined on an open bounded set GcR3, The changes of variable are determined as solutions of the nonlinear Eulerian transport equation ==v(rj(s;x),s), r)(t;x)=x, (0.1) with the location 77(0;x) in the canonical image carried to the location x in the deformed image. The variational problem then takes the form arg mm ;||2 + [ |Tor?(0;a;) - D(x)\2dx (0.2) JG where ||v|| is an appropriate norm on the velocity field v(-, •), and the second term at- tempts to enforce fidelity to the data. In this paper we derive conditions under which the variational problem described above is well posed. The key issue is the choice of the norm. Conditions are formulated under which the regularity of v(-, ■) imposed by finiteness of the norm guarantees that the associated flow is supported on a space of diffeomorphisms. The problem (0.2) can Received March 15, 1996.
    [Show full text]
  • On Generalizations of Sylow Tower Groups
    Pacific Journal of Mathematics ON GENERALIZATIONS OF SYLOW TOWER GROUPS ABI (ABIADBOLLAH)FATTAHI Vol. 45, No. 2 October 1973 PACIFIC JOURNAL OF MATHEMATICS Vol. 45, No. 2, 1973 ON GENERALIZATIONS OF SYLOW TOWER GROUPS ABIABDOLLAH FATTAHI In this paper two different generalizations of Sylow tower groups are studied. In Chapter I the notion of a fc-tower group is introduced and a bound on the nilpotence length (Fitting height) of an arbitrary finite solvable group is found. In the same chapter a different proof to a theorem of Baer is given; and the list of all minimal-not-Sylow tower groups is obtained. Further results are obtained on a different generalization of Sylow tower groups, called Generalized Sylow Tower Groups (GSTG) by J. Derr. It is shown that the class of all GSTG's of a fixed complexion form a saturated formation, and a structure theorem for all such groups is given. NOTATIONS The following notations will be used throughont this paper: N<]G N is a normal subgroup of G ΛΓCharG N is a characteristic subgroup of G ΛΓ OG N is a minimal normal subgroup of G M< G M is a proper subgroup of G M<- G M is a maximal subgroup of G Z{G) the center of G #>-part of the order of G, p a prime set of all prime divisors of \G\ Φ(G) the Frattini subgroup of G — the intersec- tion of all maximal subgroups of G [H]K semi-direct product of H by K F(G) the Fitting subgroup of G — the maximal normal nilpotent subgroup of G C(H) = CG(H) the centralizer of H in G N(H) = NG(H) the normalizer of H in G PeSy\p(G) P is a Sylow ^-subgroup of G P is a Sy-subgroup of G PeSγlp(G) Core(H) = GoreG(H) the largest normal subgroup of G contained in H= ΓioeoH* KG) the nilpotence length (Fitting height) of G h(G) p-length of G d(G) minimal number of generators of G c(P) nilpotence class of the p-group P some nonnegative power of prime p OP(G) largest normal p-subgroup of G 453 454 A.
    [Show full text]
  • LM2931-N SNOSBE5G –MARCH 2000–REVISED APRIL 2013 LM2931-N Series Low Dropout Regulators Check for Samples: LM2931-N
    LM2931-N www.ti.com SNOSBE5G –MARCH 2000–REVISED APRIL 2013 LM2931-N Series Low Dropout Regulators Check for Samples: LM2931-N 1FEATURES DESCRIPTION The LM2931-N positive voltage regulator features a 2• Very Low Quiescent Current very low quiescent current of 1mA or less when • Output Current in Excess of 100 mA supplying 10mA loads. This unique characteristic and • Input-output Differential Less than 0.6V the extremely low input-output differential required for • Reverse Battery Protection proper regulation (0.2V for output currents of 10mA) make the LM2931-N the ideal regulator for standby • 60V Load Dump Protection power systems. Applications include memory standby • −50V Reverse Transient Protection circuits, CMOS and other low power processor power • Short Circuit Protection supplies as well as systems demanding as much as • Internal Thermal Overload Protection 100mA of output current. • Mirror-image Insertion Protection Designed originally for automotive applications, the LM2931-N and all regulated circuitry are protected • Available in TO-220, TO-92, TO-263, or SOIC-8 from reverse battery installations or 2 battery jumps. Packages During line transients, such as a load dump (60V) • Available as Adjustable with TTL Compatible when the input voltage to the regulator can Switch momentarily exceed the specified maximum operating voltage, the regulator will automatically shut down to protect both internal circuits and the load. The LM2931-N cannot be harmed by temporary mirror-image insertion. Familiar regulator features such as short circuit and thermal overload protection are also provided. The LM2931-N family includes a fixed 5V output (±3.8% tolerance for A grade) or an adjustable output with ON/OFF pin.
    [Show full text]
  • Quasi P Or Not Quasi P? That Is the Question
    Rose-Hulman Undergraduate Mathematics Journal Volume 3 Issue 2 Article 2 Quasi p or not Quasi p? That is the Question Ben Harwood Northern Kentucky University, [email protected] Follow this and additional works at: https://scholar.rose-hulman.edu/rhumj Recommended Citation Harwood, Ben (2002) "Quasi p or not Quasi p? That is the Question," Rose-Hulman Undergraduate Mathematics Journal: Vol. 3 : Iss. 2 , Article 2. Available at: https://scholar.rose-hulman.edu/rhumj/vol3/iss2/2 Quasi p- or not quasi p-? That is the Question.* By Ben Harwood Department of Mathematics and Computer Science Northern Kentucky University Highland Heights, KY 41099 e-mail: [email protected] Section Zero: Introduction The question might not be as profound as Shakespeare’s, but nevertheless, it is interesting. Because few people seem to be aware of quasi p-groups, we will begin with a bit of history and a definition; and then we will determine for each group of order less than 24 (and a few others) whether the group is a quasi p-group for some prime p or not. This paper is a prequel to [Hwd]. In [Hwd] we prove that (Z3 £Z3)oZ2 and Z5 o Z4 are quasi 2-groups. Those proofs now form a portion of Proposition (12.1) It should also be noted that [Hwd] may also be found in this journal. Section One: Why should we be interested in quasi p-groups? In a 1957 paper titled Coverings of algebraic curves [Abh2], Abhyankar conjectured that the algebraic fundamental group of the affine line over an algebraically closed field k of prime characteristic p is the set of quasi p-groups, where by the algebraic fundamental group of the affine line he meant the family of all Galois groups Gal(L=k(X)) as L varies over all finite normal extensions of k(X) the function field of the affine line such that no point of the line is ramified in L, and where by a quasi p-group he meant a finite group that is generated by all of its p-Sylow subgroups.
    [Show full text]
  • A Fast Diffeomorphic Image Registration Algorithm
    www.elsevier.com/locate/ynimg NeuroImage 38 (2007) 95–113 A fast diffeomorphic image registration algorithm John Ashburner Wellcome Trust Centre for Neuroimaging, 12 Queen Square, London, WC1N 3BG, UK Received 26 October 2006; revised 14 May 2007; accepted 3 July 2007 Available online 18 July 2007 This paper describes DARTEL, which is an algorithm for diffeo- Many registration approaches still use a small deformation morphic image registration. It is implemented for both 2D and 3D model. These models parameterise a displacement field (u), which image registration and has been formulated to include an option for is simply added to an identity transform (x). estimating inverse consistent deformations. Nonlinear registration is considered as a local optimisation problem, which is solved using a Φ x x u x 1 Levenberg–Marquardt strategy. The necessary matrix solutions are ð Þ ¼ þ ð Þ ð Þ obtained in reasonable time using a multigrid method. A constant In such parameterisations, the inverse transformation is sometimes Eulerian velocity framework is used, which allows a rapid scaling and approximated by subtracting the displacement. It is worth noting that squaring method to be used in the computations. DARTEL has been this is only a very approximate inverse, which fails badly for larger applied to intersubject registration of 471 whole brain images, and the deformations. As shown in Fig. 1, compositions of these forward resulting deformations were evaluated in terms of how well they encode and “inverse” deformations do not produce an identity transform. the shape information necessary to separate male and female subjects Small deformation models do not necessarily enforce a one-to-one and to predict the ages of the subjects.
    [Show full text]
  • Deep Compositing Using Lie Algebras
    Deep Compositing Using Lie Algebras TOM DUFF Pixar Animation Studios Deep compositing is an important practical tool in creating digital imagery, Deep images extend Porter-Duff compositing by storing multiple but there has been little theoretical analysis of the underlying mathematical values at each pixel with depth information to determine composit- operators. Motivated by finding a simple formulation of the merging oper- ing order. The OpenEXR deep image representation stores, for each ation on OpenEXR-style deep images, we show that the Porter-Duff over pixel, a list of nonoverlapping depth intervals, each containing a function is the operator of a Lie group. In its corresponding Lie algebra, the single pixel value [Kainz 2013]. An interval in a deep pixel can splitting and mixing functions that OpenEXR deep merging requires have a represent contributions either by hard surfaces if the interval has particularly simple form. Working in the Lie algebra, we present a novel, zero extent or by volumetric objects when the interval has nonzero simple proof of the uniqueness of the mixing function. extent. Displaying a single deep image requires compositing all the The Lie group structure has many more applications, including new, values in each deep pixel in depth order. When two deep images correct resampling algorithms for volumetric images with alpha channels, are merged, splitting and mixing of these intervals are necessary, as and a deep image compression technique that outperforms that of OpenEXR. described in the following paragraphs. 26 r Merging the pixels of two deep images requires that correspond- CCS Concepts: Computing methodologies → Antialiasing; Visibility; ing pixels of the two images be reconciled so that all overlapping Image processing; intervals are identical in extent.
    [Show full text]
  • Integer Discrete Flows and Lossless Compression
    Integer Discrete Flows and Lossless Compression Emiel Hoogeboom⇤ Jorn W.T. Peters⇤ UvA-Bosch Delta Lab UvA-Bosch Delta Lab University of Amsterdam University of Amsterdam Netherlands Netherlands [email protected] [email protected] Rianne van den Berg† Max Welling University of Amsterdam UvA-Bosch Delta Lab Netherlands University of Amsterdam [email protected] Netherlands [email protected] Abstract Lossless compression methods shorten the expected representation size of data without loss of information, using a statistical model. Flow-based models are attractive in this setting because they admit exact likelihood optimization, which is equivalent to minimizing the expected number of bits per message. However, conventional flows assume continuous data, which may lead to reconstruction errors when quantized for compression. For that reason, we introduce a flow-based generative model for ordinal discrete data called Integer Discrete Flow (IDF): a bijective integer map that can learn rich transformations on high-dimensional data. As building blocks for IDFs, we introduce a flexible transformation layer called integer discrete coupling. Our experiments show that IDFs are competitive with other flow-based generative models. Furthermore, we demonstrate that IDF based compression achieves state-of-the-art lossless compression rates on CIFAR10, ImageNet32, and ImageNet64. To the best of our knowledge, this is the first lossless compression method that uses invertible neural networks. 1 Introduction Every day, 2500 petabytes of data are generated. Clearly, there is a need for compression to enable efficient transmission and storage of this data. Compression algorithms aim to decrease the size of representations by exploiting patterns and structure in data.
    [Show full text]
  • 7 Homomorphisms and the First Isomorphism Theorem
    7 Homomorphisms and the First Isomorphism Theorem In each of our examples of factor groups, we not only computed the factor group but identified it as isomorphic to an already well-known group. Each of these examples is a special case of a very important theorem: the first isomorphism theorem. This theorem provides a universal way of defining and identifying factor groups. Moreover, it has versions applied to all manner of algebraic structures, perhaps the most famous being the rank–nullity theorem of linear algebra. In order to discuss this theorem, we need to consider two subgroups related to any group homomorphism. 7.1 Homomorphisms, Kernels and Images Definition 7.1. Let f : G ! L be a homomorphism of multiplicative groups. The kernel and image of f are the sets ker f = fg 2 G : f(g) = eLg Im f = ff(g) : g 2 Gg Note that ker f ⊆ G while Im f ⊆ L. Similar notions The image of a function is simply its range Im f = range f, so this is nothing new. You saw the concept of kernel in linear algebra. For example if A 2 M3×2(R) is a matrix, then we can define the linear map 2 3 LA : R ! R : x 7! Ax In this case, the kernel of LA is precisely the nullspace of A. Similarly, the image of LA is the column- space of A. All we are doing in this section is generalizing an old discussion from linear algebra. Theorem 7.2. Suppose that f : G ! L is a homomorphism. Then 1.
    [Show full text]
  • Infinite General Linear Groups
    INFINITE GENERAL LINEAR GROUPS BY RICHARD V. KADISONO 1. Introduction. In this paper, we carry out the investigation, promised in [4](2), of the general linear group 5W9of a factor 5W(3), that is, the group of all invertible operators in 9ïC. The results of this investigation have much in common with the results of [4]; and, where there is some basis for com- parison, namely, in the case where M is a finite factor, there is a great deal of similarity between the present results and those of the classical finite-di- mensional case. There is much greater resemblance between *Mg and 7^, with M a factor of type IL and N a factor of type ln, than there is between the corresponding unitary groups, M« and 7^u, of such factors. This circum- stance results from the fact that the determinant function of [3] takes effect on "M, but not on 5WU(cf. [3, Theorem 4]). The spectral theoretic and ap- proximation techniques of [4] are stock methods in the present paper, and, for this reason, many of the proofs appearing here are reminiscent of those in [4]. There are, however, some striking differences in the methods required for the investigation of the general linear groups and those required for the investigation of the unitary groups. The most minor difference is the fact that the scalars involved do not form a compact set. This, in combination with the fact that the groups we handle are not locally compact, forces us, occasionally, to detour around what we have come to regard, through our experience with locally compact groups, as the most natural approach to the results obtained.
    [Show full text]
  • Math 412. Homomorphisms of Groups: Answers
    (c)Karen E. Smith 2018 UM Math Dept licensed under a Creative Commons By-NC-SA 4.0 International License. Math 412. Homomorphisms of Groups: Answers φ DEFINITION:A group homomorphism is a map G −! H between groups that satisfies φ(g1 ◦ g2) = φ(g1) ◦ φ(g2). DEFINITION: An isomorphism of groups is a bijective homomorphism. φ DEFINITION: The kernel of a group homomorphism G −! H is the subset ker φ := fg 2 G j φ(g) = eH g: φ THEOREM: A group homomorphism G ! H is injective if and only if ker φ = feGg, the trivial group. THEOREM: A non-empty subset H of a group (G; ◦) is a subgroup if and only if it is closed under ◦, and for every g 2 H, the inverse g−1 is in H. A. EXAMPLES OF GROUP HOMOMORPHISMS × 1 (1) Prove that (one line!) GLn(R) ! R sending A 7! det A is a group homomorphism. Find its kernel. (2) Show that the canonical map Z ! Zn sending x 7! [x]n is a group homomorphism. Find its kernel. × (3) Prove that ν : R ! R>0 sending x 7! jxj is a group homomorphism. Find its kernel. (4) Prove that exp : (R; +) ! R× sending x 7! 10x is a group homomorphism. Find its kernel. (5) Consider 2-element group {±} where + is the identity. Show that the map R× ! {±} sending x to its sign is a homomorphism. Compute the kernel. (6) Let σ : D4 ! {±1g be the map which sends a symmetry the square to 1 if the symmetry preserves the orientation of the square and to −1 if the symmetry reserves the orientation of the square.
    [Show full text]