On the Geometry of Data Representations, © 2020 ABSTRACT

Total Page:16

File Type:pdf, Size:1020Kb

On the Geometry of Data Representations, © 2020 ABSTRACT diss. eth no. 26998 ONTHEGEOMETRYOF DATAREPRESENTATIONS A thesis submitted to attain the degree of doctor of sciences of eth zurich¨ (Dr. sc. ETH Zurich)¨ presented by gary becigneul´ Master in Mathematics University of Cambridge born on 11.11.1994 citizen of France Accepted on the recommendation of Prof. Dr. Thomas Hofmann (ETH Zurich)¨ Prof. Dr. Gunnar Ratsch¨ (ETH Zurich)¨ Prof. Dr. Tommi Jaakkola (MIT) Prof. Dr. Regina Barzilay (MIT) 2020 Gary Becigneul:´ On the Geometry of Data Representations, © 2020 ABSTRACT The vast majority of state-of-the-art Machine Learning (ML) methods nowadays internally represent the input data as being embedded in a “continuous” space, i.e. as sequences of floats, where nearness in this space is meant to define semantic or statistical similarity w.r.t to the task at hand. As a consequence, the choice of which metric is used to measure nearness, as well as the way data is embedded in this space, currently constitute some of the cornerstones of building meaningful data representations. Characterizing which points should be close to each other in a given set essentially defines a “geometry”. Interestingly, certain geometric properties may be incompatible in a non-trivial manner − or put an- other way, selecting desired geometric properties may have non-trivial implications. The investigation of which geometric properties are de- sirable in a given context, and how to enforce them, is one of the motivations underlying this thesis. Initially motivated by uncovering how Convolutional Neural Net- work (CNN) disentangle representations to make tangled classes linearly separable, we start our investigation by studying how invariance to nuisance deformations may help untangling classes in a classification problem, by locally contracting and flattening group orbits within data classes. We then take interest into directly representing data in a Riemannian space of choice, with a particular emphasis on hyperbolic spaces, which is known to be better suited to represented tree-like graphs. We develop a word embedding method generalizing GloVe, as well as a new way of measuring entailment between concepts. In the hyperbolic space, we also develop the tools needed to build neural networks, including matrix multiplication, pointwise non-linearity, Multinomial Logistic Regression (MLR) and Gated Recurrent Unit (GRU). Since many more optimization tools are available for Euclidean do- mains than for the hyperbolic space, we needed to adapt some of the most powerful adaptive schemes − Adagrad, Adam, Amsgrad − to such spaces, in order to let our hyperbolic models have a chance to outperform their Euclidean counterparts. We also provide convergence iii guarantees for these new methods, which recover those already known for the particular case of the Euclidean space. Finally, the growing prominence of graph-like data led us to extend some of the most successful Graph Neural Network (GNN) architectures. First, we start by generalizing Graph Convolutional Network (GCN)s to hyperbolic and spherical spaces. We then leveraged Optimal Trans- port (OT) geometry to turn current architectures into a universal ap- proximator, by dispensing with the last node aggregation step yielding the final graph embedding. We hope that this work will help motivate further investigations in these geometrically-flavored directions. iv R ESUM´ E´ La grande majorite´ des methodes´ d’Apprentissage Automatique (AA) atteignant l’etat-de-l’art´ de nos jours represente´ les donnees´ rec¸ues en entree´ dans un espace “continu”, i.e. comme une sequence´ de nombres decimaux,´ dans un espace ou` la proximite´ est vouee´ a` definir´ la similarite´ semantique´ ou statistique par rapport a` la tache.ˆ En consequence,´ le choix de la metrique´ utilisee´ pour mesurer la proximite,´ ainsi que la fac¸on dont les donnees´ sont plongees´ dans cet espace, constituent actuellement l’une des pierres angulaires de l’edification´ de bonnes representations´ des donnees.´ Characteriser´ quels points doivent etreˆ proches les uns des autres dans un ensemble definit´ essentiellement ce qu’on appelle une geometrie´ . Il se trouve que certaines propriet´ es´ geom´ etriques´ peuvent etreˆ incom- patibles d’une fac¸on non-triviale − ou, dit autrement, la selection´ de cer- taines propriet´ es´ geom´ etriques´ peut avoir des implications non-triviales. La recherche de quelles propriet´ es´ geom´ etriques´ sont souhaitables dans un contexte donne,´ et de comment les obtenir, est l’une des motivations principales sous-jacentes a` cette these.` Initiallement motive´ par la decouverte´ de comment les reseaux´ de neurones a` convolutions parviennent a` desenchev´ etrerˆ les donnees´ afin de rendre des classes enchevetrˆ ees´ linearement separables,´ nous commenc¸ons notre recherche en etudiant´ comment l’invariance a` des transformations nuisibles peut aider a` desenchev´ etrerˆ les classes d’un probleme` de classification, en contractant et en applatissant locallement des orbites de groupes au sein des classes de donnees.´ Nous nous interessons´ ensuite a` representer´ directement les donnees´ dans un espace Riemannien de choix, avec une emphase particuliere` sur les espaces hyperboliques, qui sont connus pour etreˆ mieux adaptes´ a` representer´ les graphes ayant une structure arborescente sous-jacente. Nous developpons´ une methode´ de word embedding gen´ eralisant´ GloVe, ainsi qu’une nouvelle fac¸on de mesurer les relations asymetriques´ d’inclusion semantique´ entre les conceptes. Nous developpons´ egalement´ les outils necessaires´ a` la construction de reseaux´ de neurones dans les espaces hyperboliques: multiplication matricielle, application de non-linearit´ e´ ponctuelle, regression´ multi-logistique et GRU. v Comme il existe bien davantage d’outils d’optimisation pour les espaces Euclidiens que pour les espaces hyperboliques, nous avions besoin d’adapter certaines des methodes´ adaptives les plus perfor- mantes − Adagrad, Adam, Amsgrad − a` ces espaces, afin de per- mettre a` nos modeles` hyperboliques d’avoir une chance de performer superieurement´ a` leur analogue Euclidien. Nous prouvons ainsi des garanties de convergence pour ces nouvelles methodes,´ qui recouvrent celles dej´ a` connues pour le cas particulier de l’espace Euclidien. Enfin, la presence´ accrue de donnees´ sous forme de graphe nous a conduit a` etendre´ certaines des architectures de reseaux´ de neurones de graphes les plus puissantes. En premier lieu, nous commenc¸ons par gen´ eraliser´ les reseaux´ de neurones de graphes a` convolutions, aux espaces hyperboliques et spheriques.´ Ensuite, nous faisons appel a` la geom´ etrie´ du transport optimal pour transformer les architec- tures courantes en approximateur universel, en supprimant la derniere` aggregation´ des representations´ internes des noeuds du graphe qui avant resultait´ en la representation´ finale du graphe. Nous esperons´ que ceci contribuera a` motiver davantage d’explorations dans ces directions de recherche a` tendance geom´ etrique.´ vi PUBLICATIONS The material presented in this thesis has in parts been published in the following publications: 1. Octavian-Eugen Ganea 1, Gary Becigneul´ 1 and Thomas Hofmann. “Hyperbolic Neural Networks.” NeurIPS 2018: Advances in Neural Information Processing Systems. [GBH18c]. 2. Gary Becigneul´ and Octavian-Eugen Ganea. “Riemannian Adaptive Optimization Methods.” ICLR 2019: International Conference on Learning Representations. [BG19]. 3. Alexandru Tifrea 1, Gary Becigneul´ 1 and Octavian-Eugen Ganea 1. “Poincare´ GloVe: Hyperbolic Word Embeddings.” ICLR 2019: International Conference on Learning Representations. [TBG19]. 4. Gregor Bachmann 1, Gary Becigneul´ 1 and Octavian-Eugen Ganea. “Constant Curvature Graph Convolutional Networks.” ICML 2020: International Conference on Machine Learning. [BBG20]. In part currently under review: 5. Gary Becigneul´ 1, Octavian-Eugen Ganea 1, Benson Chen 1, Regina Barzilay, Tommi Jaakkola. “Optimal Transport Graph Neural Networks.” Under Review at ICLR 2021. [Bec+´ 20]. And in part unpublished: 6. Gary Becigneul.´ “On the Effect of Pooling on the Geometry of Representations.” arXiv preprint arXiv:1703.06726, 2017. [Bec´ 17]. 1Equal contribution. vii The following publications were part of my PhD research and present results that are supplemental to this work or build upon its results, but are not covered in this dissertation: 7. Octavian-Eugen Ganea, Gary Becigneul´ and Thomas Hofmann. “Hyperbolic Entailment Cones for Learning Hierarchical Em- beddings.” ICML 2018: International Conference on Machine Learning . [GBH18a]. 8. Octavian-Eugen Ganea, Sylvain Gelly, Gary Becigneul´ and Aliak- sei Severyn. “Breaking the Softmax Bottleneck via Learnable Monotonic Point- wise Non-linearities.” ICML 2019: International Conference on Machine Learning. [Gan+19]. 9. Ondrej Skopek, Octavian-Eugen Ganea and Gary Becigneul.´ “Mixed-Curvature Variational Autoencoders.” ICLR 2020: International Conference on Learning Representations. [SGB20]. 10. Foivos Alimisis, Antonio Orvieto, Gary Becigneul´ and Aurelien´ Lucchi. “A Continuous-time Perspective for Modeling Acceleration in Riemannian Optimization AISTATS 2020: International Conference on Artificial Intelligence and Statistics.” [Ali+20b]. 11. Calin Cruceru, Gary Becigneul´ and Octavian-Eugen Ganea. “Computationally Tractable Riemannian Manifolds for Graph Embeddings.” AAAI 2021: Association for the Advancement of Artificial Intelligence. [CBG21]. Currently under review are also other parts of my PhD research and present results that are
Recommended publications
  • Generalized Gyrovector Spaces and the Positive Cone of a Unital C
    Generalized gyrovector spaces and the positive cone of Contact Information: ∗ Faculty of Engineering a unital C -algebra Niigata University 8050 2no-cho ikarashi Nishi-ku Niigata-shi, Niigata- ken, Japan Toshikazu Abe Phone: +81 (25) 262 7469 Niigata University, Faculty of Engineering Email: [email protected] Abstract Definition 1. A magma (G; ⊕) is called a gyrogroup if there exists an Note that a real normed space (V; +; ×) is a GGV with ϕ = idV. Results The concept of generalized gyrovecor space (GGV, for short) is a com- element e such that the binary operation ⊕ satisfies the following (G1) On a GGV, we can define the gyromidpoint. It is a algebraic mid- mon generalization of the concept of real normed spaces and of the gyrovector spaces. The addition of a GGV is not necessarily a commutative group but is a to (G5). point on the GGV. On a real normed space, the gyromidpoint is the A Mazur-Ulam Theorem for GGV’s gyrocommutative gyrogroup. A typical example of GGV’s is the positive cone 8 2 ⊕ arithmetric mean. ∗ (G1) a G, e a = a. of a unital C -algebra. (G; ⊕; ⊗) p(a; b) We have a Mazur-Ulam type theorem for GGV’s as follows. This is a (G2) 8a 2 G, 9 ⊖ a s.t. ⊖a ⊕ a = e: Definition 4. Let be a GGV. The gyromidpoint of a; b 2 (G; ⊕; ⊗) is defined as p(a; b) = 1 ⊗ (a ⊞ b), where ⊞ is the generalization of the celebrated Mazur-Ulam theorem. A proof of this 8 2 9 2 ⊕ ⊕ ⊕ ⊕ 2 (G3) a; b; c G, ! gyr[a; b]c G s.t.
    [Show full text]
  • From the Lorentz Transformation Group in Pseudo-Euclidean Spaces to Bi-Gyrogroups
    Mathematics Interdisciplinary Research 1 (2016) 229 − 272 From the Lorentz Transformation Group in Pseudo-Euclidean Spaces to Bi-Gyrogroups Abraham A. Ungar? Abstract The Lorentz transformation of order (m = 1; n), n 2 N, is the well-known Lorentz transformation of special relativity theory. It is a transformation of m=1;n time-space coordinates of the pseudo-Euclidean space R of one time di- mension and n space dimensions (n = 3 in physical applications). A Lorentz transformation without rotations is called a boost. Commonly, the special relativistic boost is parametrized by a relativistically admissible velocity pa- n n rameter v, v 2 Rc , whose domain is the c-ball Rc of all relativistically n n admissible velocities, Rc = fv 2 R : kvk < cg, where the ambient space n R is the Euclidean n-space, and c > 0 is an arbitrarily fixed positive con- stant that represents the vacuum speed of light. The study of the Lorentz transformation composition law in terms of parameter composition reveals that the group structure of the Lorentz transformation of order (m = 1; n) induces a gyrogroup and a gyrovector space structure that regulate the pa- n rameter space Rc . The gyrogroup and gyrovector space structure of the ball n Rc , in turn, form the algebraic setting for the Beltrami-Klein ball model of n hyperbolic geometry, which underlies the ball Rc . The aim of this article is to extend the study of the Lorentz transformation of order (m; n) from m = 1 and n ≥ 1 to all m; n 2 N, obtaining algebraic structures called a bi-gyrogroup and a bi-gyrovector space.
    [Show full text]
  • The First Sharp Gyrotriangle Inequality in Möbius Gyrovector Space (D,⊕,⊗)
    Forum Geometricorum b Volume 17 (2017) 439–447. b b FORUM GEOM ISSN 1534-1178 The First Sharp Gyrotriangle Inequality in Mobius¨ Gyrovector Space (D, ⊕, ⊗) Oguzhan˘ Demirel Abstract. In this paper we present two different inequalities (with their reverse inequalities) written in a common type by using classical vector addition “+” and classical multiplication “·” in Euclidean geometry and by using Mobius¨ addi- tion “⊕” and Mobius¨ scalar multiplication “⊗” in Mobius¨ gyrovector space (D, ⊕, ⊗). It is known that this Mobius¨ gyrovector space form the algebraic setting for the Poincare´ disc model of hyperbolic geometry, just as vector spaces form the algebraic setting for the standard model of Euclidean geometry. 1. Introduction There are many fundamental inequalities in mathematics and one of them is the famous “Triangle Inequality”. In [7], Kato, Saito and Tamura presented the following “Sharp Triangle Inequality”in a Banach Space as follows: Theorem 1. For all nonzero elements x1,x2, ··· ,xn in a Banach space X, n n xj xj + n − min kxj k kx k 1≤j≤n j=1 j=1 j X X n ≤ kxjk j=1 X n n xj ≤ xj + n − max kxj k kx k 1≤j≤n j=1 j=1 j X X holds. Recently Mitani, Kato, Saito and Tamura [8] presented a new type Sharp Trian- gle Inequality as follows: Publication Date: December 5, 2017. Communicating Editor: Paul Yiu. The author would like to thank Professor Abraham Albert Ungar for his helpful and valuable comments. 440 O. Demirel Theorem 2. For all nonzero elements x1,x2, ··· ,xn in a Banach space X, n ≥ 2 —small n n k ∗ xj ∗ ∗ k xjk + k − (kx k − kx k) kx∗k k k+1 j=1 k=2 j=1 j X X X n ≤ kxjk j X=1 n n n ∗ xj ∗ ∗ ≤k xjk− k − (kx k − kx k) kx∗k n−k n−(k−1) j=1 k=2 j=n−(k−1) j X X X ∗ ∗ ∗ ∗ where x1,x2, ··· ,xn are the rearrangement of x1,x2, ··· ,xn satisfying kx1k ≥ ∗ ∗ ∗ ∗ kx2k≥···≥kxnk and x0 = xn+1 = 0.
    [Show full text]
  • Ordered Gyrovector Spaces
    S S symmetry Article Ordered Gyrovector Spaces Sejong Kim Department of Mathematics, Chungbuk National University, Cheongju 28644, Korea; [email protected] Received: 5 June 2020; Accepted: 18 June 2020; Published: 22 June 2020 Abstract: The well-known construction scheme to define a partial order on a vector space is to use a proper convex cone. Applying this idea to the gyrovector space we construct the partial order, called a gyro-order. We also give several inequalities of gyrolines and cogyrolines in terms of the gyro-order. Keywords: gyrogroup; gyrovector space; gyroline; cogyroline; partial order 1. Introduction Since finding appropriate algebraic coordinatizations in geometric settings has been attempted first by R. Descartes and P. Fermat, the study of more general algebraic structures have been actively researched. Gyrogroups and gyrovector spaces with the non-associative binary operation ⊕, introduced by A. Ungar [1], are suitable generalization of groups and vector spaces. In particular, gyrovector spaces algebraically regulate many typical examples of analytic hyperbolic geometry. For instance, the Einstein gyrovector space, Möbius gyrovector space, and Proper Velocity (PV, in short) gyrovector space provide algebraic tools to study the Beltrami–Klein, Poincaré ball models, and PV space model of hyperbolic geometry, respectively. As many recent results shows that the theory of gyrogroups and gyrovector spaces can be applied to various areas such as the loop theory, the theory of special relativity, and quantum information, it has been widely studied [2–6]. To capture fundamental analogies between gyrogroups and groups, there is another binary operation, called a coaddition in a gyrogroup: see [1] for more information.
    [Show full text]
  • Pdf 395.98 K
    Mathematics Interdisciplinary Research 1 (2016) 187 − 198 An Extension of Poincaré Model of Hyperbolic Geometry with Gyrovector Space Approach Mahfouz Rostamzadeh? and Sayed-Ghahreman Taherian Abstract The aim of this paper is to show the importance of analytic hyperbolic geometry introduced in [9]. In [1], Ungar and Chen showed that the alge- bra of the group SL(2; C) naturally leads to the notion of gyrogroups and gyrovector spaces for dealing with the Lorentz group and its underlying hy- perbolic geometry. They defined the Chen addition and then Chen model of hyperbolic geometry. In this paper, we directly use the isomorphism prop- erties of gyrovector spaces to recover the Chen’s addition and then Chen model of hyperbolic geometry. We show that this model is an extension of the Poincaré model of hyperbolic geometry. For our purpose we consider the Poincaré plane model of hyperbolic geometry inside the complex open unit disc D. Also we prove that this model is isomorphic to the Poincaré model and then to other models of hyperbolic geometry. Finally, by gyrovec- tor space approach we verify some properties of this model in details in full analogue with Euclidean geometry. Keywords: Hyperbolic geometry, gyrogroup, gyrovector space, Poincaré model, analytic hyperbolic geometry. 2010 Mathematics Subject Classification: 51M10, 51N99, 20N99. 1. Introduction Gyrogroups are noncommutative and nonassociative algebraic structures and this noncommutativity-nonassociativity turns out to be generated by the Thomas pre- cession, well-known in the special theory of relativity. Gyrogroups also revealed themselves to be specially fitting in order to deal with formerly unsolved problems in special relativity (e.g.
    [Show full text]
  • Hyperbolic Deep Neural Networks: a Survey
    JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 1 Hyperbolic Deep Neural Networks: A Survey Wei Peng, Tuomas Varanka, Abdelrahman Mostafa, Henglin Shi, Guoying Zhao* Senior Member, IEEE Abstract—Recently, there has been a rising surge of momentum for deep representation learning in hyperbolic spaces due to their high capacity of modeling data, like knowledge graphs or synonym hierarchies, possessing hierarchical structure. We refer to the model as hyperbolic deep neural network in this paper. Such a hyperbolic neural architecture can potentially lead to drastically more compact models with much more physical interpretability than its counterpart in the Euclidean space. To stimulate future research, this paper presents a coherent and a comprehensive review of the literature around the neural components in the construction of hyperbolic deep neural networks, as well as the generalization of the leading deep approaches to the hyperbolic space. It also presents current applications around various machine learning tasks on several publicly available datasets, together with insightful observations and identifying open questions and promising future directions. Index Terms—Deep Neural Networks, Hyperbolic Geometry, Neural Networks on Riemannian Manifold, Hyperbolic Neural Networks, Poincare´ Model, Lorentz Model, Negative Curvature Geometry. F 1 INTRODUCTION curvature, and it is analogous to a high-dimensional sphere with Earning semantically rich representations of data such as text constant positive curvature. L or images is a central point of interest in current machine In many domains, data is with a tree-like structure or can learning. Recently, deep neural networks [1] have showcased a be represented hierarchically [14]. For instance, social networks, great capability in extracting feature representations and have human skeletons, sentences in natural language, and evolution- dominated many research fields, such as image classification [2], ary relationships between biological entities in phylogenetics.
    [Show full text]
  • Distributivity on the Gyrovector Spaces
    KYUNGPOOK Math. J. 55(2015), 13-20 http://dx.doi.org/10.5666/KMJ.2015.55.1.13 pISSN 1225-6951 eISSN 0454-8124 ⃝c Kyungpook Mathematical Journal Distributivity on the Gyrovector Spaces Sejong Kim Department of Mathematics, Chungbuk National University, Cheongju, 362-763, Korea e-mail : [email protected] Abstract. As a vector space provides a fundamental tool for the study of Euclidean geometry, a gyrovector space provides an algebraic tool for the study of hyperbolic ge- ometry. In general, the gyrovector spaces do not satisfy the distributivity with scalar multiplication. In this article, we see under what condition the distributivity with scalar multiplication is satisfied. 1. Introduction In order to provide an algebraic tool to study Einstein's relativistic velocity sum, A. A. Ungar [2] has introduced a notion of gyrogroup and has developed to- gether the study of analytic hyperbolic geometry. The gyrogroup is a most natural extension of a group into the nonassociative algebra. The associativity (and the commutativity) of group operations is salvaged in a suitably modified form, called a gyroassociativity (and a gyrocommutativity). In Section 2 we introduce a notion of (gyrocommutative) gyrogroup with its examples. Later on it is known that gyrocommutative gyrogroups are equivalent to Bruck loops (see [1]). To elaborate a precise language, we prefix a gyro to terms that de- scribe concepts in Euclidean geometry to mean the analogous concepts in hyperbolic geometry. The prefix gyro stems from Thomas gyration, which is the mathematical abstraction of a special relativistic effect known as Thomas precession. Some gyrocommutative gyrogroups give rise to gyrovector spaces just as some commutative groups give rise to vector spaces.
    [Show full text]
  • A Characteristic of Gyroisometries in Möbius Gyrovector Spaces
    WN North-Western European Journal of Mathematics M EJ A characteristic of gyroisometries in Möbius gyrovector spaces Oğuzhan Demırel1 Received: October 18, 2019/Accepted: January 30, 2019/Online: March 1, 2020 Abstract Hugo Steinhaus (1966a, b) has asked whether inside each acute angled triangle there is a point from which perpendiculars to the sides divide the triangle into three parts of equal areas. In this paper, we prove that f : D D is a gyroisometry (hyperbolic isometry) if, and only if it is a continuous mapping! that preserves the partition of a gyrotriangle (hyperbolic triangle) asked by Hugo Steinhaus. Keywords: Gyrogroups, Möbius Gyrovector Spaces, Gyroisometry. msc: 51B10, 30F45, 20N05, 51M09, 51M10, 51M25. 1 Introduction A Möbius transformation f : C C is a mapping of the form w = [ f1g ! [ f1g (az + b)=(cz + d) satisfying ad bc , 0, where a,b,c,d C. The set of all Möbius transformations Möb(C )−is a group with respect2 to the composition and any f Möb(C ) is conformal,[ f1g i.e. it preserves angles. Let us define 2 [ f1g Ω = S C : S is an Euclidean circle or a Euclidean line . f ⊂ [ f1g [ 1g It is well known that if C Ω and f Möb(C ), then f (C) Ω. There are well- known elementary proofs2 that if f is2 a continuous[ f1g injective map2 of the extended complex plane C that maps circles into circles, then f is Möbius. In addition to this a map is Möbius[ f1g if, and only if it preserves cross ratios. In function theory, it is known that a function f is Möbius if, and only if the Schwarzian derivative of f vanishes when f 0(z) , 0.
    [Show full text]
  • 3 Finitely Generated Gyrovector Subspaces Let (G, ⊕, ⊗) Be a Gyrovector Space
    Accepted Manuscript Finitely generated gyrovector subspaces and orthogonal gyrodecomposition in the Möbius gyrovector space Toshikazu Abe, Keiichi Watanabe PII: S0022-247X(16)30729-6 DOI: http://dx.doi.org/10.1016/j.jmaa.2016.11.039 Reference: YJMAA 20892 To appear in: Journal of Mathematical Analysis and Applications Received date: 3 October 2016 Please cite this article in press as: T. Abe, K. Watanabe, Finitely generated gyrovector subspaces and orthogonal gyrodecomposition in the Möbius gyrovector space, J. Math. Anal. Appl. (2017), http://dx.doi.org/10.1016/j.jmaa.2016.11.039 This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain. Finitely generated gyrovector subspaces and orthogonal gyrodecomposition in the M¨obius gyrovector space Toshikazu Abe and Keiichi Watanabe Toshikazu Abe: College of Engineering, Ibaraki University, Hitachi, Ibaraki 316- 8511, Japan E-mail address: [email protected] Keiichi Watanabe: Department of Mathematics, Faculty of Science, Niigata Uni- versity, Niigata 950–2181, Japan E-mail address: [email protected] Key words and phrases. M¨obius gyrogroup, M¨obius gyrovector space Mathematical Subject Classification (2010). Primary 51M10; Secondary 20N05, 46C99, 51P05. Abstract. We will show that any finitely generated gyrovector subspace in the M¨obius gyrovector space coincides with the intersection of the vector subspace generated by the same generators and the M¨obius ball.
    [Show full text]
  • Differentiating Through the Fréchet Mean
    Differentiating through the Frechet´ Mean Aaron Lou * 1 Isay Katsman * 1 Qingxuan Jiang * 1 Serge Belongie 1 Ser-Nam Lim 2 Christopher De Sa 1 Abstract Recent advances in deep representation learning on Riemannian manifolds extend classical deep learning operations to better capture the geom- etry of the manifold. One possible extension is the Frechet´ mean, the generalization of the Eu- clidean mean; however, it has been difficult to apply because it lacks a closed form with an eas- ily computable derivative. In this paper, we show how to differentiate through the Frechet´ mean for arbitrary Riemannian manifolds. Then, fo- cusing on hyperbolic space, we derive explicit gradient expressions and a fast, accurate, and hyperparameter-free Frechet´ mean solver. This fully integrates the Frechet´ mean into the hyper- bolic neural network pipeline. To demonstrate this integration, we present two case studies. First, Figure 1. Depicted above is the Frechet´ mean, µ, of three points, we apply our Frechet´ mean to the existing Hyper- x1; x2; x3 in the Lorentz model of hyperbolic space. As one can see, the Frechet´ mean conforms with the geometry of the bolic Graph Convolutional Network, replacing its hyperboloid and is vastly different from the standard Euclidean projected aggregation to obtain state-of-the-art re- mean. sults on datasets with high hyperbolicity. Second, to demonstrate the Frechet´ mean’s capacity to gen- eralize Euclidean neural network operations, we space, in which distances grow exponentially as one moves develop a hyperbolic batch normalization method away from the origin. Such a geometry is naturally equipped that gives an improvement parallel to the one ob- to embed trees, since if we embed the root of the tree near served in the Euclidean setting1.
    [Show full text]
  • Multicoloured Random Graphs: the Random Dynamics Program
    Multicoloured Random Graphs: The Random Dynamics Program Sam Tarzi London, England arXiv:1407.6681v1 [physics.gen-ph] 15 Jul 2014 Copyright Notice © 2014 Sam Tarzi All Rights Reserved Sam Tarzi is identified as the author of this work in accordance with the Copyright, Designs and Patents Act, 1988. No part of this publication may be reproduced or stored in a retrieval system, nor transmitted in any form or by any means - digital scanning, mechanical, photo-copying, recording, or by any other media - without prior permission of the copyright owner. Every effort has been made to trace copyright ownership and permissions are confirmed in the text. Any omission brought to the notice of the publishers will be acknowledged in future impressions. i Abstract Multicoloured Random Graphs: The Random Dynamics Program The Random Dynamics program of Holger Nielsen and co-workers is a proposal to explain the origin of all symmetries, including Lorentz and gauge invariance without appeal to any fundamental invariance of the laws of nature, and to derive the known physical laws in such a way as to be almost unavoidable. The philosophy of Random Dynamics is that the most useful as- sumption that we can make about the fundamental laws is that they are random and then to see how known physics like mechanics and relativity follow from them. It is believed that almost any model of these regularities would appear in some limit for example as energies become small. Almost all theories or models at the fundamental level could then explain known physics and specific models would arise from chaotic fundamental physics which has as few starting assumptions as possible.
    [Show full text]
  • The Hyperbolic Superposition General Quantum State
    The Hyperbolic Superposition of a General Quantum State Author Grant Hubbard ABSTRACT: This dissertation documents an inquiry into the geometric structure of the Spaces used to represent a 2-Dimensional Quantum State Space as it relates to Quantum Computation. In particular we investigate the possible applications of Professor NJ Wildberger's Universal Hyperbolic Geometry and discover a genuinely intuitive method for constructing and hence visualising an otherwise abstract algebraic discipline. Submitted to the University of Hertfordshire in partial fulfilment of the requirements of the degree of MSc by Research 30th October 2015 Contents 1 Introduction 4 1.1 Objectives . .4 1.2 Quantum Considerations . .5 1.3 The Structure of the Thesis . .5 1.3.1 Notation . .7 2 Preliminaries 8 2.1 Fields . .8 2.1.1 Ordered Fields . .9 2.2 Spaces . .9 2.2.1 Vector Spaces . 10 2.2.2 Euclidean n-Space En ......................... 10 2.2.3 Spherical n-Space Sn .......................... 11 2.2.4 Lorentzian n-Space Ln ......................... 11 2.2.5 Hyperbolic n-Space Hn ......................... 11 2.2.6 Metric Spaces . 12 2.2.7 Euclidean Metric Spaces . 12 2.2.8 Elliptic Metric Spaces . 12 2.2.9 Hyperbolic Metric Spaces . 13 2.2.10 Complete Metric Spaces . 13 2.2.11 Hilbert Space H ............................. 14 2.3 Groups . 14 2.3.1 Orthogonal Groups . 15 2.3.2 Unitary Groups . 15 2.3.3 Matrices . 16 2.3.4 The Spinor Matrix . 17 2.4 Projective Geometry . 18 2.4.1 Consequences of Stereographic/Gnomonic Projections . 19 3 A Quantum Mechanical System 20 3.1 Introducing the Qubit .
    [Show full text]