Angular Embedding: a New Angular Robust Principal Component Analysis

Total Page:16

File Type:pdf, Size:1020Kb

Angular Embedding: a New Angular Robust Principal Component Analysis Angular Embedding: A New Angular Robust Principal Component Analysis Shenglan Liu1 *, Yang Yu1 1 School of Computer Science and Technology, Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian 116024, Liaoning, China. fliusl, [email protected] Abstract which optimizes PCs with angular density, is proposed with quadratic optimization of cosine value on hypersphere man- As a widely used method in machine learning, principal com- ifold. Based on the cosine measurement, AE enhances the ponent analysis (PCA) shows excellent properties for dimen- sionality reduction. It is a serious problem that PCA is sen- robustness of `2-norm based PCA and outperforms the ex- sitive to outliers, which has been improved by numerous isting methods. As an eigen decomposition based PCA ap- Robust PCA (RPCA) versions. However, the existing state- proach, AE is improved to reduce the computational com- of-the-art RPCA approaches cannot easily remove or toler- plexity by weighing the dimensionality and the sample size. ate outliers by a non-iterative manner. To tackle this issue, Theoretically, we prove the superiority of quadratic co- this paper proposes Angular Embedding (AE) to formulate sine optimization in AE with analyzing the effects of on the a straightforward RPCA approach based on angular density, determination of PCs and outlier suppression. And in prac- which is improved for large scale or high-dimensional data. tice, the experimental results of AE on both synthetic and Furthermore, a trimmed AE (TAE) is introduced to deal with real data show its effectiveness. Furthermore, to address data data with large scale outliers. Extensive experiments on both with large scale outliers, we propose a pre-trimming the- synthetic and real-world datasets with vector-level or pixel- level outliers demonstrate that the proposed AE/TAE outper- ory based on cosine measurement and propose trimmed AE forms the state-of-the-art RPCA based methods. (TAE). Base on TAE theory, the experiments on background modeling and shadow removal tasks show its superiority. Introduction Related Work As machine leaning is widely used in many applications, In recent years, robust extensions for PCA have attracted in- principal component analysis (PCA) (Wold, Esbensen, and creasing attention due to its wide applications. Most existing Geladi 1987) has already become a remarkable method methods could fall into three categories as follows. for dimensionality reduction (Vasan and Surendiran 2016; Adiwijaya et al. 2018), computer vision (Bouwmans et al. RPCA with low-rank representation. Robust PCA 2018), etc. However, one of the most important issues of (Candes` et al. 2011) provides a new robust approach to re- PCA is that the principal components (PCs) are sensitive cover both the low-rank and the sparse components by de- to outliers (Zhao et al. 2014), which could not be well ad- composing the data matrix as X = L + S with Principal dressed under `2-norm. RPCA based methods (Candes` et al. Component Pursuit. Based on RPCA, many improved ver- 2011; Zhao et al. 2014) enhance the robustness of PCA sions have been proposed including the tensor version (Lu through low-rank decomposition. Besides, iterative sub- et al. 2019) (require more memory for SVD results) to re- space learning (Roweis 1998; Hauberg, Feragen, and Black duce the error of low-dimensional representation caused by 2014) is another approach to realize robust PCs for data min- outliers. Methods such as (Candes` et al. 2011) require iter- arXiv:2011.11013v1 [cs.LG] 22 Nov 2020 ing tasks (e.g., data classification (Xia et al. 2013), cluster- ative calculation SVD of data with cubic time complexity. ing (Ding et al. 2006) and information retrieval (Wang et al. (Xue et al. 2018; Yi et al. 2016) utilize gradient descent to 2015)). However, most existing robust methods for PCA are avoid frequent maxtrix decompositions. Although the above iterative optimization and always free of PCs. The outliers dimensionality reduction method is robust, it is difficult to fit cannot be easily removed or tolerated by non-iterative ap- on large datasets and to design a robust mathematical model proaches, which limits the applications on real-world prob- for capturing the linear transformation matrix between latent lems (e.g., image analysis in medical science (Lazcano et al. and observed variables. All these reasons limit the applica- 2017) or FPGA (Fernandez et al. 2019)). tions of the RPCA methods above. To tackle the problems above, in this paper, a straight- RPCA for subspace learning. Subspace learning aims to forward RPCA approach named Angular Embedding (AE), obtain low-dimensional subspace which spans with robust *Corressponding author. projection matrix. `1-norm based subspace estimation (Ke Copyright © 2021, Association for the Advancement of Artificial and Kanade 2005) is a robust version for standard PCA Intelligence (www.aaai.org). All rights reserved. using alternative convex programming. The determination of robust subspaces with `1-norm based methods, however, geodesic distance Dij between any two normalized samples is computationally expensive. Besides, questionable results ui and uj on the unit (radius R = 1) hypersphere manifold, in some tasks like clustering would be produced since the global solutions are not rotationally invariant. To this end, Dij = Rβij = βij; (1) R1-PCA (Ding et al. 2006) is proposed to improve `1-norm PCA. Unfortunately, R1-PCA is also time-consuming, sim- where βij = hui; uji is in radians. That is, the angular den- ilar to another work named Deterministic High-dimensional sity between two samples can be quantified by measuring the Robust PCA (Feng, Xu, and Yan 2012) which updates the surface density on a unit hypersphere manifold of codimen- weighted covariance matrix with frequent matrix decompo- sion one in D dimensions. The input samples can thus be sition. Furthermore, based on statistical features such as co- firstly normalized by mapping the original D-dimensional variance, Roweis proposed EM PCA (Roweis 1998) using zero-mean samples into a unit hypersphere manifold. For EM algorithm to obtain PCs for Gaussian data. i = 1; 2; ··· ; n, the unit vector ui corresponding to each sample xi can be obtained by computing RPCA with angle measurement. Recently, angle-based methods (Graf, Smola, and Borer 2003; Liu, Feng, and Qiao x u = i : (2) 2014; Wang et al. 2018, 2017) become new approaches in i kx k machine learning. The angle metric between samples is less i 2 sensitive to outliers and has been applied in many domains The normalized none zero mean inputs on (D − 1)-sphere (e.g., RNA structures analysis (Sargsyan, Wright, and Lim can be marked as U = fu1; u2; ··· ; ung. 2012), texture mapping in computer vision (Wilson et al. 2014)). The non-iterative linear dimensionality reduction The leading PC. Given the definition that q is the lead- method (Liu, Feng, and Qiao 2014), which utilizes cosine ing PC, which corresponds to the position of zero angle in value to obtain robust projection matrix, motivates the angle- D dimensions, then each sample ui can be represented as a based RPCA methods. Angle PCA (Wang et al. 2017) em- directional angle θi 2 (−π; π], in radians. To further sim- ployed cotangent value and iterative matrix decomposition plify the calculation, the optimization of angular density are to realize the idea of PCA, which is robust but computa- determined by utilizing the sine value of directional variable tionally expensive. Based on EM PCA (Roweis 1998) and θi instead of the geodesic distance Dij on the hypersphere. (trimmed) averages, a more scalable approach in introduced Then the leading PC can be determined by formulating by Grassmann Average (GA) (Hauberg, Feragen, and Black 2014), which is also related to the angle measurement. For n GA, the drawback is lower parallelization for calculating X 2 q = arg min sin θi large PCs because of iterative matrix multiplications and or- q i=1 thogonalization for PCs. Actually, most RPCA methods is n based on the iterative solutions, which are limited by the X 2 : (3) = arg max cos θi time-consuming steps (e.g., matrix decomposition in PCA) q and unused computational resources for parallelization on i=1 = arg max qT UUT q CPUs and GPUs, etc. q D Angular Robust Principal Component Multiple PCs. Let Q = fq1; ··· ; qdg ⊂ R be the top- d orthogonal PCs. Then the orthogonal projection u^i = Analysis on Hypersphere Manifold T D QQ ui of the sample ui in R can be represented as ` -norm based PCA are sensitive to outliers and can be mis- d 2 u^ = P (uT q )q =(qT q ). We define θ to be the angle led by the Euclidean distance. Many researches on robust i j=1 i j j j j i between the sample ui and its projection u^i. Each qj; j 2 PCA pay attention to characterize the error with `1-norm or 2 Pd 2 other approaches. The proposed AE, which follows `2-norm, f1; ··· ; dg contributes to θi under cos θi = j=1 cos #ij, turns to determine the principal components (PCs) based on where #ij = hui; qji indicates the angle between ui and qj angular density instead of calculating distance. In practice, in D dimensions. Then PCs can be determined by formulat- angular density will not be misled by outliers with large Eu- ing angular density based PCA as clidean distance, especially for the directional outliers (far from the direction of PCs). n d X X 2 Q = arg max cos #ji Angular Density Framework on Hypersphere Q i=1 j=1 : (4) Given a set of zero-mean data X = fx1; x2; ··· ; xng ⊂ T T RD = arg max trace Q UU Q under the assumption of Gaussian distribution, the an- Q gular density can be defined as the number of samples within the unit angle.
Recommended publications
  • Persistence in Algebraic Specifications
    Persistence in Algebraic Specifications Freek Wiedijk Persistence in Algebraic Specifications Persistence in Algebraic Specifications Academisch Proefschrift ter verkrijging van de graad van doctor aan de Universiteit van Amsterdam, op gezag van de Rector Magnificus prof. dr. P.W.M. de Meijer, in het openbaar te verdedigen in de Aula der Universiteit (Oude Lutherse Kerk, ingang Singel 411, hoek Spui), op donderdag 12 december 1991 te 12.00 uur door Frederik Wiedijk geboren te Haarlem Promotores: prof. dr. P. Klint & prof. dr. J.A. Bergstra Faculteit Wiskunde en Informatica Het hier beschreven onderzoek werd mede mogelijk gemaakt door steun van de Ne- derlandse organisatie voor Wetenschappelijk Onderzoek (voorheen de nederlandse organisatie voor Zuiver-Wetenschappelijk Onderzoek) binnen project nummer 612- 317-013 getiteld Executeerbaar maken van algebraïsche specificaties. voor Jan Truijens Contents Contents Acknowledgements 1 Introduction 1 1.1 Motivation and general overview 2 1.2 Examples of erroneous specifications 13 1.3 Background and related work 22 2. Theory 25 2.1 Basic notions 27 2.1.1. Equational specifications 28 2.1.2 Term rewriting systems 29 2.2 Termination 30 2.2.1 Weak and strong termination 33 2.2.2 Path orderings 34 2.2.3 Decidability 40 2.3 Confluence 41 2.3.1 Weak and strong confluence 42 2.3.2 Overlapping 44 2.3.3 Decidability 45 2.4 Reduction strategies 46 2.5 Persistence 48 2.5.1 Weak and strong persistence 52 2.5.2 Term approximations and bases 54 2.5.3 Normal form analysis 56 2.5.4 Decidability 60 2.6 Primitive recursive
    [Show full text]
  • Computations in Algebraic Geometry with Macaulay 2
    Computations in algebraic geometry with Macaulay 2 Editors: D. Eisenbud, D. Grayson, M. Stillman, and B. Sturmfels Preface Systems of polynomial equations arise throughout mathematics, science, and engineering. Algebraic geometry provides powerful theoretical techniques for studying the qualitative and quantitative features of their solution sets. Re- cently developed algorithms have made theoretical aspects of the subject accessible to a broad range of mathematicians and scientists. The algorith- mic approach to the subject has two principal aims: developing new tools for research within mathematics, and providing new tools for modeling and solv- ing problems that arise in the sciences and engineering. A healthy synergy emerges, as new theorems yield new algorithms and emerging applications lead to new theoretical questions. This book presents algorithmic tools for algebraic geometry and experi- mental applications of them. It also introduces a software system in which the tools have been implemented and with which the experiments can be carried out. Macaulay 2 is a computer algebra system devoted to supporting research in algebraic geometry, commutative algebra, and their applications. The reader of this book will encounter Macaulay 2 in the context of concrete applications and practical computations in algebraic geometry. The expositions of the algorithmic tools presented here are designed to serve as a useful guide for those wishing to bring such tools to bear on their own problems. A wide range of mathematical scientists should find these expositions valuable. This includes both the users of other programs similar to Macaulay 2 (for example, Singular and CoCoA) and those who are not interested in explicit machine computations at all.
    [Show full text]
  • PLS Toolbox 7.0 Pltternf - Plots a 3D Ternary Diagram with Frequency of Occurrence
    plotgui - Interactive data viewer. logdecay - Mean centers and variance scales a matrix using the log plttern - Plots a 2D ternary diagram. decay of the variable axis. PLS_Toolbox 7.0 pltternf - Plots a 3D ternary diagram with frequency of occurrence. lsq2top - Fits a polynomial to the top/(bottom) of data. querydb - Executes a query on a database defined by connection mdcheck - Missing Data Checker and infiller. Quick-Reference Card string. med2top - Fits a constant to top/(bottom) of data. Copyright © Eigenvector Research, 2012 reportwriter - Write a summary of the analysis including medcn - Median center scales matrix to median zero. associated figures to html/word/powerpoint. mncn - Scale matrix to mean zero. Help and Information rwb - Red white and blue color map. mscorr - Multiplicative scatter/signal correction (MSC). setpath - Modifies and saves current directory to the MATLAB normaliz - Normalize rows of matrix. helppls - Context related help on the PLS_Toolbox. search path. npreprocess - Preprocessing of multi-way arrays. readme - Release notes for Version 4.1 of PLS_Toolbox. snabsreadr - Reads Stellarnet ABS XY files. oscapp - Applies OSC model to new data. demos - Demo list for the PLS_Toolbox. spcreadr - Reads a Galactic SPC file. osccalc - Calculates orthogonal signal correction (OSC). evricompatibility - Tests for inter-product compatibility of trendtool - Univariate trend analysis tool. poissonscale - Perform Poisson scaling with scaling offset. Eigenvector toolboxes. vline - Adds vertical lines to figure at specified locations. polyinterp - Polynomial interpolation, smoothing, and evridebug - Checks the PLS_Toolbox installation for problems. writeasf - Writes AIT ASF files from a dataset object. differentiation. evridir - Locate and or create EVRI home directory. writecsv - Export a DataSet object to a comma-separated values preprocess - Selection and application of standard preprocessing evriinstall - Install Eigenvector Research Product.
    [Show full text]
  • What Can Computer Algebraic Geometry Do Today?
    What can computer Algebraic Geometry do today? Gregory G. Smith Wolfram Decker Mike Stillman 14 July 2015 Essential Questions ̭ What can be computed? ̭ What software is currently available? ̭ What would you like to compute? ̭ How should software advance your research? Basic Mathematical Types ̭ Polynomial Rings, Ideals, Modules, ̭ Varieties (affine, projective, toric, abstract), ̭ Sheaves, Divisors, Intersection Rings, ̭ Maps, Chain Complexes, Homology, ̭ Polyhedra, Graphs, Matroids, ̯ Established Geometric Tools ̭ Elimination, Blowups, Normalization, ̭ Rational maps, Working with divisors, ̭ Components, Parametrizing curves, ̭ Sheaf Cohomology, ঠ-modules, ̯ Emerging Geometric Tools ̭ Classification of singularities, ̭ Numerical algebraic geometry, ̭ ैक़௴Ь, Derived equivalences, ̭ Deformation theory,Positivity, ̯ Some Geometric Successes ̭ GEOGRAPHY OF SURFACES: exhibiting surfaces with given invariants ̭ BOIJ-SÖDERBERG: examples lead to new conjectures and theorems ̭ MODULI SPACES: computer aided proofs of unirationality Some Existing Software ̭ GAP,Macaulay2, SINGULAR, ̭ CoCoA, Magma, Sage, PARI, RISA/ASIR, ̭ Gfan, Polymake, Normaliz, 4ti2, ̭ Bertini, PHCpack, Schubert, Bergman, an idiosyncratic and incomplete list Effective Software ̭ USEABLE: documented examples ̭ MAINTAINABLE: includes tests, part of a larger distribution ̭ PUBLISHABLE: Journal of Software for Algebra and Geometry; www.j-sag.org ̭ CITATIONS: reference software Recent Developments in Singular Wolfram Decker Janko B¨ohm, Hans Sch¨onemann, Mathias Schulze Mohamed Barakat TU Kaiserslautern July 14, 2015 Wolfram Decker (TU-KL) Recent Developments in Singular July 14, 2015 1 / 24 commutative and non-commutative algebra, singularity theory, and with packages for convex and tropical geometry. It is free and open-source under the GNU General Public Licence.
    [Show full text]
  • The Power of Pyramid Decomposition in Normaliz
    Journal of Symbolic Computation 74 (2016) 513–536 Contents lists available at ScienceDirect Journal of Symbolic Computation www.elsevier.com/locate/jsc The power of pyramid decomposition in Normaliz Winfried Bruns a, Bogdan Ichim b, Christof Söger a a Universität Osnabrück, FB Mathematik/Informatik, 49069 Osnabrück, Germany b Simion Stoilow Institute of Mathematics of the Romanian Academy, Research Unit 5, C.P. 1-764, 010702 Bucharest, Romania a r t i c l e i n f o a b s t r a c t Article history: We describe the use of pyramid decomposition in Normaliz, a soft- Received 27 January 2014 ware tool for the computation of Hilbert bases and enumerative Accepted 26 August 2015 data of rational cones and affine monoids. Pyramid decomposition Available online 16 September 2015 in connection with efficient parallelization and streamlined evalua- tion of simplicial cones has enabled Normaliz to process triangula- MSC: ≈ · 11 52B20 tions of size 5 10 that arise in the computation of Ehrhart 13F20 series related to the theory of social choice. 14M25 © 2015 Elsevier Ltd. All rights reserved. 91B12 Keywords: Hilbert basis Ehrhart series Hilbert series Rational polytope Volume Triangulation Pyramid decomposition 1. Introduction Normaliz (Bruns et al., 2015)is a software tool for the computation of Hilbert bases and enumera- tive data of rational cones and affine monoids. In the 17 years of its existence it has found numerous applications; for example, in integer programming (Bogart et al., 2010), algebraic geometry (Craw et al., 2007), theoretical physics (Kappl et al., 2011), commutative algebra (Sturmfels and Welker, 2012) or elimination theory (Emiris et al., 2013).
    [Show full text]
  • Normaliz 3.2.0 Contents
    Normaliz 3.2.0 January 22, 2017 Winfried Bruns, Tim Römer, Richard Sieg and Christof Söger Normaliz 2 team member: Bogdan Ichim http://normaliz.uos.de https://github.com/Normaliz mailto:[email protected] Contents 1. Introduction 7 1.1. The objectives of Normaliz . .7 1.2. Platforms, implementation and access from other systems . .8 1.3. Major changes relative to version 3.1.1 . .9 1.4. Future extensions . .9 1.5. Download and installation . .9 2. Normaliz by examples 10 2.1. Terminology . 10 2.2. Practical preparations . 11 2.3. A cone in dimension 2 . 12 2.3.1. The Hilbert basis . 13 2.3.2. The cone by inequalities . 14 2.3.3. The interior . 15 2.4. A lattice polytope . 17 2.4.1. Only the lattice points . 20 2.5. A rational polytope . 20 2.5.1. The rational polytope by inequalities . 23 2.6. Magic squares . 24 2.6.1. With even corners . 27 2.6.2. The lattice as input . 30 1 2.7. Decomposition in a numerical semigroup . 30 2.8. A job for the dual algorithm . 31 2.9. A dull polyhedron . 32 2.9.1. Defining it by generators . 34 2.10. The Condorcet paradoxon . 35 2.10.1. Excluding ties . 36 2.10.2. At least one vote for every preference order . 37 2.11. Testing normality . 38 2.11.1. Computing just a witness . 39 2.12. Inhomogeneous congruences . 40 2.12.1. Lattice and offset . 41 2.12.2. Variation of the signs . 41 2.13.
    [Show full text]
  • Program Synthesis
    Program Synthesis Program Synthesis Sumit Gulwani Microsoft Research [email protected] Oleksandr Polozov University of Washington [email protected] Rishabh Singh Microsoft Research [email protected] Boston — Delft Foundations and Trends® in Programming Languages Published, sold and distributed by: now Publishers Inc. PO Box 1024 Hanover, MA 02339 United States Tel. +1-781-985-4510 www.nowpublishers.com [email protected] Outside North America: now Publishers Inc. PO Box 179 2600 AD Delft The Netherlands Tel. +31-6-51115274 The preferred citation for this publication is S. Gulwani, O. Polozov and R. Singh. Program Synthesis. Foundations and Trends® in Programming Languages, vol. 4, no. 1-2, pp. 1–119, 2017. ® This Foundations and Trends issue was typeset in LATEX using a class file designed by Neal Parikh. Printed on acid-free paper. ISBN: 978-1-68083-292-1 © 2017 S. Gulwani, O. Polozov and R. Singh All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, mechanical, photocopying, recording or otherwise, without prior written permission of the publishers. Photocopying. In the USA: This journal is registered at the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items for in- ternal or personal use, or the internal or personal use of specific clients, is granted by now Publishers Inc for users registered with the Copyright Clearance Center (CCC). The ‘services’ for users can be found on the internet at: www.copyright.com For those organizations that have been granted a photocopy license, a separate system of payment has been arranged.
    [Show full text]
  • Arxiv:2010.08903V1 [Math.AC]
    STANDARD PAIRS OF MONOMIAL IDEALS OVER NON-NORMAL AFFINE SEMIGROUPS IN SageMath BYEONGSU YU ABSTRACT. We present stdPairs.spyx,a SageMath library to compute standard pairs of a monomial ideal over a pointed (non-normal) affine semigroup ring. Moreover, stdPairs.spyx provides the associated prime ideals, the correspondingmultiplicities, and an irredundant irreducible primary decomposition of a monomial ideal. The library expands on the standardPairs func- tion on Macaulay2 over polynomial rings, and is based on algorithms from [6]. We also provide methods that allow the outputs from this library to be compatible with the Normaliz package of Macaulay2 and SageMath. 1. INTRODUCTION Affine semigroup rings are objects of many studies in combinatorial commutative algebra. The goal of this article is to present the SageMath library stdPairs.spyx, which systematizes computations for monomial ideals in affine semigroup rings. The algorithms implemented here are based on the notion of standard pairs, introduced for monomial ideals in polynomial rings by [8], and generalized to the semigroup ring case in [6]. Standard pairs are a combinatorial structure that contains information on primary and irreducible decompositions of monomial ideals, as well as multiplicities. One of the main contributions of [6] is that standard pairs and the associated algebraic concepts can be effectively computed over affine semigroup rings. The SageMath library stdPairs.spyx implements the algorithms of [6] to calculate standard pairs for monomial ideals in any pointed (non-normal) affine semigroup ring. This library can be regarded as a generalization of standardPairs function in Macaulay2 implemented by [4]. This library can be obtained via https://github.com/byeongsuyu/StdPairs.
    [Show full text]
  • Arxiv:2010.15331V2 [Math.AC] 19 Nov 2020 Rlqeto Nivratter.Ti Ril Ecie Significant a Describes R Article Action This Group Theory
    THE INVARIANTRING PACKAGE FOR MACAULAY2 LUIGI FERRARO, FEDERICO GALETTO, FRANCESCA GANDINI, HANG HUANG, MATTHEW MASTROENI, AND XIANGLONG NI Abstract. We describe a significant update to the existing InvariantRing package for Macaulay2. In addition to expanding and improving the methods of the existing package for actions of finite groups, the updated package adds func- tionality for computing invariants of diagonal actions of tori and finite abelian groups as well as invariants of arbitrary linearly reductive group actions. The implementation of the package has been completely overhauled with the aim of serving as a unified resource for invariant theory computations in Macaulay2. 1. Introduction Let G be a group acting linearly on an n-dimensional vector space V over a field K via v 7→ g · v for g ∈ G and v ∈ V . The action of G on V induces an action −1 of G on the polynomial ring K[V ]= K[x1,...,xn] by (g · f)(x)= f(g · x). By a classical result of Hilbert (see [DK15, 2.2.10]), the subring of invariant polynomials K[V ]G = {f ∈ K[V ] | g · f = f, ∀g ∈ G} is always a finitely generated K-algebra provided that G is linearly reductive and V is a rational representation of G.1 A linearly reductive group is a group that can be identified with a Zariski-closed subgroup of some general linear group GLn = GLn(K) and has good representation-theoretic properties while still being general enough to encompass all finite groups, all tori (K×)r, and all semisimple Lie groups (at least if char(K) = 0).
    [Show full text]
  • An Introduction to Zeta Tobias Rossmann National University of Ireland, Galway
    An introduction to Zeta Tobias Rossmann National University of Ireland, Galway Düsseldorf September 2018 Zeta (2014–) ... is a package for Sage (http://www.sagemath.org/) (which also relies on Singular, Normaliz, LattE, ...) ... provides methods for computing various types (ask, subobject, ...) of zeta functions in "fortunate cases" ... is freely available (http://www.maths.nuigalway.ie/~rossmann/Zeta/) ... implements techniques outlined here (https://doi.org/10.1007/978-3-319- 70566-8_25) Sage (2005–) is a free open-source mathematics software system licensed under the GPL. It builds on top of many existing open-source packages [...] Mission: Creating a viable free open source alternative to Magma, Maple, Mathematica and Matlab. Freely available (http://www.sagemath.org/) Based upon Python (https://www.python.org/) Includes GAP (https://www.gap-system.org/), Singular (http://www.singular.uni- kl.de/), ... Runs on Linux, Mac OS X, and Windows Command-line ("interactive shell") or browser-based ("notebook") interface Sage as a symbolic calculator −1 M (Z ) 1−p T Recall: the ask zeta function of 1 p is (1−T)2 . In [1]: var('p t') # create symbolic variables Z = (1-p^(-1)*t)/(1-t)^2 In [2]: Z.series(t,3) # power series expansion up to given order Out[2]: 1 + (-1/p + 2)*t + (-2/p + 3)*t^2 + Order(t^3) In [3]: Z(p=1) # substitution Out[3]: -1/(t - 1) In [4]: Z+1 # arithmetic Out[4]: -(t/p - 1)/(t - 1)^2 + 1 In [5]: _.simplify_full() Out[5]: (p*t^2 - (2*p + 1)*t + 2*p)/(p*t^2 - 2*p*t + p) In [6]: _.factor() Out[6]: (p*t^2 - 2*p*t + 2*p - t)/(p*(t - 1)^2) Installing Zeta Linux-only at the moment.
    [Show full text]
  • Normaliz 2013-2016
    NORMALIZ 2013–2016 WINFRIED BRUNS, RICHARD SIEG, AND CHRISTOF SOGER¨ ABSTRACT. In this article we describe mathematically relevant extensions to Normaliz that were added to it during the support by the DFG SPP “Algorithmische und Experi- mentelle Methoden in Algebra, Geometrie und Zahlentheorie”: nonpointed cones, ratio- nal polyhedra, homogeneous systems of parameters, bottom decomposition, class groups and systems of module generators of integral closures. 1. INTRODUCTION The software package Normaliz [12] has been developed by the algebra and discrete mathematics group at Osnabr¨uck since 1998. It is a tool for the computation of lattice points in rational polyhedra. Meanwhile it has been cited about 130 times in the literature FIGURE 1. Lattice points in a polyhedron (see [12]) with applications to algebraic geometry, commutative algebra, polytope theory, arXiv:1611.07965v2 [math.CO] 3 Mar 2017 integer programming, combinatorial topology, group theory, theoretical physics and other areas. There exist interfaces to the major computer algebra systems CoCoA [6], [2], GAP [20], Macaulay2 [19] and Singular [18] and to polymake [21], a comprehensive tool for the computation of polytopes. Normaliz is used by other packages, notably by Regina [17], a tool for the exploration of 3-manifolds, and by SecDec [7] in the computation of multiscale integrals. 2010 Mathematics Subject Classification. 52B20, 13F20, 14M25, 91B12. Key words and phrases. Hilbert basis, Hilbert series, rational cone, rational polyhedron, bottom decom- position, class group, triangulation, linear diophantine system. The second author was partially supported by the German Research Council DFG-GRK 1916 and a doctoral fellowship of the German Academic Exchange Service. 1 2 W.BRUNS,R.SIEG,ANDC.SOGER¨ During the second half of the SPP 1489 “Algorithmische und Experimentelle Metho- den in Algebra, Geometrie und Zahlentheorie” Normaliz was supported by the SPP.
    [Show full text]
  • Fast Prototyping and Flexible C++ Library for Computations in Commutative Algebra
    CoCoA and CoCoALib: Fast Prototyping and Flexible C++ Library for Computations in Commutative Algebra John Abbott Anna Maria Bigatti Institut fur¨ Mathematik Dipartimento di Matematica Universitat¨ Kassel Universita` degli Studi di Genova, Italy Kassel, Germany http://www.dima.unige.it/∼bigatti http://www.dima.unige.it/∼abbott Abstract—The CoCoA project began in 1987, and conducts (with two closely related applications being the interactive research into Computational Commutative Algebra (from which system and the compute server). This initial decision, together its name comes) with particular emphasis on Grobner¨ bases with the desire to help the software prosper, has many im- of ideals in multivariate polynomial rings, and related areas. A major output of the project is the CoCoA software, including the plications: e.g. designing a particularly clean interface for all CoCoA-5 interactive system and the CoCoALib C++ library. The functions with comprehensive documentation. This cleanliness software is open-source (GPL v.3), and under continual, active makes it easy to integrate CoCoALib into other software in a development. We give a summary of the features of the software trouble-free manner. likely to be relevant to the SC-Square community. Many computer algebra systems have a two-tier structure: I. INTRODUCTION a compiled kernel with interpreter, and an external library CoCoA-5 is a well-established Computer Algebra System of interpretable code for more advanced functions. Our aim specialized in operations on polynomial ideals, and a number with CoCoALib is different: we plan to have all functions of allied operations (e.g. factorization, and linear algebra). implemented directly in the C++ library so that all features It offers a dedicated, mathematician-friendly programming are available to programs which link to CoCoALib.
    [Show full text]