A Computable Measure of Algorithmic Probability by Finite Approximations with an Application to Integer Sequences

A Computable Measure of Algorithmic Probability by Finite Approximations with an Application to Integer Sequences

Hindawi Complexity Volume 2017, Article ID 7208216, 10 pages https://doi.org/10.1155/2017/7208216 Research Article A Computable Measure of Algorithmic Probability by Finite Approximations with an Application to Integer Sequences Fernando Soler-Toscano1,2 and Hector Zenil2,3,4 1 Grupo de Logica,´ Lenguaje e Informacion,´ Universidad de Sevilla, Sevilla, Spain 2Algorithmic Nature Group, LABORES, Paris, France 3Algorithmic Dynamics Lab, Center for Molecular Medicine, Science for Life Laboratory (SciLifeLab), Department of Medicine, Solna, Karolinska Institute, Stockholm, Sweden 4Group of Structural Biology, Department of Computer Science, University of Oxford, Oxford, UK Correspondence should be addressed to Hector Zenil; [email protected] Received 19 February 2017; Revised 22 June 2017; Accepted 7 August 2017; Published 21 December 2017 Academic Editor: Giacomo Innocenti Copyright © 2017 Fernando Soler-Toscano and Hector Zenil. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Given the widespread use of lossless compression algorithms to approximate algorithmic (Kolmogorov-Chaitin) complexity and that, usually, generic lossless compression algorithms fall short at characterizing features other than statistical ones not different from entropy evaluations, here we explore an alternative and complementary approach. We study formal properties of a Levin- inspired measure calculated from the output distribution of small Turing machines. We introduce and justify finite approxima- tions that have been used in some applications as an alternative to lossless compression algorithms for approximating algorithmic (Kolmogorov-Chaitin) complexity. We provide proofs of the relevant properties of both and and compare them to Levin’s Universal Distribution. We provide error estimations of with respect to . Finally, we present an application to integer sequences from the On-Line Encyclopedia of Integer Sequences, which suggests that our AP-based measures may characterize nonstatistical patterns, and we report interesting correlations with textual, function, and program description lengths of the said sequences. 1. Algorithmic Information Measures as input and produces the integer () as output. This is usually considered a major problem, but one ought to Central to Algorithmic Information Theory is the definition expect a universal measure of randomness to have such a of algorithmic (Kolmogorov-Chaitin or program-size) com- property. plexity [1, 2]: In previous papers [4, 5], we have introduced a novel method to approximate based on the seminal concept of () = min { ,()=}, (1) algorithmic probability (or AP), introduced by Solomonoff where is a program that outputs running on a universal [6] and further formalized by Levin [3] who proposed the Turing machine and || is the length in bits of . The concept of uncomputable semimeasures and the so-called measure was first conceived to define randomness andis Universal Distribution. today the accepted objective mathematical measure of ran- Levin’s semimeasure (it is called a semimeasure because, domness, among other reasons, because it has been proven unlike probability measures, the sum is never 1. This is due to be mathematically robust [3]. In the following, we use to the Turing machines that never halt) m defines the so- () instead of () because the choice of is only relevant called Universal Distribution [7], with the value m() being up to an additive constant (invariance theorem). A technical the probability that a random program halts and produces inconvenience of as a function taking to be the length of running on a universal Turing machine .Thechoiceof is the shortest program that produces is its uncomputability. only relevant up to a multiplicative constant, so we will simply In other words, there is no program that takes a string write m instead of m. 2 Complexity It is possible to use m() to approximate () by means of prefix-free set of programs (Section 3). Then, in Section 4, the following theorem. we define a computable algorithmic probability measure based on our Turing machine formalism and prove its main Theorem 1 (algorithmic coding theorem [3]). There is a properties, both for andforfiniteapproximations.In constant such that Section5,wecompute5, compare it with our previous distribution (5) [5],andestimatetheerrorin5 as an − m () −() <. log2 (2) approximation to . We finish with some comments in Section 7. This implies that if a string has many descriptions (high value of m(),asthestringisproducedmanytimes,which − m() m() < 1 implies a low value of log2 ,giventhat ), 2. The Turing Machine Formalism it also has a short description (low value of ()). This is because the most frequent strings produced by programs of We denote by (, 2) theclass(orspace)ofall-state 2-symbol length are those which were already produced by programs Turing machines (with the halting state not included among of length −1, as extra bits can produce redundancy in the states) following the Busy Beaver Turing machine an exponential number of ways. On the other hand, strings formalism as defined by Rado[18].BusyBeaverTuring´ produced by programs of length which could not be machines are deterministic machines with a single head and a produced by programs of length −1are less frequently singletapeunboundedinbothdirections.Whenthemachine produced by programs of length , as only very specific entersthehaltingstate,theheadnolongermovesandthe programs can generate them (see Section 14.6 in [8]). This output is considered to comprise only the cells visited by theorem elegantly connects probability to complexity—the theheadpriortohalting.Formally,wehavethefollowing frequency (or probability) of occurrence of a string with definition. its algorithmic (Kolmogorov-Chaitin) complexity. It implies that [4] one can calculate the Kolmogorov complexity of a Definition 2 (Turing machine formalism). We designate as string from its frequency [4], simply rewriting the formula as (, 2) the set of Turing machines with two symbols {0, 1} and states {1,...,}plus a halting state 0. These machines have () =−log2m () +(1) . (3) 2 entries (1,1) (for ∈ {1,...,} and ∈{0,1})inthe transition table, each with one instruction that determines Thankstothiselegantconnectionestablishedby(2)between their behavior. Such entries are represented by algorithmic complexity and probability, our method can attempt to approximate an algorithmic probability measure (1,1) → (2,2,) , (4) by means of finite approximations using a fixed model of computation. The method is called the Coding Theorem Method (CTM) [5]. where 1 and 1 are, respectively, the current state and the ( , ,) In this paper, we introduce ,acomputableapprox- symbol being read and 2 2 represents the instruction imation to m which can be used to approximate by to be executed: 2 is the new state, 2 is the symbol to write, 0 =0 means of the algorithmic coding theorem. Computing () and is the direction. If 2 is the halting state ,then ; 1 −1 requires the output of a numerable infinite number of otherwise is (right) or (left). Turing machines, so we first undertake the investigation of Proposition 3. Machines in (, 2) can be enumerated from 0 finite approximations () that require only the output of (4 + 2)2 −1 machines up to states. A key property of m and is to . theiruniversality:thechoiceoftheTuringmachineusedto Proof. Given the constraints in Definition 2, for each transi- compute the distribution is only relevant up to an (additive) tion of a Turing machine in (, 2),thereare4 + 2 different constant, independent of the objects. The computability of instructions (2,2,).Theseare2 instructions when 2 =0 this measure implies its lack of universality. The same is (given that =0is fixed and 2 can be one of the two possible true when using common lossless compression algorithms symbols) and 4 instructions if 2 =0̸ (2 possible moves, to approximate , but on top of their nonuniversality in the states, and 2 symbols). Then, considering the 2 entries in the algorithmic sense, they are block entropy estimators as they transition table, traverse files in search of repeated patterns in a fixed-length window to build a replacement dictionary. Nevertheless, this 2 does not prevent lossless compression algorithms from find- |(,) 2 | = (4 +) 2 . (5) ing useful applications in the same way as more algorithmic- basedmotivatedmeasurescancontributeevenifalsolimited. These machines can be enumerated from 0 to |(, 2)|. −1 Indeed, m has found successful applications in cognitive Several enumerations are possible. We can, for example, use sciences [9–13] and in financial time series research [14] and a lexicographic ordering on transitions (4). graph theory and networks [15–17]. However, a thorough investigation to explore the properties of these measures and to provide theoretical error estimations was missing. For the current paper, consider that some enumeration We start by presenting our Turing machine formalism has been chosen. Thus, we use to denote the machine (Section 2) and then show that it can be used to encode a number in (, 2) following that enumeration. Complexity 3 3. Turing Machines as a Prefix-Free and simulating () iscomputable,giventheenumerationof Set of Programs Turing machines. As an example, this is the binary representation of the We show in this section that the set of Turing machines

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    11 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us