TA Lecture 12 Compression.Pptx

Total Page:16

File Type:pdf, Size:1020Kb

TA Lecture 12 Compression.Pptx 27.11.2012 Problem • Compress Text Algorithms (6EAP) – Text – Images, video, sound, … Compression • Reduce space, efficient communicaon, etc… Jaak Vilo – Data deduplicaon 2012 fall • Exact compression/decompression Jaak Vilo MTAT.03.190 Text Algorithms 1 • Lossy compression Links • Managing Gigabytes: • hUp://datacompression.info/ Compressing and Indexing • hUp://en.wikipedia.org/wiki/Data_compression Documents and Images • Data Compression Debra A. Lelewer and Daniel S. Hirschberg • Ian H. WiUen, Alistair Moffat, – hUp://www.ics.uci.edu/~dan/pubs/DataCompression.html Timothy C. Bell • Compression FAQ hUp://www.faqs.org/faqs/compression-faq/ • Hardcover: 519 pages Publisher: Morgan Kaufmann; 2nd Revised • Informaon Theory Primer With an ediNon ediNon (11 May 1999) Appendix on Logarithms by Tom Language English ISBN-10: Schneider hUp://www.lecb.ncifcrf.gov/~toms/paper/primer/ 1558605703 • hUp://www.cbloom.com/algs/index.html Problem What it’s about? • Informaon transmission • Eliminaon of redundancy • Informaon storage • Being able to predict… • The data sizes are huge and growing – fax - 1.5 x 106 bit/page • Compression and decompression – photo: 2M pixels x 24bit = 6MB – Represent data in a more compact way – X-ray image: ~ 100 MB? – Decompression - restore original form – Microarray scanned image: 30-100 MB – Tissue-microarray - hundreds of images, each tens of MB • Lossy and lossless compression – Large Hardon Collider (CERN) - The device will produce few peta (1015) bytes of stored – Lossless - restore in exact copy data in a year. – Lossy - restore almost the same informaon – TV (PAL) 2.7 · 108 bit/s • Useful when no 100% accuracy needed – CD-sound, super-audio, DVD, ... • voice, image, movies, ... • Decompression is determinisNc (lossy in compression phase) – Human genome – 3.2Gbase. 30x sequencing => 100Gbase + quality info (+ raw data) • Can achieve much more effecNve results – 1000 genomes, all individual genomes … 1 27.11.2012 Methods covered: Model • Code words (Huffman coding) Model Model • Run-length encoding • ArithmeNc coding • Lempel-Ziv family (compress, gzip, zip, pkzip, ...) • Burrows-Wheeler family (bzip2) Compressed data • Other methods, including images Data Data Encoder Decoder • Kolmogorov complexity • Search from compressed texts • Let pS be a probability of message S • The informaon content can be represented in terms of bits • I(S) = -log( pS ) bits • If the p=1 then the informaon content is 0 (no new informaon) – If Pr[s]=1 then I(s) = 0. – In other words, I(death)=I(taxes)=0 • I( heads or tails ) = 1 -- if the coin is fair • Entropy H(S) is the average informaon content of S – H(S) = pS · I(S) = -pS log( pS ) bits hUp://en.wikipedia.org/wiki/Informaon_entropy • Shannon's experiments with human predictors show an informaon rate of between .6 and 1.3 bits per character, depending on the experimental setup; the PPM compression algorithm can achieve a compression rao of 1.5 bits per character. 2 27.11.2012 http://prize.hutter1.net/ • The data compression world is all abuzz about Marcus HuUer’s recently announced 50,000 euro prize for record-breaking data compressors. Marcus, of the Swiss Dalle Molle InsNtute for ArNficial Intelligence, apparently in cahoots with Florida compression maven Ma Mahoney, is offering cash prizes for what amounts to the most • No compression can on average achieve beUer compression impressive ability to compress 100 MBytes of Wikipedia data. (Note that nobody is going to than the entropy win exactly 50,000 euros - the prize amount is prorated based on how well you beat the current record.) • Entropy depends on the model (or choice of symbols) • This prize differs considerably from my Million Digit Challenge, which is really nothing more • Let M={ m , .. m } be a set of symbols of the model A and let than an aempt to silence people foolishly claiming to be able to compress random data. 1 n Marcus is instead looking for the most effecNve way to reproduce the Wiki data, and he’s p(mi) be the probability of the symbol mi pung up real money as an incenNve. The benchmark that contestants need to beat is that set by Ma Mahoney’s paq8f , the current record holder at 18.3 MB. (Alexander • The entropy of the model A, H(M) is -∑i=1..n p(mi) · log( p(mi) ) Ratushnyak’s submission of a paq variant looks to clock in at a Ndy 17.6 MB, and should soon bits be confirmed as the new standard.) • So why is an AI guy inserNng himself into the world of compression? Well, Marcus realizes • Let the message S = s1, .. sk, and every symbol si be in the that good data compression is all about modeling the data. The beUer you understand the model M. The informaon content of model A is -∑ log p data stream, the beUer you can predict the incoming tokens in a stream. Claude Shannon i=1..k empirically found that humans could model English text with an entropy of 1.1 to 1.6 0.6 to (si) 1.3 bits per character, which at at best should mean that 100 MB of Wikipedia data could be reduced to 13.75 7.5 MB, with an upper bound of perhaps 20 16.25 MB. The theory is that • Every symbol has to have a probability, otherwise it cannot be reaching that 7.5 MB range is going to take such a good understanding of the data stream coded if it is present in the data that it will amount to a demonstraon of ArNficial Intelligence. hUp://marknelson.us/2006/08/24/the-huUer-prize/#comment-293 Model Stac or adapNve • Stac model does not change during the Model Model compression • AdapNve model can be updated during the process • Symbols not in message cannot have 0-probability Compressed • Semi-adapNve model works in 2 stages, off-line. Data data Data • First create the code table, then encode the message Encoder Decoder with the code table How to compare compression Shorter code words… techniques? • Rao (t/p) t: original message length • S = 'aa bbb cccc ddddd eeeeee fffffffgggggggg' • p: compressed message length • Alphabet of 8 • In texts - bits per symbol • Length = 40 symbols • The Nme and memory used for compression • Equal length codewords • The Nme and memory used for decompression • 3-bit a 000 b 001 c 010 d 011 e 100 f 101 g 110 • error tolerance (e.g. self-correcNng code) space 110 • S compressed - 3*40 = 120 bits 3 27.11.2012 Run-length encoding AlphabeNcally ordered word-lists • hUp://michael.dipperstein.com/rle/index.html • The string: • "aaaabbcdeeeeefghhhij" • may be replaced with resume 0resume • "a4b2c1d1e5f1g1h3i1j1". retail 2tail • This is not shorter because 1-leUer repeat takes more characters... retain 5n • "a3b1cde4fgh2ij" • Now we need to know which characters are followed by run-length. retard 4rd • E.g. use escape symbols. retire 3ire • Or, use the symbol itself - if repeated, then must be followed by run- length • "aa2bb0cdee3fghh1ij" Coding techniques Variable length encoders • Coding refers to techniques used to encode • How to use codes of variable length? tokens or symbols. • Decoder needs to know how long is the • Two of the best known coding algorithms are symbol Huffman Coding and ArithmeNc Coding. • Coding algorithms are effecNve at • Prefix-free code: no code can be a prefix of compressing data when they use fewer bits another code for high probability symbols and more bits for low probability symbols. Algoritm Shannon-Fano • Calculate the frequencies and probabiliNes of • Input: probabiliNes of symbols symbols: • Output: Codewords in prefix free coding • S = 'aa bbb cccc ddddd eeeeee fffffffgggggggg' freq ratio p(s) a 2 2/40 0.05 1. Sort symbols by frequency b 3 3/40 0.075 c 4 4/40 0.1 2. Divide to two almost probable groups d 5 5/40 0.125 space 5 5/40 0.125 3. First group gets prefix 0, other 1 e 6 6/40 0.15 f 7 7/40 0.175 4. Repeat recursively in each group unNl 1 g 8 8/40 0.2 symbol remains 4 27.11.2012 Example 1 Example 1 Code: a 1/2 0 a 1/2 0 b 1/4 10 b 1/4 10 c 1/8 110 c 1/8 110 d 1/16 1110 d 1/16 1110 e 1/32 11110 e 1/32 11110 f 1/32 11111 f 1/32 11111 Shannon-Fano Shannon-Fano S = 'aa bbb cccc ddddd eeeeee fffffffgggggggg' • S = 'aa bbb cccc ddddd eeeeee fffffffgggggggg' p(s) code • S in compressed is 117 bits g 0.2 00 0.2 0.525 • 2*4 + 3*4 + 4*3 + 5*3 + 5*3 + 6*3 + 7*3 + 8*2 f 0.175 010 0.175 0.325 0.15 = 117 e 0.15 011 1 d 0.125 100 • Shannon-Fano not always opNmal space 0.125 101 0.475 • SomeNmes 2 equal probable groups cannot be c 0.1 110 achieved b 0.075 1110 a 0.05 1111 • Usually beUer than H+1 bits per symbol, when H is entropy. Char Freq Code Huffman code space 7 111 Huffman example a 4 010 • Works the opposite way. e 4 000 f 3 1101 • Start from least probable symbols and separate them with 0 h 2 1010 and 1 (sufix) i 2 1000 • Add probabiliNes to form a "new symbol" with the new m 2 0111 probability n 2 0010 • Prepend new bits in front of old ones. s 2 1011 t 2 0110 l 1 11001 o 1 00110 p 1 10011 r 1 11000 u 1 00111 "this is an example of a huffman tree" x 1 10010 5 27.11.2012 Properes of Huffman coding • Huffman coding is opNmal when the • Error tolerance quite good frequencies of input characters are powers of • In case of the loss, adding or change of a two.
Recommended publications
  • Lecture 24 1 Kolmogorov Complexity
    6.441 Spring 2006 24-1 6.441 Transmission of Information May 18, 2006 Lecture 24 Lecturer: Madhu Sudan Scribe: Chung Chan 1 Kolmogorov complexity Shannon’s notion of compressibility is closely tied to a probability distribution. However, the probability distribution of the source is often unknown to the encoder. Sometimes, we interested in compressibility of specific sequence. e.g. How compressible is the Bible? We have the Lempel-Ziv universal compression algorithm that can compress any string without knowledge of the underlying probability except the assumption that the strings comes from a stochastic source. But is it the best compression possible for each deterministic bit string? This notion of compressibility/complexity of a deterministic bit string has been studied by Solomonoff in 1964, Kolmogorov in 1966, Chaitin in 1967 and Levin. Consider the following n-sequence 0100011011000001010011100101110111 ··· Although it may appear random, it is the enumeration of all binary strings. A natural way to compress it is to represent it by the procedure that generates it: enumerate all strings in binary, and stop at the n-th bit. The compression achieved in bits is |compression length| ≤ 2 log n + O(1) More generally, with the universal Turing machine, we can encode data to a computer program that generates it. The length of the smallest program that produces the bit string x is called the Kolmogorov complexity of x. Definition 1 (Kolmogorov complexity) For every language L, the Kolmogorov complexity of the bit string x with respect to L is KL (x) = min l(p) p:L(p)=x where p is a program represented as a bit string, L(p) is the output of the program with respect to the language L, and l(p) is the length of the program, or more precisely, the point at which the execution halts.
    [Show full text]
  • Shannon Entropy and Kolmogorov Complexity
    Information and Computation: Shannon Entropy and Kolmogorov Complexity Satyadev Nandakumar Department of Computer Science. IIT Kanpur October 19, 2016 This measures the average uncertainty of X in terms of the number of bits. Shannon Entropy Definition Let X be a random variable taking finitely many values, and P be its probability distribution. The Shannon Entropy of X is X 1 H(X ) = p(i) log : 2 p(i) i2X Shannon Entropy Definition Let X be a random variable taking finitely many values, and P be its probability distribution. The Shannon Entropy of X is X 1 H(X ) = p(i) log : 2 p(i) i2X This measures the average uncertainty of X in terms of the number of bits. The Triad Figure: Claude Shannon Figure: A. N. Kolmogorov Figure: Alan Turing Just Electrical Engineering \Shannon's contribution to pure mathematics was denied immediate recognition. I can recall now that even at the International Mathematical Congress, Amsterdam, 1954, my American colleagues in probability seemed rather doubtful about my allegedly exaggerated interest in Shannon's work, as they believed it consisted more of techniques than of mathematics itself. However, Shannon did not provide rigorous mathematical justification of the complicated cases and left it all to his followers. Still his mathematical intuition is amazingly correct." A. N. Kolmogorov, as quoted in [Shi89]. Kolmogorov and Entropy Kolmogorov's later work was fundamentally influenced by Shannon's. 1 Foundations: Kolmogorov Complexity - using the theory of algorithms to give a combinatorial interpretation of Shannon Entropy. 2 Analogy: Kolmogorov-Sinai Entropy, the only finitely-observable isomorphism-invariant property of dynamical systems.
    [Show full text]
  • Information Theory
    Information Theory Professor John Daugman University of Cambridge Computer Science Tripos, Part II Michaelmas Term 2016/17 H(X,Y) I(X;Y) H(X|Y) H(Y|X) H(X) H(Y) 1 / 149 Outline of Lectures 1. Foundations: probability, uncertainty, information. 2. Entropies defined, and why they are measures of information. 3. Source coding theorem; prefix, variable-, and fixed-length codes. 4. Discrete channel properties, noise, and channel capacity. 5. Spectral properties of continuous-time signals and channels. 6. Continuous information; density; noisy channel coding theorem. 7. Signal coding and transmission schemes using Fourier theorems. 8. The quantised degrees-of-freedom in a continuous signal. 9. Gabor-Heisenberg-Weyl uncertainty relation. Optimal \Logons". 10. Data compression codes and protocols. 11. Kolmogorov complexity. Minimal description length. 12. Applications of information theory in other sciences. Reference book (*) Cover, T. & Thomas, J. Elements of Information Theory (second edition). Wiley-Interscience, 2006 2 / 149 Overview: what is information theory? Key idea: The movements and transformations of information, just like those of a fluid, are constrained by mathematical and physical laws. These laws have deep connections with: I probability theory, statistics, and combinatorics I thermodynamics (statistical physics) I spectral analysis, Fourier (and other) transforms I sampling theory, prediction, estimation theory I electrical engineering (bandwidth; signal-to-noise ratio) I complexity theory (minimal description length) I signal processing, representation, compressibility As such, information theory addresses and answers the two fundamental questions which limit all data encoding and communication systems: 1. What is the ultimate data compression? (answer: the entropy of the data, H, is its compression limit.) 2.
    [Show full text]
  • Paper 1 Jian Li Kolmold
    KolmoLD: Data Modelling for the Modern Internet* Dmitry Borzov, Huawei Canada Tim Tingqiu Yuan, Huawei Mikhail Ignatovich, Huawei Canada Jian Li, Futurewei *work performed before May 2019 Challenges: Peak Traffic Composition 73% 26% Content: sizable, faned-out, static Streaming Services Everything else (Netflix, Hulu, Software File storage (Instant Messaging, YouTube, Spotify) distribution services VoIP, Social Media) [1] Source: Sandvine Global Internet Phenomena reports for 2009, 2010, 2011, 2012, 2013, 2015, 2016, October 2018 [1] https://qz.com/1001569/the-cdn-heavy-internet-in-rich-countries-will-be-unrecognizable-from-the-rest-of-the-worlds-in-five-years/ Technologies to define the revolution of the internet ChunkStream Founded in 2016 Video codec Founded in 2014 Content-addressable Based on a 2014 MIT Browser-targeted Runtime Content-addressable network protocol based research paper network protocol based on cryptohash naming on cryptohash naming scheme Implemented and supported by all major scheme Based on the Open source project cryptohash naming browsers, an IETF standard Founding company is a P2P project scheme YCombinator graduate YCombinator graduate with backing of high profile SV investors Our Proposal: A data model for interoperable protocols KolmoLD Content addressing through hashes has become a widely-used means of Addressable: connecting layer, inspired by connecting data in distributed the principles of Kolmogorov complexity systems, from the blockchains that theory run your favorite cryptocurrencies, to the commits that back your code, to Compossible: sending data as code, where the web’s content at large. code efficiency is theoretically bounded by Kolmogorov complexity Yet, whilst all of these tools rely on Computable: sandboxed computability by some common primitives, their treating data as code specific underlying data structures are not interoperable.
    [Show full text]
  • Maximizing T-Complexity 1. Introduction
    Fundamenta Informaticae XX (2015) 1–19 1 DOI 10.3233/FI-2012-0000 IOS Press Maximizing T-complexity Gregory Clark U.S. Department of Defense gsclark@ tycho. ncsc. mil Jason Teutsch Penn State University teutsch@ cse. psu. edu Abstract. We investigate Mark Titchener’s T-complexity, an algorithm which measures the infor- mation content of finite strings. After introducing the T-complexity algorithm, we turn our attention to a particular class of “simple” finite strings. By exploiting special properties of simple strings, we obtain a fast algorithm to compute the maximum T-complexity among strings of a given length, and our estimates of these maxima show that T-complexity differs asymptotically from Kolmogorov complexity. Finally, we examine how closely de Bruijn sequences resemble strings with high T- complexity. 1. Introduction How efficiently can we generate random-looking strings? Random strings play an important role in data compression since such strings, characterized by their lack of patterns, do not compress well. For prac- tical compression applications, compression must take place in a reasonable amount of time (typically linear or nearly linear time). The degree of success achieved with the classical compression algorithms of Lempel and Ziv [9, 21, 22] and Welch [19] depend on the complexity of the input string; the length of the compressed string gives a measure of the original string’s complexity. In the same monograph in which Lempel and Ziv introduced the first of these algorithms [9], they gave an example of a succinctly describable class of strings which yield optimally poor compression up to a logarithmic factor, namely the class of lexicographically least de Bruijn sequences for each length.
    [Show full text]
  • 1952 Washington UFO Sightings • Psychic Pets and Pet Psychics • the Skeptical Environmentalist Skeptical Inquirer the MAGAZINE for SCIENCE and REASON Volume 26,.No
    1952 Washington UFO Sightings • Psychic Pets and Pet Psychics • The Skeptical Environmentalist Skeptical Inquirer THE MAGAZINE FOR SCIENCE AND REASON Volume 26,.No. 6 • November/December 2002 ppfjlffl-f]^;, rj-r ci-s'.n.: -/: •:.'.% hstisnorm-i nor mm . o THE COMMITTEE FOR THE SCIENTIFIC INVESTIGATION OF CLAIMS OF THE PARANORMAL AT THE CENTER FOR INQUIRY-INTERNATIONAL (ADJACENT TO THE STATE UNIVERSITY OF NEW YORK AT BUFFALO) • AN INTERNATIONAL ORGANIZATION Paul Kurtz, Chairman; professor emeritus of philosophy. State University of New York at Buffalo Barry Karr, Executive Director Joe Nickell, Senior Research Fellow Massimo Polidoro, Research Fellow Richard Wiseman, Research Fellow Lee Nisbet Special Projects Director FELLOWS James E. Alcock,* psychologist. York Univ., Consultants, New York. NY Irmgard Oepen, professor of medicine Toronto Susan Haack. Cooper Senior Scholar in Arts (retired), Marburg, Germany Jerry Andrus, magician and inventor, Albany, and Sciences, prof, of philosophy, University Loren Pankratz, psychologist, Oregon Health Oregon of Miami Sciences Univ. Marcia Angell, M.D., former editor-in-chief. C. E. M. Hansel, psychologist. Univ. of Wales John Paulos, mathematician, Temple Univ. New England Journal of Medicine Al Hibbs, scientist, Jet Propulsion Laboratory Steven Pinker, cognitive scientist, MIT Robert A. Baker, psychologist. Univ. of Douglas Hofstadter, professor of human Massimo Polidoro, science writer, author, Kentucky understanding and cognitive science, executive director CICAP, Italy Stephen Barrett, M.D., psychiatrist, author. Indiana Univ. Milton Rosenberg, psychologist, Univ. of consumer advocate, Allentown, Pa. Gerald Holton, Mallinckrodt Professor of Chicago Barry Beyerstein,* biopsychologist, Simon Physics and professor of history of science, Wallace Sampson, M.D., clinical professor of Harvard Univ. Fraser Univ., Vancouver, B.C., Canada medicine, Stanford Univ., editor, Scientific Ray Hyman.* psychologist, Univ.
    [Show full text]
  • Marion REVOLLE
    Algorithmic information theory for automatic classification Marion Revolle, [email protected] Nicolas le Bihan, [email protected] Fran¸cois Cayre, [email protected] Main objective Files : any byte string in a computer (text, music, ..) & Similarity measure in a non-probabilist context : Similarity metric Algorithmic information theory : Kolmogorov complexity % 1.1 Complexity 1.2 GZIP 1.3 Examples GZIP : compression algorithm = DEFLATE + Huffman. Given x a file string of size jxj define on the alphabet Ax A- x = ABCA BCCB CABC of size αx. A- DEFLATE L(x) = 6 Z(-1! 1) A- Simple complexity Dictionary compression based on LZ77 : make ref- x : L(A) L(B) L(C) Z(-3!3) L(C) Z(-6!5) K(x) : Kolmogorov complexity : the length of a erences from the past. ABC ABCC BCABC shortest binary program to compute x. B- y = ABCA BCAB CABC DEFLATE(x) generate two kinds of symbol : L(x) : Lempel-Ziv complexity : the minimal number L(y) = 5 of operations making insert/copy from x's past L(a) : insert the element a in Ax = Literal. y : L(A) L(B) L(C) Z(-3!3) Z(-6!6) to generate x. Z(-i ! j) : paste j elements, i elements before = Refer- ABC ABC ABCABC ence of length j. L(yjx) = 2 yjx : Z(-12!6) Z(-12!6) B- Conditional complexity ABCABC ABCABC K(xjy) : conditional Kolmogorov complexity : the B- Complexity C- z = MNOM NOMN OMNO length of a shortest binary program to compute Number of symbols to compress x with DEFLATE L(z) = 5 x is y is furnished as an auxiliary input.
    [Show full text]
  • On One-Way Functions and Kolmogorov Complexity
    On One-way Functions and Kolmogorov Complexity Yanyi Liu Rafael Pass∗ Cornell University Cornell Tech [email protected] [email protected] September 24, 2020 Abstract We prove that the equivalence of two fundamental problems in the theory of computing. For every polynomial t(n) ≥ (1 + ")n; " > 0, the following are equivalent: • One-way functions exists (which in turn is equivalent to the existence of secure private-key encryption schemes, digital signatures, pseudorandom generators, pseudorandom functions, commitment schemes, and more); • t-time bounded Kolmogorov Complexity, Kt, is mildly hard-on-average (i.e., there exists a t 1 polynomial p(n) > 0 such that no PPT algorithm can compute K , for more than a 1 − p(n) fraction of n-bit strings). In doing so, we present the first natural, and well-studied, computational problem characterizing the feasibility of the central private-key primitives and protocols in Cryptography. ∗Supported in part by NSF Award SATC-1704788, NSF Award RI-1703846, AFOSR Award FA9550-18-1-0267, and a JP Morgan Faculty Award. This research is based upon work supported in part by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA), via 2019-19-020700006. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, either expressed or implied, of ODNI, IARPA, or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation therein. 1 Introduction We prove the equivalence of two fundamental problems in the theory of computing: (a) the exis- tence of one-way functions, and (b) mild average-case hardness of the time-bounded Kolmogorov Complexity problem.
    [Show full text]
  • Complexity Measurement Based on Information Theory and Kolmogorov Complexity
    Leong Ting Lui** Complexity Measurement Based † Germán Terrazas on Information Theory and University of Nottingham Kolmogorov Complexity ‡ Hector Zenil University of Sheffield Cameron Alexander§ University of Nottingham Natalio Krasnogor*,¶ Abstract In the past decades many definitions of complexity Newcastle University have been proposed. Most of these definitions are based either on Shannonʼs information theory or on Kolmogorov complexity; these two are often compared, but very few studies integrate the two ideas. In this article we introduce a new measure of complexity that builds on both of these theories. As a demonstration of the concept, the technique is applied to elementary cellular automata Keywords and simulations of the self-organization of porphyrin molecules. Complex systems, measures of complexity, open-ended evolution, cellular automata, molecular computation 1 Introduction The notion of complexity has been used in various disciplines. However, the very definition of complexity1 is still without consensus. One of the reasons for this is that there are many intuitive interpretations to the concept of complexity that can be applied in different domains of science. For instance, the Lorentz attractor can be simply described by three interaction differential equations, but it is often seen as producing complex chaotic dynamics. The iconic fractal image generated by the Mandelbrot set can be seen as a simple mathematical algorithm having a com- plex phase diagram. Yet another characteristic of complex systems is that they consist of many interacting components such as genetic regulatory networks, electronic components, and ecological food chains. Complexity is also associated with processes that generate complex outcomes, evolution being the best example.
    [Show full text]
  • Kolmogorov Complexity and Its Applications
    CHAPTER 4 Kolmogorov Complexity and its Applications Ming LI Aiken Computation Laboratory, Han•ard University, Cambridge, MA 02138, USA Paul M.B. VIT ANYI Centrum voor Wiskunde en biformatica, Kruislaan 413, 1098 SJ Amsterdam, Netherlands, and Facu/teit Wiskunde en Jnjormatica, Universiteic van Amsterdam, Amsterdam, Netherlands Contents L Introduction 189 2. Mathematical theory 196 3. Applications of compressibility 214 4. Example of an application in mathematics: weak prime number theorems. 221 5. Applications of incompressibility: proving lower bounds 222 6. Resource-bounded Kolmogorov complexity and its applications 236 7. Conclusion 247 Acknowledgment 247 References :48 HANDBOOK OF THEORETICAL COMPUTER SCIENCE Edited by J. van Leeuwen ((' Elsevier Science Publishers B.V., 1990 KOLMOGOROV COMPLEXITY AND ITS APPLICATIONS 189 t. Introduction In everyday language we identify the information in an individual object with the essentials of a description for it. We can formalize this by defining the amount of information in a finite object (like a string) as the size (i.e., number of bits) of the smallest program that, starting with a blank memory, outputs the string and then terminates. A similar definition can be given for infinite strings, but in this case the program produces element after element forever. Thus, 1n (a string of nones) contains little information because a program of size about log n outputs it (like "print n ls"). Likewise, the transcendental number rr = 3.1415 ... , an infinite sequence of seemingly "random" decimal digits, contains O(l) information. (There is a short program that produces the consecutive digits of rr forever.) Such a definition would appear to make the amount of information in a string depend on the particular programming language used.
    [Show full text]
  • Minimum Complexity Pursuit: Stability Analysis
    Minimum Complexity Pursuit: Stability Analysis Shirin Jalali Arian Maleki Richard Baraniuk Center for Mathematics of Information Department of Electrical Engineering Department of Electrical Engineering California Institute of Technology Rice University Rice University Pasadena, California, 91125 Houston, Texas, 77005 Houston, Texas, 77005 Abstract— A host of problems involve the recovery of struc- their linear measurements without having the knowledge of tured signals from a dimensionality reduced representation the underlying structure? Does this ignorance incur a cost in such as a random projection; examples include sparse signals the sampling rate? (compressive sensing) and low-rank matrices (matrix comple- tion). Given the wide range of different recovery algorithms In algorithmic information theory, Kolmogorov complex- developed to date, it is natural to ask whether there exist ity, introduced by Solomonoff [12], Kolmogorov [13], and “universal” algorithms for recovering “structured” signals from Chaitin [14], defines a universal notion of complexity for their linear projections. We recently answered this question in finite-alphabet sequences. Given a finite-alphabet sequence the affirmative in the noise-free setting. In this paper, we extend x, the Kolmogorov complexity of x, K(x), is defined as the our results to the case of noisy measurements. length of the shortest computer program that prints x and I. INTRODUCTION halts. In [15], extending the notion of Kolmogorov complex- ity to real-valued signals1 by their proper quantization, we Data compressors are ubiquitous in the digital world. They addressed some of the above questions. We introduced the are built based on the premise that text, images, videos, minimum complexity pursuit (MCP) algorithm for recover- etc.
    [Show full text]
  • Prefix-Free Kolmogorov Complexity
    Prefix-free complexity Properties of Kolmogorov complexity Information Theory and Statistics Lecture 7: Prefix-free Kolmogorov complexity Łukasz Dębowski [email protected] Ph. D. Programme 2013/2014 Project co-financed by the European Union within the framework of the European Social Fund Prefix-free complexity Properties of Kolmogorov complexity Towards algorithmic information theory In Shannon’s information theory, the amount of information carried by a random variable depends on the ascribed probability distribution. Andrey Kolmogorov (1903–1987), the founding father of modern probability theory, remarked that it is extremely hard to imagine a reasonable probability distribution for utterances in natural language but we can still estimate their information content. For that reason Kolmogorov proposed to define the information content of a particular string in a purely algorithmic way. A similar approach has been proposed a year earlier by Ray Solomonoff (1926–2009), who sought for optimal inductive inference. Project co-financed by the European Union within the framework of the European Social Fund Prefix-free complexity Properties of Kolmogorov complexity Kolmogorov complexity—analogue of Shannon entropy The fundamental object of algorithmic information theory is Kolmogorov complexity of a string, an analogue of Shannon entropy. It is defined as the length of the shortest program for a Turing machine such that the machine prints out the string and halts. The Turing machine itself is defined as a deterministic finite state automaton which moves along one or more infinite tapes filled with symbols from a fixed finite alphabet and which may read and write individual symbols. The concrete value of Kolmogorov complexity depends on the used Turing machine but many properties of Kolmogorov complexity are universal.
    [Show full text]