Maxwell's Demon: Entropy, Information, Computing

Total Page:16

File Type:pdf, Size:1020Kb

Maxwell's Demon: Entropy, Information, Computing Complex Social Systems a guided exploration of concepts and methods Information Theory Martin Hilbert (Dr., PhD) Today’s questions I. What are the formal notions of data, information & knowledge? II. What is the role of information theory in complex systems? …the 2nd law and life… III. What is the relation between information and growth? Complex Systems handle information "Although they [complex systems] differ widely in their physical attributes, they resemble one another in the way they handle information. That common feature is perhaps the best starting point for exploring how they operate.” (Gell-Mann, 1995, p. 21) Source: Gell-Mann, M. (1995). The Quark and the Jaguar: Adventures in the Simple and the Complex. New York: St. Martin’s Griffin. Complex Systems Science as computer science “What we are witnessing now is … actually the beginning of an amazing convergence of theoretical physics and theoretical computer science” Chaitin, G. J. (2002). Meta-Mathematics and the Foundations of Mathematics. EATCS Bulletin, 77(June), 167–179. Society computes Numerical: Analytical: Agent-based models Information Theory Why now? Thermodynamics: or ? Aerodynamics: or ? Hydrodynamics: or ? Electricity: or ? etc… Claude Shannon Alan Turing John von Neumann Bell, D. (1973). The Coming of Post- Industrial Society: A Venture in Social Forecasting. New York, NY: Basic Books. “industrial organization of goods” => “industrial organization of information” 15,000 citations 35,000 citations Beniger, J. (1986). The Control Revolution: Castells, M. (1999). The Information Technological and Economic Origins of the Age, Volumes 1-3: Economy, Society Information Society. Harvard University Press and Culture. Cambridge (Mass.); “exploitation of information” Oxford: Wiley-Blackwell. => “control social & economic processes” “society lives in networks” 3,000 citations => “information as well” Shannon, C. (1948). A Mathematical Theory of Communication. Bell System Technical Journal, 27, 379–423, 623–656. Shannon, C. E., & Weaver, W. (1949). The Mathematical Theory of Communication. University of Illinois Press. U.S. Congress, Office of Technology Assessment. (1995). Global communications : opportunities for trade and aid. (OTA-ITC-642nd ed.). U.S. Government Printing Office. From the Source Coding Theorem to the Channel Coding Theorem Information Communication Claude Shannon (1948) Information Theory Primer: information counts differences Syntactic and Semantic information o In information theory and computer science, the definition of information consists in its quantification (just like the definitions of “heat” (C○;F), “speed” (mph; km/h), or “water” (g,l) are independent from being good, bad or useless for you) o To assign a value to the meaning of information, we first require its quantity: Ralph Hartley [value / unit of information], [US$ / unit of information], or [fitness increase / unit of information], etc. (1888 – 1970) Ralph Hartley & “the difference” o Information consists in the differentiation among alternative possibilities: no information without differences! information does only exist if there are different choices can be understood as the opposite of uncertainty (the more uncertainty we have regarding different possibilities, the less information we have; the more information we have, the less uncertainty) What’s the most basic difference? o Binary: [black,white]; [up,down]; [left,right]; [head,tail]; [42,non-42]; [no-current,current]; [0,1] (!) o How much information is revealed with a binary decision (coin flip)? o To make sense, each additional unit should provide as much information as the previous: additive measure = each additional unit gives the number of choices at this additional step Source: Hartley, R. V. L. (1928). Transmission of Information. Bell System Technical Journal, International Congress of Telegraphy and Telephony, Lake Como, Italy, 1927, 535–563 Information Theory Primer: finding the right measure How many binary symbols (coin flips) are needed to describe 8 differences? ? 1st coin flip Ralph Hartley (1888 – 1970) 2nd coin flip [h,h] [h,t] [t,h] [t,t] 3rd coin flip [111] [110] [101] [100] [011] [010] [001] [000] 4th coin flip…? Answer: …through the “2-ary uncertainty” revealed at each step: 1 flip o 2 choices = 1 symbol (coin flip) o 2 = 2 choices o 4 choices = 2 symbols o 22 flips = 4 choices 2 ∗ 2 ∗ 2 = 23 = 8 o 8 choices = 3 symbols o 23 flips = 8 choices 푙표푔2 2 ∗ 2 ∗ 2 = 푙표푔28 = 3 o 16 choices = 4 symbols o 24flips = 16 choices Source: Hartley, R. V. L. (1928). Transmission of Information. Bell System Technical Journal, International Congress of Telegraphy and Telephony, Lake Como, Italy, 1927, 535–563 Information Theory Primer: base of log defines metric system How many senary (6-ary) symbols are needed to describe 36 choices? ? 1st roll of dice [1,3] [2,4] [4,4] 2nd roll of dice Base of logarithm defines metric system, just like: - feet vs. meters - liters vs. gallons 6-ary code: o 6 choices = 1 symbol (dice rolls) o 61 roll = 6 choices o 36 choices = 2 symbols o 62 rolls = 36 choices 6 ∗ 6 ∗ 6 = 63 = 216 o 216 choices = 3 symbols o 63 rolls = 216 choices 푙표푔6 6 ∗ 6 ∗ 6 = 푙표푔6216 = 3 Source: Hartley, R. V. L. (1928). Transmission of Information. Bell System Technical Journal, International Congress of Telegraphy and Telephony, Lake Como, Italy, 1927, 535–563 Information Theory Primer: Shannon’s (1948) idea of the bit Shannon’s game of twenty questions: 220 = 1,048,576 => …down to 5 miles2 ! Claude Shannon (1916 – 2001) Claude E. Shannon (1948) A Mathematical Theory of Communication, Bell System Technical Journal, Vol. 27, pp. 379–423, 623–656. A Information Theory Primer B C coding Yes D E No F G H Yes I J K No Yes L No M N Claude Shannon Yes O (1948) P A Mathematical 2 3 4 5 J Theory of 1 R bits S Communication, Yes Bell System Technical Journal, T Vol. 27, pp. 379–423, 623–656. No No U V Yes W genius: X No Y Transmit: Z Ñ “g” Yes Á No É Í Ó => “g” = yes-yes-no-no-yes = 11001 Ú A Information Theory Primer B C + coding Yes D P E R No F O G H B Yes I A J B K I No Yes L L No M Yes N I Claude Shannon O T (1948) P Y A Mathematical J 1 2 3 4 5 bits– Theory of R S D Communication, No Yes Bell System Technical Journal, T I Vol. 27, pp. 379–423, 623–656. No U S V T Yes W genius: X R No Y I Transmit: Z B Ñ U Yes Á “i” T No É Í I Ó O => “i” = yes-no-yes-yes-yes = 10111 Ú N => consider REDUNDANCY A Information Theory Primer B C + Yes D P E R No F O G H B Yes I A J B K I No Yes L L No M Yes N I O T P Y 1 2 3 4 5 Jbits– R S D Yes No T I Source Coding No U S V T Theorem: Yes W genius: X R What’s the purest No Y I form Transmit: Z B Ñ U of information? Yes Á “i” T ! ENTROPY ! No É Í I Ó O => “i” = yes-no-yes-yes-yes = 10111 Ú N => consider REDUNDANCY COMPRESSION Transmit less data, so you can transmit more information “I do not ne_d to comm_______ (…with the same amount of ev_ry det_il, _u kn_w what I data symbols…) mean, and _u rec_ive all the info____ anywa_s, right? Source Coding Probability of letters Probability of words Theorem: in English alphabet: in English language: What’s the purest form E: 13% “the”: 10% T: 10% “ of”: 5.1% of information? A: 8% “in”: 2.7% ! ENTROPY ! O: 7% … … “with”: 0.66% Q: 0.121% “from”: 0.64% Z: 0.077% ... COMPRESSION Transmit less, so you can transmit more! SourceL Purandare, A. (2008). JPEG vs JPEG2000 Comparison P.S. there is also “lossy compression”, but this is simply elimination of some information (reducing quality / detail / information). Real “data compression” into information is “loss-less”. Source Coding Theorem: What’s the purest form of information? ! ENTROPY ! 1993! telecom number of performance of sum up their product telecom devices devices Infrastructure Hardware 7% per year 28% 8% per year per year X Software = compression 10.4% per year Number of X X = (drops) bits storage number of performance of sum up their product storage devices devices Infrastructure Hardware 8% per year 25% 5% per year per year X Software = compression 11% per year Number of X X = (stones) bits The 2nd law of thermodynamics: how structure and information relate “The law that entropy always increases holds, I think, the supreme position among the laws of Nature” (Eddington, 1927) => only 1 way to => 20 ways to => 6! = 720 ways create this structure create this structure to create this structure . ? ? ? 1 2 3 4 5 6 1 2 3 4 5 6 1 2 3 4 5 6 t1 no uncertainty t2 max. uncertainty t3 = min. entropy = max. entropy Things decay as time passes = Things become more uniformly distributed as everything interacts with everything = as time goes by it is most likely (almost certain) that the most likely thing will happen… Sources: Eddington, A. S. (1927). The Nature of the Physical World. Kessinger Publishing. As time goes by entropy increases… …BUT what about Isaac Asimov’s “Last Question” ever asked: "Can entropy ever be reversed?“ Life reverses entropy all the time! Erwin Schrödinger Sources: Asimov, Isaac. The Last Question. Science Fiction Quarterly. November 1956 (1887-1961) 3/4 of 1/4 of the How does evolution the time time “communicate”? 50 % 50 % 33 % 67 % 50 % 50 % 33 % 67 % 20 % 80 % 3/4 of 1/4 of the How does evolution the time time “communicate”? 20 % 80 % 11 % 89 % 20 % 80 % 11 % 89 % 6 % 94 % How does evolution “communicate”? t1 => t2 => … => tn 25 % of the time … 75 % of the time The eco-”system” communicates an environmental structure to the system, which stores this information in its own structure A How much information B C does an evolutionary Yes D E dynamic communicate? No F G H Yes I J No K Yes L Yes No M N O P .
Recommended publications
  • Information Theory and the Length Distribution of All Discrete Systems
    Information Theory and the Length Distribution of all Discrete Systems Les Hatton1* Gregory Warr2** 1 Faculty of Science, Engineering and Computing, Kingston University, London, UK 2 Medical University of South Carolina, Charleston, South Carolina, USA. * [email protected] ** [email protected] Revision: $Revision: 1.35 $ Date: $Date: 2017/09/06 07:30:43 $ Abstract We begin with the extraordinary observation that the length distribution of 80 million proteins in UniProt, the Universal Protein Resource, measured in amino acids, is qualitatively identical to the length distribution of large collections of computer functions measured in programming language tokens, at all scales. That two such disparate discrete systems share important structural properties suggests that yet other apparently unrelated discrete systems might share the same properties, and certainly invites an explanation. We demonstrate that this is inevitable for all discrete systems of components built from tokens or symbols. Departing from existing work by embedding the Conservation of Hartley-Shannon information (CoHSI) in a classical statistical mechanics framework, we identify two kinds of discrete system, heterogeneous and homogeneous. Heterogeneous systems contain components built from a unique alphabet of tokens and yield an implicit CoHSI distribution with a sharp unimodal peak asymptoting to a power-law. Homogeneous systems contain components each built from arXiv:1709.01712v1 [q-bio.OT] 6 Sep 2017 just one kind of token unique to that component and yield a CoHSI distribution corresponding to Zipf’s law. This theory is applied to heterogeneous systems, (proteome, computer software, music); homogeneous systems (language texts, abundance of the elements); and to systems in which both heterogeneous and homogeneous behaviour co-exist (word frequencies and word length frequencies in language texts).
    [Show full text]
  • Conformational Transition in Immunoglobulin MOPC 460" by Correction. in Themembership List of the National Academy of Scien
    Corrections Proc. Natl. Acad. Sci. USA 74 (1977) 1301 Correction. In the article "Kinetic evidence for hapten-induced Correction. In the membership list of the National Academy conformational transition in immunoglobulin MOPC 460" by of Sciences that appeared in the October 1976 issue of Proc. D. Lancet and I. Pecht, which appeared in the October 1976 Natl. Acad. Sci. USA 73,3750-3781, please note the following issue of Proc. Nati. Acad. Sci. USA 73,3549-3553, the authors corrections: H. E. Carter, Britton Chance, Seymour S. Cohen, have requested the following changes. On p. 3550, right-hand E. A. Doisy, Gerald M. Edelman, and John T. Edsall are affil- column, second line from bottom, and p. 3551, left-hand col- iated with the Section ofBiochemistry (21), not the Section of umn, fourth line from the top, "Fig. 2" should be "Fig. 1A." Botany (25). In the legend of Table 2, third line, note (f) should read "AG, = -RTlnKj." On p. 3553, left-hand column, third paragraph, fifth line, "ko" should be replaced by "Ko." Correction. In the Author Index to Volume 73, January-De- cember 1976, which appeared in the December 1976 issue of Proc. Natl. Acad. Sci. USA 73, 4781-4788, the limitations of Correction. In the article "Amino-terminal sequences of two computer alphabetization resulted in the listing of one person polypeptides from human serum with nonsuppressible insu- as the author of another's paper. On p. 4786, it should indicate lin-like and cell-growth-promoting activities: Evidence for that James Christopher Phillips had an article beginning on p.
    [Show full text]
  • Copyright by Xue Chen 2018 the Dissertation Committee for Xue Chen Certifies That This Is the Approved Version of the Following Dissertation
    Copyright by Xue Chen 2018 The Dissertation Committee for Xue Chen certifies that this is the approved version of the following dissertation: Using and Saving Randomness Committee: David Zuckerman, Supervisor Dana Moshkovitz Eric Price Yuan Zhou Using and Saving Randomness by Xue Chen DISSERTATION Presented to the Faculty of the Graduate School of The University of Texas at Austin in Partial Fulfillment of the Requirements for the Degree of DOCTOR OF PHILOSOPHY THE UNIVERSITY OF TEXAS AT AUSTIN May 2018 Dedicated to my parents Daiguang and Dianhua. Acknowledgments First, I am grateful to my advisor, David Zuckerman, for his unwavering support and encouragements. David has been a constant resource for providing me with great advice and insightful feedback about my ideas. Besides these and many fruitful discussions we had, his wisdom, way of thinking, and optimism guided me through the last six years. I also thank him for arranging a visit to the Simons institute in Berkeley, where I enrich my understanding and made new friends. I could not have hoped for a better advisor. I would like to thank Eric Price. His viewpoints and intuitions opened a different gate for me, which is important for the line of research presented in this thesis. I benefited greatly from our long research meetings and discussions. Besides learning a lot of technical stuff from him, I hope I have absorbed some of his passion and attitudes to research. Apart from David and Eric, I have had many mentors during my time as a student. I especially thank Yuan Zhou, who has been an amazing collaborator and friend since starting college.
    [Show full text]
  • Shannon's Formula Wlog(1+SNR): a Historical Perspective
    Shannon’s Formula W log(1 + SNR): A Historical· Perspective on the occasion of Shannon’s Centenary Oct. 26th, 2016 Olivier Rioul <[email protected]> Outline Who is Claude Shannon? Shannon’s Seminal Paper Shannon’s Main Contributions Shannon’s Capacity Formula A Hartley’s rule C0 = log2 1 + ∆ is not Hartley’s 1 P Many authors independently derived C = 2 log2 1 + N in 1948. Hartley’s rule is exact: C0 = C (a coincidence?) C0 is the capacity of the “uniform” channel Shannon’s Conclusion 2 / 85 26/10/2016 Olivier Rioul Shannon’s Formula W log(1 + SNR): A Historical Perspective April 30, 2016 centennial day celebrated by Google: here Shannon is juggling with bits (1,0,0) in his communication scheme “father of the information age’’ Claude Shannon (1916–2001) 100th birthday 2016 April 30, 1916 Claude Elwood Shannon was born in Petoskey, Michigan, USA 3 / 85 26/10/2016 Olivier Rioul Shannon’s Formula W log(1 + SNR): A Historical Perspective here Shannon is juggling with bits (1,0,0) in his communication scheme “father of the information age’’ Claude Shannon (1916–2001) 100th birthday 2016 April 30, 1916 Claude Elwood Shannon was born in Petoskey, Michigan, USA April 30, 2016 centennial day celebrated by Google: 3 / 85 26/10/2016 Olivier Rioul Shannon’s Formula W log(1 + SNR): A Historical Perspective Claude Shannon (1916–2001) 100th birthday 2016 April 30, 1916 Claude Elwood Shannon was born in Petoskey, Michigan, USA April 30, 2016 centennial day celebrated by Google: here Shannon is juggling with bits (1,0,0) in his communication scheme “father of the information age’’ 3 / 85 26/10/2016 Olivier Rioul Shannon’s Formula W log(1 + SNR): A Historical Perspective “the most important man..
    [Show full text]
  • Information Basics Observer and Choice
    information basics observer and choice Information is defined as “a measure of the freedom from choice with which a message is selected from the set of all possible messages” Bit (short for binary digit) is the most elementary choice one can make Between two items: “0’ and “1”, “heads” or “tails”, “true” or “false”, etc. Bit is equivalent to the choice between two equally likely alternatives Example, if we know that a coin is to be tossed, but are unable to see it as it falls, a message telling whether the coin came up heads or tails gives us one bit of information 1 Bit of uncertainty 1 Bit of information H,T? uncertainty removed, choice between 2 symbols information gained recognized by an observer INDIANA [email protected] UNIVERSITY informatics.indiana.edu/rocha Fathers of uncertainty-based information Information is transmitted through noisy communication channels Ralph Hartley and Claude Shannon (at Bell Labs), the fathers of Information Theory, worked on the problem of efficiently transmitting information; i. e. decreasing the uncertainty in the transmission of information. Hartley, R.V.L., "Transmission of Information", Bell System Technical Journal, July 1928, p.535. C. E. Shannon [1948], “A mathematical theory of communication”. Bell System Technical Journal, 27:379-423 and 623-656 C. E. Shannon, “A Symbolic analysis of relay and switching circuits” .MS Thesis, (unpublished) MIT, 1937. C. E. Shannon, “An algebra for theoretical genetics.” Phd Dissertation, MIT, 1940. INDIANA [email protected] UNIVERSITY informatics.indiana.edu/rocha Let’s talk about choices Multiplication Principle “If some choice can be made in M different ways, and some subsequent choice can be made in N different ways, then there are M x N different ways these choices can be made in succession” [Paulos] 3 shirts and 4 pants = 3 x 4 = 12 outfit choices INDIANA [email protected] UNIVERSITY informatics.indiana.edu/rocha Hartley Uncertainty Nonspecificity A = Set of Hartley measure Alternatives The amount of uncertainty associated with a set of alternatives (e.g.
    [Show full text]
  • Active Regression Via Linear-Sample Sparsification
    Active Regression via Linear-Sample Sparsification Xue Chen∗ Eric Pricey [email protected] [email protected] Northwestern University The University of Texas at Austin March 25, 2019 Abstract We present an approach that improves the sample complexity for a variety of curve fitting problems, including active learning for linear regression, polynomial regression, and continuous sparse Fourier transforms. In the active linear regression problem, one would like to estimate the ∗ n×d least squares solution β minimizing kXβ − yk2 given the entire unlabeled dataset X 2 R but only observing a small number of labels yi. We show that O(d) labels suffice to find a constant factor approximation βe: 2 ∗ 2 E[kXβe − yk2] ≤ 2 E[kXβ − yk2]: This improves on the best previous result of O(d log d) from leverage score sampling. We also present results for the inductive setting, showing when βe will generalize to fresh samples; these apply to continuous settings such as polynomial regression. Finally, we show how the techniques yield improved results for the non-linear sparse Fourier transform setting. arXiv:1711.10051v3 [cs.LG] 21 Mar 2019 ∗Supported by NSF Grant CCF-1526952 and a Simons Investigator Award (#409864, David Zuckerman), part of this work was done while the author was a student in University of Texas at Austin and visiting the Simons Institute for the Theory of Computing. yThis work was done in part while the author was visiting the Simons Institute for the Theory of Computing. 1 Introduction We consider the query complexity of recovering a signal f(x) in a given family F from noisy observations.
    [Show full text]
  • Sounds of Science – Schall Im Labor (1800–1930)
    MAX-PLANCK-INSTITUT FÜR WISSENSCHAFTSGESCHICHTE Max Planck Institute for the History of Science 2008 PREPRINT 346 Julia Kursell (ed.) Sounds of Science – Schall im Labor (1800–1930) TABLE OF CONTENTS Introductory Remarks Julia Kursell 3 „Erzklang“ oder „missing fundamental“: Kulturgeschichte als Signalanalyse Bernhard Siegert 7 Klangfiguren (a hit in the lab) Peter Szendy 21 Sound Objects Julia Kursell 29 A Cosmos for Pianoforte d’Amore: Some Remarks on “Miniature Estrose” by Marco Stroppa Florian Hoelscher 39 The “Muscle Telephone”: The Undiscovered Start of Audification in the 1870s Florian Dombois 41 Silence in the Laboratory: The History of Soundproof Rooms Henning Schmidgen 47 Cats and People in the Psychoacoustics Lab Jonathan Sterne 63 The Resonance of “Bodiless Entities”: Some Epistemological Remarks on the Radio-Voice Wolfgang Hagen 73 VLF and Musical Aesthetics Douglas Kahn 81 Ästhetik des Signals Daniel Gethmann 89 Producing, Representing, Constructing: Towards a Media-Aesthetic Theory of Action Related to Categories of Experimental Methods Elena Ungeheuer 99 Standardizing Aesthetics: Physicists, Musicians, and Instrument Makers in Nineteenth-Century Germany Myles W. Jackson 113 Introductory Remarks The following collection of papers documents the workshop “Sounds of Science – Schall im Labor, 1800 to 1930,” carried out at the Max Planck Institute for the History of Science, Berlin October 5 to 7, 2006. It was organized within the context of the project “Experimentalization of Life: Configurations between Science, Art, and Technology” situated in Department III of the Institute. While the larger project has discussed topics such as “Experimental Cultures,” the “Materiality of Time Relations in Life Sciences, Art, and Technology (1830-1930),” “Science and the City,” “Relations between the Living and the Lifeless,” and the “Shape of Experiment” in workshops and conferences, this workshop asked about the role sound plays in the configurations among science, technology and the arts, focusing on the years between 1800 and 1930.
    [Show full text]
  • Backward Chaining –! Bayesian Belief Network –! Explanation
    Information, Search, and Expert Systems ! Robert Stengel! Robotics and Intelligent Systems MAE 345 ! Princeton University, 2017 •! Communication/Information Theory –! Wiener vs. Shannon –! Entropy •! Finding Decision Rules in Data –! ID3 Algorithm •! Graph and Tree Search •! Expert Systems –! Forward and Backward Chaining –! Bayesian Belief Network –! Explanation Copyright 2017 by Robert Stengel. All rights reserved. For educational use only. 1 http://www.princeton.edu/~stengel/MAE345.html “Communication Theory” or “Information Theory”? •! Prodigy at Harvard, professor at MIT •! Cybernetics •! Feedback control •! Communication theory Dark Hero Of The Information Age: In Search of Norbert Wiener, the Father of Cybernetics, Flo Norbert Wiener Conway and Jim Siegelman, 2005. Basic Books (1894-1964) •! University of Michigan, MIT (student), Bell Labs, MIT (professor) •! Boolean algebra •! Cryptography, telecommunications •! Information theory Claude The Information: A History, A Theory, A Flood, Shannon James Gleick, 2011, Pantheon. (1916-2001) 2 Information, Noise, and Observation S = Information (Signal) Power, e.g., watts N = Noise Power, e.g., watts ()S + N = Observed Power, e.g., watts 3 Communication: Separating Signals from Noise Signal-to-Noise Ratio, SNR Signal Power S SNR = ! Noise Power N 2 ! signal watts = (zero-mean), e.g., ! 2 noise watts SNR often expressed in decibels 2 Signal Power (Signal Amplitude) SNR(dB) = 10log10 = 10log10 2 Noise Power (Noise Amplitude) Signal Amplitude = 20log = S(dB) ! N(dB) 10 Noise Amplitude 4 Communication:
    [Show full text]
  • University of Central Florida School of Electrical Engineering and Computer Science EEL-6532: Information Theory and Coding. Spring 2010 - Dcm
    University of Central Florida School of Electrical Engineering and Computer Science EEL-6532: Information Theory and Coding. Spring 2010 - dcm Lecture 2 - Wednesday January 13, 2000 Shannon Entropy In his endeavor to construct mathematically tractable models of communication Shannon concentrated on stationary and ergodic1 sources of classical information. A stationary source of information emits symbols with a probability that does not change over time and an ergodic source emits information symbols with a probability equal to the frequency of their occurrence in a long sequence. Stationary ergodic sources of information have a ¯nite but arbitrary and potentially long correlation time. In late 1940s Shannon introduced a measure of the quantity of information a source could generate [13]. Earlier, in 1927, another scientist from Bell Labs, Ralph Hartley had proposed to take the logarithm of the total number of possible messages as a measure of the amount of information in a message generated by a source of information arguing that the logarithm tells us how many digits or characters are required to convey the message. Shannon recognized the relationship between thermodynamic entropy and informational entropy and, on von Neumann's advice, he called the negative logarithm of probability of an event, entropy2. Consider an event which happens with probability p; we wish to quantify the information content of a message communicating the occurrence of this event and we impose the condition that the measure should reflect the \surprise" brought by the occurrence of this event. An initial guess for a measure of this surprise would be 1=p, the lower the probability of the event the larger the surprise.
    [Show full text]
  • Claude Shannon and the Making of Information Theory
    The Essential Message: Claude Shannon and the Making of Information Theory by Erico Marui Guizzo B.S., Electrical Engineering University of Sao Paulo, Brazil, 1999 Submitted to the Program in Writing and Humanistic Studies in Partial Fulfillment of the Requirements for the Degree of Master of Science in Science Writing at the Massachusetts Institute of Technology MASSACHUSETTSINSTITUTE OFTECHNOLOGY K" September 2003 JUN 2 7 2003 LIBRARIES- © 2003 Erico Marui Guizzo. All rights reserved. The author hereby grants to MIT permission to reproduce and to distribute publicly paper and electronic copies of this thesis document in whole or in part. Signatureof Author:.................. ...................... {.. s udies Programln Writing aIf Humanistic Studies June 16, 2003 Certified and Accepted by:.......................................... ............................... Robert Kanigel Thesis Advisor Professor of Science Writing Director, Graduate Program in Science Writing ARCHIVES The Essential Message: Claude Shannon and the Making of Information Theory by Erico Marui Guizzo Submitted to the Program in Writing and Humanistic Studies on June 16, 2003 in Partial Fulfillment of the Requirements for the Degree of Master of Science in Science Writing ABSTRACT In 1948, Claude Shannon, a young engineer and mathematician working at the Bell Telephone Laboratories, published "A Mathematical Theory of Communication," a seminal paper that marked the birth of information theory. In that paper, Shannon defined what the once fuzzy concept of "information" meant for communication engineers and proposed a precise way to quantify it-in his theory, the fundamental unit of information is the bit. He also showed how data could be "compressed" before transmission and how virtually error-free communication could be achieved.
    [Show full text]
  • Hartley Oscillator - Wikipedia, the Free Encyclopedia Page 1 of 3
    Hartley oscillator - Wikipedia, the free encyclopedia Page 1 of 3 Hartley oscillator From Wikipedia, the free encyclopedia The Hartley oscillator is an electronic oscillator circuit in which the oscillation frequency is determined by a tuned circuit consisting of capacitors and inductors, that is, an LC oscillator. The circuit was invented in 1915 by American engineer Ralph Hartley. The distinguishing feature of the Hartley oscillator is that the tuned circuit consists of a single capacitor in parallel with two inductors in series (or a single tapped inductor), and the feedback signal needed for oscillation is taken from the center connection of the two inductors. Contents ■ 1 Operation ■ 2 History ■ 3 See also ■ 4 References ■ 5 External links Operation The Hartley oscillator is distinguished by a tank circuit consisting of two series-connected coils in parallel with a capacitor, with the feedback signal needed for oscillation taken from the center connection between the coils; the coils act as a voltage divider. The Hartley oscillator is the dual of the Colpitts oscillator which uses a voltage divider made of two capacitors rather than two inductors. Although there is no requirement for there to be mutual coupling between the two coil segments, the circuit is usually implemented using a tapped coil, with the feedback taken from the tap, as shown here. The optimal tapping point (or ratio of coil inductances) depends on the amplifying device used, which may be a bipolar junction Common-collector Hartley circuit transistor, FET, triode,
    [Show full text]
  • This Is IT: a Primer on Shannon's Entropy and Information
    L'Information, S´eminairePoincar´eXXIII (2018) 43 { 77 S´eminairePoincar´e This is IT: A Primer on Shannon's Entropy and Information Olivier Rioul LTCI, T´el´ecomParisTech I didn't like the term `information theory'. Universit´eParis-Saclay Claude [Shannon] didn't like it either. You see, 75013 Paris, France the term `information theory' suggests that it is a theory about information|but it's not. CMAP It's the transmission of information, not Ecole´ Polytechnique information. Lots of people just didn't Universit´eParis-Saclay understand this. 91128 Palaiseau, France Robert Fano, 2001 Abstract. What is Shannon's information theory (IT)? Despite its continued impact on our digital society, Claude Shannon's life and work is still unknown to numerous people. In this tutorial, we review many aspects of the concept of entropy and information from a historical and mathematical point of view. The text is structured into small, mostly independent sections, each covering a particular topic. For simplicity we restrict our attention to one-dimensional variables and use logarithm and exponential notations log and exp without specifying the base. We culminate with a simple exposition of a recent proof (2017) of the entropy power inequality (EPI), one of the most fascinating inequality in the theory. 1 Shannon's Life as a Child Claude Elwood Shannon was born in 1916 in Michigan, U.S.A., and grew up in the small town of Gaylord. He was a curious, inventive, and playful child, and probably remained that way throughout his life. He built remote-controlled models and set up his own barbed-wire telegraph system to a friend's house [48].
    [Show full text]