Unbridled Mental Power Artifcial Intelligence Is Set to Rival the Human Mind, Just As the Engine Did the Horse

Total Page:16

File Type:pdf, Size:1020Kb

Unbridled Mental Power Artifcial Intelligence Is Set to Rival the Human Mind, Just As the Engine Did the Horse measure for measure Unbridled mental power Artifcial intelligence is set to rival the human mind, just as the engine did the horse. José Hernández-Orallo looks at how we compare cognitive performance. here was a time when horses were a it comes to quantifying mental power. For major source of physical power. When example, Levin’s universal search makes it Tthe steam engine started to rival them, possible to define the difficulty of any inversion manufacturers wanted to know how many task. From here, the capability of a system horses a particular engine would replace. can be defined as an integral of performance James Watt soon realized how important over a range of difficulties. In this way, both these comparisons were, and conceived a new difficulty and capability are measured on a measure: the horsepower. From discussions ratio scale with the same unit: the logarithm of with millwrights, who used horses to turn the number of computational steps4. This unit their wheels, one mechanical horsepower is ultimately commensurate to bits, under the was estimated to be 33,000 foot-pounds per two terms of Levin’s universal search. minute — the measure was a great success. This conceptually appealing formulation And now, as artificial intelligence (AI) has some technical limitations. For example, emerges as an alternative source of mental without the choice of a reference machine, power, scientists are rethinking if and how Credit: Ievgen Chepil / Alamy Stock Vector the Kolmogorov complexity and the the mental capabilities of humans and logarithm of the number of computational machines can be measured. steps will depend on constants that need For the moment, humans still claim but the fact that it derives from human to be evaluated independently. One way superiority in mental power, but it’s clear populations. For obvious reasons, the notion to overcome those limitations might be that AI is becoming a rival. As in Watt’s of machine population in AI is thorny. Still, the further linking of computation and approach, one valuable comparison seems a bevy of AI competitions, benchmarks and information to physics. Indeed, there must be to be whether a particular AI system is more platforms have been recently introduced1. bounds between mental power and physical powerful than a ‘standard’ human, what is Progress is measured in terms of performance energy, and discovering them may shed light naively referred to as ‘human level’ machine on particular tasks, usually compared with on questions such as AI progress, intelligence intelligence. Perhaps project management some average human estimate. Cross-task growth, footprints on the environment and could lend us a measure: the ‘person month’. comparison remains elusive though, as many the effect of quantum computing on AI. Stimulating analogies aside, there are AI systems are specialized for a single task. By trying to identify the units of mental many differences between physical and Some would say that cognitive tasks cannot power and then linking them to physical mental work. The early psychometricians be reduced to a limited number of capabilities, units, we may well look eccentric from the pushed the analogy as far as they could, less so to a single one, because different tasks perspective of a thriving — and seemingly measuring intelligence as the capability of can never be compared. But others would unbridled — AI field. But like Watt two producing a particular kind of information- say that intelligence may have different centuries ago, sometimes we have to put the processing work. However, psychometric manifestations and a complex structure, a cart before the horse. ❐ measurement derives from human phenomenon that is common in physics. By populations. In many of its forms, it just looking at the performance of a system on José Hernández-Orallo1,2 captures a deviation from the mean, but not one set of tasks, we might be able to predict its 1Department of Computer Science, Universitat an actual magnitude. There is no such thing performance on a different set of tasks. Politècnica de València, Valencia, Spain. 2Leverhulme as an imperial foot for intelligence. The range between these two extremes — Centre for the Future of Intelligence, University of So back in the late eighteenth century, the importance of bias — is immanent Cambridge, Cambridge, UK. what Watt did was Copernican: he measured within machine learning. e-mail: [email protected] horses in terms of universal physical measures In this context, Ray Solomonoff’s prediction — feet, pounds and minutes. In Watt’s time, theory2 and Leonid Levin’s universal heuristics3 Published online: 2 January 2019 the understanding of the physical world was see Occam’s razor as a bias that emerges from https://doi.org/10.1038/s41567-018-0388-1 sufficiently mature to realize that the power algorithmic information theory — a possible needed in a mill could be compared to the foundation for computational measures of References power needed to boil a pot of water. intelligence. These theories are still insufficient 1. Castelvecchi, D. Nature 540, 323–324 (2016). 2. Solomonof, R. J. Inf. Control 7, 1–22 (1964). In contrast, even today, there is nothing like for nailing down such a measure, and they 3. Levin, L. A. in Algorithmic Probability and Friends. Bayesian a unit for mental power, which is independent are superficially different from deep learning, Prediction and Artifcial Intelligence (ed. Dowe, D.L.) of both the human and the task. In fact, the the dominant paradigm in AI today. Still, they 53–54 (Springer, Berlin, Heidelberg, 2013). 4. Hernández-Orallo, J. Te Measure of all Minds: Evaluating main problem for adapting psychometrics to have more potential than any other current Natural and Artifcial Intelligence (Cambridge University Press, AI is not only the lack of a ratio-scale measure, computational theory of intelligence when Cambridge, 2017). 106 NATURE PHYSICS | VOL 15 | JANUARY 2019 | 106 | www.nature.com/naturephysics.
Recommended publications
  • Ray Solomonoff, Founding Father of Algorithmic Information Theory [Obituary]
    UvA-DARE (Digital Academic Repository) Ray Solomonoff, founding father of algorithmic information theory [Obituary] Vitanyi, P.M.B. DOI 10.3390/a3030260 Publication date 2010 Document Version Final published version Published in Algorithms Link to publication Citation for published version (APA): Vitanyi, P. M. B. (2010). Ray Solomonoff, founding father of algorithmic information theory [Obituary]. Algorithms, 3(3), 260-264. https://doi.org/10.3390/a3030260 General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:29 Sep 2021 Algorithms 2010, 3, 260-264; doi:10.3390/a3030260 OPEN ACCESS algorithms ISSN 1999-4893 www.mdpi.com/journal/algorithms Obituary Ray Solomonoff, Founding Father of Algorithmic Information Theory Paul M.B. Vitanyi CWI, Science Park 123, Amsterdam 1098 XG, The Netherlands; E-Mail: [email protected] Received: 12 March 2010 / Accepted: 14 March 2010 / Published: 20 July 2010 Ray J.
    [Show full text]
  • The Dartmouth College Artificial Intelligence Conference: the Next
    AI Magazine Volume 27 Number 4 (2006) (© AAAI) Reports continued this earlier work because he became convinced that advances The Dartmouth College could be made with other approaches using computers. Minsky expressed the concern that too many in AI today Artificial Intelligence try to do what is popular and publish only successes. He argued that AI can never be a science until it publishes what fails as well as what succeeds. Conference: Oliver Selfridge highlighted the im- portance of many related areas of re- search before and after the 1956 sum- The Next mer project that helped to propel AI as a field. The development of improved languages and machines was essential. Fifty Years He offered tribute to many early pio- neering activities such as J. C. R. Lick- leiter developing time-sharing, Nat Rochester designing IBM computers, and Frank Rosenblatt working with James Moor perceptrons. Trenchard More was sent to the summer project for two separate weeks by the University of Rochester. Some of the best notes describing the AI project were taken by More, al- though ironically he admitted that he ■ The Dartmouth College Artificial Intelli- Marvin Minsky, Claude Shannon, and never liked the use of “artificial” or gence Conference: The Next 50 Years Nathaniel Rochester for the 1956 “intelligence” as terms for the field. (AI@50) took place July 13–15, 2006. The event, McCarthy wanted, as he ex- Ray Solomonoff said he went to the conference had three objectives: to cele- plained at AI@50, “to nail the flag to brate the Dartmouth Summer Research summer project hoping to convince the mast.” McCarthy is credited for Project, which occurred in 1956; to as- everyone of the importance of ma- coining the phrase “artificial intelli- sess how far AI has progressed; and to chine learning.
    [Show full text]
  • A Philosophical Treatise of Universal Induction
    Entropy 2011, 13, 1076-1136; doi:10.3390/e13061076 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article A Philosophical Treatise of Universal Induction Samuel Rathmanner and Marcus Hutter ? Research School of Computer Science, Australian National University, Corner of North and Daley Road, Canberra ACT 0200, Australia ? Author to whom correspondence should be addressed; E-Mail: [email protected]. Received: 20 April 2011; in revised form: 24 May 2011 / Accepted: 27 May 2011 / Published: 3 June 2011 Abstract: Understanding inductive reasoning is a problem that has engaged mankind for thousands of years. This problem is relevant to a wide range of fields and is integral to the philosophy of science. It has been tackled by many great minds ranging from philosophers to scientists to mathematicians, and more recently computer scientists. In this article we argue the case for Solomonoff Induction, a formal inductive framework which combines algorithmic information theory with the Bayesian framework. Although it achieves excellent theoretical results and is based on solid philosophical foundations, the requisite technical knowledge necessary for understanding this framework has caused it to remain largely unknown and unappreciated in the wider scientific community. The main contribution of this article is to convey Solomonoff induction and its related concepts in a generally accessible form with the aim of bridging this current technical gap. In the process we examine the major historical contributions that have led to the formulation of Solomonoff Induction as well as criticisms of Solomonoff and induction in general. In particular we examine how Solomonoff induction addresses many issues that have plagued other inductive systems, such as the black ravens paradox and the confirmation problem, and compare this approach with other recent approaches.
    [Show full text]
  • Dartmouth College Artificial Intelligence Conference the Next Fifty Years
    To proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can be so precisely described that a machine can be made to simulate it. – Goal proposed for AI by the 1956 Dartmouth Summer Research Project on Artificial Intelligence Dartmouth College Artificial Intelligence Conference The Next Fifty Years AI@50 Co-Sponsored by: Office of the Dean of the Faculty, Welcome! Dartmouth College Office of the Provost, Fifty years ago, pioneer AI researchers Dartmouth College gathered on this campus for their first-ever meeting in 1956, convening the Dartmouth Summer Research Project on Artificial The project is also sponsored in part Intelligence. The term itself had to be through a grant from the coined by John McCarthy, then a Dartmouth Defense Advanced Research Projects Agency math professor, to apply for a grant to fund The content of the information does not necessarily the original project. By end of summer, one reflect the position or the policy of the Government. of the first AI operating programs, Simon and No official endorsement should be inferred. Newell’s Logic Theorist, had presented demonstrable proof of a future for AI We wish to express our gratitude to through digital computer development. these cooperating organizations: This summer, the College commemorates that Summer Research Project by again hosting the 2006 Dartmouth Artificial Intelligence Conference: The Next Fifty Years, over July 13-15. We wish to thank the Office of the Dean of the Faculty and the Office of the Provost for their originating support toward forwarding progress on AI into the 21st century.
    [Show full text]
  • Kolmogorov Complexity and Its Applications
    CHAPTER 4 Kolmogorov Complexity and its Applications Ming LI Aiken Computation Laboratory, Han•ard University, Cambridge, MA 02138, USA Paul M.B. VIT ANYI Centrum voor Wiskunde en biformatica, Kruislaan 413, 1098 SJ Amsterdam, Netherlands, and Facu/teit Wiskunde en Jnjormatica, Universiteic van Amsterdam, Amsterdam, Netherlands Contents L Introduction 189 2. Mathematical theory 196 3. Applications of compressibility 214 4. Example of an application in mathematics: weak prime number theorems. 221 5. Applications of incompressibility: proving lower bounds 222 6. Resource-bounded Kolmogorov complexity and its applications 236 7. Conclusion 247 Acknowledgment 247 References :48 HANDBOOK OF THEORETICAL COMPUTER SCIENCE Edited by J. van Leeuwen ((' Elsevier Science Publishers B.V., 1990 KOLMOGOROV COMPLEXITY AND ITS APPLICATIONS 189 t. Introduction In everyday language we identify the information in an individual object with the essentials of a description for it. We can formalize this by defining the amount of information in a finite object (like a string) as the size (i.e., number of bits) of the smallest program that, starting with a blank memory, outputs the string and then terminates. A similar definition can be given for infinite strings, but in this case the program produces element after element forever. Thus, 1n (a string of nones) contains little information because a program of size about log n outputs it (like "print n ls"). Likewise, the transcendental number rr = 3.1415 ... , an infinite sequence of seemingly "random" decimal digits, contains O(l) information. (There is a short program that produces the consecutive digits of rr forever.) Such a definition would appear to make the amount of information in a string depend on the particular programming language used.
    [Show full text]
  • CS 4700: Foundations of Artificial Intelligence
    CS 4700: CS 4701: Foundations of Practicum in Artificial Intelligence Artificial Intelligence Fall 2017 Instructor: Prof. Haym Hirsh Lecture 1 Irving Ives, 1896-1962 CS 4700: CS 4701: Foundations of Practicum in Artificial Intelligence Artificial Intelligence Fall 2017 Instructor: Prof. Haym Hirsh Doubled in 4 years ? New CS Professors and Lecturers Hired in the Last 3 Years Class is Full All 287 seats are taken Class is Full Please drop as soon as you know you’re not taking the class Today • Overview of AI • Overview of 4700 Next Time • Introduction • Last 15 minutes: 4701 What is Artificial Intelligence? What is Intelligence? Intelligence Intelligence Manipulate Play Plan and Use Language See Learn and Move Games Reason Artificial Intelligence Manipulate Play Plan and Use Language See Learn and Move Games Reason Artificial Intelligence Natural Computer Machine Planning/ Language Robotics Games Automated Understanding Vision Learning Reasoning Artificial Intelligence Natural Computer Machine Planning/ Language Robotics Games Automated Understanding Vision Learning Reasoning (1950s) John McCarthy (1927-2011) The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it. 1. Ray Solomonoff 11. Abraham Robinson 2. Marvin Minsky 12. Tom Etter 3. John McCarthy 13. John Nash 4. Claude Shannon 14. David Sayre 5. Trenchard More 15. Arthur Samuel 6. Nathaniel Rochester 16. Shoulders 7. Oliver Selfridge 17. Shoulder's friend 8. Julian Bigelow 18. Alex Bernstein 9. W. Ross Ashby 19. Herbert Simon 10. W.S. McCulloch 20.
    [Show full text]
  • Raymond J. Solomonoff 1926–2009 Peter Gács Boston University Paul M
    11 Raymond J. Solomonoff 1926–2009 Peter Gács Boston University Paul M. B. Vitányi CWI February 7, 2011 Ray Solomonoff, the first inventor of some of the fundamen- work with the idea of applying it to classical statistical infer- tal ideas of Algorithmic Information Theory, died in December, ence [10,11]. 2009. His original ideas helped start the thriving research areas of algorithmic information theory and algorithmic inductive in- Since Solomonoff is the first inventor of Algorithmic Informa- ference. His scientific legacy is enduring and important. He was tion Theory, one can raise the question whether we ought to talk also a highly original, colorful personality, warmly about “Solomonoff complexity”. However, the remembered by everybody whose life he touched. name “Kolmogorov complexity” for shortest We outline his contributions, placing it into its his- effective description length has become well torical context, and the context of other research in entrenched and is commonly understood. Sol- algorithmic information theory. omonoff’s publications apparently received little attention until Kolmogorov started to 1. Introduction refer to them from 1968 onward. Says Kol- mogorov, “I came to similar conclusions [as Raymond J. Solomonoff died on December 7, 2009, Solomonoff], before becoming aware of Solo- in Cambridge, Massachusetts. He was the first in- monoff’s work, in 1963–1964” and “The basic ventor of some of the fundamental ideas of Algo- discovery, which I have accomplished inde- rithmic Information Theory, which deals with the pendently from and simultaneously with R. shortest effective description length of objects and Solomonoff, lies in the fact that the theory of is commonly designated by the term “Kolmogorov algorithms enables us to eliminate this arbi- complexity.“ trariness by the determination of a ‘complexi- ty’ which is almost invariant (the replacement In the 1950s Solomonoff was one of the first re- of one method by another leads only to the searchers to introduce probabilistic grammars and addition of a bounded term)” the associated languages.
    [Show full text]
  • Inventing Intelligence: on the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century
    UNIVERSITY OF CAMBRIDGE Inventing Intelligence: On the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century This thesis is submitted for the degree of Doctor of Philosophy In History and Philosophy of Science By Jonathan Nigel Ross Penn (‘Jonnie Penn’) Pembroke College, Cambridge Examiners: Professor Simon Schaffer, University of Cambridge Professor Jon Agar, University College London Supervisor: Dr. Richard Staley, University of Cambridge Advisor: Dr. Helen Anne Curry, University of Cambridge Word Count: 78,033 Date: 14 December 2020 This thesis is the result of my own work and includes nothing which is the outcome of work done in collaboration except as declared in the Preface and specified in the text. It is not substantially the same as any that I have submitted, or, is being concurrently submitted for a degree or diploma or other qualification at the University of Cambridge or any other University or similar institution except as declared in the Preface and specified in the text. I further state that no substantial part of my thesis has already been submitted, or, is being concurrently submitted for any such degree, diploma or other qualification at the University of Cambridge or any other University or similar institution except as declared in the Preface and specified in the text. It does not exceed the prescribed word limit for the Degree Committee of the Department of History and Philosophy of Science at the University of Cambridge. Copyright Jonnie Penn, 2020. All rights reserved. 2 Abstract Inventing Intelligence: On the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century In the mid-1950s, researchers in the United States melded formal theories of problem solving and intelligence with another powerful new tool for control: the electronic digital computer.
    [Show full text]
  • Abelian Varieties in Coding and Cryptography
    IEEE Information Theory Society Newsletter Vol. 61, No. 1, March 2011 Editor: Tracey Ho ISSN 1045-2362 Editorial committee: Helmut Bölcskei, Giuseppe Caire, Meir Feder, Joerg Kliewer, Anand Sarwate, and Andy Singer President’s Column Giuseppe Caire I am deeply honored to serve as your Society too often forgotten by many politicians and peo- president for 2011. I’d like to begin with ac- ple with responsibilities around the world. On knowledging my recent predecessors, Frank the other hand, we have a lot of work to do and Kschischang (Junior Past President) and Andrea we must be proactive in facing the challenges Goldsmith (Senior Past President), as well as that the future is putting in front of us. Dave Forney, who has offi cially ended his ser- vice as an offi cer of the IT Society on December While our Information Theory Transactions 31st, 2010, after a long and perhaps unmatched continues to rank at the top of the ISI Journal history of leadership and guidance, that stands Citation Report among all journals in Electri- as a reference for all of us. Under their lead, the cal and Electronic Engineering and Computer IT Society has sailed through some signifi cant Science in total citations, and fi rst among all storms, such as an unprecedented economic re- journals in Electrical Engineering, Computer cession, causing turmoil for IEEE fi nances, and a Science and Applied Mathematics according number of sad events, including the premature to the Eigenfactor tm score, the “sub-to-pub” passing of Shannon Lecturer Rudolf Ahlswede time (i.e., the length of the time between fi rst and other notables.
    [Show full text]
  • Ray Solomonoff, Founding Father of Algorithmic Information Theory
    Algorithms 2010, 3, 260-264; doi:10.3390/a3030260 OPEN ACCESS algorithms ISSN 1999-4893 www.mdpi.com/journal/algorithms Obituary Ray Solomonoff, Founding Father of Algorithmic Information Theory Paul M.B. Vitanyi CWI, Science Park 123, Amsterdam 1098 XG, The Netherlands; E-Mail: [email protected] Received: 12 March 2010 / Accepted: 14 March 2010 / Published: 20 July 2010 Ray J. Solomonoff died on December 7, 2009, in Cambridge, Massachusetts, of complications of a stroke caused by an aneurism in his head. Ray was the first inventor of Algorithmic Information Theory which deals with the shortest effective description length of objects and is commonly designated by the term “Kolmogorov complexity.” In the 1950s Solomonoff was one of the first researchers to treat probabilistic grammars and the associated languages. He treated probabilistic Artificial Intelligence (AI) when “probabilistic” was unfashionable, and treated questions of machine learning early on. But his greatest contribution is the creation of Algorithmic Information Theory. Already in November 1960, Solomonoff published [1] that presented the basic ideas of Algorithmic Information Theory as a means to overcome serious problems associated with the application of Bayes‟s rule in statistics. Ray Solomonoff was born on July 25, 1926, in Cleveland, Ohio, United States. He studied physics during 1946–1950 at the University of Chicago (he recalls the lectures of E. Fermi and R. Carnap) and obtained an M.Sc. from that university. From 1951–1958 he worked in the electronics industry doing math and physics and designing analog computers, working half-time. His scientific autobiography up to 1997 is published as [2].
    [Show full text]
  • From Artificial Intelligence to Machine Learning
    From Artificial Intelligence to Machine Learning Dr. Paul A. Bilokon, an alumnus of Christ Church, Oxford, and Imperial College, is a quant trader, scientist, and entrepreneur. He has spent more than twelve years in the City of London and on Wall Street – at Morgan Stanley, Lehman Brothers, Nomura, Citigroup, and Deutsche Bank, where he has led a team of electronic traders spread over three continents. He is the Founder and CEO of Thalesians Ltd – a consultancy and think tank focussing on new scientific and philosophical thinking in finance. Paul’s academic interests include mathematical logic, stochastic analysis, markets microstructure, and machine learning. His research work has appeared in LICS, Theoretical Computer Science, and has been published by Springer and Wiley. Ever since the dawn of humanity inventors have been dreaming of creating sentient and intelligent beings. Hesiod (c. 700 BC) describes how the smithing god Hephaestus, by command of Zeus, moulded earth into the first woman, Pandora. The story wasn’t a particularly happy one: Pandora opened a jar – Pandora’s box – realeasing all the evils of humanity, leaving only Hope inside when she closed the jar again. Ovid (43 BC – 18 AD) tells us a happier legend, that of a Cypriot sculptor, Pygmalion, who carved a woman out of ivory and fell in love with her. Miraculously, the statue came to life. Pygmalion married Galatea (that was the statue’s name) and it is rumoured that the town Paphos is named after the couple’s daughter. Stories of golems, anthropomorphic beings created out of mud, earth, or clay, originate in the Talmud and appear in the Jewish tradition throughout the Middle Ages and beyond.
    [Show full text]
  • The Evaluation Metric in Generative Grammar
    The evaluation metric in generative grammar John Goldsmith December 8, 2011 1 Introduction It is both a privilege and a great pleasure to be able to be here today, and celebrate the fiftieth anniversary of the MIT program in linguistics, and to say thank you to the great teachers and great friends and colleagues that we have known here: my own teachers, Noam Chomsky and Morris Halle and Haj Ross and Paul Kiparsky, certainly, and it is a time to remember some who cannot be here today, like Ken Hale, and Yuki Kuroda, and Jim McCawley. 1 The subject which I would like to address today is the evaluation metric in generative grammar. Why? Arguably, the evaluation metric is both the most novel and the most important concept in the development of generative grammar by Noam Chomsky. And yet it is at the same time one of the least recognized and surely most misunderstood of the core concepts of generative grammar. So there you are: the evaluation metric is critically important, it is arguably novel, it is misunderstood, and at some times and in some places, it has even been reviled. What better reasons could there be for spending our time today talking about it? In the twenty minutes available to me|hardly long enough to recite a poem, let alone discuss a misun- derstood idea|I would like, first, to explain the idea of the evaluation metric in early generative grammar; this will mean exploring the separate ideas of (1) a prior over the set of grammars and (2) a measure of goodness of fit to the data.
    [Show full text]