<<

The Blackwell Guide to the of Computing and Information Blackwell Philosophy Guides Series Editor: Steven M. Cahn, City University of New York Graduate School Written by an international assembly of distinguished philosophers, the Blackwell Philosophy Guides create a groundbreaking student resource – a complete critical survey of the central themes and issues of philosophy today. Focusing and advancing key arguments throughout, each essay incorporates essential background material serving to clarify the history and of the relevant topic. Accordingly, these volumes will be a valuable resource for a broad range of students and readers, including professional philosophers. 1 The Blackwell Guide to Edited by John Greco and Ernest Sosa 2 The Blackwell Guide to Ethical Theory Edited by Hugh LaFollette 3 The Blackwell Guide to the Modern Philosophers Edited by Steven M. Emmanuel 4 The Blackwell Guide to Philosophical Logic Edited by Lou Goble 5 The Blackwell Guide to Social and Edited by Robert L. Simon 6 The Blackwell Guide to Business Edited by Norman E. Bowie 7 The Blackwell Guide to the Edited by Peter Machamer and Michael Silberstein 8 The Blackwell Guide to Edited by Richard M. Gale 9 The Blackwell Guide to the Edited by Nigel Blake, Paul Smeyers, Richard Smith, and Paul Standish 10 The Blackwell Guide to Philosophy of Edited by Stephen P. Stich and Ted A. Warfield 11 The Blackwell Guide to the Philosophy of the Social Sciences Edited by Stephen P. Turner and Paul A. Roth 12 The Blackwell Guide to Edited by Robert C. Solomon and David Sherman 13 The Blackwell Guide to Edited by Christopher Shields 14 The Blackwell Guide to the Philosophy of Computing and Information Edited by Luciano Floridi 15 The Blackwell Guide to Edited by Peter Kivy The Blackwell Guide to the Philosophy of Computing and Information

Edited by Luciano Floridi © 2004 by Blackwell Publishing Ltd

350 Main Street, Malden, MA 02148-5020, USA 108 Cowley Road, Oxford OX4 1JF, UK 550 Swanston Street, Carlton, Victoria 3053, Australia

The right of Luciano Floridi to be identified as the Author of the Editorial Material in this Work has been asserted in accordance with the UK Copyright, Designs, and Patents Act 1988.

All rights reserved. No part of this publication may be reproduced, stored in a retrieval , or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by the UK Copyright, Designs, and Patents Act 1988, without the prior permission of the publisher.

First published 2004 by Blackwell Publishing Ltd

Library of Congress Cataloging-in-Publication Data

The Blackwell guide to the philosophy of computing and information / edited by Luciano Floridi. p. cm. — (Blackwell philosophy guides) Includes bibliographical references and index. ISBN 0-631-22918-3 (alk. paper) — ISBN 0-631-22919-1 (pbk. : alk. paper) 1. Computer science—Philosophy. 2. Information technology—Philosophy. I. Floridi, Luciano, 1964– II. Series.

QA76.167.B53 2003 004′.01—dc21 2003045339

A catalogue record for this title is available from the British Library.

Set in 9/11.5pt Galliard by Graphicraft Limited, Hong Kong Printed and bound in the United Kingdom by MPG Books Ltd, Bodmin, Cornwall

For further information on Blackwell Publishing, visit our website: http://www.blackwellpublishing.com Contents

Notes on Contributors viii Preface xi

Part I: Four 1

1 Computation 3 B. Jack Copeland 2 Complexity 18 Alasdair Urquhart 3 System: An Introduction to Science 28 Klaus Mainzer 4 Information 40 Luciano Floridi

Part II: Computers in Society 63

5 Computer Ethics 65 Deborah G. Johnson 6 Computer-mediated Communication and Human–Computer Interaction 76 Charles Ess 7 Internet Culture 92 Wesley Cooper 8 Digital Art 106 Dominic McIver Lopes

v Contents

Part III: Mind and AI 117

9 The Philosophy of AI and its Critique 119 James H. Fetzer 10 Computationalism, , and the 135 Brian P. McLaughlin

Part IV: Real and Virtual Worlds 153

11 155 Barry Smith 12 Virtual 167 Derek Stanovsky 13 The of Information 178 Eric Steinhart 14 186 Roberto Cordeschi 15 Artificial Life 197 Mark A. Bedau

Part V: Language and Knowledge 213

16 Information and Content 215 Jonathan Cohen 17 Knowledge 228 Fred Adams 18 The Philosophy of Computer Languages 237 Graham White 19 Hypertext 248 Thierry Bardini

Part VI: Logic and Probability 261

20 Logic 263 G. Aldo Antonelli 21 Probability in Artificial 276 Donald Gillies 22 Game Theory: Nash Equilibrium 289 Cristina Bicchieri

vi Contents

Part VII: Science and Technology 305

23 Computing in the Philosophy of Science 307 Paul Thagard 24 Methodology of Computer Science 318 Timothy Colburn 25 Philosophy of Information Technology 327 Carl Mitcham 26 Computational Modeling as a Philosophical Methodology 337 Patrick Grim

Index 350

vii Notes on Contributors

Fred Adams is Professor of Oregon; Adjunct Professor of Systems Science and Philosophy, and Chair of the Department at Portland State University; and Editor-in-chief of Philosophy at the University of Delaware. He of the journal Artificial Life (MIT Press). He has also taught at Augustana College, Central has published extensively on both scientific and Michigan University, Lawrence University, and philosophical aspects of artificial life. the University of Wisconsin-Madison. A sample Cristina Bicchieri is Professor of Philosophy of his publications include the co-edited book and Social and Decision Sciences at Carnegie Reflections on Philosophy, and authored or co- Mellon University. She is the author of Ration- authored articles, “Cognitive Trying,” “Causal ality and Coordination (1993, 1997), and co- Contents,” “Fodorian Semantics,” and “Vacu- author of The Dynamics of Norms (1998), The ous Singular Terms.” Logic of Strategy (2000) and Knowledge, Belief, G. Aldo Antonelli is Associate Professor of Logic and Strategic Interaction (1992). She has pub- and Philosophy of Science at the University of lished extensively in philosophy, AI, game theory, California, Irvine. He has worked in applications and the social sciences. of logic to artificial intelligence and game theory, Jonathan Cohen is Assistant Professor of non-well-founded set theories, modal logic, and Philosophy at the University of California, San and mathematics. Diego. He is the author of several articles in Thierry Bardini is Associate Professor in the philosophy of mind and philosophy of lan- Communication Department at the University of guage. Much of his recent work has concerned Montréal. He holds a degree in agronomy (ENSA the metaphysics of color. Montpellier, 1986) and a Ph.D. in sociology Timothy Colburn has been Professor of Com- (Paris X, 1991). Since 1992 he has conducted puter Science at the University of Minnesota- research on the sociological history of comput- Duluth since 1988. Prior to that, he was principal ing, with a special emphasis on the genesis of research scientist for Honeywell Inc. He is the personal computing. author of Philosophy and Computer Science (2000) and co-editor of Program Verification: Funda- Mark A. Bedau received his Ph.D. in philo- mental Issues in Computer Science (1993). sophy from University of California at Berkeley in 1985. He is currently Professor of Philosophy Wesley Cooper is Professor of Philosophy and Humanities at Reed College in Portland, at the University of Alberta in Edmonton,

viii Notes on Contributors

Alberta, Canada. He is the author of The Unity Luciano Floridi is Associate Professor of Logic of William James’s (2002) and the and Epistemology at the University of Bari, Italy, chief administrator of Alberta MOO (http:// and Markle Foundation Fellow in Information www.arts.ualberta.ca:3000). Policy at the Oxford University in the UK. His publications include over 30 articles on the philo- B. Jack Copeland is Professor of Philosophy at sophy of computing and information and Sextus the University of Canterbury, New Zealand, and Empiricus (2002), Philosophy and Computing – Director of the Turing Archive for the History An Introduction (1999), and Scepticism and the of Computing. He works in mathematical and Foundation of Epistemology (1996). He was the philosophical logic, cognitive science, and the consultant editor for the Iter Italicum on CD- history and foundations of computing, and has ROM (1995) and for the Routledge Encyclopedia numerous articles in journals including Journal of Philosophy on CD-ROM (1998). He is the of Philosophy, Mind, Analysis, and Scientific founding director of the Italian Web Site for American. He is author of Artificial Intelligence: Philosophy (www.swif.it). A Philosophical Introduction (Blackwell, 1993, 2nd ed. forthcoming) and is currently writing Donald Gillies is Professor in the Philosophy and editing several books on Turing, including Department of King’s College, University of one on Turing’s Automatic Computing Engine, London. Since 1966 he has carried out research and editing a volume entitled Colossus: The First in the philosophy of science and mathematics. Electronic Computer. His edited volume Logic He has published 5 books, and edited a collec- and Reality: Essays on the Legacy of Arthur Prior tion on “Revolutions in Mathematics.” From the appeared in 1996. late 1980s, he has taken a particular interest in interactions between philosophy and AI, and has Roberto Cordeschi is Professor of Philosophy been involved in four interdisciplinary research of Science at the University of Salerno, Italy. He projects in this area. is the author of several publications in the history of cybernetics and in the epistemological issues Patrick Grim is SUNY Distinguished Teaching of cognitive science and artificial intelligence, Professor, State University of New York at Stony including The Discovery of the Artificial (2002). Brook. He is author of The Incomplete Universe: Totality, Knowledge, and Truth (1991), co- Charles Ess is Professor of Philosophy and author of The Philosophical Computer: Exploratory Religion, and Director of the Center for Inter- Essays in Philosophical Computer Modeling (with disciplinary Studies, Drury University. Ess has Gary Mar and Paul St. Denis, 1998), and found- received awards for teaching excellence and scho- ing co-author of 22 volumes of The Philosopher’s larship, as well as a national award for his work in Annual. He has published a variety of articles hypermedia. He has published in interdisciplinary incorporating computer modeling in journals in ethics, hypertext and democratization, history of philosophy, computer science, linguistics, and philosophy, feminist biblical studies, contempor- theoretical biology. ary Continental philosophy, computer resources for humanists, and the interactions between cul- Deborah G. Johnson is the Anne Shirley Carter ture, technology, and communication. Olsson Professor of in the Depart- ment of Technology, Culture, and Commun- James H. Fetzer is McKnight Professor of ication within the School of and Philosophy at the University of Minnesota, and at the University of Virginia teaches on its Duluth campus. The founding (Charlottesville, Virginia, USA). She specializes editor of the book series Studies in Cognitive in ethical, social, and policy issues involving tech- Systems and of the journal and Machines, nology, especially issues involving computers and he has published more than 20 books and 100 the internet. articles and reviews in the philosophy of science and on the theoretical foundations of computer Dominic McIver Lopes is Associate Professor science, artificial intelligence, and cognitive of Philosophy at the University of British science. Columbia. He writes on issues at the intersection

ix Notes on Contributors of the philosophy of art and the philosophy of Derek Stanovsky is Director of Internet Studies mind, and is the author of Pic- at Appalachian State University and an Assistant tures and co-editor of the Routledge Companion Professor in the Department of Interdisciplinary to Aesthetics. He is currently working on a book Studies. His research interests include internet entitled Live Wires: The Digital Arts. studies, feminist theory, and contemporary con- tinental philosophy, and his articles have appeared Klaus Mainzer is Professor of Philosophy of in The National Women’s Studies Association Jour- Science and Director of the Institute of Inter- nal, Jouvert: A Journal of Postcolonial Studies, disciplinary Informatics (http://www.informatik. and Feminist Teacher. uni-augsburg.de/I3) at the University of Augs- burg. He is president of the German Society Eric Steinhart is Professor of Philosophy at of Complex Systems and Nonlinear Dynamics, William Paterson University. He works mainly author and editor of several books on philosophy on metaphysics and has written on possible of science, systems science, cognitive and com- worlds semantics for metaphors, the logical puter science. foundations of physical theory, theories of Brian P. McLaughlin is a Professor of Philo- transfinite computation, and the of sophy at Rutgers University and Chairperson of persons. the Department. He is the author of numerous Paul Thagard is Professor of Philosophy and papers in the philosophy of mind, metaphysics, Director of the Cognitive Science Program at epistemology, and philosophical logic. the University of Waterloo, Canada. His books Carl Mitcham is Professor of Liberal Arts and include Coherence in Thought and Action (2000) International Studies at the Colorado School and How Scientists Explain Disease (1999). of Mines. His books include Thinking through Technology: The Path between Engineering and Alasdair Urquhart is Professor of Philosophy Philosophy (1994) and a co-edited (with Stephen and Computer Science at the University of H. Cutcliffe) volume on Visions of STS: Counter- Toronto. He has published papers in non- points in Science, Technology, and Society Studies classical , algebraic logic, and complexity (2001). theory, and is the editor of volume 4 of the Collected Papers of . Barry Smith is Park Professor of Philosophy at the State University of New York at Buffalo and Graham White is Lecturer in Computer Sci- Director of the Institute for Formal Ontology ence at Queen Mary, University of London. He and Medical Information Science at the University has previously published on the history of logic of Leipzig. He is editor of The Monist and author and on the philosophy of common-sense reas- of numerous articles on ontology, the history of oning; currently he is working on the logic of Austro-, and related themes. action and its computational implementation.

x Preface

The information revolution has been changing the world produces between 1 and 2 exabytes of the world profoundly, irreversibly and problem- data, that is, roughly 250 megabytes for every atically for some time now, at a breathtaking pace human being on earth (source: Lyman & Varian, and with an unprecedented scope. Every year, online, see figure P1). An exabyte is approximately

Too much information World production of data, terabytes PC disk log 1999 estimates drives scale Grand total: 2,120,000 Tb Departmental 1,000,000 Photographs servers Camcorder tape 100,000 Enterprise X-rays servers 10,000

1,000 Office documents 100 Music CDs Newspapers DVDs Cinema Periodicals 10 Books Data CDs 1 Optical Paper Film Magnetic 0.08 0.24 Total 427 1,693 ’000Tb Source: University of California, Berkeley.

Figure P1

xi Preface

1,200 1,100 Worldwide 1,000 USA 900 Western Europe Asia-Pacific 800 700 600

Millions 500 400 300 200 100 0 1995 2000 2001 2007

Figure P2: PCs in use worldwide Source: Computer Industry Almanac Inc., .

1018 bytes, or a billion times a billion bytes, the vided a wealth of extremely powerful tools and equivalent of about 20 billion copies of this methodologies, raised a wide range of unique Guide. It has taken the entire history of human- problems and conceptual issues, and opened up ity to accumulate 12 exabytes of data. Stored on endless possibilities hitherto unimaginable. floppy disks, 12 exabytes of data would form a Inevitably, the information revolution has also stack 24 million miles high. At the rate of growth deeply affected what philosophers do, how they measured in 1999, humanity will already have think about their problems, what problems they created the next 12 exabytes by the time this consider worth their attention, how they con- Guide is published. ceptualize their views, and even the vocabulary To cope with exabytes of data, hundreds of they use (see Bynum & Moor 1998 and 2002, millions of computing machines are employed Colburn 2000, Floridi 1999, and Mitcham & every day. In 2001, the number of PCs in use Huning 1986 for references). It has made possi- worldwide reached 600M units (source: Com- ble new approaches and original investigations, puter Industry Almanac Inc., , see figure P2). By the crucial questions, and given new meaning to end of 2007, this number will have nearly classic problems and traditional topics. In short, doubled to over 1.15B PCs, at a compound information-theoretic and computational research annual growth of 11.4 percent. Of course, PCs are in philosophy has become increasingly innova- among the greatest sources of further exabytes. tive, fertile, and pervasive. It has already pro- Figures P1 and P2 give some quantitative sub- duced a wealth of interesting and important stance to the trite remark that we live in a “data- results. This Guide is the first systematic attempt based society.” They also show that the end of to map this new and vitally important area of the information society, understood as the mature research. Owing to the novelty of the field, it is stabilization in the growth of quantity of data an exploration as much as an introduction. and number of computational machines, is not As an introduction, the 26 chapters in this in sight. volume seek to provide a critical survey of the The databased society has been brought fundamental themes, problems, arguments, theor- about by the information revolution, whose main ies, and methodologies constituting the new field means have been first the PC and then the Web. of philosophy of computing and information (PCI). “Datifying” the world and human society has The chapters are organized into seven sections. created entirely new , made possible un- In Part I, four of the most crucial concepts precedented phenomena and experiences, pro- in PCI, namely computation, complexity, system,

xii Preface and information are analyzed. They are the four Blackwell and I agreed to use “philosophy of columns on which the other chapters are built, computing and information.” PCI is a new but as it were. The following six parts are dedicated still very recognizable label, which we hope will to specific areas: the information society (compu- serve both scholarly and marketing ends equally ter ethics; communication and interaction; cyber- well. In the introductory chapter, entitled “What philosophy and internet culture; and digital art); is the Philosophy of Information?,” I offer an mind and intelligence (philosophy of AI and its interpretation of the new informational critique; and computationalism, connectionism, in philosophy and argue that philosophy of infor- and the philosophy of mind); natural and mation (PI) is conceptually a much more satis- artificial realities (formal ontology; virtual reality; factory name for it, because it identifies far more the physics of information; cybernetics; and arti- clearly what really lies at the heart of the new ficial life); language and knowledge (meaning and paradigm. But much as I hope that PI will be- information; knowledge and information; formal come a useful label, I suspect that it would have languages; and hypertext theory); logic and prob- been premature and somewhat obscure as the ability (nonmonotonic logic; probabilistic reas- title for this volume. Since the chapter is meant oning; and game theory); and, finally, science, to prepare the ground for the Guide, I thought technology, and methodology (computing in the it would be convenient to make it available on philosophy of science; methodology of computer the Web free of charge (it can be found online science; philosophy of IT; and computational at http://www.blackwellpublishing.com/pci). The modeling as a philosophical methodology). Each reader may wish to consider that the project for chapter has been planned as a freestanding intro- the Guide was based on the hermeneutical frame duction to its subject. For this purpose, the outlined in that chapter. volume is further supported by an exhaustive glossary of technical terms, available online (http://www.blackwellpublishing.com/pci). Acknowledgments As an exploration, the Guide attempts to bring into a reasonable relation the many computational and informational issues with which philosophers Because of the innovative nature of the research have been engaged at least since the 1950s. The area, working on this Guide has been very chal- aim has been to identify a broad but clearly lenging. I relied on the patience and expertise of definable and well-delimited field where before so many colleagues, friends, and family members there were many special problems and that I wish to apologize in advance if I have whose interrelations were not always explicit or forgotten to mention anyone below. Jim Moor well understood. Each chapter is meant to pro- was one of the first people with whom I discussed vide not only a precise, clear, and accessible intro- the project and I wish to thank him for his time, duction but also a substantial and constructive suggestions, and support. Jeff Dean, philosophy contribution to the current debate. editor at Blackwell, has come close to instantiating Precisely because the Guide is also an explora- the Platonic of editor, with many comments, tion, the name given to the new field is some- ideas, suggestions, and the right kind of sup- what tentative. Various labels have recently been port. This Guide has been made possible also by suggested. Some follow fashionable terminology his far-sighted faith in the project. Nick Bellorini, (e.g. “cyberphilosophy,” “digital philosophy,” also editor at Blackwell, has been equally “computational philosophy”), while the major- important in the second stage of the editorial ity expresses specific theoretical orientations (e.g. project. I am also grateful to the two anonymous “philosophy of computer science,” “philosophy referees who provided constructive feedback. of computing/computation,” “philosophy of Many other colleagues, most of whom I have not AI,” “philosophy and computers,” “computing met in real life, generously contributed to the and philosophy,” “philosophy of the artificial,” shaping of the project by commenting on earlier “artificial epistemology,” “android epistemo- drafts through several mailing lists, especially logy”). For this Guide, the philosophy editors at [email protected], philinfo@yahoogroups.

xiii Preface com, [email protected], philosop@louisiana. to all contributors through this website and I edu, and [email protected]. I am grateful to the hope it will be possible to transform it into a list moderators and to Bryan Alexander, Colin permanent resource for the use of the Guide. Allen, Leslie Burkholder, Rafael Capurro, Tony The Programme in Comparative Media Law and Chemero, Ron Chrisley, Stephen Clark, Anthony Policy at Oxford University and its founding dir- Dardis, M. G. Dastagir, Bob Di Falco, Soraj ector Monroe Price greatly facilitated my work. Hongladarom, Ronald Jump, Lou Marinoff, Research for this project has been partly sup- Ioan-Lucian Muntean, Eric Palmer, Mario ported by a grant from the Coimbra Group, Piazza, John Preston, Geoffrey Rockwell, Gino Pavia University, and I wish to thank Lorenzo Roncaglia, Jeff Sanders, and Nelson Thompson. Magnani, Director of the Computational Philo- Unfortunately, for reasons of space, not all their sophy Lab (http://www.unipv.it/webphilos_lab/) suggestions could be followed in this context. and Mario Stefanelli, Director of the Medical Here are some of the topics left out or only Informatics Laboratory (http://dis.unipv.it/labs/ marginally touched upon: information science as labmed/home-e.html) for their support. Finally, applied philosophy of information; social episte- I wish to thank all the contributors for bearing mology and the philosophy of information; visual with me as chapters went through so many revi- thinking; pedagogical issues in PCI; the philo- sions; my father, for making me realize the obvi- sophy of information design and modeling; the ous, namely the exploratory nature of this project; philosophy of information economy; lambda cal- Paul Oldfield, for his copy-editing help; and my culus; linear logic; conditional reasoning; epistemic wife Kia, who not only implemented a wonder- logic; deontic logic; temporal logic; the logic of ful life for our family, but also listened to me action; fuzzy logic; situation logic; dynamic logic; patiently when things were not working, pro- common-sense reasoning and AI; causal reason- vided many good solutions to problems in which ing, the hermeneutical interpretation of AI. I had entangled myself, and went as far as to J. C. Beall, Jonathan Cohen, Gualtiero read my contributions and comment carefully Piccinini, Luigi Dappiano, and Saul Fisher sent me on their contents. The only thing she could not useful feedback on an earlier draft of the Glossary. do was to take responsibility for any mistakes Members of four research groups have still remaining. played an influential role in the development of the project. I cannot thank all of them but I Luciano Floridi wish to acknowledge the help I have received from IACAP, the International Association for Computing and Philosophy, directed by References Robert Cavalier (http://caae.phil.cmu.edu/caae/ CAP/), with its meetings at Carnegie Mellon Bynum, T. W. and Moor, J. H., eds. 1998. The (CAP@CMU); INSEIT, the International Soci- Digital Phoenix: How Computers are Changing Philosophy. Malden, MA and Oxford: Blackwell. ety for Ethics and Information Technology; the —— and ——, eds. 2002. CyberPhilosophy: The In- American Philosophical Association Committee tersection of Philosophy and Computing. Malden, on Philosophy and Computers (http://www. MA and Oxford: Blackwell. apa.udel.edu/apa/governance/committees/ Colburn, T. R. 2000. Philosophy and Computer Sci- computers/); and the Epistemology and Com- ence. Armonk, NY and London: M. E. Sharpe. puting Lab, directed by Mauro Di Giandomenico Floridi, L. 1999. Philosophy and Computing – An at the Philosophy Department of the Univer- Introduction. London and New York: Routledge. sity of Bari (http://www.ssscienza.uniba.it/ Lyman, P. and Varian, H. R. (online). “How Much index.html). I am also grateful to my College in Information?,” . made possible the organization of a website to Mitcham, C. and Huning, A., eds. 1986. Philosophy support the editorial work (http://www.wolfson. and Technology II – Information Technology and ox.ac.uk/~floridi/blackwell/index.htm). During Computers in Theory and Practice. Dordrecht/ the editorial process, files were made available Boston: Reidel.

xiv Part I Four Concepts Computation

Chapter 1 Computation B. Jack Copeland

by John von Neumann and in the UK by Max The Birth of the Modern Newman, the two mathematicians who were by Computer and large responsible for placing Turing’s abstract universal machine into the hands of electronic As everyone who can operate a personal computer engineers (Copeland 2001). By 1945, several knows, the way to make the machine perform groups in both countries had embarked on creat- some desired task is to open the appropriate ing a universal Turing machine in hardware. The program stored in the computer’s memory. Life race to get the first electronic stored-program was not always so simple. The earliest large-scale computer up and running was won by Manchester electronic digital computers, the British Colossus University where, in Newman’s Computing (1943) and the American ENIAC (1945), did not Machine Laboratory, the “Manchester Baby” ran store programs in memory (see Copeland 2001). its first program on June 21, 1948. By 1951, To set up these computers for a fresh task, it electronic stored-program computers had begun was necessary to modify some of the machine’s to arrive in the marketplace. The first model to wiring, rerouting cables by hand and setting go on sale was the Ferranti Mark I, the pro- switches. The basic principle of the modern com- duction version of the Manchester computer puter – the idea of controlling the machine’s (built by the Manchester firm Ferranti Ltd.). Nine operations by means of a program of coded of the Ferranti machines were sold, in Britain, instructions stored in the computer’s memory – Canada, Holland, and Italy, the first being was thought of by Alan Turing in 1935. His installed at Manchester University in February abstract “universal computing machine,” soon 1951. In the US, the Computer Corporation known simply as the universal Turing machine sold its first UNIVAC later the same year. The (UTM), consists of a limitless memory, in which LEO computer also made its debut in 1951; both data and instructions are stored, and a LEO was a commercial version of the prototype scanner that moves back and forth through the EDSAC machine, which at Cambridge Uni- memory, symbol by symbol, reading what it finds versity in 1949 had become the second stored- and writing further symbols. By inserting differ- program electronic computer to function. In ent programs into the memory, the machine is 1953 came the IBM 701, the company’s first made to carry out different computations. mass-produced stored-program electronic com- Turing’s idea of a universal stored-program puter (strongly influenced by von Neumann’s computing machine was promulgated in the US prototype IAS computer, which was working at

3 B. Jack Copeland

01011001

SCANNER

Figure 1.1: A Turing machine

Princeton University by the summer of 1951). consists of a tape divided into squares. Each square A new era had begun. may be blank or may bear a single symbol, “0” Turing introduced his abstract Turing or “1,” for example, or some other symbol taken machines in a famous article entitled “On Com- from a finite alphabet. The scanner is able to putable Numbers, with an Application to the examine only one square of tape at a time (the Entscheidungsproblem” (published in 1936). “scanned square”). (See figure 1.1.) The tape is Turing referred to his abstract machines simply the machine’s general-purpose storage medium, as “computing machines” – the American logician serving as the vehicle for input and output, and Alonzo Church dubbed them “Turing machines” as a working memory for storing the results of (Church 1937: 43). “On Computable Numbers” intermediate steps of the computation. The tape pioneered the theory of computation and is may also contain a program of instructions. The regarded as the founding publication of the input that is inscribed on the tape before the modern science of computing. In addition, computation starts must consist of a finite Turing charted areas of mathematics lying bey- number of symbols. However, the tape itself is ond the reach of the UTM. He showed that not of unbounded length – since Turing’s aim was to all precisely-stated mathematical problems can show that there are tasks which these machines be solved by a Turing machine. One of them is are unable to perform, even given unlimited the Entscheidungsproblem – “decision problem” working memory and unlimited time. (A Turing – described below. This discovery wreaked havoc machine with a tape of fixed finite length is called with received mathematical and philosophical a finite state automaton. The theory of finite state opinion. Turing’s work – together with contem- automata is not covered in this chapter. An intro- poraneous work by Church (1936a, 1936b) – duction may be found in Sipser 1997.) initiated the important branch of mathematical logic that investigates and codifies problems “too hard” to be solvable by Turing machine. In a The Basic Operations of single article, Turing ushered in both the mod- a Turing Machine ern computer and the mathematical study of the uncomputable. Each Turing machine has the same small repertoire of basic (or “atomic”) operations. These are logically simple. The scanner contains What is a Turing Machine? mechanisms that enable it to erase the symbol on the scanned square, to write a symbol on the A Turing machine consists of a limitless memory scanned square (first erasing any existing symbol), and a scanner that moves back and forth through and to shift position one square to the left or the memory, symbol by symbol, reading what right. Complexity of operation is achieved by it finds and writing further symbols. The memory chaining together large numbers of these simple

4 Computation

Table 1.1

State Scanned square Operations Next state a blank P[0], R b b blank R c c blank P[1], R d d blank R a

basic actions. The scanner will halt if instructed Example of a Turing machine to do so, i.e. will cease work, coming to rest on some particular square, for example the square The following simple example is from “On Com- containing the output (or if the output consists putable Numbers” (Turing 1936: 233). The of a string of several digits, then on the square machine – call it M – starts work with a blank containing the left-most digit of the output, say). tape. The tape is endless. The problem is to set In addition to the operations just mentioned, up the machine so that if the scanner is posi- erase, write, shift, and halt, the scanner is able to tioned over any square of the tape and the ma- change state. A device within the scanner is cap- chine set in motion, it will print alternating binary able of adopting a number of different positions. digits on the tape, 0 1 0 1 0 1 . . . , working to This device may be conceptualized as consisting the right from its starting place, leaving a blank of a dial with a finite number of positions, labeled square in between each digit. In order to do its “a,” “b,” “c,” etc. Each of these positions counts work M makes use of four states labeled “a,” as a different state, and changing state amounts “b,” “c,” and “d.” M is in state a when it starts to shifting the dial’s pointer from one labeled work. The operations that M is to perform can position to another. The device functions as a be set out by means of a table with four columns simple memory. As Turing said, by altering its (see table 1.1). “R” abbreviates the instruction state the “machine can effectively remember “shift right one square,” “P[0]” abbreviates some of the symbols which it has ‘seen’ (scanned) “print 0 on the scanned square,” and likewise previously” (1936: 231). For example, a dial with “P[1].” The top line of table 1.1 reads: if you two positions can be used to keep a record of are in state a and the square you are scanning is which binary digit, 0 or 1, is present on the blank, then print 0 on the scanned square, shift square that the scanner has just vacated. If a right one square, and go into state b. A machine square might also be blank, then a dial with acting in accordance with this table of instructions three positions is required. – or program – toils endlessly on, printing the Commercially available computers are hard- desired sequence of digits while leaving alternate wired to perform basic operations considerably squares blank. more sophisticated than those of a Turing Turing did not explain how it is to be brought machine – add, multiply, decrement, store-at- about that the machine acts in accordance with address, branch, and so forth. The precise list of the instructions. There was no need. Turing’s basic operations varies from manufacturer to machines are abstractions and it is not neces- manufacturer. It is a remarkable that none sary to propose any specific for of these computers can out-compute the UTM. causing the machine to follow the instructions. Despite the austere simplicity of Turing’s However, for purposes of visualization, one machines, they are capable of computing any- might imagine the scanner to be accompanied thing that any computer on the market can com- by a bank of switches and plugs resembling an pute. Indeed, because they are abstract machines, old-fashioned telephone switchboard. Arranging they are capable of computations that no “real” the plugs and setting the switches in a certain computer could perform. way causes the machine to act in accordance

5 B. Jack Copeland with the instructions in table 1.1. Other ways the human carrying it out, and (b) produces of setting up the “switchboard” cause the the correct answer in a finite number of steps. machine to act in accordance with other tables (An example of an effective method well-known of instructions. among philosophers is the truth table test for tautologousness.) Many thousands of human computers were employed in business, govern- The universal Turing machine ment, and research establishments, doing some of the sorts of calculating work that nowadays The UTM has a single, fixed table of instructions, is performed by electronic computers. Like which we may imagine to have been set into the filing clerks, computers might have little detailed machine by way of the switchboard-like arrange- knowledge of the end to which their work was ment just mentioned. Operating in accordance directed. with this table of instructions, the UTM is able The term “computing machine” was used to to carry out any task for which a Turing- refer to calculating machines that mechanized machine instruction table can be written. The elements of the human computer’s work. These trick is to place an instruction table for carrying were in effect homunculi, calculating more out the desired task onto the tape of the universal quickly than an unassisted human computer, but machine, the first line of the table occupying the doing nothing that could not in principle be first so many squares of the tape, the second done by a human clerk working effectively. Early line the next so many squares, and so on. The computing machines were somewhat like today’s UTM reads the instructions and carries them nonprogrammable hand-calculators: they were out on its tape. This ingenious idea is funda- not automatic, and each step – each addition, mental to computer science. The universal Turing division, and so on – was initiated manually machine is in the stored-program digital by the human operator. For a complex calcula- computer. tion, several dozen human computers might be Turing’s greatest contributions to the develop- required, each equipped with a desk-top com- ment of the modern computer were: puting machine. By the 1940s, however, the scale of some calculations required by physicists and • The idea of controlling the function of the engineers had become so great that the work computing machine by storing a program of could not easily be done in a reasonable time by (symbolically or numerically encoded) instruc- even a roomful of human computers with desk- tions in the machine’s memory. top computing machines. The need to develop • His proof that, by this means, a single high-speed, large-scale, automatic computing machine of fixed structure is able to carry out machinery was pressing. every computation that can be carried out by In the late 1940s and early 1950s, with the any Turing machine whatsoever. advent of electronic computing machines, the phrase “computing machine” gave way gradu- ally to “computer.” During the brief period in which the old and new meanings of “computer” Human Computation co-existed, the prefix “electronic” or “digital” would usually be used in order to distinguish When Turing wrote “On Computable Num- machine from human. As Turing stated, the new bers,” a computer was not a machine at all, but electronic machines were “intended to carry out a human being – a mathematical assistant who any definite rule of thumb process which could calculated by rote, in accordance with some have been done by a human operator work- “effective method” supplied by an overseer prior ing in a disciplined but unintelligent manner” to the calculation. A paper-and-pencil method is (Turing 1950: 1). Main-frames, laptops, pocket said to be effective, in the mathematical sense, calculators, palm-pilots – all carry out work that if it (a) demands no insight or ingenuity from a human rote-worker could do, if he or she

6 Computation worked long enough, and had a plentiful enough The UTM is able to perform any calculation supply of paper and pencils. that any human computer can carry out. The Turing machine is an idealization of the human computer (Turing 1936: 231). An equivalent way of stating the thesis is: Wittgenstein put this point in a striking way: Any effective – or mechanical – method can Turing’s “Machines.” These machines are be carried out by the UTM. humans who calculate. (Wittgenstein 1980: §1096) (“Mechanical” is a term of art in mathematics and logic. It does not carry its everyday meaning, being in its technical sense simply a synonym It was not, of course, some deficiency of for “effective.”) Notice that the converse of the imagination that led Turing to model his logical thesis – any problem-solving method that can be computing machines on what can be achieved carried out by the UTM is effective – is obvi- by a human being working effectively. The pur- ously true, since a human being can, in principle, pose for which he introduced them demanded work through any Turing-machine program, it. The Turing machine played a key role in his obeying the instructions (“in principle” because demonstration that there are mathematical tasks we have to assume that the human does not go which cannot be carried out by means of an crazy with boredom, or die of old age, or use up effective method. every sheet of paper in the universe). Church independently proposed a different way of replacing talk about effective methods with The Church–Turing Thesis formally precise language (Church 1936a). Tur- ing remarked that his own way of proceeding was “possibly more convincing” (1937: 153); The concept of an effective method is an informal Church acknowledged the point, saying that one. Attempts such as the above to explain what Turing’s concept of computation by Turing counts as an effective method are not rigorous, machine “has the advantage of making the iden- since the requirement that the method demand tification with effectiveness . . . evident immedi- neither insight nor ingenuity is left unexplicated. ately” (Church 1937: 43). One of Turing’s leading achievements – and The name “Church–Turing thesis,” now this was a large first step in the development standard, seems to have been introduced by of the mathematical theory of computation – Kleene, with a flourish of bias in favor of his was to propose a rigorously defined expression mentor Church (Kleene 1967: 232): with which the informal expression “by means of an effective method” might be replaced. The Turing’s and Church’s theses are equivalent. rigorously defined expression, of course, is “by We shall usually refer to them both as Church’s means of a Turing machine.” The importance thesis, or in connection with that one of of Turing’s proposal is this: if the proposal is its . . . versions which deals with “Turing correct, then talk about the existence and non- machines” as the Church–Turing thesis. existence of effective methods can be replaced throughout mathematics and logic by talk about Soon ample evidence amassed for the Church– the existence or non-existence of Turing machine Turing thesis. (A survey is given in chs. 12 and programs. For instance, one can establish that 13 of Kleene 1952.) Before long it was (as Turing there is no effective method at all for doing such- put it) “agreed amongst logicians” that his pro- and-such a thing by proving that no Turing posal gives the “correct accurate rendering” of machine can do the thing in question. talk about effective methods (Turing 1948: 7). Turing’s proposal is encapsulated in the (Nevertheless, there have been occasional dis- Church–Turing thesis, also known simply as senting voices over the years; for example Kalmár Turing’s thesis: 1959 and Péter 1959.)

7 B. Jack Copeland

into which of these two categories it falls. Turing Beyond the Universal Turing showed that this problem cannot be solved by Machine the UTM. The halting problem (Davis 1958) is another Computable and uncomputable example of a problem that cannot be solved by numbers the UTM (although not one explicitly consid- ered by Turing). This is the problem of deter- Turing calls any number that can be written out mining, given any arbitrary Turing machine, by a Turing machine a computable number. That whether or not the machine will eventually halt is, a number is computable, in Turing’s sense, if when started on a blank tape. The machine and only if there is a Turing machine that calcu- shown in table 1.1 is rather obviously one of lates each digit of the number’s decimal representa- those that never halts – but in other cases it is tion, in sequence. π, for example, is a computable definitely not obvious from a machine’s table number. A suitably programmed Turing machine whether or not it halts. And, of course, simply will spend all eternity writing out the decimal watching the machine run (or a simulation of representation of π digit by digit, 3.14159 . . . the machine) is of no help at all, for what can be Straight off, one might expect it to be the concluded if after a week or a year the machine case that every number that has a decimal rep- has not halted? If the machine does eventually resentation (that is to say, every real number) is halt, a watching human – or Turing machine – computable. For what could prevent there being, will sooner or later find this out; but in the case for any particular number, a Turing machine that of a machine that has not yet halted, there is no “churns out” that number’s decimal representa- effective method for deciding whether or not it tion digit by digit? However, Turing proved that is going to halt. not every real number is computable. In fact, computable numbers are relatively scarce among numbers. There are only countably many The halting function computable numbers, because there are only countably many different Turing-machine pro- A function is a mapping from “arguments” (or grams (instruction tables). (A collection of things inputs) to “values” (or outputs). For example, is countable if and only if either the collection addition (+) is a function that maps pairs of num- is finite or its members can be put into a one- bers to single numbers: the value of the function to-one correspondence with the integers, 1, + for the pair of arguments 5, 7 is the number 2, 3, . . . .) As Georg Cantor proved in 1874, 12. The squaring function maps single numbers there are uncountably many real numbers – in to single numbers: e.g. the value of n2 for the other words, there are more real numbers than argument 3 is 9. integers. There are literally not enough Turing- A function is said to be computable by Turing machine programs to go around in order for machine if some Turing machine will take in every real number to be computable. arguments of the function (or pairs of arguments, etc.) and, after carrying out some finite number of basic operations, produce the corresponding The printing problem and the value – and, moreover, will do this no matter halting problem which argument of the function is presented. For example, addition over the integers is comput- Turing described a number of mathematical able by Turing machine, since a Turing machine problems that cannot be solved by Turing can be set up so that whenever two integers are machine. One is the printing problem. Some pro- inscribed on its tape (in binary notation, say), grams print “0” at some stage in their computa- the machine will output their sum. tions; all the remaining programs never print The halting function is as follows. Assume the “0.” The printing problem is the problem of Turing machines to be ordered in some way, so deciding, given any arbitrarily selected program, that we may speak of the first machine in the

8 Computation ordering, the second, and so on. (There are vari- in his Paris lecture: “in mathematics there is no ous standard ways of accomplishing this order- ignorabimus” (there is no we shall not know) ing, e.g. in terms of the number of symbols in (Hilbert 1902: 445). each machine’s instruction table.) The arguments It is important that the system expressing the of the halting function are simply 1, 2, 3, . . . . “whole thought content of mathematics” be (Like the squaring function, the halting func- consistent. An inconsistent system – a system tion takes single arguments.) The value of the containing contradictions – is worthless, since halting function for any argument n is 1 if the any statement whatsoever, true or false, can be nth Turing machine in the ordering eventually derived from a contradiction by simple logical halts when started on a blank tape, and is 0 if steps. So in an inconsistent system, absurdities the nth machine runs on forever (as would, for such as 0 = 1 and 6 ≠ 6 are provable. An incon- example, a Turing machine programmed to pro- sistent system would indeed contain all true duce in succession the digits of the decimal rep- mathematical statements – would be complete, resentation of π). in other words – but would in addition also The theorem that the UTM cannot solve the contain all false mathematical statements. halting problem is often expressed in terms of If ignorance is to be banished absolutely, the the halting function. system must be decidable. An undecidable sys- tem might on occasion leave us in ignorance. Halting theorem: The halting function is Only if the mathematical system were decidable not computable by Turing machine. could we be confident of always being able to tell whether or not any given statement is prov- able. Unfortunately for the Hilbert program, The Entscheidungsproblem however, it became clear that most interesting mathematical systems are, if consistent, incom- The Entscheidungsproblem, or decision problem, plete and undecidable. was Turing’s principal quarry in “On Computable In 1931 Gödel showed that Hilbert’s ideal is Numbers.” The decision problem was brought impossible to satisfy, even in the case of simple to the fore of mathematics by the German math- arithmetic. He proved that the system called ematician David Hilbert (who in a lecture given Peano arithmetic is, if consistent, incomplete. in Paris in 1900 set the agenda for much of This is known as Gödel’s first incompleteness twentieth-century mathematics). Hilbert and his theorem. (Gödel later generalized this result, followers held that mathematicians should seek pointing out that “due to A. M. Turing’s work, to express mathematics in the form of a com- a precise and unquestionably adequate defini- plete, consistent, decidable formal system – a tion of the general concept of formal system can system expressing “the entire thought-content now be given,” with the consequence that incom- of mathematics in a uniform way” (Hilbert 1927: pleteness can “be proved rigorously for every 475). The project of formulating mathematics in consistent formal system containing a certain this way became known as the “Hilbert program.” amount of finitary number theory” (Gödel 1965: A consistent system is one that contains no 71).) Gödel had shown that no matter how contradictions; a complete system one in which hard mathematicians might try to the every true mathematical statement is provable. all-encompassing formal system envisaged by “Decidable” means that there is an effective Hilbert, the product of their labors would, if method for telling, of each mathematical state- consistent, inevitably be incomplete. As Hermann ment, whether or not the statement is provable Weyl – one of Hilbert’s greatest pupils – in the system. A complete, consistent, decidable observed, this was nothing less than “a catastro- system would banish ignorance from math- phe” for the Hilbert program (Weyl 1944: 644). ematics. Given any mathematical statement, one Gödel’s theorem does not mention decidabil- would be able to tell whether the statement is ity. This aspect was addressed by Turing and by true or false by deciding whether or not it is Church. Each showed, working independently, provable in the system. Hilbert famously declared that no consistent formal system of arithmetic is

9 B. Jack Copeland decidable. They showed this by proving that not mechanism, and established a fundamental result even the weaker, purely logical system presup- to the effect that the UTM can simulate the posed by any formal system of arithmetic and behavior of any machine. The myth has passed called the first-order predicate calculus is decid- into the philosophy of mind, theoretical psycho- able. Turing’s way of proving that the first- logy, cognitive science, Artificial Intelligence, order predicate calculus is undecidable involved and Artificial Life, generally to pernicious effect. the printing problem. He showed that if a Tur- For example, the Oxford Companion to the Mind ing machine could tell, of any given statement, states: “Turing showed that his very simple whether or not the statement is provable in the machine . . . can specify the steps required for first-order predicate calculus, then a Turing the solution of any problem that can be solved machine could tell, of any given Turing machine, by instructions, explicitly stated rules, or proced- whether or not it ever prints “0.” Since, as he ures” (Gregory 1987: 784). Dennett maintains had already established, no Turing machine can that “Turing had proven – and this is probably do the latter, it follows that no Turing machine his greatest contribution – that his Universal can do the former. The final step of the argu- Turing machine can compute any function that ment is to apply Turing’s thesis: if no Turing any computer, with any architecture, can com- machine can perform the task in question, then pute” (1991: 215); also that every “task for there is no effective method for performing it. which there is a clear recipe composed of simple The Hilbertian dream lay in total ruin. steps can be performed by a very simple com- Poor news though Turing’s and Church’s puter, a universal Turing machine, the universal result was for the Hilbert school, it was wel- recipe-follower” (1978: xviii). Paul and Patricia come news in other quarters, for a reason that Churchland assert that Turing’s “results entail Hilbert’s illustrious pupil von Neumann had something remarkable, namely that a standard given in 1927 (von Neumann 1927: 12): digital computer, given only the right program, a large enough memory and sufficient time, can If undecidability were to fail then mathematics, compute any rule-governed input–output func- in today’s sense, would cease to exist; its place tion. That is, it can display any systematic pat- would be taken by a completely mechanical tern of responses to the environment whatsoever” rule, with the aid of which any man would be (1990: 26). Even Turing’s biographer, Hodges, able to decide, of any given statement, whether has endorsed the myth: the statement can be proven or not. Alan had . . . discovered something almost . . . In a similar vein, the Cambridge mathematician miraculous, the idea of a universal machine G. H. Hardy said in a lecture in 1928 (Hardy that could take over the work of any machine. 1929: 16): (Hodges 1992: 109)

if there were ...a mechanical set of rules for Turing did not show that his machines can the solution of all mathematical problems . . . solve any problem that can be solved “by instruc- our activities as mathematicians would come to an end. tions, explicitly stated rules, or procedures,” and nor did he prove that the UTM “can compute The next section is based on Copeland 1996. any function that any computer, with any archi- tecture, can compute” or perform any “task for which there is a clear recipe composed of simple steps.” As previously explained, what he proved Misunderstandings of the is that the UTM can carry out any task that any Church–Turing Thesis: Turing machine can carry out. Each of the claims The Limits of Machines just quoted says considerably more than this. If what the Churchlands assert were true, then A myth has arisen concerning Turing’s work, the view that psychology must be capable of namely that he gave a treatment of the limits of being expressed in standard computational terms

10 Computation would be secure (as would a number of other thesis properly so called” for the proposition that controversial claims). But Turing had no result Turing and Church themselves endorsed. entailing that “a standard digital computer ...can compute any rule-governed input–output func- [C]onnectionist models . . . may possibly even tion.” What he did have was a result entailing challenge the strong construal of Church’s the exact opposite. The theorem that no Turing Thesis as the claim that the class of well- machine can decide the predicate calculus entails defined computations is exhausted by those of that there are rule-governed input–output func- Turing machines. (Smolensky 1988: 3) tions that no Turing machine is able to compute Church–Turing thesis: If there is a well defined – for example, the function whose output is 1 procedure for manipulating symbols, then a whenever the input is a statement that is prov- Turing machine can be designed to do the able in the predicate calculus, and is 0 for all procedure. (Henry 1993: 149) other inputs. There are certainly possible pat- terns of responses to the environment, perfectly [I]t is difficult to see how any language that systematic patterns, that no Turing machine could actually be run on a physical computer can display. One is the pattern of responses just could do more than Fortran can do. The described. The halting function is a mathemat- idea that there is no such language is called ical characterization of another such pattern. Church’s thesis. (Geroch & Hartle 1986: 539)

The first aspect that we examine of Church’s Distant cousins of the Thesis...[w]e can formulate, more precisely: Church–Turing thesis The behaviour of any discrete physical system evolving according to local mechanical laws is As has already been emphasized, the Church– recursive. (Odifreddi 1989: 107) Turing thesis concerns the extent of effective I can now state the physical version of the methods. Putting this another way (and ignoring Church–Turing principle: “Every finitely real- contingencies such as boredom, death, or insuf- izable physical system can be perfectly simu- ficiency of paper), the thesis concerns what a lated by a universal model computing machine human being can achieve when working by rote operating by finite means.” This formulation with paper and pencil. The thesis carries no im- is both better defined and more physical than plication concerning the extent of what machines Turing’s own way of expressing it. (Deutsch are capable of achieving (even digital machines 1985: 99) acting in accordance with “explicitly stated rules”). For among a machine’s repertoire of basic That there exists a most general formulation operations, there may be those that no human of machine and that it leads to a unique set of working by rote with paper and pencil can input–output functions has come to be called perform. Church’s thesis. (Newell 1980: 150) Essentially, then, the Church–Turing thesis says that no human computer, or machine that mimics a human computer, can out-compute the The maximality thesis UTM. However, a variety of other propositions, very different from this, are from time to time It is important to distinguish between the called the Church–Turing thesis (or Church’s Church–Turing thesis properly so called and what thesis), sometimes but not always with accom- I call the “maximality thesis” (Copeland 2000). panying hedges such as “strong form” and (Among the few writers to distinguish explicitly “physical version.” Some examples from the re- between Turing’s thesis and stronger proposi- cent literature are given below. This loosening tions along the lines of the maximality thesis are of established terminology is unfortunate, and Gandy 1980 and Sieg 1994.) can easily lead to misunderstandings. In what A machine m is said to be able to generate a follows I use the expression “Church–Turing certain function if m can be set up so that if m is

11 B. Jack Copeland presented with any of the function’s arguments, Understood correctly, this remark attributes to m will carry out some finite number of atomic Turing not a thesis concerning the limits of what processing steps at the end of which m produces can be achieved by machine but the Church– the corresponding value of the function (mutatis Turing thesis properly so called. mutandis in the case of functions that, like addi- The technical usage of “mechanical” tends tion, demand more than one argument). to obscure the possibility that there may be machines, or biological organs, that generate (or Maximality Thesis: All functions that can compute, in a broad sense) functions that cannot be generated by machines (working on finite be computed by Turing machine. For the ques- input in accordance with a finite program tion “Can a machine execute a procedure that of instructions) are computable by Turing is not mechanical?” may appear self-answering, machine. yet this is precisely what is asked if thesis M is questioned. In the technical literature, the word “comput- The maximality thesis (“thesis M”) admits of able” is often tied by definition to effectiveness: two interpretations, according to whether the a function is said to be computable if and only if phrase “can be generated by machine” is taken there is an effective method for determining its in the this-worldly sense of “can be generated values. The Church–Turing thesis then becomes: by a machine that conforms to the physical laws (if not to the resource constraints) of the actual Every computable function can be computed world,” or in a sense that abstracts from whether by Turing machine. or not the envisaged machine could exist in the actual world. Under the latter interpretation, Corollaries such as the following are sometimes thesis M is false. It is straightforward to describe stated: abstract machines that generate functions that cannot be generated by the UTM (see e.g. [C]ertain functions are uncomputable in an Abramson 1971, Copeland 2000, Copeland & absolute sense: uncomputable even by [Turing Proudfoot 2000, Stewart 1991). Such machines machine], and, therefore, uncomputable by any are termed “hypercomputers” in Copeland and past, present, or future real machine. (Boolos Proudfoot (1999a). & Jeffrey 1980: 55) It is an open empirical question whether or When understood in the sense in which it is not the this-worldly version of thesis M is true. intended, this remark is perfectly true. However, Speculation that there may be physical processes to a casual reader of the technical literature, such – and so, potentially, machine-operations – whose statements may appear to say more than they in behavior conforms to functions not computable fact do. by Turing machine stretches back over at least Of course, the decision to tie the term “com- five decades. (Copeland & Sylvan 1999 is a sur- putable” and its cognates to the concept of effect- vey; see also Copeland & Proudfoot 1999b.) iveness does not settle the truth-value of thesis M. A source of potential misunderstanding about Those who abide by this terminological decision the limits of machines lies in the difference will not describe a machine that falsifies thesis M between the technical and everyday meanings of as computing the function that it generates. the word “mechanical.” As previously remarked, Putnam is one of the few writers on the in technical contexts “mechanical” and “effect- philosophy of mind to question the proposition ive” are often used interchangeably. (Gandy 1988 that Turing machines provide a maximally gen- outlines the history of this usage of the word eral formulation of the notion of machine: “mechanical.”) For example: [M]aterialists are committed to the view that Turing proposed that a certain class of abstract a human being is – at least metaphorically – a machines could perform any “mechanical” machine. It is understandable that the notion computing procedure. (Mendelson 1964: 229) of a Turing machine might be seen as just a

12 Computation

way of making this materialist idea precise. entail that the brain, and indeed any biological Understandable, but hardly well thought out. or physical system whatever, can be simulated by The problem is the following: a “machine” in a Turing machine. For example, the entry on the sense of a physical system obeying the laws Turing in A Companion to the Philosophy of Mind of Newtonian physics need not be a Turing contains the following claims: “we can depend machine. (Putnam 1992: 4) on there being a Turing machine that captures the functional relations of the brain,” for so long as “these relations between input and output The Church–Turing fallacy are functionally well-behaved enough to be de- scribable by . . . mathematical relationships . . . we To commit what I call the Church–Turing fallacy know that some specific version of a Turing (Copeland 2000, 1998) is to believe that the machine will be able to mimic them” (Guttenplan Church–Turing thesis, or some formal or semi- 1994: 595). Even Dreyfus, in the course of criti- formal result established by Turing or Church, cizing the view that “man is a Turing machine,” secures the following proposition: succumbs to the belief that it is a “fundamental truth that every form of ‘information process- If the mind–brain is a machine, then the ing’ (even those which in practice can only be Turing-machine computable functions pro- carried out on an ‘analogue computer’) must in vide sufficient mathematical resources for a principle be simulable on a [Turing machine]” full account of human . (1992: 195). Searle writes in a similar fashion: Perhaps some who commit this fallacy are misled purely by the terminological practice already men- If the question [“Is comput- tioned, whereby a thesis concerning which there able?”] asks “Is there some level of description is little real doubt, the Church–Turing thesis pro- at which conscious processes and their cor- perly so called, and a nexus of different theses, related brain processes can be simulated [by a some of unknown truth-value, are all referred to Turing machine]?” the answer is trivially yes. as Church’s thesis or the Church–Turing thesis. Anything that can be described as a precise The Church–Turing fallacy has led to some series of steps can be simulated [by a Turing remarkable claims in the foundations of psycho- machine]. (Searle 1997: 87) logy. For example, one frequently encounters the Can the operations of the brain be simulated view that psychology must be capable of being on a digital computer? . . . The answer seems expressed ultimately in terms of the Turing ma- to me . . . demonstrably “Yes” . . . That is, chine (e.g. Fodor 1981: 130; Boden 1988: 259). naturally interpreted, the question means: Is To anyone in the grip of the Church–Turing there some description of the brain such that fallacy, conceptual space will seem to contain no under that description you could do a com- room for mechanical models of the mind–brain putational simulation of the operations of the that are not equivalent to a Turing machine. Yet brain. But given Church’s thesis that anything it is certainly possible that psychology will find that can be given a precise enough character- the need to employ models of human cognition ization as a set of steps can be simulated on a that transcend Turing machines (see Chapter 10, digital computer, it follows trivially that the COMPUTATIONALISM, CONNECTIONISM, AND THE question has an affirmative answer. (Searle PHILOSOPHY OF MIND). 1992: 200)

Church’s thesis properly so called does not say The simulation fallacy that anything that can be described as a precise series of of steps can be simulated by Turing A closely related error, unfortunately also com- machine. mon in modern writing on computation and the Similarly, Johnson-Laird and the Churchlands brain, is to hold that Turing’s results somehow argue:

13 B. Jack Copeland

If you assume that [consciousness] is scient- As with thesis M, thesis S is trivially false if it ifically explicable...[and] [g]ranted that the is taken to concern all conceivable processes, and [Church–Turing] thesis is correct, then the its truth-value is unknown if it is taken to con- final dichotomy rests on Craik’s functionalism. cern only processes that conform to the physics If you believe [functionalism] to be false . . . of the real world. For all we presently know, a then presumably you hold that consciousness completed neuroscience may present the mind– could be modelled in a computer program in brain as a machine that – when abstracted out the same way that, say, the weather can be from sources of inessential boundedness, such as modelled . . . If you accept functionalism, how- mortality – generates functions that no Turing ever, then you should believe that conscious- machine can generate. ness is a computational process. (Johnson-Laird 1987: 252)

Church’s Thesis says that whatever is com- The equivalence fallacy putable is Turing computable. Assuming, with some safety, that what the mind-brain Paramount among the evidence for the Church– does is computable, then it can in principle be Turing thesis properly so called is the fact that simulated by a computer. (Churchland & all attempts to give an exact analysis of the intuit- Churchland 1983: 6) ive notion of an effective method have turned out to be equivalent, in the sense that each ana- As previously mentioned, the Churchlands lysis has been proved to pick out the same class believe, incorrectly, that Turing’s “results entail of functions, namely those that are computable . . . that a standard digital computer, given only by Turing machine. (For example, there have the right program, a large enough memory and been analyses in terms of lambda-definability, sufficient time, can . . . display any systematic recursivenes, register machines, Post’s canonical pattern of responses to the environment whatso- and normal systems, combinatory definability, ever” (1990: 26). This no doubt explains why Markov algorithms, and Gödel’s notion of they think they can assume “with some safety” reckonability.) Because of the diversity of these that what the mind–brain does is computable, various analyses, their equivalence is generally for on their understanding of matters, this is to considered very strong evidence for the Church– assume only that the mind–brain is character- Turing thesis (although for a skeptical point of ized by a “rule-governed” (1990: 26) input– view see Kreisel 1965: 144). output function. However, the equivalence of these diverse The Church–Turing thesis properly so called analyses is sometimes taken to be evidence also does not entail that the brain (or the mind, or for stronger theses like M and S. This is nothing consciousness) can be simulated by a Turing more than a confusion – the equivalence fallacy machine, not even in conjunction with the belief (Copeland 2000). The analyses under discussion that the brain (or mind, etc.) is scientifically are of the notion of an effective method, not of explicable, or exhibits a systematic pattern of the notion of a machine-generable function; the responses to the environment, or is “rule- equivalence of the analyses bears only on the governed” (etc.). Each of the authors quoted issue of the extent of the former notion and seems to be assuming the truth of a close relat- indicates nothing concerning the extent of the ive of thesis M, which I call “thesis S” (Copeland latter. 2000).

Thesis S: Any process that can be given Artificial intelligence and the a mathematical description (or a “precise equivalence fallacy enough characterization as a set of steps,” or that is scientifically describable or scient- Newell, discussing the possibility of artificial ifically explicable) can be simulated by a intelligence, argues that (what he calls) a “phys- Turing machine. ical symbol system” can be organized to exhibit

14 Computation general intelligence. A “physical symbol system” Newell’s a priori argument for the claim that a is a universal Turing machine, or any equivalent physical symbol system can become generally system, situated in the physical – as opposed intelligent founders in confusion. to the conceptual – world. (The tape of the machine is accordingly finite; Newell specifies that the storage capacity of the tape [or equivalent] Conclusion be unlimited in the practical sense of finite yet not small enough to “ concern.”) Since there are problems that cannot be solved A [physical symbol] system always contains by Turing machine, there are – given the the potential for being any other system if so Church–Turing thesis – limits to what can be instructed. Thus, a [physical symbol] system accomplished by any form of machine that works can become a generally intelligent system. in accordance with effective methods. However, (Newell 1980: 170) not all possible machines share those limits. It is an open empirical question whether there are Is the premise of this pro-AI argument true? actual deterministic physical processes that, in the A physical symbol system, being a universal long run, elude simulation by Turing machine; Turing machine situated in the real world, can, and, if so, whether any such processes could use- if suitably instructed, simulate (or, metaphoric- fully be harnessed in some form of calculating ally, become) any other physical symbol system machine. It is, furthermore, an open empirical (modulo some fine print concerning storage question whether any such processes are involved capacity). If this is what the premise means, then in the working of the human brain. it is true. However, if taken literally, the premise is false, since as previously remarked, systems can be specified which no Turing machine – and References so no physical symbol system – can simulate. However, if the premise is interpreted in the Abramson, F. G. 1971. “Effective computation over former manner, so that it is true, the conclusion the real numbers.” Twelfth Annual Symposium fails to follow from the premise. Only to one who on Switching and Automata Theory. Northridge, believes, as Newell does, that “the notion of CA: Institute of Electrical and Electronics machine or determinate physical mechanism” is Engineers. “formalized” by the notion of a Turing machine Boden, M. A. 1988. Computer Models of Mind. (ibid.) will the argument appear deductively valid. Cambridge: Cambridge University Press. Newell’s defense of his view that the uni- Boolos, G. S. and Jeffrey, R. C. 1980. Computabil- versal Turing machine exhausts the possibilities ity and Logic, 2nd ed. Cambridge: Cambridge of mechanism involves an example of the equi- University Press. valence fallacy: Church, A. 1936a. “An unsolvable problem of elementary number theory.” American Journal [An] important chapter in the theory of com- of Mathematics 58: 345–63. puting...has shown that all attempts to . . . ——. 1936b. “A note on the Entscheidungs- formulate...general notions of mechanism problem.” Journal of Symbolic Logic 1: 40–1. . . . lead to classes of machines that are equival- ——. 1937. Review of Turing 1936. Journal of ent in that they encompass in toto exactly the Symbolic Logic 2: 42–3. same set of input–output functions. In effect, Churchland, P. M. and Churchland, P. S. 1983. there is a single large frog pond of functions “Stalking the wild epistemic engine.” Nous 17: no matter what species of frogs (types of 5–18. machines) is used. ...A large zoo of different —— and ——. 1990. “Could a machine think?” formulations of maximal classes of machines is Scientific American 262 (Jan.): 26–31. known by now – Turing machines, recursive Copeland, B. J. 1996. “The Church–Turing Thesis.” functions, Post canonical systems, Markov In E. Zalta, ed., The Stanford Encyclopaedia of algorithms . . . (Newell 1980: 150) Philosophy, .

15 B. Jack Copeland

——. 1998. “Turing’s O-machines, Penrose, Searle, Guttenplan, S. 1994. A Companion to the Philo- and the Brain.” Analysis 58: 128–38. sophy of Mind. Oxford: Blackwell. ——. 2000. “Narrow versus wide mechanism, Hardy, G. H. 1929. “Mathematical proof.” Mind including a re-examination of Turing’s views on 38: 1–25. the mind–machine issue.” Journal of Philosophy Henry, G. C. 1993. The Mechanism and Freedom of 97: 5–32. Repr. in M. Scheutz, ed., Computa- Logic. Lanham, MD: University Press of America. tionalism: New Directions. Cambridge, MA: MIT Hilbert, D. 1902. “Mathematical problems: lecture Press, 2002. delivered before the International Congress of ——. 2001. “Colossus and the dawning of the com- Mathematicians at Paris in 1900.” Bulletin of the puter age.” In M. Smith and R. Erskine, eds., American Mathematical Society 8: 437–79. Action This Day. London: Bantam. ——. 1927. “Die Grundlagen der Mathematik” ——. and Proudfoot, D. 1999a. “Alan Turing’s [The Foundations of Mathematics]. English trans. forgotten ideas in computer science.” Scientific in J. van Heijenoort, ed., From Frege to Gödel: American 280 (April): 76–81. A Source Book in Mathematical Logic, 1879– —— and ——. 1999b. “The legacy of Alan 1931. Cambridge, MA: Harvard University Press, Turing.” Mind 108: 187–95. 1967. —— and ——. 2000. “What Turing did after he Hodges, A. 1992. Alan Turing: The Enigma. Lon- invented the universal Turing machine.” Journal don: Vintage. of Logic, Language, and Information 9: 491–509. Johnson-Laird, P. 1987. “How could consciousness —— and Sylvan, R. 1999. “Beyond the universal arise from the computations of the brain?” In Turing machine.” Australasian Journal of Philo- C. Blakemore and S. Greenfield, eds., Mindwaves. sophy 77: 46–66. Oxford: Blackwell. Davis, M. 1958. Computability and Unsolvability. Kalmár, L. 1959. “An argument against the New York: McGraw-Hill. plausibility of Church’s thesis.” In A. Heyting, Dennett, D. C. 1978. Brainstorms: Philosophical ed., Constructivity in Mathematics. Amsterdam: Essays on Mind and Psychology. Brighton: Harvester. North-Holland. ——. 1991. Consciousness Explained. Boston: Kleene, S. C. 1952. Introduction to Meta- Little, Brown. mathematics. Amsterdam: North-Holland. Deutsch, D. 1985. “Quantum theory, the Church– ——. 1967. Mathematical Logic. New York: Wiley. Turing principle and the universal quantum com- Kreisel, G. 1965. “Mathematical logic.” In T. L. puter.” Proceedings of the Royal Society, Series A, Saaty, ed., Lectures on Modern Mathematics, 400: 97–117. vol. 3. New York: Wiley. Dreyfus, H. L. 1992. What Computers Still Can’t Langton, C. R. 1989. “Artificial life.” In Langton, Do: A Critique of Artificial Reason. Cambridge, ed., Artificial Life. Redwood City: Addison- MA: MIT Press. Wesley. Fodor, J. A. 1981. “The mind–body problem.” Mendelson, E. 1964. Introduction to Mathematical Scientific American 244 (Jan.): 124–32. Logic. New York: Van Nostrand. Gandy, R. 1980. “Church’s thesis and principles Newell, A. 1980. “Physical symbol systems.” for mechanisms.” In J. Barwise, H. Keisler, and Cognitive Science 4: 135–83. K. Kunen, eds., The Kleene Symposium. Amster- Odifreddi, P. 1989. Classical Recursion Theory. dam: North-Holland. Amsterdam: North-Holland. ——. 1988. “The confluence of ideas in 1936.” In Péter, R. 1959. “Rekursivität und Konstruktivität.” R. Herken, ed., The Universal Turing Machine: In A. Heyting, ed., Constructivity in Mathematics. A Half-century Survey. Oxford: Oxford Univer- Amsterdam: North-Holland. sity Press. Putnam, H. 1992. Renewing Philosophy. Cambridge, Geroch, R. and Hartle, J. B. 1986. “Computability MA: Harvard University Press. and physical theories.” Foundations of Physics 16: Searle, J. 1992. The Rediscovery of the Mind. Cam- 533–50. bridge, MA: MIT Press. Gödel, K. 1965. “Postscriptum.” In M. Davis, ed., ——. 1997. The Mystery of Consciousness. New York: The Undecidable. New York: Raven, pp. 71–3. New York Review of Books. Gregory, R. L. 1987. The Oxford Companion to the Sieg, W. 1994. “Mechanical procedures and Mind. Oxford: Oxford University Press. mathematical experience.” In A. George, ed.,

16 Computation

Mathematics and Mind. Oxford: Oxford Univer- Edinburgh: Edinburgh University Press, 1969. sity Press. [A digital facsimile is available in The Turing Sipser, M. 1997. Introduction to the Theory of Com- Archive for the History of Computing, .] Smolensky, P. 1988. “On the proper treatment of ——. 1950. “Programmers’ handbook for connectionism.” Behavioral and Brain Sciences Manchester electronic computer.” University of 11: 1–23. Manchester Computing Laboratory. [A digital Stewart, I. 1991. “Deciding the undecidable.” facsimile is available in The Turing Archive Nature 352: 664–5. for the History of Computing, .] with an application to the Entscheidungs- von Neumann, J. 1927. “Zur Hilbertschen problem.” Proceedings of the London Mathematical Beweistheorie” [On Hilbert’s proof theory], Society, series 2, 42 (1936–7): 230–65. Mathematische Zeitschrift 26: 1–46. ——. 1937. “Computability and λ-definability.” Weyl, H. 1944. “David Hilbert and his Mathemat- Journal of Symbolic Logic 2: 156–63. ical Work,” Bulletin of the American Mathematical ——. 1948. “Intelligent machinery.” National Society 50: 612–54. Physical Laboratory Report. In B. Meltzer Wittgenstein, L. 1980. Remarks on the Philosophy of and D. Michie, eds., Machine Intelligence 5. Psychology, vol. 1. Oxford: Blackwell.

17 Alasdair Urquhart

Chapter 2 Complexity Alasdair Urquhart

other subjects. Much of this research is centered 1 Introduction in the Santa Fe Institute in New Mexico, where work on “complex systems” of various kinds The theory of computational complexity is con- is done. The confusion between the two fields cerned with estimating the resources a computer arises from the fact that the word “complexity” needs to solve a problem. The basic resources is often used in different ways. A system or object are time (number of steps in a computation) could reasonably be described as “complex” and space (amount of memory used). There are under various conditions: if it consists of many problems in computer science, logic, algebra, and interacting parts; if it is disordered or exhibits calculus that are solvable in principle by com- high entropy; if it exhibits diversity based on puters, but, in the worst case, require completely hierarchical structure; if it exhibits detail on many infeasible amounts of space or time, so that in different scales, like fractal sets. Some of these practical terms they are insoluble. The goal of meanings of “complexity” are connected with complexity theory is to classify problems accord- the theory of computational complexity, but ing to their complexity, particularly problems that some are only tangentially related. In the present are important in applications such as cryptology, chapter, we confine ourselves to the simple quant- linear programming, and combinatorial optimiza- itative measures of time and space complexity of tion. A major result of the theory is that prob- computations. lems fall into strict hierarchies when categorized A widely accepted working hypothesis in the in accordance with their space and time require- theoretical computer science community is that ments. The theory has been less successful in practically feasible algorithms can be identified relating the two basic measures; there are major with those whose running time can be bounded open questions about problems that are solvable by a polynomial in the size of the input. For using only small space, but for which the best example, an algorithm that runs in time 10n for algorithms known use exponential time. inputs with n symbols would be very efficient; The theory discussed in this chapter should this would be described as an algorithm running be distinguished from another area often called in linear time. A quadratic time algorithm runs “complexity theory,” a loosely defined interdis- in time cn2 for some constant c; obviously such ciplinary stream of research that includes work an algorithm is considerably less efficient than a on complex dynamical systems, chaos theory, linear time algorithm, but could be quite prac- artificial life, self-organized criticality, and many tical for inputs of reasonable size. On the other

18 Complexity hand, a computer procedure requiring time ex- are colored alike (see below for more details ponential in the size of the input very rapidly on this problem). The only known general leads to infeasible running times. algorithms for such problems require exponenti- To illustrate the point of the previous para- ally long run-times in the worst case, and it is graph, consider a modern fast computer. The widely conjectured that no polynomial time algo- speed of such machines is often measured in the rithms exist for them. This conjecture is usually number of numerical operations performed per phrased as the inequality “P ≠ NP,” the central second; a commonly used standard is the num- open problem in theoretical computer science, ber of floating-point operations per second. Sup- and perhaps the most important open problem pose we have a machine that performs a million in mathematical logic. floating-point operations per second, slow by In this chapter, we begin by giving an outline current supercomputer standards. Then an algo- of the basic definitions and results of complexity rithm that requires n2 such operations for an theory, including the existence of space and time input of size n would take only a quarter of a hierarchies, then explain the basics of the theories second for an input of size 500. Even if the of NP-completeness and parallel computation. running time is bounded by n3, an input of size The chapter concludes with some brief reflec- 500 would require at most 2 minutes 5 seconds. tions on the relevance of complexity theory to On the other hand, an algorithm running in time questions in the philosophy of computing. 2n could in the worst case take over 35 years for an input of size 50. The reader can easily verify with the help of a pocket calculator that 2 Time and Space in this dramatic difference between polynomial and Computation exponential growth is robust, in the sense that a thousand-fold increase in computer speed only adds 10 to the size of the largest problem instance The theory of complexity analyzes the computa- we can solve in an hour with an exponential (2n) tional resources necessary to solve a problem. time algorithm, whereas with a quadratic (n2) The most important of these resources are time time algorithm, the largest such problem increases (number of steps in a computation) and space by a factor of over 30. (storage capacity of the computer). This chapter is The theory of computational complexity has mainly concerned with the complexity of decision provided rigorous proofs of the existence of com- problems having infinitely many instances. There putational problems for which such exponential is another approach to complexity applicable behavior is unavoidable. This means that for to individual objects, in which the complexity such problems, there are infinitely many “diffi- of an object is measured by the size of the short- cult” instances, for which any algorithm solving est program that produces a description of the problem must take an exponentially long it. This is the Kolmogorov complexity of the time. An especially interesting and important class object; Li and Vitányi (1997) give a readable of problems is the category of NP-complete and detailed introduction to this subject in their problems, of which the satisfiability problem of textbook. propositional logic is the best-known case. These The model for computation chosen here is problems all take the form of asking for a solu- the Turing machine, as defined in the preceding tion of a certain set of constraints (formulas of chapter. The time for a computation is the num- propositional logic, in the case of the satisfiabil- ber of steps taken before the machine halts; the ity problem), where a proposed solution can be space is the number of cells of the tape visited quickly checked to see if it is indeed a solu- by the reading head during the computation. tion, but in general there are exponentially many Several other models of sequential computa- candidate solutions. As an example of such a tion have been proposed. The time and space problem, consider the problem of coloring a large complexity of a problem clearly depend on the and complicated map with only three colors so machine model adopted. However, the basic that no two countries with a common border concepts of complexity theory defined here are

19 Alasdair Urquhart robust in the sense that they are the same for first against the last symbol, deleting these two any reasonable model of sequential computation. symbols, and repeating this step until either the Let Σ be a finite alphabet, and Σ* the set of empty string (with no symbols at all) or a string all finite strings in this alphabet. A subset of Σ* consisting of exactly one symbol is reached. (This is said to be a problem (often called a “lan- is an instructive exercise in Turing machine pro- guage”), and a string in Σ* an instance of the gramming.) In fact, it is not possible to do much problem. The size |s | of an instance s is its length, better than this simple algorithm. Any algorithm i.e. the number of occurrences of symbols in it. for a single-tape Turing machine requires cn2 A function f defined on Σ* and having strings as steps to solve PAL for some c > 0; for an elegant its values is computed by a Turing machine M if proof of this fact using the “incompressibility for any string s in Σ*, if M is started with s on its method” see Li and Vitányi (1997: ch. 6). tape, then it halts with f(s) on its tape. A prob- Other natural examples of computational lem L is solvable (decidable) if there is a Turing problems arise in the area of games. For example, machine that computes the characteristic func- given a chess position, consider the problem: “Is tion of L (the function f such that f(s) = 1 if s this a winning position for White?”; that is to is in L and f(s) = 0 otherwise). For example, say, does White have a plan that check- the satisfiability problem of determining whether mate no matter how Black plays? In this case, a formula of propositional logic is satisfiable there is a simple but crude algorithm to answer or not is solvable by the familiar method of any such question – simply compile a database truth-tables. of all possible board positions, then classify them Solvable problems can be classified according as winning, losing, or drawing for White by con- to the time and space required for their solu- sidering all possible continuations. Such databases tion. If f is a computable function, then we say have been compiled for the case of endgames that f is computable in time T(n) if there is a with only a few pieces (for example, queen versus Turing machine computing f that for any input s rook endgames). Can we do better than this halts with output f(s) after O(T(|s |)) steps (that brute-force approach? There are reasons to think is, there is a constant c such that M halts in at not. The results of Fraenkel and Lichtenstein most c · T(|s |) steps). Similarly, f is computable described below show that computing a perfect in space T(n) if there is a machine M computing strategy for a generalization of chess on an n by f so that for any input s, M halts after visiting n board requires time exponential in n. O(T(|s |)) squares on its tape. A problem L is One of the most significant complexity classes solvable in time T(n) if the characteristic func- is the class P of problems solvable in polynomial tion of L is computable in time T(n); L is solv- time. A function f is polynomial-time comput- able in space S(n) if the characteristic function able if there exists a polynomial p for which f is of L is computable in space S(n). For example, computable in time p(n). A problem is solvable the truth-table method shows that the satisfiabil- in polynomial time if its characteristic function ity problem can be solved in time 2n and space n is polynomial-time computable. The importance (we need only enough tape space to evaluate the of the class rests on the widely accepted working truth-table one row at a time). hypothesis that the class of practically feasible As an illustration of these rather abstract defini- algorithms can be identified with those algorithms tions, let us consider a concrete problem. Sup- that operate in polynomial time. Similarly, the pose that our alphabet contains only two symbols, class PSPACE contains those problems solvable so that Σ= {a; b}, and Σ* is the set of all finite in polynomial space. The class EXPTIME con- strings consisting of a’s and b’s. The palindrome sists of the problems solvable in exponential problem PAL is defined by letting the instances time; a problem is solvable in exponential time if k in PAL be all those strings in Σ* that read the there is a k for which it is solvable in time 2n . same forward as backwards; for example, aba The class EXPSPACE contains those problems and bbb are both palindromes, but ab and bba solvable in exponential space. The satisfiability are not. This problem can be solved in time n2 problem is in EXPTIME; whether it is in P is a by a simple strategy that involves checking the major open problem.

20 Complexity

use the function f to produce an efficient algo- 3 Hierarchies and Reducibility rithm for L1. Conversely, if there is no efficient

algorithm for L1, then there cannot be an efficient

A fundamental early result of complexity theory algorithm for L2. Notice that the class P is closed is the existence of strict hierarchies among prob- under polynomial-time reductions since if L1 is lems. So, for example, we can prove that there reducible to L2, and L2 is in P, then L1 is also are problems that can be solved in time n2, but in P. not in time n, and similar theorems hold for If C is a complexity class, and L is a problem space bounds on algorithms. To state this result in C so that any problem in C is reducible to L, in its most general form, we introduce the con- then L is said to be C-complete. Such problems cept of a space constructible function. A func- are the hardest problems in C; if any problem tion S(n) is said to be space constructible if there in C is computationally intractable, then a C- is a Turing machine M that is S(n) space complete problem is intractable. The technique bounded, and for each n there is an input of of reducing one problem to another is very flex- length n on which M actually uses S(n) tape ible, and has been used to show a large variety cells. All “reasonable” functions such as n2, n3, of problems in computer science, combinatorics, and 2n are space constructible. The space hier- algebra, and combinatorial game theory intrac- archy theorem, proved by Hartmanis, Lewis and table. We now provide some examples of such

Stearns in 1965, says that if S1(n) and S2(n) are problems. space constructible functions, and S2 grows faster The time hierarchy theorem implies that there than S1 asymptotically, so that are problems in EXPTIME that require expon- ential time for their solution, no matter what Sn() algorithm is employed. The reduction method lim inf 1 = 0, n→∞ Sn2() then allows us to draw the same conclusion for other problems. For example, let us define gen- then there exists a problem solvable in space eralized chess to be a game with rules similar to × S2(n), but not in space S1(n). A similar hierarchy standard chess, but played on an n n board, theorem holds for complexity classes defined by rather than an 8 × 8 board. Fraenkel and time bounds. The hard problems constructed in Lichtenstein (1981) used the reduction technique the proofs of the hierarchy theorems are pro- to show that generalized chess is EXPTIME- duced by diagonalizing over classes of machines, complete, and hence computationally intractable. and so are not directly relevant to problems aris- EXPSPACE-complete problems are also com- ing in practice. However, we can prove lower putationally intractable. An example of a prob- bounds on the complexity of such problems by lem of this type in classical algebra is provided by using the technique of efficient reduction. We the word problem for commutative semigroups. wish to formalize the notion that one problem Here the problem is given in the form of a finite can be reduced to another in the sense that if we set of equations formed from a set of constants had an efficient algorithm for the second prob- using a single binary operation that is assumed lem, then we would have an efficient algorithm to be associative and commutative, together with = for the first. Let L1 and L2 be problems ex- a fixed equation s t. The problem is to deter- Σ Σ = pressed in alphabets 1 and 2. L1 is said to be mine whether s t is deducible from the set of polynomial-time reducible to L2 (briefly, reduc- equations, assuming the usual rules for equality. ible to L2) if there is a polynomial-time comput- Mayr and Meyer in 1981 showed this problem Σ Σ able function f from *1 to *2 such that for any to be EXPSPACE-complete, so that any algo- Σ string s in *,1 s is in L1 if and only if f(s) is in rithm solving this problem must use an expon-

L2. Other notions of reducibility can be defined ential amount of space on infinitely many inputs. by varying the class of functions f that imple- Logic also provides a fertile source of examples ment the reduction. The importance of the con- of intractable problems. Although the decision cept lies in the fact that if we have an efficient problem for true sentences of number theory is algorithm solving the problem L2, then we can unsolvable, if we restrict ourselves to sentences 21 Alasdair Urquhart that involve only the constants 0 and 1, together examples. The implication-conjunction fragment with and the addition symbol, then there of the logic R of relevant implication was proved is an algorithm to determine whether such a sen- decidable by Saul Kripke in 1959 using a - tence is true or false, a result proved by Presburger icated combinatorial lemma. The author of the in 1930. However, in 1973 Rabin and Fischer present chapter showed (Urquhart 1999) that showed that the inherent complexity of this prob- this propositional logic has no primitive recursive lem is doubly exponential. This means that for decision procedure, so that Kripke’s intricate any machine solving this problem, there is a con- method is essentially optimal. This is perhaps the stant c > 0 so that for infinitely many sentences most complex decidable nonclassical logic known. cn the machine takes at least 22 steps to determine whether it is true or not. If we add quantification over finite sets, then 4 NP-completeness and Beyond we can prove even more powerful lower bounds. The weak monadic second-order theory of one successor (WS1S) is formulated in a second- A very common type of computational problem order language with equality, the constant 0 and consists in searching for a solution to a fixed set a successor function. In the intended interpreta- of conditions, where it is easy to check whether tion for this theory, the second-order quantifiers a proposed solution really is one. Such solutions range over finite sets of non-negative integers. may be scattered through a very large set, so The decision problem for this theory was proved that in the worst case we may be reduced to to be solvable by Büchi in 1960, but its inherent doing an exhaustive search through an exponen- complexity is very high. Albert Meyer showed in tially large set of possibilities. Many problems of 1972 that an algorithm deciding this theory must practical as well as theoretical interest can be use for infinitely many inputs of length n an described in this general setting. The theory of amount of space that is bounded from below by NP-completeness derives its central importance an iterated exponential function, where the stack in computer science from its success in provid- contains at least dn iterations, for a fixed d > 0. ing a flexible theoretical framework for this type The conclusion of the previous paragraph of problem. could be challenged by pointing out that Meyer’s A problem L belongs to the class NP if there lower bound is an asymptotic result that does not is a polynomial p and a polynomial-time com- rule out a practical decision procedure for sen- putable relation R so that a string x is in L if and tences of practically feasible size. However, a fur- only if there is a string y so that the length of y ther result shows that astronomical lower bounds is bounded by p(|x |), and R(x, y) holds. The can be proved for WS1S even if we restrict the idea behind the definition is that we think of y length of sentences. A Boolean network or cir- as a succinct proof (or ‘certificate’) that x is in P, cuit is an acyclic directed graph in which the where we insist that we can check efficiently that nodes are labeled with logical operators such an alleged proof really is a proof. as AND, OR, NOT etc. Such a network with Here are a few examples to illustrate this designated input and output nodes computes a definition. Consider the problem of determining Boolean function in an obvious way. Meyer and whether an integer in decimal notation is non- Stockmeyer showed that any such network that prime (that is to say, the strings in the problem decides the truth of all sentences of WS1S of are the decimal representations of numbers that length 616 or less must contain at least 10123 are not prime). Then a proof that a number x is nodes. Even if the nodes were the size of a not prime consists of a pair of numbers y, z > 1 proton and connected by infinitely thin wires, the so that yz = x. The satisfiability problem is also network would densely fill the known universe. easily seen to be in NP. Here a positive instance Inherently intractable problems also exist in the of the problem consists of a satisfiable formula F area of nonclassical propositional logics. The area of propositional logic; the proof that F is satisfi- of substructural logics, such as linear logic and able is simply a line of a truth-table. It is obvi- relevance logics, provides us with several such ous that we can check very quickly if a formula

22 Complexity is satisfied by an assignment; on the other hand, of the problem of factoring a number that is the the best current algorithms for satisfiability in product of two large prime numbers. The same the worst case are forced to check exponentially remarks apply to other cryptosystems, with the many possibilities, thus being not much differ- exception of the theoretically invulnerable one- ent from the crude brute-force method of trying time pad system. The fact that no such feasible all possible lines in the truth-table for a given algorithm has been found for any of these prob- formula. lems is one of the main reasons for the wide- The satisfiability problem occupies a central spread belief in the conjecture that P ≠ NP. place in theoretical computer science as the best The lower bounds described in the preceding known NP-complete problem. Any problem in section were all proved by the diagonal method. NP can be reduced efficiently to the satisfiability That is to say, the method in each case was an problem. This reflects the fact that the language adaptation of the technique originally employed of propositional logic forms a kind of universal by Cantor to prove the set of real numbers un- language for problems of this type. Given a prob- countable, and subsequently adapted by Church lem in NP, it is usually a routine exercise to see and Turing to prove the decision problem for how to translate the problem into a set of condi- predicate logic unsolvable. There are reasons to tions in propositional logic so that the problem think that this method will not succeed in resolv- has a solution if and only if the set of conditions ing the problem of whether or not P = NP. To is satisfiable. For example, consider the problem explain these reasons, we need to introduce the of coloring a map in the plane with three colors. concept of a Turing machine with an oracle. Here the problem takes the form of a map, and The concept of a Turing machine explicates the a set of three colors, say red, white, and blue, so notion of computability in an absolute sense. that adjacent countries are colored differently. Similarly, the concept of an oracle machine explic- We can formalize this problem by introducing ates the general notion of what it means for a a set of constants to stand for the countries in problem to be solvable relative to another prob- the map, and variables Rx, Wx, Bx to stand for lem (the definition of reducibility above is a spe- “Country x is colored red (white, blue).” The cial case of this general notion). If A is a set of reader should check that given a map, we can strings then a Turing machine with oracle A is quickly write down a corresponding set of con- defined to be a Turing machine with three spe- ditions in propositional logic that formalizes the cial states q?, qy, and qn. The query state q? is statement that the map can be properly colored used to ask “Is the string of nonblank symbols with the three colors. to the right of the reading head in A?” The Cook’s famous theorem of 1971 showing that answer is supplied by having the machine change satisfiability is NP-complete was quickly followed on the next move to one of the two states qy by proofs that many other well-known computa- or qn, depending on whether the answer is yes tional problems fall into this class. Since then, or no. Time and space of a computation by an thousands of significant problems have been oracle machine are computed just as for an ordin- proved NP-complete; for a partial list, see the ary Turing machine, counting the time taken for book by Garey and Johnson (1979). The ubiquity the answer to the oracle query as one step (the of NP-completeness in the theory of combinator- oracle answers any query instantaneously). ial problems means that a proof of P = NP (that We can imagine an oracle machine as rep- is to say, a proof that there is a polynomial-time resenting a situation where we have access to a algorithm for satisfiability) would have dramatic “black box” that instantaneously answers ques- consequences. It would mean that feasible solu- tions belonging to a type for which we have tions would exist for hundreds of problems that no algorithm, or for which the only known are currently intractable. For example, the RSA algorithm is very inefficient. For example, sup- cryptosystem, widely used for commercial trans- pose that the oracle (black box) can answer all actions on the internet, would immediately be queries of the form: “Do all integers n satisfy vulnerable to computer attack, since the security the property P(n)?,” where P is a decidable prop- of the system rests on the assumed intractability erty of integers. Then the black box exhibits a

23 Alasdair Urquhart kind of limited omniscience that would enable The family of algorithms operating in polyno- us to answer instantaneously many open prob- mial space appears to be a much more extensive lems of current mathematics such as Goldbach’s class than the family of algorithms operating in conjecture or the Riemann hypothesis. In spite polynomial time. However, we are unable on the of this, it is possible to show that there are prob- basis of current knowledge to refute the equality lems that such a miraculous machine cannot P = PSPACE. This illustrates the point men- answer; classical recursion theory (computability tioned in the introduction, that in contrast to theory) is largely taken up with such problems. the detailed hierarchy theorems known for time If A is any set of strings, then by imitating the and space separately, the problem of relating time definitions of the complexity classes above, but and space requirements for computations remains substituting “Turing machine with oracle A” largely unsolved. everywhere for “Turing machine” we can define relativized complexity classes P(A), NP(A), and so on. Baker, Gill, and Solovay proved in 1975 5 Parallel Computation that there is a decidable oracle A for which P(A) = NP(A), and a decidable oracle B for which P(B) ≠ NP(B). The significance of this theorem The computational model discussed in the pre- lies in the fact that known techniques of diagon- ceding sections was that of serial or sequential alization, such as are used in computability the- computation, where the machine is limited to a ory, continue to work in the presence of oracles. bounded number of actions at each step, for Thus it provides evidence that standard diagonal example, writing a symbol, moving left or right, techniques are inadequate to settle such ques- and changing internal state in the case of the tions as “P = NP?” Turing model. However, there is considerable The literature of theoretical computer sci- current interest, both theoretical and practical, ence contains many complexity classes beyond in parallel models of computation. Parallel com- the few discussed here; for details, the reader putation is attractive in applications such as should consult the collection of survey articles in searching large databases, and also is of interest Van Leeuwen (1990). We conclude this section in modeling brain function (since the brain seems, with a brief description of an important com- to a first approximation, to be some kind of plexity class that, like the classes P and NP, has parallel computer). In this section, we provide strong connections with logic. The class PSPACE a brief discussion of the complexity of parallel consists of those problems solvable using a poly- computation. nomial amount of space. It is not hard to see that In the case of parallel computation, various this class contains the class NP, since we require models have been proposed, and there is no only a small amount of space to do an exhaustive universal agreement on the best. These include search through the space of all possible strings models such as the PRAM (parallel random access that are candidates for certificates showing that a machine), where a large number of simple pro- string is a positive instance of an NP-complete cessors with limited memory have joint access to problem. This class of problems bears the same a large shared memory; various conventions on relationship to the quantified propositional cal- read–write conflicts can be adopted. For more culus as the class NP to the ordinary propositional details on these models, the readers should con- calculus. In the quantified propositional calculus, sult the articles of van Emde Boas, Karp, and we add to ordinary propositional logic quantifiers Ramachandran in Van Leeuwen (1990). We shall ranging over propositions. Thus, for example, not discuss these models further here, but instead the formula ∃p∀q(p → q) is a logical truth in describe the area of non-uniform complexity. this language. The valid (logically true) formulas The Turing model has the property that a of quantified propositional logic constitute a single machine operates on inputs of arbitrary PSPACE-complete set (that is, the problem of length. An alternative approach to measuring determining the validity of formulas in the quan- the complexity of computations is to limit our- tified language is PSPACE-complete). selves to functions of a fixed input and output

24 Complexity size – Boolean functions, in the case of decision “polynomially bounded amount of advice”; con- problems – and then estimate the minimum size versely, any problem solved by such a machine of the circuitry needed to provide a “hard-wired” has polynomial-size circuits. version of the function. The description of P/poly in the preceding We define a circuit as a finite, labeled, directed paragraph should make it clear that we are dealing graph with no directed cycles. The nodes with with an extremely powerful class of procedures, no arrows pointing in are input nodes, while the since they have the ability to answer arbitrarily nodes with no arrows pointing out are output complex questions about finite configurations in nodes. The internal nodes are considered as logic the time it takes to write down the question gates, and labeled with appropriate Boolean func- (and so should be considered as “algorithms” tions. For example, a circuit could be built from only in an extended sense). Nevertheless, it is AND gates with two inputs and one output, widely conjectured that the satisfiability problem and NOT gates with one input and one output. does not have polynomial-size circuits. Current The important complexity measures for a circuit proof techniques in the theory of Boolean cir- are its depth (length of the shortest path from cuits seem to be inadequate for resolving this an input node to an output node) and its size challenging conjecture. (number of nodes in the circuit). We can now define parallel complexity classes using the circuit model. Perhaps the most im- 6 Complexity and Philosophy portant of these is the class of problems with polynomial-size circuits, abbreviated as P/poly. Given a problem L, we can encode the strings Philosophical treatments of the concept of of L in binary notation; let us refer to this computation often ignore issues relating to com- encoded problem as Lb. Then L is said to have plexity. However, we shall argue in this section polynomial-size circuits if there is a polynomial p that such questions are directly relevant to some so that for every n there is a Boolean circuit C frequently discussed problems. Since Turing’s with size bounded by p(n) so that C gives the famous article of 1950, it has been common to output 1 exactly for those binary strings of length replace the question “Can machines think?” – n that belong to Lb, that is, exactly those strings which Turing thought too meaningless to dis- of length n that represent positive instances of cuss – with the question “Can digital computers the problem L. successfully play the imitation game?” Turing This is a much more powerful model of com- made the following optimistic prediction: putation than the standard Turing model; it is non-uniform, since we allow a different circuit I believe that in about fifty years’ time it will for each input length n. In particular, it is not be possible to programme computers, with a hard to see that in this model, an unsolvable storage capacity of about 109, to make them problem can have polynomial-size circuits. play the imitation game so well that an aver- The connection between the circuit model and age interrogator will not have more than the Turing model can be made more precise by 70 per cent chance of making the right identi- considering the oracle machines defined earlier. fication after five minutes of questioning. Given a fixed circuit, we can easily program a (Turing 1950) Turing machine to simulate its behavior, simply by encoding the circuit as a big look-up table It is clear that Turing was thinking of com- (we discuss the philosophical import of this puters as real physical devices. However, let us below). Hence, if a problem L has suppose that for the moment we think of com- polynomial-size circuits, we can program an puters as idealized mathematical machines, and oracle machine that, relative to the oracle set C (as is common in the mathematical context of representing the encoding of the family of cir- computability theory) ignore all questions of cuits solving L, solves the problem. The machine resources, efficiency, and so forth. Then it is a can be considered as a machine that takes a mathematical triviality that the answer to Turing’s

25 Alasdair Urquhart question is affirmative. Let us recall the basic gates. This model is sometimes described in the situation for the imitation game. An interrogator cognitive science literature as “parallel distributed communicates by teletype with two participants, processing” or “PDP” for short. If we take into one a human being, the other a digital computer. account speed of processing, then such models The task of the interrogator is to determine may indeed provide more accurate simulations by skillful questioning which of the two is the of processes in real brains, since neurophysiology human being. For a computer to succeed at the indicates that mammalian brains are made out of imitation game means that it can succeed in fool- relatively slow elements (neurons) joined together ing the interrogator in a substantial number of in a highly connected network. On the other cases, if the game is played repeatedly. hand, there is nothing new here as compared Turing envisages a limit of five minutes of with the classical serial model of computation, if interrogation, but for our present purposes, let we ignore limitations of time and space. Never- us suppose that we simply limit the number of theless, some of the literature in cognitive science symbols exchanged between the participants in argues otherwise. the game to some reasonably large number (bear- In their debate of 1990 with , ing in mind that all the participants have to type Paul and largely agree with at human speeds, otherwise the computer could the conclusions of Searle’s critique of classical be spotted immediately). It is now easy to see that AI (based on a serial model of computation), there is indeed a machine (in the mathematical for which Searle argues on the grounds of his sense) that can play this game with perfect suc- “” thought experiment, but dis- cess (i.e. a skilled interrogator cannot improve agree with the conclusions of his “Chinese gym” on random guessing in the game). Consider all thought experiment designed to refute the claims sequences of symbols representing a possible of parallel processors to represent the mind. The sequence of questions and answers in the imita- Churchlands first point out the implausibility of tion game. Of these, some will be bad, in the the simulation (involving huge numbers of people sense that they will easily reveal to the inter- passing messages in an enormous network), but rogator the identity of the computer, while others then continue: are good (we can imagine these to be the sort of responses produced when the computer is re- On the other hand, if such a system were to be placed by a human). Now provide the computer assembled on a suitably cosmic scale, with all with the set of all good sequences as a gigantic its pathways faithfully modeled on the human look-up table, and program the computer to case, we might then have a large, slow, oddly answer in accordance with this table. By defini- made but still functional brain on our hands. In tion, the computer must succeed perfectly at the that case the default assumption is surely that, game. given proper inputs, it would think, not that it Of course, the “machine” described in the pre- couldn’t. (Churchland & Churchland 1990) vious paragraph is a pure mathematical abstrac- tion, but it suffices to illustrate the fact that in This imaginary cosmic network is a finite philosophical, as opposed to mathematical, con- object, working according to a fixed algorithm texts, the purely abstract definition of a machine (as embodied in its pattern of connections). It is not appropriate. Similar remarks apply in the follows that it can be simulated by a serial com- case of the distinction between serial and parallel puter (in fact, all of the early research on “neural computation. nets” was carried out by writing simulation pro- It is currently fashionable to think of cognitive grams on computers of conventional design). Of processes as modeled by neural networks com- course, there will be a loss of speed, but the posed of simple elements (typically threshold Churchlands explicitly rule out speed of opera- gates of some kind), joined together in some tion as a relevant variable. It’s difficult to see, random fashion, and then “trained” on some though, why the serial computer doing the simu- family of inputs, the “learning” process consisting lation of the neural network should be ruled out of altering the strength of connections between as a “functional brain.”

26 Complexity

Let us expand a little more on the implica- imitation game), and devices that are physically tions of this analysis. The basic fact that serial constructible. machines can simulate parallel machines (a point emphasized by Searle himself) should not be References considered as an argument for or against either the Chinese-room argument or the Chinese- Churchland, P. M. and Churchland, P. S. 1990. gym argument, both of which involve obscure “Could a machine think?” Scientific American questions concerning the presence or absence 262: 32–7. [A spirited reply to John Searle’s of “mental contents” or “semantic contents.” article in the same issue describing his “Chinese Rather, it points to the difficulties of a position room” thought experiment.] that rejects a serial model of computation for Fraenkel, A. S. and Lichtenstein, D. 1981. “Com- × the mind, but accepts a parallel model, while puting a perfect strategy for n n chess requires ignoring questions of complexity and efficiency. time exponential in n.” Journal of Combinatorial Theory Series A 31: 199–214. Since we are not limited by technological Garey, M. R. and Johnson D. S. 1979. Computers feasibility, let us imagine a huge, super-fast serial and Intractability: A Guide to the Theory of NP- computer that simulates the Churchlands’ cosmic completeness. San Francisco: Freeman. [This well- network. Furthermore, to make the whole thing known text contains a very readable introduction more dramatic, let’s imagine that this marvelous to the theory of NP-completeness, as well as an machine is wired up to a gigantic cosmic network extensive list of NP-complete problems from with flashing lights showing the progress of the many areas.] computation, working so fast that we can’t tell Li, M. and Vitányi, P. 1997. An Introduction to the difference between a real cosmic network Kolmogorov Complexity and its Applications. New and the big display. Is this a “functional brain” York: Springer-Verlag. [The basic textbook on or not? It’s hard to know what the criteria are this fascinating theory; it contains detailed tech- for having a “functional brain on our hands,” nical developments as well as more discursive but without considering questions of computa- material on the general notion of complexity.] tional complexity, it is difficult to see how we Parberry, I. 1994. Circuit Complexity and Neural can reject serial candidates for “functional brains.” Networks. Cambridge, MA: MIT Press. [The first For a more detailed discussion of the “Chinese chapter is an excellent nontechnical discussion of room” argument from a complexity-theoretic the Chinese-room thought experiment from a perspective, the reader should consult Parberry complexity-theoretic point of view. The remainder 1994. (See also Chapter 9, THE PHILOSOPHY OF of the book is a more technical, but accessible, AI AND ITS CRITIQUE.) discussion of the problem of scaling in neural Current work in the philosophy of mind network theory.] manifests a fascination with far-fetched thought Turing, A. 1950. “Computing machinery and intel- ligence.” Mind 59: 433–60. [The classic article experiments, involving humanoid creatures mag- in which the great computer pioneer presented ically created out of swamp matter, zombies, and the “TuringTest” for machine intelligence.] similar imaginary entities. Philosophical discus- Urquhart, A. 1999. “The complexity of decision sion on the foundations of cognitive science also procedures in relevance logic II.” Journal of frequently revolves around implausible thought Symbolic Logic 64: 1774–1802. [This article gives experiments like Searle’s “Chinese room” argu- a detailed proof that the propositional logic of ment. The point of the simple relevant implication, restricted to conjunction and above is that unless computational resources are implication, has enormous intrinsic complexity.] considered, arguments based on such imagin- Van Leeuwen, J., ed. 1990. Handbook of Theoret- ary experiments may appear quite powerful. ical Computer Science, Volume A: Algorithms and On the other hand, by taking such resources Complexity. Amsterdam: Elsevier. [A collection into account, we can distinguish between objects of detailed survey articles by leading researchers that exist in the purely mathematical sense (such covering many topics, including parallel complex- as the Turing machine that succeeds at the ity and cryptology.]

27 Klaus Mainzer

Chapter 3 System: An Introduction to Systems Science Klaus Mainzer

The first section of this chapter defines the basic Introduction concept of a dynamical system. The dynamics of systems is measured by time series and modeled Dynamical systems, with their astonishing variety in phase spaces, which are introduced in sec- of forms and functions, have always fascinated tion 2. Phase spaces are necessary to recognize scientists and philosophers. Today, structures attractors of a system, such as chaos. In the case and laws in nature and society are explained by of chaos, severe restrictions on long-term pre- the dynamics of complex systems, from atomic dictions and systems control must be taken into and molecular systems in physics and chemistry account. But, in practice, there are only finitely to cellular organisms and ecological systems many measurements and observations of a time in biology, from neural and cognitive systems series. So, in section 3, time-series analysis is in brain research and cognitive science to introduced in order to reconstruct phase spaces societies and market systems in sociology and and attractors of behavior. Section 4 presents ex- economics. In these cases, complexity refers amples of complex systems in nature and society. to the variety and dynamics of interacting ele- From a philosophical point of view, dynamical ments causing the of atomic and systems in nature and society can be considered molecular structures, cellular and neural patterns, as information and computational systems. This or social and economic order (on computa- deep insight of modern systems science is dis- tional complexity see Chapter 1, COMPUTATION). cussed in the last section. Computational systems can simulate the self- organization of complex dynamical systems. In these cases, complexity is a measure of computa- 1 Basic Concepts of Systems tional degrees for predictability, depending on Science the information flow in the dynamical systems. The philosophy of modern systems science aims to explain the information and computational A dynamical system is characterized by its ele- dynamics of complex systems in nature and ments and the time-depending development of society. their states. In the simple case of a falling stone,

28 Introduction to Systems Science one may consider for example only the accelera- position and momentum of a particle) depend- tion of a single element. In a planetary system, ing on Heisenberg’s principle of uncertainty. The the states of planets are also determined by their latter principle allows only probabilistic forecasts position and momentum. The states can also of future states. refer to moving molecules in a gas, the excita- tion of neurons in a neural net, nutrition of populations in an ecological system, or products 2 Dynamical Systems, Chaos, and in a market system. The dynamics of a system, Other Attractors that is, the change of system states depending on time, is mathematically described by differen- tial equations. For deterministic processes (e.g., During the centuries of classical physics, the motions in a planetary system), each future state universe was considered as a deterministic and is uniquely determined by the present state. A conservative system. The astronomer and math- conservative (Hamiltonian) system, e.g. an ideal ematician Pierre-Simon Laplace (1814), for pendulum, is determined by the reversibility of example, assumed the total computability and time direction and conservation of . Con- predictability of nature if all natural laws and ini- servative systems are closed and have no ener- tial states of celestial bodies are well known. The getic dissipation with their environment. Thus, Laplacean spirit well expressed philosophers’ faith conservative systems in the strict sense exist only in and computability of the world as approximations like, e.g., an ideal Thermos during the eighteenth and nineteenth centuries. bottle. In our everyday world, we mainly observe Laplace was right about linear and conserva- dissipative systems with a distinct time direction. tive dynamical systems. A simple example is a so- Dissipative systems, e.g., a real pendulum with called harmonic oscillator, like a mass attached friction, are irreversible. to a spring oscillating regularly without friction. In classical physics, the dynamics of a system Let us consider this example in more detail. It is analyzed as a continuous process. In a famous will help us to introduce the basic notions of quotation, Leibniz assumed that nature does time series, phase space, and trajectory, essential not jump (natura non facit saltus). However, to understand the structure and development of continuity is only a mathematical idealization. dynamical systems. In general, a linear relation Actually, a scientist deals with single observa- means that the rate of change in a system is tions or measurements at discrete time points proportional to its cause: small changes cause that are chosen equidistant or defined by other small effects, while large changes cause large measurement devices. In discrete processes, there effects. In the example of a harmonic oscillator, are finite differences between the measured states, a small compression of a spring causes a small no infinitely small differences between the meas- oscillation of the position of a mass, while a large ured states, and no infinitely small differences compression causes a large oscillation, following (differentials) that are assumed in a continuous Hooke’s law. Changes of a dynamical system process. Thus, discrete processes are mathemat- can be modeled in one dimension by changing ically described by difference equations. values of a time-depending quantity along the Random events (e.g., Brownian motion in time axis (time series). In figure 3.1a, the posi- a fluid, mutation in evolution, in tion x(t) of a mass attached to a spring is oscil- economy) are represented by additional fluctu- lating in regular cycles along the time axis t. x(t) ation terms. Classical stochastic processes, e.g. the is the solution of a linear equation, according to billions of unknown molecular states in a fluid, Hooke’s law. Mathematically, linear equations are defined by time-depending differential equa- are completely computable. This is the deeper tions with distribution functions of probabilistic reason for Laplace’s philosophical assumption to states. In quantum systems of elementary particles, be right for linear and conservative systems. the dynamics of quantum states is defined by In systems theory, the complete information Schrödinger’s equation with observables (e.g., about a dynamical system at a certain time is

29 Klaus Mainzer

v(t) 1 x(0) = 0.5

π ω 2 / = x[0] x(0) 1.0 0

0 t

–1 –x[0] –1 0 1 x(t) (a) (b)

Figure 3.1a: A solution x(t) of a linear equation as time series (Kaplan 1995: 211) Figure 3.1b: Trajectories of two solutions of a linear equation in a 2-dimensional phase space (Kaplan 1995: 212) determined by its state at that time. In the ex- complex, depending on more quantities, with pat- ample of a harmonic oscillator, the state of the terns of behavior that are not as regular as in the system is defined by the position x(t) and the simple case of a harmonic oscillator. It is a main velocity v(t) of the oscillating mass at time t. Thus, insight of modern systems theory that the beha- the state of the system is completely determined vior of a dynamic system can only be recognized if by a pair of two quantities that can be represented the corresponding state space can be reconstructed. geometrically by a point in a 2-dimensional At the end of the nineteenth century, Henri phase space, with a coordinate of position and a Poincaré (1892–3) discovered that celestial coordinate of velocity. The dynamics of a system mechanics is not a completely computable clock- refers to the time-depending development of its work, even if it is considered as a deterministic states. Thus, the dynamics of a system is illus- and conservative system. The mutual gravita- trated by an orbit of points (trajectory) in a phase tional interactions of more than two celestial space corresponding to the time-depending de- bodies (“Many-bodies-problem”) correspond to velopment of its states. In the case of an harmonic nonlinear and non-integrable equations with oscillator, the trajectories are closed ellipses instabilities and irregularities. According to the around a point of stability (figure 3.1b), cor- Laplacean view, similar causes effectively deter- responding to the periodic cycles of time series, mine similar effects. Thus, in the phase space, oscillating along the time axis (figure 3.1a). trajectories that start close to each other also Obviously, the regular behavior of a linear and remain close to each other during time evolu- conservative system corresponds to a regular and tion. Dynamical systems with deterministic chaos stable pattern of orbits. So, the past, present, exhibit an exponential dependence on initial and future of the system are completely known. conditions for bounded orbits: the separation In general, the state of a system is determined of trajectories with close initial states increases by more than two quantities. This means that exponentially (figure 3.2). higher dimensional phase space is required. From Tiny deviations of initial data lead to expon- a methodological point of view, time series and entially increasing computational efforts to ana- phase spaces are important instruments to study lyze future data, limiting long-term predictions, systems dynamics. The state space of a system con- although the dynamics is in principle uniquely tains the complete information of its past, present determined. This is known as the “butterfly and future behavior. The dynamics of real sys- effect”: initial, small, and local causes soon lead to tems in nature and society is, of course, more unpredictable, large and global effects (see figure

30 Introduction to Systems Science

x1(t) x2 ∆(t) ~ ∆(0)exp(λt)

x2(t)

x1(0)

∆(0) x2(0)

x1

Figure 3.2: Exponential dependence of two trajectories x1(t) and x2(t) on initial conditions

20 10

Dt 0 –10 –20 0 500 1000 1500 2000 t Figure 3.3: A measured signal Dt from a chaotic system (time series)

3.3). According to the famous KAM Theorem of periodic or quasi-periodic. A periodic orbit is a A. N. Kolmogorov (1954), V. I. Arnold (1963), closed trajectory into which all trajectories depart- and J. K. Moser (1967), trajectories in the phase ing from an adjacent region converge. For a space of classical mechanics are neither completely simple dynamical system with only two degrees regular, nor completely irregular, but depend of freedom and continuous time, the only pos- sensitively on the chosen initial conditions. sible attractors are fixed points or periodic limit Dynamical systems can be classified on the basis cycles. An example is a Van der Pol oscillator of the effects of the dynamics on a region of the modeling a simple vacuum-tube oscillator circuit. phase space. A conservative system is defined by In continuous systems with a phase space of the fact that, during time evolution, the volume dimension n > 2, more complex attractors are of a region remains constant, although its shape possible. Dynamical systems with quasi-periodic may be transformed. In a dissipative system, dy- limit cycles show a time evolution that can be namics causes a volume contraction. An attractor decomposed into different periodic parts with- is a region of a phase space into which all traject- out a unique periodic regime. The correspond- ories departing from an adjacent region, the so- ing time series consist of periodic parts of called basin of attraction, tend to converge. There oscillation without a common structure. Never- are different kinds of attractors. Fixed points form theless, closely starting trajectories remain close the simplest class of attractors. In this case, all to each other during time evolution. The third trajectories of adjacent regions converge to a class contains dynamical systems with chaotic point. An example is a dissipative harmonic attractors that are nonperiodic, with an exponen- oscillator with friction: the oscillating system is tial dependence on initial conditions for bounded gradually slowed down by frictional forces and orbits. A famous example is the chaotic attractor finally comes to a rest in an equilibrium point. of a Lorenz system (Lorenz 1963) simulating Conservative harmonic oscillators without the chaotic development of weather caused by friction belong to the second class of attractors local events, which cannot be forecast in the with limit cycles, which can be classified as being long run (“butterfly effect”) (figure 3.4b).

31 Klaus Mainzer

Dt Z

Y

Dt–8 X Dt–16 (a) (b)

Figure3.4a: The reconstructed trajectory of a measured series (fig. 3.3) in an embedding phase space of three dimensions with time lag Figure 3.4b: The trajectory of the original phase space of the chaotic system (Kaplan 1995: 310)

The goal for this kind of time-series analysis is 3 Dynamical Systems and Time- comparable to constructing a computer program series Analysis without any knowledge of the real system from which the data come. As a black box, the com- We have started by seeing the kind of mathemat- puter program would take the measured data ical equations of dynamical systems required to as input and provide as output a mathematical derive their patterns of behavior; the latter have model describing the data. But, in this case, it is been characterized by time series and attractors difficult to identify the meaning of components in in phase spaces, such as fixed points, limit cycles, the mathematical model without understanding and chaos. This top-down approach is typically the dynamics of the real systems. Thus, the top- theoretical: we use our understanding of real sys- down and bottom-up approach, model-building tems to write dynamical equations. In empirical and time-series analysis, expert knowledge in the practice, however, we must take the opposite fields of application, and mathematical and pro- bottom-up approach and start with finite sequences gramming skills, must all be integrated in an of measurements, i.e. finite time series, in order to interdisciplinary research strategy. find appropriate equations of mathematical models In practice, only a time series of a single (one- with predictions that can be compared with meas- dimensional) measured variable is often given, urements made in the field of application. although the real system is multidimensional. Measurements are often contaminated by un- The aim of forecasting is to predict the future wanted noise, which must be separated from the evolution of this variable. According to Takens’ signals of specific interest. Moreover, in order to theorem (1981), in nonlinear, deterministic, and forecast the behavior of a system, the develop- chaotic systems, it is possible to determine the ment of its future states must be reconstructed in structure of the multidimensional dynamic system a corresponding phase space from a finite sequence from the measurement of a single dynamical of measurements. So time-series analysis is an variable (figure 3.3). immense challenge in different fields of research Takens’ method results in the construction such as climatic data in meteorology, ECG- of a multidimensional embedding phase space for signals in cardiology, EEG-data in brain research, measured data (figure 3.4a) with a certain time or economic data of business cycles in economics. lag in which the dynamics of attractors is similar

32 Introduction to Systems Science to the orbits in the phase space of the chaotic have been discovered in the self-organized system (figure 3.4b). behavior of quite different systems in physics, The disadvantage of Takens’ theorem is that chemistry, biology, and sociology. These it does not detect and prove the existence of a multicomponent systems consist of many units chaotic attractor. It only provides features of an like elementary particles, atoms, cells, or organ- attractor from measured data, if the existence of isms. Properties of these elementary units, such the attractor is already guaranteed (Grassberger as their position and momentum vectors, and & Procaccia 1983). The dimension of an attractor their local interactions constitute the microscopic can be determined by a correlation integral level of description (imagine the interacting defining the different frequency with which a molecules of a liquid or gas). The global state of region in an attractor is visited by the orbits. the complex systems results from the collective Thus, the correlation integral also provides a configurations of the local multicomponent states. method to study the degrees of periodicity and At the macroscopic level, there are few collective aperiodicity of orbits and measured time series. (“global”) quantities like, for instance, pressure, The Lyapunov spectrum shows us the depend- density, temperature, and entropy characterizing ence of dynamics from initial data. The so-called observable collective patterns or figures of the Lyapunov exponents measure the averaged ex- units. ponential rates of divergence or convergence of If the external conditions of a system are neighboring orbits in phase space. If the largest changed by varying certain control parameters Lyapunov exponent is positive, the attractor is (e.g., temperature), the system may undergo a chaotic, and the initial small difference between change in its macroscopic global states at some two trajectories will diverge exponentially (figure threshold value. For instance, water as a com- 3.2). If the largest exponent is zero and the rest plex system of molecules changes spontaneously is negative, the attractor is a periodic limit cycle. from a liquid to a frozen state at the critical If there is more than one exponent equal to value of temperature with zero celsius. In physics, zero, the rest being negative, the behavior is those transformations of collective states are quasi-periodic. If the exponents are all negative, called phase transitions. Obviously, they describe the attractor is a fixed point. In general, for dis- a change of self-organized behavior between the sipative systems, the sum of Lyapunov exponents interacting elements of a complex system. is negative, despite the fact that some exponents According to Landau and Lifshitz (1959), the could be positive. suitable macrovariables characterizing this change of global order are denoted as “order para- meters.” In statistical mechanics the order trans- ition of complex systems like fluids, gases, etc. is 4 Dynamical Systems in Nature modeled by differential equations of the global and Society state. A paradigmatic example is a ferromagnet consisting of many elementary atomic mag- Structures in nature and society can be explained nets (“dipoles”). The two possible local states by the dynamics and attractors of complex sys- of a dipole are represented by upwards- and tems. They result from collective patterns of downwards-pointing arrows. If the temperature interacting elements that cannot be reduced to (“control parameter”) is annealed to the thermal the features of single elements in a complex sys- equilibrium, in this case the Curie point, then tem. Nonlinear interactions in multicomponent the average distribution of upwards and down- (“complex”) systems often have synergetic ef- wards pointing dipoles (“order parameter”) is fects, which can neither be traced back to single spontaneously aligned in one regular direction causes nor be forecasted in the long run. The (figure 3.5). This regular pattern corresponds to mathematical formalism of complex dynamical the macroscopic state of magnetization. Obvi- systems is taken from statistical physics. In gen- ously, the emergence of magnetization is a self- eral, the theory of complex dynamical systems organized behavior of atoms that is modeled by deals with profound and striking analogies that a phase transition of a certain order parameter,

33 Klaus Mainzer

Figure 3.5: Phase transition in a 2-dimensional Ising model of a ferromagnet (Mainzer 1997: 134) the average distribution of upwards and down- the atomic system during a critical transition. wards pointing dipoles. The atoms have to “obey” the orders of order Landau’s scheme of phase transitions cannot parameters. This mathematical scheme has a very be generalized to all cases of phase transitions. comfortable consequence: it is not necessary (and A main reason for its failure results from an not possible) to compute all microstates of atoms inadequate treatment of fluctuations, which are in a complex system; just find the few macro- typical for many multicomponent systems. Never- scopic order parameters, and you understand the theless, Landau’s scheme can be used as a heuristic dynamics of a complex system. device to deal with several non-equilibrium trans- Actually, the corresponding equations describe itions. In this case, a complex system is driven a competition of several order parameters among away from equilibrium by increasing energy (not each other. The atoms will then obey that order decreasing energy, as in the case of equilibrium parameter that wins the competition. A typical transitions like freezing water or magnetizing example is a Bénard experiment analyzing the ferromagnets). The phase transitions of nonlinear emergence of convection rolls in a fluid layer at dissipative complex systems far from thermal a critical value of a control parameter (temperat- equilibrium can be modeled by several math- ure). The layers of the atmosphere provide fur- ematical methods (Haken 1983, Mainzer 1997, ther examples. In this case, the order parameters Glansdorff & Prigogine 1971). correspond to the two possible rolling directions: As an example, consider a solid-state laser. This “left” or “right” of the convection rolls. During consists of a set of laser-active atoms embedded the phase transition of increasing temperature in a solid-state configuration. The laser end-faces it cannot be forecast which of the two possible act as mirrors and serve two purposes: they order parameters will win the competition, select light modes in axial direction and with because it depends on tiny initial fluctuations on discrete frequencies. If the laser atoms are excited the molecular level. Thus, this phase transition only weakly by external sources (“control para- corresponds to a spontaneous breaking meters”), the laser acts as an ordinary lamp. The of two possible orders. Fluctuations are the driv- atoms, independently of each other, emit wave- ing forces of the system’s evolution. tracks with random phases. The atoms, visualized Simplifying, we may say that old structures as oscillating dipoles, are oscillating completely become unstable, broken down by changing con- at random. If the level of excitement is further trol parameters, and new structures and attractors increased, the atomic dipoles spontaneously oscil- are achieved. If, for example, the fluid of a stream late in phase, although they are excited completely is driven further and further away from thermal at random. Obviously, the atoms show a self- equilibrium, for example by increasing fluid organized behavior of great regularity. The extra- velocity (control parameter), then fluid patterns ordinary coherence of laser light results from the of increasing complexity emerge from vortices collective cooperation of the atomic dipoles. of fixed points, periodic and quasi-periodic oscilla- The laser shows features of phase transitions. tions to chaotic turbulence. Order parameters describe mode amplitudes of More mathematically, stochastic nonlinear dif- the light field becoming unstable at a critical ferential equations (e.g. Fokker–Planck equations, value of pumping. These slowly varying ampli- Master equations) are employed to model the tudes now “slave,” as Haken (1983) claimed, dynamics of complex systems. The dominating

34 Introduction to Systems Science order parameters are founded by the adiabatic signals or orbits in a phase space. There is no elimination of fast-relaxing variables of these commanding cell, but an attractor of collective equations. The reason is that the relaxation time behavior (“order parameter”) dominating the of unstable modes (order parameters) is very long, beating regime of the heart from healthy oscilla- compared to the fast-relaxing variables of stable tions to dangerous chaos. ones, which can therefore be neglected. Thus, In brain research, the brain is considered as a this concept of self-organization can be illustrated complex dynamical system of firing and nonfiring by a quasi-biological slogan: long-living systems neurons, self-organizing in macroscopic pat- dominate short-living systems. terns of cell assemblies through neurochemical Dynamical systems and their phase transitions interactions. Their dynamical attractors are cor- deliver a successful formalism to model the emer- related with states of perception, motion, emotion, gence of order in nature and society. But these , or even consciousness. There is no methods are not reduced to special laws of “mother neuron” that can feel, think, or at least physics, although their mathematical principles coordinate the appropriate neurons. The famous were first discovered and successfully applied in binding problem of pixels and features in a per- physics. Methodologically, there is no , ception is explained by clusters of synchronously but an interdisciplinary approach to explain the firing neurons dominated by learned attractors increasing complexity and differentiation of forms of brain dynamics. by phase transitions. The question is how to The self-organization of complex systems select, interpret, and quantify the appropriate can also be observed in social groups. If a group variables of dynamical models. Let us consider a of workers is directed by another worker, the few examples. so-called foreman, then we get an organized Thermodynamic self-organization is not suf- behavior to produce some product that is by no ficient to explain the emergence of life (see also means self-organized. Self-organization means Chapter 14 in this volume, CYBERNETICS). As that there are no external orders from a foreman, nonlinear mechanism of genetics we use the but that the workers work together by some autocatalytic process of genetic self-replication. kind of mutual understanding, each one doing his The evolution of new species by mutation and job according to a collective concept dominating selection can be modeled by nonlinear stochastic their behavior. equations of second-order non-equilibrium phase In a political community, collective trends transitions. Mutations are mathematized as “fluc- or majorities of opinions can be considered as tuating forces” and selections as “driving forces.” order parameters produced by mutual discussion Fitness degrees are the order parameters dominat- and interaction of the people in a more or less ing the phase transitions to new species. During “heated” situation. They can even be initiated evolution a sensible network of equilibria between by some few people in a critical and unstable populations of animals and plants has developed. (“revolutionary”) situation of the whole com- The nonlinear Lotka–Volterra equations (Lotka munity. There may be a competition of order 1925, Volterra 1931) model the ecological equi- concepts during heavy fluctuations. The essential librium between prey and predator populations point is that the winning concept of order will which can be represented by oscillating time series dominate the collective behavior of the people. of population growth or limit cycles around Thus, there is a kind of feedback: the collective points of stability. Open dissipative systems of order of a complex system is generated by the ecology may become unstable because of local interactions of its elements (“self organization”). perturbations, e.g., pollution of the atmosphere, On the other hand, the behavior of the elements leading to global chaos of the atmosphere in the is dominated by the collective order. People have sense of the butterfly effect. their individual will to influence collective trends In cardiology, the heart is modeled as a com- of society. But, they are also driven by attractors plex dynamical system of electrically interacting of collective behavior. cells producing collective patterns of beating, In classical economics, an economy was be- which are then represented by time series of ECG lieved to be a conservative equilibrium system.

35 Klaus Mainzer

According to Adam Smith (1976), the market ministic, stochastic, or quantum computers, com- is self-organized by an “invisible hand,” tending puting information about present or future to the equilibrium of supply and demand. In states from initial conditions by the correspond- the age of globalization, markets are open, non- ing dynamical equations. In the case of quantum equilibrium systems at the edge of chaos (in the systems, the binary concept of information is technical sense of the word seen above), with replaced by quantum information with superposi- sensible dependence on local perturbations (but- tion of binary digits. Thus, quantum informa- terfly effect). The time series of stock markets tion only provides probabilistic forecasts of future and business cycles are examples of economic states. signals. The complex system approach offers a research Another application of social dynamics is program to bridge the gap between brain research the behavior of car drivers. In automobile traffic and cognitive science. In a famous metaphor, systems, a phase transition from nonjamming to Leibniz compared the machinery of a human jamming phases depends on the averaged car brain and body with the machinery of a mill density as control parameter. At a critical value, that can be explored inside and observed in its fluctuations with fractal or self-similar features behavior. In modern brain research, the inter- can be observed. The term self-similarity states acting cogs of the mill are the firing and nonfiring that the time series of measured traffic flow looks neurons which could be technically constructed the same on different time scales, at least from a by a neural net. If the human brain is considered qualitative point of view with small statistical as a complex dynamical system, then emergence deviations. In the theory of complex systems, of mental states can be modeled by phase trans- self-similarity is a (not sufficient) hint at chaotic itions of macroscopic order parameters which are dynamics. These signals can be used by traffic achieved by collective nonlinear interactions of guiding systems. neurons, but which are not reducible to micro- scopic states of the system: A single neuron can- not think or feel. The complex system approach is an empirical research program that can be 5 Dynamical, Information, and specified and tested in appropriate experimental Computational Systems applications to understand the dynamics of the human cognitive system. Further on, it gives heur- Dynamical systems can be characterized by in- istic devices to construct artificial systems with formation and computational concepts. A dynam- cognitive features in robotics (see Chapters 8, ical system can be considered as an information 13–16 in this volume). processing machine, computing a present state as In a dramatic step, the complex systems output from an initial state of input. Thus, the approach has been enlarged from neural networks computational efforts to determine the states of a to global computer networks like the World Wide system characterize the complexity of a dynam- Web. The internet can be considered as a com- ical system. The transition from regular to chaotic plex open computer network of autonomous nodes systems corresponds to increasing computational (hosts, routers, gateways, etc.), self-organizing problems, according to increasing degrees in the without central control mechanisms. The informa- computational theory of complexity (see Chapter tion traffic is constructed by information packets 1 in this volume, COMPUTATION). In statistical with source and destination addresses. Routers mechanics, the information flow of a dynamical are nodes of the network determining the local system describes the intrinsic evolution of statist- path of each packet by using local routing tables ical correlations. In chaotic systems with sensitiv- with cost metrics for neighboring routers. A ity to the initial states, there is an increasing loss router forwards each packet to a neighboring of information about the initial data, according router with lowest costs to the destination. As a to the decay of correlations between the entire router can only deal with one packet, other arriv- past and future states of the system. In general, ing packets at a certain time must be stored in a dynamical systems can be considered as deter- buffer. If more packets arrive than a buffer can

36 Introduction to Systems Science store, the router discards the overflowed packets. distributed in the intelligent environments of our Senders of packets wait for confirmation mess- everyday world. Like GPS (the Global Position age from the destination host. These buffering System) in car traffic, things in everyday life could and resending activities of routers can cause interact telematically by sensors. The real power congestion in the internet. A control parameter of of the concept does not come from any one of data density is defined by the propagation of these single devices. In the sense of complex sys- congestion from a router to neighboring routers tems, the power emerges from the collective inter- and dissolution of the congestion at each router. action of all of them. For instance, the optimal The cumulative distribution of congestion dura- use of energy could be considered as a macro- tion is an order parameter of phase transition. At scopic order parameter of a household realized a critical point, when the congestion propagation by the self-organizing use of different household rate is equal to congestion dissolution, fractal goods according to less consumption of electri- and chaotic features can be observed in data city during special time-periods with cheap prices. traffic. Congested buffers behave in surprising The processors, chips, and displays of these smart analogy to infected people. If a buffer is over- devices don’t need a user interface like a mouse, loaded, it tries to send packets to the neighbor- windows, or keyboards, but just a pleasant and ing routers. Therefore the congestion spreads effective place to get things done. Wireless com- spatially. On the other hand, routers can recover puting devices on small scales become more and when the congestion from and to the own subnet more invisible to the user. Ubiquitous comput- are lower than the service rate of the router. ing enables people to live, work, use, and enjoy That is not only an illustrative metaphor, but things directly without being aware of their com- hints at nonlinear mathematical models describ- puting devices. ing true epidemic processes like malaria extension What are the human perspectives in these devel- as well as the dynamics of routers. Computer net- opments of dynamical, information, and com- works are computational ecologies. The capability putational systems? Modern societies, economies, to manage the complexity of modern societies and information networks are highly dimensional depends decisively on effective communication systems with a complex nonlinear dynamics. networks. From a methodological point of view, it is a chal- The transformation of the internet into a sys- lenge to improve and enlarge the instruments of tem with self-organizing features of learning and modelization (cf. sections 1–3 above) from low- adapting is not merely a metaphor. Information to high-dimensional systems. Modern systems retrieval is already realized by neural networks science offers an interdisciplinary methodology to adapting to the information preferences of a understand the typical features of self-organizing human user with synaptic plasticity. In sociobio- dynamics in nature and society. As nonlinear logy, we can learn from populations of ants and models are applied in different fields of research, termites how to organize traffic and information we gain general insights into the predictable processing by swarm intelligence. From a tech- horizons of oscillatory chemical reactions, fluctu- nical point of view, we need intelligent programs ations of species, populations, fluid turbulence, distributed in the nets. There are already more and economic processes. The emergence of sun- or less intelligent virtual organisms (“agents”), spots, for instance, which was formerly analyzed learning, self-organizing, and adapting to our by statistical time-series methods, is by no means individual preferences of information, selecting a random activity. It can be modeled by a non- our e-mails, preparing economic transactions, linear chaotic system with several characteristic or defending against attacks by hostile com- periods and a strange attractor, allowing bounded puter viruses, like the immune system of our forecasts of the variations. In nonlinear models body. Complexity of global networking not only of public opinion formation, for instance, we means increasing numbers of PCs, workstations, may distinguish a predictable stable state before servers, and supercomputers interacting via data public voting (bifurcation) when neither of two traffic in the internet. Below the complexity of possible opinions is preferred, a short interval of a PC, low-power, cheap, and smart devices are bifurcation when tiny unpredictable fluctuations

37 Klaus Mainzer may induce abrupt changes, and a transition to a with their activities to the collective macrostate stable majority. The situation can be compared of society representing cultural, political, and eco- to growing air bubbles in turbulently boiling nomic order (order parameters). Yet, macrostates water: When a bubble has become big enough, of a society, of course, don’t simply average over its steady growth on its way upward is predict- its parts. Its order parameters strongly influ- able. But its origin and early growth is a ques- ence the individuals of the society by orientating tion of random fluctuation. Obviously, nonlinear (enslaving) their activities and by activating or modeling explains the difficulties of the modern deactivating their attitudes and capabilities. This sibyls of demoscopy. kind of feedback is typical for complex dynam- Today, nonlinear forecasting models don’t ical systems. lf the control parameters of the always deliver better and more efficient predic- environmental conditions attain certain critical tions than the standard linear procedures. Their values due to internal or external interactions, main advantage is the explanation of the actual the macrovariables may move into an unstable nonlinear dynamics in real processes, the identi- domain out of which highly divergent alternat- fication and improvement of local horizons with ive paths are possible. Tiny unpredictable micro- short-term predictions. But first of all the phase fluctuations (e.g., actions of a few influential space and an appropriate dynamical equation people, scientific discoveries, new technologies) governing a time series of observations must be may decide which of the diverging paths in an reconstructed to predict future behavior by solv- unstable state of bifurcation society will follow. ing that equation. Even in the natural sciences, So, the paradigm of a centralized control must be it is still unclear whether appropriate equations given up by the insights in the self-organizing for complex fields such as earthquakes can be dynamics of highly dimensional systems. By derived. We may hope to set up a list in a com- detecting global trends and the order para- puter memory with typical nonlinear equations meters of complex dynamics, we have the chance whose coefficients can be automatically adjusted of implementing favorite tendencies. By under- for the observed process. Instead, to make an standing complex systems we can make much exhaustive search for all possible relevant para- more progress in evaluating our information tech- meters, a learning strategy may start with a crude nologies and choosing our next steps. Under- model operating over relatively short times, and standing complex systems supports deciding and then specify a smaller number of parameters in a acting in a complex world. relatively narrow range of values. An improve- ment of short-term forecasting has been realized by the learning strategies of neural networks. On the basis of learned data, neural nets can References weight the input data and minimize the fore- casting errors of short-term stock quotations by Arnold, V. I. 1963. “Small denominators II. Proof of a theorem of A. N. Kolmogorov on the self-organizing procedures. So long as only some preservation of conditionally-periodic motions stock-market advisers use this technical support, under a small perturbation of the Hamiltonian.” they may do well. But if all agents in a market Russian Mathematical Surveys 18(5). [Proof of use the same learning strategy, the forecasting KAM theorem; graduate level.] will become a self-defeating prophecy. The reason Glansdorff, P. and Prigogine, I. 1971. Thermo- is that human societies are not complex systems dynamic Theory of Structures, Stability and Fluctu- of molecules or ants, but the result of highly ations. New York: Wiley. [Basic textbook of intentional acting beings with a greater or lesser dissipative structures; graduate level.] amount of . A particular kind of self- Grassberger, P. and Procaccia, I. 1983. “Charac- fulfilling prophecy is the Oedipus effect: like the terization of strange attractors.” Physical Revue legendary Greek king, people try, in vain, to Letters 50: 346–9. [Theorem of characterizing change their future as forecasted to them. chaotic attractor in time series; graduate level.] From a macroscopic viewpoint we may, of Haken, H. 1983. : Nonequilibrium course, observe single individuals contributing Phase Transitions and Self-organization in Physics,

38 Introduction to Systems Science

Chemistry, and Biology, 3rd ed. Berlin: Springer. Mainzer, K. 1997. Thinking in Complexity: The [Basic textbook on synergetics; undergraduate Complex Dynamics of Matter, Mind, and Man- level.] kind. 3rd ed. Berlin: Springer. [Interdisciplinary Holland, J. H. 1992. Adaption in Natural and and philosophical introduction to complex sys- Artifical Systems. Cambridge MA: MIT Press. [An tems; undergraduate level.] introduction; undergraduate level.] Moser, J. 1967. “Convergent series expansions of Kaplan, D. and Glass, L. 1995. Understanding quasi-periodic motions.” Mathematical Annals Nonlinear Dynamics. New York: Springer. [Intro- 169: 163. [Proof of KAM theorem; graduate duction to nonlinear dynamics; undergraduate level.] level.] Poincaré, H. 1892–3. Les Méthodes nouvelles de la Kolmogorov, A. N. 1954. “On conservation of mécanique célèste I–III. Paris: Gauthier-Villars. conditionally-periodic motions for a small change [Historically important source of chaos theory; in Hamilton’s function.” Dokl. Akad. Nauk SSSR graduate level.] 98: 527–30. [Proof of KAM theorem: graduate Smith, A. 1976. An into the Nature and level.] Causes of the Wealth of Nations. Chicago: Uni- Landau, L. D. and Lifshitz, E. M. 1959. Course of versity of Chicago Press. [Historically important , vol. 6: Fluid Mechanics. Lon- source on economic self-organization; under- don: Pergamon Press. [Famous fluid mechanics graduate level.] textbook; graduate level.] Takens, F. 1981. “Detecting strange attractors in Laplace, P.-S. de. 1814. Essai philosophique sur les turbulence.” In D. A. Rand and L. S. Young, probabilités. Paris. [Historically important essay; eds., Dynamical Systems and Turbulence. Berlin: undergraduate level.] Springer, pp. 336–86. [Takens’ theorem for Lorenz, E. N. 1963. “Deterministic nonperiodic detecting chaotic attractors in time series; gra- flow.” Journal of Atmospheric Science 20: 130–41. duate level.] [Detection of Lorenz’s attractor: graduate level.] Volterra, V. 1931. Leçons sur la théorie Lotka, A. J. 1925. Elements of Mathematical Biology. mathématique de la lutte pour la vie. Paris. New York: Dover. [Historically important eco- [Historically important ecological systems science logical systems science textbook; undergraduate textbook; graduate level.] level.]

39 Luciano Floridi

Chapter 4 Information Luciano Floridi

capture all major concepts of information (from 1 Introduction Shannon’s to Baudrillard’s, from genetic to neural), but also sufficiently specific to discriminate Information “can be said in many ways,” just as between semantic nuances. They attempt to show being can (, Metaphysics Γ.2), and the that all kinds of information are ultimately reduc- correlation is probably not accidental. Informa- ible conceptually, genetically, or genealogically tion, with its cognate concepts like computa- to some Ur-concept, the mother of all instances. tion, data, communication, etc., plays a key role The development of a systematic UTI is a matter in the ways we have come to understand, model, of time, patience, and ingenuity. The ultimate and transform reality. Quite naturally, informa- UTI will be hierarchical, linear (even if probably tion has adapted to some of being’s contours. branching), inclusive, and incompatible with any Because information is a multifaceted and poly- alternative model. valent concept, the question “what is informa- Reductionist strategies are unlikely to succeed. tion?” is misleadingly simple, exactly like “what Several surveys have shown no consensus or even is being?” As an instance of the Socratic ques- convergence on a single, unified definition of tion “ti esti . . . ?,” it poses a fundamental and information (see for example Braman 1989, complex problem, intrinsically fascinating and Losee 1997, Machlup 1983, NATO 1974, 1975, no less challenging than “what is truth?,” “what 1983, Schrader 1984, Wellisch 1972, Wersig is virtue?,” “what is knowledge?,” or “what is & Neveling 1975). This is hardly surprising. meaning?” It is not a request for dictionary Information is such a powerful and flexible con- explorations but an ideal point of intersection cept and such a complex phenomenon that, as of philosophical investigations, whose answers an explicandum, it can be associated with several can diverge both because of the conclusions explanations, depending on the level of abstrac- reached and because of the approaches adopted. tion adopted and the cluster of requirements and Approaches to a Socratic question can usually desiderata orientating a theory. Claude Shannon be divided into three broad groups: reductionist, (1993a: 180), for one, was very cautious: antireductionist, and nonreductionist. Philosoph- ical theories of information are no exception. The word “information” has been given dif- Reductionists support the feasibility of a ferent meanings by various writers in the gen- “unified theory of information” (UTI, see the eral field of . It is likely that UTI website for references), general enough to at least a number of these will prove sufficiently

40 Information

useful in certain applications to deserve further ant, and the “periphery” plays a counterbalancing study and permanent recognition. It is hardly role. Depending on the orientation, information to be expected that a single concept of information is seen as interpretation, power, narrative, mes- would satisfactorily account for the numerous or medium, conversation, construction, a possible applications of this general field. [italics commodity, and so on. Thus, philosophers like added] Baudrillard, Foucault, Lyotard, McLuhan, Rorty, and Derrida are united by what they dismiss, if At the opposite end, antireductionists stress not challenge: the predominance of the factual. the multifarious nature of the concept of informa- For them information is not in, from, or about tion and of the corresponding phenomena. They reality. They downplay the aboutness of informa- defend the radical irreducibility of the differ- tion and bend its referential thrust into a self- ent species to a single stem, objecting especially referential circle of hermeneutical communication. to reductionist attempts to identify Shannon’s Their classic target is Cartesian quantitative concept of information as the re- seen as the clearest expression of a hierarchical quired Ur-concept and to ground a UTI on the and authoritarian approach to the genesis, justi- mathematical theory of communication. Anti- fication, and flow of information. Disoriented, reductionist strategies are essentially negative they mistake it (Cartesian foundationalism) as the and can soon become an impasse rather than a only alternative to their fully decentralized view. solution. They allow specialized analyses of the Centralized approaches interpret the vari- various concepts of information to develop inde- ous meanings, uses, applications, and types of pendently, thus avoiding the vague generaliza- information as a system gravitating around a core tions and mistaken confusions that may burden notion with theoretical priority. The core notion UTI strategies. But their fragmented works as a hermeneutical device that influences, remains unsatisfactory insofar as it fails to account interrelates, and helps to access other notions. for the ostensible connections permeating and In metaphysics, Aristotle held a similar view about influencing the various ways in which informa- being, and argued in favor of the primacy of tion qua information “can be said.” Connections, the concept of substance. In the philosophy mind, not Wittgensteinian family resemblances. of information, this “substantial” role has long The genealogical analogy would only muddy been claimed by factual or epistemically oriented the waters here, giving the superficial impression semantic information. The basic idea is simple. of having finally solved the difficulty by merely In order to understand what information is, the hiding the actual divergences. The die-hard best thing to do is to start by analyzing it in terms reductionist would still argue that all informa- of the knowledge it can yield about its reference tion concepts descend from the same family, (the “abouted”). This epistemic approach is not while the unrepentant antireductionist would still without competitors. Weaver (1949), for example, object that we are facing mere resemblances, and supported a tripartite analysis of information in that the various information concepts truly have terms of (1) technical problems concerning the different roots. quantification of information and dealt with by Nonreductionists seek to escape the dichotomy Shannon’s theory; (2) semantic problems relat- between and antireductionism by ing to meaning and truth; and (3) what he called replacing the reductionist hierarchical model with “influential” problems concerning the impact and a distributed network of connected concepts, effectiveness of information on human behavior, linked by mutual and dynamic influences that which he thought had to play an equally import- are not necessarily genetic or genealogical. This ant role. Moreover, in pragmatic contexts, it is “hypertextual analysis” can be centralized in vari- common to privilege a view of information as ous ways or completely decentralized and per- primarily a resource for decision-making pro- haps multicentered. cesses. One of the tasks of this chapter is to show According to decentralized or multicentered how in each case the centrality of epistemically approaches, there is no key concept of informa- oriented semantic information is presupposed tion. More than one concept is equally import- rather than replaced.

41 Luciano Floridi

We are now well placed to look at the struc- in a range of codes and formats, embedded in ture of this chapter. In the following pages the physical implementations of different kinds. They question “what is information?” is approached can variously be produced, processed, commun- from a nonreductionist and epistemically cent- icated, and accessed. The Cambridge Dictionary ralized perspective. In section 2, the concept of Philosophy, for example, defines information of semantic information is reviewed assuming thus: that factual information is the most important and influential sense in which information qua an objective (mind independent) entity. It can information “can be said.” However, no attempt be generated or carried by messages (words, is made to reduce all other concepts to factual sentences) or by other products of cognizers information. Factual information is like the capital (interpreters). Information can be encoded and of the informational archipelagos, crucially posi- transmitted, but the information would exist tioned to provide both a clear grasp of what independently of its encoding or transmission. information is and a privileged gateway to other important concepts that are interconnected but Examples of information in this broad sense not necessarily reducible to a single Ur-concept. are this Guide, Edgar Allan Poe’s The Raven, To show this in practice and to enrich our under- Verlaine’s Song of Autumn, the Rosetta Stone standing of what else information may be, we and the movie Fahrenheit 451. shall look at two neighboring areas of great Over the last three decades, many analyses have importance. Section 3 summarizes the mathemat- converged on a General Definition of Informa- ical theory of communication, which studies the tion (GDI) as semantic content in terms of statistical behavior of uninterpreted data, a much- data + meaning (see Floridi forthcoming a for impoverished concept of information. Section 4 extended bibliography): outlines some important philosophical programs of research that investigate a more enriched con- GDI) σ is an instance of information, under- cept of semantic information. Space constraints stood as objective semantic content, if prevent discussion of several other important con- and only if: cepts of information, but some of them are briefly GDI.1) σ consists of n data (d), for n ≥ 1; mentioned in the conclusion. GDI.2) the data are well-formed (wfd); GDI.3) the wfd are meaningful (mwfd =δ).

GDI has become an operational standard especi- 2 Semantic Information ally in fields that treat data and information as reified entities (consider, for example, the now In this section, a general definition of semantic common expressions “data mining” and “informa- information is introduced, followed by a special tion management”). Examples include Informa- definition of factually oriented semantic infor- tion Science; Information Systems Theory, mation. The contents of the section are based Methodology, Analysis, and Design; Information on Floridi forthcoming a and c. The approach (Systems) Management; Database Design; and is loosely connected with the methodology . Recently, GDI has begun developed in situation logic (see section 3.2). to influence the philosophy of computing and information (Floridi 1999 and Mingers 1997). According to GDI, information can consist of 2.1 Semantic information as content different types of data δ. Data can be of four types (Floridi 1999): Information is often used in connection with communication phenomena to refer to objective δ.1) primary data. These are the principal (in the sense of mind-independent or external, data stored in a database, e.g. a simple and informee-independent) semantic contents. array of numbers. They are the data These can be of various size and value, formulated an information-management system is

42 Information

generally designed to convey to the user data on the occurrence of differences variously in the first place. implementable physically, explain why informa- δ.2) metadata. These are secondary indications tion can be decoupled from its support. Inter- about the nature of the primary data. They pretations of this support-independence vary describe properties such as location, format, radically because Dd leaves underdetermined not updating, availability, copyright restrictions, only the logical type to which the relata belong and so forth. (see TyN), but also the classification of the relata δ.3) operational data. These are data regard- (taxonomic neutrality), the kind of support re- ing usage of the data themselves, the quired for the implementation of their inequality operations of the whole data system and (ontological neutrality), and the dependence of the system’s performance. their semantics on a producer (genetic neutrality). δ.4) derivative data. These are data that can Consider the taxonomic neutrality (TaN) first. be extracted from δ.1–δ.3, whenever the A datum is usually classified as the entity exhib- latter are used as sources in search of pat- iting the anomaly, often because the latter is terns, clues, or inferential evidence, e.g. perceptually more conspicuous or less redundant for comparative and quantitative analyses than the background conditions. However, the (ideometry). relation of inequality is binary and symmetric. A white sheet of paper is not just the necessary GDI indicates that information cannot be data- background condition for the occurrence of a less, but it does not specify which types of data black dot as a datum, it is a constitutive part of constitute information. This typological neutral- the datum itself, together with the fundamental ity (TyN) is justified by the fact that, when relation of inequality that couples it with the dot. the apparent absence of data is not reducible to Nothing is a datum per se. Being a datum is an the occurrence of negative primary data, what external property. GDI endorses the following becomes available and qualifies as information is thesis: some further nonprimary information µ about σ constituted by some nonprimary data δ.2–δ.4. TaN) A datum is a relational entity. For example, if a database query provides an answer, it will provide at least a negative answer, So, no data without relata, but GDI is neutral e.g. “no documents found.” If the database pro- with respect to the identification of data with vides no answer, either it fails to provide any specific relata. In our example, GDI refrains from data at all, in which case no specific information identifying either the black dot or the white sheet σ is available, or it can provide some data δ to of paper as the datum. establish, for example, that it is running in a Understood as relational entities, data are loop. Likewise, silence, as a reply to a question, constraining affordances, exploitable by a sys- could represent negative information, e.g. as tem as input of adequate queries that correctly implicit assent or denial, or it could carry some semanticize them to produce information as out- nonprimary information µ, e.g. the person has put. In short, information as content can also be not heard the question. described erotetically as data + queries (Floridi Information cannot be dataless but, in the 1999). I shall return to this definition in sec- simplest case, it can consist of a single datum tion 3.2. (d). A datum is reducible to just a lack of uni- Consider now the ontological neutrality (ON). formity between two signs. So our definition of By rejecting the possibility of dataless informa- a datum (Dd) is: tion, GDI endorses the following modest thesis:

Dd) d = (x ≠ y), where the x and the y are two ON) No information without data representa- uninterpreted variables. tion.

The dependence of information on the occur- Following Landauer and Bennett 1985 and rence of syntactically well-formed data, and of Landauer 1987, 1991, and 1996, ON is often

43 Luciano Floridi interpreted materialistically, as advocating the deep bottom, in most instances – an impossibility of physically disembodied informa- immaterial source and explanation; that tion, through the equation “representation = which we call reality arises in the last physical implementation”: analysis from the posing of yes–no ques- tions and the registering of equipment- ON.1) No information without physical evoked responses; in short, that all implementation. things physical are information-theoretic in origin and that this is a participatory ON.1 is an inevitable assumption when working universe” (Wheeler 1990, 5); on the physics of computation, since computer science must necessarily take into account the and physical properties and limits of the data carriers. Thus, the debate on ON.1 has flourished especi- ON.3) “[information is] a name for the con- ally in the context of the philosophy of quantum tent of what is exchanged with the outer computing (see Landauer 1991, Deutsch 1985, world as we adjust to it, and make our 1997; Di Vincenzo & Loss 1998; Steane 1998 adjustment felt upon it.” (Wiener 1954, provides a review). ON.1 is also the ontological 17). “Information is information, not assumption behind the Physical Symbol System matter or energy. No which Hypothesis in AI and Cognitive Science (Newell does not admit this can survive at the & Simon 1976). But ON, and hence GDI, does present day” (Wiener 1961: 132). not specify whether, ultimately, the occurrence of every discrete state necessarily requires a mater- ON.2 endorses an information-theoretic, meta- ial implementation of the data representations. physical : the universe’s essential nature is Arguably, environments in which all entities, digital, being fundamentally composed of informa- properties, and processes are ultimately noetic tion as data instead of matter or energy, with (e.g. Berkeley, Spinoza), or in which the material material objects as a complex secondary mani- or extended universe has a noetic or non- festation (a similar position has been defended extended matrix as its ontological foundation (e.g. more recently in physics by Frieden 1998, whose Pythagoras, , Descartes, Leibniz, Fichte, work is based on a Platonist perspective). ON.2 Hegel), seem perfectly capable of upholding may but does not have to endorse a strictly ON without necessarily embracing ON.1. The computational view of information processes. relata in Dd could be monads, for example. ON.3 advocates a more pluralistic approach along Indeed, the classic realism debate can be recon- similar lines. Both are compatible with GDI. structed in terms of the possible interpretations A final comment concerning GDI.3 can be of ON. introduced by discussing a fourth slogan: All this explains why GDI is also consistent with two other popular slogans this time favorable to the proto-physical nature of information and ON.4) “In fact, what we mean by information hence completely antithetic to ON.1: – the elementary unit of information – is a difference which makes a differ- ON.2) “It from bit. Otherwise put, every ‘it’ ence.” (Bateson 1973: 428) – every particle, every field of force, even the space-time continuum itself – ON.4 is one of the earliest and most popular derives its function, its meaning, its very formulations of GDI (see for example Franklin existence entirely – even if in some con- 1995, 34 and Chalmers 1997: 281; note that the texts indirectly – from the apparatus- formulation in MacKay 1969, that is, “informa- elicited answers to yes-or-no questions, tion is a distinction that makes a difference,” binary choices, bits. ‘It from bit’ sym- predates Bateson’s and, although less memorable, bolizes the idea that every item of the is more accurate). A “difference” is just a dis- physical world has at bottom – a very crete state (that is, a datum), and “making a

44 Information difference” simply means that the datum is 2.2 Semantic information as “meaningful,” at least potentially. factual information Finally, let us consider the semantic nature of the data. How data can come to have an We have seen that semantic information is usu- assigned meaning and function in a semiotic ally associated with communication. Within this system in the first place is one of the hardest context, the most important type of semantic problems in semantics. Luckily, the point in information is factual information, which tells question here is not how but whether data con- the informee something about something else, stituting information as semantic content can be for example where a place is, what the time is, meaningful independently of an informee. The whether lunch is ready, or that penguins are birds. genetic neutrality (GeN) supported by GDI states Factual information has a declarative (Kant’s that: judicial) nature, is satisfactorily interpretable in terms of first-order, classic predicate logic, is cor- GeN) δ can have a semantics independently of rectly qualifiable alethically, and can be appro- any informee. priately analyzed in the following form “a’s being (of type) F carries the information that b is G” Before the discovery of the Rosetta Stone, (Dretske 1981, Barwise & Seligman 1997). Egyptian hieroglyphics were already regarded Does GDI provide a definition of factual as information, even if their semantics was bey- information? Some philosophers (Barwise & ond the comprehension of any interpreter. The Seligman 1997, Dretske 1981, Floridi forthcom- discovery of an interface between Greek and ing a and c. Grice 1989) have argued that it Egyptian did not affect the semantics of the hiero- does not, because otherwise false information glyphics but only its accessibility. This is the weak, would have to count as a type of factual informa- conditional-counterfactual sense in which GDI. tion, and there are no convincing reasons to 3 speaks of meaningful data being embedded believe it does, while there are compelling rea- in information carriers informee-independently. sons to believe that it does not (for a detailed GeN supports the possibility of information with- analysis see Floridi forthcoming a). As Dretske out an informed subject, to adapt a Popperian and Grice have put it: “false information and phrase. Meaning is not (at least not only) in the mis-information are not kinds of information – mind of the user. GeN is to be distinguished any more than decoy ducks and rubber ducks from the stronger, realist thesis, supported for are kinds of ducks” (Dretske 1981: 45) and example by Dretske (1981), according to which “False information is not an inferior kind of data could also have their own semantics inde- information; it just is not information” (Grice pendently of an intelligent producer/informer. 1989: 371). Let us look at the problem in more This is also known as environmental informa- detail. tion, and a typical example given is the series The difficulty lies here with yet another import- of concentric rings visible in the wood of a cut ant neutrality in GDI. GDI makes no comment tree trunk, which may be used to estimate its on the truthfulness of data that may comprise age. information (alethic neutrality AN): To summarize, GDI defines information broadly understood as semantic content com- AN) Meaningful and well-formed data qualify prised of syntactically well-formed and meaning- as information, no matter whether they ful data. Its four types of neutrality (TyN, TaN, represent or convey a truth or a falsehood ON, and GeN) represent an obvious advantage, or have no alethic value at all. as they make GDI perfectly scalable to more complex cases and reasonably flexible in terms of Verlaine’s Song of Autumn counts as informa- applicability and compatibility. The next question tion even if it does not make sense to ask whether is whether GDI is satisfactory when discussing it is true or false, and so does every sentence in the most important type of semantic information, Old Moore’s Almanac, no matter how downright namely factual information. false. Information as purely semantic content is

45 Luciano Floridi completely decoupled from any alethic consid- SDI.2) the data are well-formed (wfd); eration (Colburn 2000, Fetzer forthcoming, and SDI.3) the wfd are meaningful (mwfd =δ); Fox 1983 can be read as defending this perspec- SDI.4) the δ are truthful. tive). However, if GDI is also taken to define factual information, then Factual information encapsulates truthfulness, which does not contingently supervene on, but a) false information about the world (includ- is necessarily embedded in it. And since informa- ing contradictions), i.e. misinformation, be- tion is “said primarily in factual ways,” to put it comes a genuine type of factual information; in Aristotelian terms, false information can be dis- b) tautologies qualify as factual information; missed as no factual information at all, although c) “it is true that p” where p can be replaced it can still count as information in the trivial by any instance of genuine factual informa- sense of semantic content. tion, is no longer a redundant expression, e.g. “it is true” in the conjunction “‘the earth is round’ qualifies as information and 3 The Mathematical Theory of it is true” cannot be eliminated without Communication semantic loss; and finally d) it becomes impossible to erase factual Some features of information are intuitively information semantically (we shall be more quantitative. Information can be encoded, stored, and more informed about x, no matter what and transmitted. We also expect it to be additive the truth value of our data about x is). and non-negative. Similar properties of informa- tion are investigated by the mathematical theory None of these consequences is ultimately defens- of communication (MTC) with the primary aim ible, and their rejection forces a revision of GDI. of devising efficient ways of encoding and trans- “False” in “false information” is used attribut- ferring data. ively, not predicatively. As in the case of a false MTC is not the only successful mathematical constable, false information is not factual informa- approach to information theory, but it certainly tion that is false, but not factual information is the best and most widely known, and the one at all. So “false information” is, like “false evid- that has had the most profound impact on philo- ence,” not an oxymoron, but a way of specifying sophical analyses. The name for this branch of that the informational contents in question do probability theory comes from Shannon’s seminal not conform to the situation they purport to work (Shannon 1948, now Shannon & Weaver map (or “to about”), and so fail to qualify as 1998). Shannon pioneered this field and obtained factual information. Well-formed and meaning- many of its principal results, but he acknow- ful data may be of poor quality. Data that are ledged the importance of previous work done by incorrect (vitiated by errors or inconsistencies), other researchers at Bell laboratories, most not- imprecise (precision is a measure of the repeatab- ably Nyquist and Hartley (see Cherry 1978 and ility of the collected data), or inaccurate (accuracy Mabon 1975). After Shannon, MTC became refers to how close the average data value is to known as information theory, an appealing but the actual value) are still data and may be re- unfortunate label, which continues to cause end- coverable. But, if they are not truthful, they can less misunderstandings. Shannon came to regret only amount to semantic content at best and its widespread popularity, and we shall avoid misinformation at worst. using it in this context. The special definition of information (SDI) This section outlines some of the key ideas needs to include a fourth condition about the behind MTC, with the aim of understanding positive alethic nature of the data in question: the relation between MTC and the philosophy of information. The reader with no taste for SDI) σ is an instance of factual information mathematical formulae may wish to go directly if and only if: to section 3.2, where some implications of SDI.1) σ consists of n data (d), for n ≥ 1; MTC are discussed. The reader interested in

46 Information

ALPHABET

Sent signal Received signal

INFORMER ENCODING DECODING INFORMEE Information TRANSMITTER RECEIVER Information

CHANNEL

NOISE Information source

Message INFORMANT

Figure 4.1: Communication model (adapted from Shannon 1948, 1998) knowing more can start by reading Weaver 1949 knowledge, it is obvious that a unary device pro- and Shannon 1993b, then Schneider 2000, duces zero amount of information. Simplifying, Pierce 1980, and Jones 1979, and finally Cover we already know the outcome so our ignorance & Thomas 1991. cannot be decreased. Whatever the informational state of the system, asking appropriate questions to the raven does not make any difference. Note 3.1 The quantification of raw that a unary source answers every question all information the time with only one symbol, not with silence or symbol, since silence counts as a signal, as we MTC has its origin in the field of electrical com- saw in section 2.1. It follows that a completely munication, as the study of communication limits. silent source also qualifies as a unary source. It develops a quantitative approach to informa- Consider now a binary device that can pro- tion as a means to answer two fundamental prob- duce two symbols, like a fair coin A with its two lems: the ultimate level of data compression and equiprobable symbols {h, t}; or, as Matthew 5:37 the ultimate rate of data transmission. The two suggests, “Let your communication be Yea, yea; solutions are the entropy H in equation (4.9) Nay, nay: for whatsoever is more than these (see below) and the channel capacity C. The rest cometh of evil.” Before the coin is tossed, the of this section illustrates how to get from the informee (for example a computer) is in a state problems to the solutions. of data deficit greater than zero: the informee Imagine a very boring device that can pro- does not “know” which symbol the device will duce only one symbol, like Poe’s raven, who can actually produce. Shannon used the technical answer only “nevermore.” This is called a unary term “uncertainty” to refer to data deficit. In a device. Even at this elementary level, Shannon’s nonmathematical context this is a misleading term simple model of communication applies (see because of its strongly semantic connotations, figure 4.1). The raven is the informer, we are especially from a Cartesian perspective. Recall the informee, “nevermore” is the message (the that the informee can be a very simple machine, informant), there is a coding and decoding pro- and psychological, mental, or doxastic states are cedure through a language (English), a channel clearly irrelevant. Once the coin has been tossed, of communication, and some possible noise. the system produces an amount of raw informa- Informer and informee share the same tion that is a function of the possible outputs, in background knowledge about the collection of this case 2 equiprobable symbols, and equal to usable symbols (the alphabet). Given this a priori the data deficit that it removes.

47 Luciano Floridi

Table 4.1: Examples of communication devices and their informational power

Device Alphabet Bits of information per symbol

Poe’s raven (unary) 1 symbol log(1) = 0 1 coin (binary) 2 equiprobable symbols log(2) = 1 2 coins 4 equiprobable symbols log(4) = 2 1 die 6 equiprobable symbols log(6) = 2.58 3 coins 8 equiprobable symbols log(8) = 3

Let us build a slightly more complex system, units in bits. The base is partly a matter of con- made of two fair coins A and B. The AB system vention, like using centimeters instead of inches, can produce 4 ordered outputs: , , partly a matter of convenience, since it is useful , . It generates a data deficit of 4 when dealing with digital devices that use binary units, each couple counting as a symbol in the codes to represent data. Given an alphabet of source alphabet. In the AB system, the occur- N equiprobable symbols, we can rephrase some rence of each symbol <_, _> removes a higher examples more precisely (table 4.1) by using data deficit than the occurrence of a symbol in equation (4.1): the A system. In other words, each symbol pro- = vides more raw information. Adding an extra log2(N ) bits of information per symbol coin would produce a 8 units of data deficit, (4.1) further increasing the amount of information carried by each symbol in the ABC system, and The basic idea is all in equation (4.1). Raw in- so on. formation can be quantified in terms of decrease We are ready to generalize the examples. Call in data deficit (Shannon’s uncertainty). Unfortun- the number of possible symbols N. For N = 1, ately, real coins are always biased. To calculate the amount of information produced by a unary how much information they produce one must device is 0. For N = 2, by producing an rely on the frequency of the occurrences of equiprobable symbol, the device delivers 1 unit symbols in a finite series of tosses, or on their of information. And for N = 4, by producing an probabilities, if the tosses are supposed to go on equiprobable symbol the device delivers the sum indefinitely. Compared to a fair coin, a slightly of the amount of information provided by coin biased coin must produce less than 1 bit of A plus the amount of information provided by information, but still more than 0. The raven coin B, that is, 2 units of information, although produced no information at all because the the total number of symbols is obtained by occurrence of a string S of “nevermore” was not multiplying A’s symbols by B’s symbols. Now, informative (not surprising, to use a more intu- our information measure should be a continuous itive but psychologistic vocabulary), and that and monotonic function of the probability of is because the probability of the occurrence of the symbols. The most efficient way of satisfying “nevermore” was maximum, so overly predict- these requirements is by using the logarithm to able. Likewise, the amount of raw information the base 2 of the number of possible symbols produced by the biased coin depends on the (the logarithm to the base 2 of a number n is average informativeness (also known as average the power to which 2 must be raised to give the surprisal, another unfortunate term to refer to = 3 = number n, for example log28 3, since 2 8). the average statistical rarity) of the string S of h Logarithms have the useful property of turn- and t produced by the coin. The average informat- ing multiplication of symbols into addition of iveness of the resulting string S depends on the information units. By taking the logarithm to probability of the occurrence of each symbol. the base 2 (henceforth log simply means log2) The higher the frequency of a symbol in S, the we have the further advantage of expressing the less raw information is being produced by the

48 Information coin, up to the point when the coin is so biased h type of symbol occurs added to the numbers to produce always the same symbol and stops of times the t type of symbol occurs. Generaliz- being informative, behaving like the raven. So, ing for i types of symbols: to calculate the average informativeness of S we need to know how to calculate S and the N = informativeness of an i th symbol in general. This SS∑ i (4.5) i =1 requires understanding what the probability of an i th symbol (P) to occur is. i Putting together equations (4.4) and (4.5) we The probability P of the i th symbol can be i see that the average informativeness for a string “extracted” from equation (4.1), where it is of S symbols is the sum of the informativeness of embedded in log(N ), a special case in which each symbol divided by the sum of all symbols: the symbols are equiprobable. Using some ele- mentary properties of the logarithmic function N we have: ∑ Suii = i 1 (4.6) ⎛ 1 ⎞ N log(N ) =−log(N −1) =−log ⎜ ⎟ =−log(P) ∑ S ⎝ ⎠ i N i =1 (4.2) Formula (4.6) can be simplified thus: The value of 1/N = P can range from 0 to 1. If the raven is our source, the probability of “good N Si morning” is 0. In the case of the coin, P(h) + ∑ ui (4.7) = S P(t) = 1, no matter how biased the coin is. Prob- i 1 ability is like a cake that gets sliced more and Now S /S is the frequency with which the i th more thinly depending on the number of guests, i symbol occurs in S when S is finite. If the length but never grows beyond its original size. More of S is left undetermined (as long as one wishes), formally: then the frequency of the i th symbol becomes its probability P . So, further generalizing formula N i = (4.7) we have: ∑ Pi 1 (4.3) i =1 N ∑ Pu (4.8) The sigma notation simply means that if we add ii i =1 all probabilities values from i = 1 to i = N the sum is equal to 1. Finally, by using equation (4.4) we can substi- We can now be precise about the raven: “never- tute for ui and obtain more” is not informative at all because Pnevermore = 1. Clearly, the lower the probability of occur- N rence of a symbol, the higher is the informative- =− H ∑Pi log Pi (bits per symbol) (4.9) ness of its actual occurrence. The informativeness i =1 u of an i th symbol can be expressed by analogy with −log (P) in equation (4.2): Equation (4.9) is Shannon’s formula for H = uncertainty, which we have called data deficit =− ui log(Pi) (4.4) (actually, Shannon’s original formula includes a positive constant K which amounts to a choice Next, we need to calculate the length of a gen- of a unit of measure, bits in our case; apparently, eral string S. Suppose that the biased coin, tossed Shannon used the letter H because of R. V. L. 10 times, produces the string: . The (length of the) string S (in our case ates that the quantity of raw information pro- equal to 10) is equal to the number of times the duced by a device corresponds to the amount of

49 Luciano Floridi data deficit erased. It is a function of the average its symbols is to associate each of them with two informativeness of the (potentially unlimited) digits: string of symbols produced by the device. It is easy to prove that, if symbols are equiprobable, = 00 (4.9) reduces to (4.1) and that the highest quan- = 01 tity of raw information is produced by a system = 10 whose symbols are equiprobable (compare the = 11 fair coin to the biased one). To arrive at (4.9) we have used some very Call this Code 1. In Code 1 a message conveys simple examples: a raven and a handful of coins. 2 bits of information, as expected. Do not con- Things in life are far more complex. For example, fuse bits as bi-nary units of information (recall we have assumed that the strings of symbols are that we decided to use log2 also as a matter of ergodic: the probability distribution for the occur- convenience) with bits as bi-nary digits, which is rences of each symbol is assumed to be stable what a 2-symbols system like a CD-ROM uses through time and independently of the selec- to encode a message. Suppose now that the AB tion of a certain string. Our raven and coins are system is biased, and that the four symbols occur discrete and zero-memory sources. The successive with the following probabilities: symbols they produce are statistically independ- ent. But in real life occurrences of symbols are = 0.5 often interdependent. Sources can be non-ergodic = 0.25 and have a memory. Symbols can be continuous, = 0.125 and the occurrence of one symbol may depend = 0.125 upon a finite number n of preceding symbols, in which case the string is known as a Markov This system produces less information, so by chain and the source an nth order Markov source. using Code 1 we would be wasting resources. A Consider for example the probability of being more efficient Code 2 should take into account sent an “e” before or after having received the the symbols’ probabilities, with the following string “welcom.” And consider the same example outcomes: through time, in the case of a child learning how to spell English words. In brief, MTC develops = 0 0.5 × 1 binary digit = .5 the previous analysis to cover a whole variety of = 10 0.25 × 2 binary digits = .5 more complex cases. We shall stop here, how- = 110 0.125 × 3 binary digits = .375 ever, because in the rest of this section we need = 111 0.125 × 3 binary digits = .375 to concentrate on other central aspects of MTC. The quantitative approach just sketched plays In Code 2, known as Fano Code, a message a fundamental role in coding theory (hence in conveys 1.75 bits of information. One can prove cryptography) and in data storage and transmis- that, given that probability distribution, no other sion techniques. Recall that MTC is primarily a coding system will do better than Fano Code. study of the properties of a channel of commun- On the other hand, in real life a good codification ication and of codes that can efficiently encipher is also modestly redundant. Redundancy refers to data into recordable and transmittable signals. the difference between the physical representation Since data can be distributed either in terms of of a message and the mathematical representa- here/there or now/then, diachronic communica- tion of the same message that uses no more bits tion and synchronic analysis of a memory can than necessary. Compression procedures work by be based on the same principles and concepts reducing data redundancy, but redundancy is not (our coin becomes a bi-stable circuit or flip-flop, always a bad thing, for it can help to counteract for example), two of which are so important to equivocation (data sent but never received) and deserve a brief explanation: redundancy and noise. noise (received but unwanted data). A message + Consider our AB system. Each symbol occurs noise contains more data than the original mess- with 0.25 probability. A simple way of encoding age by itself, but the aim of a communication

50 Information process is fidelity, the accurate transfer of the equivocation). If H > C it is possible to encode original message from sender to receiver, not the source so that the equivocation is less than data increase. We are more likely to reconstruct a H − C +ε where ε is arbitrarily small. There message correctly at the end of the transmission is no method of encoding which gives an if some degree of redundancy counterbalances equivocation less than H − C. (Shannon & the inevitable noise and equivocation introduced Weaver 1998: 71) by the physical process of communication and the environment. Noise extends the informee’s Roughly, if the channel can transmit as much or freedom of choice in selecting a message, but it more information than the source can produce, is an undesirable freedom and some redundancy then one can devise an efficient way to code and can help to limit it. That is why, in a crowded transmit messages with as small an error probab- pub, you shout your orders twice and add some ility as desired. These two fundamental theorems gestures. are among Shannon’s greatest achievements. And We are now ready to understand Shannon’s with our message finally sent, we may close this two fundamental theorems. Suppose the 2-coins section. biased system produces the following message: . Using Fano Code we obtain: 11001111010. The next step 3.2 Some conceptual implications is to send this string through a channel. Channels of MTC have different transmission rates (C), calculated in terms of bits per second (bps). Shannon’s For the mathematical theory of communication fundamental theorem of the noiseless channel (MTC), information is only a selection of one states that symbol from a set of possible symbols, so a simple way of grasping how MTC quantifies raw Let a source have entropy H (bits per symbol) information is by considering the number of yes/ and a channel have a capacity C (bits per sec- no questions required to guess what the source ond). Then it is possible to encode the output is communicating. One question is sufficient to of the source in such a way as to transmit at guess the output of a fair coin, which therefore the average rate of C/H −ε symbols per sec- is said to produce 1 bit of information. A 2-fair- ond over the channel where ε is arbitrarily coins system produces 4 ordered outputs (, , , ) and therefore requires average rate greater than C/H. (Shannon & two questions, each output containing 2 bits of Weaver 1998: 59) information, and so on. This erotetic analysis clarifies two important points. In other words, if you devise a good code you First, MTC is not a theory of information in can transmit symbols over a noiseless channel at the ordinary sense of the word. The expression an average rate as close to C/H as one may wish “raw information” has been used to stress the but, no matter how clever the coding is, that fact that in MTC information has an entirely average can never exceed C/H. We have already technical meaning. Consider some examples. Two seen that the task is made more difficult by equiprobable “yes” contain the same quantity of the inevitable presence of noise. However, the raw information, no matter whether their cor- fundamental theorem for a discrete channel with responding questions are “would you like some noise comes to our rescue: tea?” or “would you marry me?” If we knew that a device could send us with equal probab- Let a discrete channel have the capacity C and ilities either the movie Fahrenheit 451 or this a discrete source the entropy per second H. whole Guide, by receiving one or the other we If H ≤ C there exists a coding system such would receive many bytes of data but only one that the output of the source can be transmit- bit of raw information. On June 1, 1944, the ted over the channel with an arbitrarily small BBC broadcasted a line from Verlaine’s Song of frequency of errors (or an arbitrarily small Autumn: “Les sanglots longs des violons de

51 Luciano Floridi

Autumne.” The message contained almost 1 bit messages constituted by uninterpreted symbols of information, an increasingly likely “yes” to encoded in well-formed strings of signals, it is the question whether the D-Day invasion was commonly described as a study of information imminent. The BBC then broadcasted the sec- at the syntactic level. MTC can be successfully ond line “Blessent mon coeur d’une longueur applied in ICT (information and communica- monotone.” Another almost meaningless string tion technologies) because computers are syntact- of letters, but almost another bit of information, ical devices. What remains to be clarified is how since it was the other long-expected “yes” to the H in equation (4.9) should be interpreted. question whether the invasion was to take place Assuming the ideal case of a noiseless channel immediately. German intelligence knew about of communication, H is a measure of three the code, intercepted those messages, and even equivalent quantities: notified Berlin, but the high command failed to alert the Seventh Army Corps stationed in a) the average amount of raw information per Normandy. Hitler had all the information in symbol produced by the informer, or Shannon’s sense of the word, but failed to under- b) the corresponding average amount of data stand the real meaning and importance of those deficit (Shannon’s uncertainty) that the two small bits of data. As for ourselves, we were informee has before the inspection of the not surprised to conclude that the maximum output of the informer, or amount of raw information is produced by a c) the corresponding informational potential- text where each character is equally distributed, ity of the same source, that is, its informa- that is by a perfectly random sequence. tional entropy. Second, since MTC is a theory of informa- tion without meaning, and information minus H can equally indicate (a) or (b) because, by meaning = data, mathematical theory of data selecting a particular alphabet, the informer auto- communication is a far more appropriate descrip- matically creates a data deficit (uncertainty) in the tion than information theory. In section 2.1, we informee, which then can be satisfied (resolved) saw that information as semantic content can in various degrees by the informant. Recall the also be described erotetically as data + queries. erotetic game. If you use a single fair coin, I Imagine a piece of information such as “the earth immediately find myself in a 1-bit deficit predica- has only one moon.” It is easy to polarize almost ment. Use two fair coins and my deficit doubles, all its semantic content by transforming it into a but use the raven, and my deficit becomes null. query + binary answer: “does the earth have only My empty glass is an exact measure of your capa- one moon? + yes.” Subtract the “yes” and you are city to fill it. Of course, it makes sense to talk of left with virtually all the semantic content, fully raw information as quantified by H only if one de-alethicized (the query is neither true nor false). can specify the probability distribution. The datum “yes” works as a key to unlock the Regarding (c), MTC treats raw information information contained in the query. MTC studies like a physical quantity, such as mass or energy, the codification and transmission of raw informa- and the closeness between equation (4.9) and the tion by treating it as data keys, as the amount of formulation of the concept of entropy in statist- details in a signal or message or memory space ical mechanics was already discussed by Shannon. necessary to unlock the informee’s knowledge. As The informational and the thermodynamic con- Weaver (1949: 12) remarked, “the word informa- cepts of entropy are related through the concepts tion relates not so much to what you do say, as of probability and randomness (“randomness” is to what you could say. MTC deals with the car- better than “disorder” since the former is a syn- riers of information, symbols and signals, not tactical concept whereas the latter has a strongly with information itself. That is, information is semantic value), entropy being a measure of the the measure of your freedom of choice when amount of “mixed-up-ness” in processes and sys- you select a message.” tems bearing energy or information. Entropy can Since MTC deals not with information itself also be seen as an indicator of reversibility: if but with the carriers of information, that is, there is no change of entropy then the process is

52 Information reversible. A highly structured, perfectly organ- MTC, with the aim of incorporating the semantic ized message contains a lower degree of entropy dimension. Donald M. MacKay (1969) proposed or randomness, less raw information, and causes a quantitative theory of qualitative information a smaller data deficit – consider the raven. The that has interesting connections with situation higher the potential randomness of the symbols logic (see below), whereas Doede Nauta (1972) in the alphabet, the more bits of information developed a semiotic-cybernetic approach. Nowa- can be produced by the device. Entropy assumes days, few philosophers follow these lines of its maximum value in the extreme case of uni- research. The majority agree that MTC provides a form distribution. Which is to say that a glass of rigorous constraint to any further theorizing on water with a cube of ice contains less entropy all the semantic and pragmatic aspects of informa- than the glass of water once the cube has melted, tion. The disagreement concerns the crucial issue and a biased coin has less entropy than a fair coin. of the strength of the constraint. At one extreme In thermodynamics, we know that the greater of the spectrum, a theory of semantic informa- the entropy, the less available the energy. This tion is supposed to be very strongly constrained, means that high entropy corresponds to high perhaps even overdetermined, by MTC, some- energy deficit, but so does entropy in MTC: what as mechanical engineering is by Newtonian higher values of H correspond to higher quant- physics. Weaver’s interpretation of Shannon’s ities of data deficit. work is a typical example. At the other extreme, a theory is supposed to be only weakly constrained, perhaps even completely underdetermined, by MTC, somewhat as tennis is constrained by 4 Some Philosophical Approaches Newtonian physics, that is, in the most uninter- to Semantic Information esting, inconsequential, and hence disregardable sense (see for example Sloman 1978 and Thagard The mathematical theory of communication 1990). The emergence of MTC in the 1950s approaches information as a physical phenomenon. generated earlier philosophical enthusiasm that Its central question is whether and how much has gradually cooled down through the decades. uninterpreted data can be encoded and trans- Historically, philosophical theories of semantic mitted efficiently by means of a given alphabet information have moved from “very strongly con- and through a given channel. MTC is not inter- strained” to “only weakly constrained,” becom- ested in the meaning, aboutness, relevance, use- ing increasingly autonomous from MTC (for a fulness, or interpretation of information, but only review, see Floridi forthcoming b). in the level of detail and frequency in the un- Popper (1935) is often credited as the first interpreted data, being these symbols, signals philosopher to have advocated the inverse rela- or messages. On the other hand, philosophical tion between the probability of p and the amount approaches seek to give an account of informa- of semantic information carried by p. However, tion as semantic content, investigating questions systematic attempts to develop a formal calculus like “how can something count as information? were made only after Shannon’s breakthrough. and why?,” “how can something carry informa- MTC defines information in terms of probab- tion about something else?,” “how can semantic ility space distribution. Along similar lines, the information being generated and flow?,” “how probabilistic approach to semantic information is information related to error, truth and know- defines the semantic information in p in terms of ledge?,” “when is information useful?” Philo- logical probability space and the inverse relation sophers usually adopt a propositional orientation between information and the probability of p. and an epistemic outlook, endorsing, often im- This approach was initially suggested by Bar- plicitly, the prevalence of the factual (they analyze Hillel and Carnap (Bar-Hillel & Carnap 1953, examples like “The Bodleian library is in Ox- Bar-Hillel 1964) and further developed by ford”). How relevant is MTC to similar analyses? Hintikka (especially Hintikka & Suppes 1970) In the past, some research programs tried and Dretske 1981 (on Dretske’s approach see to elaborate information theories alternative to also Chapters 16 and 17, on MEANING and on

53 Luciano Floridi

KNOWLEDGE). The details are complex but the The informational content of p is not determined original idea is simple. The semantic content a priori, through a calculus of possible states (CONT) in p is measured as the complement of allowed by a representational language, but in the a priori probability of p: terms of factual content that p carries with respect to a given situation. Information tracks possible CONT(p) = 1 − P(p) (4.10) transitions in a system’s states space under normal conditions. Both Dretske and situation theories CONT does not satisfy the two requirements of require some presence of information already additivity and conditionalization, which are immanent in the environment (environmental satisfied by another measure, the informativeness information), as nomic regularities or constraints. (INF) of p, which is calculated, following equa- This “semantic ” can be controversial tions (4.9) and (4.10), as the reciprocal of P(p), both epistemologically and ontologically. Finally, expressed in bits, where P(p) = 1 − CONT(p) : the inferential approach defines information in terms of entailment space: information depends 1 on valid inference relative to a person’s theory INF(p) = log =− logPp ( ) (4.11) 1 − CONT or epistemic state. Most approaches close to MTC assume the Things are complicated by the fact that the con- principle of alethic neutrality, and run into the cept of probability employed in equations (4.10) difficulties I outlined in section 2.2 (Dretske and (4.11) is subject to different interpretations. and Barwise are important exceptions; Devlin In Bar-Hillel and Carnap the probability distribu- rejects truthfulness as a necessary condition). As tion is the outcome of a logical construction of a result, the semantic approach (Floridi forth- atomic statements according to a chosen formal coming a and c) adopts SDI and defines factual language. This introduces a problematic reliance information in terms of data space. on a strict correspondence between observational Suppose there will be exactly three guests for and formal language. In Dretske, the solution is dinner tonight. This is our situation w. Imagine to make probability values refer to states of affairs that you are told that (s) of the world observed: T) there may or may not be some guests for I(s) =−log P(s) (4.12) dinner tonight; or V) there will be some guests tonight; or The modal approach modifies the probabilistic P) there will be three guests tonight. approach by defining semantic information in terms of modal space and in/consistency. The The degree of informativeness of T is zero because, information conveyed by p becomes the set of as a tautology, T applies both to w and to ¬w. V all possible worlds or (more cautiously) the set performs better, and P has the maximum degree of all the descriptions of the relevant possible of informativeness because, as a fully accurate, states of the universe that are excluded by p. precise, and contingent truth, it “zeros in” on its The systemic approach, developed especially in target w. Generalizing, the more distant a true situation logic (Barwise & Perry 1983, Israel & σ is from its target w, the larger is the number Perry 1990, Devlin 1991; Barwise & Seligman of situations to which it applies, the lower its 1997 provide a foundation for a general theory degree of informativeness becomes. A tautology of information flow) also defines information in is a true σ that is most “distant” from the world. terms of states space and consistency. However, Let us use the letter ϑ to refer to the distance it is less ontologically demanding than the between a true σ and w. Using the more precise modal approach, since it assumes a clearly limited vocabulary of situation logic, ϑ indicates the domain of application, and it is compatible with degree of support offered by w to σ. We can Dretske’s probabilistic approach, although it does now map on the x axis the values of ϑ given a not require a probability measure on sets of states. specific σ and a corresponding target w. In our

54 Information

y

ι(σ) = 1 −ϑ(σ)2 1

0.75

0.5

0.25

x

−1 −0.75 −0.5 −0.25 0.25 0.5 0.75 1 ϑ(σ)

Figure 4.2: degree of informativeness (ι) of σ example, we know that ϑ(T) = 1 and ϑ(P) = 0. that is, the definite integral of the function ι(σ) For the sake of simplicity, let us assume that on the interval [0, 1]. As we know, the max- ϑ(V) = 0.25 (see Floridi forthcoming c on how imum quantity of semantic information (call it to calculate ϑ values). We now need a formula to α) is carried by P, whose ϑ= 0. This is equivalent calculate the degree of informativeness ι of σ in to the whole area delimited by the curve. Gen- relation to ϑ(σ). It can be shown that the most eralizing to σ we have: elegant solution is provided by the complement of the square value of ϑ(σ), that is y = 1 − x 2. 1 2 ισ ()dx == α (4.14) Using the symbols jut introduced we have: 0 3

ι(σ) = 1 −ϑ(σ)2 (4.13) Figure 4.3 shows the graph generated by equa- tion (4.14). The shaded area is the maximum Figure 4.2 shows the graph generated by equa- amount of semantic information α carried by σ. tion (4.13) when we also include negative values Consider now V, “there will be some guests of distance for false σ (ϑ ranges from −1 = tonight.” V can be analyzed as a (reasonably contradiction to 1 = tautology). finite) string of disjunctions, that is V = [“there If σ has a very high degree of informativeness will be one guest tonight” or “there will be two ι (very low ϑ) we want to be able to say that it guests tonight” or . . . “there will be n guests contains a large quantity of semantic informa- tonight”], where n is the reasonable limit we tion and, vice versa, the lower the degree of wish to consider (things are more complex than informativeness of σ is, the smaller the quantity this, but here we only need to grasp the general of semantic information conveyed by σ should principle). Only one of the descriptions in V be. To calculate the quantity of semantic informa- will be fully accurate. This means that V also tion contained in σ relative to ι(σ) we need to contains some (perhaps much) information that calculate the area delimited by equation (4.13), is simply irrelevant or redundant. We shall refer

55 Luciano Floridi

y ι(σ) = 1 −ϑ(σ)2

1

0.75 1 2 ι (σ) dx =α= 0 3

0.5

0.25

x

−1 −0.75 −0.5 −0.25 0.25 0.5 0.75 1 ϑ(σ)

Figure 4.3: maximum amount of semantic information α carried by σ

to this “informational waste” in V as vacuous Note the similarity between 4.14 and 4.16. When information in V. The amount of vacuous ϑ(σ) = 1, that is, when the distance between σ information (call it β) in V is also a function of and w is maximum, then α=β and γ(σ) = 0. the distance ϑ of V from w, or more generally This is what happens when we consider T. T is so distant from w as to contain only vacuous ϑ information. In other words, T contains as ισ ()dx = β (4.16) 0 much vacuous information as P contains relevant information. Since ϑ(V) = 0.25, we have A final comment, before closing this section. Each of the previous extentionalist approaches 025. can be given an intentionalist interpretation by ι ()V dx = 0 . 24479 (4.17) 0 considering the relevant space as a doxastic space, in which information is seen as a reduction in Figure 4.4 shows the graph generated by equa- the degree of personal uncertainty given a state tion (4.17). The shaded area is the amount of of knowledge of the informee. vacuous information β in V. Clearly, the amount of semantic information in V is simply the difference between α (the maximum amount 5 Conclusion of information that can be carried in principle by σ) and β (the amount of vacuous informa- tion actually carried by σ), that is, the clear area In this chapter, we have been able to visit only a in the graph of figure 4.4. More generally, the few interesting places. The connoisseur might amount of semantic information γ in σ is: be disappointed and the supporter of some local interests appalled. To try to appease both and to γ(σ) = (α−β) (4.18) whet the appetite of the beginner here is a list of

56 Information

y ι(σ) = 1 −ϑ(σ)2

1 0.25 ι (V) dx = 0.24479 0 0.75

0.5 γ=α−β 0.25

x

−1 −0.75 −0.5 −0.25 0.25 0.5 0.75 1 ϑ(σ)

Figure 4.4: Amount of semantic information γ carried by σ some very important concepts of information that where characteristics such as relevance, timeliness, have not been discussed: updatedness, cost, significance, and so forth are crucial (see Chapter 22, on GAME THEORY); informational complexity (Kolmogorov and valuable information in ethical contexts (see Chaitin, among others), a measure of the com- Chapter 5, on COMPUTER ETHICS, and Floridi plexity of a string of data defined in terms of the 2003); length of the shortest binary program required to compute that string. Note that Shannon’s H environmental information, that is, the possible can be considered a special case of Kolmogorov location and nature of information in the world complexity K, since H ≈ K if the sequence is (Dretske 1981 and see Chapters 11–13, on drawn at random from a probability distribution ONTOLOGY, on VIRTUAL REALITY, and on THE with entropy = H; PHYSICS OF INFORMATION, respectively); instructional information (imagine a recipe, an physical information and the relation between algorithm, or an order), a crucial concept in fields being and information (see Leff & Rex 1990 like computer science, genetics, biochemistry, and again Chapters 11–13); neuroscience, cognitive science, and AI (see biological information (see Chapter 15, on Chapters 1 and 2, on COMPUTATION and on ARTIFICIAL LIFE). The biologically minded reader COMPLEXITY); will notice that the 4 symbols in the AB system pragmatic information, central in any theory we built in section 3.1 could be adenine, guanine, addressing the question of how much informa- cytosine, and thymine, the four bases whose order tion a certain informant carries for an informee in the molecular chain of DNA or RNA codes in a given doxastic state and within a specific genetic information. informational environment. This includes useful information, a key concept in economics, informa- The nature of these and other information tion management theory, and decision theory, concepts, the analysis of their interrelations and

57 Luciano Floridi of their possible dependence on MTC, and the UTI, the Unified Theory of Information website, investigation of their usefulness and influence in contains documents, and links about the develop- the discussion of philosophical problems are some ment of UTI: of the crucial issues that a philosophy of informa- very interesting and important work to do.

References Acknowledgments Bar-Hillel, Y. 1964. Language and Information. Reading, MA and London: Addison Wesley. This chapter is based on Floridi 2003 and forth- [Collection of influential essays on semantic coming a. I am very grateful to Mark Bedau, information, graduate level.] John Collier, Phil Fraundorf, Ken Herold, James —— and Carnap, R. 1953. “An outline of a theory Fetzer, and Kia Nobre for their very valuable of semantic information,” repr. in Bar-Hillel comments on earlier drafts. 1964: 221–74. [One of the first and most influ- ential attempts to develop a quantitative analysis of semantic information, graduate level.] Some Web Resources Barwise, J. and Perry, J. 1983. Situations and Atti- tudes. Cambridge, MA: MIT Press. [Influential There are many useful resources freely available text in situation logic, graduate level.] —— and Seligman, J. 1997. Information Flow: The on the web, of which the following have been Logic of Distributed Systems. Cambridge: Cam- used in writing this chapter: bridge University Press. [Innovative approach to information flow, graduate level, but the modular Feldman D., A Brief Tutorial on Information structure contains some very accessible chapters.] Theory, Excess Entropy and Statistical Complex- Bateson, G. 1973. Steps to an Ecology of Mind. ity: of the ecological approach to mind and informa- Fraundorf P., Information-Physics on the Web: tion, undergraduate level.] munications Policy 13: 233–42. Chalmers, D. J. 1997. The Conscious Mind: In Search Introduction to Information Theory, by Lucent of a Fundamental Theory. Oxford: Oxford Univer- Technologies Bell Labs : Cherry, C. 1978. On Human Communication, MacKay J. C., A Short Course in Information 3rd ed. Cambridge, MA: MIT Press. [Clear and Theory: accessible introduction to communication theory, Colburn, T. R. 2000. “Information, thought, and Shannon C. E. 1948. A Mathematical Theory of knowledge.” Proceedings of the World Multicon- Communication: ference on Systemics, Cybernetics and Informatics [The classic text on the as justified true belief from an informational per- mathematical theory of information, graduate spective, very accessible.] level.] Cover, T. and Thomas, J. A. 1991. Elements of Schneider T. 2000. Information Theory Primer – Information Theory. New York: Chichester, Wiley. With an Appendix on Logarithms: [Standard textbook in the field, requires a solid [A very clear and accessible see Jones 1979 for a more accessible text, or introduction, undergraduate level.] Pierce 1980.]

58 Information

Deutsch, D. 1985. “Quantum theory, the Church– Minds and Machines. Preprint available at . Computer.” Proceedings of the Royal Society [Develops a truth-based approach to semantic 400: 97–117. [Information and computation in information, graduate level.] quantum computing, requires a solid mathemat- Fox, C. J. 1983. Information and Misinformation ical background, graduate level only.] – An Investigation of the Notions of Information, ——. 1997. The Fabric of Reality. London: Penguin. Misinformation, Informing, and Misinforming. [On the ontological implications of quantum Westport, CN: Greenwood Press. [Analysis of physics, advanced undergraduate level.] information based on information science, under- Devlin, K. 1991. Logic and Information. Cam- graduate level.] bridge: Cambridge University Press. [Reviews and Franklin, S. 1995. Artificial Minds. Cambridge, MA: improves on situation logic, undergraduate level.] MIT Press. [Undergraduate level.] Di Vincenzo, D. P. and Loss, D. 1998. “Quantum Frieden, B. R. 1998. Physics from Fisher Informa- information is physical.” Superlattices and Micro- tion: A Unification. Cambridge: Cambridge Uni- structures 23: 419–32; special issue on the occa- versity Press. [Controversial attempt to provide sion of Rolf Landauer’s 70th birthday; also an interpretation of physics in terms of informa- available at . [Reviews the debate on the physical graduate level only.] aspects of information, graduate level.] Grice, P. 1989. Studies in the Way of Words. Cam- Dretske, F. 1981. Knowledge and the Flow of bridge MA: Harvard University Press. [Collection Information. Cambridge, MA: MIT Press, rep. of Grice’s influential works, advanced under- Stanford: CSLI, 1999. [Classic informational ana- graduate level.] lysis of knowledge, advanced undergraduate level.] Hanson, P., ed. 1990. Information, Language Fetzer, J. H. (forthcoming). “Information misin- and Cognition. Vancouver: University of British formation, and disinformation.” Forthcoming in Columbia Press. [Important collection of essays, Minds and Machines. [Criticizes Floridi’s view most at graduate level.] that information encapsulates truth and develops Hintikka, J. and Suppes, P. 1970. Information and an alternative account; accessible.] Inference. Dordrecht: Reidel. [Important collec- Floridi, L. 1999. Philosophy and Computing – An tion of philosophical essays on information theory, Introduction. London and New York: Routledge. graduate level.] [Textbook that complements this Guide, element- Israel, D. and Perry, J. 1990. “What is informa- ary undergraduate level.] tion?” In Hanson 1990: 1–19. [Analyzes informa- ——. 2003. “On the intrinsic value of information tion on the basis of situation logic, graduate level.] objects and the infosphere.”. Ethics and Informa- Jones, D. S. 1979. Elementary Information Theory. tion Technology 4, no. 4: 287–304. Preprint avail- Oxford: Clarendon Press. [Brief textbook on able at . [Develops an ethical approach to & Thomas 1991, but still more demanding than information environments, advanced undergradu- Pierce 1980.] ate level.> Landauer, R. 1987. “Computation: a fundamental ——. (forthcoming a). “Is semantic information physical view.” Physica Scripta 35: 88–95. [Gradu- meaningful data?” Forthcoming in Philosophy ate level only.] and Phenomenological Research. Preprint avail- ——. 1991. “Information is physical.” Physics Today able at . [Defines semantic information as ——. 1996. “The physical nature of information.” well-formed, meaningful, and truthful data; Physics Letter A217: 188. [Graduate level only.] graduate level.] —— and Bennett, C. H. 1985. “The fundamental ——. (forthcoming b). “Information, semantic con- physical limits of computation.” Scientific Amer- ceptions of.” Stanford Encyclopedia of Philosophy. ican July: 48–56. [A more accessible presentation [Reviews philosophical conceptions of semantic of the view that information requires a physical information, undergraduate level.] implementation, undergraduate level.] ——. (forthcoming c) “Outline of a theory of Leff, H. S. and Rex, A. F. 1990. Maxwell’s Demon: strongly semantic information.” Forthcoming in Entropy, Information, and Computing. Bristol:

59 Luciano Floridi

Hilger. [Collection of essays on this classic prob- Popper, K. R. 1935. Logik der Forschung: zur lem, graduate level.] Erkenntnistheorie der modernen Naturwissenschaft. Losee, R. M. 1997. “A discipline independent Vienna: J. Springer; tr. The Logic of Scientific definition of information.” Journal of the Amer- Discovery. London: Hutchinson, 1959. [Popper’s ican Society for Information Science 48(3): 254– classic text, graduate level.] 69. [Undergraduate level.] Schrader, A. 1984. “In search of a name: informa- Mabon, P. C. 1975. Mission Communications: tion science and its conceptual antecedents.” The Story of Bell Laboratories. [Very readable Library and Information Science Research 6: account of the people and the discoveries that 227–71. [Undergraduate level.] made information theory possible, undergradu- Schneider, T. 2000. “Information theory primer ate level.] – with an appendix on logarithms.” Version Machlup, F. 1983. “Semantic quirks in studies of 2.48, postscript version , web version . Messages. New York: John Wiley, pp. 641–71. [A very clear and simple introduction that [Advanced undergraduate level.] can also be consulted for further clarification MacKay, D. M. 1969. Information, Mechanism and about the mathematics involved, undergraduate Meaning. Cambridge, MA: MIT Press. [Develops level.] an alternative view of information to Shannon’s, Shannon, C. E. 1993a. Collected Papers, eds. N. J. graduate level.] A. Sloane and A. D. Wyner. Los Alamos, CA: Mingers, J. 1997. “The nature of information and IEEE Computer Society Press. [Mostly graduate its relationship to meaning.” In R. L. Winder level only.] et al., eds., Philosophical Aspects of Information ——. 1993b. Article on “Information Theory,” Systems. London: Taylor and Francis, pp. 73–84. Encyclopedia Britannica, repr. in his Collected [Analyzes information from a system theory per- Papers, pp. 212–20. [A brief and accessible pre- spective, advanced undergraduate level.] sentation of information theory by its founding NATO. 1974. Advanced Study Institute in Informa- father, undergraduate level.] tion Science, Champion, 1972. Information —— and Weaver, W. 1998 [first published 1948]. Science: Search for Identity, ed. A. Debons. New The Mathematical Theory of Communication, with York: Marcel Dekker. a foreword by R. E. Blahut and B. Hajek. Urbana ——. 1975. Advanced Study Institute in Informa- and Chicago, IL: University of Illinois Press. [The tion Science, Aberystwyth, 1974. Perspectives in classic text in information theory, graduate level; Information Science, eds. A. Debons and W. J. Shannon’s text is also available on the web – see Cameron. Leiden: Noordhoff. above.] ——. 1983. Advanced Study Institute in Informa- Sloman A. 1978. The Computer Revolution in tion Science, Crete, 1978. Information Science Philosophy. Atlantic Highlands: Humanities in Action: Systems Design, eds. A. Debons and A. Press. [One of the earliest and most insightful G. Larson. Boston: Martinus Nijhoff. discussions of the informational/computation Nauta, D. 1972. The Meaning of Information. The turn in philosophy, most chapters undergraduate Hague: Mouton. [Reviews various analyses of level.] information, advanced undergraduate level.] Steane, A. M. 1998. “Quantum computing.” Newell, A. and Simon, H. A. 1976. “Computer Reports on Progress in Physics 61: 117–73. [A science as empirical inquiry: symbols and search.” review, graduate level, also available online at Communications of the ACM, 19 March: 113– .] 26. [The classic paper presenting the Physical Thagard, P. R. 1990. “Comment: concepts of Symbol System Hypothesis in AI and cognitive information.” In Hanson 1990. science, graduate level.] Weaver, W. 1949. “The mathematics of commun- Pierce, J. R. 1980. An Introduction to Information ication.” Scientific American 181(1): 11–15. Theory: Symbols, Signals and Noise. New York, [Very accessible introduction to Shannon’s theory, Dover Publications. [Old but still very valuable undergraduate level.] introduction to information theory for the non- Wellisch, H. 1972. “From information science to mathematician, undergraduate level.] informatics.” Journal of Librarianship 4: 157–87.

60 Information

Wersig, G. and Neveling, U. 1975. “The phenom- Cybernetics and Society, 2nd ed.; reissued in 1989 ena of interest to information science.” Informa- with a new introduction by Steve J. Heims. Lon- tion Scientist 9: 127–40. don: Free Association. [A very early discussion of Wheeler, J. A. 1990. “Information, physics, quan- the ethical and social implications of the com- tum: the search for links.” In W. H. Zureck, ed., puter revolution, undergraduate level.] Complexity, Entropy, and the Physics of Information. ——. 1961. Cybernetics or Control and Commun- Redwood City, CA: Addison Wesley. [Introduces ication in the Animal and the Machine, 2nd ed. the “It from Bit” hypothesis, graduate level.] Cambridge, MA: MIT Press. [The foundation of Wiener, N. 1954. The Human Use of Human Beings: cybernetics, graduate level.]

61 Part II Computers in Society Computer Ethics

Chapter 5 Computer Ethics Deborah G. Johnson

The first is to organize the ethical issues by 1 Introduction type of technology and its use. When computers were first invented, they were understood to be From the moment of their , computers essentially sophisticated calculating machines, but have generated complex social, ethical, and value they seemed to have the capacity to do that which concerns. These concerns have been expressed was thought to be uniquely human – to reason in a variety of ways, from the science fiction and exhibit a high degree of rationality; hence, stories of Isaac Asimov (1970) to a dense three- there was concern that computers threatened volume treatise on social theory by Manuel ideas about what it means to be human. In the Castells (1996, 1997, 1998), and with much in shadow of the Second World War, concerns between. Generally, the literature describes the quickly turned to the use of computers by gov- social consequences of computing, speculates on ernments to centralize and concentrate power. the meaning of computation and information These concerns accompanied the expanding use technology in human history, and creatively of computers for record-keeping and the expon- predicts the future path of development of com- ential growth in the scale of databases, allow- puter technology and social institutions around ing the creation, maintenance, and manipulation it. A small, though steadily increasing, number of huge quantities of personal information. This of philosophers has focused specifically on the was followed by the inception of software con- ethical issues. trol systems and video games, raising issues of As computer technology evolves and gets de- accountability–liability and property rights. This ployed in new ways, certain issues persist – issues evolution of computer technology can be fol- of privacy, property rights, accountability, and lowed through to more recent developments social values. At the same time, seemingly new including the internet, simulation and imaging and unique issues emerge. The ethical issues can technologies, and virtual reality systems. Each be organized in at least three different ways: one of these developments was accompanied by according to the type of technology; according conceptual and moral uncertainty. What will this to the sector in which the technology is used; or that development mean for the lives and values and according to ethical concepts or themes. In of human beings? What will it do to the relation- this chapter I will take the third approach. How- ship between government and citizen? Between ever, before doing so it will be useful to briefly employer and employee? Between businesses and describe the other two approaches. consumers?

65 Deborah G. Johnson

A second enlightening approach is to organize software? Is there any harm in me taking on a the issues according to the sector in which they pseudo-identity in an online chatroom? Should occur. Ethical issues arise in real-world contexts, companies doing business online be allowed to and computer-ethical issues arise in the con- sell the transaction-generated information they texts in which computers are used. Each context collect? These are examples of policy vacuums or sector has distinctive issues, and if we ignore created by computer technology. this context we can miss important aspects of Moor’s account of computer ethics has shaped computer-ethical issues. For example, in dealing the field of computer ethics with many com- with privacy protection in general, we might miss puter ethicists understanding their task to be that the special importance of privacy protection for of helping to fill policy vacuums. Indeed, one of medical records where confidentiality is so essen- the topics of interest in computer ethics is to tial to the doctor–patient relationship. Similarly, understand this activity of filling policy vacuums. one might not fully understand the appropriate This will be addressed later on. role for computers in education were one not sensitive to distinctive goals of education. Both of these approaches – examining issues 2.1 The connection between technology by types and uses of particular technologies, and and ethics sector by sector – are important and illuminat- ing; however, they take us too far afield of the While Moor’s account of computer ethics remains philosophical issues. The third approach – the influential, it leaves several questions unanswered. approach to be taken in this chapter – is to em- Hence, discussion and debate continue around phasize ethical concepts and themes that persist the question of why there is or should be a field across types of technology and sectors. Here the of computer ethics and what the focus of the issues are sorted by their philosophical and eth- field should be. ical content. In this chapter I divide the issues In one of the deeper analyses, Floridi (1999) into two broad categories: (1) metatheoretical argues for a metaphysical foundation for com- and methodological issues, and (2) traditional puter ethics. He provides an account of com- and emerging issues. puter ethics in which information has status such that destroying information can itself be morally wrong. In my own work I have tried to establish the foundation of computer ethics in the non- 2 Metatheoretical and obvious connection between technology and Methodological Issues ethics (Johnson 2001). Why is technology of relevance to ethics? What difference can tech- Perhaps the deepest philosophical thinking on nology make to human action? To human affairs? computer-ethical issues has been reflection on the To moral concepts or theories? field itself – its appropriate subject matter, its Two steps are involved in answering these relationship to other fields, and its methodology. questions. The first step involves fully recogniz- In a seminal piece entitled “What is Computer ing something that Moor’s account acknow- Ethics?” Moor (1985) recognized that when ledges, namely that technology often makes it computers are first introduced into an environ- possible for human beings to do what they could ment, they make it possible for human beings not do without it. Think of spaceships that take (individuals and institutions) to do things they human beings to the moon; think of imaging couldn’t do before, and this creates policy technology that allows us to view internal organs; vacuums. We do not have rules, policies, and or think of computer viruses that wreak havoc conventions on how to behave with regard to on the internet. the new possibilities. Should employers monitor Of course, it is not just that human beings can employees to the extent possible with com- do what they couldn’t do before. It is also that we puter software? Should doctors perform surgery can do the same sorts of things we did before, remotely? Should I make copies of proprietary only in new ways. As a result of technology, we

66 Computer Ethics can travel, work, keep records, be entertained, others before computer technology, but not in communicate, and engage in warfare in new ways. this way and perhaps not quite so easily. Computer When we engage in these activities using com- technology is not unique among technologies puter technology, our actions have different prop- in this respect; other technologies have turned erties, properties that may change the character simple movements of the body into powerful of the activity or action-type. Consider the act actions, e.g. dynamite, automobiles. of writing with various technologies. When I write Recognizing the intimate connection between with paper and pencil, the pencil moves over technology and human action is important for paper; when I write using a typewriter, levers and stopping the deflection of human responsibility gears move; when I write using a computer, elec- in technology-instrumented activities, especially tronic impulses change configurations in micro- when something goes wrong. Hence, the hacker chips. So, the physical events that take place when cannot avoid responsibility for launching a virus I write are very different when I use computer on grounds that he simply moved his fingers technology. while sitting in his home. Technology does noth- Using action theory, the change can be char- ing independent of human initiative; though, of acterized as a change in the possible act tokens course, sometimes human beings cannot foresee of an act type. An act type is a kind of action what it is they are doing with technology. (e.g. reading a book, walking) and an act token Thus, the first step in understanding the con- is a particular instance of an act type. An act nection between computer technology and ethics token is an instance of the act type performed is to acknowledge how intimate the connection by a particular person, at a particular time, and between (computer) technology and human in a particular place. For example, “Jan is, at this action can be. The second step is to connect moment, playing chess with Jim in Room 200 human action to ethics. This step may seem too of Thornton Hall on the campus of University obvious to be worthy of mention since ethics is of Virginia” is an act token of the act type “play- often understood to be exclusively the domain ing chess.” When technology is involved in the of human action. Even so, computer technology performance of an act type, a new set of act changes the domain of human action; hence, it tokens may become possible. It is now possible, is worth asking whether these changes have moral for example, to “play chess” while sitting in significance. Does the involvement of computer front of a computer and not involving another technology – in a human situation – have moral human being. Instead of manually moving three- significance? Does the instrumentation of human dimensional pieces, one presses keys on a key- action affect the character of ethical issues, the board or clicks on a mouse. Thus, when human nature of ethical theory, or ethical decision- beings perform actions with computers, new sets making? of tokens (of act types) become possible. Most The involvement of computer technology important, the new act tokens have properties has moral significance for several reasons. As men- that are distinct from other tokens of the same tioned earlier, technology creates new possibilities act type. for human action and this means that human Computer technology instruments human beings face ethical questions they never faced action in ways that turn very simple movements before. Should we develop biological weapons into very powerful actions. Consider hardly- and risk a biological war? Should I give my organs visible finger movements on a keyboard. When for transplantation? In the case of computer the keyboard is connected to a computer and the technology, is it wrong to monitor keystrokes of computer is connected to the internet, and when employees who are using computers? To place the simple finger movements create and launch cookies on computers when the computers are a computer virus, those simple finger movements used to visit a website? To combine separate can wreak havoc in the lives of thousands (even pieces of personal data into a single comprehens- millions) of people. The technology has instru- ive portfolio of a person? mented an action not possible without it. To be When technology changes the properties of sure, individuals could wreak havoc on the lives of tokens of an act type, the moral character of the

67 Deborah G. Johnson act type can change. In workplace monitoring, we to conceptualize a search engine? A cookie? for example, while it is generally morally accept- A virtual harm? In other words, computer ethi- able for employers to keep track of the work of cists do more than “apply” principles and theor- employees, the creation of software that allows ies; they do conceptual analysis. Moreover, the the employer to record and analyze every key- analysis of a computer-ethical issue often involves stroke an employee makes raises the question in a synthesis, synthesis that creates an understanding new way. The rights of employers and employees of both the technology and the ethical situation. have to be reconsidered in light of this new A fascinating illustration of this is the case of a possibility. Or to use a different sort of example, virtual rape (Dibbell 1993). Here a character in when it comes to property rights in software, a multi-user virtual reality game rapes another the notion of property and the stakes in owning character. Those participating in the game are and copying are significantly different when it outraged and consider the behavior of the real comes to computer software because computer person controlling the virtual characters offens- software has properties unlike that of anything ive and bad. The computer ethical issue involves else. Most notably, software can be replicated figuring out what, if anything, wrong the real with no loss to the owner in terms of possession person controlling the virtual character has done. or usefulness (though, of course, there is a loss This involves understanding how the technology in the value of the software in the marketplace). works, what the real person did, figuring out how So, computers and ethics are connected inso- to characterize the actions, and then recommend- far as computers make it possible for humans ing how the behavior should be viewed and re- to do things they couldn’t do before and to do sponded to. Again, analysis of this kind involves things they could do before but in new ways. more than simply “applying” principles and theor- These changes often have moral significance. ies. It involves conceptual analysis and interpreta- tion. Indeed, the synthetic analysis may have implications that reflect back on the meaning of, or our understanding of, familiar moral principles 2 Applied and Synthetic Ethics and theories. To be sure, philosophical work in computer To say that computer technology creates new ethics often does involve drawing on and extend- tokens of an act type may lead some to categorize ing the work of well-known philosophers and computer ethics as a branch of applied or practical making use of familiar moral concepts, principles, ethics. Once a computer ethical issue is under- and theories. For example, computer ethical issues stood to involve familiar act types, it might be have frequently been framed in utilitarian, presumed, all that is necessary to resolve the deontological, and theory. Many issue is to use moral principles and theories that scholars writing about the internet have drawn on generally apply to the act type. For example, if the work of existentialist philosophers such as the situation involves honesty in communicating Søren Kierkegaard (Dreyfus 1999; Prosser & Ward information, simply follow the principle, “tell 2000) and Gabriel Marcel (Anderson 2000). The the truth,” with all its special conditions and work of Jürgen Habermas has been an import- caveats. Or, if the situation involves producing ant influence on scholars working on computer- some positive and negative effects, simply do the mediated communication (Ess 1996). Recently utilitarian calculation. This account of computer van den Hoven (1999) has used Michael Walzer’s ethics is, however, as controversial as is the notion “spheres of justice” to analyze the information of “applied ethics” more generally. society; Cohen (2000) and Introna (2001) have For one thing, computer technology and the used Emmanuel Levinas to understand internet human situations arising around it are not always communication; Adams and Ofori-Amanfo (2000) so easy to understand. As Moor has pointed out, have been connecting feminist ethics to compu- often there are conceptual muddles (1985). What ter ethics; and Grodzinsky (1999) has developed is software? What is a computer virus? How are virtue theory to illuminate computer ethics.

68 Computer Ethics

Nevertheless, while computer ethicists often 3.1 Ethics for computer professionals draw on, extend, and “apply” moral concepts and theories, computer ethics involves much more In an information society, a large number of than this. Brey (2000) has recently argued for individuals are educated for and employed in an approach that he labels “disclosive computer jobs that involve development, maintenance, ethics.” The applied ethics model, he notes, buying and selling, and use of computer and emphasizes controversial issues for which the eth- information technology. Indeed, an information ical component is transparent. Brey argues that society is dependent on such individuals – there are many nontransparent issues, issues that dependent on their special knowledge and are not so readily recognized. Analysis must be expertise and on their fulfilling correlative social done to “disclose” and make visible the values responsibilities. Expertise in computing can be at stake in the design and use of computer tech- deployed recklessly or cautiously, used for good nology. A salient example here is work by Introna or ill, and the organization of information tech- and Nissenbaum (2000) on search engines. They nology experts into occupations/professions is show how the design of search engines is laden an important social means of managing that with value choices. In order to address those expertise in ways that serve human well-being. value choices explicitly, the values embedded An important philosophical issue here has to in search engine design must be uncovered and do with understanding and justifying the social disclosed. This may sound simple but in fact responsibilities of computer experts. Recogniz- uncovering the values embedded in technology ing that justification of the social responsibilities involves understanding how the technology of computer experts is connected to more general works and how it affects human behavior and notions of duty and responsibility, computer ethi- human values. cists have drawn on a variety of traditional philo- Setting aside what is the best account of sophical concepts and theories, but especially computer ethics, it should be clear that a major social contract theory. concern of the field is to understand its domain, Notice that the connection between being a its methodology, its reason for being, and its computer expert and having a duty to deploy that relationship to other areas of ethical inquiry. As expertise for the good of humanity cannot be computer technology evolves and gets deployed explained simply as a causal relationship. For one in new ways, more and more ethical issues are thing, one can ask “why?” Why does the role of likely to arise. computer expert carry with it social responsibil- ities? For another, individuals acting in occupa- tional roles are typically not acting simply as individual autonomous moral agents; they act as 3 Traditional and Emerging employees of companies or agencies, and may Issues not be involved in the decisions that most critic- ally determine project outcomes. Hence, there is “Information society” is the term often used a theoretical problem in explaining why and to (especially by economists and sociologists) to what extent individuals acting in occupational characterize societies in which human activity roles are responsible for the effects of their work. and social institutions have been significantly Social contract theory provides an account of transformed by computer and information tech- the connection between occupational roles and nology. Using this term, computer ethics can social responsibilities. A social contract exists be thought of as the field that examines ethical between members of an occupational group and issues distinctive to “an information society.” the communities or societies of which they are Here I will focus on a subset of these issues, a part. Society (states, provinces, communities) those having to do with professional ethics, allows occupational groups to form professional privacy, cyber crime, virtual reality, and general organizations, to make use of educational institu- characteristics of the internet. tions to train their members, to control admission,

69 Deborah G. Johnson and so on, but all of this is granted in exchange on. Thus, it would be a mistake to think that all for a commitment to organize and control the of the ethical issues surrounding computer and occupational group in ways that benefit society. information technology are simply ethical issues In other words, a profession and its members for computer professionals. Computer experts acquire certain privileges in exchange for accept- face many complex and distinctive issues, but ing certain social responsibilities. these are only a subset of the ethical issues sur- The substantive content of those responsibil- rounding computer and information technology. ities has also been a topic of focus for computer ethicists. Computer professional groups have developed and promulgated codes of professional 3.2 Privacy and ethical conduct that delineate in broad terms what is and is not required of computer experts. In an “information society” privacy is a major See, for example, the ACM Code of Ethics and concern in that much (though by no means all) Professional Conduct or the Code of Conduct of the information gathered and processed is of the British Computer Society. Since these information about individuals. Computer tech- codes are very general, there has been a good nology makes possible a previously unimaginable deal of discussion as to their appropriate role magnitude of data collection, storage, retention, and function. Should they be considered com- and exchange. Indeed, computer technology has parable to law? Should there be enforcement made information collection a built-in feature of mechanisms and sanctions for those who violate many activities, for example, using a credit card, the code? Or should codes of conduct aim at making a phone call, browsing the web. Such inspiration? If so, then they should merely con- information is often referred to as transaction- sist of a statement of ideals and need not be generated information or TGI. followed “to the letter” but only in spirit. Computer ethicists often draw on prior philo- At least one computer ethicist has gone so far sophical and legal analysis of privacy and focus as to argue that the central task of the field of on two fundamental questions: What is privacy? computer ethics is to work out issues of pro- Why is it of value? These questions have been fessional ethics for computer professionals. contentious and privacy often appears to be an Gotterbarn (1995: 21) writes that the “only way elusive concept. Some argue that privacy can be to make sense of ‘Computer Ethics’ is to narrow reduced to other concepts such as property or its focus to those actions that are within the liberty; some argue that privacy is something control of the individual moral computer in its own right and that it is intrinsically valu- professional.” able; yet others argue that while not intrinsically While Gotterbarn’s position is provocative, it valuable, privacy is instrumental to other things is not at all clear that it is right. For one thing, that we value deeply – friendship, intimacy, and many of the core issues in computer ethics are democracy. social value and policy issues, such as privacy Computer ethicists have taken up privacy and property rights. These are issues for all cit- issues in parallel with more popular public con- izens, not just computer professionals. Moreover, cerns about the social effects of so much personal many of the core issues faced by computer pro- information being gathered and exchanged. The fessionals are not unique to computing; they are fear is that an “information society” can easily similar to issues facing other occupational groups: become a “surveillance society.” Here computer What do we owe our clients? Our employers? ethicists have drawn on the work of Bentham When are we justified in blowing the whistle? and Foucault suggesting that all the data being How can we best protect the public from risk? gathered about individuals may create a world Furthermore, since many computer professionals in which we effectively live our daily lives in a work in private industry, many of the issues they panopticon (Reiman 1995). “Panopticon” is face are general issues of business ethics. They the shape of a structure that Jeremy Bentham have to do with buying and selling, advertising, designed for prisons. In a panopticon, prison cells proprietary data, competitive practices, and so are arranged in a circle with the inside wall of

70 Computer Ethics each cell made of glass so that a guard, sitting in ican approach has been piecemeal, with separate a guard tower situated in the center of the circle, legislation for different kinds of records (i.e., can see everything that happens in each and every medical records, employment histories, credit cell. The effect is not two-way; that is, the pris- records), whereas several European countries have oners cannot see the guard in the tower. In fact, comprehensive policies that specify what kind a prison guard need not be in the guard tower of information can be collected under what for the panopticon to have its effect; it is enough conditions in all domains. Currently the policy that prisoners believe they are being watched. debates are pressured by the intensification of When individuals believe they are being watched, global business. Information-gathering organiza- they adjust their behavior accordingly; they take tions promise data subjects that they will only into account how the watcher will perceive their use information in certain ways; yet, in a global behavior. This influences individual behavior and economy, data collected in one country – with a how individuals see themselves. certain kind of data protection – can flow to While computerized information-gathering another country where there is no or different does not physically create the structure of a protection. An information-gathering organiza- panopticon, it does something similar insofar as tion might promise to treat information in a it makes a good deal of individual behavior avail- certain way, and then send the information able for observation. Thus, data collection activ- abroad where it is treated in a completely differ- ities of an information society could have the ent way, thus breaking the promise made to the panopticon effect. Individuals would know that data subject. To assure that this does not hap- most of what they do can be observed and this pen, a good deal of attention is currently being could influence how they behave. When human focused on working out international arrange- behavior is monitored, recorded, and tracked, ments and agreements for the flow of data across individuals could become intent on conforming national boundaries. to norms for fear of negative consequences. If this were to happen to a significant extent, it might incapacitate individuals in acting freely and 3.3 Cybercrime and abuse thinking critically – capacities necessary to real- ize democracy. In this respect, the privacy issues While the threats to privacy described above arise around computer technology go to the heart of from uses of computer and information techno- freedom and democracy. logy, other threats arise from abuses. As individuals It might be argued that the panoptic effect and companies do more and more electronically, will not occur in information societies because their privacy and property rights become ever data collection is invisible so that individuals are more important, and these rights are sometimes unaware they are being watched. This is a pos- threatened by individuals who defy the law or sibility, but it is also possible that as individuals test its limits. Such individuals may seek personal become more and more accustomed to informa- gain or may just enjoy the challenge of figuring tion societies, they will become more aware of out how to crack security mechanisms. They are the extent to which they are being watched. They often called hackers or crackers. The term hacker may come to see how information gathered in used to refer to individuals who simply loved the various places is put together and used to make challenge of working on programs and figuring decisions that affect their interactions with gov- out how to do complex things with computers, ernment agencies, credit bureaus, insurance com- but did not necessarily break the law. Crackers panies, educational institutions, employers, etc. were those who broke the law. However, the Concerns about privacy have been taken up terms are now used somewhat interchangeably in the policy arena, with a variety of legislation to refer to those who engage in criminal activity. controlling and limiting the collection and use The culture of hackers and crackers has been of personal data. An important focus here has of interest not only because of the threat posed been comparative analyses of policies in different by their activities, but also because the culture countries – for they vary a good deal. The Amer- of hackers and crackers represents an alternative

71 Deborah G. Johnson vision of how computer technology might be Computer ethicists have taken up the task of developed and used, one that has intrigued philo- trying to understand and conceptualize cyber- sophers. (See Chapter 7 on INTERNET CULTURE.) crimes as well as determining how to think about Hackers and crackers often defend their behavior their severity and appropriate punishment. Crim- by arguing for a much more open system of com- inal behavior is nothing new, but in an informa- puting with a freer flow of information, creating tion society new types of crimes are made possible, an environment in which individuals can readily and the actions necessary to catch criminals and share tools and ideas. In particular, the culture prevent crimes are different. suggests that a policy of no ownership of soft- ware might lead to better computing. This issue goes to the heart of philosophical theories of 3.4 Internet issues property, raising traditional debates about the foundations of property, especially intellectual Arguably the internet is the most powerful property. technological development of the late twentieth Some draw on Locke’s labor theory of property century. The internet brings together many and argue that software developers have a natural industries, but especially the computer, telecom- right to control the use of their software. Others, munications, and media enterprises. It brings such as me, argue that while there are good together and provides a forum for millions of utilitarian reasons for granting ownership in soft- individuals and businesses around the world. It ware, natural rights arguments do not justify is not surprising, then, that the internet is cur- private ownership of software (Johnson 2001). rently a major focus of attention for computer There is nothing inherently unfair about living ethicists. The development of the internet has in a world in which one does not own and cannot involved moving many basic social institutions control the use of software one has created. from a paper and ink medium to the electronic Nevertheless, currently, in many industrialized medium. The question for ethicists is this: is there countries there are laws against copying and dis- anything ethically distinctive about the internet? tributing proprietary software, and computer ethi- (A parallel question was asked in the last section cists have addressed issues around violations of with regard to cybercrime.) these laws. Conceptually, some have wondered The internet seems to have three features that whether there is a difference between familiar make it unusual or special. First, it has an unusual crimes such as theft or harassment and parallel scope in that it provides many-to-many commun- crimes done using computers. Is there any ication on a global scale. Of course, television morally significant difference between stealing and radio as well as the telephone are global in (copying and selling copies of) a software pro- scale, but television and radio are one-to-many gram and stealing a car? Is harassment via the forms of communication, and the telephone, internet morally any different than face-to-face which is many-to-many, is expensive and more harassment? The question arises because actions difficult to use. With the internet, individuals and interactions on the internet have some dis- and companies can have much more frequent tinguishing features. On the internet, individuals communication with one another, in real time, can act under the shroud of a certain kind of at relatively low cost, with ease and with visual as anonymity. They can disguise themselves through well as sound components. Second, the internet the mediation of computers. This together with facilitates a certain kind of anonymity. One can the reproducibility of information in computer communicate extensively with individuals across systems makes for a distinctive environment for the globe (with ease and minimal cost), using criminal behavior. One obvious difference in pseudonyms or real identities, and yet one never cybertheft is that the thief does not deprive the has to encounter the others face-to-face. This owner of the use of the property. The owner type of anonymity affects the content and nature still has access to the software, though of course of the communication that takes place on the the market value of the software is diminished internet. The third special feature of the internet when there is rampant copying. is its reproducibility. When put on the internet,

72 Computer Ethics text, software, music, and video can be duplicated vested interests in the information being discussed ad infinitum. They can also be altered with ease. (e.g., a participant may be an employee of the Moreover, the reproducibility of the medium company/product being discussed). When one means that all activity on the internet is recorded can’t determine the real source of information and can be traced. or develop a history of experiences with a source, These three features of the internet – global it is impossible to gauge the trustworthiness of many-to-many scope, anonymity, and reproduc- the information. ibility – have enormous positive as well as negat- Like global scope and anonymity, reproducib- ive potential. The global, many-to-many scope ility also has benefits and dangers. Reproducibility can bring people from around the globe closer facilitates access to information and commun- together, relegating geographic distance to insig- ication; it allows words and documents to be nificance. This feature is especially freeing to those forwarded (and downloaded) to an almost for whom travel is physically challenging or infinite number of sites. It also helps in tra- inordinately expensive. At the same time, these cing cybercriminals. At the same time, however, potential benefits come with drawbacks; one of reproducibility threatens privacy and property the drawbacks is that this power also goes to rights. It adds to the problems of accountability those who would use it for heinous purposes. and integrity of information arising from anony- Individuals can – while sitting anywhere in the mity. For example, when I am teaching a class, world, with very little effort – launch viruses and students can now send their assignments to disrupt communication between others. They can me electronically. This saves time, is convenient, misrepresent themselves and dupe others on a saves paper, etc. At the same time, however, the much larger scale than before the internet. reproducibility of the medium raises questions Similarly, anonymity has both benefits and about the integrity of the assignments. How can dangers. The kind of anonymity available on the I be sure the student wrote the paper and didn’t internet frees some individuals by removing bar- download it from the web? riers based on physical appearance. For example, When human activities move to the internet, in contexts in which race and gender may get in features of these activities change and the changes the way of fair treatment, the anonymity provided may have ethical implications. The internet has by the internet can eliminate bias; for example, in led to a wide array of such changes. The task of on-line education, race, gender, and physical ap- computer ethics is to ferret out these changes pearance are removed as factors affecting student- and address the policy vacuums they create. to-student interactions as well as the teacher evaluations of students. Anonymity may also facilitate participation in beneficial activities such 3.5 Virtual reality as discussions among rape victims or battered wives or ex-cons where individuals might be re- One of the most philosophically intriguing luctant to participate unless they had anonymity. capacities of computer technology is “virtual real- Nevertheless, anonymity leads to serious ity systems.” These are systems that graphically problems of accountability and for the integrity and aurally represent environments, environments of information. It is difficult to catch criminals into which individuals can project themselves and who act under the shroud of anonymity. And interact. Virtual environments can be designed anonymity contributes to the lack of integrity to represent real-life situations and then used to of electronic information. Perhaps the best illus- train individuals for those environments, e.g., pilot tration of this is information one acquires in training programs. They can also be designed to chatrooms on the internet. It is difficult (though do just the opposite, that is, to create environ- not impossible) to be certain of the identities ments with features radically different from the of the persons with whom one is chatting. The real world, e.g., fantasy games. Ethicists have same person may be contributing information just begun to take up the issues posed by virtual under multiple identities; multiple individuals may reality and the issues are deep (Brey 1999). The be using the same identity; participants may have meaning of actions in virtual reality is what is at

73 Deborah G. Johnson stake as well as the moral accountability of indi- in virtual systems one “does” something, though This is the Computer and Information Ethics it is not the action represented. For example, Resources portion of the website of the Centre killing a figure in a violent fantasy game is not for Applied Ethics of the University of British the equivalent of killing a real person. Neverthe- Columbia. The site includes Starting Points in less, actions in virtual systems can have real-world Computer Ethics/Info-Tech Ethics, a set of consequences; for example, violence in a fantasy papers to read, and a bookstore showing books game may have an impact on the real player or, in computer and information ethics which are as another example, the pilot flying in the flight linked to Amazon.com. simulator may be judged unprepared for real flight. As human beings spend more and more This website is the work of Luciano Floridi. It time in virtual systems, ethicists will have to contains his paper entitled “Information ethics: analyze what virtual actions mean and what, if on the philosophical foundation of computer any, accountability individuals bear for their ethics,” and includes a list of resources as well as virtual actions. (See Chapter 12 for more on links to other projects and papers by Floridi. VIRTUAL REALITY.) This is the site for the Centre for Computing and Social Responsibility (CCSR), in the UK, 4 Conclusion provides access to a variety of useful materials including a list of conferences, a discussion forum and links to other sites. This chapter has covered only a selection of the topics addressed by philosophers working in the field of computer ethics. Since computers and This website is devoted broadly to engineering information technology are likely to continue to and computer ethics, and contains bibliographic evolve and become further integrated into the materials and case studies, as well as links to human and natural world, new ethical issues are other sites. likely to arise. On the other hand, as we become more and more accustomed to acting with and References through computer technology, the difference between “ethics” and “computer ethics” may well Adams, A. and Ofori-Amanfo, J. 2000. “Does disappear. gender matter in computer ethics?” Ethics and Information Technology 2(1): 37–47. Anderson, T. C. 2000. “The body and commu- Websites and Other Resources nities in cyberspace: a Marcellian analysis.” Ethics and Information Technology 2(3): 153–8. Ethics and Information Technology, an interna- Asimov, I. 1970. I, Robot. Greenwich, CT: Fawcett tional quarterly journal by Kluwer Academic Publications. Publishers (and the only journal devoted speci- Brey, P. 1999. “The ethics of representation and fically to moral philosophy and information and action in virtual reality.” Ethics and Information communication technology; first published in Technology 1(1): 5–14. 1999), contains articles on a variety of topics. Brey, P. 2000. “Disclosive computer ethics.” Com- puters & Society, Dec.: 10–16. Tavani, H. 1996. “Bibliography: a computer ethics Castells, M. 1996. The Rise of the Network Society. bibliography.” Computers & Society SIGCASE Malden, MA: Blackwell Publishers. Reader 1996. New York, NY: ACM Inc. This is ——. 1997. The Power of Identity. Malden, MA: an extremely useful resource which Tavani con- Blackwell Publishers. tinues to update at: . Blackwell Publishers.

74 Computer Ethics

Cohen, R. A. 2000. “Ethics and cybernetics: Introna, L. D. 2001. “Proximity and simulacra: Levinasian reflections.” Ethics and Information ethics in an electronically mediated world.” Philo- Technology 2(1): 27–35. sophy in the Contemporary World (forthcoming). Dibbell, J. 1993. “A rape in cyberspace: how an —— and Nissenbaum, H. 2000. “Shaping the web: evil clown, a Haitian trickster spirit, two wizards, why the politics of search engines matters.” The and a cast of dozens turned a database into a Information Society 16(3): 169–85. society.” The Village Voice, Dec. 23: 36–42. Johnson, D. G. 2001. Computer Ethics, 3rd ed. Dreyfus, H. L. 1999. “Anonymity versus commit- Upper Saddle River, NJ: Prentice Hall. ment: the dangers of education on the internet.” Moor, J. 1985. “What is computer ethics?” Ethics and Information Technology 1(1): 15–21. 16(4): 266–75. Ess, C., ed. 1996. Philosophical Perspectives on Prosser, B. T. and Ward, A. 2000. “Kierkegaard and Computer-Mediated Communication. Albany: the internet: existential reflections on education State University of New York Press. and community.” Ethics and Information Tech- Floridi, L. 1999. “Information ethics: on the philo- nology 2(3): 167–80. sophical foundation of computer ethics.” Ethics Reiman, J. H. 1995. “Driving to the panopticon: a and Information Technology 1(1): 37–56. philosophical exploration of the risks to privacy Gotterbarn, D. 1995. “Computer ethics: respons- posed by the highway technology of the future.” ibility regained.” In D. G. Johnson and H. Computer and High Technology Law Journal 11: Nissenbaum, eds., Computers, Ethics and Social 27–44. Values. Englewood Cliffs, NJ: Prentice Hall, van den Hoven, J. 1999. “Privacy and the varieties pp. 18–24. of informational wrongdoing.” Australian Grodzinsky, F. S. 1999. “The practitioner from Journal of Professional and Applied Ethics 1(1): within: revisiting the virtues.” Computers & Soci- 30–43. ety 29(1): 9–15.

75 Charles Ess

Chapter 6 Computer-mediated Communication and Human–Computer Interaction Charles Ess

and so forth (on this new methodological turn Introduction: CMC and see also Chapter 26, PHILOSOPHY OF INFORMA- Philosophy TION TECHNOLOGY). To see how this is so, this chapter begins with From Anaximander through Kant, philosophers some elementary definitions. The second section have recognized that knowing a thing involves provides an analysis of some of the key philo- knowledge of its limits, i.e., the boundaries or sophical issues that are illuminated through vari- edges that define (delimit) both what a thing is ous disciplinary approaches that incorporate CMC and what it is not. Information and Computing technologies. This discussion is organized in Technologies (ICT) give philosophers powerful terms of the fundamental elements of worldview, new venues for examining previously held be- i.e., of ontology, epistemology (including semi- liefs concerning what delimits human beings, for otics, hypertext, and logic), the meaning of iden- instance, artificial vis-à-vis natural intelligence. tity and personhood (including issues of gender As we will see, ICT further allow us to test long- and embodiment), and ethical and political values debated claims regarding human nature and thus (especially those clustering about the claim that politics, that is, questions like whether we are these technologies will issue in a global demo- capable of democratic governance or we require cracy vs. the correlative dangers of commercial- authoritarian control. Computer-Mediated Com- ization and a “computer-mediated colonization”). munication (CMC) and Human–Computer Inter- In the last section, some suggestions of possible action (HCI) provide philosophers with new research directions for, and potential contribu- laboratories in which claims that previously rested tions of, “computer-mediated philosophy” are primarily on the force of the best arguments can offered, in view of a philosophical inquiry ori- now be reevaluated empirically, in light of the ented towards the sorts of theories, praxis, and attempts made to implement these assumptions interdisciplinary dialogues described here. Perhaps in the praxis of human–machine interaction, in most importantly, philosophers may be able to the potentially democratizing effects of CMC, contribute to a renewed education – one taking

76 Computer-mediated Communication

Socrates as its model – that is required for cul- so the need increased for more “user-friendly” tural flows facilitated by CMC. interface design, thus requiring greater attention to HCI issues (Bardini 2000, Hollan 1999, Suchman 1987). Finally, as Carleen Maitland (2001) points out, 1 Some Definitions the research area of computer-supported co- operative work (CSCW) may be included as a CMC may be defined as interactive communica- subarea of CMC/HCI. tion between two or more intelligent agents that relies on ICT – usually personal computers and networks – as its primary medium. Examples in- 2 Philosophical Perspectives: clude: e-mail, chatrooms, USENET newsgroups, Worldview MUDs and MOOs, listserves, “instant messaging” services (ICQ, AOL Instant Messager, etc.), audio- and video-teleconferencing, shared virtual While extensive and growing almost as explos- reality systems, and other ways of sharing files ively as the internet and the web themselves, and information via networks and the internet, both scholarly and popular literatures in CMC, including peer-to-peer file transfers (via a service HCI, and CSCW remain primarily within the such as Gnutella, ), boundaries of the disciplines of computer sci- and the multimedia communication of the web ence, “human factors” as understood in terms (e.g., personal homepages, folder- and link- of ergonomics, communication theory, cultural sharing via , photo- studies, and such social sciences as ethnography, file sharing on commercial servers, etc.). This anthropology, psychology, and, especially, in the definition allows for the possibility of humans case of CSCW, the social psychology of group communicating with intelligent but artificial work (Hakken 1999; Bell 2000). Some theorists agents via computers and networks and, as we and designers exploit the theoretical frameworks will see below, thus points towards artificial in- and insights of , cognitive telligence and related developments as limiting science, artificial intelligence, and so forth, thus issues for CMC (see Herring 2002 for a more approaching more directly philosophical domains. complete description and history of the most Finally, some examples represent an explicit dia- significant examples of CMC). logue between CMC and HCI on the one hand, HCI may be construed as a narrowly defined and philosophical concerns on the other. The variant of CMC. While CMC refers to any com- communication theorists Chesebro and Bertelson, munication between intelligent agents mediated for example, utilize a theory of communication ori- by computers, such communication usually in- ginally developed by Innis, Eisenstein, McLuhan, cludes, and thus presupposes, successful interac- and Ong, that sees communication as a techno- tion between human agents and the mediating logy that in turn centrally defines culture, in order technologies. Such interaction requires an inter- to explicitly address philosophical concerns with face design that, ideally, allows for “seamless” or epistemology, ontology, critical reasoning, etc. “intuitive” communication between human and (Chesebro & Bertelson 1996; see Ess 1999). machine. The design of such interfaces, and the Taken together, these contribute significantly to correlative investigations into human and machine the characteristically philosophical projects of un- capacities, cognitive abilities, and possible ways covering and articulating basic worldview assump- of interacting with the world and one another tions such as epistemology (including questions constitute the subject matter of HCI. While HCI concerning the nature of truth, whether truth is incipient in every computer design, early HCI may have a universally valid status, etc.), ontology literature largely assumed that machines would (including questions concerning the reality and be used by an elite of technical experts; but as meaning of being human), ethics, politics (includ- computing technologies became more ubiquitous, ing issues of democracy and justice), and so forth

77 Charles Ess

(see also in this volume Chapters 11 and 12, machine will thus turn on a variety of physical ONTOLOGY and VIRTUAL REALITY). devices (most famously, Engelbart’s mouse) and multiple senses (including a graphical user interface that exploits the visual organization of 2.1 Ontology, epistemology, personhood information). The AI orientation tended to min- imize matching human and computer in terms In this chapter the term “ontology” is used in a of interface, partly because any human–machine broad sense, one that includes more traditional symbiosis was seen as only an intermediate stage metaphysics. This category raises questions about on the way to machines replacing human beings the nature of the real, including both internal (Bardini 2000: 21). Engelbart’s “coevolutionary” entities (such as a self, mind, and/or spirit), and approach to HCI, by contrast, rests on an ana- external realities as well as an external world or logous dialogue between disciplines. He was worlds, including persons, transcendental realities directly influenced by linguist Benjamin Whorf (mathematical, ethical [e.g., values and rights that and the recognition of the role of natural lan- are not reducible to the strictly material], religious, guage in shaping worldview (Bardini 2000: 36). etc.), causal and other possible relationships. Worldview is thus the conceptual interface Beyond questions regarding ontology and vir- between HCI, linguistics, and philosophy. tual reality, questions concerning human nature Winograd and Flores (1986) more explicitly and the self are among the most prominent take up the philosophical dimensions of the split ontological questions evoked by, and explored in HCI between AI and Engelbart. They explore in, CMC and HCI. These questions are perhaps the intersections between computer technology as ancient as speculation concerning the Golem and the nature of human existence, including and automata in the fifteenth and sixteenth cen- “the background of a tacit understanding of turies. In any case, directions for design of HCI human nature and human work.” They clarify were defined from the 1950s on by two distinct that “in designing tools we are designing ways philosophical visions. The first (originally, the of being” (1986: xi). That is: tools are designed minority position represented by Douglas Engel- with the goal of making specific actions and pro- bart) was a more humanistic – indeed, classically cesses easier and thus their design reflects a range Enlightenment/Cartesian – vision of using com- of assumptions, including worldview assumptions puting technologies as slaves, in a symbiosis regarding what is valuable, what is possible and intended to augment, not replace, human intel- easy for the users involved, and what are the ligence. The second (originally more domin- preferred ways of facilitating these processes. As ant) vision of the AI community was to build they make certain actions and processes easier, superior replacements of the human mind. This tools thus embody and embed these assump- general project is commonly characterized by a tions, while excluding others. In doing so, they Cartesian dualism, one that regards the mind as thus bias their users in specific directions and, reason divorced from body and whose primary in this way, shape our possible ways of being. mode of knowledge is mathematical and sym- Following Bardini’s analysis of the dominance of bolic (see, however, Floridi 1999 for a more AI-oriented approaches in earlier HCI, Winograd extensive analysis of the philosophical assump- and Flores interpret the worldview of much com- tions underlying so-called strong AI, one that puter design as “rationalistic,” “because of its argues against the view that AI rests on Cartesian emphasis on particular styles of consciously roots). The former emphasized the need for HCI rationalized thought and action” (1986: 8). They design to accommodate the machine to the seek to correct its “particular blindness about human by recognizing that the machine differs the nature of human thought and language” by from the human in important ways. Its binary highlighting how language and thought depend language and symbolic processes do not neatly on social interaction, an analysis based on the match human natural language, and the human philosophical traditions of and phe- “interface” with our world includes that of an nomenology and including Heidegger, Austin, embodied mind, one whose interaction with the Searle, and Habermas (1986: 9).

78 Computer-mediated Communication

Winograd and Flores’s project of unveiling the as marking out a cultural shift as revolutionary established but tacit background knowledge of as the printing press, if not the invention of fire computer designers regarding what it means to (e.g., Lyotard 1984, Bolter 1984, 1991, Landow be human anticipates a burgeoning discussion in 1992, 1994). That is, to emphasize the radical CMC, HCI, and CSCW literatures in the 1990s difference between print culture and what Ong concerning specific conceptions of personhood has called the “secondary orality” of electronic and identity presumed by various design philo- media and culture (1988) requires us to establish sophies. A central focal point for this discussion a dualistic opposition between these two cultural is the notion of the cyborg, the human–machine stages, one fostered by especially post-modernist symbiosis originally figuring in science fiction, emphases on such a radical dichotomy between perhaps most prominently as the Borg in Star modernity and postmodernity. This emphasis on Trek: The Next Generation. The Borg can rep- the radical/revolutionary difference between past resent humanity’s worst fears about technology. and future is, precisely, consistent with Haraway’s Once the boundary between humanity and early “cyber-gnosticism,” the equally dualistic machinery is breached, the machinery will irre- presumption that the mind/persona in cyberspace sistibly take control, destroying our nature (spe- is radically divorced from the body sitting back cifically, the capacities for independent agency at the keyboard. Such cyber-gnosticism takes and compassion towards others) in the process. political expression in the libertarian hopes for By contrast, Donna Haraway’s “Cyborg Mani- a complete liberation from the chains of mod- festo” (1990) argues that women as embodied ernity and the constraints of what John Perry creatures are thus trapped in a real world of Barlow so contemptuously called “meatspace” patriarchal oppression, one in which women, (1996). body, and sexuality are demonized. Women (and The difficulties of dualism and Gnosticism, men) can thus find genuine equality and libera- however, are well known, ranging from the mind– tion only as disembodied minds in cyberspace, body problem (in Descartes’s terms, how does as cyborgs liberated rather than dehumanized mind as a thinking, non-extended substance com- through technology. municate with and affect the body as a non- Philosophers will recognize in Haraway’s thinking, extended substance?) to what Nietzsche vision of technologically mediated liberation a identified as “the metaphysics of the hangman,” dualism that echoes Descartes’ mind–body split. i.e., the objection that especially Christian dualisms For historians of religion, such dualism further result in a denigration of body, sexuality, women, recalls Gnostic beliefs. Gnostics held that the and “this life” in general (1954: 500). In light human spirit is a kind of divine spark fallen from of these classical difficulties, the more recent turn heaven and trapped within the debased materiality from such dualisms in the literatures of CMC and of the human body. For such a spirit – as onto- HCI is not surprising. To begin with, alternatives logically and ethically opposed to the body – to the Cartesian/AI conceptions of knowledge salvation can come only through liberation from began to emerge within the literatures of cyber- the body. Such Gnosticism appears to be at work netics and HCI, e.g. in Bateson’s notion of dis- in numerous visions of liberation through CMC tributed cognition (1972, 1979) and Engelbart’s technologies, including explicitly religious ones emphasis on kinesthetic knowledge (Bardini (O’Leary & Brasher 1996; Wertheim 1999). As 2000: 228f.; Hayles 1999: 76–90; cf. Suchman Katherine Hayles (1999) has documented, this 1987). A more recent example of this turn is dualism emerges in the foundational assumptions Hayles’ version of the “posthuman,” as charac- of cybernetics and a conception of formalistic terized by an explicit epistemological agenda: rationality in AI, one that issues most famously “reflexive epistemology replaces . . . in Hans Moravec’s hope that humans will soon embodiment replaces a body seen as a support be able to download their consciousness into system for the mind; and a dynamic partner- robotic bodies that will live forever (1988). This ship between humans and intelligent machines dualism, moreover, can be seen at work in the replaces the liberal humanist subject’s manifest relatively early celebration of hypertext and CMC destiny to dominate and control nature” (1999:

79 Charles Ess

288). That is, Hayles foregrounds here the shift shaped by users. Especially given a postmodernist from an objectivist epistemology, based on a or social-constructivist epistemology that minim- dualistic separation of subject–object (and thus izes the role of any external givens as constrain- between subjective vs. objective modes of know- ing our knowledge, time and space may become ledge, so as to then insist that only “objective” our own creations, the result of aesthetic choices modes of knowledge are of value), to an epi- and our narrative and cooperative imagination. stemology which (echoing Kant) emphasizes the On the other hand, the renewed stress on the inevitable interaction between subject and object ontological/epistemological connections between in shaping our knowledge of the world. Know- mind and world and the corresponding ethical ledge is not an “either/or” between subjective and political responsibilities entailed by such and objective, it is both subjective and objective. connections further parallel observations that, In the same way, Hayles further focuses precisely contra the ostensibly transnational character of on the meanings of embodiment in what many the web and the Net, social and national bound- now see as a post-Cartesian understanding of aries are in fact observed in cyberspace (Halavais mind-and-body in cyberspace (Bolter 2001; 2000), with potentially imperialistic consequences Brown & Duguid 2000; Dertouzos 2001). These (Barwell & Bowles 2000). As we will see in the shifts, finally, undercut the Cartesian project of discussion of politics, the strength of the connec- using technology to “master and possess nature” tions between physical spaces and cyberspace is (Descartes 1637: 35). Such a project makes sense further apparent if we examine the role of diverse for a Cartesian mind radically divorced from its cultures in resisting a “computer-mediated colon- own body and nature, indeed, a mind for whom ialism,” i.e., the imposition of Western values nature is seen as inferior and dependent (1637: and communication preferences across the globe 19). But as the environmental crises of our own as these values and preferences are embedded in day make abundantly clear, as embodied beings the current technologies of CMC and CSCW. (not just “brains on a stick”) we are intimately Recent work documents the many ways in which interwoven with a richly complex natural order, especially Asian cultures – whose cultural values contra the Cartesian dualisms underlying what and communicative preferences perhaps most Hayles calls the liberal project of mastery and clearly diverge from those embedded in Western domination of nature. More broadly, especially CMC and CSCW technologies – are able to as the demographics of the Net change, and reshape CMC and CSCW technologies in order women are now the majority of users, it seems better to preserve and enhance distinctive cultural likely that the literature on gender, cyborgs, and values and identity. personhood will continue to offer new philo- sophical insights and directions. There emerges here then a series of debates 2.2 Epistemology: , hypertext, between postmodern/dualistic emphases on rad- and logic ical difference between mind and body, humanity and nature, electronic and print cultures, etc., The notion of “communication” in CMC com- and more recent reconsiderations that stress con- bines philosophical and communication theoret- nection between these dyadic elements. These ical views. For example, Shank and Cunningham debates are further at work in philosophical con- (1996) argue that CMC requires moving from a siderations of space and place. On the one hand, Cartesian view, according to which autonomous the very term “cyberspace” indicates that our minds transfer information across a transparent ordinary conceptions cannot fully apply to the medium, to a theoretical approach reflecting both new sorts of individual and social spaces enabled communication theories that stress intersubject- by these technologies. Similarly, Mike Sandbothe ivity (as instantiated in dialogues and multilogues, (1999), partly relying on Rorty and Derrida, has in contrast with a monologue) and C. S. Peirce’s argued that the internet and the web collapse semiotics, which emphasizes the emergence of “natural” senses of time into the virtually instan- meaning out of an interconnected triad of objects, taneous, thus making the experience of time one signs, and persons (or, more generally what Peirce

80 Computer-mediated Communication calls “interpretants”). Peirce remains an import- antiphilosophical, or at least “differently” philo- ant point of dialogue between philosophers and sophical. For example, if we live increasingly in CMC theorists (Mayer 1998, Groleau & Cooren a style of multitasking and “partial attention” 1999). (Friedman 2001), how well will complex philo- Chapter 19 takes up hypertext in greater de- sophical arguments requiring sustained intel- tail. Here we can note that David Kolb (1996) lectual attention remain accessible to novice has explored how hypertextual CMC techno- philosophers? Similarly, traditional philosophical logies may preserve and expand the discursive conceptions of the self include a singular agent, moves of argument and criticism, in a domain as a moral agent responsible for its acts over time that is both hypertextual and, using Ong’s terms, or as an epistemological agent, such as Kant’s “oral” (i.e., ostensibly marked by greater equal- transcendental unity of apperception, whose unit- ity and participation than in the more hierar- ary nature is inferred from the coherence of an chical societies of literate and print cultures) experiential stream of sense-data that otherwise vis-à-vis the ostensive linearity of print. Against tends to scatter centrifugally. Of course, post- the postmodern emphasis on hypertext as radi- counters with notions of multiple, cally overturning ostensibly modern and/or solely decentered, fragmented selves. Postmodernist literate modes of reasoning and knowledge, Kolb theories dominated early CMC literature, cel- argues that hypertext can facilitate especially the ebrating the hypertextual web of cyberspace dialectical patterns of Hegelian and Nietzschean precisely as it appeared to instantiate such con- argument. But contra postmodern celebrations ceptions of self. Should our immersion into of hypertext as exploding all print-based con- cyberspace issue exactly in such decentered selves, straints, Kolb emphasizes the reality of humans however, the philosophical debates between as finite creatures. In the face of the “informa- modernists and postmodernists concerning the tion flood” of an exponentially expanding web self may become irrelevant. Selves that are de of argumentative hypertexts, Kolb (rightly) pre- facto decentered and fragmented would be dicts that the finitude of human knowers will incapable of the sustained attention required for require new centers of access to exploding infor- complex philosophical arguments – as well as mation flows, thus engendering new forms of incapable of acting as singular epistemological hypertextual discourse. and moral agents. Such selves would not dem- Herbert Hrachovec (2001) has explored CMC onstrate the cogency of the postmodern concept as a potential “space of Reason,” one whose as resulting from rigorous philosophical debate hypertextual dimensions either (a) reinstantiate between moderns and postmoderns. Rather, such traditional print-based modes of knowledge rep- selves would represent only a technologically resentation and argument (the Talmud, indices, aided self-fulfilling prophecy, i.e., the result of cross-references, use of images in medieval manu- adopting such technologies in the first place scripts, etc.) and/or (b) fundamentally challenge because we uncritically and without further and surpass traditional forms of knowledge, argument presume the truth of the postmodern argument, and reason (for similar sorts of dis- notions of self as justification for immersing cussion concerning how computer technologies ourselves in the technologies that produce such may reshape received notions of logic, see Scaltsas selves. As this last phrase tries to make clear, 1998, Barwise & Etchemendy 1998). such self-fulfilling prophecies are, in logical terms, Some famous (but controversial) studies have viciously circular arguments, for their conclusions documented negative social consequences corre- are already asserted in their premises. At stake in lating with increased participation in cyberspace. the debate, however, is nothing less than our Even such prominent proponents as Jay David most fundamental conceptions of what it means Bolter (2001) acknowledge that electronic envir- to be a human and/or a person. Both these con- onments favor the personal and playful rather ceptions and the consequences of uncritically than abstract reasoning. These observations raise accepting a given (e.g., postmodernist) concep- additional questions as to how CMC technologies tion over another are too important to have them may be shaping consciousness in ways potentially decided for us on the basis of circular argument

81 Charles Ess and self-fulfilling prophecy, rather than through a full-fledged public sphere on the internet and more logically sound philosophical debate. its complete absence in a postmodernist em- phasis on fragmentation and decentering (Becker & Wehner 2001; cf. Jones’s conceptions of 2.3 Ethics and politics: “micropolis” and “compunity,” 2001: 56–7; democratization vs. the panopticon Stevenson 2000). and modernism vs. postmodernism Examining how CMC technologies are imple- mented in praxis further illuminates this debate, Perhaps the single most important claim made where the emphasis on testing theory by attempt- in the effort to legitimate – if not simply sell – ing to realize it precisely within the particulars CMC technologies is that they will democratize, of everyday life is itself a Habermasian – indeed, in the sense of flatten, both local (including Aristotelian – requirement. Specific instances of corporate) and global hierarchies, bringing about decision-making facilitated by CMC technolo- greater freedom and equality. These claims gies appear to approximate the ideal speech situ- obviously appeal to Western – specifically, both ation and realize at least a partial public sphere modern liberal and postmodernist – values, but (Ess 1996: 218–20; Becker & Wehner 2001; require philosophical scrutiny. To begin with, Sy 2001). At the same time, however, counter- much CMC and popular literature assumes that examples abound, including cases of CMC “democracy” means especially a libertarian form technologies serving authoritarian ends and of democracy, in contrast with communitarian preserving cultural hierarchies of power, status, and pluralist forms (Ess 1996: 198–202; Hamelink privilege, etc. (Yoon 1996: 2001). There are also 2000: 165–85). Much of the theoretically in- middle grounds, with examples of CMC tech- formed debate turns on especially Habermasian nologies leading to partial fulfillment of hopes conceptions of democracy, the public sphere, and for democracy and equality in cultural contexts a notion of communicative reason which, cou- previously marked by more centralized and pled with the rules of discourse, may achieve, in hierarchical forms of government (Dahan 1999, an ideal speech situation, the freedom, equality, Hongladarom 2000, 2001, Wheeler 2001). These and critical rationality required for democracy diverse results suggest that realizing the demo- (Ess 1996: 203–12; Hamelink 2000: 165–85). cratic potentials of CMC will require conscious Seen as simply a final expression of modern attention to the social context of use, including Enlightenment, however, Habermas is criticized education, a point we shall return to below. by feminists and postmodernists for attempting to save a notion of reason that, at best, may be simply a male form of “rationality” and, at worst, 2.4 Globalization, commercialization, contrary to its intentions to achieve freedom and and commodification vs. individual, equality, threatens instead to become its own local identity form of totalitarian power (e.g. Poster 1997: 206–10). Habermas responds to these critiques Economic and infrastructure realities dramatically by incorporating especially feminist notions of call into question the assumption that CMC rep- solidarity and perspective-taking, and by criticiz- resents a democratizing technology insofar as it ing postmodernism in turn as ethically relativistic is interactive and can place a printing press in and thus unable to sustain its own preferences the hands of anyone who can afford a computer for democratic polity over other forms (Ess 1996: and internet access. Currently, less than 7 per- 212–16; Hamelink 2000: 55–76). More recent cent of the world’s population enjoys such access debate between Habermas and (see ). Commercialization and commodifica- the former’s conception of democracy and the tion work against any such democratization public sphere. Habermas’s conception of “partial effect (Poster 1997, Stratton 1997, McChesney publics” (Teilöffentlichkeiten) survives here as 2000, Willis 2000, Yoon 2001; see Plant 2000 something of a theoretical compromise between for a discussion of Irigaray’s notion of the

82 Computer-mediated Communication commodification of women in the “specular cultural values and preferences (e.g., Harris economy”). In particular, Sy (2001) describes et al. 2001). the “commodification of the lifeworld,” drawing on Habermas to understand how CMC tech- nologies in the Philippines threaten to override 3 Interdisciplinary Dialogue and local cultural values and communication prefer- Future Directions in Philosophy ences. This is a process now well-documented for numerous cultures. In the context of India, for example, Keniston (2001) analyzes commodifica- Philosophers have much to gain from the theory tion and other forces contributing to an emer- and praxis of the many disciplines clustered about ging, culturall homogenous “McWorld,” a threat CMC technologies. Despite fledgling (Ess 2001) to local and regional identity that understand- and more considered work (Borgman 1984, 1999, ably evokes sometimes violent but fragmenting Graham 1999), philosophers yet have much to efforts of preservation (Sardar 2000). Hamelink contribute to an interdisciplinary dialogue with (1999) refers to this process as the “Disneyfica- theorists and practitioners in CMC. The follow- tion scenario” (cf. Bukatman 2000). ing is only a brief overview of three key areas of Nevertheless, recent research shows how local research. or “thick” cultures both resist a computer- mediated colonization of the lifeworld and re- shape extant CMC and CSCW technologies to 3.1 Critical reflection and history better preserve and enhance distinctive com- of ideas municative preferences and cultural values. In the literature of CSCW, for example, Lorna Heaton To begin with, philosophers can extend – and, (2001) documents how Japanese CSCW resear- when necessary, amplify and challenge – the chers developed their own CSCW technologies developing histories and conceptual frameworks to capture the many elements of nonverbal com- of CMC, especially as these intersect questions of munication crucial in Japanese culture (gesture, epistemology, ethics, and ontology. Researchers gaze, etc.). Similarly, in Thailand (Hongladarom in communication theory, cultural studies, HCI, 2000, 2001) and the Philippines (Sy 2001) it etc., are not as fully versed in the history of ideas appears that any emerging global culture remains and the often complex arguments more familiar “thin” in Walzer’s sense, i.e., it provides no sense to philosophers. These limitations can result in of historical/spatial location nor any of the lacunae, oversimplifications, and errors of fact “thick” moral commitments and resources that and logic that philosophers can amend, thereby distinguish the practices and preferences of one adding greater accuracy and conceptual strength culture from the next (cf. Hamelink 1999). The to the discussion and development of CMC. dangers and problems of globalization, especially Specifically, beyond issues of epistemology, em- as fostered by the rapid diffusion of CMC bodiment, and what it means to be a person, technologies – including the presumption of a philosophers may also contribute to the related consumerist, resource-intensive, and thus non- theoretical-metatheoretical issue of what we mean sustainable lifestyle – are not to be dismissed. by culture (see Ess 2001: 20–2). However, contra the claims of technological determinism, these and similar reports suggest that CMC technologies will not inevitably over- 3.2 Uncovering worldview run diverse cultural values and preferences. Rather, especially when implemented in ways that CMC technologies force us to articulate and, attend to the social context of use, including perhaps, alter and transcend the most basic ele- education, these technologies may be appropri- ments of our worldview, including our presump- ated by diverse cultures in ways that make both tions of identity, ontology, and epistemology global (but “thin”) communication and culture (Sandbothe 1999). At the same time, the aban- possible without compromising local/“thick” doning of Cartesian dualism in early Haraway

83 Charles Ess and Barlow involves a renewed interest in phe- Republic remains a vital metaphor for both philo- nomenological and hermeneutical approaches sophy and education as processes of making that emphasize connectedness between body and tacit assumptions explicit and thereby enabling a mind and between the individual and a larger critical examination of worldview. Philosophical community as shaped by history, tradition, cul- education moves us from the ethnocentrism of ture, etc. Thus, Paul Ricoeur is enjoying a new the cave to more encompassing and finally dia- currency (Richards 1998; Bolter 2001), as are logical conceptions of human beings (Ess 2001). Husserl and Nozick (McBeath & Webb 2000). Cees Hamelink (2000: 182ff.), in his many re- In this light, Winograd and Flores, in their appeal commendations for how to democratize tech- to the hermeneutical/phenomenological philo- nology choices, calls for an explicitly “Socratic sophies of Gadamer and Heidegger, were con- education,” one that stresses critical thinking siderably ahead of their time. about the risks of deploying information and communication technologies. Hamelink appeals for an education that will “prepare people for 3.3 Contributing to global dialogue the ‘culture of dialogue’ that the democratic process requires,” a (partially Habermasian) dia- Sandbothe (1999) takes up Rorty’s hope that logue that will be based on citizens’ “capacity to the new media may lead to a transcultural com- reason through their own positions and justify munication, one that will help us become more their preferences” as they jointly “deliberate and empathic, understanding, and receptive towards reflect on the choices that optimally serve the others. Sandbothe argues that as internet com- common interest” (184). Drawing on John munication forces users to articulate our most Dewey and Martha Nussbaum, Hamelink sees basic assumptions about identity, time, and space, such education as vital to sustaining a demo- it thereby also helps us recognize the contingent cratic society as now centrally engaged with the (i.e. non-universal) character of these most basic technologies of CMC. One could add that such presumptions. Such communication thereby education is simultaneously vital to any hopes issues in a kind of epistemological humility. This for intercultural dialogue and democracy on a should short-circuit ethnocentrisms that other- global scale. In addition to historical and con- wise root both tacit and overt forms of cultural ceptual metaphors of the postmodern and imperialism, thereby contributing to the genu- posthuman, philosophical education in intercul- ine dialogue across and between cultures required tural values may contribute to a new Renaissance for the much-prophesied global village of online of cultural flows facilitated by CMC technologies communities that extend beyond specific cultural in dramatic new ways. boundaries (see also Ess 2001).

References, Resources 3.4 Education for an intercultural global village? Bardini, T. 2000. Bootstrapping: Douglas Engelbart, Coevolution, and the Origins of Personal Comput- By engaging in an interdisciplinary theory and ing. Stanford: Stanford University Press. [A highly praxis of CMC, philosophers may contribute to readable and important account of the history a specific sort of education for the citizens of an and behind some of the most im- intercultural electronic village that is required to portant debates concerning the development of avoid the cultural homogenization of McWorld HCI, including those elements such as the mouse and the radical fragmentation of Jihad. and the Graphic User Interface originally devel- While Plato (at least in a straw-man form) is oped by Engelbart and now taken for granted by routinely targeted especially by postmodernist most personal computer users. Undergraduates, critics for an alleged dualism that then grounds graduates.] subsequent dualisms in Western thought, one Barlow, J. 1996. “A declaration of the independ- can argue that his allegory of the cave in the ence of cyberspace,”

84 Computer-mediated Communication

Censorship/Internet_censorship_bills/barlow_ North Carolina Press. [One of the first and still 0296.declaration>. [A frequently-cited expression most influential cross-disciplinary explorations of the libertarian insistence on individual free- of CMC, written by a classics scholar with an dom from all constraints – first of all, with regard advanced degree in computer science. Under- to free speech – that dominated US internet cul- graduates, graduates.] ture and thus the mid-1990s internet and related ——. 1991. Writing Space: The Computer, Hypertext, CMC writing. Undergraduates, graduates.] and the History of Writing. Hillsdale, NJ: Law- Barwell, G. and Bowles, K. 2000. “Border crossings: rence Erlbaum. [A central document in the pre- the internet and the dislocation of citizenship.” vailing arguments that the structures of hypertext In D. Bell and B. Kennedy, eds., The Cybercultures conspicuously cohere with the then-emerging Reader. London and New York: Routledge, postmodernist themes (decentering, fragmenta- pp. 703–11. [Succinctly raises the question of tion, etc.). Undergraduates, graduates.] whose cultures will be lost if all cultural differ- ——. 2001. “Identity.” In T. Swiss, ed., Unspun. ence is erased (the underside of the promise of New York: New York University Press, pp. 17– the Net to eliminate the boundaries between here 29. Available online: . [A highly read- graduates.] able update of Bolter’s views, including a clear Barwise, J. and Etchemendy, J. 1998. “Computers, attack on a Cartesian sense of self as affiliated visualization, and the nature of reasoning.” In with the culture of print vs. a postmodern “fluid T. Bynum and J. Moor, eds., The Digital Phoenix: cognitive psychology” affiliated with identity on How Computers are Changing Philosophy. Oxford: the web – issuing in the insistence that CMC Blackwell, pp. 93–116. [A foundational account cannot undermine our identities as embodied. by two pioneers in the use of computers to assist Undergraduates, graduates.] teaching formal logic, in part through visualiza- Borgmann, A. 1984. Technology and the Character tion of formal relationships. Undergraduates, of Contemporary Life. Chicago: University of graduates.] Chicago Press. [One of the most significant vol- Bateson, G. 1972. Steps to an Ecology of Mind. New umes in ; Borgmann York: Ballantine Books. Also: Bateson, G. 1979. offers his notion of “focal practices” to help off- Mind and Nature: A Necessary Unity. New York: set what he argues are the humanly debilitating Bantam Books. [Two foundational volumes in consequences of increasing reliance on techno- the development of connectionist . logy. Advanced undergraduates, graduates.] Undergraduates, graduates.] ——. 1999. Holding onto Reality: The Nature of Becker, B. and Wehner, J. 2001. “Electronic net- Information at the Turn of the Millennium. Chi- works and civil society: reflections on structural cago: University of Chicago Press. [One of the changes in the public sphere.” In C. Ess, ed., most significant contributions to philosophically Culture, Technology, Communication: Towards coming to grips with “the Information Age.” an Intercultural Global Village. Albany, NY: Borgmann seeks to develop a theory of informa- State University of New York Press, pp. 65–85. tion – first of all, by developing his own theory [Provides an excellent overview of the extensive of signs and kinds of information – in order to research and scholarship pertinent to Habermas’s then establish an ethics of information intended conception of the public sphere and how far to help us recover “the good life” (in the deepest it may be instantiated in CMC environments. philosophical sense). Advanced undergraduates, Advanced undergraduates, graduate students.] graduates.] Bell, D. 2000. “Approaching cyberculture: intro- Brown, J. S. and Duguid, P. 2000. The Social Life duction.” In D. Bell and B. Kennedy, eds., The of Information. Stanford: Stanford Uuniversity Cybercultures Reader. London and New York: Press. [A thoughtful debunking of the “techno- Routledge, pp. 25–8. [An introduction to five enthusiasm” that prevails in especially US-centered essays on cyberculture, identifying representative discourse concerning “information” and the approaches and viewpoints. Undergraduates, Information Age – notable because its authors graduates.] enjoy exceptional credentials: in particular, Brown Bolter, J. D. 1984. Turing’s Man: Western Culture is chief scientist at Xerox and director of its Palo in the Computer Age. Chapel Hill: University of Alto Research Center. Undergraduates, graduates.]

85 Charles Ess

Bukatman, S. 2000. “Terminal penetration.” In ——. 1999. “Critique in communication and D. Bell and B. Kennedy, eds., The Cybercultures philosophy: an emerging dialogue?” Research in Reader. London and New York: Routledge, Philosophy and Technology 18: 219–26. [A review pp. 149–74. [Uses contemporary Continental of Chesebro & Bertelsen 1996 that examines their philosophy – including Merleau-Ponty – to effort to conjoin philosophy and communication develop an account of the “phenomenal body” theory from a viewpoint primarily grounded in that emerges in cyberspace and virtual reality, philosophy. Undergraduates, graduates.] vis-à-vis significant science-fiction and cyberpunk ——. 2001. “What’s culture got to do with it? texts and movies, as well as Disney’s Epcot Center. Cultural collisions in the electronic global Advanced undergraduates, graduates.] village, creative interferences, and the rise of Chesebro, J. W. and Bertelsen, D. A. 1996. Analyz- culturally-mediated computing.” Introduction, in ing Media: Communication Technologies as Sym- C. Ess, ed., Culture, Technology, Communication: bolic and Cognitive Systems. New York: Guilford Towards an Intercultural Global Village. Albany, Press. [An important contribution towards an NY: State University of New York Press, pp. 1– interdisciplinary of media, includ- 50. [Summarizes important cultural differences ing CMC. Advanced undergraduates, graduates.] that become apparent in the implementation of Communication Institute for Online Research, Western CMC systems in diverse cultural settings, . [CIOS offers a number so as to argue that CMC technologies embed of comprehensive databases in communication specific cultural values and communicative pref- theory and research; a modest subscription fee is erences, but that these can be resisted and over- required for full-text access. Undergraduates, come, especially through conscious attention to graduates.] the social contexts of use, including education of Dahan, M. 1999. “National security and democracy CMC users. Undergraduates, graduates.] on the Internet in Israel.” Javnost – The Public Floridi, L. 1999. Philosophy and Computing: An VI(4): 67–77. [Documents the role of the Introduction. London and New York: Routledge. internet in contributing to greater openness in [Argues against the view that Cartesian philosophy Israeli society. Undergraduates, graduates.] – specifically dualism and the resulting mind– Dertouzos, M. 2001. The Unfinished Revolution: body split – plays the role often attributed to it Human-centered Computers and What They Can in the development of and debates concerning Do For Us. New York: HarperCollins. [Dertouzos Artificial Intelligence. Undergraduates, graduates.] (head of the MIT Laboratory for Computer Friedman, T. 2001. “Cyber-serfdom.” New York Science and a thoughtful commentator) argues Times, Jan. 30, p. A27. [A prominent, techno- optimistically that a “human-centric” approach logically savvy, and attentive journalist’s account to computing, coupled with additional advances, of changing sentiments regarding information will overcome contemporary feelings of enslave- technology among some of the world’s most ment to the machines. Undergraduates, graduates.] influential business leaders. Undergraduates, Descartes, Rene. [1637] 1998. Discourse on Method graduates.] and Meditations on First Philosophy, tr. D. A. Graham, G. 1999. The Internet: A Philosophical Cress, 4th ed. Indianapolis: Hackett (original Inquiry. London and New York: Routledge. [One work published 1637). [Foundational works for of the few book-length sustained examinations , specifically the epistemo- of the internet, notable for its somewhat more logical foundations of modern science and the pessimistic conclusions, in contrast to the optim- (in)famous mind–body split. Undergraduates, istic claims of enthusiasts, regarding the tech- graduates.] nology’s promise as an agent of democratization Ess, C. 1996. “The political computer: democracy, and community. Undergraduates, graduates.] CMC, and Habermas.” In C. Ess, ed., Philo- Groleau, C. and Cooren, F. 1999. “A socio-semiotic sophical Perspectives on Computer-mediated Com- approach to computerization: bridging the gap munication. Albany, NY: State University of between ethnographers and systems analysts.” New York Press, pp. 197–230. [An early effort The Communication Review 3(1,2): 125–64. [An to examine the democratization claims of CMC example of using C. S. Peirce’s theory of semiotics proponents in light of Habermas’s theory of com- to increase the theoretical purchase of ethno- municative reason. Undergraduates, graduates.] graphy on HCI. Graduates.]

86 Computer-mediated Communication

Hakken, D. 1999. Cyborgs@cyberspace?: An Ethno- the technology and its uses are shaped by the grapher Looks to the Future. New York: Routledge. community rather than vice versa. Undergradu- [An extended examination by a prominent ates, graduates.] ethnographer – one of the first to document the Hayles, K. 1999. How We Became Posthuman: role of culture in CMC. Hakken establishes in Virtual Bodies in Cybernetics, Literature, and multiple ways a more skeptical estimate of the Informatics. Chicago: University of Chicago Press. future of CMC environments. Advanced under- [Centrally devoted to the questions of embodi- graduates, graduates.] ment, with a solid grounding in the histories Halavais, A. 2000. “National borders on the World of science, technology, and culture; Hayles over- Wide Web.” New Media and Society 2(1): 7–28. comes the usual Manichean dichotomies to [Quantitatively documents the correlation be- establish a positive and promising middle ground tween culture and webpage production and con- between modernity and more extreme versions sumption. Advanced undergraduates, graduates.] of postmodernism. Advanced undergraduates, Hamelink, C. 2000. The Ethics of Cyberspace. Lon- graduates.] don: Sage. [A very readable summary of both Heaton, L. 2001. “Preserving communication con- theoretical discussion and pertinent research con- text: virtual workspace and interpersonal space in cerning multiple ethical issues – and at the same Japanese CSCW.” In C. Ess, ed., Culture, Tech- time a significant contribution towards more nology, Communication: Towards an Intercultural ethically-informed approaches to the design and Global Village. Albany, NY: State University of implementation of CMC technologies, especially New York Press, pp. 213–40. [Provides an over- if they are to fulfill their promises of democrat- view of the (scant) CSCW literature that recog- ization (where democracy for Hamelink includes nizes the role of culture, and documents important Habermasian dimensions). For both classroom design projects in Japan that reflect distinctive and research use, undergraduate and graduate.] communicative preferences and cultural values. Haraway, D. 1990. “A cyborg manifesto: science, Undergraduates, graduates.] technology, and socialist-feminism in the late Herring, S. 2002. “Computer-mediated commun- twentieth century.” In D. Haraway, ed., Simians, ication on the internet.” In B. Cronin, ed., Annual Cyborgs, and Women: The Reinvention of Nature. Review of Information Science and Technology, New York: Routledge, pp. 149–81. [A seminal vol. 36. Medford, NJ: Information Today Inc./ essay in “cyber-feminism.” While Haraway later American Society for Information Science and modifies her views, this essay is still widely quoted Technology. [A masterful overview by one of the and appealed to. Undergraduates, graduates.] pioneers of CMC research, providing descriptions Harcourt. W., ed. 1999. Women@internet: Creating and definitions of diverse forms of CMC and a New Cultures in Cyberspace. London and New comprehensive discussion of research on com- York: Zed Books. [Takes a global, interdisciplin- munication in CMC environments as both similar ary, and culturally oriented approach to women’s to and distinctively different from previous modes engagement with the internet, in order to exam- and media of communication. Undergraduates, ine preexisting barriers to women’s involvement graduates.] and their multiple ways of reshaping the use of Hollan, J. D. 1999. “Human–computer interac- the technology to empower women both locally tion.” In R. Wilson and F. Keil, eds., The MIT and globally. Undergraduates, graduates.] Encyclopedia of the Cognitive Sciences. Available Harris, R., Bala, P., Sonan, P., Guat Lien, E. K., and online: . Trang, T. 2001. “Challenges and opportunities [Undergraduates, graduates.] in introducing information and communication Hongladarom, S. 2000. “Negotiating the global and technologies to the Kelabit community of north the local: how Thai culture co-opts the internet.” central Borneo.” New Media and Society 3(3) First Monday 5(8) (July); . to implement internet connections within a re- [A continuation of Hongladarom’s analysis, focus- mote village, significant for its efforts to do so by ing on the Thai online community of . Undergraduates, graduates.] community and to involve the community in ——. 2001. “Global culture, local cultures and the implementation process so as to ensure that the internet: the Thai example.” In C. Ess, ed.,

87 Charles Ess

Culture, Technology, Communication: Towards an traditional print and hypertext, Kolb undertakes Intercultural Global Village. Albany, NY: State a nuanced philosophical inquiry into the matches University of New York Press, pp. 307–24. and disparities between diverse argument styles and [Documents how Thai users make use of CMC hypertext structures, one that balances enthusi- to preserve distinctive Thai cultural values, and asm for their possibilities with recognition of their provides a model (based on Michael Walzer) limits, as well as the limits resulting from human for both local and global uses of CMC that beings as finite. Undergraduates, graduates.] avoid cultural homogenization. Undergraduates, Landow, G. 1992. Hypertext: The Convergence of graduates.] Contemporary Critical Theory and Technology. Hrachovec, H. 2001. “New kids on the net: Baltimore: Johns Hopkins University Press. Also: Deutschsprachige Philosophie elektronisch.” In Landow, G., ed. 1994. Hyper/Text/Theory. Bal- C. Ess, ed., Culture, Technology, Communication: timore: Johns Hopkins University Press. [Seminal Towards an Intercultural Global Village. Albany, texts by one of the most significant and influential NY: State University of New York Press, pp. 129– theorists of hypertext. Undergraduates, graduates.] 49. [Documents some of the earliest efforts to Lyotard, J.-F. 1979 [1984]. The Postmodern Con- exploit CMC for the sake of philosophical discus- dition: A Report on Knowledge, tr. G. Bennington sion in German-speaking countries and specific and B. Massumi. Minneapolis: University of Min- cultural barriers encountered therein. Under- nesota Press. [Perhaps the single most important graduates, graduates.] foundation and springboard for postmodernism Human–Computer Interface Bibliography, . [One of the most extensive graduates, graduates.] online resources – noncommercial! – devoted to Maitland, C. 2001. Personal communication. HCI. Undergraduates, graduates.] Mayer, P. A. 1998. “Computer-mediated inter- Jones, S. 2001. “Understanding micropolis and activity: a social semiotic perspective.” Conver- compunity.” In C. Ess, ed., Culture, Technology, gence: The Journal of Research into New Media Communication: Towards an Intercultural Global Technologies 4(3): 40–58. [An example of using Village. Albany, NY: State University of New York C. S. Peirce’s semiotics as a way of analyzing Press, pp. 51–66. [An excellent representation CMC. Advanced undergraduates, graduates.] of postmodern approaches to CMC as well as McBeath, G. and Webb, S. A. 2000. “On the nature an insightful and creative contribution to discus- of future worlds?: considerations of virtuality and sion of such basic issues as privacy, property, etc., utopias.” Information, Communication & Society and the prospects for online community. Online 3(1): 1–16. [Connects contemporary discourse communities, Jones argues, are only partially suc- on CMC as creating utopian spaces with tradi- cessful, and they introduce in turn new difficul- tional utopian thinking and draws on Husserl (as ties distinctive to cyberspace. Undergraduates, opposed to what the authors see as the dominant graduates.] Heideggerian approach) and Nozick to argue for Keniston, K. 2001. “Language, power, and soft- designing software in specific ways in order to ware.” In C. Ess, ed., Culture, Technology, Com- sustain defensible computer-mediated utopias. munication: Towards an Intercultural Global Advanced undergraduates, graduates.] Village. Albany, NY: State University of New York McChesney, R. 2000. “So much for the magic Press, pp. 281–306. [An extensive examination of technology and the free market: the world of both the need for and the multiple barriers to wide web and the corporate media system.” In software localization in India – an important test A. Herman and T. Swiss, eds., The World Wide case for the democratization claim, insofar as In- Web and Contemporary Cultural Theory. New dia is the world’s largest democracy, and also the York and London: Routledge, pp. 5–35. [A most diverse in terms of languages and cultures. highly critical evaluation of the role of commer- Undergraduates, graduates.] cial interests in shaping CMC technologies and Kolb, D. 1996. “Discourse across links.” In C. Ess, their uses, by a prominent technology analyst. ed., Philosophical Perspectives on Computer- Undergraduates, graduates.] mediated Communication. Albany, NY: State Uni- Moravec, H. 1988. Mind Children: The Future of versity of New York Press, pp. 15–26. [One of Robot and Human Intelligence. Cambridge, MA: the few philosophers who has published both in Harvard University Press. [Moravec’s optimistic

88 Computer-mediated Communication

arguments and visions for the development of up Irigarary’s notion of the “specular economy” artificial intelligence and robots that will surpass, (that characterizes capitalism and patriarchy as a but not necessarily enslave, human beings. Under- matter of trading women’s bodies) and argues – graduates, graduates.] somewhat parallel to early Haraway – that women Nietzsche, Friederich. [1954] 1988. “Twilight of the can escape this enslavement in cyberspace. Plant Idols” [Götzen-Dämmerung]. In The Portable is also of interest as she takes up Maturana’s work Nietzsche, tr. W. Kaufmann. New York: Penguin, in relation to cybernetics – thus continuing a pp. 463–563. [One of Nietzsche’s last works, discussion thread begun by Winograd and Flores. this collection of aphorisms provides an unparal- Advanced undergraduates, graduates.] leled overview and self-evaluation of his work, its Richards, C. 1998. “Computer mediated com- sources, and its hoped-for impacts – along with, munication and the connection between virtual for example, a number of historical and cultural utopias and actual realities.” In C. Ess and observations from this self-proclaimed “untimely F. Sudweeks, eds., Proceedings of the First man.” Undergraduates, graduates.] International Conference on Cultural Attitudes O’Leary, S. and Brasher, B. 1996. “The unknown Towards Technology and Communication. Sydney: God of the internet: religious communication Key Centre Design Computing, pp. 129–40. from the ancient agora to the virtual forum.” In [One of the (relatively) early analyses of the C. Ess, ed., Philosophical Perspectives on Computer- utopian/dystopian dichotomy characteristic of Mediated Communication. Albany, NY: State much, especially postmodern, discourse con- University of New York Press, pp. 223–69. [One cerning CMC that takes up a phenomenological of the first systematic overviews of “religion framework (Ricoeur) in order to find a more online,” seen in part through the lens of rhetor- balanced view of the utopian possibilities of ical practice. While helpfully critical of “cyber- CMC. Undergraduates, graduates.] gnosticism” – a quest for “the information that Sandbothe, M. 1999. “Media temporalities of the saves” that fails to understand the difference internet: philosophies of time and media in between information and wisdom – the authors Derrida and Rorty.” AI and Society 13(4): 421– take an optimistic stance regarding the expansion 34. [Sandbothe summarizes diverse philosophies of online spirituality and religious communities. of time in the twentieth century, and argues that Undergraduates, graduates.] the experience of time in CMC environments Ong, W. 1988. Orality and Literacy: The Technolo- favors constructivist views – views that are fur- gizing of the Word. London: Routledge. [A key ther consonant with Derrida and Rorty in argu- work in the development of the Eisenstein/Innis/ ing for a decentered, pluralist world of mutual McLuhan/Ong approach to communication, understanding rather than dominance by a single beginning with oral communication, as a form culture and its values. Advanced undergraduates, of technology – one that shapes fundamental graduates.] assumptions regarding reason, the nature and Sardar, Z. 2000. “ALT.CIVILIZATIONS.FAQ: role of argument vis-à-vis narrative, and the very cyberspace as the darker side of the west.” In structures (egalitarian vs. hierarchical) of societies. D. Bell and B. Kennedy, eds., The Cybercultures Undergraduates, graduates.] Reader. London and New York: Routledge, Poster, M. 1997. “Cyberdemocracy: internet and pp. 732–52. [An exceptionally powerful critique the public sphere.” In D. Porter, ed., Internet of the overt and covert forms of colonialism Culture. New York: Routledge, pp. 201–17. [One the author sees in the technologies and dis- of the most significant analyses of the issues sur- courses surrounding cyberspace. Undergraduates, rounding democracy and CMC from postmodern graduates.] and feminist perspectives (including an important Scaltsas, T. 1998. “Representation of philosophical discussion and critique of Habermas’s notion of argumentation.” In T. Bynum and J. Moor, eds., the public sphere). Undergraduates, graduates.] The Digital Phoenix: How Computers are Chang- Plant, S. 2000. “On the matrix: cyberfeminist simu- ing Philosophy. Oxford: Blackwell, pp. 79–92. [An lations.” In G. Kirkup et al., eds., The Gendered account of the Archelogos Project, which uses Cyborg. London and New York: Routledge, computers to help uncover and visually repres- pp. 265–75. [A complex article by one of the ent argument structures in ancient Greek texts. best-known writers on cyberfeminism, Plant takes Undergraduates, graduates.]

89 Charles Ess

Shank, G. and Cunningham, D. 1996. “Mediated the textual – as a possible alternative to the spread phosphor dots: toward a post-Cartesian model of CMC as otherwise a colonization of the of CMC via the semiotic superhighway.” In lifeworld. Undergraduates, graduates.] C. Ess, ed., Philosophical Perspectives on Computer- Wertheim, M. 1999. The Pearly Gates of Cyberspace: mediated Communication. Albany, NY: State A History of Space from Dante to the Internet. New University of New York Press, pp. 27–41. York: Norton. [By developing an account of space [Using C. S. Peirce’s notion of semiotics, the from the Middle Ages through modern science authors critique prevailing distinctions between and cyberspace, Wertheim argues strongly against orality and textuality and argue that a semiotic the claims that cyberspace may hold transcend- conception of self and communication may lead ental or redemptive potentials. Undergraduates, to an “age of meaning” rather than a putative graduates.] Information Age. Undergraduates, graduates.] Wheeler, D. 2001. “New technologies, old culture: Stevenson, N. 2000. “The future of public media a look at women, gender, and the internet in cultures: cosmopolitan democracy and ambival- Kuwait.” In C. Ess, ed., Culture, Technology, Com- ence.” Information, Communication & Society munication: Towards an Intercultural Global Vil- 3(2): 192–214. [A theoretically and practically lage. Albany, NY: State University of New York informed analysis of the conditions required for Press, pp. 187–212. [One of the very few ana- a cosmopolitan democracy in a global media cul- lyses of the impacts of CMC among women in ture, concluding with policy recommendations the Muslim world. Wheeler finds that CMC par- for fostering the author’s “cautious cosmopolitan- tially fulfills the democratization promise of its ism.” Undergraduates, graduates.] Western proponents, thereby changing the social Stratton, J. 1997. “Cyberspace and the globalization patterns of interaction between men and women of culture.” In D. Porter, ed., Internet Culture. in Kuwait. Undergraduates, graduates.] New York and London: Routledge, pp. 253–75. Willis, A. 2000. “Nerdy no more: a case study of [A helpful critique of the assumptions under- early Wired (1993–96).” In F. Sudweeks and lying American optimism regarding the internet C. Ess, eds., Second International Conference – including a misleading nostalgia for lost com- on Cultural Attitudes Towards Technology and munity, and the presumption that American values Communication 2000. Murdoch, Australia: (human rights, , and democracy School of Information Technology, Murdoch – coupled with capitalism) are rightly “the University, pp. 361–72. Available online: (271). Stratton argues that such globalizing/ [A detailed analysis of Wired magazine, arguing homogenizing bases, however, threaten to exclude that, contrary to the magazine’s ostensibly demo- a plurality of cultures, peoples, and non-economic cratic/egalitarian ideology, the prevailing content interests. Undergraduates, graduates.] and images reinforce the power and viewpoints Suchman, L. 1987. Plans and Situated Actions: The of upper-class white males. Undergraduates, Problem of Human–Machine Communication. graduates.] Cambridge and New York: Cambridge Univer- Wilson, R. A. and Keil, F., eds. 1999. The MIT sity Press. [Argues against a European view of Encyclopedia of the Cognitive Sciences. Cambridge, planning and action as exclusively rational and MA: MIT Press. Available online (subscription abstract – a view further at work in the design of fee for full-text access): . [An invaluable resource, made up account of purposeful actions as situated, i.e., of both general summary articles and highly actions “taken in the context of particular, con- detailed discussions and analyses of specific crete circumstances” (p. viii), as a more fruitful issues, topics, and figures in cognitive science. basis for design of computers and HCI. Advanced Undergraduates, graduates.] undergraduates, graduates.] Winograd, T. and Flores, F. 1986. Understanding Sy, P. 2001. “Barangays of IT: Filipinizing mediated Computers and Cognition: A New Foundation for communication and digital power.” New Media Design. Reading, MA: Addison-Wesley. [A semi- and Society 3(3) (Sept.): 297–313. [Draws on nal effort to conjoin biologically based learning Habermas and Borgmann to argue for a “cyber- theory with philosophical epistemology in order barangay” – in part, as a synthesis of the oral and to articulate the philosophical foundations for

90 Computer-mediated Communication

Human–Computer Interface design. Advanced Technology, Communication: Towards an Inter- undergraduates, graduates.] cultural Global Village. Albany, NY: State Uni- Yoon, S. 1996. “Power online: a poststructuralist versity of New York Press, pp. 241–60. [Using perspective on CMC.” In C. Ess, ed., Philosophical both theoretical approaches (Bourdieu, Foucault) Perspectives on Computer-mediated Communica- and ethnographic interviews to examine the role tion. Albany, NY: State University of New York of Korean journalism and the function of such Press, pp. 171–96. [Takes up Foucault’s notion social dynamics as “power distance” (Hofstede) of positive power to analyze ways in which CMC to document the antidemocratic impacts of CMC in Korea work in antidemocratic directions. as it extends into Korean youth culture, especi- Undergraduates, graduates.] ally as these are fostered by commercialization. ——. 2001. “Internet discourse and the habitus of Undergraduates, graduates.] Korea’s new generation.” In C. Ess, ed., Culture,

91 Wesley Cooper

Chapter 7 Internet Culture Wesley Cooper

for surfers, a commercial space much in need of Introduction zoning, the mother of all Swiss Army knives, a tool palette for artists, a lucid dream or magic, a The internet is a magnet for many metaphors. telephone or newspaper or holodeck, a monster It is cyberspace or the matrix, the “information that has escaped DARPA’s control, the Linux superhighway” or infobahn or information hair- penguin, sliced bread, an addiction, the Grand ball, a looking-glass its users step through to Canyon, and on and on. meet others, a cosmopolitan city with tony and Before attempting to think through these shady neighborhoods, a web that can withstand metaphors, it is worthwhile to note at the outset nuclear attack, electric Gaia or God, The World that we regular users of the internet are only a Wide Wait, connective tissue knitting us into a minority, even in societies that have passed group mind, an organism or “vivisystem,” a petri through industrialization and are now exploring dish for viruses, high seas for information pirates, economies in which information technology has a battleground for a war between encrypters and become central. There are important technical, decrypters, eye candy for discreet consumers of moral, and political issues about conversion of a tsunami of pornography, a haven for vilified this minority into a majority, including whether minorities and those who seek escape from stulti- that would be a good thing, whether it is fying real-world locales, a world encyclopedia required by fairness, how much priority should or messy library or textbook or post office, chat be given to information technology in develop- “rooms” and schoolrooms and academic confer- ing countries especially relative to processes of ences, a vast playground or an office complex, a industrialization, and so forth. It is clear, how- cash cow for the dot.coms, The Widow Maker, ever, that desire for connection to the Net is not training wheels for new forms of delinquency a minority taste, something only for a military or practiced by script kiddies and warez d00des, a academic elite, but rather it corresponds closely wild frontier with very little law and order, the to the enormous demand for the ubiquitous com- glimmer in the eyes of virtual-reality creators, a puter itself at every social level. So the prospect workshop for Open Source programmers, a poll- of a global electronic metropolis, in which citizens ing booth for the twenty-first century, a market- can reliably be expected to be netizens, is not an place for mass speech, a jungle where children idle dream, or nightmare. The internet is so new are prey, a public square or global village, a mall that we don’t know yet whether it has an Aristo- or concert hall, a stake for homesteaders, a safari telian telos of some benign or malign nature, or

92 Internet Culture whether instead it will always be a loose and that the self is an outmoded social construct, disjointed Humean thing, evading every attempt and so forth. Postmodernist theory should be to discern an underlying unity. sharply distinguished from postmodern culture. Although the internet is bringing us together, The latter, however it is to be characterized in it also keeps us apart in two general ways. First, detail, is a large social fact; the former, whether time spent online is inevitably time spent in a it is true or false or meaningful or nonsensical, is greater or lesser degree of detachment from one’s precisely a theory; one can be a participant in physical surroundings, including local others. postmodern culture without espousing post- Second, the connection to distant others is itself modernist doctrine. Interpretations of postmodern a form of detachment, as coolly a matter of busi- culture, including many insightful ones, point to ness as online banking or as etiolated a form of the need for a theory of personhood and personal sociability as a chat room. The major issue about identity that does full justice to the changes in the former is simply time management. Almost us, and gives us a way of thinking constructively everyone has decided that local detachment is all about them. Many different disciplines, from right, because we do it when reading, watching philosophy to psychology, from linguistics to television, listening to music, and so forth. But sociology, from anthropology to literary studies, there are still questions about how internet use should converge so as to develop such a theory. will impact on these other forms of detachment, The internet is changing our relationship to for instance in reading less or, worse, less well. nature, not only in the way that postmodernist The latter issue is more complex. Detachment theorists emphasize, by “thickening” the layers of from distant others can be valued for purposes of images that mediate our perception of the external efficiency, as with banking, or because it affords world and our interactions with it, but also by anonymity to members of unpopular subcultures, starting to lessen the stress on nature caused by as with some chat rooms, or because one happens the technologies of the industrial revolution. The to find that level of sociability to one’s liking. two are related. The thickened layers can include There is not anything evidently wrong with any the images that constitute the emerging tech- of this, putting criminal or pathological cases aside nology of teleconferencing, and the lessened – hacking into banks, planning terrorist attacks, stress, we have reason to hope, will take the escaping from life, and so on. Perhaps the major form of reduced environmental damage caused issue about online detachment will have to do by planes, trains, and automobiles; alternatives with its transformation as more “bandwidth” gets to fossil fuel will depend, either at the research piped into our ever more versatile computers, stage or in implementation, on digital techno- giving them audiovisual and even tactile powers logy to harness the energy of the sun, the wind, to create experiences that are very different from hydrogen, and so forth. The layers can include invoking File Transfer Protocol from a command the electronic paper that is clearly visible now line to send scientific data from node A to node B. on the technological horizon, and the relief for The internet is also changing us. Users of the nature will be felt by our forests. The power of internet are not the people they would have been computer modeling should also be mentioned, a in the absence of the computer revolution. At new way of representing the world that is prov- one level this is a truism: experiences change us. ing its value for understanding, monitoring, and But many interpreters of postmodern culture, controlling natural processes, from the human the culture of postindustrial societies particularly genome to the weather; it is changing the way as influenced by information technology such as traditional sciences are undertaken as well as the internet (and computers, CDs, etc.), detect a birthing relatively new sciences such as cognitive change in us that is understated even by emphas- psychology, artificial intelligence, and nanotech- izing that our personalities have become differ- nology. These changes in the images or representa- ent. Some of these interpretations are pretentious tions that we rely upon are introducing social babble, including much theorizing that passes as changes as well, ranging from less reliance on postmodernist philosophy or psychology when the amenities of cities for educational and enter- it opines that there is nothing outside the text, tainment purposes, to new forms of populism as

93 Wesley Cooper groups organize on the internet despite lack of be . . . the thinking that inquires into the truth access to high-cost tools. The great engine of of Being and so defines man’s essential abode acculturation, schooling, is now producing genera- from Being and toward Being is neither ethics tions for whom computer use is second nature, nor ontology” (1977: 87). This chapter returns a presence in the classroom since the first year. to Heidegger under the heading of dystopian This large fact presents a challenge to the exist- inherence, making the case that Heidegger’s ence of a “mainstream culture,” since this gen- philosophy of technology does indeed betray a eration will be influenced by such various cultural significant value commitment, contrary to its aim forces that even the cultural fragmentation occa- at something more profound, a commitment that sioned by the 500-channel television will only undermines its authority as a model for under- hint at the upshot. Let this thought be the back- standing the internet. ground to the question whether the internet and The general Heideggerian idea of a value information technology not only are having im- inherent in technology is instanced in the state- pact on the larger culture, but also whether they ment that the high technology of factory farming, have a culture of their own. or “agribusiness,” is inseparable from a bad way of relating to nature, understanding it, and treat- ing it simply as something to be processed in wholesale fashion for satisfaction of human Internet Culture? appetites. Heidegger’s idea has been adopted mainly by dystopian theorists like his translator Is there internet culture, something more sub- Michael Heim, who argues in The Metaphysics stantial than shared mastery of the email or of Virtual Reality that the “Boolean logic” of chatroom “smiley,” or is that an oxymoron? Is the computer marks a “new psychic framework” the internet a tool, or something more? Is the that “cuts off the peripheral vision of the mind’s internet improving education or corrupting it? eye” and generates infomania (1993: 22, 25), Is the space of cyberspace a place to explore as he indicates in the following passage. utopian possibilities, or a wrecking yard for traditional culture, or something as neutral with Note already one telltale sign of infomania: respect to questions of value as a screwdriver? the priority of system. When system precedes These are some of the questions that a philosophy relevance, the way becomes clear for the of internet culture should address. The answers primacy of information. For it to become to be found in a large and diverse literature on manipulable and transmissible as information, the subject are classifiable as utopian, dystopian, knowledge must first be reduced to homogen- or instrumental. A utopian view sees the internet ized units. With the influx of homogenized as good, perhaps profoundly so, or at least good- bits of information, the sense of overall signi- on-balance. As dystopian, it is profoundly bad ficance dwindles. This subtle emptying of or at least bad-on-balance. And as instrumental, meaning appears in the Venn diagrams that the Net is a tool, perhaps merely a tool or at graphically display Boolean logic. (1993: 17) least a tool that does not harbor profoundly good or evil values. Heim’s profound or inherence dystopianism The notion of profundity in this trichotomy may be contrasted with on-balance or simply acknowledges the influence of balance dystopianism, exemplified by Sven on the philosophy of technology, especially his Birkerts’ The Gutenberg Elegies: The Fate of Read- The Question Concerning Technology. Many inter- ing in an Electronic Age, and particularly by the preters of the internet have borrowed from him cost–benefit analysis of the computer revolution the idea that a technology can be inseparable from that he provides in the following passage. a value commitment. Heidegger would not have liked the term “value.” In “Letter on ” We can think of the matter in terms of gains he writes, “Every valuing, even where it values and losses. The gains of electronic postmodern- positively, is subjectivising. It does not let beings: ity could be said to include, for individuals,

94 Internet Culture

(a) an increased awareness of the “big bionic convergence is a matter of words. The picture,” a global perspective that admits the meanings of “mechanical” and “life” are both extraordinary complexity of interrelations; stretching until all complicated things can be (b) an expanded neural capacity, an ability to perceived as machines, and all self-sustaining accommodate a broad range of stimuli simul- machines can be perceived as alive. Yet beyond taneously; (c) a relativistic comprehension of semantics, two concrete trends are happening: situations that promotes the erosion of old (1) Human-made things are behaving more biases and often expresses itself as tolerance; lifelike, and (2) Life is becoming more engin- and (d) a matter-of-fact and unencumbered eered. The apparent veil between the organic sort of readiness, a willingness to try new situ- and the manufactured has crumpled to reveal ations and arrangements. that the two really are, and have always been, In the loss column, meanwhile, are (a) a of one being. What should we call that com- fragmented sense of time and a loss of the mon soul between the organic communities so-called duration of experience, that depth we know of as organisms and ecologies, and phenomenon we associate with reverie; (b) a their manufactured counterparts of robots, reduced attention span and a general impati- corporations, economies, and computer cir- ence with sustained inquiry; (c) a shattered cuits? I call those examples, both made and faith in institutions and in the explanatory born, “vivisystems” for the lifelikeness each narratives that formerly gave shape to subject- kind of system holds. (1994: 3) ive experience; (d) a divorce from the past, from a vital sense of history as a cumulative or The inherent value for Kelly is the value of a organic process; (e) an estrangement from vivisystem, as revelatory of a hidden connection geographic place and community; and (f) an between the natural and the mechanical. Kelly’s absence of any strong vision of a personal or focus on vivisystems is comparable to historian collective future. (Birkerts 1994: 27) Bruce Mazlish’s reconstruction of how we have overcome the fourth discontinuity, between our- Note that the distinction between inherence and selves and machines, the earlier discontinuities balance dystopians concerns the form of argu- having been overcome when Copernicus showed mentation rather than conclusions about the tech- that our earth was not the center of the universe, nology, which may be similar. Heim would agree when Darwin showed that man did not have a with Birkerts that, as the latter writes, “We are at privileged place in creation, and when Freud a watershed point. One way of processing informa- showed that our rationality is not so perfect as tion is yielding to another. Bound up with to set us apart from the other animals. Kelly’s each is a huge array of aptitudes, assumptions, vivisystems allow Mazlish’s point to be put and understandings about the world” (1994: 27). positively, in terms of continuity rather than dis- But Heim has an extra reason for that conclu- continuity: the range of manmade and natural sion, the profound one about the “infomania” vivisystems reveals the continuity between our- value inherent in the new technology. selves and machines. Heidegger’s idea, this extra reason, can be Vivisystems figure in the version of James extended to utopianism. An inherence utopian Lovelock’s Gaia Hypothesis that Kelly endorses. about the internet, on this extension, is one who This is the hypothesis, that, in Lovelock’s words, believes that there is something good about it “The entire range of living matter on Earth, beyond a simple toting up of gains and losses. from whales to viruses, from oaks to algae, could For instance, Wired magazine editor Kevin Kelly’s be regarded as constituting a single living entity, Out Of Control: The New Biology of Machines, capable of manipulating the Earth’s atmosphere Social Systems, and the Economic World theorizes to suit its overall needs and endowed with facult- the internet as a vivisystem, and as such an ies and powers far beyond those of its constitu- instance of, in his words, ent parts” (Kelly 1994: 83). (Kelly is quoting from Lovelock’s The Ages of Gaia.) Although [t]he overlap of the mechanical and the life- there may be controversy about whether Gaia is like [that] increases year by year. Part of this an organism, Kelly thinks there should be no

95 Wesley Cooper doubt that, as he writes, “it really is a system that brains can be replaced by their Turing- that has living characteristics. It is a vivisystem. machine essence: we will be brought back to life It is a system that is alive, whether or not it as programs suitable for generating the virtual- possesses all the attributes needed for an organ- reality renderings that capture our lived experi- ism” (1994: 84). Gaia is not only alive but it is ence, with the unpleasant bits trimmed away and coming to have a mind, thanks to the internet desirable additions inserted, perhaps additions and other networking technologies. Kelly makes from program-based future societies, if we can the point in dramatic language. tolerate the culture shock. The details can be found in Frank J. Tipler’s The Physics of Immor- There is a sense in which a global mind also tality and David Deutsch’s The Fabric of Reality. emerges in a network culture. The global mind This much will serve to introduce a frame- is the union of computer and nature – of work for understanding internet culture and telephones and human brains and more. It is a the theorizing that surrounds it: the utopian/ very large complexity of indeterminate shape dystopian/instrumental trichotomy and the governed by an invisible hand of its own. We balance/inherence dichotomy. The stage is set humans will be unconscious of what the global for a critical illustration of balance utopianism, mind ponders. This is not because we are not in the next section; then inherence dystopianism; smart enough, but because the design of a and then inherence ; and fin- mind does not allow the parts to understand ally some concluding remarks, including some the whole. The particular thoughts of the caveats and qualifications about the framework global mind – and its subsequent actions – just bruited. will be out of our control and beyond our understanding. Thus network economics will breed a new spiritualism. Our primary difficulty in comprehending the Balance Utopianism global mind of a network culture will be that it does not have a central “I” to appeal to. No headquarters, no head. That will be most The advent of the internet took Sherry Turkle exasperating and discouraging. In the past, by surprise. She had published The Second Self in adventurous men have sought the holy grail, 1984, describing the identity-transforming power or the source of the Nile, or Prester John, or of the computer at that stage of the computer the secrets of the pyramids. In the future the revolution. Reflecting on her experience and the quest will be to find the “I am” of the global experience of others with the new Apple and mind, the source of its coherence. Many souls IBM PC computers, she conceived of the rela- will lose all they have searching for it – and tionship of a person to her computer as one-on- many will be the theories of where the global one, a person alone with a machine. By 1995, mind’s “I am” hides. But it will be a never- when Life on the Screen appeared, she was writ- ending quest like the others before it. (1994: ing about something quite different, “a rapidly 202) expanding system of networks, collectively known as the internet, [which] links millions of people Another inherence-utopian vision incorporates in new spaces that are changing the way we think, the internet’s group mind as only a minor fore- the nature of our sexuality, the form of our com- shadowing of an end-of-time God, intelligent munities, our very identities” (1995: 9). life connected throughout the universe, as a Though Turkle speaks neutrally here of result of colonization of space (and so forth). It “change” in these matters, she fits into the will tap into the energy created by gravity’s “utopian” category of her trichotomy between “divergence towards infinity” in the Big Crunch utopian, apocalyptic, and utilitarian evaluations so as to reproduce all past experience in massive of the internet. The computer is a new and computations that generate the requisite virtual important tool, most assuredly, but the internet realities. Construing our brains as virtual reality makes it “even more than a tool and mirror: We generators themselves, these theorists prophesy are able to step through the looking glass. We

96 Internet Culture are learning to live in virtual worlds. We may zone. For instance, a player who manipulates a find ourselves alone as we navigate virtual oceans, character or avatar in an online virtual reality unravel virtual mysteries, and engineer virtual such as a Multi-User Dungeon (MUD) is dis- skyscrapers. But increasingly, when we step tinctly located in that zone. By contrast, traveling through the looking glass, other people are there to Rome or viewing someone’s movie about as well” (1995: 9). Whereas apocalyptic theorists Rome, even when doing so is “virtually like diagnose this as stepping through the looking- being there,” is safely on one side or the other glass to cultural impoverishment or a new form of the real/virtual line, awakening no awareness of mental illness, Turkle theorizes the new experi- of the zone being constructed and explored by ences by reference to colonization of a new land. Turkle’s colonists. This metaphor of colonization should be Second, constructing identity involves some- understood carefully, however, as she is not sug- thing like the notion of a dimension as it was gesting that Sherry Turkle, sociologist and MIT just introduced: although Turkle is distinctly on professor, should be left behind in favor of a the “real” side of the real/virtual continuum, new life as the cybernaut ST on LambdaMOO. she now builds her identity partially by reference That suggestion comes from an extreme form of to dimensions of herself that owe their existence inherence utopianism about the internet, or it to activity in the border zone. To the degree is the equally extreme suggestion of inherence that MUDing is important to her, for instance, dystopian theorists, like Mark Slouka in War of to that degree it is constitutive of who she is. the Worlds, who diagnose the internet experi- This is a high-technology application of the gen- ence as equivalent to wholescale departure from eral principle that we are self-defining creatures. everyday reality. More in Turkle’s spirit is the It is not the idea that crossing the postmodern thought that a new dimension of human life is divide has somehow destroyed personal identity. being colonized, and although that raises a host Although some psychologists and sociologists of new issues about budgeting time and effort, adopt the conceit of speaking this way, it is no and even about physical and mental health, Turkle more than acknowledging the complexity of self- is not proposing that it be undertaken in the definition in modern society; or else this way of spirit of these extreme forms of utopianism. speaking falsely equates personal identity with She does indeed characterize her colonists as a soul-pellet or Cartesian Thinking Substance, “constructing identity in the culture of simula- in which case it is broadcasting the stale news tion,” in a cultural context of “eroding boundar- that such conceptions of the self are largely dis- ies between the real and the virtual, the animate credited. Turkle discusses the phenomenon of and the inanimate, the unitary and the multiple Multiple Personality Disorder, and it may be that self” (1995: 10), a context in which experiences MPD is more common because of the stresses on the internet figure prominently but share a of modern life, and not because, say, the medic- cultural drift with changes in art, such as the alization of human experience leads us to find postmodern architecture that the cultural critic mental illnesses today that weren’t there yester- Fredric Jameson studies; science, such as research day. But constructing identity is, and always has in and elsewhere inspired by been, distinct from going crazy, even when the connectionist models of the mind/brain; and building material is a new high-tech dimension. entertainment, such as films and music videos in This is not to say that Turkle always gets this which traditional narrative structure is hard to exactly right. Setting out some of her interviews discern. Constructing identity in the culture of with students who play MUDs, she writes that simulation – our postmodern culture, as Turkle “as players participate, they become authors not interprets it – involves two closely related ideas. only of text but of themselves, constructing new First, there is the idea that we are newly aware selves through social interaction. One player says, of a rich continuum of states between the real and ‘You are the character and you are not the char- the virtual, the animate and the inanimate, the acter, both at the same time.’ Another says, ‘You unitary and the multiple self. A boundary that are who you pretend to be.’” Analyzing these may have been a sharp line is now a complex interviews, she continues, “MUDs make possible

97 Wesley Cooper the creation of an identity so fluid and multiple down from high-level programming languages that it strains the limits of the notion. Identity, to assembler to machine language, and wanting after all, refers to the sameness between two to know as well the engineering architecture of qualities, in this case between a person and his the hardware. (Hackers who loved to code but or her persona. But in MUDs one can be many” knew little about hardware were called “soft- (1995: 12). The short path out of these woods ies.”) By contrast the consumer computers that is to deny that a person and his or her persona were brought to the market in the mid-seventies are identical: you are not who you pretend to to early eighties, first by Apple and then by IBM be, but rather you are pretending to be some- and many others, made computers accessible one in such a way as to call upon your verbal, far beyond the military, industry, and academe. emotional, and imaginative resources to accom- For Turkle the Apple Macintosh’s graphical plish the pretense. user interface, as well as its presenting itself One of Turkle’s major themes is the trans- as “opposed and even hostile to the traditional ition from modern to postmodern culture, modernist expectation that one could take a which she glosses as follows, beginning with a technology, open the hood, and see inside” set of ideas that have come to be known as (1995: 35), are crucial developments, giving “postmodernism.” the computer massive popular appeal to many who preferred “horizontal understanding,” of an These ideas are difficult to define simply, but operating system’s or an application’s interface, they are characterized by such terms as surface over depth. “decentered,” “fluid,” “nonlinear,” and “opaque.” They contrast with modernism, the The power of the Macintosh was how its classical world-view that has dominated West- attractive simulations and screen icons helped ern thinking since the Enlightenment. The organize an unambiguous access to programs modernist view of reality is characterized by and data. The user was presented with a scin- such terms as “linear,” “logical,” “hierarch- tillating surface on which to float, skim, and ical,” and by having “depths” that can be play. There was nowhere visible to dive. (1995: plumbed and understood. MUDs offer an 34) experience of the abstract postmodern ideas that had intrigued yet confused me during my intellectual coming of age. In this, MUDs The massive growth of internet culture, from its exemplify a phenomenon we shall meet often roots in the MIT/ARPANET connection and the in these pages, that of computer-mediated UNIX/USENET connection, into the behemoth experiences bringing philosophy down to earth. we see now, turned on the fact that a lot of (1995: 17) people want to be pilots, not mechanics. Turkle acknowledges that even her beloved It does so, Turkle suggests, because the trans- Macintosh ultimately requires the skills and tools ition from modernism to post-modernism, from of modernist culture, but it strove to make these the early post-Second World War years onward, “irrelevant” to the user, and in this way “the tools is paralleled in the world of computers by a trans- of the modernist culture of calculation became ition from a culture of calculation to a culture layered underneath the experience of the culture of simulation. For those caught up in the war of simulation” (1995: 34). This is an important effort, like John von Neumann, the new com- point, and one that she may not have developed puters were objects to calculate with, specifically sufficiently. The culture of simulation requires a to make the staggeringly complex calculations modernist spine. It requires technicians to keep that would tell whether an implosion device its computer network running, for one thing, would detonate an atomic bomb. Even the relat- but it also needs inventors and theoreticians to ively carefree hackers at the MIT AI Lab in the explore its possibilities. More generally, it needs fifties and sixties were privy to this culture, a background of a world that is external to its prizing what Turkle calls “vertical” understanding rapidly thickening layers of images and other of the computer: understanding it all the way representations, a world that is best disclosed by

98 Internet Culture the sciences, in contradistinction to the post- technocratic bureaucrats who are concerned modern conceit that there is nothing outside only with efficient administration. There are in the text, that science is just one among many fact two fallacies here. First, efficient administra- narratives in an anarchic cacophony, etc. Often tion with respect to programming might be put enough to counsel attention, modernist values to the service of organizations that are devoted consort with plain truths. (This of course rejects to human rights and opposed to technocratic the postmodernist theoretician’s notion that truth manipulation of citizens. To suppose the con- reduces to what passes for true, which is a func- trary to is to commit the simple Frame Fallacy. tion of which community’s values you subscribe Additionally, Evans makes an unwitting philo- to.) The plain truth of science’s superior track sophical pun – a fallacy of equivocation – on record consorts with the modernist value that the term efficiency. The two fallacies blend in a discerns a hierarchy in which science ranks higher spectacular howler. than, say, wishful thinking in its power to reveal Evans’s Psychology and : A Genealo- the nature of things. The plain truth that there gical Critique of the Computational Model of Mind is an external world consorts with the modernist argues that “technocratic rationality” is a secret value of depth, in this case a depth beyond our value presupposed by the computer model of images, symbols, and other representations. The mind, which he takes to be the model that defines modernist value of prudence, of rational self- cognitive science and cognitive psychology. His interest which gives equal weight to each mo- fear (“the crisis of modernity”) is that conscious- ment of one’s life, consorts with the plain truth ness itself “might be reduced to just those para- about personal identity that I canvassed earlier. meters necessary for the continued reproduction The value and the fact are not the same: one can of restrictive and univocal social, cultural, and eco- grant that there is personal identity through time nomic systems” (1993: 2). In this way the com- and rational concern about it, without embrac- puter model of cognitive psychology “serves the ing the modernist conception of prudence that interest of the new technocratic elite by emulat- requires one to be a shepherd, so to speak, for a ing their style of thinking” (1993: 7). Assimilating whole human life. For instance, it is not irra- us to machines, cognitive psychology implicitly tional, on certain conceptions of rationality, to denies those cultural values that affirm and celeb- severely discount one’s distant future. But such rate life, and consequently it is “nihilist.” conceptions aren’t those that have had influence Evans’s main argument is as follows. in building our senses of ourselves and our social institutions, like social and medical insurance. Because we can precisely state its properties, Those reflect modernist values. we shall use the Turing machine as our for- malization and the idealization of “analytic discourse.” Like analytic discourse, the Turing machine divides its subject matter into a set of Inherence Dystopianism discrete entities, maintains a strict separation between its program (language) and the domain over which it operates (the same pro- A leitmotiv of some dystopian critique is a fal- gram can imitate many different machines), lacy: an inference from features of computation adheres to an ideal of transparency in its code to features of the media that the computation and in what it codifies, and subordinates its enables. Call this the Frame Fallacy, after the subject matter to the achievement of a pre- mistake of inferring from the fact that a movie is established goal that requires no change in the made up of discrete frames, the conclusion that basic rules and symbols of the Turing machine’s the experience of watching a movie is the experi- own program (the ideal of “domination” or ence of a series of discrete frames. “administration”). For both analytic discourse For instance, Fred Evans makes observations and the Turing machine, the ideal is to trans- about the algorithmic character of computation form everything into an “effective procedure,” and infers from this that computer scientists and this is exactly the task of technocratic ration- and cognitive psychologists are in league with ality. In more historical terms, the Turing

99 Wesley Cooper

machine transforms the “clearness and distinct- heaved and folded up in a third. But clearly ness” dictum of Descartes into “imitatable by technological information is the most pro- the Turing machine.” (1993: 64) minent layer of the contemporary cultural landscape, and increasingly it is more of a At bottom, this argument is a bad pun. Evans is flood than a layer, a deluge that threatens to erode, suspend, and dissolve its predecessors. equivocating on “effective procedure,” between (1999: 2) cost-effective administration on one hand, and algorithm on the other. It is the same sort of mistake as supposing that, since the Bank of This has led some disciples of Borgmann to Montreal and the bank of the Saskatchewan River eschew all digitally recorded music, insisting are both banks, it must follow that they both on listening only to live performances. Another make financial transactions. Effective procedures example of inherence dystopianism leading to in the sense that interested Alan Turing are fea- blanket evaluations is Neil Postman’s Technopoly: tures of mathematical reasoning, not features of The Surrender of Culture to Technology, which administration of people. Evans’s mistaken in- indicts the United States as a “technopoly,” ference from features of computation to features along with “Japan and several European nations of the research communities that make use of that are striving to become Technopolies as them is egregiously abetted by his equivocation well” (1993: 48–9). A Technopoly does no less, on “effective procedure.” according to Postman, than eliminate “alternat- One reason to be wary of utopian or dystopian ives to itself in precisely the way Aldous Huxley inherence theories is that they encourage a tend- outlined in Brave New World” (1993: 48). ency toward blanket denunciation and renunci- An object lesson about the wholesale approach ation of the internet, or the blanket opposite, can be drawn from Richard Bernstein’s analysis when what is needed is a piecemeal evaluation of the father of dystopian theories of high tech- of this or that use of it, this or that tool that nology, Heidegger. In “Heidegger’s Silence?: is enabled by the internet metatool. A striking Ethos and Technology” Bernstein makes the case contemporary instance of the blanket approach that the great German philosopher’s brief but is the Montana philosopher Albert Borgmann’s active support of Hitler and the Nazis, during position, in Holding on to Reality, that digitally the 10-month period when he served as Rector generated information is incapable of making of the University of Freiburg between April 1933 a positive contribution to culture, but on the and February 1934, is symptomatic of a philo- contrary threatens to dissolve it, by introducing sophical failing that expresses itself in what he information as reality to compete with the picture said and did before and after those 10 months, of the world that is drawn from natural informa- notably in his silence about the Holocaust after tion (information about reality, as in weather the war, when there were no longer any serious reports) and cultural information (information doubts about the full horror of the Nazi regime. for reality, as in recipes for baking things). “But we are delivered over to [technology] in the worst possible way,” Heidegger writes in The Question Concerning Technology, “when we The technological information on a compact regard it as something neutral; for this concep- disc is so detailed and controlled that it addresses us virtually as reality. What comes tion of it, to which today we particularly like to from a recording of a Bach cantata on a CD is do homage, makes us blind to the essence of not a report about the cantata nor a recipe technology” (1977: 91). – the score – for performing the cantata, it is According to Bernstein’s account of the link in the common understanding music itself. between his biography and his philosophy, Information through the power of technology Heidegger conceals and passes over in silence steps forward as a rival of reality. the importance for the Greeks, specifically Today the three kinds of information are Aristotle, of phronesis, the state of the soul that layered over one another in one place, grind pertains to praxis. He refers to a discussion by against each other in a second place, and are Aristotle in Nicomachean Ethics that has “special

100 Internet Culture importance,” but his reference is partial and one- corpses in gas chambers and extermination sided, bringing out the role of techne in relation camps, the same as blockading and starving of to poiesis, as sketched above, but not tracking nations [it was the year of the Berlin blockade], the full discussion, which Aristotle summarizes the same as the manufacture of hydrogen in the following passage. bombs. (cited in Bernstein 1992: 130)

Then let us begin over again, and discuss these Bernstein understands this grotesque passage states of the soul. Let us say, then, that there as a natural expression of Heidegger’s reaction are five states in which the soul grasps the against the “correct” definition of technology truth [aletheia] in its affirmations or denials. as a neutral instrument which can be used for These are craft [techne], scientific knowledge benign ends of increased food production or [episteme], [practical] intelligence [phronesis], the malignant end of extermination of human wisdom [sophia], and understanding [nous]... beings. (cited in Bernstein 1992: 121) But if we focus on the essence of technology Bernstein asks, “Why should we think that the then these differences are “non-essential.” The response that modern technology calls forth is manufacturing of corpses in gas chambers more to be found by “re-turning” to techne and poiesis, fully reveals the essence of technology . . . rather than phronesis and praxis?” He objects Unless we fully acknowledge and confront the that Heidegger does not even consider this pos- essence of technology, even in “manufacturing sibility, writing that “[t]he entire rhetorical con- of corpses in gas chambers,” unless we realize struction of The Question Concerning Technology that all its manifestations are “in essence the seduces us into thinking that the only alternative same,” we will never confront the supreme to the threatening danger of Gestell is poiesis. It danger and the possible upsurgence of the excludes and conceals the possibility of phronesis saving power. (1992: 131) and praxis” (1992: 122). Bernstein urges that our destiny rests not solely with the thinkers and Bernstein concludes that the deleted passage is the poets who are guardians of the abode in not simply some insensitive remark but rather a which man dwells, but with the phronesis of necessary consequence of the very way in which ordinary citizens’ contribution to public life. The Heidegger characterizes Gestell, as an unconceal- possible upsurgence of the saving power may ment that claims man and over which he has be revealed in action (praxis) and not only in no control. He sets out a formulaic pattern in “poetic dwelling.” Heidgegger’s thinking, Bernstein asks again, “Why is Heidegger blind to those aspects of praxis and phronesis high- a pattern that turns us away from such “mun- lighted by Taminiaux, Gadamer, Arendt, and dane” issues as mass extermination, human Habermas?” He agrees with Habermas’s sugges- misery, life and death, to the “real” plight, the tion: that Heidegger is guilty of “a terrible intel- “real” danger – the failure to keep meditative thinking alive . . . It is as if in Heidegger’s lectual hubris” when he suggests that the only obsession with man’s estrangement from proper and authentic response to the supreme Being, nothing else counts as essential or true danger is to prepare ourselves to watch over except pondering one’s ethos . . . It becomes unconcealment. clear that the only response that is really Bernstein next draws attention to an unpub- important and appropriate is the response to lished manuscript of the 1949 lecture that became the silent call of Being, not to the silent The Question Concerning Technology, which con- screams of our fellow human beings . . . when tains the following passage that has been deleted we listen carefully to what he is saying, when from the published text. we pay attention to the “deepest laws of Heideggerian discourse” then Heidegger’s Agriculture is now motorized food industry – “silence” is resounding, deafening, and damn- in essence the same as the manufacturing of ing. (1992: 136)

101 Wesley Cooper

Bernstein’s analysis and conclusions suggests a of us. Those whom you would have in your moral critique of utopian and dystopian theories ideal world are also free to imagine and relocate, of internet culture. Although none of the theories perhaps to a world of their own imagining. There I have reviewed is so damned by its inherence could be an incessant churn of relocation, all arguments as Heidegger’s, which blinded him to worlds being ephemeral, or some stable worlds the specific evil of the Holocaust, yet a Postmanian might emerge in which everyone would choose to anti-Technopolist may be blinded in parallel remain. There will be no one in a stable associ- fashion to something good or bad about this ation who wants out, and no one will be in whose or that specific aspect of American culture; a presence is not valued by the others. Libertari- Borgmannian may be blinded to something anism may be bad politics, but its conception of specifically good or bad about some digitally utopia is a plausible model of the internet. generated artifact; and a gung-ho cybernaut of The claim that inherence instrumentalism the Leary persuasion may be blinded to the makes to being “value free” is provocative, defy- old-fashioned pleasures of embodiment. ing a post-Weberian tradition of deconstructing such claims with a view to revealing hidden value commitments, an argumentative strategy that bears Heidegger’s imprimatur, as noted above. Inherence Instrumentalism It may be helpful to clarify the claim with an analogy to a box of paints and a variety of paint- Turkle’s category of utilitarian interpretation ings made with them, some of them good paint- understands the internet as a tool. The version ings, some of them bad, some of them so-so. It scouted here under the rubric of inherence would be a logical error, a “category mistake” in instrumentalism interprets the internet as essen- Rylean terminology, to evaluate the box of paints tially a metatool for creating tools. This general as a good, poor, or so-so painting. It is not a idea derives from Robert Nozick’s discussion of painting at all. Classification of the internet as a a libertarian utopia in Anarchy, State and Utopia. metatool aims at a similar conclusion. Corres- Although he did not have the internet in mind, ponding to the variety of paintings in the ana- what he says there about a framework for utopia logy is the variety of content on the internet. transfers quite naturally to the internet, as well as None of this content is value free in the sense having greater plausibility there than in political that is being reserved for the internet as a philosophy for the real world. metatool. Content in the middle of the con- The internet does not have a culture of simu- tinuum from poor to good might be deemed lation, on this metatool account, because it is a value free in the sense that it excites no judg- tool for creating a variety of subcultures, some ments of praise or condemnation with respect to of which may fit Turkle’s description of internet this or that value; such internet content might culture, some of which will not, not to mention be described as bland. But the sense in which the variety of internet activity, like setting up a the internet is value free is not like this. Rather, webpage for lecture notes, that does not amount it is like the freedom of the box of paints from to creating a subculture. The internet is the Swiss being judged a good, bad, or so-so painting. It Army knife of information technology. is not a bland painting, and the internet is not Libertarians sometimes think of utopia in this bland internet content, on the inherence instru- way: ideally, everyone would be free – would mentalist account. Inherence dystopians and have the Lockean “natural right” – to migrate utopians purport to find something deeply good or emigrate as he or she chose. The worlds that or bad about the internet, but on an instrument- result from such to-ing and fro-ing they call alist diagnosis either they become so deep that associations. Acknowledging that there is no they loose touch with the truth, as illustrated by single world that’s everyone’s perfect cup of tea, the attempt to tie the computer inevitably to a the libertarian is inspired by a utopia which is a society of technocratic administration, or else they set of possible worlds, with permeable borders, are guilty of a part–whole fallacy, judging the in which one world is the best imaginable for each whole internet by some of its uses. Even if all

102 Internet Culture uses had some bad value or effect X, that would channeled by physical neighborhood, and with the ground only a balance-of-reasons judgment that development of automatic translation software a one should or should not use the internet, great obstacle to cross-cultural friendships, namely depending on whether X outweighs the good lack of a common language, is being removed. value or effect Y. New motivations and organizational structures for work are being discovered on the internet, notably the “open source” initiative associated with Linus Torvalds, Eric Raymond, and a legion Conclusion of true hackers, showing how psychic rewards can replace monetary ones in high-quality soft- To illustrate once more the dystopian/ ware development within the internet milieu; if instrumental/utopian continuum and the work is understood as paid employment, con- balance/inherence vectors that can be traced tributions to such software development is not by reference to it, consider the changes being work, whereas if it is understood as activity that wrought in work and leisure by the computer is instrumental to some further end, such as a revolution. Offices have been transformed by new Linux kernel, it is work calling for a high the computer over the past two decades, while level of skill. This raises the question whether web surfing, computer gaming, and internet the suffusion of IT into work and leisure will chatrooms have become significant leisure activit- eventually lead to their transcendance in “mean- ies. As recent events in Afghanistan testify, even ingful work” that is pursued because of its war, that most regrettably necessary form of work, intrinsic motivations, not extrinsic ones such as must be fought with sophisticated information money. Is there something about information technology in order to achieve success in the technology that makes it inherently amenable to battlefield of the twenty-first century; the leisure meaningful work? The case could be made that activity of correspondence is migrating from the will do so by following a negative and a positive pen and the typewriter to computer email, a trans- path. The via negativa is the elimination of ition from manipulating matter to manipulating “agonistic work,” work that one would gladly digital bytes that is as significant as any preceding avoid if it weren’t necessary. The via positiva is revolution in communication technology. Despite the creation of attractive environments in which the uneven track record of “dot.coms,” business one is always able to work “just as one has a activity on the internet is starting to take giant mind.” Marx had such an environment in mind strides; new communities are being formed on when he speculated about the higher stages of the internet, like Multi-User Dungeons (MUDs), communism, in which the division of labor char- Internet Relay Chat (IRC), and so on, online acteristic of capitalism has been overcome and “third places” between work and home that one’s distinctively human powers are fully real- allow users of the Net a respite from the demands ized, without the compulsion of necessity. In of office and household. Work as traditional as The German Ideology he made the point like this: farming is becoming reliant on the boost to organization and efficiency that computers make For as soon as the distribution of labor comes possible; games like chess, go, poker, and bridge into being, each man has a particular, exclus- are just as likely to play out on the internet as in ive sphere of activity which is forced upon him physical spaces. Computers and the internet are and from which he cannot escape. He is a opening up new employment opportunities, new hunter, a fisherman, a shepherd, or a critical tools, and new media for artists; correspondingly, critic, and must remain so if he does not want creating and maintaining a personal webpage has to lose his means of livelihood; while in com- become an art that many pursue in their free munist society, where nobody has one exclus- time. Telecommuting and teleconferencing are ive sphere of activity but each can become becoming more widespread, with potentially accomplished in any branch he wishes, society enormous implications for city design and trans- regulates the general production and thus makes portation systems; making friends is no longer it possible for me to do one thing today and

103 Wesley Cooper

another tomorrow, to hunt in the morning, to be monitored carefully in order to make prac- fish in the afternoon, rear cattle in the even- tically wise choices about its use. As the “Mount ing, criticize after dinner, just as I have a mind, Carmel Declaration on Technology and Moral without ever becoming hunter, fisherman, Responsibility” observed in 1974 in its eighth shepherd, or critic. (Tucker, ed., 1974: 124) article, “We need guardian disciplines to mon- itor and assess technological innovations, with Add to Marx’s flight of fancy the thought that especial attention to their moral implications” information technology will be the means by (Hester & Ford 2001: 38). No technology is which “society regulates the general production,” morally neutral if that means freedom from moral and you have a form of inherence utopianism evaluation. But there is no inherent reason why about IT. However, given the failure of com- that evaluation should be pro or con. mand economies in real-world tests such as the USSR, Heideggerian inherence dystopianism may recommend itself instead. IT will have taught Acknowledgments us, on this account, to view nature as so much “standing reserve” and not even the overcoming I wish to thank Luciano Floridi for invaluable of the division of labor will protect us from a philosophical and editorial comments about mental architecture that we should want to avoid. earlier drafts of this chapter. Another inherence-dystopian option argues that a core value of our civilization, to which our self-respect is inexorably tied, is agonistic work; References IT, by showing us how to eliminate such work, will have the unintended consequence of remov- Bernstein, Richard. 1992. The New Constellation: ing the bases of our self-esteem. The aspect of The Ethico-political Horizons of Modernity/ technological determinism is noticeable in these Postmodernity. Cambridge, MA: MIT Press. [For three options. An alternative is the outlook that graduates; a lucid exposition and critique of advocated in The Open Society and Heidegger’s philosophy of technology.] its Enemies and elsewhere, which views with Birkerts, Sven. 1994. The Gutenberg Elegies: The Fate suspicion ideas about the necessity of history’s of Reading in an Electronic Age. London: Faber unfolding and recommends instead that oppor- & Faber. [For the educated public. It evokes the tunities for change be monitored for unintended depth and inwardness that reading can obtain, consequences, so that choices can be made that and expresses concern about the future of read- reflect knowledge of where change is going ing in an information age.] wrong. The current debate about genetically Borgmann, Albert. 1999. Holding on to Reality. modified foods is an example of such monitor- Chicago: University of Chicago Press. [For gradu- ing; it also illustrates a tendency for inherence ates and ambitious undergraduates. He makes an engaging case for focal values in life that keep us voices to emerge at the dystopian and utopian close to nature and our communities.] extremes. Deutsch, David. 1997. The Fabric of Reality. The Popperian outlook may be viewed as Harmondsworth: Penguin. [For graduates and contributing to an inherence-instrumentalist ambitious undergraduates. He brings virtual interpretation of internet culture, wherein the reality to the center of his interpretation of metatool character of the technology acknow- quantum mechanics.] ledges dystopian fears and utopian hopes with Evans, Fred. 1993. Psychology and Nihilism: A respect to particular content. At the metalevel, Genealogical Critique of the Computational Model however, the internet is neither good nor bad of Mind. New York: SUNY Press. [For graduates. nor in-between; at the level of specific content, This is an interpretation of cognitive psychology it may be any of these things. The Popperian inspired by Nietzsche.] contribution theorizes the internet, not as his- Heidegger, Martin. 1973. “Overcoming meta- torical inevitability to be deplored or valorized physics.” In The End of Philosophy. New York: holus-bolus, but rather as a locus of possibilities, Harper and Row. [For graduates. Bernstein draws

104 Internet Culture

on this essay for his interpretation of Heidegger’s New Haven and London: Yale University Press. philosophy of technology.] [For graduates. This is an historian’s scholarly ——. 1977. “Letter on humanism.” In Martin account of the convergence between humans and Heidegger, Basic Writings. New York: Harper and machines.] Row. [For graduates. Bernstein draws on this Nozick, Robert. 1974. Anarchy, State and Utopia. essay for his interpretation of Heidegger’s philo- New York: Basic Books. [For graduates. This is a sophy of technology.] remarkable defense of libertarianism that poses ——. 1977. The Question Concerning Technology. hard questions for liberals and Marxists, while New York: Harper and Row. [For graduates. This painting a minimal state as a veritable utopia.] is the inspiration for many interpretations of Popper, Karl. 1971. The Open Society and its En- technology in recent decades.] emies. Princeton: Princeton University Press. [For Heim, Michael. 1993. The Metaphysics of Virtual the educated public. This is his contribution to Reality. Oxford: Oxford University Press. [For the war effort in the Second World War, tracing undergraduates. He writes vividly about how we communism and fascism to philosophical roots are being transformed by information technology, in Plato, Hegel, and Marx.] and about the long-term prospects for virtual Postman, Neil. 1993. Technopoly: The Surrender of reality.] Culture to Technology. New York: Knopf. [For Hester, D. Micah and Ford, Paul J. 2001. Com- the educated public. This is an articulate call puters and Ethics in the Cyberage. Englewood, to arms against the influence of technology on NJ: Prentice-Hall. [For undergraduates. This is culture.] typical of many accessible anthologies of essays Slouka, Mark. 1995. War of the Worlds. New York: on technology and computers from authors in a Basic Books. [For the educated public. He argues variety of fields.] that computer games like MUDs are making us Jameson, Fredric. 1991. Postmodernism, or, The lose touch with reality.] Cultural Logic of Late Capitalism. Durham, NC: Tipler, Frank J. 1994. The Physics of Immortality. Duke University Press. [For graduates. This is an New York: Doubleday. [For graduates. Deutsch important source of theory about postmodernism.] (see above) draws on Tipler for some of his Kelly, Kevin. 1994. Out Of Control: The New Bio- cosmological speculations.] logy of Machines, Social Systems, and the Economic Tucker, Robert C. 1972. The Marx–Engels Reader. World. New York: Addison-Wesley. [For under- New York: W. W. Norton & Co. [For graduates. graduates and the educated public. This is imagin- This is typical of many anthologies of Marx’s and ative, mind-stretching scientific journalism about Engels’ writings.] convergence of the natural and the human-made.] Turkle, Sherry. 1995. Life on the Screen. New York: Lovelock, James. 1988. The Ages of Gaia: A Bio- Simon & Schuster. [For the educated public. This graphy of our Living Earth. New York: W. W. is an indispensable interpretation of the cultures Norton. [For undergraduates. Kelly (see above) that have grown up around the computer.] weaves Lovelock’s ideas into his vision of where Woolley, Benjamin. 1992. Virtual Worlds. Oxford: the Earth is heading.] Blackwell. [For the educated public. This is a witty Mazlish, Bruce. 1993. The Fourth Discontinuity: The and philosophically informed series of essays on Co-evolution of Humans and Machines. matters pertaining to virtual reality.]

105 Dominic McIver Lopes

Chapter 8 Digital Art Dominic McIver Lopes

traditional art media and practices. Art appreci- Introduction ators, once passive receptacles of aesthetic delight, may finally participate actively in the art process. Artworks are artifacts, their making always Pronouncements such as these spring less from involves some technology, and much new art careful study and more from marketing forces exploits and explores new technologies. There and simple misunderstandings of a complex and would be no novels without inexpensive print- multifaceted technology. An accurate conception ing and book binding. The modern skyscraper is of the nature of digital art and its potential may a product of steel manufacture. Jazz married the channel without dousing the enthusiasm that European technology of the diatonic scale to attends any innovation. At the same time, it coun- African rhythms. A factor in the origins of Im- terbalances some cultural critics’ jeremiads against pressionism was the manufacture of ready-made digital art. Radical antihype often depends for its oil paints in tubes, which facilitated painting rhetorical force on our reaction to hype. When outdoors in natural light. As soon as computers we are told that electronic music or fractal art became available, they were used to make art – or virtual-reality goggles are the future of art, we the first computer-based artwork was created as are given good reason to doubt the credibility early as 1951 (Reffen Smith 1997: 99) – and of our informant and this doubt may engender since then the body of digital artworks has grown blanket skepticism about digital art. But while by leaps and bounds. But although the first philo- most digital art is admittedly dreadful, this does sophical paper on “cybernetic art” appeared in not show that it never has value or interest. The 1961 (Parkinson 1961), philosophers are only correct lesson to draw is that we should proceed now beginning to address in depth the ques- with caution. tions raised by digital art. What is digital art? This chapter is divided into three sections. The How, if at all, is it new and interesting as an art first reports on the use of computers as tools in medium? Can it teach us anything about art as art-making. The second describes some artworks a whole? that capitalize on the distinctive capabilities of Answering these questions provides an antidote digital computers and digital networks. To make to the hype that frequently attaches to digital sense of these works we must define digital art art. We hear that computer art is overhauling and consider whether it is a new art medium. The our culture and revolutionizing the way we think third reviews the use of computers as instruments about art. It frees artists from the materiality of that yield general insights into art-making. This

106 Digital Art three-section division is one case of a useful way however, the computer merely realizes efficiencies of thinking about any use of computers, not just in art-making or art distribution. Digital tech- in the arts. For example, a philosophy of artificial nology, including digital networking and the intelligence might begin by discussing computers compact disk, is used to store music, as did vinyl as cognitive aids (e.g. to help with calculations), records, but in a format that is considerably more then consider whether computers possess a kind portable and transmissible without introducing of intelligence, and close with a discussion of noise. Musical recordings that once required the use of computer models of the human mind live musicians, a studio, and several technicians, in cognitive psychology. can now be made at a fraction of the cost by one person in her garage with a keyboard and a computer. Computers sometimes make it easier for artists 1 Making Art Digitally to work and, by reducing the technical demands of the craft underlying an art medium, they some- The digital computer has occasioned two quite times make it easier for untutored novices to distinct kinds of innovation. It has automated make art. In addition, some uses of computers and sped up many tasks, especially routine ones, in making and distributing art cause artworks to that were once relatively difficult or slow. It has have properties they would not otherwise have. also made some activities possible that were pre- The use of typewriters by some modernist writers viously impossible or else prohibitively difficult. in the early twentieth century influenced the Most discussions of digital art are captivated by character of their writing. Relatively inexpensive the latter kind of innovation; however, the im- digital movie editing encourages film-makers to pact of the former should not be ignored. If art experiment with faster pacing and more complex always involves some craft then the practice of sequencing. Poor musical technique is now no that craft may incorporate the use of computers. barrier to recording music and distributing it Moreover, a clear view of the uses of computers worldwide from one’s desk. Tod Machover’s as art-making tools can help crystallize a con- hyperinstruments can be played in interesting ception of the kind of innovation that involves ways – some, for instance, are soft toys whose opening up new possibilities for art. sound depends on how they are squeezed – and When the craft underlying an art medium has can be used to make music whose sound reflects practical, non-art applications, digital technology its instrumentation (see ). What properties artworks of an of that craft easier and more efficient. Here the era possess depends in part upon the techno- use of computers in making art simply extends logies employed in making art during that era. their use in other areas of human endeavor. The Art’s history is partly driven by technological first computer imagining technologies, output innovation. plotter drawings, were developed for engineering While the kinds of innovations discussed so and scientific uses, but were quickly adopted by far generate artworks with new properties, they artists in the early 1960s. It hardly needs to be neither beget new art media nor change our pointed out that word-processors have proved standards for evaluating artworks. An aesthetic as much a boon to literary authors as to office evaluation of a performance of a pop song need managers. Software created for aeronautical not take into account whether the recording of design paved the way for the stunning, complex it is analog, digitally remastered, or direct-to- curves that characterize Frank Gehry’s recent digital, and whether that recording is played back buildings, notably the Guggenheim Bilbao. Since from a vinyl record, a reel of magnetic tape, a digital sound processing and the MIDI protocol compact disk, or an MP3 file. The relative ease were developed specifically with music in mind, of online publication means that much more is music is an exception to the rule that digital art published, but the nature of literature and its technologies adapt technologies fashioned for aesthetically relevant properties endure. A novel some non-art purpose. In each of these cases, is a novel and is as good or as bad as it is whether

107 Dominic McIver Lopes it is printed and bound into a book or emailed not necessary if our aim is to characterize digital to one’s friends. It is important to recognize how art. We may assume that digital art is a kind of computers have found their way into artists’ art and concentrate our efforts on what distin- studios – or made the resources of a studio guishes it from other kinds of art. And although more widely and cheaply available. But this is no we may proceed with either an evaluative or revolution in the nature of the arts. descriptive characterization of art, it is wiser to characterize digital art as art in the descriptive sense, so as not to beg any questions about its quality. 2 The Digital Palette The assumption that digital art should be considered art, even when art is characterized Computers ease the performance of some tasks descriptively, is not uncontroversial. One theorist but they also equip us to undertake new tasks. asks of digital graphic art, Exploiting this, artists may invent new varieties of art, including what we may designate the whether we should call it “art” at all. In treating “digital arts.” One question to be answered is it as art we have tended to weigh it down with what is characteristic of digital art media. Theor- the burden of conventional art history and art ists typically propose that digital art is novel in criticism. Even now – and knowing that the use two ways, the first deriving from virtual-reality of computing will give rise to developments technologies and the second deriving from the that are as far from conventional art as com- capacity of computers to support interactivity. puters are from the abacus – is it not too late Something must be said about what virtual real- for us to think of “computer art” as some- ity and interactivity are, and it will be helpful to thing different from “art”? As something that describe some artistic uses of each. But since our perhaps carries with it parallel aesthetic and goal is to devise a theory of digital art, it is prudent emotional charges but having different and to begin by considering what an adequate theory more appropriate aims, purposes and cultural of any art medium should look like. baggage? (Lansdown 1997: 19) Art media are species of a genus that com- prises all and only works of art. This genus can There are two reasons that this objection should be characterized either evaluatively or descript- not give us pause, however. Even granting that ively. According to evaluative characterizations, what we count as art depends on a welter of works of art are necessarily good as works of art, social practices and institutions, art status is not and “art” is an essentially honorific term. Some a matter for deliberate legislation. More import- theorists who write about digital art (especially antly, the objection misses an important fact its critics) have this characterization in mind. about art. We never judge or see an artwork Brian Reffen Smith, himself a computer artist, merely as art but always as some kind of artwork dismisses much of what goes under the banner – as belonging to some art medium. If digital of digital art as “graphic design looking a bit art is art, it remains an open question whether like art” (Reffen Smith 1997: 102). He does not it is an art medium that inherits the history, allow that the works in question are poor art, purposes, standards of criticism, and “cultural for art, he assumes, is necessarily good as art. baggage” of any other art media. Descriptive conceptions of art allow that some In his classic paper “Categories of Art,” works may be failures as works art and yet Kendall Walton maintains that we perceive every deserve the name, so that to call something “art” work of art as belonging to some category of is not necessarily to commend it but merely to art, where art categories are defined by three acknowledge its membership in the class of kinds of properties: standard, variable, and con- artworks, good and bad. It is a matter of con- trastandard properties (Walton 1970). Standard siderable controversy how to characterize the properties of works in a category are ones in conditions of membership in this class (see Carroll virtue of which they belong to the category; lack- 2000, Davies 2000). Fortunately, consensus is ing a feature standard for a category would tend

108 Digital Art to disqualify a work from the category (“having evaluate works of art are fluid and can be shaped an unhappy ending” is a standard property of by technological forces. tragedies). We discriminate among works in a As the examples given indicate, there are category with respect to their variable properties several schemes of categories into which artworks (“featuring an indecisive prince” is a variable can be portioned. One scheme comprises the property of works in the category of tragedies). art media – music, painting, literature, theater, Contrastandard properties of works with respect and the like. Another scheme comprises genres to a category are the absence of standard fea- of art, such as tragedy and melodrama; works tures in respect of the category. A tragedy may in these categories may belong to different art have the contrastandard feature of having an media. A third scheme, that of styles, also cuts ending that is not unhappy. But why perceive a across media and genres. There are postmodernist work in a category when it has properties that parodies and postmodernist comedies; some of are contrastandard with respect to that category? the former are musical while others are architec- For Walton, at least four factors determine what tural and some of the latter are literary while category we should perceive a work as belong- others are pictorial. How, then, should we char- ing to: the work’s having a relatively large number acterize the scheme of art categories that com- of properties that are standard for the category, prises the art media? For it is within this scheme the artist’s intention or expectation that the work that we might expect to make room for a cat- be perceived as in the category, the existence egory of digital art. of social practices that place it in the category, One way to characterize the art media is with and the aesthetic benefits to be gleaned from reference to their physical bases. Musical works perceiving the work as being in the category – a are sounds; pictures are flat, colored surfaces; drama with a happy ending that is inventive, and theatrical performances consist in human even shocking, when viewed as tragedy, may seem bodies, their gestures and speech, together with old hat when viewed as comedy. the spaces in which they are located. Indeed, we Art categories provide a context within which use the term “medium” ambiguously to name we appropriately interpret and evaluate artworks. an art form and its physical embodiment. The To appreciate a work of art one must know “medium of pictures” can denote the pictorial art how it resembles and differs from other works of form or it can denote the stuff of which pictures art, but not every resemblance or difference is are made – oil paint, acrylic, encaustic, ink, and aesthetically significant. There is only a point to the like. Nevertheless, ordinary usage notwith- noticing differences among works that belong standing, we should distinguish art media from to a kind and to noticing similarities among works what I shall call, following Jerrold Levinson, their when the similarity is not shared by everything “physical dimensions” (Levinson 1990: 29). The of its kind. Acid jazz differs from opera, but to reason is that works in different art media may appreciate John Scofield’s “Green Tea” as a work share the same physical dimension and works in of acid jazz it is not enough to hear how it the same art medium may have different physical differs from Rigoletto – one must recognize how dimensions. The case of literature is instructive. it differs from other works of acid jazz. Literary works can have many physical dimen- Moreover, what properties are standard, sions, for they can be recited from memory as contrastandard, and variable with respect to a well as printed on paper. Moreover, when novels category is subject to change. Suppose that it is are printed on paper they have the same physical a standard property of photography that photo- dimension as many pictures, but although some graphs accurately record visible events. As the artworks are both literary and pictorial (visual use of software for editing rasterized photographs poems for instance), printed volumes of Lady increases, this may become a variable property Chatterley’s Lover are not pictures. of the category. Digital image doctoring may The medium of literature is independent of thereby change how we see all photographs any particular physical dimension because works (Mitchell 1992; Savedoff 1997). The lesson is of literature are made up of bits of language and that contexts within which we appreciate and language is independent of any particular physical

109 Dominic McIver Lopes dimension. Yet there is a sense, however makes possible an art medium with distinctive stretched, in which every art medium comprises properties. a “language,” understood as embodied in a set Some claim that virtual reality uniquely gen- of practices that govern how the materials of the erates an illusion that the user is in the computer- medium are worked. This is all we need in order generated environment, perceiving it. But what to characterize the art media. Artworks standardly is meant by “illusion”? On the one hand, it does belong to the same art medium when and only not appear that even the most sophisticated when they are produced in accordance with a set virtual-reality set-ups normally cause their users of practices for working with some materials, to believe, mistakenly, that they are part of and whether physical, as in sculpture, or symbolic, as perceiving the computer-generated environment. in literature. These materials together with the On the other hand, any imagistic representation practices of shaping them determine what works elicits an experience like that of perceiving the are possible in an art medium. Call the materials represented scene, even images (e.g. outline and the practices governing how they can be drawings) that are far from realistic. Virtual worked the art medium’s “palette.” reality could be redescribed without loss as The digital palette comprises a suite of tech- “realistic imaging” and classified with other real- nologies and ways of using them that determine istic imaging such as cinema or three-imensional what properties digital artworks can possess, in- (stereoscopic) cinema. If virtual reality offers any- cluding those properties that are standard and thing new it is the possibility for interaction with variable with respect to the category of digital the occupants and furniture of the computer- art. Since computers can be programmed to serve generated environment. As Derek Stanovsky puts indefinitely many tasks, the digital palette is un- the point, “computer representations are dif- bounded. But we can discern, if only in outline, ferent because people are able to interact with some of the potential of the digital palette by them in ways that resemble their interaction with canvassing some typical cases of innovative the genuine articles” (see Chapter 12, VIRTUAL digital art. We should keep in mind throughout REALITY). Virtual reality as realistic imaging that the point of thinking of digital artworks as should not be confused with interactivity. belonging to a digital art medium is that we The interactivity of computers capitalizes on properly appreciate and evaluate digital artworks their ability to implement complex control struc- only when we perceive them within the category tures and algorithms that allow outputs to be or medium of digital art, as it is characterized by fine tuned in response to different histories of the digital palette. inputs. What properties a work of interactive One digital technology that is much discussed digital art possesses depends on the actions of in recent years among media theorists and that its user. The point is not that every user has a is thought to engender a new digital art form is different experience when engaging with an art- virtual reality. This is standardly defined as a “syn- work – that is arguably true of our experiences thetic technology combining three-dimensional of all artworks. The point is rather that the struc- video, audio, and other sensory components to tural properties of the work itself, not just how achieve a sense of immersion in an interactive, our experience represents the work, depend on computer-generated environment” (Heim 1998: how we interact with it (Lopes 2001). Defined 442). The vagueness of this definition accurately in this way, digital interactive art is something reflects the wide range of technologies that are new and it exists precisely because of the special called virtual reality. “Three-dimensional video” capabilities of computing technology. can denote the use of perspective animations A hypertext story, such as Michael Joyce’s to represent three-dimensional scenes on two- widely read Afternoon, A Story of 1987, is inter- dimensional computer monitors, often with ex- active because it allows the reader to follow aggerated foreshortening (as in most computer multiple narrative pathways, so that the story games), or it can denote the use of stereoscopic goes differently on each reading. But there is no animations viewed through virtual-reality goggles. reason that hypertext need involve hyperlinked The question to ask is whether virtual reality text that the user selects. Simon Bigg’s Great

110 Digital Art

Wall of China of 1999 (at ) transforms a display work given performances that are, qua perform- of the text of Kafka’s story in accordance with ances, good but not redeeming. movements of the user’s mouse. The reader of According to Timothy Binkley, the aesthetic- Jeffrey Shaw’s 1989 Legible City sits on a fixed ally relevant features of a predigital artwork are bicycle which he or she uses to navigate a land- features of its physical embodiment (Binkley 1997, scape built of words, each route through the 1998a, 1998b). To make an artwork is tradition- landscape telling a story about a city. Indeed, ally to “maculate” some physical substance, shap- the input of users to interactive artworks can ing it into the work. But digital artworks are not take a variety of forms: gesture, movement, physical objects, for the computer “computes sound, drawing, writing, and mere physical pres- abstract numbers with mathematical algorithms ence have all been used. Nor is interactive art rather than plying physical material with manual always narrative in form. Avatar technologies and implements” (Binkley 1998a: 413). Instead of synchronous remote puppeteering enable users making things, digital artists manipulate data to act in represented performance spaces. Peter structures; they “mensurate” symbols instead of Gabriel’s Xplora 1 CD-ROM of 1993 allows its “maculating” physical stuff. Of course, Binkley owner to remix Gabriel’s music so that it has realizes that the data structures making up di- different sound properties from one occasion of gital artworks always take some physical, usually interaction to the next. Robert Rowe’s Cypher electronic, embodiment; his point is that the data and George Lewis’s Voyager are computer pro- and its structure is independent of any particular grams that improvise music in real time as part physical embodiment. For this reason digital of an ensemble that includes human musicians. art “bears no telltale traces of the magnetism, Since what music the computer makes depends electricity, or cardboard that might happen to on what the other players in the ensemble do, the host its abstract symbols” (Binkley 1998b: 48). computer is as interactive as musicians jamming Digital artworks are therefore types. Their aes- with each other. thetically relevant features are not features of One way to see what is special about the works physical objects. They are indefinitely reusable just described is to consider their ontology. and can be copied with perfect accuracy (think Artworks can have, broadly speaking, one of two of a digital image sent by email from one person . Some artworks, paradigmatically to many others). Binkley concludes that digital paintings, have a unitary ontology: the work just art diminishes the importance of art’s physical is the painting, a spatio-temporally bounded par- dimension (Binkley 1997: 114; Binkley 1998b: ticular. Multiple-instance artworks, paradigmatic- 50). It is, he writes, “an art form dedicated to pro- ally works of music and literature, have a dual cess rather than product” (Binkley 1998a: 413). ontology: they are types whose instances are The claim that digital artworks are types is tokens. Most musical performances, for example, instructive, as is the observation that they are for are tokens of types that are musical works. The this reason indefinitely reusable and perfectly work type determines the properties which any- reproducible. Also instructive, however, are two thing must possess in order to count as instances related mistakes in Binkley’s account. Binkley’s of it, yet we apprehend the work through its first mistake is to take painting’s ontology as instances. In the case of music, we typically paradigmatic of all art – that is, by assuming abstract the musical work from performances that all nondigital artworks are physical objects. of it by stripping from them properties of the Literature, as we have seen, is a clear counter- performances themselves. This explains how it is example. Musical works, if they are types tokened in possible for a work and its instances to have individual performances or playings, are another different as well as shared properties, especially counterexample. When I listen to a performance different aesthetic properties. We evaluate per- of “Summertime” I am hearing two things. One formances as aesthetic objects in their own right is the performance, which is a physical , and and yet we evaluate a work performed without the second is the song itself, which is not ident- thereby evaluating any performance of it. A good ical to the performance though I apprehend its

111 Dominic McIver Lopes features by listening to the performance. The Saltz interactive works are not tokens of aes- case of music indicates Binkley’s second mistake. thetically interesting types. However, the virtue From the fact that digital artwork types are of the application of the type–token distinction nonphysical it does not follow that their tokens to art is that it allows for dual objects of aesthetic are not physical. Performances of “Summertime” attention. We usually attend simultaneously to are physical events and our aesthetic interest in properties of a performance qua performance and them is partly an interest in their physical qual- to properties of the work performed. The fact ities. Binkley thinks of a computer as a central that we direct our attention upon interactive processing unit and a digital artwork as the data processes, or upon our own actions as interactors, structure a CPU processes. But this ignores two does not show that we cannot and do not simul- additional and essential components of the taneously attend to properties of a work type computer, the input and output transducers. A with which we are interacting. Saltz is right that digital image is a data structure but it is tokened there is no interactive work type understood as only by being displayed on an appropriate what is indicated by a script or score. But it does device, usually a printer or monitor. Indeed, our not follow that we cannot descry features of an aesthetic interest in the image is an interest above interactive work type through instances of inter- all in properties of the physical embodiment of action with it. The contours of the work type are its tokens. drawn by what interactions it makes possible. David Saltz identifies three design elements Afternoon is many stories, but it is important to essential to digital interactivity: a sensing device know what set of stories it tells and how: these (such as a keyboard or mouse) that transduces give access to properties of Afternoon itself, not user actions as inputs, a computational process the individual stories our interactions with it that systematically relates inputs to outputs, and generate. Moreover, we miss something import- a display mechanism that transduces outputs into ant if we do not view interaction instances as something humanly perceptible (Saltz 1997: 118). instances of a work type, since to fully appreciate All three elements must be in place in order for an interaction as an interaction, one must regard an interactive piece to vary in its content or ap- it as means of discerning the work’s properties. pearance with human interaction. For this reason, As one commentator puts the point, “the inter- Saltz models the ontology of interactive art on active art experience is one that blends together that of performance art. An interaction is perform- two individualized narratives. The first is the story ative, according to Saltz, “when the interaction of mastering the interface and the second is about itself becomes an aesthetic object . . . interactions uncovering the content that the artist brings to are performative to the extent that they are about the work” (Holmes 2001: 90). their own interactions” (Saltz 1997: 123). The Interactive work instances are not tokened by aesthetically relevant properties of performative performance or playing (as in live and recorded interactions are properties of the interactor in the music) and they are not tokened by recital or work, who plays a role in the interaction’s unfold- printing (as in literature); they are tokened by ing. But there is no work type of which individual our interaction with them. The way instances interactions are tokens since the interactions are of an interactive work are tokened cannot be unscripted, and in the performing arts it is the modeled on the way musical or literary works script (or score or choreography) that identifies are tokened. In place of the score, the script, individual performances as tokens of one work and the text we have the individual user’s inter- type. Saltz infers that “to interact with a work of action (Lopes 2001). This is one way of seeing computer art does not produce a token of the what is new about interactive digital art. It gives work the way performing a dramatic or musical a role to its user, not just in interpreting and work does” (Saltz 1997: 123). experiencing the work but in generating instances Neither Binkley’s nor Saltz’s view adequately of it, that users of no other art media enjoy. An describes the ontology of interactive digital art. interactor tokens an interactive artwork in a way According to Binkley, only digital work types that a reader or spectator of a non-interactive are objects of aesthetic attention; according to artwork does not.

112 Digital Art

Interactivity, unlike virtual reality, is distinctive by developing computer models of those pro- of the digital palette, but not all digital art is cesses, they may learn about the cognitive under- interactive. There are many rather more mun- pinnings of art-making by building art-making dane functions that computers perform and that computers. Computers have been programmed provide resources for the digital palette. Word- as a means to learn about drawing, musical com- processors routinely check the spelling of docu- position, poetic writing, architectural style, and ments: Brian Reffen Smith has created artworks artistic in general. by first running a text in English through a French One may immediately object to the viability spell checker, which substitutes orthographically of this enterprise. Artworks are necessarily arti- similar French words for the English originals, and artifacts are the products of intentional and then translating the French words back into action, but if “art”-making computers have no English (Reffen Smith 1997: 101–2). So-called intentions, then they cannot make artworks. If interface artworks are applications that change they cannot make artworks, it is pointless to use the way familiar graphical user interfaces work. them to study art-making processes. The objection I/O/D’s Web Stalker provides an alternative, does not assume that no computers or robots exploded perspective on websites, for instance can have intentions. It assumes only that the com- (see ). Like much art puters that have been programmed to make “art” of the past century that takes as one of its main are not intentional agents – and this is a plausible subjects the technical basis of its own medium, assumption. The drawing system described below some digital art uses digital technologies in order can be downloaded from the internet and installed to represent or draw our attention to features of on a computer that can otherwise do nothing the digital art medium. more than send email and word-process. Early explorations of a new medium tend to Granting that artworks are intentionally made imitate the media from which it sprang. Photo- artifacts, two replies can be made to this objec- graphy aspired at first to the look of painting tion. The first challenges the objection directly and it was only after several decades that photo- by arguing that computer-made “art” is art graphers made unabashedly photographic photo- indeed. Typical acts of art-making involve two graphs. One explanation for this is that a new intentions: an artist intends to make an object medium must establish its status as art by asso- that has certain intrinsic properties (e.g. a given ciating itself with a recognized art medium. arrangement of colors, a meaning) and further Another explanation is that it is difficult to dis- intends, typically through the realization of the cern the full potential of a medium’s palette in first intention, to make a work of art. Distin- advance of actually using it to make art. What- guishing these intentions makes sense of some ever the explanation, it is only with time that we atypical acts of art-making. An artist selects a can expect digital art to look less like other kinds piece of driftwood, mounts it, and labels it of art and to acquire a character of its own. This (alluding to Duchamp’s snow shovel) Notes in process involves coming to see what standard Advance of a Broken Arm. If Notes is a work and variable properties characterize the digital of art, it is a work of art in the absence of an medium and how they are determined by the intention to create an object with the physical digital palette. It culminates in our evaluating features possessed by the driftwood. Qua drift- digital art on its own terms, as digital art. wood, the object is not an artifact, yet it is an artifact qua artwork, since it is mounted and displayed with the intention that it be a work of art. We may view a drawing made by a computer 3 Computing Creativity as, like the driftwood, shaped by a force of nature, and yet deem it art since we intend that Making art is a cognitive activity, as well as a it be displayed as art. The second reply concedes physical and a social activity. Just as philo- that computer-made “art” is not art but suggests sophers and behavioral scientists study cognitive that it is quasi-art instead. Computer drawing is processes such as vision or language acquisition sufficiently like human drawing that we can use

113 Dominic McIver Lopes the former to study the latter. We cannot use to a printer or a plotter. It has also been supplied what a computer does to study that part of the with a set of “cognitive primitives” – the basic art-making process that depends on agency or elements of line pattern and coloration that form on social institutions, but that is no limitation the universal building blocks of pictures. A set of we need worry about. behavioral rules governs how the system deploys Early experiments in computer creativity ex- the cognitive primitives in response to feedback tend a venerable tradition of automatic art. Wind from the work in progress. Finally, a second set of chimes or aeolian harps are designed to make behavioral rules directs the system’s work in light music, but the particular music they make is not of knowledge of how things look in the world – composed. Humans can be involved in making knowledge, for instance, of human anatomy. automatic art when they do nothing more than While these rules might be devised so as to pro- implement an algorithm. Mozart’s “Musikalisches duce only realistic images in canonical perspective, Würfelspiel” requires its players to role dice that AARON is able to produce images that fit into determine how the music goes. In the surrealists’ a variety of human drawing systems, including game of Exquisite Corpse, each player draws on those favored by children of different ages. part of a surface the rest of which is blocked from AARON models an isolated artist, one who view, making part of an image that, as a single works outside a drawing tradition. David Cope’s image, nobody drew. During the 1950s and EMI is designed to write music that mimics music 1960s, the heyday of “systems art,” composers in the style of historical composers on the basis such as Iannis Xénakis and John Cage created of “listening” to a selection of their work (Cope algorithms for music generation that were im- 1991). EMI’s top-level algorithm comprises six plemented on computers. A currently popular steps: encoding input works by a target com- form of automatic art is genetic art, in which a poser into a format it can manipulate, running a computer randomly propagates several mutations pattern matcher on the input, finding the patterns of a form, of which humans select one, that pro- that make up the composer’s stylistic “signature,” vides the material for another round of mutation composing some music in accordance with an and selection (e.g. Sims 1991). appropriate set of rules, overlaying the composer’s Clearly, not all computer-based automatic art “signature” upon the newly composed music, illuminates processes of human art-making. What and finally adding musical textures that conform is required is, first, that the computer’s computa- to the composer’s style. The technologies em- tional architecture be designed to model that ployed include rule-based expert systems, pattern of humans, at least at relatively high levels of recognition neural nets, LISP transition networks, abstraction, and, second, that the choice of algo- and a style dictionary. The results are remarkable: rithms be constrained so as to produce works expert audiences are unable to reliably distinguish that resemble those made by humans. Whereas EMI’s versions of music in the styles of Mozart automatic art looks, sounds, or reads like auto- and Rachmaninoff from the originals. matic art, art made by computers designed to Specialized applications of this technology model human art-making should pass an aesthetic enable systems to improvise music in real time version of Turing’s imitation game. with or without human musicians. These systems Harold Cohen’s AARON, a version of which incorporate real-time listening, musical analysis, can be installed as a screen saver on personal and classification with real-time music genera- computers, draws convincing figures – figures tion. Moreover, since the music generated at a that are sufficiently charming that they have been given time must be recognizably related in an exhibited in art galleries (see ). The system’s four- these systems have been developed in tandem component architecture reflects some of what with computational theories of improvisation we would have to know in order to understand (Johnson-Laird 1993). Analogous style recogni- how we make images (Burton 1997). AARON tion and art-production systems have been possesses a way of creating physical images, either designed for architecture (e.g. Stiny & Mitchell by coloring pixels on a screen or by sending data 1980) and poetry. Here is a haiku written by

114 Digital Art

Ray Kurzweil’s Cybernetic Poet in imitation appear to meet these conditions, and so suggest of the style of Wendy Dennis (): the ultimate adequacy of Boden’s account but its value as an illustration of the prospects of Page developing a theory of creativity by modeling it Sashay down the page computationally. through the lioness It is tempting to assume that the cutting-edge nestled in my soul applications of digital technologies are exclusively scientific or industrial. Artists have explored the Supposing AARON, EMI, and the Cybernetic potential of computers since their invention, Poet make art, or quasi-art, it does not follow sometimes using them in surprising ways. We that their activities are creative. This means it is might learn something about computers from possible to study what creativity is by consider- their use by artists. Yet a great deal of computer- ing the possibility of creative computers. Margaret based art is pure techno-spectacle that has not Boden approaches the topic of creativity in sci- much more to offer us than the shiny newness ence and art by asking: can computation help of its technology. Digital technology is as much us understand creativity? can computers appear a challenge as well as an opportunity. creative? can computers appear to recognize cre- ativity? can computers be creative? (Boden 1994: 85; Boden 1998). The point is not to answer References these questions primarily so as to understand the capabilities of computers but rather so as to Binkley, T. 1997. “The vitality of digital creation.” gain a deeper understanding of creativity itself. Journal of Aesthetics and Art Criticism 55(2): Boden, for example, draws a distinction be- 107–16. tween historical creativity, a property of a valu- ——. 1998a. “Computer art.” In M. Kelly, ed., able idea that nobody has ever had before, and Encyclopedia of Aesthetics, vol. 1. Oxford: Oxford psychological creativity, a property of a valuable University Press, pp. 412–14. idea that could not have arisen before in the ——. 1998b. “Digital media: an overview.” In mind of the thinker who has the idea (Boden M. Kelly, ed., Encyclopedia of Aesthetics, vol. 2. 1994: 76). Computers can clearly originate his- Oxford: Oxford University Press, pp. 47–50. torically creative ideas; it is their capacity for ori- Boden, M. 1994. “What is creativity?” In M. Boden, ginating psychologically creative ideas that is in ed., Dimensions of Creativity. Cambridge, MA: question. To resolve this question we need to MIT Press. know what it means to say an idea “could not” ——. 1998. “Computing and creativity.” In T. W. have arisen before in a thinker. A creative idea is Bynum and J. H. Moor, eds., The Digital Phoenix: How Computers are Changing Philosophy. Oxford: not merely a novel idea in the sense that a com- Blackwell. putational system is said to be able to generate Burton, E. 1997. “Representing representation: novel outputs. I have never before written the artificial intelligence and drawing.” In S. Mealing, previous sentence but the sentence is hardly ed., Computers and Art. Exeter: Intellect. creative, for my capacity to write the sentence is Carroll, N., ed. 2000. Theories of Art Today. a computational capacity to generate novel sen- Madison: University of Wisconsin Press. tences. Boden proposes that a system is creative Cope, D. 1991. Computers and Musical Style. only when it can change itself so as to expand Madison: A-R Editions. the space of novel ideas it is capable of generat- Davies, S. 2000. “Definitions of art.” In B. Gaut ing. In order to change itself in this way, it and D. M. M. Lopes, eds., Routledge Companion must represent its own lower-level processes for to Aesthetics. London: Routledge. generating ideas and it must have some way of Heim, M. 1998. “Virtual reality.” In M. Kelly, ed., tweaking these processes. Genetic algorithms, Encyclopedia of Aesthetics, vol. 4. Oxford: Oxford which enable a system to rewrite its own code, University Press, pp. 442–4.

115 Dominic McIver Lopes

Holmes, T. 2001. “Rendering the viewer conscious: Parkinson, G. H. R. 1961. “The cybernetic approach interactivity and dynamic seeing.” In R. Ascott, to aesthetics.” Philosophy 36: 49–61. ed., Art, Technology, Consciousness, mind@large. Reffen Smith, B. 1997. “Post-modem art, or: virtual Exeter: Intellect. reality as trojan donkey, or: horsetail tartan lit- Johnson-Laird, P. 1993. “Jazz improvisation: a erature groin art.” In S. Mealing, ed., Computers theory at the computational level.” In P. Howell, and Art. Exeter: Intellect. R. West, and I. Cross, eds., Representing Musical Saltz, D. Z. 1997. “The art of interaction: inter- Structure. London: Academic. activity, performativity, and computers.” Journal Lansdown, J. 1997. “Some trends in computer of Aesthetics and Art Criticism 55: 117–27. graphic art.” In S. Mealing, ed., Computers and Savedoff, B. E. 1997. “Escaping reality: digital Art. Exeter: Intellect. imagery and the resources of photography.” Levinson, J. 1990. “Hybrid art forms.” In Music, Journal of Aesthetics and Art Criticism 55: Art, and Metaphysics. Ithaca: Cornell University 201–14. Press. Sims, K. 1991. “Artificial evolution for computer Lopes, D. M. M. 2001. “The ontology of inter- graphics.” Computer Graphics 25: 319–28. active art.” Journal of Aesthetic Education 35(4): Stiny, G. and Mitchell, W. J. 1980. “The grammar of 65–82. paradise: on the generation of Mughul gardens.” Mitchell, W. J. 1992. The Reconfigured Eye: Visual Environment and Planning B 7: 209–26. Truth in the Post-Photographic Era. Cambridge, Walton, K. 1970. “Categories of art.” Philosophical MA: MIT Press. Review 79: 334–67.

116 Part III Mind and AI The Philosophy of AI and its Critique

Chapter 9 The Philosophy of AI and its Critique James H. Fetzer

(1912–54), a brilliant British mathematician, cry- Historical Background ptographer, theoretician, and philosopher. Some of Turing’s most important research concerned Prior to the advent of computing machines, theor- the limitations of proof within mathematics, izing about the nature of mentality and thought where he proposed that the boundaries of the was predominantly the province of philosophers, computable (of mathematical problems whose among whom perhaps the most influential his- solutions were obtainable on the basis of finite torically has been René Descartes (1596–1650), applications of logical rules) were the same as often called “the father of modern philosophy.” those that can be solved using a specific kind of Descartes advanced an ontic (or ontological) thesis problem-solving machinery. about the kind of thing minds are as features of Things of this kind, which are known as the world and an epistemic (or epistemological) Turing machines, consist of an arbitrarily long thesis about how things of that kind could be segmented tape and a device capable of four known. According to Descartes, who advocated operations upon that tape, namely: making a a form of dualism for which mind and body are mark, removing a mark, moving the tape one mutually exclusive categories, “minds” are things segment forward, and moving the tape one seg- that can think, where access to minds can be ment backward. (The state of the tape before a secured by means of a faculty known as “intro- series of operations is applied can be referred to spection,” which is a kind of inward perception as “input,” the state of the tape after it has been of a person’s own mental states. applied as “output,” and the series of instruc- Descartes’s approach exerted enormous influ- tions as a “program.”) From the perspective of ence well into the twentieth century, when the these machines, it became obvious there are development of digital computers began to cap- mathematical problems for which no finite or tivate the imagination of those who sought a more computable solutions exist. Similar results relat- scientific and less subjective conception of the ing effective procedures to computable problems nature of thinking things. The most important were concurrently obtained by the great Amer- innovations were introduced by Alan Turing ican logician Alonzo Church.

119 James H. Fetzer

behavioral rather than introspective criterion for The Turing Test supporting inferences to the existence of thinking things, making the study Church’s work was based on purely mathematical of the mind appear far less subjective. assumptions, while Turing’s work appealed to a very specific kind of machine, which provided an abstract model for the physical embodiment of Physical Machines the procedures that suitably define “(digital) com- puters” and laid the foundation for the theory of computing. Turing argued that such procedures Descartes’s conception of human minds as impose limits upon human thought, thereby com- thinking things depends upon actually having bining the concept of a program with that of a thoughts, which might not be the case when mind in the form of a machine which in principle they are unconscious (say, asleep, drugged, could be capable of having many types of physical or otherwise incapable of thought), since their implementation. His work thus introduced what existence as things that think would not then has come to be known as the computational con- be subject to introspective verification, which ception of the mind, which inverts the Cartesian supports hypothesis (h1): account of machines as mindless by turning minds themselves into special kinds of machines, where (h1) (Conscious) human minds are the boundaries of computability define the bound- thinking things (Descartes); aries of thought. Turing’s claim to have fathered AI rests upon Analogously, Turing’s conception of these the introduction of what is known as the Turing machines as thinking things depends upon the test, where a thing or things of one kind are exercise of the capacity to manipulate marks as a pitted against a thing or things of another kind. sufficient condition for the possession of intel- Adapting a party game where a man and a woman ligence which could be comparable to that of might compete to see whether the man could humans, suggesting hypothesis (h2): deceive a contestant into mistaking him for the woman (in a context that would not give the (h2) Turing machines manipulating marks game away), he proposed pitting a human being possess intelligence (Turing); against an inanimate machine (equipped with a suitable program and mode of communication). where the identification of intelligence with Thus, if an interlocutor could not differentiate mentality offers support for the conclusion that between them on the basis of the answers they suitably programmed and properly functioning provided to questions that they were asked, then Turing machines might qualify as manmade those systems should be regarded as equal (or thinking things or, in the phrase of John equipotent) with respect to (what he took to McCarthy, as “artificial intelligence.” be) intelligence (Turing 1950). As idealized devices that are endowed with This represented a remarkable advance over properties that physical systems may not possess, Cartesian conceptions in three different respects. including segmented tapes (or “memories”) of First, it improved upon the vague notion of a arbitrary length and perfection in performance, thinking thing by introducing the precise notion however, Turing machines are abstract entities. of a Turing machine as a device capable of mark Because they do not exist in space/time, they are manipulation under the control of a program. incapable of exerting any causal influence upon Second, it implied a solution to the mind/body things in space/time, even though, by defini- problem, according to which hardware is to soft- tion, they perform exactly as intended (Fetzer ware as bodies are to minds, that was less meta- 1988). The distinction is analogous to that phorical and more scientific than the notion of between numbers and numerals, where numbers bodies with minds. Third, it appealed to a are abstract entities that do not exist in space/

120 The Philosophy of AI and its Critique time, while numerals that stand for them are where, as for Turing, the phrase “intelligent physical things that do exist in space/time. thing” means the same as “thinking thing.” Roman numerals, Arabic numerals, and such have There is an ambiguity about the words “sym- specific locations at specific times, specific shapes bol systems” as systems that process symbols and and sizes, come into and go out of existence, as the systems of symbols which they process, none which is true of numbers as timeless and where Newell and Simon focused more attention unchanging abstract entities. on the systems of symbols that machines process These “machines,” nevertheless, might be sub- than they did upon the systems that process those ject to at least partial implementations as phys- symbols. But there can be no doubt that they ical things in different ways employing different took for granted that the systems that processed materials, such as by means of digital sequences of those symbols were physical. It therefore becomes 0s and 1s, of switches that are “on” or “off,” or important, from this point hence, to distinguish of higher and lower voltage. Some might be con- between “Turing machines” as abstract entities structed out of vacuum tubes, others made of and “digital computers” as physical implemen- transistors or silicon chips. They then become tations of such machines, where digital com- instances of physical things with the finite propert- puters, but not Turing machines, possess finite ies of things of their kinds. None of them performs memories and potential to malfunction. Newell exactly as intended merely as a matter of definition: and Simon focused upon computers as physical all of them have the potential for malfunction machines, where they sought to clarify the and variable performance like aircraft, automo- status of the “marks” that computers subject to biles, television sets, and other physical devices. manipulation. Their memories are determined by specific phys- They interpreted them as sets of physical ical properties, such as the size of their registers; patterns they called “symbols,” which can occur and, while they may be enhanced by the addi- in components of other patterns they called tion of more memory, none of them is infinite. “expressions” (or “symbol structures”). Relative to sets of alpha-numerical (alphabetical and numerical) characters (ASCII or EBCDIC, for example), expressions are sequences of symbols Symbol Systems understood as sequences of characters. Their “symbol systems” as physical machines that While (some conceptions of) God might be manipulate symbols thus qualify as necessary advanced as exemplifying a timeless and unchang- and sufficient for intelligence, as formulated by ing thinking thing, the existence of entities of hypothesis (h4): that kind falls beyond the scope of empirical and scientific . Indeed, within computer sci- (h4) (Being a) symbol system is both neces- ence, the most widely accepted and broadly influ- sary and sufficient for intelligence (Newell ential adaptation of Turing’s approach has been by and Simon); means of the physical symbol system conception Alan Newell and Herbert Simon have advanced, where which, even apart from the difference between symbol systems are physical machines – possibly Turing machines as abstractions and symbol sys- human – that process physical symbol structures tems as physical things, turns out to be a much through time (Newell & Simon 1976). These stronger claim than (h2) or even (h3). Those are special kinds of digital machines that qualify hypotheses do not imply that every thinking thing as serial processing (or von Neumann) machines. has to be a digital computer or a Turing machine. Thus, they implement Turing’s conception by (h2) and (h3) are both consistent with the exist- means of a physical machine hypothesis (h3), ence of thinking things that are not digital com- puters or Turing machines. But (h4) does not (h3) Physical computers manipulating allow for the existence of thinking things that symbols are intelligent (Newell and Simon); are not digital machines.

121 James H. Fetzer

inputs, and “replications,” where systems replic- The Chinese Room ate one another when they yield the same out- puts from the same inputs by means of the same The progression of hypotheses from (h1) to (h2) processes or procedures. In this language, Searle to (h3) and perhaps (h4) appears to provide sig- shows that, even if the Turing test is sufficient nificant improvement on Descartes’s conception, for comparisons of input/output behavior especially when combined with the Turing test, (simulations), it is not sufficient for comparisons since they not only clarify the nature of mind and of the processes or procedures that yield those elucidate the relation of mind to body, but even outputs (replications). explain how the existence of other minds might be known, a powerful combination of ontic and epistemic theses that seems to support the pros- Weak AI pects for artificial intelligence. As soon as com- puting machines were designed with performance capabilities comparable to those of human beings, The force of Searle’s critique becomes apparent it would be appropriate to ascribe to those in- in asking which scenario, C or D, is more like animate entities the mental properties of think- the performance of a computer executing a pro- ing things. Or so it seemed, when the philosopher gram, which might be implemented as an auto- John Searle advanced a critique of the prospects mated look-up table: in response to inputs in for AI that has come to be known as “the Chinese the form of sequences of marks, a computer pro- Room” and cast it all in doubt (Searle 1980). cesses them into outputs in the form of other Searle proposed a thought experiment involv- sequences of marks on the basis of its program. ing two persons, call them “C” and “D,” one (C) So it appears appropriate to extend the compar- fluent in Chinese, the other (D) not. Suppose C ison to yet a third scenario, call it “E,” where a were locked in an enclosed room into which suitably programmed computer takes the same sequences of marks were sent on pieces of paper, inputs and yields the same outputs. For just as to which C might respond by sending out other the performance of D might simulate the per- sequences of marks on other pieces of paper. If formance of C, even though D knows no Chi- the marks sent in were questions in Chinese and nese, so the performance of E might simulate the marks sent out were answers in Chinese, the performance of D, even though E possesses then it would certainly look as though the occup- no mentality. Mere relations of simulation thus ant of the room knew Chinese, as, indeed, by appear too weak to establish that systems are hypothesis, he does. But suppose instead D were equal relative to their intelligence. locked in the same room with a table that allowed Searle also differentiated between what he him to look up sequences of marks to send out called “strong AI” and “weak AI,” where weak in response to sequences of marks sent in. If he AI maintains that computers are useful tools in were very proficient at this activity, his perform- the study of the mind, especially in producing ance might be the equal of that of C, who knows useful models (or simulations), but strong AI Chinese, even though D, by hypothesis, knows maintains that, when they are executing pro- no Chinese. grams, computers properly qualify as minds (or Searle’s argument was a devastating counter- replications). Weak AI thus represents an example to the Turing test, which takes for epistemic stance about the value of computer- granted that similarities in performance indicate based models or simulations, while strong AI similarities in intelligence. In the Chinese Room represents an ontic stance about the kinds of scenario, the same “inputs” yield the same things that actually are instances of minds. Pre- “outputs,” yet the processes or procedures that sumably, strong AI implies weak AI, since actual produce them are not the same. This suggests instances of minds would be suitable subjects that a distinction has to be drawn between “sim- in the study of mind. Practically no one objects ulations,” where systems simulate one another to weak AI, however, while strong AI remains when they yield the same outputs from the same controversial on many grounds.

122 The Philosophy of AI and its Critique

That does not mean it lacks for passionate ad- representations of aspects of itself. But, while vocates. One of the most interesting introductions self-awareness and self-consciousness are often to artificial intelligence has been co-authored by taken to be important kinds of intelligence or Eugene Charniak and Drew McDermott (1985). mentality, they do not appear to be essential to Already in their first chapter, the authors define having intelligence or mentality in general as “artificial intelligence” as the study of mental opposed to having intelligence or mentality of faculties through the use of computational specific kinds. There may be various kinds of models. The tenability of this position, no doubt, mentality or intelligence – mathematical, verbal, depends upon the implied premise that mental and artistic, for example – but presumably they faculties operate on the basis of computational share certain core or common properties. processes, which, indeed, they render explicit by There would seem to be scant room for doubt similarly postulating that what brains do “may that, if artificial machines are going to qualify as be thought of at some level as a kind of com- comparable to human beings relative to their putation” (Charniak & McDermott 1985: 6). mental abilities, they must have the same or similar The crucial distinction between “weak” and capacities to use and manipulate internal repres- “strong” AI, however, depends upon whether entations, at least with respect to some specified brains actually qualify as computers, not whether range – presumably, alpha-numeric – of tasks. they may be thought to be. They must take the same or similar external inputs (or “stimuli”), process them by means of the same or similar “mental” mechanisms, and pro- duce the same or similar external outputs (or Strong AI “responses”). While Charniak and McDermott may aspire to build an artificial animal, the AI Charniak and McDermott also maintain “the community at large, no doubt, would settle for ultimate goal of research in AI is to build a building an artificial thinking thing, presuming person or, more humbly, an animal.” Their gen- that it is possible to create one without the other. eral conception is that the construction of these artificial things must capture key properties of their biological counterparts, at least with respect Folk Psychology to kinds of input, kinds of processing, and kinds of output. Thus, the “inputs” they consider in- clude vision (sights) and speech (sounds), which There is an implied presumption that different are processed by means of internal modules for systems that are subject to comparison are oper- learning, deduction, explanation, and planning, ating under the same or similar causally relevant which entail search and sort mechanisms. These background conditions. No one would suppose combine with speech and motor capabilities to that a computer with a blown motherboard yield “outputs” in the form of speech (sounds) should yield the same outputs from the same and behavior (motions), sometimes called “ro- inputs as a comparable computer with no hard- botics.” The crucial issue thus becomes whether ware breakdown, even when they are loaded with these “robots” are behaving like human beings the same programs. Analogously, no one would as (mindless) simulations or instead embody assume that a human being with a broken arm, (mindful) replications. for example, should display the same behavior in Their attention focuses upon what goes on in response to the same stimuli (say, a ball coming “the black box” between stimulus and response, straight toward him while seated in the bleachers where those with minds depend upon and utilize at a game) as another person without a broken internal representations as states of such systems arm. But that does not mean that they are not that describe or otherwise represent various processing similar stimuli by means of similar aspects of the world. Indeed, some of these as- representations. pects could be internal to the system itself and Human beings are complicated mechanisms, thus represent its own internal states as internal whether or not they properly qualify as

123 James H. Fetzer

“machines” in the sense that matters to AI tends that the categories of motives and beliefs, Indeed, the full range of causally relevant factors among others, are destined for a similar fate that make a difference to human behavior appears as neuroscience develops. Churchland admits to include motives, beliefs, ethics, abilities, cap- he cannot guarantee that this will occur, where abilities, and opportunities (Fetzer 1996). Dif- the in this instance might ferent persons with the same or similar motives instead simply reflect some adjustment in folk- and beliefs, for example, but who differ in their psychological principles or dispensing with some morals, may be expected to display different of its concepts. behavior under conditions where ethics makes a The deeper problem that confronts eliminative difference, even though they may have similar materialism, however, appears to be the same abilities and are not incapacitated from the exer- problem confronting classic forms of reduction- cise of those abilities. As we all know, human ism, namely, that without access to information beings consume endless hours endeavoring to relating brain states to mind states, on the one explain and predict the behavior of others and hand, and mind states to behavioral effects, on themselves, employing a framework of causally the other, it would be impossible to derive pre- relevant factors of this kind, which has come to dictive inferences from brain states to behavioral be known as “folk psychology.” effects. If those behavioral effects are manifesta- No doubt when appraised from the perspective tions of dispositions toward behavior under spe- of, say, the conditions of adequacy for scientific cific conditions, moreover, then it seems unlikely theories – such as clarity and precision of lan- that a “mature” neuroscience could accomplish guage, scope of application for explanation and its goals if it lacked the capacity to relate brain prediction, degree of empirical support, and the states to behavioral effects by way of dispositions, economy, simplicity, or elegance with which these because there would then be no foundation for results are attained – folk psychology appears relating mind states to brain states and brain to enjoy a high degree of empirical support by states to human behavior. virtue of its capacity to subsume a broad range In the case of jealousy (hostility, insincerity, and of cases within the scope of its principles. Some so on) as causal factors that affect our behavior of that apparent success, however, may be due in the folk-psychological scheme of things, if we to the somewhat vague and imprecise character want to discover the brain states that underlie of the language upon which it depends, where these mind states as dispositions to act jealous there would appear to be opportunity for revi- (to act hostile, and so forth) under specific condi- sion and refinement to enhance or confine its tions, which include our other internal states, then scope of application. Yet some researchers argue a rigorous science of human behavior might be for its elimination altogether. developed by searching for and discovering some underlying brain states, where those dispositions toward behavior were appropriately (presumably, lawfully) related to those brain states. Sometimes brain states can have effects upon human behavior that are not mediated by mind states, as in the Paul Churchland, for example, maintains that case of brain damage or mental retardation. For folk psychology is not only incomplete but also neurologically normal subjects, mind states are inaccurate as a “misrepresentation” of our internal able to establish connections between brain states states and mental activities. He goes so far as to and their influence on behavior. suggest that progress in neuroscience should lead, not simply to the refinement of folk psychology, but to its wholesale elimination (Churchland Processing Syntax 1984: 43). The model Churchland embraces thus follows the pattern of elimination of “phlogiston” from the language of chemistry and of “witches” The predominant approach among philosophers from the language of psychology. He thus con- eager to exploit the resources provided by the

124 The Philosophy of AI and its Critique computational conception, however, has been in (h5) Physical computers processing syntax the direction of refining what it takes to have a possess minds (STM); mind rather than the relationship between minds, bodies, and behavior. While acknowledging these which may initially appear much stronger than connections are essential to the adequacy of any (h3). But Newell and Simon’s notion of account, they have focused primarily upon the “symbol” is defined formally and their “symbol prospect that language and mentality might be systems” are also computing machines. Both adequately characterized on the basis of purely approaches run the risk of identifying “thinking formal distinctions of the general kind required things” with mindless machines. by Turing machines – the physical shapes, sizes, and relative locations of the marks they manip- ulate – when interpreted as the alpha-numeric Semantic Engines characters that make up words, sentences, and other combinations of sentences as elements of a language. Systems of marks with rules for their manipula- , for example, has observed that tion are examples of (what are known as) formal computational conceptions of language and men- systems, the study of which falls within the tality entail the thesis that “mental processes have domain of pure mathematics. When those formal access only to formal (nonsemantic) properties systems are subject to interpretations, especially of the mental representations over which they with respect to properties and objects within are defined” (Fodor 1980: 307). He elaborates the physical world, their study falls within the upon the relationship between the form (syntax) domain of applied mathematics. A debate has and the content (semantics) of thoughts, main- raged within computer science over whether taining (a) that thoughts are distinct in content that discipline should model itself after pure only if they can be identified with distinct repres- or applied mathematics (Colburn et al. 1993). entations, but without offering an explanation But whatever the merits of the sides to that dis- of how it is (b) that any specific thoughts can be pute, there can be scant room for doubt that identified with any specific representations, a mere mark manipulation, even in the guise of problem for which he elsewhere offers a solution syntax processing, is not enough for thinking known as “the language of thought.” But any things. Thoughts possess content as well as form, account maintaining that the same syntax always where it is no stretch of the imagination to sug- has the same semantics or that the same semantics gest that, regarding thought, content dominates always has the same syntax runs afoul of prob- form. lems with ambiguity on the one hand, and with The STM, which makes syntax processing synonymy on the other. sufficient for the possession of mentality, thus Nevertheless, the strongest versions of com- appears to be far too strong, but a weaker ver- putational conceptions tend to eschew concern sion might still be true. The ability to process for semantics and focus instead on the centrality syntax might be necessary for mentality instead, of syntax. Stephen Stich has introduced the syn- as indeed hypothesis (h3) implies, when Newell tactic theory of the mind (STM) as having an and Simon’s “symbols” are properly understood agnostic position on content, neither insisting as marks subject to manipulation. Thus, a more that syntactic state types (as repeatable patterns plausible version of (h5) should maintain instead of syntax) have no content nor insisting that (h6): syntactic state tokens (specific instances of syn- tactic state types) have no content: “It is simply (h6) (Conscious) minds are physical com- silent on the whole matter . . . [T]he STM is in puters processing syntax; effect claiming that psychological theories have no need to postulate content or other semantic where syntax consists of marks and rules for their properties” (Stich 1983: 186). STM is thereby manipulation that satisfy constraints that make committed to hypothesis (h5): them meaningful. But since there are infinitely

125 James H. Fetzer many possible interpretations of any finite unpalatable choice between an endless hierarchy of sequence of marks, some specific interpretation successively richer and richer metalanguages for (or class of interpretations) requires specifica- specifying the meaning of lower-level languages tion as “the intended interpretation.” Marks can and a base language that is unlearned, Fodor only qualify as syntax relative to specific inter- opts for the existence of an innate and inborn pretations in relation to which those marks language of thought. become meaningful. The process of relating a learned language to From this point of view, a (properly function- the language of thought turns human beings ing) computing machine can be qualified as an into semantic engines, which may be rendered automatic formal system when it is executing a by hypothesis (h8) as follows: program, but becomes meaningful only when its syntax satisfies the constraints of an intended inter- (h8) Human beings are semantic engines pretation. Indeed, an automatic formal system with a language of thought (Fodor). where “the semantics follows the syntax” has been designated “a semantic engine” by Daniel Fodor commits a mistake in his argument, how- Dennett. This supports the contention some have ever, by overlooking the possibility that the kind called the basic idea of cognitive science: that of prior understanding which is presupposed by intelligent beings are semantic engines, that is, language learning might be nonlinguistic. Chil- automatic formal systems under which they con- dren learn to suck nipples, play with balls, and sistently make sense (Haugeland 1981: 31). (h6) draw with crayons long before they know that thus requires qualification to incorporate the role what they are doing involves “nipples,” “balls,” of interpretation as (h7): or “crayons.” Through a process of interaction with things of those kinds, they acquire habits (h7) Semantic engines are necessary and of action and habits of mind concerning the sufficient for intelligence; actual and potential behavior of things of those kinds. Habits of action and habits of mind that where, as in the case of Newell and Simon, obtain for various kinds of things are concepts. “intelligent things” are also “thinking things” Once that nonlinguistic understanding has been and “(conscious) minds,” understood as physical acquired, the acquisition of linguistic dispositions computers processing syntax under an inter- to describe them appears to be relatively unprob- pretation. The problem is to “pair up” the syntax lematical (Fetzer 1990). and the semantics the right way. One of the remarkable features of Fodor’s conception is that the innate and inborn lan- guage of thought possesses a semantic richness such that this base language has to be sufficiently The Language of Thought complete to sustain correlations between any natural language (French, German, Swahili, and Jerry Fodor (1975) has advanced an argument such) at any stage of historical development (past, hypothesizing the existence of an innate language, present, and future). This means that mentalese which is species-specific and possessed by every not only has to supply a foundation for everyday neurologically normal human being. He calls it words, such as “nipple,” “ball,” and “crayon” in mentalese (or “the language of thought”). He English, for example, but also those for more contends the only way to learn a language is to advanced notions, such as “jet propulsion,” learn the truth conditions for sentences that “polio vaccine,” and “color television,” since occur in that language: “learning (a language) L otherwise the language of thought could not involves learning that ‘Px’ is true if and only if fulfill its intended role. Among the less plausible x is G for all substitution instances. But notice consequences of this conception turn out to be that learning that could be learning P (learning that, since every human has the same innate lan- what P means) only for an organism that already guage, which has to be complete in each of its understood G” (Fodor 1975: 80). Given the instantiations, unsuccessful translations between

126 The Philosophy of AI and its Critique different languages and the evolution of language than those of automated formal systems as an- across time are both impossible, in principle, other case of abstract idealization. By comparison, which are difficult positions to defend. digital machines and human beings possess no more than (limited) computational competence (Fetzer 1992). The properties of formal systems – such as incompleteness for systems richer than Formal Systems first-order monadic logic, which Kurt Gödel established – that might be supposed to impose Fodor’s approach represents an extension of limits on mental processes and have attracted the work of Noam Chomsky, who has long interest by such scholars as J. R. Lucas (1961) championed the conception of an innate syntax, and Douglas Hofstadter (1979), appear to have both inborn and species-specific, to which Fodor slight relevance to understanding the nature of has added a semantics. Much of Chomsky’s cognition. Formal systems are useful in model- work has been predicated upon a distinction ing reasoning, but reasoning is a special case between competence and performance, where of thinking. And if we want to understand the differences between the grammatical behavior of nature of thinking, we have to study thinking different language users, which would otherwise things rather than the properties of formal sys- be the same, must be accounted for by circum- tems. Thinking things and formal systems are stantial differences, say, in physiological states not the same. or psychological context. In principle, every user of language possesses what might be described as (unlimited) computational competence, where Mental Propensities infinitely many sentences can be constructed from a finite base by employing recursive procedures of the kind that were studied by Church Roger Penrose has suggested that thinking may and Turing in their classic work on effective be a quantum phenomenon and thereby qualify procedures. as non-algorithmic (Penrose 1989: 437–9). The Fodor and Zenon Pylyshyn (1988) adopt con- importance of this prospect is that algorithms ditions for the production of sentences by lan- are commonly understood as functions that map guage users implying that the semantic content single values within some domain onto single of syntactic wholes is a function of the semantic values within some range. If mental processes content of their syntactic parts as their principle are algorithmic (functions), then they must be of the compositionality of meaning and that mole- deterministic, in the sense that the same mental- cular representations are functions of other state cause (completely specified) invariably brings molecular or atomic representations as a principle about the same mental-state effect or behavioral of recursive generability. These conditions are response. Since quantum phenomena are not obvious counterparts of distinctions between deterministic, if mental phenomena are quantum structurally atomic and structurally molecular processes, they are not functions – not even par- representations as a precondition for a language tial functions, for which, when single values of thought that is modeled on formal systems, within a domain happen to be specified, there such as sentential calculus. The principles of those exist single values in the corresponding range, formal systems, automated or not, may or may but where some of the values in the domain not transfer from abstract to physical contexts, and range of the relevant variables might not be not least because physical systems, including specified. digital machines, are limited in their capacities. Systems for which the presence or the absence Turing machines with infinite tapes and infal- of every property that makes a difference to an lible performance are clearly abstract idealizations outcome is completely specified are said to be compared to digital machines with finite memor- “closed,” while those for which the presence or ies that can malfunction. The physical properties absence of some properties that make a differ- of persons and computers are decidedly different ence to the outcome are unspecified are said to

127 James H. Fetzer be “open.” The distinction between determin- which they can effect transitions between pat- istic and (in this case) probabilistic causation is terns of activation, and the plasticity and strength that, for closed systems, for deterministic causal of their behavior responses. processes, the same cause (or complete set of Concerns about variation in such types of causa- conditions) invariably (or with universal strength tion also arise within the context of the study of u) brings about the same effect, whereas for mental models or representations of the world, probabilistic causal processes, the same cause vari- specifically, what has be known as the frame prob- ably (with probabilistic strength p) brings about lem, which Charniak and McDermott describe one or another effect within the same fixed class as the need to infer explicitly that one or more of possible outcomes. A polonium218 atom, for states will not change across time, which forms a example, has a probability for decay during a “frame” within which other states may change 3.05 minute interval of 1/2. (1985: 418) While the frame problem has proven The determination that a system, such as an amenable to many different characterizations – a atom of polonium218, is or is not a closed system, variety of which may be found, for example, in of course, poses difficult epistemic problems, Ford and Hayes 1991 – one important aspect of which are compounded in the case of human the problem is the extent to which a knowledge beings, precisely because they are vastly more base permits the prediction and the explanation complex causal systems. Moreover, probabilistic of systems when those systems are not known to systems have to be distinguished from (what are be open or closed. called) chaotic systems, which are deterministic Indeed, from this point of view, the frame systems with “acute sensitivity to initial condi- problem even appears to instantiate the classic tions,” where the slightest change to those con- encountered in attempting ditions can bring about previously unexpected to predict the future based upon information effects. A tiny difference in hundreds of thou- about the past, which was identified by David sands of lines of code controlling a space probe, Hume (1711–76), a celebrated Scottish philo- for example, consisting of the occurrence of only sopher. Thus, Hume observed that there are one wrong character, a single misplaced comma, no deductive guarantees that the future will re- caused Mariner 1, the first United States inter- semble the past, since it remains logically possible planetary spacecraft, to veer off course and then that, no matter how uniformly the occurrence have to be destroyed. of events of one kind have been associated with events of another, they may not continue to be. If the laws of nature persist through time, how- ever, then, in the case of systems that are closed, The Frame Problem it should be possible to predict – invariably or probabilistically – precisely how those systems Indeed, there appear to be at least three contexts will behave over intervals of time, so long as the in which probabilistic causation may matter to complete initial conditions and laws of systems human behavior, namely: in processing sensory of that kind are known. data into patterns of neural activation; in trans- itions between one pattern of activation and another; and in producing sounds and other Minds and Brains movement as a behavioral response. Processes of all three kinds might be governed by probabilistic or by chaotic deterministic processes and there- Because connectionism appeals to patterns of fore be more difficult to explain or predict, even activation of neural nodes rather than to indi- when the kind of system under consideration vidual nodes as features of brains that function happens to be known. The most important dif- as representations and affect behavior, it appears ferences between species appear to concern the to improve upon computationally-based concep- range and variety of sensory data they are cap- tions in several important respects, including per- able of processing, the speed and reliabilty with ceptual completions of familiar patterns by filling

128 The Philosophy of AI and its Critique in missing portions, the recognition of novel Semiotic Systems patterns even in relation to previously unfamiliar instances, the phenomenon known as “graceful degradation,” and related manifestations of men- The conception of minds as semiotic (or as tality (Rumelhart et al. 1986: 18–31). Among sign-using) systems advances an alternative to the most important differences is that connec- computational accounts that appears to fit the tionist “brains” are capable of what is known connectionist model of the brain like a hand in a as parallel processing, which means that, unlike glove. It provides a noncomputational framework (sequential) Turing machines, they are capable for investigating the nature of mind, the relation of (concurrently) processing more than one of mind to body, and the existence of other stream of data at the same time. minds. According to this approach, minds are This difference, of course, extends to physical things for which something can stand for some- computers, which can be arranged to process thing else in some respect or other (Fetzer 1990, data simultaneously, but each of them itself 1996). The semiotic relation, which was elabor- remains a sequential processor. The advantages ated by the American philosopher Charles S. of parallel processing are considerable, especially Peirce (1839–1914), is triadic (or three-placed), from the point of view of evolution, where involving a relation of causation between signs detecting the smells and the sounds of predators and their users, a (crucial) relation of grounding before encountering the sight of those predators, between signs and that for which they stand, for example, would afford adaptive advantages. and an interpretant relation between signs, what Moreover, learning generally can be understood they stand for, and the users of signs. as a process of increasing or decreasing activa- There are three branches of the theory of semi- tion thresholds for specific patterns of nodes, otics, which include syntax as the study of the where classical and operant conditioning may relations between signs and how they can be be accommodated as processes that establish combined to create new signs, semantics as the association between patterns of activation and study of the relations between signs and that for make their occurrence, under similar stimulus which they stand, and pragmatics as the study of conditions, more (or less) probable, where the the relations between signs, what they stand for, activation of some patterns tends to bring about and sign users. Different kinds of minds can then speech and other behavior. be classified on the basis of the kinds of signs they Those who still want to defend computational are able to utilize – such as icons, which resemble conceptions might hold that, even if their internal that for which they stand (similar in shapes, sizes, representations are distributed, human beings are and such); indices, which are causes or effects of semantic engines (h9): that for which they stand (ashes, fires, and smoke); and symbols, which are merely habitually (h9) Human beings are semantic engines associated with that for which they stand (words, with distributed representations; sentences, and things) – as iconic, indexical, and symbolic varieties of mentality, respectively. but the rationale for doing so becomes less and Meanings are identified with the totality of less plausible and the mechanism – more and possible and actual behavior that a sign user more “independent but coordinated” serial pro- might display in the presence of a sign as a func- cessors, for example – appears more and more ad tion of context, which is the combination of hoc. For reasons that arose in relation to eliminat- motives, beliefs, ethics, abilities, and capabilities ive materialism, however, no matter how suc- that sign-users bring to their encounters with cessful connectionism as a theory of the brain, signs. And patterns of neural activation can func- it cannot account for the relationship between tion as internal signs, where (all and only) think- bodies and minds without a defensible conception ing things are semiotic systems, (h10): of the mind that should explain why symbol systems and semantic engines are not thinking (h10) Thinking things, including human things. beings, are semiotic systems.

129 James H. Fetzer

This approach can explain what it is to be con- This difference applies even when these systems scious relative to a class of signs, where a system are processing marks by means of the same pro- is conscious with respect to signs of that kind cedures. A computer processing a tax return can when it has the ability to utilize signs of that yield the same outputs from the same inputs, yet kind and is not inhibited from the exercise of they mean nothing to that system as income, that ability. And it supports the conception of deductions, or taxes due. A distinction must be cognition as an effect that is brought about drawn between those marks that are meaningful (possibly probabilistically) by interaction between for use by a system and marks that are meaningful signs and sign-users when they are in suitable for the users of that system. They can function as causal proximity. signs for those users without having to function as signs for those systems. “Symbols” in this sense of semiotic systems must therefore be clearly distinguished from Critical Differences “symbols” in the sense of symbol systems, which can be meaningless marks, lest one mistake Among the most important differences between symbol systems in Newell and Simon’s sense for semiotic systems and computational accounts (symbol-using) semiotic systems, as has John becomes apparent at this point, because the se- McCarthy (McCarthy 1996: ch. 12). This re- mantic dimension of mentality has been encom- flects (what might be called) the static difference passed by the definition of systems of this kind. between computer systems and thinking things. Observe, for example, the difference between Another is that digital machines are under the symbol systems and semiotic systems in figures control of programs as causal implementations 9.1 and 9.2, where semiotic systems reflect a of algorithms, where “algorithms” in turn are grounding relationship that symbol systems lack. effective decision procedures. Effective decision procedures are completely reliable in producing solutions to problems within appropriate classes (“Input”) of cases that are invariably correct and they do s in a finite number of steps. If these machines are under the control of algorithms but minds are causal not, then there is a dynamic difference that may relation be more subtle but is not less important as well. Indeed, there are many kinds of thinking – from dreams and daydreams to memory and z x (“Program”) (“Output”) perception as well as ordinary thought – that do not satisfy the constraints imposed by effect- computational relation ive decision procedures. They are not reliable Figure 9.1: Symbol systems problem-solving processes and need not yield definitive solutions to problems in a finite number of steps. The causal links that affect transitions (“Sign”) between thoughts appear to be more dependent s upon our life histories and associated emotions (our pragmatic contexts) than they do on syntax causal grounding and semantics per se. Even the same sign, such relation relation as a red light at an intersection, can be taken as an icon (because it resembles other red lights), as an index (as a traffic control device that is z x (“User”) (“Something”) malfunctioning), or as a symbol (where drivers should apply the breaks and come to a complete interpretant relation halt) as a function of a sign user’s context at the Figure 9.2: Semiotic systems time. Anyone else in the same context would

130 The Philosophy of AI and its Critique

(probabilistically) have interpreted that sign the kinds of things, where thinking things are unlike same way. digital machines. Ultimately, of course, the adequacy of a hinges upon the adequacy of the theory of meaning it provides that relates brains, minds, The Hermeneutic Critique and behavior. The crucial consideration appears to be that, whether bodies and minds are deter- Whether or not the semiotic conception ulti- ministic, chaotic, or probabilistic systems, it must mately prevails, current research makes it increas- provide a completely causal account of how the ingly apparent that an adequate account of signs that minds employ make a difference to mentality will have to satisfy many of the con- the behavior of those systems that is sufficient to cerns raised by the hermeneutic critique advanced sustain an inference to the existence of mentality by Hubert Dreyfus (1979). Dreyfus not only as the best explanation for the data. One way in objected to the atomicist conception of rep- which that may occur emerges from the differ- resentation that became the foundation for the ent ways in which sensations affect behavior, compositionality of meaning and recursive gener- where the dog barked at the bush when the wind ability theses that Fodor and Pylyshyn embraced blew, because he mistook it for a stranger; where but also emphasized the importance of the role Mary rushed to the door at the sound of the of bodies as vehicles of meaning, especially knock, because she thought her friend had come; through interactions with the world, very much or where Bob slowed down when the light turned in the spirit of Peirce, which whom he shares red, because he knew that he should apply the much in common. Thus, the very idea of creat- breaks and bring the car to a complete halt. ing artificial thinking things whose minds are not inextricably intertwined with their bodies and capable of interacting with the world across time Conventions and Communication becomes increasingly implausible. It has become clear that differences between Turing machines, digital computers, and human Because different users can use different signs beings go considerably beyond those addressed with the same meaning and the same signs with above, where the semiotic conception of con- different meaning, it is even possible for a sign sciousness and cognition, for example, offers the user to use signs in ways that, in their totality, capacity to make a mistake as a general criterion are not the same as those of any other user. This of mentality, where making a mistake involves implies that social conceptions of language, taking something to stand for something else, according to which private languages are impos- but doing so wrongly, which is the right result. sible, are not well-founded from the perspective From this point of view, there appear to be three of semiotic systems. A person who found himself most important differences (see table 9.1). Even abandoned on a deserted island, for example, apart from a specific theory of representation might while away the time by constructing an intended to account for the meaning of the marks elaborate system of classification for its flora machines can manipulate, it appears evident from and fauna. Even though that system of signs the table that these are three distinctly different might therefore have unique significance for that

Table 9.1: Three distinctly different kinds of things

(Abstract) (Physical) (Actual) Turing machines Digital computers Human beings

Infinite capacities: Yes No No Subject to malfunction: No Yes Yes Capable of mistakes: No No Yes

131 James H. Fetzer

(“Sign”) Other Minds s

s stands for x for z1 s stands for y for z2 The adoption of an approach of this general kind clarifies and illuminates distinctively mental aspects of various sorts of causal processes. When causal relations occur (when causes such as inputs z1 – x y – z2 bring about effects such as outputs) and those Does x = y? inputs and outputs do not serve as signs for a Figure 9.3: Communication situations system, they may then be classified as stimuli. When effects are brought about by virtue of their grounding (because they stand for those things individual user, that system of signs, presumably, in those respects) for the systems that use them, would still be learnable in the sense that there is they may properly be classified as signs. And when no reason why it could not be taught to others. semiotic relations occur (when signs being used It would simply be the case it never had. by one user are interpreted by another) between In communication situations, whether spoken, systems that use them, they may be further clas- written, or otherwise, different sign users tend sified as signals. Sometimes the signals we send are to succeed when they use signs the same way or intentional (successful, and so on), sometimes to the extent to which they mean the same things not. Every sign must be a stimulus and every by them. The question that arises is whether the signal must also be a sign, but not vice versa. same sign s stands for the same thing x for differ- Every human being, (other) animal, and in- ent sign users z1 and z2 under specific conditions animate machine capable of using signs thereby (see figure 9.3). When z1 and z2 speak different qualifies as a thinking thing on the semiotic con- languages, such as English and German, the suc- ception. This realization thus explains why dreams cess of a translation can be difficult to ascertain. and daydreams, memory and perception, and But it can also be difficult when very similar ordinary thought are mental activities, while tooth sounds are associated with meanings that may decay, prostate cancer, and free-fall, by com- not mean the same thing for every user. parison, are not. Whether or not the semiotic There are circumstances under which we may conception emerges as the most adequate prefer for our signs to be confidential. Turing among the alternative conceptions, it has become himself, for example, spent time successfully apparent that an adequate account ought to be cracking the Enigma cipher during the Second one that is at least very much like it, especially World War, enabling the British to understand in accommodating crucial differences between the German’s coded messages. Other circum- Turing machines, digital computers, and human stances, however, encourage the use of the same beings. It has become equally apparent, I surmise, signs in the same ways, such as in the case of a that minds are not machines. If thinking were community of members with common object- governed by mental algorithms, as such accounts ives and goals. Systems of public schools, for imply, then minds simply follow instructions example, are commonly financed with the pur- mechanically, like robots, and have no need for pose, among others, of instilling the members of insight, ingenuity, or invention. Perhaps we deny the community with a common understanding that we are nothing but robots because our of the language they use, which promotes com- mental activities involve so much more. Indeed, munication and cooperation between them. Some some of the most distinctive aspects of thought nations, such as the United States, have benefited tend to separate minds from machines. immeasurably from their standing as “melting Simulations are clearly too weak and emula- pots” where people from many countries come tions, which yield the same inputs from the same together and are united by reliance upon English, outputs by means of the same processes and are in the absence of which this country would no made of the same matter, are clearly too strong. doubt tend toward Balkanization. But the shoals are treacherous. ,

132 The Philosophy of AI and its Critique for example, has argued that, for some systems, mutation, natural selection, genetic drift, sexual simulations are replications, on the presumption reproduction, sexual selection, group selection, that the same psychophysical laws will be operat- artificial selection and genetic engineering, of ive. Thus, if the transition from an initial state S1 course, biological evolution, including of our own at time t1 yields a final state Sn at tn, where the species, continues to this day, bringing about intermediate steps involved in the transition be- more complex forms of semiotic systems with tween them, say, S2 at t2 through Sn−1 at tn−1, abilities to use more signs of similar kinds and are the same, then properties that are lawfully other signs of various different kinds. related to them, such as consciousness, must As manmade connectionist systems of (dry) come along with them, even when they are made neural networks are developed, it should not be of different stuff (Chalmers 1996). But that will too surprising if they reach a point where they be true only if the difference in matter does not can be appropriately classified as artificial think- affect the operation of those laws themselves. ing things. Whether that point will ever come In cases where it does, replications may require depends upon advances in science and techno- emulations. logy over which philosophers have no control. While the conception of symbol systems and even semantic engines appear to fall short of captur- ing the character of thinking things, this does Intelligent Machines not mean that they fail to capture the character of intelligent machines. To the extent to which An approach of this kind can explain why symbol machines properly qualify as “intelligent” when systems and semantic engines are not thinking they have the ability to process complex tasks things. Their properties account for the form of in a reliable fashion, the advent of intelligent thoughts but not their content, or else cannot machines came long ago. The seductive concep- account for the transitions between thoughts tual temptation has been to confuse intelligent themselves. Turing machines, with which we machines with thinking things. began, are not even physical things and cannot sustain the existence of finite minds that can See also Chapter 1, COMPUTATION; Chapter 2, malfunction and can make mistakes. The con- COMPLEXITY; and especially Chapter 10, nectionist conception of brains as (wet) neural COMPUTATIONALISM, CONNECTIONISM, AND THE networks supplies a crucial foundation for re- PHILOSOPHY OF MIND. thinking the nature of the mind, but requires supplementation by an account of the nature of the mind that is noncomputational. An appropri- References ate conception of mental causation – including the processes of perception, of thought trans- Chalmers, D. 1996. The Conscious Mind: In Search ition, and of response behavior – should permit of a Fundamental Theory. New York, NY: Oxford those kinds of processes to be computational University Press. [One of the most sophisticated but not require it. Computing is merely one discussions of the nature of mind by a leading special kind of thinking. representative of the computational conception. Not the least of the benefits that are thereby Intermediate.] derived is an account of mentality that can be Charniak, E. and McDermott, D. 1985. Introduc- reconciled with biology and evolution. Primitive tion to Artificial Intelligence. Reading, MA: organisms must have had extremely elementary Addison-Wesley. [An encompassing and sophisti- semiotic abilities, such as sensitivity to light by cated introduction to this discipline. Introduct- means of single cells with flagella to bring about ory to advanced.] motion. If moving toward the light promotes Churchland, P. 1984. Matter and Consciousness: survival and reproduction, then that behavior A Contemporary Introduction to the Philosophy would have adaptive benefits for such simple of Mind. Cambridge, MA: MIT Press. [A lucid systems. Under the combined influence of genetic discussion of the philosophy of mind and AI with

133 James H. Fetzer

emphasis on eliminativism and the relationship Mind Design: Philosophy, Psychology, Artificial of minds and bodies. Introductory.] Intelligence. Cambridge, MA: MIT Press, pp. 1– Colburn, T. et al., eds. 1993. Program Verification: 34. [A brilliant introduction to a popular and Fundamental Issues in Computer Science. influential volume. Introductory.] Dordrecht, the Netherlands: Kluwer Academic. Hofstadter, D. 1979. Gödel, Escher, Bach: An [A collection of the most important papers. Level Eternal Golden Braid. New York: Basic Books. varies.] [A Pulitzer Prize winning study of art, music, Dreyfus, H. 1979. What Computers Still Can’t Do: and mathematics which explores the potential for A Critique of Artificial Reason, rev. ed. New York: AI. Introductory to advanced.] Harper & Row. [A critique of the foundations Lucas, J. R. 1961. “Minds, machines, and Gödel.” of AI from multiple philosophical perspectives. Philosophy 36: 112–27. [Deploys Gödel against Intermediate.] the conception of minds as machines. Advanced.] Fetzer, J. H. 1988. “Program verification: the very McCarthy, J. 1996. Defending AI Research. Stanford: idea.” Communications of the ACM 31: 1048– CSLI Lecture Notes. [A collection of essays, prin- 63. [A study of the applicability of formal meth- cipally book reviews, in which one of the original ods within computer science that ignited a contributors to AI explains why he disagrees with controversy in the field. Advanced.] other views. Intermediate.] ——. 1990. Artificial Intelligence: Its Scope and Newell, A. and Simon, H. 1976. “Computer science Limits. Dordrecht, the Netherlands: Kluwer as empirical inquiry: symbols and search.” In Academic. [A sustained study of the theoretical J. Haugeland, ed., Mind Design: Philosophy, foundations of artificial intelligence. Intermedi- Psychology, Artificial Intelligence. Cambridge, MA: ate to advanced.] MIT Press, 1981, pp. 35–66. [A classic paper ——. 1992. “Connectionism and cognition: why that has exerted great influence among com- Fodor and Pylyshyn are wrong.” In A. Clark and puter scientists and deserves more study from R. Lutz, eds., Connectionism in Context. Berlin: philosophers. Intermediate.] Springer-Verlag, pp. 37–56.[A critique of Fodor Penrose, R. 1989. The Emperor’s New Mind: and Pylyshyn’s attempt to reject connectionism Concerning Computers, Minds, and the Laws of on formal-system principles. Intermediate.] Physics. New York: Oxford University Press. [An ——. 1996. Philosophy and Cognitive Science, 2nd attack on the notion that all of human thought is ed. St. Paul, MN: Paragon House. [An introduc- algorithmic. Intermediate.] tion to cognitive science. Introductory.] Rumelhart, D. et al. 1986. Parallel Distributed Fodor, J. 1975. The Language of Thought. Cam- Processing: Explorations in the Microstructure of bridge, MA: MIT Press. [An influential position Cognition, vols. 1 (Foundations) and 2 (Psycho- that both fascinates and infuriates. Intermediate.] logical and Biological Models). Cambridge, MA: ——. 1980. “Methodological as a research MIT Press. [The studies that introduced neural strategy in cognitive psychology.” In J. Haugeland, networks, with great sensitivity to philosophical ed., Mind Design: Philosophy, Psychology, Artifi- implications. Intermediate to advanced.] cial Intelligence. Cambridge, MA: The MIT Press, Searle, J. 1980. “Minds, brains, and programs.” pp. 307–38. [Interesting reflections on methodo- Behavior and Brain Sciences 3: 417–57. [This logy. Intermediate.] classic paper provides a systematic response to —— and Pylyshyn, Z. 1988. “Connectionism and Turing (1950), including replies to Turing’s cognitive architecture: a critical analysis.” Cogni- responses to criticism. Introductory.] tion 28: 3–71. [A criticism of connectionism Stich, S. 1983. From Folk Psychology to Cognitive rooted in a formal systems conception of lan- Science: The Case Against Belief. Cambridge, MA: guage and mentality. Advanced.] MIT Press. [A brilliant analysis of alternative Ford, K. M. and Hayes, P., eds. 1991. Reasoning accounts of the mind that challenges ordinary Agents in a Dynamic World: The Frame Problem. folk psychology. Intermediate.] Greenwich, CT: JAI Press. [A broad range of Turing, A. M. 1950. “Computing machinery and conceptions of the frame problem are presented intelligence.” Mind LIX: 433–60. [The classic and explored. Intermediate to advanced.] paper in which Turing introduces his comparat- Haugeland, J. 1981. “Semantic Engines: an intro- ive test, including analysis of eight lines of objec- duction to mind design.” In J. Haugeland, ed., tion to his test. Introductory to intermediate.]

134 Computationalism, Connectionism

Chapter 10 Computationalism, Connectionism, and the Philosophy of Mind Brian P. McLaughlin

and (2) to explain how those abilities are exer- Introduction cised via operations of the system. The tasks are related. The explanation of how operations of The central questions of the philosophy of mind the system constitute exercises of our mental are the nature of mental phenomena, and how abilities will justify the claim that our possession mental phenomena fit into the causal structure of those abilities consists in our being at least of reality. The computational theory of mind partly constituted by the system. aims to answer these questions. The central tenet Computationalists hold that the functional of the theory is that a mind is a computer. Ac- architecture of the computing system that cording to the theory, mental states and events grounds our mental abilities resides in our enter into causal relations via operations of the brains. There is, however, no consensus as to computer. The main aim of the theory is to say what even the general character of that architec- what kind of computer – what kind of computa- ture is. The symbols-system paradigm and the tional mechanism – a mind is. The answer is still connectionist paradigm are the two dominant unknown. Pursuing it is the main research pro- research within the computational gram of the theory. theory of mind. They differ primarily in what In the most general sense, a computer is, kind of computer the mind is assumed to be, roughly, a system of structures functionally and thus in the kinds of functional architectures organized in such a way as to be able to com- explored. The symbol-system paradigm pre- pute. The structures, their functional organiza- supposes that the mind is a kind of automatic tion, and the basic modes of operation of the formal system, while the connectionist paradigm system when it computes comprise the functional presupposes that it is a system of connectionist architecture of the computer. The two tasks of networks. These paradigms will be discussed in the computational theory of mind are: (1) to due course. First, however, further general dis- identify the functional architecture of the com- cussion of the computational theory of mind is puting system that grounds our mental abilities in order.

135 Brian P. McLaughlin

school or by a series of additions. In fact, for any 1 The Computational Theory computable function, there will be infinitely many of Mind algorithms for computing it. The reason is trivial: we can always add intermediate steps to an algo- Cognitive functions rithm for computing the function that contribute nothing to its computation. Having a mind consists in having mental abilities How a function is computed will depend on such as, for example, the ability to think. Accord- the functional architecture of the machine (the ing to the computational theory of mind, the computing system) that computes it, including exercise of mental abilities consists in the com- the representational code used in the machine. putation of certain functions – cognitive func- Thus, for example, the algorithms for doing tions. The theory is thus concerned with what addition in base 10 differ from those for doing functions are cognitive functions and how they addition in binary notion, which uses only 1’s are computed. and 0’s. Likewise, the algorithms for doing addi- Functions are relations, but not every relation tion, multiplication, and division using Roman is a function. A function is a one–one or many– numerals are different from those for computing one relation. Functions have arguments and these functions in the Arabic numerals we were values. The set of arguments of the function is taught in grade school. its domain and the set of values is its range. For These methods of calculating are algorithms each argument in the domain, there is a unique for manipulating numerals (symbols that repres- value in the range. Functions are thus extension- ent numbers), as are the methods for calculation ally characterized as sets of ordered pairs, the using other kinds of numeral systems. However, first element of each ordered pair being an argu- the symbols manipulated by an algorithm can be ment of the function, the second, the unique symbols for anything that can be represented: value of the function for that argument. Addi- persons, places, things, etc. And since a symbol tion, multiplication, and division, for example, can purport to represent something yet fail, an are mathematical functions that take order pairs algorithm can manipulate symbols that purport of numbers as arguments and have numbers as to represent something yet fail. values. A computable function is one that can be Symbols have formal as well as semantic prop- computed. Not all functions are computable. A erties. The formal properties of symbols are in- function is computable if and only if there is an trinsic nonsemantic properties, such as shape or algorithm for finding a value of the function for syntactic structure. Symbolic algorithms operate any argument of the function. An algorithm is a on symbols in virtue of certain of their formal kind of procedure, a way of getting something properties, rather than their semantic ones. Thus, done; but not every procedure is an algorithm. An for example, the mathematical operations on algorithm is an effective procedure, a sure-fire Arabic numerals are defined over the shapes of way of getting something done in a finite number the numerals, and likewise for the mathematical of steps. It is sure-fire in that if each step is operations on Roman numerals. The beauty of followed exactly, success is guaranteed. Each step symbolic algorithms is that they can involve the of an algorithm is, moreover, easy in the sense manipulation of symbols in ways that makes sense that it requires no ingenuity or intelligence to given what the symbols represent. Thus, while carry it out, and thus “a mere mechanism” could the algorithms for adding Arabic numerals are execute it. different from those for adding Roman num- To compute is to compute a function, and to erals, the symbol transitions that occur during compute a function is to execute an algorithm. the execution of the algorithms in question can More than one algorithm can be used to com- be systematically interpreted as adding the num- pute a function. For example, the multiplication bers represented. function – understood extensionally as a set of Not all algorithms are symbol manipulation ordered pairs – can be computed using either the algorithms. An algorithm for getting all of the partial-product method taught in elementary sides of a Rubik’s Cube to match in color, for

136 Computationalism, Connectionism example, involves the manipulation of the squares electronic circuitry of such a device acts com- of the Cube, not the manipulation of symbols pletely mindlessly. representing them (Cummins & Schwarz 1991). The computational theory of mind is concerned with the algorithms by which we compute cog- The computational, algorithmic, and nitive functions. As will become apparent in due implementational levels course, while the symbol-system paradigm appeals to symbol manipulation algorithms to explain how David Marr (1982) perspicuously distinguished cognitive functions are computed, the connec- three levels of analysis for addressing a computa- tionist computational paradigm often appeals to tional problem. At the computational level of nonsymbolic algorithms to do so. analysis, one specifies what cognitive function is When we execute the partial-product algo- being computed. At the algorithmic level, one rithm using pencil and paper, for instance, each describes how the function is being computed, step of the process is guided by our intentions the algorithm used to compute it. And at the and decisions; likewise, when we execute by hand implementational level, one describes how the an algorithm for getting all of the sides of a steps of that algorithm are implemented, that is, Rubik’s Cube to match in color. Computation- the underlying mechanism by whose operations alists maintain that the execution of basic cog- they are taken. nitive algorithms (ones not executed by means The three levels are relative to a computa- of the execution of other cognitive algorithms) tional problem. What is at the implementational can be described and explained without invok- level relative to one computational problem can ing intelligent guidance for any step, and so be at either the computational or algorithmic without having to invoke “homunculi” – little level relative to another. The reason is that the intentional agents. This ultimate elimination of implementation of a step of an algorithm might homunculi is essential since computationalism itself be characterized as the computation of a aims to reductively explain intentional abilities. function, one executed by a different algorithm, (See the discussion of homunculi in Sterelny whose steps are themselves implemented some- 1990.) how; and each step of the latter algorithm might To see how an algorithm can be executed com- itself be characterized as the computation of pletely mindlessly, consider electronic circuits that a function, and so on, and so forth. Different function as truth-functional gates. For example, algorithms are executed at different scales in the an AND-gate has two input wires and one out- brain. It is an unanswered empirical question put wire. When it receives current along both of whether all cognitive functions are executed by its input wires it closes, thereby sending current algorithms involving elements at the same scale along its output wire. When it receives current (Churchland & Sjenowski 1993). from only one input wire or does not receive Marr (1982) suggested that computational current from either input wire, it remains open level analysis could be carried out largely inde- and no current is sent along its output wire. This pendently of algorithmic level analysis, and the circuit is called an “AND-gate” because using T latter largely independently of implementational to represent current being sent current along a analysis. Now it is certainly true that the com- wire and F to represent the absence of current putational level underdetermines the algorithmic being sent along a wire, we can construct a ma- level: infinitely many algorithms can compute the chine table representing how this circuit works same function. Moreover, the algorithmic level that will exactly resemble the truth-table for con- in turn underdetermines the implementational junction, the truth function expressed by central level: a given algorithm can be implemented in uses of “and” in English. An AND-gate can thus infinitely many possible ways. Nonetheless, the be interpreted as computing the truth-function computationalist’s concern is not merely to dis- conjunction. Indeed for any truth table, no mat- cover what cognitive functions are computed, ter how complex, there is a formally equivalent but also to discover the algorithms used to com- electronic circuit design, and the converse. The pute them. Since what algorithms a machine can

137 Brian P. McLaughlin execute will depend on its functional architec- every cognitive function is Turing-computable ture, attention to likely modes of implementation (see Chapter 1, COMPUTATION). can help in discovering those algorithms. Since not all functions are computable by a What algorithms a machine can execute Turing machine, a way to try to falsify the com- supervenes on the functional architecture of putational theory of mind is by showing that the machine. (A-properties supervene on B- there are cognitive functions that are not so com- properties just in case the two things cannot differ putable. Whether there are such cognitive func- in respect of A-properties without differing in tions is a subject of dispute, and so this remains respect of B-properties.) Thus, there cannot be a one line of attack on computationalism (Lucas difference in what algorithms two machines can 1961; Penrose 1989; van Gelder 1995; McCall compute without a difference in their functional 1999; Copeland 2000; see Lewis 1979 for a architectures. It follows that two machines with response to Lucas; and see Chalmers 1996 for exactly the same functional architecture can ex- responses to Penrose). ecute exactly the same algorithms. Similarly, what functions a machine can compute will supervene on what algorithms it can execute, and so two The Holy Grail of artificial machines that can execute exactly the same algo- intelligence rithms can compute exactly the same functions. Since supervenience is transitive, what functions Since the functional architecture of our minds a machine can compute will supervene on the resides in our brains, has called functional architecture of a machine. As indicated our minds “meat machines.” The Holy Grail of above, however, functional architecture does not the field of artificial intelligence (see Chapter 9, supervene on computational power. Machines THE PHILOSOPHY OF AI AND ITS CRITIQUE) is to with very different architectures can have exactly build something with mental abilities out of the same computational power. something other than living tissue. In pursuit of this Holy Grail, AI presupposes the com- putational theory of mind. If the computational Turing machines and Turing theory of mind is correct, then it is at least logic- computability ally possible for a mind essentially like ours to be made of quite different stuff from ours. The In his early machine-state psychofunctionalism, reason is that what is essential are the structures (1975) claimed that we are finite and their functional organization, not the material state automata, by which he meant that we are of which the elements of the structures are made. Turing machines with finite tapes. The idea that If the computational theory is right, then it is at our minds have the functional architecture of least logically possible that the relevant structures a Turing machine with a finite tape is not to be with the relevant functional organization can be taken seriously. No computationalist thinks the realized in something that is, for example, silicon mind has a Turing machine architecture. One of based, rather than carbon based. Putnam’s essential points, however, was that we AI has been impressively successful in design- have the computational capacities of a Turing ing machines that can perform difficult tasks with- machine with a finite tape, in that our cognitive out our supervision. Moreover, it has been at functions are computable by a Turing machine. the cutting edge of research into how cognitive According to the computational theory of mind, functions might be computed. Finding the Holy all cognitive functions can be computed by Grail, however, remains an unrealized dream. We algorithms – effective procedures. According to have computers that play master’s-level chess and the Church–Turing thesis, every function that teams of robots that play soccer, but there are can be computed by an algorithm can be com- no artifacts that are even remotely plausible can- puted by a Turing machine. If the Church– didates for being the subjects of mental abilities. Turing thesis is correct, then the computational And despite the optimism of some AI researchers, theory of mind is committed to the thesis that there do not seem to be any in the works. It is

138 Computationalism, Connectionism important to note, then, that the computational which discrete items are automatically manip- theory of mind is not committed to the success ulated according to the rules – like a chess set of this AI project. It leaves it an open empirical that plays chess by itself. An interpreted auto- question whether the dream will ever be realized. matic formal system is an automated formal For all we know, it may be that given the laws of system in which the discrete items that are nature and the initial conditions of the universe manipulated include symbols or representations: it is physically impossible for a mind (one essen- discrete items with semantic and formal prop- tially like ours) to be realized in something com- erties. The rules by which they are manipulated posed entirely of (e.g.) silicon. Such a mind might prescribe algorithmic operations on them. Thus, be impossible in the way that a 90-foot tall this paradigm is sometimes called “the rules and human being is impossible. The computational representations” paradigm. theory leaves this empirical issue open. Turing machines are interpreted automatic Nor is the computational theory of mind com- formal systems. There are many computationally mitted to Turing’s (1950) would-be test (see equivalent machines that are automatic formal Chapter 9) for determining whether something systems with different functional architectures is genuinely intelligent (as opposed to merely from that of a Turing machine (see Chapter 1, appearing intelligent). The theory leaves open COMPUTATION). One such machine is a von that there is more to intelligence than matching Neumann machine, so named after John von the dispositions to verbal or written behavior of Neumann. Virtually all commercial computers an intelligent human being. It is not committed are von Neumann machines. The functional archi- to the view that there is some pattern of dis- tecture of such a machine includes a Central positions to outward or peripheral behavior that Processing Unit (CPU), an arithmetic unit, suffices for the possession of genuine intelligence memory locations, and two kinds of memory (Block 1981; Jackson 1993; McLaughlin & access. The CPU has access to memory loca- O’Leary-Hawthorne 1994). tions and current active data structures and determines what operations the computer is to perform by consulting instructions (programs) located in memory. A somewhat similar func- 2 The Symbol-system Paradigm tional architecture is implicit in some symbolic models of mental abilities: online processing is Mind as an interpreted automatic done using a short-term memory store that holds formal system information relevant to the process being carried out; and online processing influences and is According to proponents of the symbol-system influenced by long-term memory. Nevertheless, paradigm, the mind is an interpreted automatic it is universally accepted by computationalists formal system (Haugeland 1985). More precisely, that the functional architecture of the mind is having mental abilities consists in being consti- not von Neumann architecture. tuted, at least partly, by an interpreted auto- Higher-level programming languages such as matic formal system. A formal system is a system Basic, Pascal, FORTRAN, Cobol, Java, and C++, for manipulating discrete items in virtue of their and Lisp (List Processing) are (or describe) uni- formal properties; an automatic formal system is versal machines that are automatic formal systems. one that does so automatically. Many games, Such languages are very useful since, as John among them chess and checkers, are formal sys- Haugeland (1985) has aptly noted, some ma- tems. Games that are formal systems share the chines are easy to build and some are easy to following features: they are finitely playable, program. So we build the machines that are easy and are played by manipulating discrete items to build and use them to simulate the machines according to the rules of the game. In chess for that are easy to program. Higher-level program- instance, the discrete items are the chess pieces, ming languages are thus simulated machines rel- which are manipulated according to the rules ative to the machine language of the computers of chess. An automated formal system is one in we actually build; they are thus virtual machines

139 Brian P. McLaughlin relative to the machine language of the computer. and representations paradigm presupposes that (A complier is a program that translates the the mind is “a syntactic engine.” higher-level language into the machine language of the actual physical machine – in the case of von Neumann machines, into instructions en- The language of thought coded into strings of 1’s and 0’s.). No one thinks that any of these higher-level programming lan- Jerry Fodor and Zenon Pylyshyn have articulated guages characterize the functional architecture a research program in the symbolic paradigm for of the mind. reductively explaining propositional attitudes and Production Systems are Turing-machine mental processes involving them, a program that equivalent higher-level programming languages invokes the hypothesis that there is “a language used in research in the symbolic paradigm. Pro- of thought” (Fodor 1975; Pylyshyn 1986; Fodor duction Systems consist of rules specifying actions & Pylyshyn 1988). Propositional attitudes have to be performed when certain conditions are an intentional mode and a propositional content. met. While some computationalists think that at The intentional modes include belief, desire, hope, least some cognitive modules may have a Pro- wish, fear, intention, and the like. The proposi- duction System architecture, it is widely held by tional contents of propositional attitudes are researchers in the symbolic paradigm that we do expressed using that-clauses. Thus, the belief that not yet know what the functional architecture of that the cat is on the mat has the propositional the mind is. The fundamental research objective content that the cat is one the mat; and the hope of the symbolic paradigm is to determine what that it will not rain has the propositional content kind of automatic formal system the mind is. that it will not rain. According to the language- of-thought hypothesis, some mental symbols are atomic, and some are molecular in that they Mind as syntactic engine contain other mental symbols as constituents. Moreover, the mental symbol system has a com- Recall that while a symbolic algorithm is defined positional semantics: the content or meaning of over the formal properties of symbols (e.g., their a molecular symbol is a function of its syntactic shapes or syntactic structures), symbolic algo- structure and the contents of its constituent atomic rithms can govern symbol-to-symbol transitions symbols. The contents of propositional attitudes that make sense given the meanings of the sym- are expressed by sentence-like mental symbols; the bols. In proof theory, logical relations are treated contents of concepts, by word-like symbols. On as moves in a formal game; the formal moves this view, to be in a state within a certain inten- preserve truth: they never take one from a truth tional mode is to be disposed to compute in a to a falsehood. An algorithm can preserve truth: certain way with an amalgam of concepts with a if one begins with a true symbol, the algorithm sentence-like syntactic structure – a mentalese will take one only to a true symbol. The symbol- sentence. Thus, to believe that p is to be disposed system paradigm is thus often said to presuppose to compute in a certain way with a mentalese sen- a proof-theoretic conception of mind. Moreover, tence that means that p; to desire that p is to be algorithms can be defined over the formal prop- disposed to compute in a certain other way with erties of symbols so as to preserve other sorts a mentalese sentence that means that p, and so on of semantic properties of the symbols, so that for the other propositional attitudes. It is a topic symbol transitions can implement reasoning pro- for empirical investigation what the grammar of cesses of all sorts, not only deductive reasoning, the language of thought is and what, exactly, but , analogical reasoning, these ways of being disposed to compute are. decision-making, etc. Haugeland (1985) suggests As Fodor and Pylyshyn (1988) point out, if the following as the symbolist motto: “Take care the mental symbol system includes a language of of the syntax (the formal operations), and the thought in the above sense, then we can appeal semantics will take care of itself.” As Daniel to it to explain, among other things, the system- Dennett has aptly noted, the symbolic or rules aticity and productivity of thought. They claim

140 Computationalism, Connectionism that thought is productive in that one can, ide- “all” and “some”) – can be defined by patterns ally, think an infinite number of thoughts; one is of inferential relations. However, it seems that prevented from doing so only by the fact that most symbols (e.g., “cow” and “run”) cannot. one has a finite lifespan and finite memory re- Semantic externalists claim that the meanings sources. And they claim that thought is system- of many mental symbols fail to supervene on atic in that pairs of abilities to have thoughts in anything that goes on in our heads, since they the same intentional mode and with related con- depend on environmental relations. Externalists tents are counterfactually dependent, in that one typically claim that a computational theory of would have one member of the pair if one would must thus be supplemented with have the other. Thus, normally, an individual an externalist naturalistic account of meaning that that can think that Romeo loves Juliet can think makes no appeal to intentional notions. There are that Juliet loves Romeo; and, normally, an indi- various programs for attempting to explain the vidual that can hope the rock is next to the tree semantic properties of mental symbols in purely can hope the tree is next to the rock. naturalistic terms. However, the naturalistic rela- tions appealed to (e.g., causal relations, coun- terfactual dependencies, information-theoretic Psychosemantics relations, processes of natural selection), all appear to leave the semantic properties of mental The question naturally arises of course as to representations underdetermined. Dual-aspect how mental symbols – mental representations – semanticists hope to combine inferential or con- acquire their contents or meanings. The num- ceptual role semantics with a naturalistic externalist erals of various numeral systems such as Arabic account to determine the semantic properties of numerals and Roman numerals derive their mean- mental representations (Block 1986). This project, ings from the linguistic conventions of commu- however, faces the problem of specifying which nities. The current along the input and output inferential relations are constitutive of the internal wires of an electronic computer circuit derive component of meaning (Fodor & LePore 1992). their meanings from our intentional stipulative Moreover, it is uncertain whether any such com- assignments; as do the keys of hand-calculators bination will yield determinate meanings for and the displays on their screen. If intentionality mental symbols. Saul Kripke (1981) has argued is to be explained by appeal to mental representa- that since our inferential dispositions are in fact tions, then we need an account of their meaning finite (due to limitations in memory, our finite that makes no appeal to linguistic conventions or lifespan, etc.), they will leave the meanings of semantic intentions, for these presuppose inten- symbols radically indeterminate. Thus, suppose tionality (Fodor 1990). that one is in fact disposed to perform only 500 Procedural semantics and inferential or con- million calculations with a symbol “+”. Even if ceptual role semantics attempt to explain how the output of each of these 500 million calcula- mental representations derive their meaning from tions can be systematically interpreted as a com- their participation in computational processes. putation of the plus function, one’s dispositions John Searle’s “Chinese Room” argument (see to calculate will be compatible with the symbols Chapter 9 above) makes a case that the semantic expressing instead the “quus function,” which properties of symbols do not supervene on their can be defined as follows: x quus y = x plus y, for syntactic relations, and so that intentional phe- any y < 500,000,001, otherwise x quus y = 5. nomena such as (e.g.) understanding Chinese Conceptual role theorists take this problem to cannot be explained solely by appeal to computa- be one of how conceptual roles should be ideal- tional relations (see e.g. Searle 1999). If he is ized. But no entirely satisfactory solution to the right about this failure of supervenience, then the Kripke’s problem has yet been proposed. Suffice kinds of semantics in question cannot succeed. it to note that the problem of how mental rep- Perhaps logical symbols – truth-functional sym- resentations acquire their meanings is perhaps bols (e.g., “and” and “or”) and symbols for the the deepest problem that any reductive theory universal and existential quantifiers (respectively, of intentionality faces.

141 Brian P. McLaughlin

program stored in memory; the program is im- 3 The Connectionist plicit in the pattern of connectivity exhibited by Computational Paradigm the units. Many units process information simul- taneously, and so the network as a whole engages The functional architecture of our minds is some- in parallel distributed processing (PDP) (see how realized in our brains. One of the few things McClelland & Rumelhart et al. 1986; Rumelhart we know about the brain that seems to have a & McClelland et al. 1986). bearing on mentality is that it is, inter alia, a Units have states of activation. Depending on system of neural networks. Neural networks are the network architecture, a unit may have only not well understood; indeed neurons themselves two states of activation, “on” and “off,” three are not well understood. There are somewhere or more discrete states of activation, or continu- between 50 and 500 different kinds of cortical ous levels of activation, bounded or unbounded. neurons, and these different kinds of neurons Units process information by changing or re- appear to perform specialized functions not yet taining their state of activation in response to understood (Churchland & Sejnowski 1993). activation signals from their sending units. Nonetheless, just as top-down considerations of The connections among units can be of vari- what cognitive functions are being computed can ous strengths or weights, which affect the extent guide research into how mental abilities are seated to which the activity output of a sending unit in the brain, bottom-up considerations of the influences the activation states of receiving units. workings of neural networks can guide research Connections can, moreover, be excitatory or into what algorithms are used to compute the inhibitory. The connection from a unit Ui to a functions in question. They can serve as a guide unit Uj is excitatory if the activation output of to discovering the functional architecture of the Ui tends to increase the level of activation of Uj, mind. and inhibitory if it tends to decrease it. The While the operations of actual neural networks causal influence exerted by a sending unit on a are not well understood, the connectionist com- receiving unit thus depends on two intrinsic prop- putational paradigm explores functional archi- erties of their connection: its weight and whether tectures that are at least “neurally inspired.” Some it is excitatory or inhibitory. The notation “wij” connectionist networks, called “artificial neural is used to stand for a real number that indexes networks,” are specifically designed to approx- the connection from unit Ui to unit Uj by its imate various kinds of real neural networks in weight and kind. The number is positive when various respects: the units (or nodes) in artificial the connection is excitatory, negative when it is neural networks are analogous to neurons, the inhibitory. The weight of the connection is the connections among units analogous to axons, absolute value of wij. In many networks, the and the weights or strengths of the connections extent and kind of causal influence a unit Ui analogous to synapses. These networks are extens- exerts on a unit Uj is indexed by the product of ively employed in the growing field of computa- wij and the activation value of Ui. tional neuroscience (see Churchland & Sejnowski Whether a unit changes or retains its state 1993). The connectionist networks employed to of activation is a function of three factors: the model the computation of cognitive functions totality of input to the unit, the unit’s current typically make no attempt at neural realism. activation state, and the unit’s bias (if any), which Nonetheless, they have at least “a neural flavor.” may be positive or negative. A unit thus com- putes an activation function that maps its total network activity input (and external input if any), Connectionist networks its current activation state, and its bias (if any) to an activation state. The total network activity

A connectionist network is composed of inter- input to Uj is the sum of all of the causal influ- connected units (or nodes). Individual units ences from sending units: it is the sum of the do all the information processing: there is thus product of each activation value from a sending no executive or CPU. There is, moreover, no unit and the real number indexing the weight

142 Computationalism, Connectionism and kind of connection the sending unit bears inhibitory, while units outside of the pool bear to Ui. While the output function of a unit can excitatory connections to one or more units in be linear, it is typically a threshold function, and the pool. In recurrent networks, there are con- thus nonlinear. If a unit has a negative bias, it nection patterns that contain loops, so that a may send 0, the null signal, unless its activation unit is either related to itself as an input unit or value exceeds a certain threshold. If it has a posit- there is a series of connections from the unit ive bias, it may send a certain non-0 activation back to itself, so that the output of a unit at one value unless its activation state value falls below time can causally influence its activation state at a certain threshold. In Hopfield networks, units another. In auto-associative networks, each unit have a sigmoidal (S-shaped) response to net is connected to every other unit, including itself. input: their output only increases by a given These are only some of the many kinds of patterns amount given an increase in net input; after that, of connectivity that are possible. they increase no further. But units can also have The behavior of a network as a whole is a Gaussian (bell-shaped) output functions and consequence of the pattern of connectivity ex- other sorts of nonlinear ones as well. hibited by its units at a time and the global activa- The input units of a network receive signals tion state of the network at that time. A vector directly from the environment of the network, (ordered list) of real numbers is used to index while the output units send signals directly to the global activation state. The activation value the environment. Since input and output units of each unit in the network at that time is indexed directly interact with the environment, they are by one and only one element of the vector, and called “visible units.” So-called “hidden units” so the number of elements in the vector will be only directly interact with other units. equal to the number of units in the network. Some networks, called “perceptrons,” have The set of all possible global activation states of only two layers of units: a layer of input units a network is its activation space, whose dimension- and a layer of output units. Minsky and Pappert ality is exactly equal to the number of units in (1969) showed perceptrons are limited in their the network. A network with n units will thus computational power: for example, they cannot be indexed by an n-tuple of real numbers that compute XOR (exclusive-or). The reason is that identifies a position in an n-dimensional vector the problem of determining the value of the space. That position represents the global activa- XOR function is a linearly inseparable problem tion state of the network at a time. A temporal and perceptrons cannot solve any such problem. series of global activation states will trace a path Rumelhart et al. 1986 demonstrated, however, through a vector space. Network information that networks with three or more layers can processing is characterized as the evolution compute XOR and, more generally, can solve through time of global patterns of activation. decidable linearly inseparable problems. Networks can be systematically interpreted as Feedforward networks with one or more performing mathematical operations on matrices layers of hidden units are called “multilayered” such as matrix multiplication (e.g., a computing networks. The Hamming net, for instance, is inner product). a widely used feedforward network with three Explicit representations in a connectionist layers of units, one of which is a layer of hidden network can be either local or distributed. Local units. Feedforward networks are non-interactive: representations are individual units, or individual activation flows from the input units through units at certain levels of activation. Distributed each layer of hidden units to the output units. representations are patterns of activity over a In interactive networks, two units can mutually group of units. The pattern of connectivity of a influence each other, and thus a unit can be network is sometimes characterized as implicitly related to another both as an input unit and as representing. an output unit. In interactive networks, two-way As a result of the pattern of connectivity connections between units are often symmetrical, exhibited by its units, a network as a whole = so that wij wji. In competitive networks, units can behave in rule-like ways to compute func- form pools: the units in a pool are all mutually tions. When a feedforward network computes a

143 Brian P. McLaughlin function, the arguments of the function are rep- their simultaneous activations (Rumelhart et al. resented by different patterns of activity over the 1986). The Delta rule is an error correction rule input units, and the values of those arguments that changes the weights leading from units send- for the function by patterns of activity over out- ing signals to output units on the basis of the put units. In computing the function, the net- discrepancy between the actual output and the work acts as a look-up table. Input activation desired one. Backpropagation is a generalization patterns function like questions posed to the of the Delta rule, and is widely used in multi- network (for example, “To what category does layered networks. The actual output activation this belong?”), and output patterns function like pattern for a given input activation pattern is answers to the question (“To category C”). compared with the desired output. The differ- Unlike in a symbolic look-up table, however, ence between the two is then propagated back the answers are not stored as data structures; into whatever connections were used to get the rather they are implicitly represented in the actual output activation pattern. The connec- pattern of connectivity. (Sometimes, however, tions among units that contributed to the actual hidden units are explicit representations.) In output are strengthened (increased in weight) interactive units, the argument of a function when the match is good, and are reduced in might be represented by an initial pattern of strength (decreased in weight) when it is poor. activation over units in the network, and the The weights are thus adjusted so as to reduce value of the function for that argument by a the margin of error between the actual output pattern of activation over those same units, a and the desired one. And in this way, the network pattern the network “settles” or “relaxes” into learns. after information processing. Network training can be supervised or un- The “neural flavor” of connectionist networks supervised. In supervised learning, the network is by no means their only attraction for computa- is provided explicit feedback from an external tionalists. Connectionist networks are good at source about what output is desired as a response pattern recognition tasks: pattern matching, pat- to a certain input. The Delta rule and back- tern completion, and pattern association. They propagation are used in supervised learning. In degrade gracefully in their performance of a task unsupervised learning, no such external feedback as a network is damaged (e.g., when a unit is lost) is provided to the network; rather, the network and in the face of noisy or incomplete data. They monitors its own performance through internal can learn. That is to say, with proper training, feedback. There are a number of unsupervised they can improve their performance of various learning algorithms; the Kohonen algorithm is tasks. Moreover, they are very efficient at solving one such (see Beale & Jackson 1990, ch. 5). connected problems and at arriving at optimal or near optimal solutions to best-match problems. Of these attractions, more below. Pattern recognition

NETtalk offers a vivid example of the ability of Learning networks to learn to recognize patterns. Designed by Terrence Sejnowski and C. R. Rosenberg The weights of the connections in a network are (1987), it is a network that learns to map letters not architecturally fixed. Learning consists in onto phonemes. NETtalk is a three-layered changes in the weights. There are various learn- feedforward network, whose input units repres- ing rules that govern weight change. The Hebb ent letters (individual letters are represented by Rule is based on Donald Hebb’s (1949) hypo- patterns of activation over 29 input units and thesis that the connections between two neurons there are 7 such groups of 29 input units) and might strengthen whenever they fire simultane- whose output units represent phonemes. The ously. According to the Hebb rule, the weight of network feeds into a synthesizer. After sufficient a connection between units should be increased training using backpropagation, when presented or decreased in proportion to the products of strings of letters comprising the words of actual

144 Computationalism, Connectionism

English text, the network drives the synthesizer visit each of a number of cities, while visiting each to sound like a robotic voice literally reading the city only once. Since which city the salesman text. visits depends on which cities he has already Some studies have compared networks using visited, the problem does not divide into inde- backpropagation with various members of the class pendently solvable subproblems. The traveling of “top-down inductive decision trees” (“deci- salesman problem is decidable, and so can be sion trees” for short) in respect of accuracy in solved by a symbol system. But if, for instance, pattern recognition and the ability to deal with there are 20 cities to visit, there is a minimum of noisy, incomplete data gracefully, as well as in 653,837,184,000 possible routes to take (Raggett other respects (see Sethi & Jain 1991). Decision & Bains 1992). Thus, as the number of cities trees are production-rule systems, and thus sym- increases there is an exponential increase in the bolic systems, that excel at pattern recognition computational resources required to solve the tasks (see Utgoff 1999). One comparative study traveling salesman problem within a symbol sys- by Shavlik et al. (1991) included a comparison tem (see Chapter 2, COMPLEXITY). A Hopfield between a decision tree system called “ID3” and network is able to find a solution in a small a network using backpropagation on both the number of training cycles. NETtalk A data set and the NETtalk full data Networks are also especially efficient at finding set. Shavlik et al. noted: “for the most part [both] optimal or near optimal solutions to what Minsky learning systems are similar with respect to accur- and Papert (1969) called “best-match problems” ately classifying novel examples” (1991: 120). – problems whose solution involves assessing the They also compared ID3 and the network on satisfaction of multiple soft (i.e., non-mandatory) three kinds of noisy data: inaccurate feature values, constraints. Hinton (1977) developed the first missing feature values, and insufficient number network approach to solving best-match prob- of features. The network performed slightly lems. The following description of it paraphrases better than ID3 on inaccurate feature values and the description provided in McClelland and missing feature values, but when trained on Rumelhart 1988. Units in the network have one insufficient numbers of feature values, “ID3 and of two states of activation (“on” or “off ”), and backpropagation degrade at roughly the same the network contains local representations, each rate as the number of features is reduced” (1991: unit representing a different hypothesis. The 130). Several other comparative studies have connections in the network implicitly represent found that networks using backpropagation are evidential relationships among the hypotheses. roughly comparable to various other members of Thus if a hypothesis H is evidence for another the family of decision trees as concerns accuracy hypothesis H′, the connection from the unit in pattern recognition and graceful degradation representing H to the unit representing H′ is in response to noisy, incomplete data. Decisions positive; if H is evidence against H′, the connec- trees are, however, much faster at learning than tion is negative. The stronger the confirming or networks using backpropagation (see Sethi & disconfirming relationship the one hypothesis Jain 1991; Marinov 1993; and McLaughlin & bears to the other, the stronger the connection Warfield 1994). between the units that represent them. If one hypothesis H entails another H′, then the con- nection between the unit representing H and Connected problems, best-match the unit representing H′ will be such that the problems, and multiple soft constraint second unit is on whenever the first is on. When satisfaction two hypotheses are incompatible, they are con- nected in such a way that they cannot both be Networks are especially efficient at solving con- on. Since the network contains input units that nected problems – problems that do not divide receive signals from the environment, hypo- into independently solvable subproblems, like the theses can receive confirming or disconfirming traveling salesman problem. The goal is to find evidence directly from the environment. The fact the shortest route that a salesman can take to that different hypotheses have different a priori

145 Brian P. McLaughlin probabilities is captured by biasing the relevant machine employs a computational analog of the units. The overall goodness of fit of a particular metallurgical process of annealing. Annealing is hypothesis to evidence is measured by the sum a process whereby metals are heated to a little of the individual constraints on the activation below their melting point and then cooled very value of the unit representing the hypothesis. slowly so that all their atoms have time to settle The activation values of units range between a into a single orientation. The analog of heat in minimum and a maximum. The maximum value the Boltzmann machine is random noise that is means that the hypothesis should be accepted, introduced into network activity. The function the minimum that it should be rejected, and of the noise is to jar the network out of local intermediate values correspond to intermediate valleys, so that it can explore other parts of the levels of certainty. The constraint satisfaction energy landscape to find the lowest valley, thereby problem is thus reduced to one of maximizing achieving the global maximum of fit. When the this overall goodness of constraint fit. network reaches a stable state, it has settled or Of course, this approach to best-match prob- relaxed into a solution. Given sufficient time, lems may well not seem interestingly different the Boltzmann machine can find the energy min- from a symbolic approach; indeed it might be imum for any best-match problem. viewed as a way of implementing a symbolic decision procedure on a network. The computa- tionally relevant formal property of a symbol Networks and cognitive abilities can be a unit’s being activated, and symbols can participate in probabilistic algorithmic processes. There are network models of certain aspects of There are other networks, however, that solve motor control and low-level perception. For best-match problems in strikingly different, and example, there is a network model for the vesti- striking efficient ways. bular ocular reflex, which enables eyes to track A problem that arises for procedures for an object as the head moves (Churchland & solving best-match problems is that of avoiding Sejnowski 1993). As concerns low-level visual local maxima of goodness of constraint fit. It perception, the linear models color-constancy can be characterized as an energy minimization algorithm, for instance, is characterized in con- problem (McClelland & Rumelhart 1988). The nectionist terms (Maloney & Wandell 1986). analog of the goodness maximum is the energy Moreover, opponent processing theory, the lead- minimum, and the analogs of local goodness ing neuro-computational theory of color experi- maxima are local energy minima. The situation ence, is easily understood in connectionist terms is easy to visualize as an energy landscape. In an (Hurvich 1981). According to the theory, there energy landscape, the goodness maximum cor- are pairs of opponent channels that respond responds to the lowest valley in the landscape, differently to the outputs of our three types of while local goodness minima correspond to local cones (L-cones, M-cones, and S-cones): the RED valleys in the landscape. The problem of avoid- channel and the GREEN channel, the BLUE ing local goodness maxima is thus the problem channel and the YELLOW channel, and the of avoiding settling into a local valley, rather BLACK channel and the WHITE channel. Activ- than into the lowest valley in the energy land- ity in one channel inhibits activity in its oppon- scape. John Hopfield (1982) observed that some ent channel. The theory explains the fact that networks behave in such a way as to minimize a nothing looks bluish-yellow as a result of the measure over the whole network, one he aptly fact that activity in the BLUE channel inhibits called “energy.” The behavior of these networks activity in the YELLOW channel and conversely. can be characterized as moving through an energy And the theory explains the appearance of unique landscape. The Boltzmann machine, developed blue (blue that is not at all reddish or greenish) by Hinton and Sejnowski (1986) and named as the result of the activation of the BLUE chan- after the physicist Ludwig Boltzmann, is such a nel, the YELLOW channel being deactivated, network. To handle the problem of local maxima and the RED and GREEN channels being in of goodness of constraint fit, the Boltzmann equilibrium. These channels can, of course, be

146 Computationalism, Connectionism understood as pools of interconnected neuron- the subsymbolic paradigm. Some connectionists like units. have suggested that the relationship between the Connectionist modeling has by no means been symbolic paradigm and the connectionist (sub- restricted to motor control and low-level per- symbolic) paradigm is analogous to the relation- ception. It has been extended to most areas ship that Newtonian physics bears to quantum of cognition. Connectionist networks have been mechanics (Rumelhart et al. 1986). But pro- appealed to as mechanisms for processes of ponents of the implementational view of connec- categorization in terms of resemblance to proto- tionism will claim that the more apt analogy is types (Churchland 1995). And there are, more- the relationship chemistry bears to quantum over, connectionist models of various aspects of mechanics. Atoms are constituted by subatomic language comprehension and production. For particles, and chemical processes are implemented example, Jeffrey Elman (1990) has explored how by quantum mechanical ones (indeed all causal recurrent nets can learn to become sensitive to processes are ultimately implemented by quan- anaphoric relations. And Alan Prince and Paul tum mechanical ones). Perhaps, analogously, Smolensky’s (1993) optimality theory, for in- atomic symbols are constituted by patterns of stance, outlines how a multiple soft constraint activation, and connectionist processes implement satisfaction might serve as a natural language symbolic algorithmic processes (McLaughlin parser might work by employing multiple soft 1993a). constraint satisfaction. This sample represents Yet another possibility is that the functional only a tiny fraction of connectionist cognitive architecture of the mind is either thoroughly sym- modeling. (For a recent very brief overview, see bolic or thoroughly connectionist. Proponents McClelland 1999.) of this view see the two paradigms as in a sort of zero-sum competition. The leading objection to the view that the functional architecture of the mind is thoroughly connectionist is due to 4 How are Paradigms Related? Fodor and Pylyshyn (1988). They argue that a connectionist architecture cannot explain the The relationship between the symbolic and systematicity of thought without implementing connectionist paradigms is a topic of heated a symbolic architecture. A large literature has dispute. It will have to suffice here simply to list arisen in response to this objection. Some deny some possibilities. that thought is systematic. Some concede that One possibility is that the functional architec- thought is systematic, but maintain that it is not ture of the mind is hybrid: some aspects of cog- the job of a theory of cognitive architecture to nition are symbolic and some are connectionist. explain systematicity (see McLaughlin 1993b for For example, perhaps motor control modules and a discussion). And some attempt to show how low-level perceptual modules have a connectionist a connectionist architecture could explain the architecture, while linguistic modules and reason- systematicity of thought without implementing ing in central processing have a symbolic archi- a symbolic one (see e.g. Smolensky 1991; for tecture. On this view, the mind is not a single a response, see Fodor & McLaughlin 1990; kind of computer, but has different kinds of com- McLaughlin 1997a; for a counter-response, see partments that are different kinds of computers. Cummins 2000). Another possibility is that the functional Another possibility still is that the functional architecture of the mind is a symbolic architec- architecture of the mind somehow integrates ture, but this architecture is implemented by a features of symbolic architectures and features connectionist one. (Both Turing machines and of connectionist architectures. Smolensky (1994, Production Systems, it should be mentioned, 1995) has sketched an architecture he calls “an have been implemented by connectionist net- Integrated Connectionist Symbol architecture works.) Connectionism, it is often claimed, is con- (ICS)” that attempts to include features of both cerned with microcognition; and its advocates connectionist and symbol architectures, and which often characterize the connectionist paradigm as he claims will explain the systematicity and

147 Brian P. McLaughlin productivity of thought. But it remains a question Churchland, P. M. 1995. The Engine of Reason, the whether when developed sufficiently to explain Seat of the Soul: A Philosophical Journey Into the systematicity and productivity, it would collapse Brain. Cambridge, MA: MIT/Bradford. [Dis- into an implementation architecture for a sym- cussion of the role of networks in understanding bolic one (McLaughlin 1997a). mentality. Accessible to advance undergraduates.] Of course, not all of these possibilities are Churchland, P. S. and Sejnowski, T. J. 1993. The exclusive. It may be that some cognitive modules Computational Brain. Cambridge, MA: Bradford/ are thoroughly connectionist, and that various MIT. [An introduction to neural network model- other modules are symbolic but implemented ing in the field of computational neurscience. Accessible to advanced undergraduates.] by connectionist networks. There is, it should Copeland, J. B. 2000. “Narrow versus wide mechan- be noted, a growing body of work in the field ism: including a re-examination of Turing’s views of machine learning that attempts to develop on the mind–machine issue.” Journal of Philo- machines with a mixed architecture, including sophy 1: 5–32. [A defense of the view that a mech- symbolic and connectionist subcomponents. Of anistic view of the mind need not treat minds as course, this work in machine learning is not Turing equivalent computational mechanisms. concerned with either psychological or neuro- Accessible to advanced undergraduates.] logical realism (but rather with building better Cummins, R. and Schwarz, G. 1991. “Connection- machines). Nonetheless, it demonstrates the use- ism, computation, and cognition.” In Horgan & fulness of such mixed architectures for computa- Tienson 1991: 60–73. [A discussion of the role tional purposes. of connectionism in the computational theory of Suffice it to note that the functional architec- mind. Accessible to advanced undergraduates.] ture of the mind remains an open question. What Elman, J. L. 1990. “Finding structure in time.” the future will bring remains to be seen. Cognitive Science 14: 179–211. [An attempt to show how connectionist networks can learn to compute certain linguistic functions. Accessible References to graduate students.] Fodor, J. 1975. The Language of Thought. New York: Beale, R. and Jackson, T. 1990. Neural Comput- Crowell. [A seminal defense of the language- ing. Bristol: Adam Hilger. [An excellent intro- of-thought hypothesis. Accessible to advanced duction to neural network computing that is undergraduates.] accessible to advanced undergraduates.] ——. 1990. A Theory of Content and Other Essays. Bechtel, W. and Abrahamsen, A. 1991. Connection- Cambridge, MA: MIT/Bradford. [A seminal dis- ism and the Mind. Oxford: Blackwell. [An excel- cussion of issues in psychosemantics. Accessible lent introduction to connectionist networks and to advanced undergraduates.] their contribution to the understanding of men- —— and LePore, E. 1992. . Malden, MA: tality, and a major source of the discussion of the Blackwell. [Contains criticisms of inferential/ connectionist paradigm in this chapter. Access- conceptual role semantics. Accessible to advanced ible to advanced undergraduates.] undergraduates.] Block, N. 1981. “Psychologism and .” —— and McLaughlin, B. P. 1990. “Connection- Philosophical Review 90: 5–43. [A defense of the ism and the problem of systematicity: why view that mentality does not supervene on any Smolensky’s solution doesn’t work.” Cognition pattern of dispositions to outward or peripheral 35: 183–204. Repr. in Horgan & Tienson 1991. behavior. Accessible to advanced undergraduates.] [A response to Smolensky 1991. Accessible to ——. 1986. “An advertisement for a semantics for advanced undergraduates.] psychology.” Midwest Studies in Philosophy 10: —— and Pylyshyn, Z. 1988. “Connectionism and 615–78. [A seminal defense of conceptual role cognitive architecture.” Cognition 28: 183–204. semantics. Accessible to advanced undergraduates.] [A classical attack on the view that cognitive Chalmers, D., ed. 1996. Psyche, vol. 2, . [Contains responses Accessible to advanced undergraduates.] to Penrose 1989. Accessible to graduate students Haugeland, J. 1985. Artificial Intelligence: The and professionals.] Very Idea. Cambridge, MA: MIT/Bradford.

148 Computationalism, Connectionism

[An excellent introduction to the symbolic connectionist approach to the mind. Accessible approach to artificial intelligence. Accessible to to advanced undergraduates.] undergraduates.] Maloney, L. T. and Wandell, B. A. 1986. “Color ——, ed. 1997. Mind Design II. Cambridge, MA: constancy: a method for recovering surface spec- MIT/Bradford. [Contains classical articles in the tral reflectance.” Journal of the Optical Society of field of artificial intelligence and the computa- America A 3: 29–33. [Presents the linear models tional theory of mind. Accessible to advanced color-constancy algorithm. Accessible to graduate undergraduates.] students.] Hebb, D. 1949. The Organization of Behavior. New Marinov, M. S. 1993. “On the spuriousness of York: Wiley & Sons. [A classical work on neural the symbolic/subsymbolic distinction.” Minds networks. Accessible to advanced undergraduates.] and Machines 3: 253–70. [Contains a discussion Hinton, G. E. and Sejnowski, T. J. 1986. “Learn- of comparative studies of connectionist networks ing and relearning in Boltzmann machines.” In and decision trees. Accessible to advanced Rumelhart et al. 1986: 282–317. [A presentation undergraduates.] of Boltzmann machines. Accessible to graduate Marr, D. 1982. Vision. New York: Freeman. [A students.] seminal work on low-level vision. Accessible to Hopfield, J. J. 1982. “Neural networks and physical graduate students.] systems with emergent collective computational McCall, S. 1999. “Can a Turing machine know abilities.” Proceedings of the National Academy of that the Gödel sentence is true?” Journal of Philo- Sciences, USA, 79: 2554–8. [Presents what have sophy 10: 525–32. [Appeals to results of Kurt come to be called “Hopfield nets.” Accessible to Gödel to argue that we are not computationally graduate students.] equivalent to Turing machines. Accessible to Horgan, T. and Tienson, J., eds. 1991. Connec- graduate students.] tionism and the Philosophy of Mind. Dordrecht: McClelland, J. L. 1999. “Cognitive modeling, Kluwer. [A collection of essays on philosophical connectionist.” In Wilson & Kiel 1999: 137–41. issues concerning connectionism. Accessible to [A brief overview of connectionist modeling. advanced undergraduates.] Accessible to advanced undergraduates.] Hurvich, L. M. 1981. Color Vision. Sunderland, —— and Rumelhart, D. E. 1988. Explorations MA: Sinauer. in Parallel Distributed Processing: A Handbook of Jackson, J. 1993. “Appendix A [for philosophers].” Models, Programs, and Exercises. Cambridge, MA: Philosophy and Phenomenological Research 53: MIT/Bradford. [An introduction to connectionist 901–6. [Argues that mentality does not super- modeling, and a major source of the discussion vene on dispositions to outward or peripheral of the connectionist paradigm in this chapter. behavior. Accessible to advanced undergraduates.] Accessible to advanced undergraduates.] Kripke, S. 1982. Wittgenstein on Rules and Private ——, ——, and the PDP Group, eds. 1986. Paral- Language. Cambridge, MA: Harvard University lel Distributed Processing: Explorations in the Press. [Contains a presentation of the plus–quus Microstructure of Cognition, vol. 2, Psychological problem for computationalists. Accessible to and Biological Models. Cambridge, MA: MIT/ advanced undergraduates.] Bradford. [One of the two “bibles” of connection- Lewis, D. 1979. “Lucas against mechanism II.” ism. This seminal collection of essays that intro- Canadian Journal of Philosophy 9: 373–6. [Re- duces and develops the connectionist paradigm. sponse to Lucas 1961. Accessible to graduate Some of the chapters are accessible to advanced students and professionals.] undergraduates, but some are more appropriate Lucas, J. R. 1961. “Minds, machines and Gödel.” for graduate students and professionals.] Journal of Philosophy 36: 112–27. [Appeals to McLaughlin, B. P. 1993a. “The connectionism/ the results of Kurt Gödel to argue that we are classicism battle to win souls.” Philosophical Studies not computationally equivalent to Turing ma- 70: 45–72. [An examination of the debate between chines. Accessible to graduate students.] the symbolic paradigm and the connectionist para- Macdonald, C. and Macdonald, G., eds. 1995. digm. Accessible to advanced undergraduates.] Connectionism: Debates on Psychological Explana- ——. 1993b. “Systematicity, conceptual truth, and tion. Malden, MA: Blackwell. [A collection of evolution.” In C. Hookway and D. Peterson, articles on philosophical issues surrounding the eds., Royal Institute of Philosophy, Supplement 34:

149 Brian P. McLaughlin

217–34. [Examines some attempts to show that Pylyshyn, Z. 1986. Computation and Cognition. it is not the job of the computational theory of Cambridge, MA: MIT/Bradford. [A seminal dis- mind to explain the systematicity of thought. cussion of the symbolic paradigm, and a major Accessible to advanced undergraduates.] source for the discussion of that paradigm in this ——. 1997a. “Classical constituents in Smolensky’s chapter. Accessible to advanced undergraduates.] ICS architecture.” In M. L. dalla Chiara, K. Raggett, J. and Bains, W. 1992. Artificial Intelli- Doetes, D. Mundici, and J. van Bentham, eds., gence From A to Z. London: Chapman & Hall. The Tenth International Congress of Logic, Meth- [A wonderful source of very simple explanations odology and Philosophy of Science, Florence, August of technical terms in the field of artificial intelli- 1995, vol. 2, Structures in Science. Dordrecht: gence. Accessible to advanced undergraduates.] Kluwer, 331043. [Argues that Smolensky’s ICS Ramsey, W., Stich, S., and Rumelhart, D., eds. architecture can containmental representations 1991. Philosophy and Connectionism Theory. with classical constituents. Accessible to advanced Hillsdale, NJ: Erlbaum. [A collection of articles undergraduates.] on the connectionist paradigm. Accessible to ——. 1997b. “Connectionism.” In E. Craig, ed., advanced undergraduates.] The Routledge Encyclopedia of Philosophy. London: Rumelhart, D. E., McClelland, J. L., and the PDP Routledge. [An introduction to the connectionist Research Group, eds. 1986. Parallel Distributed paradigm. A source for the discussion of the Processing: Explorations in the Microstructure of connectionist paradigm in this chapter.] Cognition, vol. 1, Foundations. Cambridge, MA: —— and O’Leary-Hawthorne, J. 1994. “Dennett’s MIT/Bradford. [One of the two “bibles” of logical behaviorism.” Philosophical Topics 22: 189– connectionism. A classical introduction of the 258. [Presents ’s intentional sys- connectionist paradigm, and a major source for tems theory, and argues that, contra that theory, the discussion of that paradigm in this chapter. intentionality does not supervene on any pattern Some chapters are accessible to advanced under- of dispositions to outward or peripheral behavior. graduates, but some are most appropriate for Accessible to advanced undergraduates.] graduate students and professionals.] —— and Warfield, T. A. 1994. “The allure of Searle, J. 1999. “Chinese room argument.” In connectionism reexamined.” Synthese 101: 365– Wilson & Keil 1999: 115–16. [Searle’s most 400. [Contains a discussion of studies compar- recent presentation of his well-known Chinese ing connectionist networks and decision trees. room argument. Accessible to undergraduates.] Accessible to advanced undergraduates.] Sejnowski, J. and Rosenberg, C. R. 1987. “Parallel Minsky, M. A. and Pappert, S. 1969. Perceptrons. networks that learn to pronounce English text.” Cambridge, MA: MIT Press. [A critical examina- Complex Systems 1: 145–68. [Introduces NETtalk. tion of perceptrons, two-layered connectionist Accessible to graduate students.] networks. Accessible to graduate students.] Sethi, I. K. and Jain, A. K., eds. 1991. Artificial Penrose, R. 1989. The Emperor’s New Mind: Concern- Neural Networks and Statistical Pattern Recogni- ing Computers, Minds, and the Laws of Physics. tion. Amsterdam: Elsevier/North Holland. [A New York: Oxford University Press. [Includes collection of essays including some comparative an appeal to results of Kurt Gödel to argue that studies of connectionist networks and decision our minds are not computationally equivalent to trees. Accessible to graduate students.] Turing machines. Accessible to graduate students.] Shavlik, J. W., Mooney, R. J., and Towell, G. G. Prince, A. and Smolensky, P. 1993. Optimality 1991. “Symbolic and neural learning algorithms: Theory: Constraint and Interaction in Generative an experimental comparison.” Machine Learning Grammar. Technical report TR-2, New Bruns- 6: 111–44. wick, NJ: Rutgers Center for Cognitive Sciences. Smolensky, P. 1991. “The constituent structure of [A seminal presentation of optimality theory. connectionist mental states: a reply to Fodor and Accessible to graduate students.] Pylyshyn.” In Horgan & Tienson 1991: 281– Putnam, H. 1975. “Minds and machines.” In his 308. [A reply to Fodor & Pylyshyn 1988. Access- Philosophical Papers, vol. 2, Mind, Language, and ible to advanced undergraduates.] Reality. Cambridge: Cambridge University Press. ——. 1994. “Computational models of mind.” In [A classical paper on machine state psychofunc- Guttenplan 1994: 176–84. [An introduction to tionalism. Accessible to advanced undergraduates.] the connectionist paradigm that includes a sketch

150 Computationalism, Connectionism

of Smolensky’s ICS architecture. Accessible to Utgoff, P. 1999. “Decision trees.” In Wilson & advanced undergraduates.] Kiel 1999: 223–5. [An introduction to decision ——. 1995. “Reply: constitutive structure and ex- trees. Accessible to advanced undergraduates.] planation in an integrated connectionist/symbolic Van Gelder, T. 1995. “What might cognition be, if cognitive architecture.” In Macdonald & not computation?” Journal of Philosophy 7: 345– Macdonald 1995: 223–90. 81. [A defense of the view that cognition should Sterelny, K. 1990. The Representational Theory of be understood in dynamical systems terms, rather Mind: An Introduction. Oxford: Blackwell. [An ex- than computationally. Accessible to advanced cellent introduction to the representational theory undergraduates.] of mind. Accessible to advanced undergraduates.] Wilson, R. A. and Keil, F. C., eds. 1999. The MIT Turing, A. 1950. “Computing machinery and in- Encyclopedia of the Cognitive Sciences. Cambridge, telligence.” Mind 59: 433–60. [A truly ground- MA: MIT/Bradford. [Contains many entries rel- breaking article that literally began the field of evant to topics discussed in this article. Entries artificial intelligence. Accessible to undergraduates.] are accessible to advanced undergraduates.]

151 Part IV Real and Virtual Worlds Ontology

Chapter 11 Ontology Barry Smith

needed for a complete description and explana- Philosophical Ontology tion of all the goings-on in the universe? Or: What classes of entities are needed to give an Ontology as a branch of philosophy is the science account of what makes true all truths? It of what is, of the kinds and structures of objects, should be exhaustive in the sense that all types properties, events, processes, and relations in of entities should be included in the classifica- every area of reality. “Ontology” is often used tion, including also the types of relations by by philosophers as a synonym of “metaphysics” which entities are tied together to form larger (a label meaning literally: “what comes after the wholes. Physics”), a term used by early students of Aris- Different schools of philosophy offer different totle to refer to what Aristotle himself called approaches to the provision of such classifica- “first philosophy.” Sometimes “ontology” is used tions. One large division is that between what in a broader sense, to refer to the study of what we might call substantialists and fluxists, which might exist; “metaphysics” is then used for the is to say between those who conceive ontology study of which of the various alternative possible as a substance- or thing- (or continuant-) based ontologies is in fact true of reality (Ingarden discipline and those who favor an ontology 1964). The term “ontology” (or ontologia) was centered on events or processes (or occurrents). coined in 1613, independently, by two philo- Another large division is between what we might sophers, Rudolf Göckel (Goclenius) in his Lexicon call adequatists and reductionists. Adequatists philosophicum and Jacob Lorhard (Lorhardus) in seek a taxonomy of the entities in reality at all his Theatrum philosophicum. Its first occurrence levels of aggregation, from the microphysical to in English as recorded by the Oxford English the cosmological, and including also the middle Dictionary appears in Bailey’s dictionary of 1721, world (the mesocosmos) of human-scale entities which defines ontology as “an Account of being in between. Reductionists see reality in terms of in the Abstract.” some one privileged level of existents; they seek Ontology seeks to provide a definitive and to establish the “ultimate furniture of the uni- exhaustive classification of entities in all spheres verse” by decomposing reality into its simplest of being. The classification should be definitive constituents, or they seek to “reduce” in some in the sense that it can serve as an answer to other way the apparent variety of types of en- such questions as: What classes of entities are tities existing in reality.

155 Barry Smith

It is the work of adequatist philosophical Ontological Commitment ontologists such as Aristotle, Ingarden (1964), and Chisholm (1996) which will be of primary importance for us here. Their taxonomies are in To create effective representations it is an advant- many ways comparable to the taxonomies pro- age if one knows something about the things duced by sciences such as biology or chemistry, and processes one is trying to represent. (We though they are of course radically more general might call this the Ontologist’s Credo.) The at- than these. Adequatists transcend the dichotomy tempt to satisfy this credo has led philosophers between substantialism and fluxism, since they to be maximally opportunistic in the sources they accept categories of both continuants and have drawn upon in their ontological explorations occurrents. They study the totality of those of reality and in their ontological theorizing. objects, properties, processes, and relations that These have ranged all the way from the prepara- make up the world on different levels of focus tion of commentaries on ancient texts to reflec- and granularity, and whose different parts and tion on our linguistic usages when talking about moments are studied by the different scientific entities in domains of different types. Increasingly, disciplines. Ontology, for the adequatist, is then however, philosophers have turned to science, a descriptive enterprise. It is thus distinguished embracing the assumption that one (perhaps the from the special sciences not only in its radical only) generally reliable way to find out some- generality but also in its goal or focus: it seeks thing about the things and processes within a not predication and explanation but rather given domain is to see what scientists say. Some taxonomy and description. philosophers have indeed thought that the way to do ontology is exclusively through the invest- igation of scientific theories. With the work of Quine (1953) there arose in Methods of Ontology this connection a new conception of the proper method of ontology, according to which the The methods of ontology – henceforth in philo- ontologist’s task is to establish what kinds of sophical contexts always used in the adequatist entities scientists are committed to in their sense – are the methods of philosophy in general. theorizing. The ontologist studies the world by They include the development of theories of drawing conclusions from the theories of the wider or narrower scope and the testing and natural sciences, which Quine takes to be our refinement of such theories by measuring them best sources of knowledge as to what the world up, either against difficult counter-examples or is like. Such theories are extensions of the theor- against the results of science. These methods ies we develop and use informally in everyday were familiar already to Aristotle himself. life, but they are developed with closer attention In the course of the twentieth century a range to those special kinds of evidence that confer a of new formal tools became available to ontolo- higher degree of probability on the claims made. gists for the development and testing of their Quine – or at least the Quine of 1953 (I am theories. Ontologists nowadays have a choice here leaving aside Quine’s views on such matters of formal frameworks (deriving from algebra, as ontological relativity and the indeterminacy category theory, mereology, set theory, topo- of translation) – still takes ontology seriously. logy) in terms of which their theories can be His aim is to use science for ontological pur- formulated. These new formal tools, along with poses, which means: to find the ontology in the language of formal logic, can in principle scientific theories. Ontology is then a network allow philosophers to express intuitive ideas of claims, derived from the natural sciences, about and definitions in clear and rigorous fashion, what exists, coupled with the attempt to establish and, through the application of the methods what types of entities are most basic. Each nat- of formal semantics, they can allow also for the ural science has, Quine holds, its own preferred testing of theories for logical consistency and repertoire of types of objects to the existence completeness. of which it is committed. Each such science

156 Ontology embodies only a partial ontology. This is defined reality. Quine himself would resist this latter by the vocabulary of the corresponding theory conclusion. For him the best we can achieve in and (most importantly for Quine) by its canon- ontology lies in the quantified statements of par- ical formalization in the language of first-order ticular theories, theories supported by the best logic. Note that ontology is for Quine himself evidence we can muster. We have no way to rise not the metalevel study of the ontological com- above the particular theories we have; no way to mitments or presuppositions embodied in the harmonize and unify their respective claims. different natural-scientific theories. Ontology is rather these commitments themselves. Quine moves to the metalevel, making a semantic as- Internal vs. External Metaphysics cent to consider the statements in a theory, only in setting out to establish those expressions which definitively carry its commitments. Quine fixes Quine is a realist philosopher. He believes in a upon the language of first-order logic as the world beyond language and beliefs, a world which medium of canonical representation not out of the theories of natural science give us the power dogmatic devotion to this particular form, but to illuminate. There is, however, another tend- rather because he holds that this is the only ency in twentieth-century , a really clear form of language. First-order logic is tendency often associated with Quine but in- itself just a regimentation of corresponding parts spired much rather by Kant and promulgated by of ordinary language, a regimentation from thinkers such as Carnap and Putnam. According which, in Quine’s eyes, logically problematic to these thinkers ontology is a metalevel dis- features have been excised. It is then, Quine cipline which concerns itself not with the world argues, only the bound variables of a theory that itself but rather only with theories or languages carry its definitive commitment to existence. It or systems of beliefs. Ontology as a first-level is sentences like “There are horses,” “There are science of reality – ontology as what these numbers,” “There are electrons,” that do this philosophers call “external metaphysics” – is job. His so-called “criterion of ontological com- impossible. The best we can achieve, they hold, mitment” is captured in the slogan: To be is to be is internal metaphysics, which means precisely the the value of a bound variable. This should not study of the ontological commitments of spe- be understood as signifying some reductivistic cific theories or systems of beliefs. Strawsonian conception of existence itself as a merely logico- descriptive metaphysics is one example of such linguistic matter. Rather it is to be interpreted in internal metaphysics. Model-theoretic semantics, practical terms: to determine what the ontological too, is often implicitly understood in internal- commitments of a scientific theory are, it is neces- metaphysical terms – the idea being that we can- sary to determine the values of the quantified not understand what a given language or theory variables used in its canonical formalization. is really about, but we can build models with Quine’s approach is thus most properly more or less nice properties. What we can never conceived not as a reduction of ontology to the do is compare these models to some reality bey- study of scientific language, but rather as a ond. Ontology in the traditional philosophical continuation of ontology in the traditional sense. sense thus comes to be replaced by the study of When viewed in this light, however, it can be how a given language or science conceptualizes seen to be in need of vital supplementation. For a given domain. It becomes a theory of the the objects of scientific theories are discipline- ontological content of certain representations. specific. This means that the relations between Traditional ontologists are seeking principles that objects belonging to different disciplinary do- are true of reality. The practitioners of internal mains fall out of bounds for Quinean ontology. metaphysics, in contrast, are seeking to elicit prin- Only something like a philosophical theory of how ciples from subjects or theories. The elicited prin- different scientific theories (or their objects) ciples may or may not be true, but this, to the relate to each other can fulfill the task of provid- practitioner of internal metaphysics, is of no con- ing an inventory of all the types of entities in cern, since the significance of these principles

157 Barry Smith lies elsewhere – for instance in yielding a cor- Ontology and Information Science rect account of the taxonomical system used by speakers of a given language or by the scientists working in a given discipline. In a related development, also hardly noticed by In a development that has hardly been noted philosophers, the term “ontology” has gained by philosophers, a conception of the job of the currency in recent years in the field of computer ontologist close to that of Carnap and Putnam and information science (Welty & Smith 2001). has been advanced in recent years also in certain The big task for the new “ontology” derives extraphilosophical disciplines, as linguists, psy- from what we might call the Tower of Babel prob- chologists, and anthropologists have sought to lem. Different groups of data- and knowledge- elicit the ontological commitments (“ontologies,” base system designers have their own idiosyncratic in the plural) of different cultures and groups. terms and concepts by means of which they build Thus, they have sought to establish the ontology frameworks for information representation. Dif- underlying common-sense or folk theories of ferent databases may use identical labels but with various sorts by using the standard empirical different meanings; alternatively the same mean- methods of the cognitive sciences (see for ing may be expressed via different names. As example Keil 1979, Spelke 1990). Researchers ever more diverse groups are involved in sharing in psychology and anthropology have sought and translating ever more diverse varieties of to establish what individual human subjects, or information, the problems standing in the way entire human cultures, are committed to, ontolo- of putting this information together within a gically, in their everyday cognition, in much the single system increase geometrically. Methods same way in which Quine-inspired philosophers must be found to resolve the terminological and of science had attempted to elicit the ontolo- conceptual incompatibilities which then inevit- gical commitments of the natural sciences. ably arise. It was still reasonable for Quine to identify Initially, such incompatibilities were resolved the study of ontology – the search for answers on a case-by-case basis. Gradually, however, it to the question: what exists? – with the study of was recognized that the provision, once and for the ontological commitments of natural scient- all, of a common reference ontology – a shared ists. This is because it is a reasonable hypothesis taxonomy of entities – might provide significant that all natural sciences are in large degree con- advantages over such case-by-case resolution, and sistent with each other. Moreover, the identifica- the term “ontology” came to be used by informa- tion of ontology with ontological commitments tion scientists to describe the construction of a continues to seem reasonable when one takes canonical description of this sort. An ontology into account not only the natural sciences but is in this context a dictionary of terms formu- also certain commonly shared commitments of lated in a canonical syntax and with commonly common sense – for example that tables and accepted definitions designed to yield a lexical chairs and people exist. For the common-sense or taxonomical framework for knowledge rep- taxonomies of objects such as these are compat- resentation which can be shared by different ible with those of scientific theory if only we are information-systems communities. More ambi- careful to take into account the different granu- tiously, an ontology is a formal theory within larities at which each operates (Forguson 1989, which not only definitions but also a supporting Omnès 1999, Smith & Brogaard 2001). framework of axioms is included (perhaps the Crucially, however, the running together of axioms themselves provide implicit definitions of ontology and ontological commitments becomes the terms involved). strikingly less defensible when the ontological The methods used in the construction of commitments of various specialist groups of ontologies thus conceived are derived on the nonscientists are allowed into the mix. How, onto- one hand from earlier initiatives in database logically, are we to treat the commitments of management systems. But they also include astrologists, or clairvoyants, or believers in lepre- methods similar to those employed in philosophy chauns? We shall return to this question below. (as described in Hayes 1985), including the

158 Ontology methods used by logicians when developing measure, quantity, functional dependence, pro- formal semantic theories. cess, event, attribute, boundary, and so on. (See for example .) The top-level ontology would then be designed to serve as common neutral backbone, which would be sup- Upper-level Ontologies plemented by the work of ontologists working in more specialized domains on, for example, The potential advantages of ontology thus con- ontologies of geography, or medicine, or ecology, ceived for the purposes of information manage- or law, or, still more specifically, ontologies of ment are obvious. Each group of data analysts built environments (Bittner 2001), or of surgical would need to perform the task of making its deeds (Rossi Mori et al. 1997). terms and concepts compatible with those of other such groups only once – by calibrating its results in the terms of the single canonical back- Uses of Ontology bone language. If all databases were calibrated in terms of just one common ontology (a single consistent, stable, and highly expressive set of The initial project of building one single onto- category labels), then the prospect would arise logy, even one single top-level ontology, which of leveraging the thousands of person-years of would be at the same time nontrivial and also effort that have been invested in creating separate readily adopted by a broad population of differ- database resources in such a way as to create, in ent information-systems communities, has how- more or less automatic fashion, a single integrated ever largely been abandoned. The reasons for knowledge base of a scale hitherto unimagined, this can be summarized as follows. The task of thus fulfilling an ancient philosophical dream of ontology-building proved much more difficult a Great Encyclopedia comprehending all know- than had initially been anticipated (the difficult- ledge within a single system. ies being at least in part identical to those with The obstacles standing in the way of the con- which philosophical ontologists have grappled struction of a single shared ontology in the sense for some 2000 years). The information-systems described are unfortunately prodigious. Consider world itself, on the other hand, is very often the task of establishing a common ontology of subject to the short time horizons of the com- world history. This would require a neutral and mercial environment. This means that the require- common framework for all descriptions of his- ments placed on information systems change torical facts, which would require in turn that all at a rapid rate, so that already for this reason legal and political systems, rights, beliefs, powers, work on the construction of corresponding and so forth, be comprehended within a single, ontological translation modules has been unable perspicuous list of categories. to keep pace. Added to this are the difficulties which arise Yet work in ontology in the information- at the point of adoption. To be widely accepted systems world continues to flourish, and the prin- an ontology must be neutral as between different cipal reason for this lies in the fact that its focus data communities, and there is, as experience on classification (on analysis of object types) and has shown, a formidable trade-off between this on constraints on allowable taxonomies has constraint of neutrality and the requirement that proved useful in ways not foreseen by its initial an ontology be maximally wide-ranging and progenitors. The attempt to develop termino- expressively powerful – that it should contain logical standards, which means the provision of canonical definitions for the largest possible explicit specifications of the meanings of the terms number of terms. One solution to this trade-off used in application domains such as medicine or problem is the idea of a top-level ontology, which air-traffic control, loses nothing of its urgency would confine itself to the specification of such even when it is known in advance that the more highly general (domain-independent) categories ambitious goal of a common universal ontology as: time, space, inherence, instantiation, identity, is unlikely to be realized.

159 Barry Smith

Consider the following example, due to representation technologies necessary to create Guarino. Financial statements may be prepared decision automation systems – “white collar under either the US GAAP or the IASC stand- robots” – for large-scale industrial enterprises in ards (the latter being applied in Europe and many fields such as aerospace and defense. Realizing other countries). Cost items are often allocated that the ontology required to build such systems to different revenue and expenditure categories would need to embrace in a principled way the under the two standards, depending on the tax entire gamut of entities encompassed by these laws and accounting rules of the countries businesses in a single, unified framework, Ontek involved. So far it has not been possible to de- approached this problem by systematically ex- velop an algorithm for the automatic conversion ploiting the resources of ontology in the tradi- of income statements and balance sheets between tional (adequatist) philosophical sense. A team the two systems, since so much depends on highly of philosophers (including David W. Smith and volatile case law and on the subjective inter- Peter Simons) collaborated with software engin- pretation of accountants. Not even this relatively eers in constructing the system PACIS (for simple problem has been satisfactorily resolved, Platform for the Automated Construction of though this is prima facie precisely the sort of Intelligent Systems), which is designed to topic where ontology could contribute some- implement a comprehensive theory of entities, thing of great commercial impact. ranging from the very concrete (aircraft, their structures, and the processes involved in design- ing and developing them), to the somewhat abstract (business processes and organizations, If Ontek did not Exist, it would their structures, and the strategies involved in be Necessary to Invent It creating them), to the exceedingly abstract for- mal structures which bring all of these diverse Perhaps the most impressive attempt to develop components together. an ontology – at least in terms of sheer size – is Ontek has thus realized in large degree the the CYC project (http://www.cyc.com), which project sketched by Hayes in his “Naïve Physics grew out of an effort initiated by Doug Lenat Manifesto,” of building a massive formal theory in the early 1980s to formalize common-sense of (in Hayes’s case) common-sense physical real- knowledge in the form of a massive database of ity (in Ontek’s case this is extended to include axioms covering all things, from governments to not only airplane wings and factories but also mothers. The resultant ontology has been criti- associated planning and accounting procedures). cized for what seems to be its lack of principle As Hayes insisted, if long-term progress in artifi- in the ways in which new terms and theories cial intelligence is to be achieved it is necessary come to be added to the edifice of the theory. to put away the toy worlds of classical AI re- CYC takes the form of a tangled hierarchy, with search and to concentrate instead on the task of a topmost node labeled Thing, beneath which formalizing the ontological features of the world are a series of cross-cutting total partitions, itself, as this is encountered by adults engaged including: Represented Thing vs. Internal Machine in the serious business of living. Thing, Individual Object vs. Collection, Intan- The Leipzig project in medical ontology (see gible vs. Tangible Object vs. Composite Tangible ), too, is based on a realist and Intangible Object. Examples of Intangible methodology close to that of Ontek, and some- Objects (Intangible means: has no mass) are sets thing similar applies also to the work of Guarino and numbers. A person, in the CYC ontology, is and his colleagues in Italy. Most prominent a Composite Object made up of a Tangible body information-systems ontologists in recent years, and an Intangible mind. however, have abandoned the Ontologist’s Credo More important for our purposes here is the and have embraced instead a view of ontology as work of the firm Ontek – short for “ontological an inwardly directed discipline (so that they have technology” – which since 1981 has been devel- in a sense adopted an epistemologized reading oping database programming and knowledge of ontology analogous to that of Carnap and

160 Ontology

Putnam). They have come to hold that onto- firmly entrenched in many information-systems logy deals not with reality itself but rather with circles. Gruber is to be given credit for having “alternative possible worlds,” worlds which are crystallized the new sense of the term by relating indeed defined by the information systems them- it to the technical definition of “conceptualiza- selves. This means not only that only those en- tion” introduced by Genesereth and Nilsson in tities exist which are represented in the system their Logical Foundation of Artificial Intelligence (Gruber 1995), but also that the entities in ques- (1987). In his 1993 article Gruber defines an tion are allowed to possess only those properties ontology as “the specification of a conceptual- which the system itself can recognize. It is as if ization.” Genesereth and Nilsson conceive con- Hamlet, whose hair (we shall suppose) is not ceptualizations as extensional entities (they are mentioned in Shakespeare’s play, would be not defined in terms of sets of relations), and their merely neither bald nor nonbald, but would work has been criticized on the grounds that somehow have no properties at all as far as hair is this extensional understanding makes concep- concerned. (Compare Ingarden 1973 on the “loci tualizations too remote from natural language, of indeterminacy” within the stratum of rep- where intensional contexts predominate (see resented objects of a literary work.) What this Guarino, Introduction to 1998). For present means, however, is that the objects represented purposes, however, we can ignore these issues, in the system (for example, people in a database) since we shall gain a sufficiently precise under- are not real objects – the objects of flesh and standing of the nature of “ontology,” as Gruber blood we find all around us. Rather, they are conceives it, if we rely simply on the account denatured surrogates, possessing only a finite of conceptualizations which he himself gives in number of properties (sex, date of birth, social passages such as the following: security number, marital status, employment status, and the like), and being otherwise entirely A conceptualization is an abstract, simplified indeterminate with regard to those sorts of prop- view of the world that we wish to represent erties with which the system is not concerned. for some purpose. Every knowledge base, Information-systems ontologies in the sense knowledge-based system, or knowledge-level of Gruber are, we see, not oriented around the agent is committed to some conceptualization, world of objects at all. Rather, they are focused explicitly or implicitly. (Gruber 1995) on our concepts or languages or mental models (or, on a less charitable interpretation, the two The idea is as follows. As we engage with the realms of objects and concepts are simply con- world from day to day we participate in rituals fused). It is in this light that we are to interpret and we tell stories. We use information systems, passages such as the following: databases, specialized languages, and scientific instruments. We buy insurance, negotiate traffic, an ontology is a description (like a formal invest in bond derivatives, make supplications to specification of a program) of the concepts the gods of our ancestors. Each of these ways of and relationships that can exist for an agent or behaving involves, we can say, a certain concep- a community of agents. This definition is con- tualization. What this means is that it involves a sistent with the usage of ontology as set-of- system of concepts in terms of which the corres- concept-definitions, but more general. And it ponding universe of discourse is divided up into is certainly a different sense of the word than objects, processes, and relations in different sorts its use in philosophy. (Gruber, n.d.) of ways. Thus in a religious ritual setting we might use concepts such as salvation and purification; in a scientific setting we might use concepts such as virus and nitrous oxide; in a story-telling set- Conceptualizations ting we might use concepts such as leprechaun and dragon. Such conceptualizations are often The newly fashionable usage of “ontology” as tacit; that is, they are often not thematized in meaning just “conceptual model” is by now any systematic way. But tools can be developed

161 Barry Smith to specify and to clarify the concepts involved and rather like the attempt to find some highest to establish their logical structure, and in this common denominator that would be shared in way we are able to render explicit the underlying common by a plurality of true and false theories. taxonomy. We get very close to the use of the Attempts to construct such an ontology must term “ontology” in Gruber’s sense if we define fail if they are made on the basis of a methodo- an ontology as the result of such clarification – logy which treats all application domains on an as, precisely, the specification of a conceptualiza- equal footing. Instead, we must find ways to do tion in the intuitive sense described in the above. justice to the fact that the different conceptual- Ontology now concerns itself not with the izations which serve as inputs to ontology are question of ontological realism, that is with the likely to be not only of wildly differing quality question whether its conceptualizations are true but also mutually inconsistent. of some independently existing reality. Rather, it is a strictly pragmatic enterprise. It starts with conceptualizations, and goes from there to the What can Information Scientists description of corresponding domains of objects Learn from Philosophical (also called “concepts” or “classes”), the latter Ontologists? being conceived as nothing more than nodes in or elements of data models devised with specific practical purposes in mind. As we have seen, some ontological engineers have In very many cases the domains addressed by recognized that they can improve their models ontological engineers are themselves the pro- by drawing on the results of the philosophical ducts of administrative fiat. The neglect of truth work in ontology carried out over the last 2000 to independent reality as a measure of the cor- years. This does not in every case mean that rectness of an ontology is then of little import. they are ready to abandon their pragmatic per- In such cases the ontologist is called upon merely spective. Rather, they see it as useful to employ to achieve a certain degree of adequacy to the a wider repertoire of ontological theories and specifications laid down by the client, striving as frameworks and, like philosophers themselves, best he can to do justice to the fact that what they are willing to be maximally opportunistic the client says may fall short, for example, when in their selection of resources for purposes of measured in terms of logical coherence. Truth ontology-construction. Guarino and Welty (or the lack of truth) can be a problem also in (2000), for example, use standard philosophical non-administrative domains. Bad conceptualiza- analyses of notions such as identity, part, set- tions abound (rooted in error, myth-making, membership, and the like in order to expose hype, bad linguistics, or in the confusions of inconsistencies in standard upper-level ontologies ill-informed “experts” who are the targets of such as CYC, and they go on from there to knowledge-mining). Conceptualizations such as derive metalevel constraints which all ontologies these may deal only with created (pseudo-) must satisfy if they are to avoid inconsistencies domains, and not with any transcendent reality of the sorts exposed. beyond. They call for a quite different approach Given what was said above, it appears fur- than is required in those areas – above all in the ther that information ontologists may have areas addressed by the natural sciences – where sound pragmatic reasons to take the philo- the striving for truth to independent reality is a sopher ontologist’s traditional concern for truth paramount constraint. Yet this difference in ques- more seriously still. For the very abandonment tion has hardly been noted by those working on of the focus on mere conceptualizations and information-systems ontology – and this gives on conceptualization-generated object-surrogates us one clue as to why the project of a common may itself have positive pragmatic consequences. reference ontology applicable in domains of many This applies even in the world of admin- different types should thus far have failed. istrative systems, for example in relation to the Considered against this background the project GAAP/IASC integration problem referred to of developing a top-level ontology begins to seem above. For ontologists are here working in a

162 Ontology type of theoretical context where they must move an incremental process of theory construction, back and forth between distinct conceptualiza- criticism, testing, and amendment. As suggested tions, and where they can find the means to in Smith and Mark (2001), there may be reasons link the two together only by looking at their to look beyond natural science, above all where common objects of reference in the real, flesh- we are dealing with objects (such as societies, and-blood world of human agents and financial institutions, and concrete and abstract artifacts) transactions. existing at levels of granularity distinct from those Where ontology is directed in this fashion, which readily lend themselves to natural-scientific not towards a variety of more or less coherent inquiry. Our best candidates for good conceptu- surrogate models, but rather towards the real alizations will, however, remain those of the nat- world of flesh-and-blood objects in which we all ural sciences – so that we are, in a sense, brought live, then this itself reduces the likelihood of back to Quine, for whom the job of the ontolo- inconsistency and systematic error in the theor- gist coincides with the task of establishing the ies which result; and, conversely, it increases the ontological commitments of scientists, and of likelihood of our being able to build a single scientists alone. workable system of ontology that will be at the same time nontrivial. On the other hand, however, the ontological project thus conceived What Can Philosophers Learn will take much longer to complete and it will from Information-systems face considerable internal difficulties along the Ontologists? way. Traditional ontology is a difficult busi- ness. At the same time, however, it has the potential to reap considerable rewards – not least Developments in modal, temporal, and dynamic in terms of a greater stability and conceptual logics as also in linear, substructural, and paracon- coherence of the software artifacts constructed sistent logics have demonstrated the degree to on its basis. which advances in computer science can yield To put the point another way: it is precisely benefits in logic – benefits not only of a strictly because good conceptualizations are transparent technical nature, but also sometimes of wider to reality that they have a reasonable chance of philosophical significance. Something similar being integrated together in robust fashion into can be true, I suggest, in relation to the devel- a single unitary ontological system. If, however, opments in ontological engineering referred to we are to allow the real world to play a significant above. These developments can first of all help role in ensuring the unifiability of our separate to encourage existing tendencies in philosoph- ontologies, then this will imply that those who ical ontology (nowadays often grouped under the accept a conceptualization-based methodology heading “analytic metaphysics”) towards open- as a stepping stone towards the construction of ing up new domains of investigation, for example adequate ontologies must abandon the attitude the domain of social institutions (Mulligan 1987, of tolerance towards both good and bad con- Searle 1995, Smith 2002), of patterns (Johansson ceptualizations. It is this very tolerance which 1998), of artifacts (Dipert 1993, Simons & is fated to undermine the project of ontology Dement 1996), of boundaries (Smith 2001), of itself. dependence and instantiation (Mertz 1996), of Of course to zero-in on good conceptualiza- holes (Casati & Varzi 1994), and parts (Simons tions is no easy matter. There is no Geiger- 1987). Secondly, it can shed new light on the counter-like device which can be used for many existing contributions to ontology, from automatically detecting truth. Rather, we have Aristotle to Goclenius and beyond (Burkhardt to rely at any give stage on our best endeavors – & Smith 1991), whose significance was for a long which means concentrating above all on the work time neglected by philosophers in the shadow of of natural scientists – and proceed in careful, Kant and other enemies of metaphysics. Thirdly, critical, and fallibilistic fashion from there, hop- if philosophical ontology can properly be con- ing to move gradually closer to the truth via ceived as a kind of generalized chemistry, then

163 Barry Smith information systems can help to fill one important Bibliography gap in ontology as it has been practiced thus far, which lies in the absence of any analog of chem- Bittner, Thomas. 2001. “The qualitative structure ical experimentation. For one can, as C. S. Peirce of built environments.” Fundamenta Informaticae remarked (1933: 4.530), “make exact experi- 46: 97–126. [Uses the theory of fiat boundaries ments upon uniform diagrams.” The new tools to develop an ontology of urban environments.] of ontological engineering might help us to real- Brentano, Franz. 1981. The Theory of Categories. ize Peirce’s vision of a time when operations The Hague/Boston/London: Martinus Nijhoff. upon diagrams will “take the place of the experi- [Defends a classification of entities, and a new mereological view of substances and their accid- ments upon real things that one performs in ents, based on Aristotle.] chemical and physical research.” Burkhardt, Hans and Smith, Barry, eds. 1991. Finally, the lessons drawn from information- Handbook of Metaphysics and Ontology, 2 systems ontology can support the efforts of those vols. Munich/Philadelphia/Vienna: Philosophia. philosophers who have concerned themselves [Reference work on philosophical ontology and not only with the development of ontological ontologists.] theories, but also – in a field sometimes called Casati, Roberto and Varzi, Achille C. 1994. Holes “applied ontology” (Koepsell 1999, 2000) – with and Other Superficialities. Cambridge, MA: MIT the application of such theories in domains such Press. [On the ontology and cognition of holes.] as law, or commerce, or medicine. The tools of Chisholm, Roderick M. 1996. A Realistic Theory of philosophical ontology have been applied to solve Categories: An Essay on Ontology. Cambridge: practical problems, for example concerning the Cambridge University Press. [Defends a classifica- nature of intellectual property or concerning tion of entities based on Aristotle and Brentano.] the classification of the human fetus at different Dement, Charles W., Mairet, Charles E., Dewitt, stages of its development. Collaboration with Stephen E., and Slusser, Robert W. 2001. Mereos: information-systems ontologists can support such Structure Definition Management Automation ventures in a variety of ways, first of all because (MEREOS Program Final Technical Report), the results achieved in specific application do- available through the Defense Technical Informa- mains can provide stimulation for philosophers, tion Center (http://www.dtic.mil/), Q3 2001. but also – and not least importantly – because [Example of Ontek work in public domain.] information-systems ontology is itself an enorm- Dipert, Randall R. 1993. Artefacts, Art Works and ous new field of practical application that is Agency. Philadelphia: Temple University Press. [Defends a view of artifacts as products of crying out to be explored by the methods of human intentions.] rigorous philosophy. Forguson, Lynd. 1989. Common Sense. London/ New York: Routledge. [Survey of recent work Acknowledgments on common sense as a universal of human devel- opment and on the relevance of this work for the philosophy of common-sense realism.] This chapter is based upon work supported by Genesereth, Michael R. and Nilsson, L. 1987. the National Science Foundation under Grant Logical Foundation of Artificial Intelligence. Los No. BCS-9975557 (“Ontology and Geographic Altos, CA: Morgan Kaufmann. [On logic in Categories”) and by the Alexander von AI. Contains definition of “conceptualization” Humboldt Foundation under the auspices of its in extensionalist terms.] Wolfgang Paul Program. Thanks go in addition Gruber, T. R. 1993. “A translation approach to to Thomas Bittner, Charles Dement, Andrew portable ontology specifications.” Knowledge Frank, Angelika Franzke, Wolfgang Grassl, Nicola Acquisition 5: 199–220. [Account of the lan- Guarino, Kathleen Hornsby, Ingvar Johansson, guage ontolingua as an attempt to solve the Kevin Mulligan, David W. Smith, William portability problem for ontologies.] Rapaport, Chris Welty, and Graham White for ——. 1995. “Toward principles for the design of helpful comments. They are not responsible for ontologies used for knowledge sharing.” Inter- any errors which remain. national Journal of Human and Computer Studies

164 Ontology

43(5/6): 907–28. [An outline of motivations for Johansson, Ingvar. 1989. Ontological Investigations: the development of ontologies.] An Inquiry into the Categories of Nature, Man ——. [n.d.] “What is an Ontology?” [Wide-ranging study of realist ontology.] [Summary statement of Gruber’s definition of ——. 1998. “Pattern as an ontological category.” ontology as a specification of a conceptualization.] In Guarino ed. 1998: 86–94. [Ontology of Guarino, Nicola. 1995. “Formal ontology, con- patterns.] ceptual analysis and knowledge representation.” Keil, Frank. 1979. Semantic and Conceptual Devel- International Journal of Human-Computer opment: An Ontological Perspective. Cambridge, Studies 43: 625–40. [Arguments for the sys- MA: Harvard University Press. [Study of cog- tematic introduction of formal ontological nitive development of category knowledge in principles in the current practice of knowledge children.] engineering.] Koepsell, David R., ed. 1999. Proceedings of the ——, ed. 1998. Formal Ontology in Information Buffalo Symposium on Applied Ontology in the Systems. Amsterdam: IOS Press (Frontiers in Social Sciences (The American Journal of Eco- Artificial Intelligence and Applications). [Influ- nomics and Sociology, 58[2]). [Includes studies of ential collection.] geographic ontology, the ontology of economic ——. 1999. “The role of identity conditions in objects, the ontology of commercial brands, the ontology design.” In C. Freksa and D. M. Mark, ontology of real estate, and the ontology of eds., Spatial Information Theory: Cognitive and television.]***** Computational Foundations of Geographic Inform- ——. 2000. The Ontology of Cyberspace: Law, Philo- ation Science. Berlin/New York: Springer Verlag, sophy, and the Future of Intellectual Property. pp. 221–34. [On constraints on ontologies.] Chicago: Open Court. [A contribution to ap- —— and Welty, C. 2000. “A formal ontology of plied legal ontology with special reference to the properties.” In R. Dieng and O. Corby, eds., opposition between patent and copyright.] Knowledge Engineering and Knowledge Manage- Mertz, D. W. 1996. Moderate Realism and Its Logic. ment: Methods, Models and Tools. 12th Interna- New Haven, CT: Yale University Press. [Study tional Conference (EKAW 2000). Berlin/New of the logic and ontology of instantiation.] York: Springer Verlag, pp. 97–112. [A common Mulligan, Kevin, 1987. “Promisings and other problem of ontologies is that their taxonomic social acts: their constituents and structure.” In structure is poor and confusing as a result of the Kevin Mulligan, ed., Speech Act and Sachverhalt. unrestrained use of subsumption to accomplish a Reinach and the Foundations of Realist Phenom- variety of tasks. The paper provides a framework enology. Dordrecht/Boston/Lancaster: D. Reidel, for solving this problem.] pp. 29–90. [On the ontology of speech acts as Hayes, Patrick J. 1985. “The second naive physics a foundation for a general ontology of social manifesto.” In J. R. Hobbs and R. C. Moore. institutions.] eds., Formal Theories of the Common-sense World. Omnès, Roland. 1999. Understanding Quantum Norwood, NJ: Ablex, pp. 1–36. [Against toy Mechanics. Princeton: Princeton University Press. worlds methods in AI, in favor of massive axiom- [Introduction to the consistent histories inter- atization of common-sense physics along lines pretation of quantum mechanics.] similar to much current work in ontology.] Peirce, C. S. 1933. Collected Papers. Cambridge, Ingarden, Roman. 1964. Time and Modes of Being, MA: Harvard University Press. tr. Helen R. Michejda. Springfield, IL: Charles Quine, W. V. O. 1953. “On what there is,” repr. Thomas. [Translated extracts from a masterly in his From a Logical Point of View. New York: four-volume work in realist ontology entitled The Harper & Row. [Defends a view of ontology as Problem of the Existence of the World (the original the study of the ontological commitments of nat- published in Polish and in German).] ural science.] ——. 1973. The Literary Work of Art: An Investiga- Rossi Mori, A., Gangemi, A., Steve, G., Consorti, tion on the Borderlines of Ontology, Logic, and F., and Galeazzi, E. 1997. “An ontological ana- Theory of Literature. Evanston, IL: Northwestern lysis of surgical deeds.” In C. Garbay et al., eds., University Press. [Ontology applied to cultural Proceedings of Artificial Intelligence in Europe objects.] (AIME ’97). Berlin: Springer Verlag.

165 Barry Smith

Searle, John R. 1995. The Construction of Social —— and Brogaard, Berit. 2002. “Quantum mereo- Reality. New York: Free Press. [An ontology topology.” Annals of Mathematics and Artificial of social reality as the product of collective Intelligence 7: 591–612. [A theory of the relations intentionality.] between partitions at different levels of granularity.] Simons, Peter M. 1987. Parts: An Essay in Ontology. —— and Mark, David M. 2001. “Geographical Oxford: Clarendon Press. [Logical and philo- categories: an ontological investigation.” Inter- sophical study of mereology.] national Journal of Geographic Information Science —— and Dement, Charles W. 1996. “Aspects of 15(7). [Study of naive subjects’ conceptualizations the mereology of artifacts.” In Roberto Poli and of the geographical realm.] Peter Simons, eds., Formal Ontology. Dordrecht: Spelke, Elizabeth S. 1990. “Principles of object Kluwer, pp. 255–76. perception.” Cognitive Science 14: 29–56. [On Smith, Barry. 2001. “Fiat objects.” Topoi 20(2). the role of objects (as the cohesive bounded foci [On fiat and bona fide boundaries, with illustra- of local action) in human cognitive development.] tions in geography and other domains.] Welty, C. and Smith, B., eds. 2001. Formal Ontology ——. 2002. “John Searle: from speech acts to and Information Systems. New York: ACM Press. social reality.” In B. Smith, ed., John Searle. Cam- [Articles survey current work in information bridge: Cambridge University Press. systems ontology.]

166 Virtual Reality

Chapter 12 Virtual Reality Derek Stanovsky

sort of reality is virtual reality? Does the advent Introduction of virtual reality mark an extension, revision, expansion, or addition to reality? That is, is virtual “Virtual reality” (or VR) is a strangely oxy- reality real? Or is virtual reality more virtual than moronic term. “Virtual,” with its sense of “not real and, thus, not a significant new metaphys- actual” is jarringly juxtaposed with “reality” and ical problem itself? How else might the links its opposing sense of “actual.” Undoubtedly the between “reality” and “virtuality” be understood term has gained such currency at least partly and negotiated? Perhaps even more importantly, because of this intriguing provocation. “Virtual do the possible metaphysical challenges presented reality” is currently used to describe an increas- by virtual reality necessitate any changes in exist- ingly wide array of computer-generated or medi- ing metaphysical views, or shed any light on other ated environments, experiences and activities metaphysical problems? ranging from the near ubiquity of video games, to This chapter approaches some of these ques- emerging technologies such as tele-immersion, tions, focusing on three main issues within the to technologies still only dreamed of in science tremendously open field of inquiry laying at the fiction and only encountered in the novels of intersection of metaphysics with virtual reality. William Gibson or Orson Scott Card, on the First, the technology of virtual reality, along with Holodeck of television’s Star Trek, or at the some of the issues arising from this technology, movies in The Matrix of the Wachowski brothers, will be situated and examined within the Western where existing VR technologies make possible a philosophical tradition of metaphysics stretching narrative about imagined VR technologies. The from ancient to modern and postmodern times. term “virtual reality” covers all of this vast, and Next, the issues raised by virtual reality for per- still rapidly expanding, terrain. sonal identity and the subject will be explored and “Metaphysics” too is an expansive term (see examined, beginning with Cartesian for example Chapter 11, ONTOLOGY, and Chapter and moving through poststructuralist theories of 13, THE PHYSICS OF INFORMATION). Setting itself the subject and their various implications for vir- the enormous task of investigating the funda- tual reality. Finally, these metaphysical considera- mental nature of being, metaphysics inquires into tions and speculations will be brought to bear what principles may underlie and structure all of on the current economic realities of globalization reality. Some questions about virtual reality from and the emerging information economy, which the perspective of metaphysics might be: What have become inextricably bound up with both

167 Derek Stanovsky the metaphysics and politics of virtual reality as Virtual reality emerged from an unlikely hybrid it exists today. of technologies developed for use by the military Since metaphysics itself is one of the broadest and aerospace industries, Hollywood, and the subjects, it seems odd to restrict the discussion computer industry, and was created within con- of virtual reality only to one of its narrower texts ranging from the cold war to science senses. Therefore, virtual reality too will be con- fiction’s cyberpunk subculture. The earliest forms strued as broadly as possible, and not confined of virtual reality were developed as flight simu- to any one particular technological implementa- lators used by the US military and NASA to train tion, either existing or imagined. However, the pilots. This technology led to the head-mounted insights concerning virtual reality gleaned in this displays and virtual cockpit environments used manner should also find application in many by today’s fighter pilots to control actual aircraft. of its narrower and more restricted domains as Another source of VR lies in the entertainment well. One final qualification: since metaphysics industry’s search for ever more realistic movie inquires into the fundamental structures of reality, experiences beginning with the early Cinerama, and since it is unclear at this stage how virtual stereo sound, and 3D movies, and leading to reality is to be located within reality, it might further innovations in the production of realistic be more appropriate if the present inquiry into images and audio. Add to this a whole host the metaphysics of virtual reality were described of developments in computer technology. For instead as an exercise in “virtual metaphysics.” instance, computer-aided design programs, such It may be that what virtual reality requires is as AutoCAD, made it possible to use computers not so much a place within the history of West- to render and manipulate three-dimensional rep- ern metaphysics as it does a metaphysics all of resentations of objects, and graphical computer its own. interfaces pioneered by Xerox and popularized by Apple and Microsoft have all but replaced text-based computer interfaces and transformed the way people interact with computers. All of Virtual Reality these trends and technologies conspired to create the technology that has come to be known Virtual reality has been described in a variety of as “virtual reality” (for more on the genesis ways. In one of the earliest book-length treat- and genealogy of VR see Rheingold 1991 and ments of virtual reality, Howard Rheingold writes: Chesher 1994). “One way to see VR is as a magical window There is not, or at least not yet, any fixed set onto other worlds . . . Another way to see VR is of criteria clearly defining virtual reality. In his to recognize that in the closing decades of the book The Metaphysics of Virtual Reality, Michael twentieth century, reality is disappearing behind Heim identifies a series of “divergent concepts a screen” (Rheingold 1991: 19). This framing currently guiding VR research” each of which of virtual reality is a useful one for our purposes “have built camps that fervently disagree as to in that it helps to clarify and highlight one of what constitutes virtual reality” (Heim 1993: the central issues at stake. Does virtual reality 110). The cluster of features considered in this provide us with new ways to augment, enhance, section concern computer-generated simulations and experience reality, or does it undermine and which are interactive, which may be capable of threaten that reality? Virtual reality is equally being shared by multiple users, may provide fully prone to portrayals as either the bearer of bright realistic sensory immersion, and which may allow utopian possibilities or dark dystopian nightmares, for forms of telepresence enabling users to com- and both of these views have some basis to re- municate, act, and interact over great distances. commend them. Before exploring these issues Although not all of these elements exist in every further, it will be helpful to describe and explain version of virtual reality, taken together, these the origins of virtual reality, what virtual reality features have come to characterize virtual reality. is currently, and what it may become in the At one end of the spectrum, technologies future. allowing interactions with any representation or

168 Virtual Reality simulation generated by means of a computer video screens, or realized through more ambi- are capable of being described as virtual reality. tiously realistic and robustly immersive technolo- Thus, a video game simulation of Kung-Fu gies, there are other elements that may also play fighting, or the icons representing “documents” a part in virtual reality. Perhaps the most import- on a simulated computer “desktop” might both ant of these is provided by the capability of com- be cases where computers create a virtual reality puters to be networked so that multiple users with which people then interact in a variety of can share a virtual reality and experience and ways. What makes these candidates for virtual interact with its simulations simultaneously. The reality is not simply the fact that they are repres- possibility for virtual reality to be a shared experi- entations of reality. Paintings, photographs, tele- ence is one of the principal features by which vision, and film also represent reality. Computer virtual reality can be distinguished from fantasy. representations are different because people are One of the tests of reality is that it be available able to interact with them in ways that resemble intersubjectively. Thus, what is unreal about their interactions with the genuine articles. In fantasy is not necessarily that the imagined experi- short, people can make the computer simulations ences do not exist; it is that they do not exist do things. This is something that does not hap- for anyone else. Dreams are private experiences. pen with other forms of representation. This form On the contrary, the shared availability of virtual of virtual reality can already be provided by reality makes possible what William Gibson de- existing computer technologies and is becoming scribes so vividly in his early cyberpunk novels of increasingly commonplace. a computer-generated “consensual hallucination” At the other end of the spectrum lie tech- (Gibson 1984: 51). The ability to share virtual nologies aimed at fuller sensory immersion. reality sets the stage for a wide variety of human Head-mounted displays, datagloves, and other interactions to be transplanted into virtual reality, equipment translate body, eye, and hand move- and opens opportunities for whole new avenues ments into computer input and provide visual, of human activity. Communication, art, politics, audio, and even tactile feedback to the user. This romance, and even sex and violence are all hu- type of virtual reality aims at being able to pro- man activities that have found new homes in duce and reproduce every aspect of our sensory virtual reality. The possibility for the creation of world, with users interacting with virtual reality entirely new forms of human interactions and in many of the same ways they interact with practices that have no analog or precedence out- reality, e.g. through looking, talking, listening, side of virtual reality always remains open. touching, moving, etc. (even tasting and smell- Another feature that may be encountered in ing may find homes in virtual reality one day). virtual reality is that of “telepresence” or presence Virtual reality in this vein aims at creating simu- at distance, now frequently shortened simply to lations that are not only perceptually real in how “presence.” E-mail, video conferencing, distance they look and sound, but also haptically and education, and even telephones, all enable types proprioceptively real in how they feel to users as of telepresence. In each of these cases, the tech- well. As Randal Walser, a developer of virtual- nology allows users to communicate with distant reality systems, has written: “Print and radio tell; people as if they were in the physical presence of stage and screen show,” while virtual reality each other. Such communication is so common- “embodies” (quoted in Rheingold 1991: 192). At place in so much of the world today, it hardly the imagined limit of such systems lie the virtual- seems strange anymore that it is possible to com- reality machines of science fiction, with Star Trek’s municate with people who are thousands of miles Holodeck and the computer-generated world of away. More sophisticated, realistic, and immersive The Matrix producing virtual realities that are technologies both exist, and can be imagined, perceptually and experientially indistinguishable that allow not only for written or spoken com- from reality. No such technology exists today, munication over great distances, but also for but some elements of it are already possible. other types of interactions as well. For instance, In addition to the virtual reality of interactive the military use of remotely controlled aircraft simulations, whether confined to two-dimensional and missiles, or the use of unmanned spacecraft

169 Derek Stanovsky for exploration where humans might see, move, horizon within which any metaphysics of virtual control, and use instruments to explore far-flung reality must take place. destinations in the solar system are both examples which allow human presence virtually. Other ex- amples can be found in medicine, where surger- Virtual Metaphysics ies are now performed via computer-controlled instruments, and surgeons interface with a video screen rather than a patient. These examples It is possible to recapitulate a large portion of the illustrate ways in which human presence, action, history of Western metaphysics from the vantage- and interaction can be created virtually, and such point offered by virtual reality. The debates over examples are becoming more, rather than less, , , realism, , mater- common. ialism, nominalism, phenomenology, possible Virtual reality not only creates new virtual worlds, supervenience, space, and time, to name spaces to inhabit and explore, but creates the just a few, can all find new purchase, as well as possibility of virtual time as well. With the cre- some new twists, in this brave new world of ation of computer-generated simulations came computer-generated virtual reality. This section a bifurcation of time such that one now needs traces some of the most influential Western meta- to distinguish between time in the simulated, physical views concerning the distinction between virtual world and time in the rest of the world. appearance and reality and explores their pos- Thus, only with the advent of the artificially sible relevance to virtual reality. This discussion created worlds of virtual reality does the concept by no means exhausts the metaphysical possib- of “real time” (RT) enter into general parlance. ilities of virtual reality. In addition to the many Communications and interactions in virtual real- strands of Western (henceforth this qualification ity (as opposed to IRL, “in real life”) may be will be omitted) metaphysics left untouched, synchronous (as in video-conferencing and there remain vast areas of metaphysical thought chatrooms) and coincide closely with real time, that could also be fruitfully explored, including or asynchronous (as in e-mail exchanges) and long and rich traditions of African, Chinese, diverge widely and unpredictably from the pas- Indian, and Latin American metaphysics. sage of time in other virtual interactions or with Distinguishing between appearance and reality time outside the simulation. Time may even stop, is perhaps one of the most basic tasks of meta- or go backwards, within virtual reality. For in- physics, and one of the oldest, dating back at stance, a simulation might be paused indefinitely, least to Thales and his pronouncement that de- or reset to some previous state to allow users to spite the dizzying variety in how things appear, experience a part of a simulation again. Time in reality “All is water.” This desire to penetrate may also vary simply as a result of the techno- behind the appearances and arrive at the things logy used. This might happen when faster ma- themselves is one of the most persistent threads chines are networked with computers operating in metaphysics. Virtual reality presses at the very at lower MHz, or utilizing slower modems. In limits of the metaphysical imagination and fur- such cases, this can mean that some objects are ther tangles and troubles long standing prob- rendered faster and changed and updated more lems concerning how things seem versus how frequently than others, giving an oddly disjointed they really are. For instance, puzzles concerning sense of time, as objects in the same simulation mirrors and dreams and the ways in which they move at distinctly different rates of time. These can confound our understanding of reality have variations and complications in time emerge a long history and haunt the writings of many alongside and with virtual reality. metaphysicians. Virtual reality complicates these Not all of these elements exist in every ver- puzzles still further. sion of virtual reality. However, taken together, “But suppose the reflections on the mirror they provide the background against which cur- remaining and the mirror itself not seen, we rent virtual-reality systems are being invented and would never doubt the solid reality of all that reinvented. These same elements also trace the appears” (III. 6 [13]). This passage from Plotinus

170 Virtual Reality comes wonderfully close to describing the cur- reality, the image need not be limited to sight. rent possibilities of virtual reality. Virtual reality In virtual reality, the representation may pass may be very like the images in a mirror persist- scrutiny from any angle using any sense. As for ing even after the mirror disappears. In the case transience, while the images in virtual reality may of mirrors, such a possibility remains only hypo- disappear at any moment, they also may be just thetical. Plotinus assumes that in most cases the as permanent and long-lived as any real object difference between reality and the reflection of or event. Moreover, mirrors can only reflect the reality presented by a mirror is easy to discern. images of already existing things. Virtual reality After all, it is only Lewis Carroll’s Alice who peers has no such constraint. Objects in virtual reality into a looking glass and takes what she sees to may be copies of other things, but they also may be a room “just the same as our drawing-room, be their own unique, individual, authentic objects only the things go the other way” (Carroll 1871: existing nowhere else. This last point means that 141). Such a confusion seems amusingly child- the grounds for needing to distinguish image ish and naive. So confident is Plotinus in this from reality have changed. It is not simply that distinction between real objects and their unreal the representations of virtual reality are false (not mirror images that he uses it as an analogy in genuine) like the reflections in a mirror. It is not support of his claim that reality lies with form even analogous to Plato’s view of theater, which rather than matter. However, what is more strik- was to be banned from his Republic because of ing is that Plotinus allows that under certain its distortions and misrepresentations of reality. circumstances (if the image in the mirror endured, Instead, virtual reality may summon up a whole and if the mirror itself was not visible) these new reality, existing without reference to an reflections might fool us as well. Indeed, it is external reality, and requiring its own internal our inability to distinguish image from reality methods of distinguishing true from false, what that lends interest to such spectacles as fun is genuine or authentic from what is spurious houses, with their halls of mirrors, and the illu- or inauthentic. sions performed by magicians. In these cases, Dreams too can provide occasions where we do make the same mistake as Alice. It is this perception and reality become interestingly possibility of fundamentally conflating image, or entangled and may be one of the best, and most representation, with reality that lends mirrors familiar, comparisons for virtual reality. Dreams their metaphysical interest. possess many (although not all) of the elements Virtual reality may present us with a new sort of virtual reality. Dreams are immersive, match- of mirror; one with the potential to surpass even ing in sensory clarity and distinctness even the the finest optical mirrors. If so, then virtual real- most optimistic science fiction accounts of ity may fatally complicate the usual mechanisms virtual reality. In his Meditations, Descartes fam- used to distinguish image from reality, and ously entertains the possibility that there may representation from what is represented. For be no certain method for distinguishing dreams Plotinus, it is the limitations of the mirror image from reality. He writes: “How often, asleep at that reveals its status as a reflection of reality. It night, am I convinced of just such familiar events is only because images in a mirror are transient – that I am here in my dressing gown, sitting (fleeting, temporary, failing to persist over time by the fire – when in fact I am lying undressed or cohere with the rest of our perceptions) and in bed!” and finds such anecdotes sufficiently because the mirror itself does not remain invis- persuasive to conclude that “I see plainly that ible (its boundaries glimpsed, or reflecting sur- there are never any sure signs by means of which face flawed or otherwise directly perceptible) that being awake can be distinguished from being enables us to tell the difference between image asleep” (Descartes 1641: 77). Here, Descartes and reality. One of the inherent limitations of seems to suggest that dreams and reality can any mirror is that it is necessarily confined to actually be confused, unlike Plotinus, who optical representations. Reaching out to touch viewed the confusion of images in a mirror with an object in a mirror always reveals the decep- reality as only a hypothetical possibility at best. tion. However, in immersive versions of virtual Descartes, however, is unwilling to allow this

171 Derek Stanovsky much uncertainty into his philosophical system Other metaphysical systems plot more subtle and so appends the following curious solution and complex relationships between appearance to the dream problem in the last paragraph of and reality. Kantian metaphysics occupies a pivotal his last Meditation. “But when I distinctly see place in the history of metaphysics providing, as where things come from and where and when it does, a continuation of important strands of they come to me, and when I can connect my debate from antiquity, the culmination of sev- perceptions of them with the whole of the rest eral disputes within the modern period, and the of my life without a break, then I am quite cer- origin of many contemporary discussions in the tain that when I encounter these things I am field. Can the Kantian system help provide a not asleep but awake” (Descartes 1641: 122). more sophisticated description of the status of Along with clarity and distinctness, Descartes adds virtual reality? coherence as a final criterion for certainty, in an Kant’s revolves around effort to resolve the doubts raised by the dream the view that things in themselves are unknowable problem. This is despite the fact that one of the in principle and that human knowledge is only chief strengths of the dream problem, as he put of appearances. Just like Descartes, Kant holds it forward, lay in the fact that dreams often could that we are epistemically acquainted with only be fit coherently into waking life. our own perceptions. However, unlike Descartes, Virtual reality also can pass these tests of clar- for Kant perceptual objects are nothing other ity, distinctness, and coherence. Beyond this, VR, than these patterns of representation encountered unlike a dream, is able to satisfy the requirement by the mind. Thus, Kant believes it is possible to of intersubjective availability that only “real” overcome the epistemological problems intro- reality is generally assumed to possess. That is, duced by the division between appearance and whereas a dream can only be experienced by a reality. This is because, for Kant, the mind plays single person, virtual reality is available to any- an active, constitutive role in structuring reality. one. At this point, Descartes’s dream problem Chief among these contributions are the intui- takes on new life. Just as was true of the com- tions of space and time. Space and time are not parison with images in a mirror, the need to themselves “things” that are directly perceptible, distinguish virtual reality from nonvirtual reality and yet, it is impossible for human beings to seems to dissolve. If virtual reality is not “real,” experience objects outside of space and time. it must be on some basis other than those con- What this means, according to Kant, is that “Both sidered so far. Distinguishing dream from real- space and time . . . are to be found only in us” ity, for Descartes, just like distinguishing image (1781: A 373). In this way, Kant hopes to over- from reality for Plotinus, takes on importance come the epistemological divide between em- precisely because, without some reliable means piricism and rationalism by restricting knowledge of discrimination, such confusions run the risk to objects of experience, while at the same time of infecting an otherwise easily recognized real- granting an active role to the mind in structur- ity with instances of unreality. This would render ing that experience. reality a concept of dubious usefulness, for it Given a Kantian view, the objects encoun- could no longer clearly be distinguished from tered in virtual reality may not pose any signific- its opposite, from the unreal, from appearance, antly new metaphysical challenges. Since things from image, or from dream. Descartes and in themselves are never the direct objects of Plotinus both identify permanence and coher- human knowledge, the fact that experiences ence as criteria of the real and transience as the in virtual reality fail to correspond to objects mark of the merely apparent. However, such outside the mind in any simple, straightforward solutions work even less well in the case of way is not necessarily a problem. Every object of virtual reality. At this point the name “virtual human knowledge, whether actual or virtual, is reality” starts to become justified. Virtual reality nothing other than just such an organized col- takes on an existence with a distinctly different lection of perceptual representations. This means character from dreams, images, and other mere that virtual reality can be admitted to the world representations. of empirical human experience on more or less

172 Virtual Reality equal footing with the more usual forms of expense of the reality of all experience. It is not experience. Another way of stating this might just Descartes and Kant who find a need to be that, for Kant, all experience is essentially accord an increased status to ideas and percep- virtual. It is not epistemic contact with, or know- tion. Even in Heidegger’s existentialist meta- ledge of, things as they exist apart from the mind physics there is always not only the object, but that ever characterizes any human experience. also the encounter of the object; and these two What is known is only how those things appear moments remain distinct, and distinctly import- to the mind. Given this, the fact that virtual ant. This experiential aspect of virtual reality is reality exists for the mind (and can be made to something that invites a more phenomenolo- exist for more than one mind) is sufficient to gically oriented approach. It may be tempting qualify those experiences as “real.” One may still to see virtual reality as a vindication of Platonist need to exercise some care in using and applying metaphysics, where the world of ideas is brought the empirical knowledge gained by way of vir- to fruition and the less-than-perfect world of tual reality. Likewise, inferences based on that bodies and matter can be left behind. Others knowledge must be confined to their appropri- argue that rather than demonstrating the truth ate domain. However, this holds true for any of Platonic idealism, or marking the completion piece of empirical knowledge no matter how it of the Cartesian project of separating the mind is acquired. from the body, virtual reality instead illustrates Kantian metaphysics may also help explain why the inseparability of mind from body and the human interactions with computers have con- importance of embodiment for all forms of jured up these strange new frontiers of virtual human experience and knowledge. After all, even space and virtual time. If it is true, as Kant con- in the noncorporeal world of virtual reality, jectures, that the mind cannot experience things virtual bodies had to be imported, re-created, outside of space and time, then any new experi- and imposed in order to allow for human ences will also have to be fit within these schemas. interaction with this new virtual world. This Although the mind does not possess innate ideas tends to point to the necessity of embodiment or any other particular content, it does provide as a precondition for, rather than an impedi- a formal structure that makes possible any ex- ment to, experience and knowledge (see Heidt perience of the world. Presumably, this remains 1999). true of computer-generated worlds as well. Once There are many other possible approaches computer-mediated experiences become a tech- to the metaphysics of VR. For instance, Jean nical possibility, the mind also structures, organ- Baudrillard’s theories of simulation and hyper- izes, and interprets these experiences within the reality seem readymade for virtual reality, point- necessary framework. Thus, virtual reality may ing to a metaphysics where contemporary social be a predictable artifact of the mind’s ordering reality could be understood as having already of these computer-generated experiences. Virtual fallen prey to the order of simulation made space and virtual time may be the necessary increasingly available by virtual reality. From forms of apprehension of virtual reality, just as Baudrillard’s vantage-point, simulations, like space and time are necessary to the apprehension those of VR, mark the end of our ability to dis- of reality. In the case of virtual reality, the claim tinguish between appearance and reality, reduc- that space and time are “found only in us” seems ing everything to a depthless hyperreality (see much less contentious. Given these possibilities Baudrillard 1983). Another possibility would be and connections, virtual reality may turn out to follow Jacques Derrida’s critique of the meta- to provide a laboratory for the exploration of physics of presence onto the terrain of virtual Kantian metaphysics. reality where the absence of presence could be At this point one may wish to retreat to the marked in new, high-tech ways. However, rather relative safety of a more thoroughgoing materi- than pursuing additional examples, at this point alism, where what is real is only the circuits it is better to inquire into a different, although and wires that actually produce virtual reality. closely related, set of metaphysical problems con- However, the cost of such a move comes at the cerning the identity of the self.

173 Derek Stanovsky

the mirror stage as an identification, in the full Virtual Identity sense that analysis gives to the term: namely, the transformation that takes place in the subject In addition to raising questions about the nature when he assumes an image” (Lacan 1977: 2). and status of external reality, virtual reality also The subject is thus produced by an identifica- raises difficult questions concerning the nature tion with an image, an image that is not the of the subject, or self. Despite the differences in subject and yet which is mistaken to be identical the metaphysical views discussed up to this point, with it. If identity is based on identifications, there is one area of general agreement. Whether and identification is always an identification with Platonist, Cartesian, or Kantian in orientation, something one is not, then one’s identity will in all of these systems there is a shared notion of always be something that is at odds with itself. a unified, and unifying, subject whose existence Elsewhere, Lacan explicitly relies on an example provides a ground for knowledge, action, and of a trick done with mirrors to illustrate the personal identity. Such a conception of the sub- situation of the human subject. Here, the illu- ject has been complicated in recent years. In sion of a vase filled with flowers is produced. particular, poststructuralist accounts of a divided For Lacan, it is the illusion of the self that is and contingent subject have raised questions produced. (See figure 12.1.) about the adequacy of previous views. Virtual In the figure, the subject occupies the posi- reality also complicates assumptions concerning tion of the viewer (symbolized by a barred S to a unified subject. The example discussed above re-emphasize this division which founds the sub- of images in a mirror can be used again to ap- ject), and the ego is represented by the virtual proach these questions surrounding the subject, image of the inverted vase seen in the mirror. this time through the work of Jacques Lacan. Lacan is proposing that a mistake worse than Lacan’s influential formulation of the “mirror that made by Alice with the looking glass is not stage” pushes the notion of the knowing subject merely commonplace, but constitutive of human to its limits. Inverting traditional Cartesian epi- subjectivity. The self, emerging over time as the stemology, the subject, instead of being the first result of a series of identifications with others, and most surely known thing, becomes the first is, like the image of the vase in the mirror, not misrecognized and misknown thing. This is an actual but virtual. even more radical mistake than that made by Virtual reality compounds this dilemma. If in Alice in her trip through the looking glass. At reality the subject is already not where it thinks least when Alice looked in the mirror and saw itself to be, in virtual reality the situation becomes a girl very much like herself, she still took it to even worse. Virtual reality provides an open field be a different little girl and not herself. For for various and even multiple identities and iden- Descartes, this would amount to a mistake in tifications. In virtual environments, people are the one thing he thought he could be certain of, not confined to any one stable unifying subject the cogito. Given Lacan’s view, “I think, there- position, but can adopt multiple identities (either fore I am” becomes an occasion for error when serially or simultaneously). From the graphical pronounced while looking into a mirror. In this avatars adopted to represent users in virtual en- case, the I of thinking can differ from the I of vironments, to the handles used in chatrooms, existing (the I of consciousness thinks, therefore to something as simple as multiple e-mail ac- the I in the mirror exists). Lacan reworks the counts, all of these can be used to produce and slogan to read, “I think where I am not, there- maintain virtual identities. Identity in virtual fore I am where I do not think” (Lacan 1977: reality becomes even more malleable than in real 166). Such a formulation could never serve as life, and can be as genuine and constitutive of Descartes’s foundation for knowledge once this the self as the latter. Sexual and racial identities division is introduced within the subject. can be altered, edited, fabricated, or set aside This divide within the subject is precisely what entirely. Identities can be ongoing, or adopted is highlighted in Lacan’s discussion of the mirror only temporarily. Thus, virtual reality opens the stage. Lacan writes: “We have only to understand possibility not only of recreating space and time,

174 Virtual Reality

x′

Mirror $ S, I y ′ a a

i′ (a)

O C

x

y′

Figure 12.1: The illusion of a vase/the illusion of the self (Lacan 1978) Source: “Diagram on p. 145,” from The Four Fundamental Concepts of Psycho-analysis by Jacques Lacan, tr. Alan Sheridan. Copyright © 1973 by Editions du Seuil. English tr. copyright © 1977 by Alan Sheridan. Used by permission of W. W. Norton & Co. Inc. but the self as well. The subject is produced process of producing and reshaping vast areas anew as it comes to occupy this new space. In of our social reality. If virtual reality has yet to her influential book Life on the Screen, Sherry supplant more traditional modes of human in- Turkle argues that online identities make “the teraction with the physical world, with each other, Gallic abstractions” of French theorists like Lacan and even with oneself, there is one arena in “more concrete,” writing: “In my computer- which virtual reality has already made startling mediated worlds, the self is multiple, fluid, and and astonishingly swift inroads, and that is in constituted in interaction with machine connec- the realm of economics. From ATM machines tions; it is made and transformed by language” and electronic transfers to the dot-com boom (Turkle 1995: 15). For Turkle, the divisions and and bust, global capital has not been shy about fragmentations that mark every identity take on leaping into the virtual world of e-commerce. new prominence and find new uses in the virtual Why has global capital been able to find a home reality of online society. in this new virtual economic space with such ease and rapidity? What does this colonization of virtual reality portend for other noncommodity possibilities of virtual reality? Economic Reality Globalization is a process that has certainly been facilitated by the information economy of The metaphysics of virtual reality may strike some the digital age. Mark Poster has described this as the most esoteric of topics, far removed from situation as “Capitalism’s linguistic turn” as the everyday life and practical human concerns. How- industrial economy segued into the information ever, metaphysical views often have a surprising economy (Poster 2001: 39). Capital has been reach and can make their influence felt in unsus- instrumental in producing and disseminating the pected ways. In the case of virtual reality, these technologies that have made this process pos- metaphysical attachments are currently in the sible. The coining of the phrase “virtual reality”

175 Derek Stanovsky is most often attributed to Jaron Lanier, a devel- remains and the mirror disappears. The current oper and entrepreneur of virtual-reality systems, electronic transfer of funds around the globe to use as part of a marketing strategy for his comes close to realizing this goal (for a further software company. The potential of e-mail as an discussion of “digital gold” in the information advertising medium was pioneered early on when, age, see Floridi 1999: 113ff). It may be that in 1994, a pair of enterprising green-card attor- this spectral nature of money means that capital neys became the first to use e-mail as a form of is uniquely adapted for virtual reality. Money is direct marketing. Computer sales, driven by the already the virtual expression of value. expansion of the internet, fueled the expansion For capital, the additional “metaphysical sub- of the high-tech economy to such an extent that tleties” tacked on by virtual reality may scarcely the internet service provider America Online matter. The already virtual existence of money could afford to buy media giant Time Warner. has facilitated the migration of capital into vir- Virtual reality has created new commodities, tual reality with nothing lost in the transition. which have quickly become new economic real- The online virtual reality of the internet was once ities. Capital has also tended to transplant and home to a variety of small, but close-knit, virtual reproduce already existing social and economic communities. This has changed. Now the char- inequalities into this new virtual world. For in- acter and function of the internet more closely stance, there has been much discussion of the resembles a virtual shopping mall as advertise- “digital divide” between those with access to ments appear everywhere and the identity of con- global information networks and those without. sumer overtakes every other online identity. We This divide falls along the well-worn demarca- may currently be living through a process of tions of race and gender, but even more starkly, virtual primitive accumulation, or a kind of elec- along class lines. The divide between rich and tronic enclosure movement, as the free associ- poor, both within and between nations, has been ation and utopian possibilities offered by online mapped onto the very foundations of the informa- virtual reality are driven out by the commodi- tion age. These capitalist origins of virtual reality fication imposed by global capital. Capital, long should not be forgotten. a kind of universal solvent for social relations, is Capital organizes economic and social life currently transforming the virtual social relations around the production and consumption of com- of online life at a breathtaking pace. However, modities. Marx writes that the commodity form this process does not occur without active resist- raises a whole host of “metaphysical subtleties ance (see Chesher 1994, and Dyer-Witheford and theological niceties” (Marx 1867: 163). 1999). It is here that the urgency of these other- Relationships between commodities become wise abstract metaphysical speculations can be “dazzling” in their variety and movements, while felt. The metaphysics of virtual reality provides the social relationships between producers and the horizon on which a host of new ethical and consumers become obscured behind the appear- political questions will take shape and within ances of wages and prices (Marx 1867: 139). which they must be answered. For Marx, the value of a commodity only emerges virtually. The value of one commodity finds ex- pression only in the body of another commodity References through the relationship of exchange. Thus, the value of a watch might be expressed in its ex- Baudrillard, J. 1983. Simulations, tr. P. Foss, change for a cellphone. This system of exchange P. Patton, and P. Beitchman. New York: finds its culmination in money, a commodity Semiotext(e). whose function is to provide a mirror for the Carroll, L. 2000 [1871]. The Annotated Alice: value of every other commodity. The particular Alice’s Adventures in Wonderland and Through commodity serving as money changes over time, the Looking-glass. New York: W. W. Norton. from gold and silver to paper and plastic, as Original works published 1865 and 1871. money asymptotically approaches the perfect mir- Chesher, C. 1994. “Colonizing virtual reality: ror described by Plotinus, where only the image construction of the discourse of virtual reality,

176 Virtual Reality

1984–1992.” Cultronix 1(1) (Fall 1994). Kant, I. 1996 [1781]. Critique of Pure Reason: . Unified Edition, tr. W. S. Pluhar. Indianapolis: Descartes, R. 1988 [1641]. Descartes: Selected Philo- Hackett. Original works published 1781 and sophical Writings, tr. J. Cottingham, R. Stoothoff, 1787. and D. Murdoch. Cambridge: Cambridge Uni- Lacan, J. 1977. Écrits: A Selection, tr. A. Sheridan. versity Press. Meditations on First Philosophy ori- New York: W. W. Norton. ginally published 1641. ——. 1978. The Four Fundamental Concepts of Dyer-Witheford, N. 1999. Cyber-Marx: Cycles and Psycho-analysis, tr. A. Sheridan. New York: Circuits of Struggle in High-Technology Capital- W. W. Norton. ism. Urbana: University of Illinois Press. [This Marx, K. 1977 [1867]. Capital, Volume One, tr. book provides a thorough and detailed account B. Fowkes. New York: Vintage Books. Original of the vicissitudes of class struggle within the work published 1867. global capitalist information economy from an Plotinus. 1992 The Enneads, tr. S. MacKenna. autonomist Marxist perspective.] Burdett, NY: Paul Brunton Philosophic Floridi, L. 1999. Philosophy and Computing: An Foundation. Introduction. London: Routledge. [This textbook Poster, M. 2001. What’s the Matter with the provides a clear and accessible overview of philo- Internet? Minneapolis: University of Minnesota sophical problems relating to computers and Press. [Mark Poster’s book provides a sophistic- information theory ranging from the internet ated inquiry into the culture and politics of the to artificial intelligence.] internet and covers topics on critical theory, Gibson, W. 1984. Neuromancer. New York: Ace postmodernism, globalization, and democracy.] Books. Rheingold, H. 1991. Virtual Reality. New York: Heidt, S. 1999. “Floating, flying, falling: a philosoph- Summit Books. [This book is a lively, and access- ical investigation of virtual reality technology.” ible, journalistic account of the people and Inquiry: Critical Thinking Across the Disciplines history behind the early development of the 18(4): 77–98. technologies of virtual reality.] Heim, M. 1993. The Metaphysics of Virtual Reality. Turkle, S. 1995. Life on the Screen: Identity in the New York: Oxford University Press. [Michael Age of the Internet. New York: Touchstone. Heim’s book gives an accessible introduction [Sherry Turkle’s book is one of the new classics to some of the philosophical issues arising from of internet studies. Drawing on psychoanalysis virtual reality and explores the changes this tech- and French theory to explore online identities, nology may have introduced into reality.] Turkle examines the fragmented nature of the self in postmodern culture.]

177 Eric Steinhart

Chapter 13 The Physics of Information Eric Steinhart

theory with an apparently consistent notion of 1 Introduction limits and the theory of physical computations should be able to take advantage of that idea. This chapter has two goals. The first is to analyze Failure to consider infinities adequately has led physical concepts like space, time, and many thinkers to regard finite Turing comput- in informational and computational terms (the ability as some sort of necessary upper bound to informational/computational nature of physics). physical computation. This would be a mistake. The other is to explain some key informational It is possible that there are physical hypercom- and computational concepts in physical terms (the puters far more powerful than classical Turing physical nature of information/computation). machines (even if none exist in our universe). These two goals are philosophically interesting (4) Philosophical efforts to analyze the mind– for at least six reasons. (1) Philosophers have brain relation in terms of programs and comput- always been interested in the logical structure of ers (e.g. functionalism) seem to have introduced physical reality, even metaphorically. The images a kind of dualism between software and hard- of the physical universe as an arrangement of ware. The ontology of software is unnecessarily geometrical figures or a clock have been super- vague. “Virtual” software objects are often de- seded by the image of the universe as a computer. scribed as if they were nonphysical objects that Metaphors apart, we shall see that, strictly speak- nevertheless participate in spatio-temporal-causal ing, the universe is a computer if and only if its processes. Hence the “virtual” can become a physics is recursive. (2) Claims about the com- strange category, neither concrete nor abstract. putational powers of physical systems appear (5) Sometimes, philosophers make false claims in many philosophical arguments. Some argu- about computers. One hears about continuous ments in the philosophy of mind, for example, “analog” computers even though all the quant- depend on the computational powers of physical ities involved are known to be discrete. Or one systems like the human body and brain. (3) The is told that computers manipulate information role of the transfinite in the philosophical con- as if it were some kind of immaterial stuff (like ception of computation requires clarification. If pneuma or ether). Finally, (6) theoretical efforts an idealized Turing machine has infinitely many to understand physical reality in computational squares on its tape, then it ought to be able to terms are often confused with attempts to devise have infinitely many 1s or to make infinitely many simulations of physical systems. However, physi- moves. The calculus has long provided physical cists who wonder whether the universe is a

178 The Physics of Information computer are not concerned with virtual reality each square in G(n) into four equal squares is a models or simulations, they are concerned with grid G(n + 1). Figure 13.1 shows the series of the spatio-temporal-causal structure of the phys- grids G(0), G(1), G(2). The definition can be ical universe itself. further extended to generate a 3D lattice whose points are at the corners of cubical cells. We can then add a fourth dimension. If this is time, the result is a 4D space-time structure in which all 2 Programs and Theories the spatial and temporal facts (distances and durations) are recursive. A program can be described as a recursive defini- A recursive definition associates each finite tion of some property P, consisting of at least whole number with some set of facts. The basis two parts: a basis clause, which states some ini- clause associates 0 with some facts; the first ap- tial fact about P, and a recursion clause, which plication of the recursion clause associates 1 with defines new facts about P in terms of old facts. some facts; the nth application associates n with Properties that have recursive definitions are sim- some facts, and so forth. The recursive definition ply referred to as “recursive.” If P is recursive, of is-a-chain, for example, associates 0 with O, 1 then the set of all objects that have P is also with OO, 2 with OOO, and so on. If R is some described as recursive, and each object in that recursive definition, let R(n) be the nth set of set is said to be a recursive object. Many physical facts generated by R. So, R(n) is the nth chain, things are recursive. For example, the property or the nth grid. A set of facts is a state of affairs. is-a-chain can be defined recursively thus: (1) a A state of affairs is finitely recursive if and only if link O is a chain; (2) if X is a chain and O is a there is some recursive definition R and some link, then X attached to O is a chain. The defini- finite whole number n such that the state of tion generates a series of chains: O, OO, OOO, affairs is the set of facts R(n). etc., where each chain is a linear or one-dimen- To extend recursive definitions to the infinite sional series of neighboring points (the links). one needs to define a state of affairs at infinity. This recursive definition can be extended to gen- The result is a definition by transfinite recursion. erate discrete structures with more dimensions If some finite recursive definition R associates D. A grid is a 2D arrangement of neighboring each finite number n with R(n), then one way points, i.e. a set of points and a recursive neighbor to extend R transfinitely is to add a limit clause relation that determines a distance relation. All that associates infinity (∞) with the limit of R(n) geometric facts about the grid are recursive. Here as n goes to infinity. R(∞) is then the limit of is an informal recursive definition of the prop- R(n) as n increases without bound. For example: erty is-a-grid: (1) a set of four points occupying recall Zeno’s paradox of the racecourse, in which the corners of a square is a grid G(0); (2) if G(n) Achilles starts at 0 and runs to 1 by always going is a grid, then the set of points made by dividing halfway first. Achilles traverses the points 1/2,

G(0) G(1) G(2)

Figure 13.1: A recursive series of discrete 2D spaces

179 Eric Steinhart

3/4, 7/8, and so on. The property is-a-Zeno- 3 Finite point is defined by finite recursion like this: (1) R(0) is the Zeno point 1/2; (2) if R(n) is the Zeno point P/Q, then R(n + 1) is the Zeno The set of logically possible programs is infinite point P/Q + P/2Q. This definition generates and hence much larger than the set of programs the series: 1/2, 3/4, 7/8, and so on. The theory that can actually be written by human beings. of infinite series from the calculus shows that All programs have possible physical models. Since the limit of the series of Zeno points is 1. So we some of the programs we write are realized by add the limit clause (3) R(∞) is the limit of artificial computers (which are physical parts of R(n) as n goes to ∞. The result is a transfinite our universe), some programs are at least ap- recursive definition. If recursive definitions that proximately true of some parts of our universe. take limits are available, then much of the calcu- Consider now that some states of affairs in our lus is also available. However, it is not necessary universe are finitely recursive, and that our uni- to restrict limit clauses to limits as defined in verse is a large state of affairs, but it may be calculus. We can also let R(∞) be the state of finite (Finkelstein 1995; Steinhart 1998). Finite affairs that contains all the facts in every R(n) digital physics argues that there is some recursive for which n is finite. R(∞) can be the union of definition R and some finite number N, such that all the R(n) for n finite. We can thus define R(N) entails all and only the physical facts about recursive states of affairs transfinitely. For instance, our universe. The recursive definition R(N) is a if the grid G(∞) is the union of all the G(n) for program P that runs for N steps. Each step defines n finite, then G(∞) is an infinitely subdivided some change (some state-transition) of our uni- space. It is a grid such that between any two verse. So finite digital physics suggests that there points there is another. Generally, a state of is some finite P that is exactly instantiated by our affairs is transfinitely recursive if and only if whole universe. P is the ultimate theory for our there is some transfinite recursive definition R universe. We certainly cannot run P on any part of such that the state of affairs is R(∞). our universe; we may not be able to write P; what A theory T can be described as a collection of finite digital physics argues for is that P exists. facts that entails further facts. Every recursive A physically possible universe U is digital if definition is a theory. T is ultimate for some and only if it is finitely recursive. Since all finitely physical system S whenever T entails all and only recursive quantities are digital, if U is finitely the physical facts about S. A physical system S is recursive then all its physical quantities are digital, recursive if and only if there is some recursive i.e. discrete (measured by an integer variable) and theory that is ultimate for S, i.e. there is some with only finitely many values (finite upper and program that generates all and only the facts lower bounds). Discrete quantities are contrasted about S. If a physical system is finitely recursive with dense quantities (measured by rational num- this means that there is some recursive definition bers) and continuous quantities (measured by R and some finite number n such that R(n) is real numbers). Since almost all variables in the ultimate for S; if it is transfinitely recursive this classical (Newton–Maxwell) physical theory of means that there is some recursive definition R our universe are real number variables that refer and some transfinite number N such that R(N) to continuous physical quantities, and since is ultimate for S; otherwise it is nonrecursive. It almost all the equations in classical physics are is known that there are entities whose definitions differential equations that refer to continuous are nonrecursive. Nonrecursive systems typically transformations of those quantities, it might be involve the real numbers or logical undecidabil- argued that our universe is far too mathemat- ities. The three kinds: finitely recursive, trans- ically complex to be finitely recursive. However, finitely recursive, and nonrecursive are of course differential equations may be over-idealizations. logically exhaustive. Since our universe is a phys- The Lotka–Volterra differential equations, for ical system, it necessarily falls under one of those example, describe the interactions of predator– three kinds. Where our universe lies is an open prey populations as if they were continuous, question. yet animals come in discrete units. Moreover,

180 The Physics of Information classical physics has been replaced by quantum space-times. All distances in discrete space-times physics, where quantities (charge, mass, length, are integers, but a square whose sides have unit time) are quantized into discrete units (quanta). distance has a diagonal, and since the Pythag- Thus, analytic work on the foundations of quan- orean theorem shows that such diagonals are tum physics motivates John Wheeler’s theory irrational (non-integer) numbers, the theorem (Zurek 1990) that physical reality emerges from cannot be true for discrete space-times. If we try binary alternatives. Wheeler himself has referred to avoid this problem by treating the lengths to this position by means of the slogan “Its from of sides and the lengths of diagonals as funda- Bits.” Zeilinger (1999) argues that the quantiza- mental lengths (of which we allow integer mul- tion of information (as bits) entails the quantiza- tiples), then we have two incommensurable tion of all measurable physical quantities, and distances. Space is no longer uniform. Second: that this is a fundamental feature of physical discrete space-times have very limited internal systems. Whether all the fundamental physical symmetries for physical rotations and reflections. quantities of our universe are digital is an open Square lattices allow only 90-degree rotations; scientific question. Since it is a question about triangular or hexagonal lattices allow only 60- the form of the ultimate scientific theory, it is degree rotations. Nevertheless, actual physics not clear whether there are experiments that can seems to demand rotations of any degree. Finally: empirically decide this question. it is difficult to translate continuous differential If U is digital, then space-time has finitely equations into discrete difference equations. It is many dimensions and it is finitely divided into not known whether these mathematical features atomic (0-dimensional) point-instants (cells). A are genuine obstacles, or whether they are merely digital space-time is a network of finitely many inconveniences for scientists used to thinking of cells, in which each cell has links to finitely many space-time as continuous. The study of discrete spatial and temporal neighbors. Motion is not space-times is an active research area and it is continuous, time proceeds in clock ticks, and likely that lattice quantum field theories will solve space proceeds in steps. There is a maximal rate the problems associated with digital space-times. of change (a speed of light), namely one step Whether our universe has digital space-time re- per tick. Physical quantities in U are associated mains an open scientific question. with geometrical complexes of cells (e.g. the 0- If U is digital, then the causal regularities in dimensional cells in U, the 1D links between U are finitely recursive. The laws of nature are cells, the 2D areas bounded by links, the 3D finitely recursive transformations of digital phys- volumes bounded by areas, etc.). Each cell in a ical quantities and all the basic quantities stand digital universe is associated with at most finitely to one another in finitely recursive arithmetical many quantities, e.g. some digital mass. Links relations. Since all quantities are digital, an arith- between cells are associated with digital spins or metic transformation is possible (is consistent with other forces; areas bounded by links are associ- the fact that U is digital) if and only if that ated with digital charges. operation takes some finitely bounded integers Space and time in our universe could be finite as inputs and produces some finitely bounded and discrete. Relativity theory permits space to integers as outputs. More precisely: all the pos- be finitely extended and, according to quantum sible laws of nature for U must be recursive func- mechanics, space is finitely divided with a minimal tions on the integers, and they must produce length (the Planck length, about 10−35 meters) outputs within the finite upper and lower bounds and time has a minimal duration (the Planck on quantities in U. For example, suppose that U time, about 10−43 seconds). Furthermore, theor- consists of a space that is a 2D grid, like a chess- ies of loop-quantum gravity predict that space is board. Each cell on the board is associated with a discrete “spin-network.” some quantity of matter (some mass either 0 or Discrete space-times have mathematical fea- 1). The assignment of masses to cells is a discrete tures that may make them unsuitable for use as (binary) mass field. As time goes by (as the clock actual physical structures. First, discrete space- ticks), the mass field changes according to some times have different geometries than continuous causal operator. A causal operator for the mass

181 Eric Steinhart

field of U is digital if and only if it defines the our universe is transfinitely recursive. If a uni- mass field at the next moment in terms of the verse U is transfinitely recursive this means that mass field at some previous moments. The causal there is some recursive definition R and some operator on a binary mass field is a Boolean transfinite N such that R(N) is ultimate for U. function of bits that takes 0s and 1s as inputs The class of hyperdigital universes is very large. and produces 0s and 1s as outputs. A dynamical Hyperdigital physics permits physical infinities so system is a physical system whose states (synchronic long as they do not introduce logical inconsist- distributions of quantities) and transitions encies. While finite recursion subdivides space- (diachronic transformations of quantities) are time finitely many times, transfinite recursion recursively defined. It repeatedly applies a causal subdivides space-time infinitely. An infinitely sub- operator to its initial state to produce its next divided space-time seems to be consistent. If U states. Such dynamical repetition or iteration is is finitely recursive, then each quantity in U is recursive change. If U is a digital universe and measured by only finitely many digits; but if U is its causality is recursive and discrete, then U is a transfinitely recursive, it can contain quantities discrete dynamical system. Discrete dynamical sys- measured by infinitely long series of digits. It is tems are an area of active physical research. Our easy to define transformations on infinitely long universe could be a discrete dynamical system, series of digits by defining them in terms of in which case, all differential equations that relate endless repetitions of operations on single digits. continuous rates of change are ultimately based A quantity measured by an infinitely long series on finite difference equations involving digital of digits is infinitely precise. Infinitely precise quantities in digital space-time. Whether our uni- physical arithmetic seems to be consistent, that verse has digital causality is an open question. is, it seems that physical quantities can become There are many classes of digital universes. arbitrarily large or small without introducing any Cellular automata (CAs) are the most familiar contradiction. Still, it is necessary to be extremely digital universes. Conway’s Game of Life is a careful whenever introducing infinities into phys- popular (see Chapter 15, ical systems. For if any quantity is actually infin- ARTIFICIAL LIFE). Space, time, causality, and all itely large or small, then every quantity to which physical quantities in CAs are finite and discrete. it is arithmetically related must also be either CAs are computational field theories: all quantities actually infinitely large or small. Infinities often and transformations are associated with space- entail physical inconsistencies. time cells. Causality in CAs is additionally con- Since the transfinite includes the finite, if U strained to be local: the quantities associated with is any hyperdigital universe, then all the funda- each cell are recursively defined in terms of the mental physical quantities in U are finitely or quantities associated with the spatio-temporal transfinitely recursive. Say a physical quantity is neighbors of that cell. CA theory has seen great hyperdigital if and only if it is measured by some development (Toffoli & Margolus 1987), and infinitely long series of finite digits. Integers and CAs have seen extensive physical application rational numbers (fractions) are hyperdigital. A (Chopard & Droz 1998). There are many gen- real number is hyperdigital (it is a recursive real eralizations of CAs: lattice gasses (Wolf-Gladrow number) if and only if there is some recursive 2000), lattice quantum CAs and lattice quantum rule for generating its series of digits. For exam- field theories are currently active research areas in ple: π is hyperdigital since there is a recursive physics. Fredkin (1991) argues that our universe rule for generating each digit of the infinite is a finitely complex CA. series 3.14159. . . . If U is hyperdigital, then physical quantities (e.g. mass, charge, length, time) can be infinitely precise. Arithmetical operations in U, for example, can be infinitely 4 Transfinite Digital Physics precise manipulations of fractions. It is possible for space and time in hyperdigital Our universe may be too complex to be only universes to be infinitely subdivided (infinite finitely recursive. Hyperdigital physics argues that extension raises subtle problems). One way to

182 The Physics of Information define an infinitely subdivided space recursively complex structural features of concrete systems is by endlessly many insertions of cells between and are true of universes with infinitely complex cells. Recall the construction of the finite grids space-times (such as infinitely subdivided space- G(i) shown in Figure 13.1. We extend the con- times) or infinitely complex causal regularities. struction to the transfinite by adding a limit clause If U is hyperdigital, then its physical quantities, for the infinitely subdivided grid G(∞). G(∞) is space-time, and causal laws are all defined by the union of all the G(i) for finite i. G(∞) is a transfinite recursion. Consider a universe defined dense 2D cellular space: between any 2 cells, there as follows: (1) the space of U is an infinitely is always another cell. The topology of G(∞) is subdivided 3D grid like G(∞); (2) the time is not the same as the topology of an infinitely infinitely subdivided so that U is made up of extended 2D lattice (e.g. an infinitely large chess infinitely many space-time point-instants (cells); board). An infinitely extended 2D lattice is not (3) infinitely precise physical quantities (rational dense. Even though G(∞) is dense, each cell numbers or recursive real numbers) are associated (each point in G(∞)) still has exactly 8 neighbors. with geometrical complexes of cells (e.g. the 0- Since every cell has 8 neighbors, it is possible to dimensional cells in U, the 1D links between cells, run rules from finite 2D cellular automata (CAs) the 2D areas bounded by links of cells, the 3D on G(∞). If time is kept discrete, then Conway’s volumes bounded by areas of cells, etc.); (4) all the Game of Life CA can run on G(∞). It is also physical laws in U are transfinitely recursive func- possible to recursively define dense time by always tions on the rational numbers or recursive reals. inserting moments between moments on the Dense space-times whose causal laws are ATM temporal dimension. It is easy to build lattice transformations of infinitely long digit sequences gasses and other CAs on the dense grid. There is quantities are examples of hyperdigital universes. a large class of digital universes on dense lattices. These universes have infinitely complex dynamics. One way to define an infinitely subdivided 5 The Physics of Computation time is to use acceleration: each change happens twice as fast. For example: accelerating Turing machines (ATMs) are infinitely complex dynam- If our universe is digital, then it is possible that ical systems (Copeland 1998). An ATM consists some things in it are digital computers. Our of a Turing machine read/write head running universe obviously contains classical physical over an actually infinitely long tape. An ATM realizations of finite Turing machines (TMs). tape can have infinitely many 1s, unlike a clas- Therefore, it is at least finitely recursive. Since sical TM tape. An ATM is able to accelerate. If classical TMs have potentially infinitely long tapes, it performs an act at any speed, it can always and can operate for potentially infinitely long perform the next act twice as fast. ATMs can periods of time, finite TMs are not really even as perform supertasks (Koetsier & Allis 1997). An powerful as Turing machines. So far, all efforts

ATM starts with some initial tape-state T0 at to build computers more powerful than finite time 0. It computes at Zeno points. It performs TMs have failed. Quantum computers do not the first computation in 1/2 seconds that prints exceed the limits of finite Turing machines

T1 at time 1/2; it performs a second computa- (Deutsch 1985). Some suggestions for making tion in 1/4 seconds that prints T2 at 3/4; it hypercomputers involve accelerating the machin- performs its nth computation in 1/2n seconds ery to the speed of light or the use of unusually n − n to print Tn at time (2 1)/2 . At 1 second, the structured space-times. However, such sugges- ATM has computed infinitely many operations. tions are matters for science fiction. Since an At the limit time 1, an ATM outputs the limit ATM accelerates past any finite bound, it re- of the tape-states sequence {T0, T1, T2,...}, quires infinitely much energy to perform any if the series converges, or a blank tape-state if infinite computation. If our universe is digital, it does not converge. Copeland shows that then all the things in it are too, including hu- ATMs are more powerful than classical Turing man bodies and brains. If it is hyperdigital then machines. Programs for ATMs describe infinitely it is possible that some things in it (some proper

183 Eric Steinhart parts of it) are hyperdigital computers. How- alternatives, either true (1) or false (0). Wheeler’s ever, hyperdigital computers run into the lower slogan “Its from Bits” implies that physical real- limits imposed by quantum mechanics (e.g. the ity (the “Its”) is generated from binary alternat- Planck time or length) or into the upper limits ives (the “Bits”). Therefore, the “Its from Bits” imposed by relativity theory (e.g. the speed of program naturally hooks up with the metaphysics light). An accelerating Turing machine does not of possible worlds. If there are finitely many appear to be physically possible in our universe. propositions in some world, and if the logical If our universe is nonrecursive, then it phys- relations among them are recursive, then that ically realizes properties that have neither finite world is digital. If we want to link finite digital nor transfinite recursive definitions. Perhaps, it physics to the metaphysics of possible worlds, physically instantiates the nonrecursive real num- we need to define the propositions physically. ber continuum. Analog computers are possible One way to do this is via Quine’s suggestion of in universes that instantiate the nonrecursive con- a Democritean world (Quine 1969: 147–52). It tinuum. However, the continuum is not math- has been argued that Quine’s theory of Demo- ematically well understood (e.g. its cardinality critean physics leads to a Pythagorean vision of is undetermined; it has unmeasurable subsets; physical reality as an entirely mathematical sys- it is supposed to be well-ordered but no well- tem of numbers. If some universe is recursive, ordering is known; etc.). Attempts to define then there is some (finitely or transfinitely) com- analog computation in our universe (e.g. con- putable system of numbers that is indiscernible tinuously varying electrical current) conflict with from it. A recursive universe is Pythagorean: the laws of quantum mechanics. If quantum physical structures are identical with numerical mechanics is a correct description of our uni- structures. The mystery of the link between the verse, then is unlikely that there are any analog material and the mathematical is thereby solved: computers in our universe. If today’s biology is the material is reducible to the mathematical. right, then real neural networks are not analog Since our brains and bodies are physical things, machines. Perhaps our universe is nonrecursive they are finitely recursive, transfinitely recursive, because its structure is logically undecidable. Just or nonrecursive. If human beings are somehow as Gödel’s theorems prove (roughly) that there more than physical, then their transcendence of are facts about an arithmetic structure S that are material reality is reflected by their computational not decidable within S, it may be that analogous abilities. If physics is digital, then we transcend theorems tell us that there are facts about our it exactly insofar as we are hyperdigital or even universe that are not decidable by any given axi- nonrecursive. If physics is hyperdigital, then we omatic physical theory. The physical structure of transcend it exactly insofar as we are nonrecursive. our universe may not be axiomatizable at all. It Perhaps discussions of free will or our math- may be deeply undecidable. Physical computation ematical capacities aim to find the degree by in our universe, as far as we presently know, is which we surpass physical reality. It is possible limited to finite Turing computability. Whatever that we are parts of hyperdigital computers even the upper bound on physical computation in if we are only digital. The limits of our cognitive our universe, it seems clear that this bound is powers may be the limits of the computers that contingent. While hypercomputers seem both contain us, even if we are only parts of those mathematically and physically possible, the in- machines, and even if those machines infinitely ternal limitations of our universe might actually transcend physical computability. If physics is prevent it from containing any. recursive, then there is some recursive property (a program) that is exactly instantiated by each person’s body over the whole course of its life. The history or fate of each person’s body is a 6 Conclusion program. If such programs exist, they are multi- ply realizable; so, if physics is recursive, and if all Philosophers have long defined possible worlds possible recursive worlds exist, then our lives (and as sets of propositions. Propositions are binary all variations of them) are endlessly repeated

184 The Physics of Information within the system of digital or hyperdigital uni- the thesis that nature is finitely recursive hence verses. One could hardly hope for a richer kind a CA.] of personal immortality. So far from eliminating Koetsier, T. and Allis, V. 1997. “Assaying the soul, recursive physics may show that it has supertasks.” Logique et Analyse 159: 291–313. entirely natural realizations. [An excellent analysis of transfinite operations.] Quine, W. V. 1969. Ontological Relativity and Other Essays. New York: Columbia University Press. [Discusses Democritean worlds.] References Steinhart, E. 1998. “Digital metaphysics.” In T. Bynum and J. Moor, eds., The Digital Phoenix: Chopard, B. and Droz, M. 1998. Cellular Auto- How Computers are Changing Philosophy. Oxford mata Modeling of Physical Systems. New York: and Malden, MA: Blackwell, pp. 117–34. [An ana- Cambridge University Press. [An advanced text lysis of the thesis that nature is finitely recursive, that discusses the use of CAs in many aspects of with an extensive bibliography on finitely recurs- physical theory.] ive physics.] Copeland, B. J. 1998. “Super Turing-machines.” Toffoli, T. and Margolus, N. 1987. Cellular Auto- Complexity 4(1); 30–2. [A brief introduction mata Machines: A New Environment for Modeling. to accelerating Turing machines, with many Cambridge, MA: MIT Press. [A classic work on references.] the use of CAs in physical theory; deals nicely Deutsch, D. 1985. “Quantum theory, the Church– with CAs and differential equations.] Turing principle and the universal quantum com- Wolf-Gladrow, D. 2000. Lattice-gas Cellular Auto- puter.” Proceedings of the Royal Society, Series A, mata and Lattice Boltzmann Models: An Intro- 400: 97–117. [The classical discussion of quan- duction. Lecture Notes in Mathematics, vol. 1725. tum computing.] New York: Springer-Verlag. Finkelstein, D. 1995. “Finite physics.” In R. Herken, Zeilinger, A. 1999. “A foundational principle for ed., The Universal Turing Machine: A Half- quantum mechanics.” Foundations of Physics Century Survey. New York: Springer-Verlag, 29(4): 631–43. [Discusses the use of informa- pp. 323–47. [A discussion of physical theories tion theory as a basis for quantum mechanics; based on the assumption that nature is finite.] references to the “It from Bit” program.] Fredkin, E. 1991. “Digital mechanics: an informa- Zurek, W. H., ed. 1990. Complexity, Entropy, and tional process based on reversible universal cel- the Physics of Information. SFI Studies in the lular automata.” In H. Gutowitz, ed., Cellular Sciences of Complexity, vol. 8. Reading, MA: Automata: Theory and Experiment. Cambridge, Addison-Wesley. [The classic work on the physics MA: MIT Press, pp. 254–70. [A discussion of of information, with many important essays.]

185 Roberto Cordeschi

Chapter 14 Cybernetics Roberto Cordeschi

sciences, as it developed in a highly interdiscip- Introduction linary approach, aimed at seeking common con- cepts and methods in rather different disciplines. The term cybernetics was first used in 1947 by In classical cybernetics, this ambition did not with reference to the centri- produce the desired results and new approaches fugal governor that James Watt had fitted to his had to be attempted in order to achieve them, steam engine, and above all to Clerk Maxwell, at least partially. who had subjected governors to a general math- In this chapter, we shall focus our attention ematical treatment in 1868. Wiener used the word in the first place on the specific topics and key “governor” in the sense of the Latin corruption concepts of the original program in cybernetics of the Greek term kubernetes, or “steersman.” As and their significance for some classical philo- a political metaphor, the idea of steersman was sophic problems (those related to ethics are dealt already present in A. M. Ampère, who in 1843 with in Chapter 5, COMPUTER ETHICS, and Chap- had defined cybernetics as the “art of govern- ter 6, COMPUTER-MEDIATED COMMUNICATION AND ment.” Wiener defined cybernetics as the study HUMAN–COMPUTER INTERACTION). We shall then of “control and communication in the animal examine the various limitations of cybernetics. and the machine” (Wiener 1948). This definition This will enable us to assess different, more captures the original ambition of cybernetics recent, research programs that are either ideally to appear as a unified theory of the behavior of related to cybernetics or that claim, more living organisms and machines, viewed as sys- strongly, to represent an actual reappraisal of it tems governed by the same physical laws. on a completely new basis. The initial phase of cybernetics involved dis- ciplines more or less directly related to the study of such systems, like communication and con- 1 The Basic Idea behind Classical trol engineering, biology, psychology, logic, and Cybernetics neurophysiology. Very soon, a number of at- tempts were made to place the concept of con- trol at the focus of analysis also in other fields, The original research program of classical cyber- such as economics, sociology, and anthropology. netics was strongly interdisciplinary. The research The original ambition of “classical” cybernetics fields within which cybernetics interacted can thus seemed to involve also several human be grouped under three headings: engineering/

186 Cybernetics biology, philosophy/psychology, and logic/ tiny size of their movements, cannot be cor- neuroscience. rectly listed among the ancestors of cybernetic machines. These automata are merely “mechanic” and lack the fundamental self-regulating capa- 1.1 Cybernetics between engineering city typical of feedback control systems. and biology The study of the different feedback control mechanisms was common in Wiener’s times, The study of automatic control devices in ma- as was the analysis of self-regulation in living chines attained full maturity around the middle organisms. In the latter case, the existence of of the twentieth century. The essence of auto- such systems, which may be compared to negat- matic control resides in the capacity of a (usually ive feedback devices, had already been described electromechanical) system S to attain a goal-state in modern physiology, in particular by Claude G (the Greek word for goal is telos) set by a Bernard and Walter B. Cannon. Examples in- human operator, without the latter having to clude systems that automatically maintain at a intervene any further to modify the behavior of constant level body temperature, breathing, S as it attains G. In this case, one may also speak and blood pressure. In the late 1920s, Cannon of closed loop or feedback control. Engineers have referred to these systems as homeostatic systems. mathematically described different types of closed Wiener’s definition of cybernetics thus finds loop, which have been used in both electronic its initial justification in the converging of two and control engineering. A typical example is research areas that, although having developed the positive feedback used in oscillators, or in separately within engineering and biology, in the so-called regenerative receivers in the early Wiener’s times seemed to share an essential core radios, where part of the output signal is fed of common problems all strictly related to the back in such a way as to increase the input study of control and information transmission signal. Of greater interest in this context is processes, abstracted from mechanical or biolo- negative feedback. The behavior of a negative gical implementations. In 1943 Wiener, together feedback system is governed by the continuous with Arturo Rosenblueth, a physiologist and one comparison made between the current state C of Cannon’s pupils, and the engineer Julian and the state established as a reference para- Bigelow, wrote a brief article, entitled “Behavior, meter G, in such a way that the system uses this Purpose, and Teleology,” in which the unified error information to avoid ever wandering too study of living organisms and machines, which far from the latter. Watt’s governor is an exam- a few years later was to suggest the term “cyber- ple of such a system: it maintains the speed of netics,” was set out explicitly (Rosenblueth, rotation of the driving shaft of a steam engine Wiener, & Bigelow 1943). approximately constant in the face of load vari- ations. It is thus capable of regulating itself automatically (self-regulation) without the need 1.2 Cybernetics between philosophy for any intervention by human operators once and psychology the latter have set the reference parameter (in this case G = the desired speed). In their 1943 article, Rosenblueth et al. actually Devices like Watt’s governor are the genuine provided considerably more than a comparative and influential precursors of cybernetic machines. analysis of the self-regulating mechanisms in liv- Examples of such self-regulating systems were ing organisms and machines. They supported a known long before Watt’s governor, as far back view that was immediately perceived as provocat- as the period of ancient Greece (Mayr 1970). ive by numerous philosophers and which gave On the contrary, the clockwork automata of the rise to a very lively debate. The three authors, eighteenth century – such as the androids con- after summarizing the fundamental theoretical structed by the Swiss watchmaker Pierre Jaquet- issues involved in the study of the new control Droz and his son Henri-Louis – although devices, claimed that science could now reap- astonishing in the realistic reproduction and the praise the vocabulary of teleology, which included

187 Roberto Cordeschi such terms as purpose, goal, end, and means. and ultimately, physical microexplanation of the According to them, teleological analyses had been teleological behavior itself. Cybernetics could “discredited” from the scientific point of view then provide further support for the Unitary because of the Aristotelian notion of purpose as Science proposed by logical neopositivism, ac- final cause. The term “final cause” suggests that cording to which it was ultimately possible to the purpose is supposed to guide the behavior hypothesize the reduction to physics of the con- directed towards its attainment, despite the fact cepts and methods of the other sciences. that, insofar as the purpose is a state to be at- Clearly, cybernetics was reproposing the idea tained (end state), it is a future state. Compared of the organism-machine of the old mechanist with the ordinary causal explanation, in which tradition in a completely new context. The idea the cause always precedes the effect, the teleolo- had already been implicit in Descartes who, in gical explanation seems to give rise to a puzzle, the Traité de l’homme (1664), had described that of the reversal of causal order. The hypo- the functioning of the human body in terms of thesis advanced by the founders of cybernetics hydraulic automatisms. Descartes had argued was that the vocabulary of teleology might be for a fundamental distinction between human revaluated by means of an objective or opera- beings and true automata, which represent the tional definition of its terms that allows the puzzle nonhuman animals in the living world. How- introduced by the notion of final cause to be ever, La Mettrie, in his Homme machine (1748), avoided. In the definition they proposed, the dropped Descartes’s dualism and claimed that man “purpose,” i.e. the final state G pursued by a himself is merely a machine with a particularly system S, either natural or artificial, is the state complex organization. Mechanistic conceptions that serves as a reference parameter for S, and of the living were proposed in the eighteenth S’s teleological behavior is nothing else but S’s century also by other authors, such as George behavior under negative feedback control. This Cabanis, while Thomas Huxley, referring back was a provocative idea since psychologists and to Descartes’s theory, claimed in the following vitalist philosophers saw purposeful action as char- century that man was nothing but an automaton acterizing only the world of living organisms, possessing consciousness. The animal-automaton and opposed the latter both to the world of theory was then revived, in the interpretations artificial or synthetic machines and to the phys- of animal behavior, in terms of chains of reflexes ical world in general (see, for instance, McDougall in psychology and philosophy, between the eight- 1911). In fact, the new feedback machines, by eenth and the nineteenth centuries (see Fearing interacting with the external environment, are 1930). capable of automatically changing the way they It is again the new idea of a cybernetic function in view of a given purpose. For philo- machine capable of interacting with the environ- sophers concerned with a materialistic solution ment that abated interest in the reflex-arch con- of the mind–body problem, cybernetics thus sug- cept rampant in conventional neurological and gests how certain behavior regularities, usually psychological mechanisms. Indeed, instead of the classified as teleological to distinguish them from simple stimulus–response relationship typical of causal regularities in physics, may be described the reflex arch, the interest was now focused on using purely physical laws and concepts. a circular relationship, or loop, through which As pointed out by the logical positivist philo- the response could be fed back as the effect of sopher Herbert Feigl, a champion of the mater- the stimulus itself. Behavioristic psychologists ialist thesis of the identity between types of like E. Thorndike and Clark L. Hull had already mental states and types of brain states, with the pointed out this aspect. Thorndike had explicitly advent of cybernetics the concept of teleological formulated a trial-and-error learning Law of machine was no longer a contradiction in terms Effect, in which it was precisely the effect that (Feigl 1967). In addition, according to Feigl, reinforced the correct response among the many cybernetics suggested the possibility of integrat- possible responses attempted at random by the ing the various levels of explanation, the mental organism during the learning phase. Between the and the physical, in view of a future neurological, 1920s and 1930s, Hull proposed an ambitious

188 Cybernetics research program, which he himself defined as viewed as a threshold unit, that is, functioning a “robot approach,” which foreshadowed that according to the “all-or-nothing law” (a neuron of cybernetics. The aim of the robot approach fires or does not fire according to whether the was the construction of machines that actually pulses it receives exceed a certain threshold or worked and hence could be viewed as mechanical not). Neurons of this type could be intercon- models of trial-and-error learning and learning nected to form networks, whose functioning by conditioning. By constructing these models could then be explored according to the laws of (which were actually very simple electromechan- classic propositional logic. McCulloch and Pitts’ ical devices), Hull set out to prove that it was article forms the basis of the development of useless to employ vitalist entities/concepts to artificial neural networks as well as computer sci- attempt to account for mental life. Indeed, if a ence. John von Neumann, for example, adopted machine behaves like an organism – in Hull’s its symbolic notation in 1945 in his well-known view – the behavior of an organism may be ac- First Draft, in which he described the computer counted for by means of the same physical laws architecture that was later named after him (all as those used to explain the machine’s behavior. ordinary PCs have a von Neumann architecture). The reductionism underlying this thesis explains Neurology had already suggested to psycholo- why Hull subscribed to the logical positivist gists laws of learning based on the assumption hypothesis of Unitary Science. that the physical basis of learning is to be sought Kenneth Craik, the Cambridge psychologist in the presence, in the central nervous system, of who, at the dawn of cybernetics, described neurons whose reciprocal connections may be several models of adaptation and learning based strengthened or weakened according to the on different types of feedback, pointed out that stimuli received by the organism from the out- Hull’s position actually represented an innovation side world. The tradition of connectionism, which of mechanistic tradition. Unlike the supporters dates back to Thorndike, was revived in the 1940s of mechanistic conceptions of life such as Cabanis in the research carried out by Donald Hebb. and others, based on the man-machine metaphor, Unlike the preceding connectionism, Hebb’s Hull had endeavored to construct learning approach supported a new interpretation of the models that, insofar as they were not generic nervous system containing reverberating neural metaphors but working implementations, allowed loops. The presence of such loops in the brain the hypothesis of the mechanical nature of this tissue had been observed, among others, by the phenomenon to be tested (Craik 1943: 52). This neurologist R. Lorente de Nó. This new rep- observation by Craik on the nature of models is resentation of the nervous system now tended fundamental, as it sheds light on the simulative to replace that of the quasi-linear connections methodology later developed by cybernetics between stimulus and response, which were pre- and the mental life sciences that followed his viously predominant. Within this new paradigm, teachings. Hebb formulated the learning law named after him, according to which a connection between two neurons activated at short time intervals tends 1.3 Cybernetics between logic and the to be strengthened (Hebb 1949). neurosciences After the official birth of cybernetics, neuro- logical connectionism comes into contact with The interaction with logic and neurology is the neural networks à la McCulloch and Pitts in another feature of classical cybernetics. The bio- the work done by Frank Rosenblatt, the builder physicist Nicolas Rashevsky had already made a of one of the best-known machines of the clas- mathematical analysis of several nervous func- sical cybernetics era, the Perceptron. Constructed tions. However, it was the article published in at Cornell University in 1958, the Perceptron 1943 by Warren McCulloch and that displays an elementary form of learning, consist- introduced logic into cybernetics (Anderson & ing in learning to discriminate and classify visual Rosenfeld 1988). The article proposed a “formal” patterns, such as letter-shaped figures (Anderson neuron, a simplified analog of a real neuron, & Rosenfeld 1988). In its simplified version, the

189 Roberto Cordeschi

Perceptron consists of three layers: a first layer, equilibration the allegedly “teleological” processes the analog of a retina, collects the input stimuli of the adaptation of organisms to the environ- and is composed of several units or neurons à la ment, anticipating the aforementioned claim of McCulloch and Pitts, randomly connected to one Rosenblueth, Wiener, and Bigelow. According or more units of the second layer, the associ- to Ashby, trial-and-error adaptation “is in no ation system. The units comprising the latter, or way special to living things, . . . it is an element- association units, are then connected to a third ary and fundamental property of matter” (Ashby layer, the effector system, which comprises the 1945: 13). In order to test this hypothesis, Ashby response units. The connections among the asso- constructed a machine that he described in his ciation units are modifiable: learning actually book Design for a Brain (Ashby 1952), as the occurs through the modification of their strength “Homeostat,” with obvious reference to Cannon. or “weight.” In the first Perceptron experiment, The Homeostat embodied a new and important learning was based essentially on reinforcement concept, that of “ultrastability,” in addition to rules. Further experiments led to the formula- that of feedback control or “stability.” In Ashby’s tion of a supervised learning rule, used by the definition, a system is said to be “ultrastable” Perceptron to modify the weighting of the con- when it is not only capable of self-correcting its nection in the case of an incorrect response, own behavior (as in the case of feedback control leaving it unchanged when it was correct. Other systems), but is also capable of changing its own neural networks have embodied quantitative internal organization in such a way as to select statements of the Hebb rule or its modifications. the response that eliminates a disturbance from Other learning models developed during the the outside from among the random responses classical cybernetics period were the mobile ro- that it attempts. In this way, a system such as bots, such as those simulating an animal learn- the Homeostat is capable of spontaneously re- ing to go through a maze. Thomas Ross invented establishing its own state of equilibrium: it thus the forerunner of this type of synthetic animal, displays a simple form of self-organization. The which was influenced by Hull’s robot approach. notion of ultrastability was deemed more inter- In collaboration with the behavioristic psycholo- esting than that of simple stability based on gist Stevenson Smith, in 1935 Ross constructed feedback control because it pointed the way to a “robot rat” at Washington University. Much simulating in an artifact some features of the more interesting as models of simple learning plasticity and variability of response typical of forms, as well as being more popular, are the animal behavior. For example, according to robots constructed by Walter Grey Walter at Ashby, ultrastability could be considered on the the Burden Neurological Institute, in England, basis of Thorndike’s Law of Effect. the electronic “tortoises.” The simplest of these could successfully avoid any obstacles in their path; other, more complex ones, learned by con- 2 Limits of Classical Cybernetics ditioning to react to different visual and audit- and New Developments ory stimuli (Walter 1953). The tortoises had a very simple structure. They were composed of only a small number of internal units, and Grey All these lines of research soon entered into Walter considered this to be confirmation of the crisis and were drastically curtailed, when not assumption that, in order to account for relat- actually abandoned, between the 1960s and ively complex behavior by organisms, it is not so 1970s. This happened mainly because of the early much the number of neurons as the number of successes of a new discipline, which resumed the their connections that accounts for the relatively ambition of cybernetics to strive for the unified complex behavior of living organisms. study of organisms and machines, although on In the newborn field of cybernetics, again in a completely different basis, namely Artificial England, William Ross Ashby was perhaps the Intelligence (AI). Subsequently, several research first to investigate the physical bases of learn- programs typical of cybernetics were resumed, ing. As early as 1940, he described in terms of including an explicit attempt to reformulate a

190 Cybernetics

“new cybernetics.” In the present and the fol- them, the problem-solver constructs an internal lowing section, we shall examine the concepts representation of the problem space and works and principal results characterizing these differ- out plans aimed at finding a solution, or the final ent phenomena and their significance for the state or goal, within this space. In these cases, it philosophy of mind and epistemology. is not necessary for the teleological activity to be guided by a final state that actually exists in the external environment (see Pylyshyn 1984 for 2.1 Teleological machines and further details). computer programs As regards the simulation of cognitive pro- cesses, the introduction of the concept of algo- The claim that purpose may be defined opera- rithm, which underlies the concept of program, tionally, by means of the negative feedback represented an undisputed step forward and notion, was challenged by many philosophers, led to the development of Cognitive Science. who argued that the latter does not really fulfill Prompted by the notion of algorithm, or, more all the conditions for appropriately considering precisely, of a Turing machine (see Chapter 1, a behavior pattern as purposeful. In the first COMPUTATION), is a philosophic position critical instance, such a definition is always relative to of reductionist materialism in the mind–body the external observer who attributes purposes to problem. This is functionalism, which was intro- the system, while it tells us nothing about the duced in the philosophy of mind by Putnam in purposes of the system. Furthermore, in any such his seminal article “Minds and Machines” (1960). system it is the feedback signals from the object Putnam argued that mental states could be stud- or the state pursued as a goal existing in the ied not by referring them directly to brain states, external environment that guides the system’s but on the basis of their functional organization, purposeful behavior. In the case of non-existent that is, of their reciprocal interactions and inter- objects, which may nevertheless be the content actions with the sensory inputs and behavioral of beliefs or desires of the system, the cybernetic outputs (see Chapter 9). approach seems to have nothing to say (see, for example, Taylor 1966). Pioneers of AI further criticized the incapa- 2.2 Neural networks city of the artifacts proposed by cybernetics, such as neural networks or systems with simple The early success of AI in constructing computer self-organizing capability, to simulate cognitive programs that could tackle significant cognitive processes. They pointed out that, in order to tasks further hindered research on neural net- reproduce artificial teleological behavior in a works, the descendants of the Perceptron. The system, such as making inferences or problem- decline in research on neural networks became solving, it was necessary to study selection and generalized after the publication of Minsky & action-planning procedures that, in the case of Papert 1969, which demonstrated the effective an artificial system, could be realized by a com- difficulties encountered by the Perceptrons in puter program (see Chapter 9, THE PHILOSOPHY discerning even very simple visual stimuli. Despite OF AI AND ITS CRITIQUE). Actually, early AI pro- these early failures, several researchers in differ- grams were considered teleological systems, al- ent countries, such as James Anderson, Eduardo though in this case the purposes were represented Caianiello, Stephen Grossberg, and Teuvo as symbol structures holding information about Kohonen, continued to work on neural networks the goals pursued. Other symbol structures were (Anderson & Rosenfeld 1988). Rosenblatt’s work used to organize the system’s behavior into com- was finally vindicated in the early 1980s by two plex hierarchies, such as processes for creating events, accompanied by the development of large subgoals, selecting the methods for attempting computers, allowing the hitherto impossible them, testing them, and so on. Two good ex- simulation of complex neural networks. John amples are chess playing and theorem proving, Hopfield demonstrated that symmetrical neural the task environments preferred by early AI. In networks necessarily evolve towards steady states,

191 Roberto Cordeschi later called “attractors” in dynamical system theory Rodney Brooks has pointed out the limits of (see Chapter 3, SYSTEM: AN INTRODUCTION TO both AI robotics and cybernetic robotics clearly. SYSTEMS SCIENCE), and can function as associative First, cybernetic robotics did not take into con- memories (Anderson & Rosenfeld 1988). David sideration fully the possibility of decomposing Rumelhart and collaborators published a series complete behavior into simpler modules with the of papers based on a “parallel distributed process- task of controlling actions that are more element- ing” (PDP) approach to information, showing ary. Secondly, cybernetic robotics either did not how a learning algorithm based on error correc- recognize or else underestimated the potential tion, known as “back-propagation,” made it pos- of digital computation and its greater flexibility sible to overcome the main limitations of neural vis-à-vis analog computation. In conclusion, “the networks reported by Minsky and Papert. These mechanisms of cybernetics and the mechanisms limitations were actually found to apply only to of computation were intimately interrelated in networks with a single associative-unit layer, such deep and self-limiting ways” (Brooks 1995: 38). as the simple Perceptron, but not to multilayer The new architecture proposed by Brooks nets, that is, networks with more than one layer appears as a radical alternative to the AI robotics of associative units “hidden” between the input approach and at the same time represents an layer and the output layer. Since the 1980s, attempt to identify a level of abstraction that research on neural networks differing even more would allow the limitations of cybernetic robot- substantially from the original Perceptrons has ics to be overcome. Brooks’ “subsumption archi- flourished, and numerous models are currently tecture” describes the agent as composed of available both in neurology and psychology (see functionally distinct control levels, a possibility Chapter 10, COMPUTATIONALISM, CONNECTIONISM, ignored in cybernetic robotics. These control AND THE PHILOSOPHY OF MIND). This research, levels then act on the environment without to the extent to which it proposes models with being supervised by a centralized control and an architecture closer to that of the real brain action planning center, as is the case instead in than the algorithmic models proposed by AI and AI robotics. In the subsumption architecture, Cognitive Science, seems to provide strong the low-level control routines, operating via con- arguments to reject functionalism. The debate has tinuous feedback loops with the environment, resuscitated materialist-reductionist solutions of are connected to high-level routines that con- the mind–body problem (e.g. Churchland 1986) trol a more complex behavior. For instance, the that are reminiscent of the kind of positions we robot Allen, the first member of this generation have seen above to be popular during the age of of new robots or “creatures,” is capable of avoid- classical cybernetics. ing different persons and obstacles in its path (a low-level, essentially reactive task) while con- tinuing to pursue a goal assigned to it (that is, 2.3 New robotics a higher level task). Brooks’ approach and that of behavior-based robotics in general, are con- The construction of mobile robots such as the strained by the fact that, in the end, it is not easy “tortoises” very soon came to a standstill owing to integrate an increasing number of element- to the above-mentioned predominance in AI of ary modules to obtain behaviors that are more interest in the procedures of reasoning, plan- complex. Evolutionary robotics, based on genetic ning, and problem-solving. A new kind of robot algorithms, is an attempt to get round these dif- was constructed based on the idea that an agent ficulties. In general, these approaches to robotics must have an explicit symbolic representation, have several advantages, such as robustness and or centralized model of the world, in order to the capability of real-time response. However, act in it successfully. The rather disappointing the trade-off consists of limitations imposed on results obtained in this sector in the 1970s planning and reasoning capabilities. encouraged a different kind of approach that, Behavior-based robotics and evolutionary ro- although connected to cybernetics, acknowledged botics have had the merit of attracting attention its limitations and tried to overcome them. to the importance of several aspects neglected

192 Cybernetics by early AI and by radical functionalism, namely of cybernetics, i.e. feedback and information con- developmental issues in cognition and the fact trol, could be applied also to the study of a very that the intelligence of an agent cannot easily wide range of all sorts of forms of interaction be “disembodied,” since it is also the result of among organisms or agents. In this way, cyber- the deep interaction between the agent and its netics began to be used as a meeting ground for environment. This accounts for the importance specialists in widely differing disciplines, as is acquired by “situated” cognition and for the shown by the Macy Foundation Conferences revaluation of the role of perception and of the held in New York between 1946 and 1953. The body in general, as well as for the attention involvement of neurologists and psychologists devoted to what Steven Harnad has defined proved inevitable from the outset. “He who stud- the “grounding problem,” i.e. the problem of ies the nervous system cannot forget the mind, grounding the meaning of symbols in an artifi- and he who studies the mind cannot forget the cial system on reality (Clark 1997). nervous system,” said Wiener, so that “the vocabulary of the engineers soon became con- taminated with the terms of the neurophysiolo- gists and the psychologists” (Wiener 1948: 18, 3 Self-organization and 15). In addition to the presence of neurologists Complexity (e.g. Rafael Lorente de Nó) and psychologists (e.g. Kurt Lewin), those historic interdisciplinary In the preceding section, we have discussed the seminars were also attended by pioneers of com- limits of the cybernetics program. In doing so, puter science and of information theory (e.g. we have identified several research programs that Claude Shannon), as well as by sociologists (e.g. were developed in opposition to this program, Paul Lazarsfeld), ecologists (e.g. George E. as in the case of symbolic AI, or else could be Hutchinson), and social scientists (e.g. Gregory ideally linked to this program, such as the new Bateson). The negative-feedback principle soon neural networks and the new robotics approaches. became a universal principle by means of which The latter research programs are able to over- to interpret the evolution towards an equilib- come at least some of the limitations of early rium state of a wide range of complex systems – cybernetics, and do so in open opposition to social, political, pedagogical, economic, industrial, symbolic AI. In the present section, we shall and ecological. Laws belonging to specific dis- look at other developments that, again in opposi- ciplines, such as Maupertuis’s principle in physics tion to symbolic AI, are explicit projects for a or that of Le Châtelier in chemistry, as well as new cybernetics. Before doing so, however, some different laws describing optimization phenom- further developments of the original cybernetics ena in economics and interspecies interaction in program must be briefly sketched. biology, were to appear as examples of this unique universal principle. Inevitably, parallel to, and often mingled with, the work of the various 3.1 Cybernetics and the human researchers, who were trying out new conceptual sciences synthesis tools on specific problems, there arose a popular philosophy of cybernetics that some- The value of Wiener’ cybernetics hypothesis was times ended up employing cybernetic concepts confirmed by the development of metaphorically, going as far as to interpret the and the spread of the new negative feedback notion of feedback as the “revealer of nature’s mechanisms, and by the discovery of automatic secret.” It was Wiener himself who appealed regulatory processes in living organisms, com- against the “excessive optimism” of all those who, parable to those of negative feedback. This led like Bateson and the anthropologist Margaret to several attempts to develop cybernetic models Mead, believed it possible to apply the ideas of of the functions of living organisms (McFarland cybernetics to anthropology, sociology, and eco- 1971). Soon, however, a much more radical ap- nomics, overestimating “the possible homeostatic proach began to gain popularity: the basic ideas elements in the community,” and ultimately

193 Roberto Cordeschi turning them into the cornerstone of an approach Other authors shift the emphasis away from to complexity. More generally, while cybernetics the notion of control, as introduced by Wiener, was to suggest an extension of the natural-science on to the concepts of self-organization and auto- method to the human sciences, in the hope of nomy. These authors are closer to Ashby, who repeating in the latter field the successful results had insisted on the centrality of these notions. obtained in the former, to the excessive optim- They focus their attention on a classic topic in ism was actually added a “misunderstanding of the philosophy of knowledge: the relationship the nature of all scientific achievement” (Wiener between the subject, or observer, and the object, 1948: 162). Cybernetics, which “is nothing if it or what is observed. According to these “new is not mathematical,” would end up by encour- cyberneticians,” Wienerian cybernetics, although aging a fashion, already rampant according to acknowledging that the agent and its environ- Wiener, consisting of the inappropriate use of ment must be viewed as a single system, fails to mathematics in the human sciences: “a flood place sufficient emphasis on the autonomous or of superficial and ill-considered work” (Wiener “autopoietic” nature, to use the expression coined 1964: 88). by and (1987), of this interaction. In this view, reality itself becomes an interactive object, as observer 3.2 Systems theory and second-order and the observed exist in a perpetually unbroken cybernetics circular system. The new cyberneticians thus criti- cize philosophic realism, which they claim was Wiener’s call for caution did not prevent others not completely ruled out by Wienerian cyber- from transferring the fundamental concepts of netics, and in fact is a distinctive feature of cybernetics to wider-ranging, different interdis- symbolic AI because of its representational view ciplinary projects. The project for a “general sys- of mind. These authors consider the activity of tem theory,” initially proposed by the biologist knowing not as an act of duplicating or replicat- , is a good example. ing, through internal (symbolic) representations, Bertalanffy, while emphasizing the interdiscip- what is supposed to be already in the outside linary nature of the cybernetic approach, also world, but as a process built up by the observer. argued against what he believed to be its limits. They want to break free from what they claim His approach was not based on a homeostatic to be the scientific-philosophic “dogma” par system that can be described in terms of feed- excellence, that is, that the aim of science should back control, but on a system that exchanges be to approach as closely as possible a fully matter and energy with the environment, the preconstituted reality alleged to exist as such, only system that may be defined as thermo- independently of the observer. dynamically open. Moreover, in its more general The criticism of these epistemological claims definition, a system is a complex of elements in has its starting points in ’s dynamic interaction. Bertalanffy’s idea was that “second-order cybernetics” and Silvio Ceccato’s the cybernetic model presupposes this more “operational methodology” (Somenzi 1987). general definition insofar as the feedback occurs Criticisms of this kind also give rise to a re- as a “secondary regulation.” It comes into play appraisal of hermeneutic positions based on the in order to stabilize elements of the system that central role of interpretation and language in are already part of the dynamic interaction knowledge. The outcome is twofold. On the that characterizes the “primary regulation” of a one hand, there is the “radical constructivism” thermodynamically open system such as a living of , according to which it is organism, a social body, a biological species, an the subject S that constructs what S knows, and industrial organization, and so on (Bertalanffy S does so on the basis of S’s own experience – 1968). Ilya Prigogine has further developed this the only “world” in which S is capable of living approach in the study of systems far from equi- (von Glasersfeld 1995). On the other hand, there librium, and by theories studying chaotic systems are more general worldviews (those suggested, and complex dynamic systems (see Chapter 3). for instance, by Winograd & Flores 1986, and

194 Cybernetics above all by Varela, Thompson, & Rosch 1991), Embodied, Situated Agents. Hillsdale, NJ: Erlbaum. in which situated cognition and constructivism, [This essay includes a criticism of the so-called autopoiesis and the hermeneutics of Hans “symbolic” view of intelligence from the view- Gadamer, the philosophy of Martin Heidegger point of the new robotics, of which the author is and are occasionally gath- one of the main proponents.] ered together in a criticism of the alleged, West- Churchland, P. S. 1986. Neurophilosophy: Toward a ern, “scientist” or “rationalist” tradition, variously Unified Science of the Mind-Brain. Cambridge, defined as “Cartesian” or “Leibnizian.” It is still MA: MIT Press. [A book that includes both an unclear whether these positions bring any ad- introduction to the neuroscience addressed to philosophers and a criticism of the main claims vancement in our understanding of cybernetic- of classical Cognitive Science.] related phenomena. On the other hand, many Clark, A. 1997. Being There: Putting Brain, Body, important and legitimate requirements under- and World Together Again. Cambridge, MA: MIT lying these positions seem to be already fulfilled Press. [Debates the main claims of situated cogni- by the very tradition that they are challenging, tion and gives a well-balanced critical judgment.] whenever the latter is not caricatured or over- Craik, K. J. W. 1943. The Nature of Explanation. simplified (see, for example, Vera & Simon 1993). Cambridge: Cambridge University Press. [A book that foreruns several issues of both cybernetics and cognitive science.] Acknowledgments Fearing, F. 1930. Reflex Action: A Study in the History of Physiological Explanation. Cambridge, I am grateful to Giuseppe Trautteur and to MA: MIT Press. [This book is still an excellent Luciano Floridi for their comments on a previ- introduction to the historical and the epistemo- ous version of this paper. logical issues of mechanism in neurophysiology.] Feigl, Herbert. 1967. The “Mental” and the “Physical”: The Essay and a Postscript. Minneapolis: References University of Minnesota Press. Hebb, D. O. 1949. The Organization of Behavior. Anderson, J. A. and Rosenfeld, E., eds. 1988. New York and London: Wiley and Chapman, Neurocomputing. Cambridge, MA: MIT Press. New York and London. [The book that is cur- [This book collects classical works in the history rently the reference text of new connectionists.] of neural networks, from those by McCulloch Hook, S., ed. 1960. Dimensions of Mind: A Sym- & Pitts, Hebb, and Rosenblatt, to those by posium. New York: New York University Press. Anderson, Caianiello, Kohonen, Grossberg, [A reading including classical articles by Feigl, Hopfield, Rumelhart, and others. A second Putnam, McCulloch, and others that demonstrate volume was published in 1990.] the influence of cybernetics and Turing-machine Ashby, W. R. 1945. “The physical origin of functionalism on the debate regarding the mind– adaptation by trial and error.” Journal of General body problem.] Psychology 32: 13–25. [One of the clearest state- Maturana, H. and Varela, F. J. 1987. The Tree ment of the newborn cybernetics.] of Knowledge: The Biological Roots of Human ——. 1952. Design for a Brain. London: Chapman Understanding. Boston: New Science Library. and Hall; 2nd ed. New York: Wiley, 1960. [This [Maturana, a pioneering cybernetician, develops book is usually considered a classic of cyber- with Varela the notion of “autopoietic circle.”] netics. It synthesizes Ashby’s research on the Mayr, O. 1970. The Origins of Feedback Control. physical bases of adaptation and learning and on Cambridge, MA: MIT Press. [A historical survey the concept of self-organization.] of several feedback control systems before the Bertalanffy, L. von 1968. General System Theory. advent of cybernetics.] New York: Braziller. [The reference work on sys- McDougall, W. 1911. Body and Mind: A History tem theory.] and a Defense of Animism. London: Methuen. Brooks, R. A. 1995. “Intelligence without reason.” [A passionate defense of and a criticism In L. Steels and R. Brooks, eds., The Artificial of the different mechanistic solutions of the mind– Life Route to Artificial Intelligence: Building body problem.]

195 Roberto Cordeschi

McFarland, D. J. 1971. Feedback Mechanism in Experience. Cambridge, MA; MIT Press. Animal Behavior. New York: Academic Press. [An [Heidegger and Buddha against Descartes and approach to the life sciences based on feedback Leibniz. A criticism of classical cognitive science, control models.] partially based on certain claims of cybernetics.] Minsky, M. L. and Papert, S. 1969. Perceptrons. Vera, A. H. and Simon, H. A. 1993. “Situated Cambridge, MA: MIT Press; repr. 1988, with action: a symbolic interpretation.” Cognitive the authors’ Preface and Postfaction. [The classic Science 17: 7–48. [A lively response to recent critical analysis of the limitations of the early criticisms of the so-called “classical paradigm” in Perceptrons.] cognitive science.] Putnam, Hilary. 1960. “Minds and machines.” In von Glasersfeld, E. 1995. Radical Constructivism. Sidney Hook, ed., Dimensions of Mind: A Sym- Brighton, UK: Falmer Press. [The author posium. New York: New York University Press, introduces radical constructivism and examines 1960. the constructivist strand in the history of Pylyshyn, Z. W. 1984. Computation and Cogni- philosophy.] tion: Toward a Foundation for Cognitive Science. Walter, W. G. 1953. The Living Brain. London: Cambridge, MA: MIT Press. [An attempt to give Duckworth. [The book includes a clear descrip- a foundation to classical cognitive science, as tion of the famous electronic “tortoises” within opposed to both behaviorism and connectionism.] the framework of the mechanistic hypotheses of Rosenblueth, A., Wiener, N., and Bigelow, J. 1943. cybernetics.] “Behavior, purpose and teleology.” Philosophy of Wiener, N. 1961 [1948]. Cybernetics, or Control Science 10: 18–24. [The manifesto of the new- and Communication in the Animal and the born cybernetics.] Machine, 2nd ed. Cambridge, MA: MIT Press. Somenzi, V. 1987. “The ‘Italian Operative School.’” [The book that made cybernetics popular, writ- Methodologia 1: 59–66. [This paper analyzes ten by its founder.] Ceccato’s claims, who was, with von Foerster, ——. 1964. God & Golem, Inc. Cambridge, MA: one of the first critics of realism in the epistemo- MIT Press. [A clear exposition of the ideas of logy of cybernetics.] cybernetic, of its hopes and fears.] Taylor, R. 1966. Action and Purpose. Englewood Winograd, T. and Flores, F. 1986. Understanding Cliffs, NJ: Prentice-Hall. [A philosophical criticism Computers and Cognition: A New Foundation for of the mechanistic and cybernetic interpretations Design. Norwood, NJ: Ablex. [A criticism of the of purpose.] “classical paradigm” of cognition, of which one Varela, F. J., Thompson, E., and Rosch, E. 1991. The of the authors (Terry Winograd) has been one of Embodied Mind: Cognitive Science and Human the most authoritative proponent.]

196 Artificial Life

Chapter 15 Artificial Life Mark A. Bedau

Artificial life (also known as “ALife”) is a broad, Artificial life has broad intellectual roots, and interdisciplinary endeavor that studies life and shares many of its central concepts with other, life-like processes through simulation and syn- older disciplines: computer science, cybernetics, thesis. The goals of this activity include modeling biology, complex systems theory, and artificial and even creating life and life-like systems, as intelligence, both symbolic and connectionist (on well as developing practical applications using these topics see Chapter 3, SYSTEM: AN INTRO- and methods taken from living sys- DUCTION TO SYSTEMS SCIENCE, Chapter 9, THE tems. Artificial life both illuminates traditional PHILOSOPHY OF AI AND ITS CRITIQUE, and Chap- philosophical questions and raises new philo- ter 14, CYBERNETICS). sophical questions. Since both artificial life and John von Neumann (1966) implemented the philosophy investigate the essential nature of first artificial-life model (without referring to certain fundamental aspects of reality like life it as such), with his famous creation of a self- and adaptation, artificial life offers philosophy a reproducing, computation-universal entity, using new perspective on these phenomena. This chap- cellular automata. Von Neumann was trying to ter provides an introduction to current research understand some of the fundamental properties in artificial life and explains its philosophical of living systems, such as self-reproduction and implications. the evolution of complex adaptive structures. His approach was to construct simple formal systems that exhibited those properties. This constructive and abstract methodology typifies contemporary The Roots of Artificial Life artificial life, and cellular automata are still widely used in the field. The phrase “artificial life” was coined by Chris- At about the same time, cybernetics (Wiener topher Langton. He envisioned a study of life as 1948) applied two new tools to the study of it could be in any possible setting, and he organ- living systems: information theory and the ana- ized the first conference that explicitly recog- lysis of self-regulatory processes (homeostasis). nized this field (Langton 1989). There has since One of the characteristics of living systems is been a regular series of conferences on artificial their spontaneous self-regulation: their capacity life and a number of academic journals have been to maintain an internal equilibrium in the face launched to publish work in this new field. of changes in the external environment. This

197 Mark A. Bedau capacity is still a subject of investigation in artifi- The Methodology of Artificial Life cial life. Information theory concerns the trans- mission of signals independently of their physical representation. The abstract and material- The computer-model methodology of artificial independent approach of information theory life has several virtues. The discipline of express- is characteristic of artificial life. ing a model in feasible computer code requires Biology’s contribution to artificial life includes precision and clarity. It also ensures that hypothes- a wealth of information about the life forms ized mechanisms are feasible. Computer models found on Earth. Artificial life seeks to under- also facilitate the level of abstraction required of stand all forms of life that could exist anywhere maximally general models of phenomena. The in the universe, and detailed information about bottom-up architecture of artificial-life models life on Earth is one good clue about this. Bio- creates an additional virtue. Allowing microlevel logy has also provided artificial life with models entities continually to affect the context of their that were originally devised to study a specific own behavior introduces a realistic complexity biological phenomenon. For example, random that is missing from analytically studied math- Boolean networks (discussed below), which were ematical models. Analytically solvable mathemat- originally devised by as models ical models can reveal little about the global effects of gene regulation networks, are now a para- that emerge from a web of simultaneous non- digm of artificial-life research. linear interactions. The obvious way to study the Physics and mathematics have also had a strong effects of these interactions is to build bottom- influence on artificial life. One example is the up models and then empirically investigate their study of cellular automata as exemplars of com- emergent global behavior through computer plex systems (Wolfram 1994). In addition, artifi- simulations. cial life’s methodology of studying model systems There is an important difference between the that are simple enough to have broad generality modeling strategies AI and ALife typically em- and to permit quantitative analysis was pioneered ploy. Most traditional AI models are top-down- in statistical mechanics and dynamical systems. specified serial systems involving a complicated, For example, the Ising model consists of a lattice centralized controller that makes decisions based of up and down “spins” that have simple local on access to all aspects of global state. The con- interactions and that are randomly perturbed by troller’s decisions have the potential to affect “thermal” fluctuations. This model is so abstract directly any aspect of the whole system. On the that it contains almost none of the detailed other hand, many natural living systems exhibit- internal physical structure of such materials as a ing complex autonomous behavior are parallel, cup of water or a bar of iron. Nevertheless, the distributed networks of relatively simple low-level model provides a precise quantitative description “agents” that simultaneously interact with each of how liquid water turns into water vapor or a other. Each agent’s decisions are based on infor- bar of iron loses its magnetization as temperature mation about only its own local situation, and rises. its decisions directly affect only its own local Artificial life also has deep roots in artificial situation. ALife’s models characteristically follow intelligence (AI). Living and flourishing in a nature’s example. The models themselves are changing and uncertain environment seems to bottom-up-specified parallel systems of simple require at least rudimentary forms of intelligence. agents interacting locally. The models are Thus, the subject-matter of artificial life and AI repeatedly iterated and the resulting global overlap. Their methodology is also similar, since behavior is observed. Such lower-level models both study natural phenomena by building com- are sometimes said to be “agent-based” or putational models. The computational meth- “individual-based.” The whole system’s behavior odology of artificial life is especially close to the is represented only indirectly. It arises out of the connectionist movement that has recently swept interactions of a collection of directly represented through AI and cognitive science. parts (“agents” or “individuals”). Two ALife

198 Artificial Life models illustrating this pattern are described in agents’ output is to perform actions in their this section below. environment, and those actions have direct The parallel, distributed character of ALife consequences for the agents’ well-being. Thus models is similar to the structure of the models their output has an intrinsic meaning regard- studied in the connectionist (parallel distributed less of human interpretation. processing, neural network) movement. Both • Third, artificial life and connectionism typic- involve bottom-up models in which a population ally seek different kinds of dynamical beha- of autonomous agents follows simple local rules. vior. Much connectionist modeling aims to In fact, the agents in many artificial-life models produce behavior that settles into an equi- are themselves controlled by internal connectionist librium. This is because both learning and nets. But there are at least three important dif- applying knowledge are conceived as fixed ferences between typical artificial-life models and and determinate goals. By contrast, artificial the connectionist models that have attracted the life views much of the distinctive behavior most attention, such as feedforward networks of living systems as a process of continual that learn by the back-propagation algorithm. creative evolution, so the aim of many ALife models is an open-ended evolutionary dy- • First, artificial life and connectionism depend namic that is forever far from equilibrium. on different kinds of learning algorithms. Con- nectionist models often employ supervised The biological world is often viewed as a learning algorithms like back-propagation. nested hierarchy of levels. These levels include These learning algorithms are typically turned (among other things) chemicals, organelles, cells, on when the network is learning and then organs, organisms, and ecologies. Artificial-life turned off when the acquired information is models usually explicitly represent one level with applied. This distinction between training and the aim of generating the characteristic phenom- application phases is sometimes unnatural. ena of a higher level. One of the ambitious goals In addition, supervised learning algorithms of artificial life is the search for a single model require an omniscient teacher, which is also that generates the behavior of all these levels often unnatural. By contrast, the learning from the explicit specification of only the lowest algorithms employed in artificial-life models level. So far, the field has had difficulty produc- usually avoid these criticisms. They are ing a model that generates even two levels of typically unsupervised and in continual opera- emergent phenomena. tion. Often the algorithm is simply natural The most primitive phenomenon explored by selection. some artificial-life models is self-organization. • Second, human intervention and interpreta- Such models study how structure can emerge tion play different roles in artificial life and from unstructured ensembles of initial condi- connectionism. Typical connectionist models tions, such as models of chemical soups in which passively receive sensory information prepack- fundamental structures such as self-maintaining aged by a human designer and produce out- autocatalytic networks might be seen to emerge. put that must be interpreted by a human A host of models target the organismic level, designer. In artificial-life models, on the other sometimes with significant interactions between hand, a microlevel agent’s sensory input comes organisms. These models typically allow changes directly from the environment in which the in the organisms as part of the system’s dy- agent lives. In many cases, this environment namics (e.g., through a genetic mechanism). The is itself part of the computer model. A most common goal of research using these human designer originally creates the model, models is to identify and elucidate structure that of course, but the specific way it impinges emerges in the ensuing evolutionary process. on the agents is typically the result of an un- Some models fit in between the chemical level predictable collection of low-level interactions and the organismic level, aiming to understand in the model. In ALife models the microlevel development by modeling interacting cells. Other

199 Mark A. Bedau models are interorganismic, in the sense that memory and consuming CPU time as a resource. they aim explicitly to model interactions between A Tierran genotype consists of a string of machine different types of organisms or agents. These code, and each Tierran creature is a token of a models often contain elements of game theory. Tierran genotype. A simulation starts when com- Many artificial-life models are designed not to puter memory is inoculated with a single self- represent known biological systems but to gen- replicating program, the ancestor, which is then erate wholly new and extremely simple instances left to self-replicate on its own. The ancestor of life-like phenomena. The simplest example of and its descendants repeatedly replicate, until the such a system is the so-called “Game of Life,” available memory space is teeming with cre- devised by the mathematician John Conway in atures that all share the same ancestral genotype. the 1960s (Berlekamp et al. 1982). Conway’s To create space in memory for new descendants, Game of Life can be thought of as a model at older creatures are continually removed from the the physical or chemical level, embodying an system. Errors (mutations) sometimes occur when extremely simple and unique form of “chem- a creature replicates, so the population of Tierra ical” interactions. However, the self-organization creatures evolves by natural selection. If a muta- exhibited in the Game of Life is not a repres- tion allows a creature to replicate faster, that entation of chemical self-organization in the genotype tends to take over the population. Over real world but a wholly new instance of this time, the ecology of Tierran genotypes becomes phenomenon. The Game of Life is a two-state, remarkably diverse. Quickly reproducing parasites two-dimensional cellular automaton with a trivial that exploit a host’s genetic code evolve, and nearest-neighbor rule. Think of this “game” as this prompts the evolution of new creatures that taking place on a two-dimensional rectangular resist the parasites. After millions of CPU cycles, grid of cells, analogous to a huge checker-board. Tierra typically contains many kinds of creatures Time advances in discrete steps, and a cell’s state exhibiting a variety of competitive and cooperat- at a given time is determined by the states of its ive ecological relationships. 8 neighboring cells according to the following Computer simulation is crucial for the study simple “birth–death” rule: a “dead” cell becomes of complex adaptive systems. It plays the role that “alive” if and only if exactly 3 neighbors are just observation and experiment play in more con- “alive,” and a “living” cell “dies” if and only if ventional science. The complex self-organizing fewer than 2 or more than 3 neighbors are just behavior of the Game of Life would never have “alive.” From inspection of the birth–death rule, been discovered without simulating thousands nothing particular can be discerned regarding of generations for millions of sites. Similarly, it how the whole system will behave. But when would have been impossible to discover the emer- the system is simulated, a rich variety of complic- gence of complex ecological interactions in Tierra ated dynamics can be observed and a complex without simulating many millions of generations. zoo of structures can be identified and classified Simulation of large-scale complex systems is the (blinkers, gliders, glider guns, logic switching single most crucial development that has enabled circuits, etc.). It is even possible to construct a the field of artificial life to flourish and distinguish universal Turing machine in the Game of Life, itself from precursors such as cybernetics. by cunningly positioning the initial configura- Rather than merely producing computer tion of living cells. In such constructions gliders simulations, some artificial-life research aims to perform a role of passing signals. Analyzing the implement systems in the real world. The pro- computational potential of cellular automata on ducts of this activity are physical devices such as the basis of glider interactions has become a robots that exhibit characteristic life-like behavior. major research thrust. Some of these implementations are motivated An example of an organismic level artificial- by the concern to engineer practical devices that life system is Tierra (Ray 1992). This ALife have some of the useful features of living systems, system consists of “organisms” that are actually such as robustness, flexibility, and autonomy. But simple, self-replicating computer programs popu- some of this activity is primarily theoretical, mot- lating an environment consisting of computer ivated by the belief that the best way to confront

200 Artificial Life the hard questions about how life occurs in the gent property. So the first step toward solving physical world is to study real physical systems. the problem of emergence is to explain the notion Again, there is an analogy with biological levels. of an emergent property. There are three main The “chemical” level is represented by work on views of what an emergent property is. evolvable hardware, often using programmable According to the first view, emergent proper- logic arrays, which attempts to use biologically ties apply only to “wholes” or “totalities,” not inspired adaptive processes to shape the config- to their component “parts” considered in isola- uration of microelectronic circuitry. The “organ- tion (e.g., Harré 1985, Baas 1994). For example, ismic” level is represented by new directions the constituent molecules in a cup of water, con- in biologically inspired robotics, such as using sidered individually, do not have properties like evolutionary algorithms to automate the design fluidity or transparency, though these properties of robotic controllers. A swarm of robots com- do apply to the whole cup of water. The “wholes” municating locally to achieve some collective goal at one level of analysis are sometimes “parts” of is an example at the “population” level. An “eco- a larger “whole” at a higher level of analysis, so logical” level might be represented by the internet a hierarchy can contain successive levels of this along with its interactions with all its users on sort of emergence. This view easily explains the computers distributed around the world. two hallmarks of emergence. Macrolevel emer- gent phenomena are constituted from and gen- erated by microlevel phenomena in the trivial sense that wholes are constituted and generated Emergence by their constituents; and emergent phenomena are autonomous from underlying phenomena in Both living systems and artificial-life models are the straightforward sense that emergent proper- commonly said to exhibit emergent phenomena; ties do not apply to the underlying entities. This indeed, many consider emergence to be a defin- notion of emergence is very broad, applies to a ing feature of life. However, the notion of emer- large number of intuitive examples of emergent gence remains ill defined. In general, emergent phenomena, and corresponds to the compelling phenomena share two broad hallmarks: they are picture of reality consisting of autonomous levels constituted by and generated from underlying of phenomena. Its breadth is its greatest weak- phenomena, and yet they are also autonomous ness, however, for it applies to all macroproperties from those underlying phenomena. There are that are not possessed by microentities. Macro- abundant examples of apparent emergent phe- properties are usually classified into two kinds: nomena, and most involve life or mind. Yet the genuine emergent properties and mere “result- two hallmarks of emergence seem inconsistent ant” properties. Resultant properties are those or metaphysically illegitimate: How can some- that can be predicted and explained from the thing be autonomous from underlying phenom- properties of the components. For example, a ena if it is constituted by and generated from circle consists of a collection of points, and the them? This is the problem of emergence. A individual points have no shape. So being a solution would both dissolve the appearance of circle is a property of a “whole” but not its illegitimate metaphysics and enfold emergence constituent “parts.” Thus being a circle is an in constructive scientific explanations of phenom- emergent property according to the first view. ena involving life and mind. However, if you know that all the points in a One can distinguish emergent properties, geometrical figure are equidistant from a given emergent entities, and emergent phenomena. point, then you can conclude that the figure is a Being alive, for example, is an emergent property, circle. So being a circle is a resultant property. an organism is an emergent entity, and the life To distinguish emergent from resultant proper- history of an organism is an emergent phenom- ties one must turn to other views. enon. An entity with an emergent property is an The second main view construes emergent emergent entity, and an emergent phenomenon properties as supervenient properties with causal involves an emergent entity possessing an emer- powers that are irreducible to the causal powers

201 Mark A. Bedau of microlevel constituents (e.g., Kim 1999). On of emergent phenomena to the complex con- this view, supervenience explains the sense in sequences of myriad, nonlinear, and context- which the underlying processes constitute and dependent local microlevel interactions. Emergent generate the emergent phenomena, and irreduc- phenomena can have causal powers on this view, ible causal powers explain the sense in which but only by means of aggregating microlevel they are autonomous from underlying phenom- causal powers. There is nothing inconsistent or ena. These irreducible causal powers give emer- metaphysically illegitimate about underlying pro- gent properties a dramatic form of ontological cesses constituting and generating phenomena by novelty that many people associate with the most iteration and aggregation. Furthermore, this form puzzling kinds of emergent phenomena, such of emergence is prominent in scientific accounts as consciousness. However, an irreducible but of exactly the natural phenomena like life and supervenient causal power by definition cannot mind that apparently involve emergence. How- be explained in terms of the aggregation of the ever, this form of emergence sheds no light microlevel potentialities. No evident mechanism on those mysterious emergent phenomena, like explains these irreducible supervenient powers, consciousness, that science still cannot explain. so they must be viewed as primitive or “brute” In addition, the autonomy of these kinds of facts of nature. In addition, this strong form of emergent phenomena seems to be merely epi- emergence seems to be scientifically irrelevant. stemological rather than ontological. Emergent Illustrations of it in recent scientific literature phenomena are epistemologically autonomous in almost universally focus on one isolated example: the sense that knowledge of the underlying phe- Sperry’s explanation of consciousness from over nomena does not provide knowledge about the 30 years ago (Sperry 1969). There is little if any emergent phenomena. However, metaphysically, evidence that this form of emergence is empiric- the emergent phenomena seem wholly depend- ally relevant in the sciences studying emergent ent on the constituent phenomena, since emer- phenomena. gent causal powers result from microlevel causal A third view of emergence is poised midway powers. This will not satisfy those who think between the first two. It refers to the resultant emergent phenomena have a strong form of aggregate global behavior of complex systems. ontological autonomy. In this sense, a system’s macrostate is emergent Artificial life can be expected to play an active just in case it can be derived from the system’s role in the future philosophical debate about boundary conditions and its microlevel dynam- emergence and related notions like supervenience, ical process but only through the process of iter- reduction, complexity, and hierarchy. Living sys- ating and aggregating all the microlevel effects tems are one of the primary sources of emergent (e.g., Bedau 1997a). In this case, the microlevel phenomena, and artificial life’s bottom-up models phenomena clearly constitute and generate the generate impressive macrolevel phenomena wholly macrolevel phenomena. At the same time, the out of microlevel interactions. Exploration and macrolevel phenomena are autonomous in that modification of these models is a constructive the only way to recognize or predict them is by way to analyze the nature and causes of different empirically observing the macrolevel effect of kinds of emergent phenomena. aggregating all the microlevel phenomena. In effect, this view identifies emergent properties with a special subset of resultant properties: those Adaptationism that cannot be predicted or explained except by empirically aggregating the interactions among microlevel entities. This form of emergence is Adaptive evolutionary explanations are familiar common in complex systems found in nature. from high-school biology. It is a cliché to explain Artificial life’s models also exhibit it, since their the giraffe’s long neck as an adaptation for brows- bottom-up behavior consists of the continual ing among the tops of trees, on the grounds iteration of microlevel interactions. This view attri- that natural selection favored longer-necked butes the unpredictability and unexplainability giraffes over their shorter-necked cousins. But the

202 Artificial Life scientific legitimacy of these adaptive explanations trait is an adaptation. Different specific adaptive is controversial, largely because of a classic paper hypotheses usually have different observable con- by Stephen Jay Gould and Richard Lewontin sequences. A general adaptive hypothesis entails (1979). Gould and Lewontin directly challenge that some specific adaptive hypothesis is true, but adaptationism: the thesis that the activity of it gives no indication which one is true. So the pursuing adaptive explanations of biological traits general adaptive hypothesis makes no particular is a legitimate part of empirical science. They empirical prediction. Dawkins admits that general accept that adaptive explanations are appropriate adaptive hypotheses cannot be tested. “It is true in some contexts, but they despair of identifying that the one hypothesis that we shall never test those contexts in any principled and rigorous is the hypothesis of no adaptive function at all, way. Biology provides many alternatives to adapt- but only because that is the one hypothesis in ive explanations, such as explanations appealing this whole area that really is untestable” (1983: to allometry, genetic drift, developmental con- 361). Dawkins can defend the appeal to adaptive straints, genetic linkage, epistasis, and pleiotropy. explanations when a specific adaptive hypothesis But Gould and Lewontin complain that those has been corroborated. But in the absence of this alternatives receive only lip-service. The pre- – which is the typical situation – Dawkins must supposition that a trait is an adaptation and concede Gould’s and Lewontin’s fundamental so deserves an adaptive explanation is treated challenge. as untestable. The fundamental challenge for Artificial life has been used to develop and adaptationism raised by Gould and Lewontin, illustrate a new defense of adaptationism. It is then, is to find some empirical method for test- argued that it is possible to test general adaptive ing when an adaptive explanation is needed. This hypotheses empirically, by recording and analyz- problem is often especially acute in artificial life. ing so-called “evolutionary activity” information Those studying artificial models have the luxury collected from the evolving system (Bedau 1996, of being able to collect virtually complete data, Bedau & Brown 1999). The fundamental intui- but this mass of information only compounds tion behind this method is that we can detect the problem of identifying which evolutionary whether an item (gene, gene complex, genotype, changes are adaptations. etc.) is an adaptation by observing the extent to The canonical response to Gould and which it persists in the face of selection pressures. Lewontin makes two claims. The first claim is Whenever an item that is subject to heritable that specific adaptive hypotheses, hypotheses variation is “active” or expressed, natural selection about the specific nature of a character’s adapta- has an opportunity to provide feedback about its tion, are testable. Second, although the general adaptive value, its costs and benefits. If it per- hypothesis that a trait is an adaptation might sists and spreads through a population when it itself not be testable, it is a working hypothesis is repeatedly active, and especially if it exhibits and empirical science normally treats working significantly more activity than one would expect hypotheses as untestable. For example, Richard to see if it had no adaptive value, then we have Dawkins claims that “hypotheses about adapta- positive evidence that the item is persisting tion have shown themselves in practice, over and because of its adaptive value. This means that we over again, to be easily testable, by ordinary, have positive evidence that it is an adaptation mundane methods of science” (Dawkins 1983: and deserves an adaptive explanation, even if we 360ff). Dawkins’s point is that specific adaptive have no idea about its specific adaptive function. hypotheses have observable consequences that Since natural selection is not instantaneous, can be checked. The canonical response reflects maladaptive items persist for a while before they and explains evolutionary biology’s emphasis on are driven out by natural selection. Adaptations formulating and testing specific adaptive hypo- are distinguished by accruing much more activity theses. But this response does not address the than would be expected in a nonadaptive item. fundamental challenge to adaptationism, for that A general way to measure the activity expected challenge is about the of general adapt- of nonadaptive items is to construct a “neutral ive hypotheses, hypotheses to the effect that a shadow” of the target system – that is, a system

203 Mark A. Bedau that is similar to the target in all relevant re- life models he studies are abstract enough to apply spects except that none of the items in it have to a wide variety of contexts. Random Boolean any adaptive significance. The activity in the neut- networks are one such class of models. These ral shadow is a no-adaptation null hypothesis consist of a finite collection of binary (ON, OFF) for the target system. If the target system shows variables with randomly chosen input and out- significantly more activity than the neutral put connections. The state of each variable at shadow, this excess activity must be due to nat- each step in discrete time is governed by some ural selection and the target system must contain logical or Boolean function (AND, OR, etc.) of adaptations. The evolutionary-activity method the states of variables that provide input to it. responds directly to Gould and Lewontin. It pro- The network is started by randomly assigning vides an empirical method for determining when states to each variable, and then the connections evolution is creating adaptations. Rather than and functions in the network determine the suc- just assuming that traits are adaptations, it puts cessive state of each variable. Since the network this assumption to the empirical test. Another is finite, it eventually reaches a state it has previ- advantage of the activity method is that stat- ously encountered, and from then on the network istics based on activity information can be used will forever repeat the same cycle of states. Dif- to measure various aspects of the dynamics of ferent network states can end up in the same adaptive evolution, thus allowing the process of state cycle, so a state cycle is called an attractor. adaptation in different systems to be classified Kauffman found that the number of variables in and quantitatively compared (Bedau et al. 1997, the network, the number of connections between Bedau et al. 1998). One weakness of the evolu- the variables, and the character of the Boolean tionary activity method is that practical problems functions determine many biologically crucial sometimes make activity data difficult to collect. properties of the networks. These properties Another weakness is that genetic hitchhikers – include the number and length of attractors, the nonadaptive or maladaptive traits that persist stability of attractors to perturbation and muta- because of a genetic connection to an adaptive tion, etc. If the variables are highly connected, trait – can accumulate more activity than expected then the network’s attractors contain so many in a neutral shadow. Thus, a trait that is not an states that the time it takes to traverse the adaptation can have significant excess activity if attractor vastly exceeds the lifetime of the entire it is connected to a trait that is an adaptation. universe. Furthermore, any perturbation or muta- Significant excess activity in a cluster of traits tion in the network causes a vast change in its shows that there are adaptations in the cluster, behavior. For all practical purposes, the network but it does not separate out the hitchhikers. behaves chaotically. The network acts differently The adaptationist perspective on evolution when each variable takes input from only a bio- emphasizes natural selection’s role in creating logically plausible number of other variables and the complex adaptive structures found in living when the variables are governed by biologically systems. Artificial life has been the source of a realistic Boolean functions. In this case, the net- new and fundamental challenge to this whole work has a tiny number of attractors, it main- perspective. Stuart Kauffman (1993, 1995) has tains homeostatic stability when perturbed, and used artificial-life models to show that many fea- mutations have limited consequences; in other tures of metabolisms, genetic networks, immune words it exhibits “order for free.” Furthermore, systems, and ecological communities should be these biologically realistic Boolean networks viewed not as the products of selection but largely explain a number of empirically observed features as the spontaneous, self-organized behaviors of of biological systems, such as how the number certain abstract complex systems. Kauffman also of different cell types and cell replication times argues that spontaneous self-organized structures vary as a function of the number of genes per – what he calls “order for free” (Kauffman 1995) cell. Kauffman’s nonadaptationist explanations – explain both life’s origin and its subsequent of the origins of order are controversial, partly ability to evolve. Kauffman can make sweeping because of the sweeping scope of his analysis. claims about order for free because the artificial- But the suggestion that self-organization rather

204 Artificial Life than natural selection can explain much of the Stephen Jay Gould (1989) devised an ideal structure in living systems is plausible. The issue way to address this issue, namely the thought is not whether self-organization explains struc- experiment of replaying the tape of life. Imagine ture, but how much. that the process of evolution left a record on The problem of adaptationism is as acute in a tape. Gould’s thought experiment consists in artificial life as it is in biology. Artificial life can rewinding the evolutionary process backward make a distinctive contribution to the debate, in time and then replaying it again forward in for the evolutionary processes studied by artifi- time but allowing different accidents, different cial life provide many diverse examples of the contingencies to reshape the evolution of life. process of adaptation. Furthermore, the systems The evolution of life is rife with contingencies. can be analyzed with the kind of detail and rigor Repeatedly replaying the tape of life with novel that is simply impossible to achieve in the bio- contingencies could produce as large a sample sphere, because the historical data are unavail- of evolutionary histories as desired. It would be able or impractical to examine. For analogous relatively straightforward to determine whether reasons, we can expect artificial life to contribute a general pattern emerges when all the evolu- to our understanding of many other funda- tionary trajectories are compared. mental issues in the , such There is substantial controversy about the as the nature of functions, the nature of species, outcome of Gould’s thought experiment. Gould whether and how selection operates at different himself suggests that “any replay of the tape biological levels, the nature of the niche, and would lead evolution down a pathway radically the nature of the relationship between organisms different from the road actually taken” (1989: and their environment. 51). He concludes that the contingency of evolution will debar general laws like the hypo- thesized ladder of complexity. Daniel Dennett (1995) draws exactly the opposite conclusion. Evolutionary Progress Dennett argues that certain complex features like sophisticated sensory processing provide a The evolution of life shows a remarkable growth distinct adaptive advantage. Thus, natural selec- in complexity. Simple prokaryotic one-celled life tion will almost inevitably discover significantly led to more complex eukaryotic single-celled advantageous features that are accessible from life, which then led to multicellular life, then to multiple evolutionary pathways. Examples of mul- large-bodied vertebrate creatures with complex tiple independent evolutionary convergence, such sensory processing capacities, and ultimately to as flight and eyesight, illustrate this argument. highly intelligent creatures that use language and Dennett concludes that replaying life’s tape will develop sophisticated technology. This illustra- almost inevitably produce highly intelligent cre- tion of evolution’s creative potential has led some atures that use language and develop sophistic- to propose a ladder-of-complexity hypothesis ated technology. according to which open-ended evolutionary Artificial life can make a number of contribu- processes have an inherent, law-like tendency to tions to this debate. Experience in artificial life create creatures with increasingly complicated has shown time and again that expectations about adaptive structure. But the evolution of life is the outcome of thought experiments like replay- equally consistent with the denial of the ladder ing life’s tape are highly fallible. The only sure of complexity. The observed progression could way to determine what to expect is to create the be a contingent result of evolution rather than a relevant model and observe the results of re- reflection of any inherent tendency. The ladder- peated simulation. In fact, artificial life is exactly of-complexity hypothesis is difficult to test where this sort of modeling activity occurs. A because we do not have a variety of different central goal of artificial life is to discover the histories of life to compare. A sample size of one inherent trends in evolving systems by devising makes it difficult to distinguish inherent trends a model of open-ended evolution, repeatedly from artifacts. replaying life’s tape with different historical

205 Mark A. Bedau contingencies and searching for patterns that hold 1 Living systems have an enormously complex across all the results. The best evidence in favor and adaptive organization. of the ladder-of-complexity hypothesis would 2 Organisms are composed of a chemically come from showing that a tendency toward in- unique set of macromolecules. creasing adaptive complexity is the in such 3 Living phenomena are predominantly qual- ALife models. However, no one has yet con- itative, not quantitative. ducted the experiment of replaying life’s tape, 4 Living systems consist of highly variable because no one has yet been able to create a groups of unique individuals. system that exhibits continual open-ended evolu- 5 Organisms engage in purposeful activities by tion of adaptive complexity. Achieving this goal means of evolved genetic programs. is one of the key open problems in artificial life 6 Classes of organisms have historical connec- (Bedau et al. 2000). All conjectures about the tions of common descent. ladder of complexity will remain unsettled until 7 Organisms are the product of natural one can actually replay the tape of life. selection. 8 Biological processes are especially unpre- dictable. The Nature of Life Cluster conceptions of life account for the char- acteristic hallmarks of life, although they do this Philosophy traditionally addressed the nature of merely by fiat. Lists like Mayr’s raise rather than life but most philosophers ignore the issue today, answer the question why this striking collection perhaps because it seems too “scientific.” At the of features is present in an indefinite diversity of same time, most biologists also ignore the issue, natural phenomena. The main drawback of all perhaps because it seems too “philosophical.” The cluster conceptions is that they inevitably make advent of artificial life raises the question anew, life seem rather arbitrary or mysterious. A clus- for two reasons. Modeling the fundamental fea- ter conception cannot explain why any particu- tures of living systems presupposes an under- lar cluster of properties is a fundamental and standing of life, and new artificial-life systems ubiquitous natural phenomenon. push the boundaries of what life could be. Schrödinger illustrated the second view of life There are three prominent views about the when he proposed persistence in the face of the nature of life: life as a cluster of properties, life second law of thermodynamics by means of the as metabolization, and life as evolution. The process of metabolization as the defining feature cluster conception takes two forms, depending of life. on whether the properties in the cluster are taken to be individually necessary and jointly sufficient It is by avoiding the rapid decay into the inert for life. Skeptics argue that life is characterized state of “equilibrium” that an organism appears merely by a loose cluster of properties typically so enigmatic; . . . How does the living organism but not necessarily possessed by living entities. avoid decay? The obvious answer is: By eating, This view treats something as alive if it possesses drinking, breathing and (in the case of plants) a sufficient number of properties in the cluster, assimilating. The technical term is metabolism. but no precise number of properties is sufficient. (Schrödinger 1969: 75) On this view, the diversity of living forms have only a family resemblance. Viewing life as a loose Living systems need some way to self-maintain cluster of properties provides a natural explanation their complex internal structure. So metaboliza- of why life has vague boundaries and borderline tion seems to be at least a necessary condition of cases. Life is also sometimes characterized by a all physical forms of life. The view that life cent- list of properties intended to provide something rally involves the process of metabolization also much closer to individually necessary and jointly nicely explains our that a crystal is not sufficient conditions. Ernst Mayr (1982) pro- alive. There is a metabolic flux of molecules only duced a comprehensive list of such properties: at the crystal’s edge, not inside it. One drawback

206 Artificial Life of metabolization as an all-encompassing con- are two main objections to viewing life as evolu- ception of life is that many metabolizing entities tion. The first is that it seems to be entirely seem not to be alive and not to involve life in contingent that life forms were produced by an any way. Standard examples include a candle evolutionary process. The Biblical story of Adam flame, a vortex, and a convection cell. A second and Eve shows that is easy to imagine life forms problem is whether metabolization can explain in the absence of any evolutionary process. A sec- the hallmarks of life (recall Mayr’s list). It is ond objection calls attention to evolving systems doubtful whether metabolization can explain that seem devoid of life. Viruses and prions evolve those characteristics on Mayr’s list that depend but are questionably alive, and cultural evolution on evolution. provides much starker counterexamples. The third main conception of life focuses on The advent of artificial life has revitalized the evolutionary process of adaptation. The investigation into the nature of life. This is partly central idea is that what is distinctive of life is because one can simulate or synthesize living the way in which adaptive evolution automatic- systems only if one has some idea what life ally fashions new and intelligent strategies for essentially is. Artificial life’s self-conscious aim to surviving and flourishing as local environments discern the general nature of life as it could be change. As John Maynard Smith explains: encourages liberal experimentation with novel life-like organizations and processes. Thus, arti- We shall regard as alive any population of ficial life both fosters a broad perspective on life entities which has the properties of multiplica- and has the potential to create radically new forms tion, heredity and variation. The justification of life. In the final analysis, the nature of life will for this definition is as follows: any population be settled by whatever provides the best explana- with these properties will evolve by natural tion of the rich range of natural phenomena selection so as to become better adapted to its that seem to characterize living systems. Better environment. Given time, any degree of adapt- understanding of how to explain these phenom- ive complexity can be generated by natural ena will also help resolve a cluster of puzzles selection. (Maynard Smith 1975: 96ff) about life. These puzzles include whether life admits of degrees, how the notion of life applies The view of life as evolution has two forms. at different levels in the biological hierarchy, Maynard Smith illustrates one form, according whether life is essentially connected with mental to which living systems are the entities in an capacities, and the relationship between the evolving population. Recently, Bedau (1996, material embodiment of life and the dynamical 1998) has argued that, in fact, an evolving sys- processes in those materials. tem itself should be viewed as alive in the primary sense. One virtue of the conception of life as evolution is that it explains why Mayr’s hall- Strong Artificial Life marks of life coexist in nature. We would expect life to involve the operation of natural selection producing complex adaptive organization in Artificial life naturally raises the question whether historically connected organisms with evolved artificial constructions could ever literally be alive. genetic programs. The random variation and his- Agreement about the nature of life would make torical contingency in the evolutionary process this question easier to answer. For example, if explains why living phenomena are especially the defining property of living systems were the qualitative and unpredictable and involve unique process of sustaining a complex internal organ- and variable individuals with frozen accidents like ization through a metabolism, then the issue chemically unique macromolecules. This view can would be whether an artificially created system also explain why metabolism is so important in could literally exhibit this property (see Boden living systems, for a metabolism is a physically 1999 for discussion). But the debate over creat- necessary prerequisite in any system that can sus- ing real but artificial life currently proceeds in tain itself long enough to adapt and evolve. There the absence of agreement about what life is.

207 Mark A. Bedau

It is important to distinguish two questions physical process. Furthermore, as emphasized about creating artificial life. The first concerns earlier, artificial-life models are often intended whether it is possible to create a physical device not as simulations or models of some real-world such as a robot that is literally alive. Aside from living system but as novel examples of living controversy about what life is, the challenge here systems. Conway’s Game of Life (Berlekamp is less philosophical than scientific. It concerns et al. 1982), for example, is not a simulation or our ability to synthesize the appropriate materials model of any real biochemical system. Rather, and processes. The philosophically controversial it is a simple system that exhibits spontaneous question is whether the processes or entities macroscopic self-organization. Similarly, Ray’s inside a computer that is running an artificial-life Tierra (Ray 1992) is not a simulation or model of model could ever literally be alive. This is the the ecology and evolution of some real biological issue of whether so-called “strong” artificial life is system. Instead, it is an instance of ecological possible. Strong ALife is contrasted with “weak” and evolutionary dynamics in a digital domain. ALife, the uncontroversial thesis that computer So, when the Game of Life and Tierra are actu- models are useful for understanding living systems. ally running in computers, they are new physical The strong ALife question is sometimes put instances of self-organization and evolution. Pro- in terms of computer simulations: can a computer cesses like self-organization and evolution are simulation of a living system ever literally be alive? multiply realizable and can be embodied in a wide This formulation prompts the response (e.g., variety of different media, including the physical Pattee 1989, Harnad 1994) that it is a simple media of suitably programmed computers. So, category mistake to confuse a simulation of some- to the extent that the essential properties of living thing with a realization of it. A flight simula- systems involve processes like self-organization tion for an airplane, no matter how detailed and and evolution, suitably programmed computers realistic, does not really fly. A simulation of a will actually be novel realizations of life. Models hurricane does not create real rain driven by real that merely represent some phenomenon differ gale-force winds. Similarly, a computer simula- from models that actually generate it. For ex- tion of a living system produces merely a symbolic ample, a two-dimensional model of a branching representation of the living system. The intrinsic process with random pruning can be viewed as a ontological status of this symbolic representa- description of the evolution of more or less com- tion is nothing more than certain electronic states plex insects, if one dimension is taken to rep- inside the computer (e.g., patterns of high and resent time and the other is taken to represent low voltages), and this constellation of electronic complexity. But exactly the same branching pro- states is no more alive than is a series of English cess can equally be viewed as a description of the sentences describing an organism. It seems alive evolution of more or less tall humans. It can even only when it is given an appropriate interpreta- be viewed as a description of various nontemporal tion. This interpretation might be fostered if the and nonbiological processes, such as the pattern description dynamically reflects how the living of tributaries in a geography. In itself, the model system changes over time and if the simulation does not intrinsically involve any of these things. produces a vivid life-like visualization, but it is By contrast, a glider in Conway’s Game of Life still only an interpretation. is not an electronic pattern that is merely inter- A number of considerations can blunt this pretable as a self-sustaining dynamic collective. charge of category mistake. It is important to It really is an electronic self-sustaining collective, recognize that an artificial-life model that is whether or not anyone notices it and regards it actually running on a computer consists of a real as such. Likewise, the self-replicating machine- physical process occurring in a real physical language programs in Ray’s Tierra genuinely medium consuming real physical resources. The evolve by natural selection and genuinely en- software specifying the model might be a static gage in host/parasite relations. The nature of abstract entity with the ontological nature of ALife’s key problem of modeling the open-ended a Platonic universal, but an actual simulation evolution of adaptive complexity can be appreci- of the model has the ontological status of any ated in this light. It is easy to make a model that

208 Artificial Life can be interpreted as exhibiting this phenom- entities, metabolization creates a continual flux enon; the challenge is to make a model that of molecules through living systems. Thus, life actually generates it. seems more like a kind of a process than a kind The Turing test in artificial intelligence was of material entity. This implies that life could be an attempt to settle whether computing could realized in a variety of media, perhaps including be indistinguishable from thinking in the absence suitably programmed computer hardware. This of any agreement about the nature of thinking motivation for strong ALife prompts a function- itself. Thus the proposal to settle the strong ALife alist and computationalist view of life, analogous debate with a “Turing test” for life often arises to contemporary functionalism and computation- in artificial life. Some (e.g., Sober 1992) warn alism with respect to mind. Sober (1992) points that the Turing test in AI is an insufficient test out that many essential properties of organisms for intelligence because it is possible in principle involve their interaction with the environment. for an unthinking device to pass the test. A typ- Thus, the computational character of the pro- ical example of such a hypothetical device is a cesses inside organisms would not alone sup- machine that stores an appropriate output for all port functionalism and computationalism about the different input that might be encountered. life. But since many artificial-life models situate The characteristic drawback of such devices is artificial organisms in an artificial environment, that, even to exhibit modest capabilities, the artificial life still promotes functionalism and number of pieces of information they must store computationalism. Bedau (1997b) argues that is larger than the number of elementary particles artificial life’s models generate macrolevel dy- in the entire universe. Though possible in prin- namics with a suppleness that is distinctive of ciple, such a device is clearly impossible in prac- adaptive intelligence and that cannot be captured tice. Artificial life’s computational methodology by any fixed algorithm. The models are imple- demands models that actually produce the phe- mented in a computer but adaptive processes like nomenon of interest. In this context, what is natural selection continually change the micro- possible in principle but impossible in practice level rules that govern the system. Thus, the is irrelevant. So the experience in ALife prompts macrolevel processes that emerge are noncom- one to ignore unfeasible counterexamples to putational. This perspective still supports func- Turing tests. Harnad (1994) has advocated eco- tionalism with respect to life, but a form of logical and evolutionary indistinguishability from functionalism divorced from computationalism. biological life as a Turing test for life. The motiva- Artificial-life models generate behavior that is tion for this test for life is that it would be characteristic of living systems, so the practice of arbitrary to deny life to anything that is indistin- artificial life will continually raise the question guishable ecologically and evolutionarily from whether a computer model of life could literally biological life. But this test is biased against life be alive. By continually challenging the bound- forms that are isolated from the biosphere. Sys- aries between life and nonlife, artificial life will tems existing inside computers running artificial- also spur novel perspectives on the issue. The life models might exhibit all the ecological and debate about strong ALife will also enliven and evolutionary richness found in the biosphere. Yet inform many related issues in the philosophy they might not interact with biological life, so of mind and artificial intelligence, including they might fail Harnad’s test for life. Thus, functionalism, computationalism, intelligence, Harnad’s test begs the question against some intentionality, and representationalism. forms of artificial life. The debate about strong artificial life is inter- twined with philosophical questions about func- Philosophical Methodology tionalism and computation. A significant source of support for strong ALife is the belief that life concerns form more than matter. Although Artificial life also has implications for the meth- certain carbon-based macromolecules play a cru- odology of philosophy. Philosophy and artificial cial role in the vital processes of all known living life are natural partners. Both seek to understand

209 Mark A. Bedau phenomena at a level of generality that is suffi- ——. 1998. “Four puzzles about life.” Artificial ciently deep to ignore contingencies and reveal Life 4: 125–40. [An explanation of how the view essential natures. In addition, artificial life’s com- of life as supple adaptation (Bedau 1996) ex- putational methodology is a direct and natural plains four puzzles about life. For postsecondary extension of philosophy’s traditional methodo- school audiences.] logy of a priori thought experiments. In the —— and Brown, C. T. 1999. “Visualizing evolu- attempt to capture the simple essence of vital tionary activity of genotypes.” Artificial Life 5: processes, artificial-life models abstract away as 17–35. [Shows how evolutionary activity graphs many details of natural living as possible. These reveal the dynamics of adaptive evolution of geno- types in an artificial life model. For postsecondary models are for exploring the consequences of school audiences.] certain simple ideas or premises. They are ——, McCaskill, J. S., Packard, N. H., Rasmussen, “thought experiments” explored with the help S., Adami, C., Green, D. G., Ikegami, T., Kaneko, of a computer. Like the traditional armchair K., and Ray, T. S. 2000. “Open problems in thought experiments employed in philosophy, artificial life.” Artificial Life 6: 363–76. [Describes artificial-life simulations attempt to answer 14 grand challenges in artificial life, each of which “What if X?” questions. Artificial life’s thought requires a major advance on a fundamental issue experiments are distinctive in that they can be for its solution. Intelligible to secondary school explored only by computer simulation; armchair audiences.] analysis is simply inconclusive. Synthesizing ——, Snyder, E., Brown, C. T., and Packard, N. thought experiments on a computer can bring H. 1997. “A comparison of evolutionary activity a new clarity and constructive evidence to bear in artificial evolving systems and the biosphere.” in philosophy (see Chapter 26, COMPUTATIONAL In P. Husbands and I. Harvey, eds., Proceedings MODELING AS A PHILOSOPHICAL METHODOLOGY). of the Fourth European Conference on Artificial Life, ECAL97. Cambridge, MA: MIT Press, pp. 125–34. [Comparison of evolutionary activ- References ity in two artificial life models and in the fossil record reveals qualitative differences. For post- Baas, N. A. 1994. “Emergence, hierarchies, and graduate audiences.] hyperstructures.” In C. G. Langton, ed., Artificial ——, ——, and Packard, N. H. 1998. “A classifica- Life III. Redwood City, CA: Addison-Wesley, tion of long-term evolutionary dynamics.” In C. pp. 515–37. [A mathematical and technical pre- Adami, R. Belew, H. Kitano, and C. Taylor, eds., sentation and illustration of the view of emergent Artificial Life VI. Cambridge, MA: MIT Press, properties as novel macroproperties.] pp. 228–37. [Evolutionary activity is used to Bedau, M. A. 1996. “The nature of life.” In classify qualitatively different kinds of evolving M. Boden, ed., The Philosophy of Artificial Life. systems. For postgraduate audiences.] Oxford: Oxford University Press, pp. 332–57. [A Berlekamp, E. R., Conway, J. H., and Guy, R. K. defense of the view of life as supple adaptation or 1982. Winning Ways for your Mathematical Plays, open-ended evolution, illustrated in artificial life vol. 2: Games in Particular. New York: Academic models. For postsecondary school audiences.] Press. [Chapter 25 is the authoritative descrip- ——. 1997a. “Weak emergence.” Philosophical tion of the Game of Life, including a description Perspectives 11: 375–99. [A defense of emergence of how to embed a universal Turing machine in as complicated iteration and aggregation of the Game of Life. For a general audience.] microlevel interactions, illustrated in artificial life Boden, M. A. 1999. “Is metabolism necessary?” models. For postsecondary school audiences.] British Journal of the Philosophy of Science 50: 231– ——. 1997b. “Emergent models of supple dy- 48. [Distinguishes three senses of metabolism and namics in life and mind.” Brain and Cognition examines their implications for strong artificial 34: 5–27. [Describes a characteristic suppleness life. For postsecondary school audiences.] of the dynamics of mental states, argues that arti- Dawkins, R. D. 1983. “Adaptationism was always ficial life models capture this kind of dynamics, predictive and needed no defense.” Behavioral and draws out the implications for functionalism and Brain Sciences 6: 360–1. [A defense of adapta- about life. For postsecondary school audiences.] tionism. For postsecondary school audiences.]

210 Artificial Life

Dennett, D. C. 1995. Darwin’s Dangerous Idea: annotated bibliography of works relevant to arti- Evolution and the Meanings of Life. New York: ficial life. For a general audience.] Simon and Schuster. [An extended essay on how Maynard Smith, J. 1975. The Theory of Evolution, 3rd natural selection transforms our view of human- ed. New York: Penguin. [A classic, for a general ity’s place in the universe. Accessible to a general audience.] audience.] Mayr, E. 1982. The Growth of Biological Thought: Div- Gould, S. J. 1989. Wonderful Life: The Burgess Shale ersity, Evolution, and Inheritance. Cambridge, MA: and the Nature of History. New York: Norton. Harvard University Press. [A history of the ideas [A defense of radical contingency in the evolution in biology, for postsecondary school audiences.] of life, in the context of detailed examination of Pattee, H. H. 1989. “Simulations, realizations, and fossils in the Burgess shale. Accessible to a general theories of life.” In C. G. Langton, ed., Artificial audience.] Life: The Proceedings of an Interdisciplinary Work- —— and Lewontin, R. C. 1979. “The spandrels of shop on the Synthesis and Simulation of Living San Marco and the Panglossian paradigm: a cri- Systems. Redwood City: Addison-Wesley, pp. 63– tique of the adaptationist programme.” Proceed- 78. [A criticism of strong artificial life. For ings of the Royal Society B 205: 581–98. [The postsecondary school audiences.] classic criticism of the use of adaptive explanations Ray, T. S. 1992. “An approach to the synthesis of in biology. For postsecondary school audiences.] life.” In C. Langton, C. Taylor, D. Farmer, and Harnad, S. 1994. “Levels of functional equivalence S. Rasmussen, eds., Artificial Life II. Redwood in reverse bioengineering.” Artificial Life 1: 293– City: Addison-Wesley, pp. 371–408. [The classic 301. [Distinguishes synthetic and virtual artificial discussion of one of the best-known artificial-life life, argues that virtual artificial life is impossible, models. For postsecondary school audiences.] and suggests a Turing test for life to settle whether Schrödinger, E. 1969. What is Life? Cambridge: synthetic artificial life is possible. For postsec- Cambridge University Press. [The classic state- ondary school audiences.] ment of the metabolic view of life. For a general Harré, Rom. 1985. The Philosophies of Science. audience.] Oxford: Oxford University Press. [An introduc- Sober, E. 1992. “Learning from functionalism – tion to the philosophy of science, for secondary prospects for strong artificial life.” In C. Langton, and postsecondary school audiences.] C. Taylor, D. Farmer, and S. Rasmussen, eds., Kauffman, S. A. 1993. The Origins of Order: Self- Artificial Life II. Redwood City: Addison-Wesley, organization and Selection in Evolution. New pp. 749–65. [Explores what might be concluded York: Oxford University Press. [A technical about strong artificial life from recent related work presentation of Kauffman’s views of the place of in the philosophy of mind. For postsecondary spontaneous order in living systems. For post- school audiences.] graduate audiences.] Sperry, R. W. 1969. “A modified concept of con- ——. 1995. At Home in the Universe: The Search sciousness.” Psychological Review 76: 532–6. [A for the Laws of Self-Organization and Complexity. classic psychological presentation of consciousness New York: Oxford University Press. [Kauffman’s as an emergent phenomenon. For postsecondary views presented to a general audience.] audiences.] Kim, J. 1999. “Making sense of emergence.” Von Neumann, J. 1966. Theory of Self-reproducing Philosophical Studies 95: 3–36. [An investigation Automata. Urbana-Champaign: University of of the credibility of emergence with special refer- Illinois Press. [Von Neumann’s classic work on ence to the philosophy of mind. For postsecond- self-reproducing automata, completed and edited ary school audiences.] after his death by Arthur Burks. For postgradu- Langton, C. G., ed. 1989. Artificial Life: The Pro- ate audiences.] ceedings of an Interdisciplinary Workshop on the Wiener, N. 1948. Cybernetics, or Control and Com- Synthesis and Simulation of Living Systems. Red- munication in the Animal and the Machine. New wood City: Addison-Wesley. [Proceedings of the York: Wiley. [The classic work on cybernetics.] first “artificial life” conference identified as such. Wolfram, S. 1994. Cellular Automata and Com- The editor’s introduction (expanded in Boden plexity. Reading: Addison-Wesley. [A collection of 1996) is a classic introductory overview of the technical papers on cellular automata as complex field. Includes 25 technical papers and a 40-page systems. For postsecondary school audiences.]

211 Part V Language and Knowledge Information and Content

Chapter 16 Information and Content Jonathan Cohen

Mental states differ from most other entities in aboutness is real, it must be really something the world in having semantic or intentional prop- else. (Fodor 1987: 97) erties: they have meanings, they are about other things, they have satisfaction- or truth-conditions, Some philosophers have reacted to this clash they have representational content. Mental states by giving up one of the two views generating are not the only entities that have intentional the tension. For example, Churchland (1981) properties – so do linguistic expressions, some opts for intentional irrealism in order to save paintings, and so on; but many follow Grice ontological , while McDowell (1994) (1957) in supposing that we could understand abandons naturalism (in the sense under discus- the intentional properties of these other entities as sion) in favor of a kind of intentional dualism as derived from the intentional properties of mental a way of preserving intentional realism. (Termino- states (viz., the mental states of their producers). logical caution: unfortunately, McDowell reserves Of course, accepting this supposition leaves the term ‘naturalism’ for his own view, and refers us with a puzzle about how the non-derivative to the sort of naturalism under discussion here – bearers of intentional properties (mental states) which he rejects – as ‘bald naturalism’.) How- could have these properties. In particular, inten- ever, others hope to find a way of reconciling tional properties seem to some to be especially their naturalistic ontology with intentional real- difficult to reconcile with a robust commitment ism. In particular, many propose to locate inten- to ontological naturalism – the view that the tionality within the natural order by naturalizing natural properties, events, and individuals are the the intentional – by offering an account of inten- only properties, events, and individuals that exist. tional features in naturalistically respectable terms. Fodor puts this intuition nicely in this oft-quoted Many philosophers pursuing this project think passage: that an appeal to information might satisfy their needs – they believe that this notion is both I suppose that sooner or later the physicists naturalistically acceptable and adequate to the will complete the catalogue they’ve been com- analysis of the intentional. In this chapter I shall piling of the ultimate and irreducible pro- review several attempts to naturalize the inten- perties of things. When they do, the likes of tional in terms of information. In section 1 I’ll spin, charm, and charge will perhaps appear lay down some conditions of adequacy on in- upon their list. But aboutness surely won’t; formational theories of content that will be useful intentionality simply doesn’t go that deep . . . If in evaluating the theories presented later. Next,

215 Jonathan Cohen in section 2, I’ll consider Dretske’s influential that are equivalent in any of a number of senses early formulation of an informational theory. I’ll (e.g., logical, analytic, metaphysical, or nomic move on in sections 3–4 to discuss some elabora- equivalence). Many follow Frege in arguing for tions of the view that incorporate notions of this conclusion from the observation that it is epistemic optimality and teleology. Finally, in possible rationally to believe that p and dis- section 5 I’ll discuss another informational theory, believe that q, even when p and q are equivalent due to Fodor, that turns on his notion of asym- in one of these ways; for example, even assum- metric dependence. (Please note that there is a ing that it is metaphysically necessary that the short glossary of technical terms at the end of Morning Star is identical to the Evening Star, it the chapter.) is possible rationally to believe that the Morning Star is wet and simultaneously disbelieve that the Evening Star is wet (you might be in such a state if you didn’t know that the Morning Star 1 Adequacy Conditions is identical to the Evening Star). A theory of content, therefore, must make it possible to dis- It will be useful to structure the discussion around tinguish between contents that are logically, ana- a number of adequacy conditions that an accept- lytically, metaphysically, or nomically equivalent. able naturalistic theory of content must meet: Misrepresentation: An acceptable theory of Naturalism: To vindicate the naturalism in whose intentional content must make it possible that service it is deployed, an informational theory of intentional contents can represent the world content must show how intentional properties erroneously – that they can represent the world as can be characterized in naturalistic terms. While being a certain way even when, in fact, the world it is not obvious what terms should count as is not that way (the problem of formulating an acceptably naturalistic, proposed naturalistic ac- account that makes room for misrepresentation counts of intentional properties should not pre- is sometimes called “the problem of misrepres- suppose that intentional properties are compatible entation” or “the problem of error”). with naturalism (on of begging the very question such theories are intended to answer). Consequently, the necessary and sufficient con- ditions for having intentional features required 2 Dretske and the Flow by a naturalistic theory of content must not them- of Information selves employ intentional idioms, nor should they employ any other notions whose compatibility One of the earliest systematic attempts to under- with naturalism is in doubt. That is to say, the stand intentional content in terms of informa- naturalism constraint is intended to force ana- tion occurs in ’s seminal Knowledge lyses of intentional features to break out of the and the Flow of Information (1981). (Dretske’s circle of intentional features, and therefore pro- was not, however, the first such proposal; posed analyses of these features in terms of others antecedents include Stampe 1975 and 1977, and of these features – satisfaction, truth, content, some suggestive remarks about natural meaning believing that p, etc. – will count for present from Peirce 1931 and Grice 1957.) (Dretske’s purposes as unacceptably non-naturalistic. On the account is also discussed in Chapter 17, KNOWL- other hand, the insistence on naturalism is not EDGE.) Dretske wants to understand the inten- intended to preclude abstract entities (e.g., laws, tional content of a signal (e.g., a ) in properties, numbers) from appearing in the ana- terms of the information that that signal carries lysis of content. under certain circumstances. He understands in- Grain: Since at least Frege (1892), philosophers formation in terms of objective conditional prob- of mind and language have made much of the abilities between events (I shall ignore problems extremely fine grain of individuation for con- concerning the interpretation of these probabil- tent. In particular, there can be distinct contents ities raised in Loewer 1983 and 1987): he writes

216 Information and Content

(p. 65) that a signal r carries the information interpretation of probabilities raised by Loewer that p just in case the conditional probability of – see above), but it faces challenges concerning p, given r (and k, the knowledge of the receiver the desiderata of grain and misrepresentation. of r), is 1 (but given k alone, less than 1). First consider the problem of grain. While But we cannot straightforwardly identify in- Dretske’s requirement of counterfactual support formation carried with intentional content because allows him to set aside merely correlated events in a signal will usually carry too much information. determining the content of a signal, it will not For example, an acoustic signal carrying the in- allow him to choose between properties whose formation that the doorbell is ringing will typic- covariation is necessary. (I assume, following ally also carry the information that the doorbell’s Dretske, that our underlying theory of event indi- button is being pressed. In contrast, it seems viduation allows for a distinction between instan- that intentional content is more constrained: I tiations of properties that necessarily covary.) can have a belief that the doorbell is ringing with- Suppose that a signal carries the information that out having the belief that the doorbell’s button p, but that it is (nomically, metaphysically, ana- is being pressed. Therefore, to give an account lytically, or logically) necessary that p covaries of intentional content, Dretske needs to rule out with a distinct property q; in any of these cases, the sort of informational nesting found in the it will be nomically necessary that p covaries example described. To do this, he stipulates that with q. (A special case comes from Quine’s fam- the information that p is nested in the informa- ous “gavagai” puzzle from Quine 1964.) Quine tion that q just in case q carries the information argued that if a field linguist encountered the that p (71), and then claims that a signal S has term “gavagai” in an unfamiliar language and the fact that p as its semantic content iff: noted that natives assented to the use of this term when and only when rabbits were present, 1. S carries the information that p, and there would be no fact of the matter which of 2. S carries no other information, q, such that many incompatible English translations of the the information that p is nested (nomically term is correct. Live possibilities for the transla- or analytically) in q. (p. 185) tion, according to Quine, include the following: “rabbit,” “undetached rabbit part,” “instantane- Dretske emphasizes that his notion of in- ous temporal stage of a rabbit,” “instance of the formation transmission – hence also his notion universal rabbithood,” and “part of the scattered of intentional content – presupposes a counter- mereological sum of all rabbits.” This provides factual-supporting connection between the signal a special case of the problem under discussion and the information it carries. As a result, a signal because it is necessary that the properties picked correlated with p will fail to carry the informa- out by these expressions covary. (See the discus- tion that p if the correlation is merely accidental sion of the “gavagai” puzzle in the context of or statistical: my thermometer carries informa- informational theories of content in Gates 1996.) tion about the temperature of my room and not In this case, the signal will also carry the informa- yours, even if the two rooms are the same tem- tion that q. Presumably the intentional content perature, because the state of my thermometer that p can be distinct from the intentional con- supports counterfactuals about the temperature tent that q; so which is the intentional content of my room but not about the temperature of of the signal? Dretske is prepared to admit that a your room (that is to say, it is a true generaliza- signal cannot have one of these contents with- tion that if the temperature of my room were out having the other (p. 264 n. 2). different, the state of my thermometer would be This admission has struck many as counter- different; in contrast, it is not generally true that intuitive; however, the problem is even more if the temperature of your room were different, serious than Dretske’s admission would suggest. the state of my thermometer would be different). In fact, Dretske’s account has the result that (not The view of intentional content set out so far both, but) neither of the two pieces of informa- is plausibly thought of as meeting the naturalism tion considered can be the content of any signal. requirement (putting aside worries about its In the case described, the information that p is

217 Jonathan Cohen nomically nested in the information that q, and Dretske’s answer to the problem of misrep- the information that q is nomically nested in the resentation raises a number of problems of its information that p. But on Dretske’s account, own. First, it has seemed to many implausible that the intentional content of a signal cannot be a there is anything like a principled distinction fact that is nomically nested in some other piece between the learning period and the nonlearn- of information carried by that signal, so no signal ing period for most signals. A second concern can have either the content p or the content q. is that, even if there is a learning period for (It is left open that a signal could have the dis- signals, this period must be characterized non- junctive content [p ∨ q].) This problem is obvi- intentionally if the naturalism constraint is to ously quite general, and so is a serious objection be respected. This requirement precludes under- against Dretske’s account. standing the learning period for a signal simply Second, it seems that the account spelled out as the period leading up to that signal’s having so far cannot accommodate the possibility of mis- the intentional content p, and it is not obvious representation. This is because, according to that that there is an alternative naturalistically accept- account, a signal S with the intentional content able understanding in the offing. Third, relying p must carry the information that p, and this on the learning period to explain misrepresenta- requires that the conditional probability of p given tion leaves Dretske without an account of how S is 1 – i.e., p must be true. (Cf. the discussion unlearned (innate) signals could misrepresent. of the so-called “disjunction problem” in Fodor Fourth, as Loewer 1997 notes, the account in 1990d: if both p and q can cause Ss, how can a terms of a learning period is implausible for many theory of content make it the case that q-caused signals; for instance, it seems possible that a child Ss have the erroneous content p rather than the could learn that the linguistic symbol “aardvark” always veridical disjunctive content [p ∨ q]?) Thus, has aardvarks (not pictures of aardvarks) as its on the theory we have considered so far, inten- content, even if all tokens of “aardvark” in the tional contents cannot misrepresent the world. learning period carry information about pictures Dretske is aware of the problem of misrepres- of aardvarks rather than aardvarks. entation, and attempts to answer it by propos- ing that only some of the tokenings of a signal carry the information that determines that signal’s 3 Epistemic Optimality content. In particular, he proposes that there is a learning period for a signal, during which that signal carries the information that p, and that Many are convinced by consideration of these the signal’s intentional content is given only in difficulties that content can’t be reconstructed terms of the information it carries in its learn- in terms of information alone. However, many ing period. After the learning period, when the believe these problems can be solved by a hybrid intentional content of the signal has already theory that appeals to both an informational been fixed as p, tokenings of that signal can fail factor and some other (not strictly informational) to carry the information that p, and so can be factor. The thought is that, so long as both the erroneous: informational and the non-informational factors can be given naturalistic explications, they can In the learning situation special care is taken to work together to provide a naturalistic account see that incoming signals have an intensity, a of content that escapes the vulnerabilities of more strength, sufficient unto delivering the required strict informational theories. piece of information to the learning subject One family of theories of this sort appeals . . . But once we have meaning, once the sub- to a notion of epistemic optimality to take up ject has articulated a structure that is select- the slack left by information. On these views, a ively sensitive to [the information that p]..., signal S has the content p iff there are epistemic-

instances of this structure, tokens of this ally optimal conditions Cp for p such that if Cp type, can be triggered by signals that lack the obtained, then S would nomologically covary appropriate piece of information. (pp. 194–5) with p.

218 Information and Content

Proponents of such accounts, including Stampe a microscope, but the optimality conditions for (1975, 1977), Dretske (1983), and Fodor the belief that there’s a paramecium in the room (1990b), hope that their appeal to epistemic require looking through a microscope. But, of optimality might resolve the problem of mis- course, epistemic optimality versions of informa- representation; this thought is motivated by the tional theories explain the content of signals in (reasonable) suggestion that misrepresentation terms of the optimality conditions for that con- occurs when cognitive systems attempt to repres- tent. Consequently, the understanding of a signal’s ent the world while operating in epistemically having the content p provided by an epistemic suboptimal conditions. Thus, for example, tokens optimality theory must be stated in terms of the of the linguistic symbol “aardvark” that are caused content p, contrary to the naturalism require- by armadillos seen on dark nights have as their ment. Second, insofar as belief fixation is widely content the property aardvark – rather than thought to be a holistic enterprise, it is equally armadillo on a dark night or aardvark or plausible that the epistemic optimality condi- armadillo on a dark night – because the symbol tions for a content p cannot be stated without “aardvark” would covary only with instances adverting to contents other than p. For example, of aardvark in epistemically optimal conditions my tokenings of the linguistic symbol “there are (epistemically optimal conditions for aardvark aardvarks in the room” won’t nomologically presumably involve the lighting being up, the covary with the presence of aardvarks in the room subject’s attentively looking in the right direc- if I believe that aardvarks are not macroscopic- tion, and so on). ally observable. But if so, then an epistemic It has also been suggested that appeals to optimality theory’s unpacking of what it is for a optimality could solve the problem of grain. signal to have the content that p must advert to For even if p and q covary, we could say that states that must be characterized by their con- a symbol has p rather than q as its content if S tents. And once again, this seems a clear violation would nomologically covary with p but not q in of the naturalism constraint. epistemically optimal conditions for p. However, it is unclear that this appeal to epistemic optimal- ity resolves the problem of grain. For one thing, 4 Teleology this solution will fail if the optimal conditions for p (Cp) are identical to the optimal conditions for q (Cq); for in this case, S would, once again, An alternative elaboration of the informational nomologically covary with both p and q in Cp approach to content, appealing to considerations

(and Cq, of course), and so would not deter- of teleology, rather than epistemic optimality, is minately have the content p. A variant of this advocated by several philosophers (see Fodor worry arises when “p iff q” is necessary. For, 1984 and 1990b, Millikan 1984, 1986, and here again, S would covary with both p and q in 1989, Dretske 1988, Papineau 1993). Here, too, all possible conditions, a fortiori in Cp (assum- the hope is that the theory’s non-informational ing condition Cp is possible; if not, then it would factor – its appeal to teleology – will resolve the not be true that S nomologically covaries with p problems of misrepresentation and grain that in Cp, so S could not have the content that p). plague stricter informational accounts. Roughly Moreover, the naturalistic credentials of put, the thought is that misrepresentation occurs epistemic optimality theories are questionable as when a signal S covaries with the information well. This can be seen in two ways. First, it is that q even though S’s teleological function is plausible that the epistemic optimality conditions to carry the information that p (p ≠ q). Sim- for a content p cannot be stated without advert- ilarly for the problem of grain: we could say that ing to the content p itself, since what is epistem- a symbol means p rather than q (even if p is ically optimal seems to depend on what content satisfied when and only when q is satisfied) if the we’re hoping to reconstruct. For example, the teleological function of the symbol is to carry the optimality conditions for the belief that there’s information that p. (These formulations assume an aardvark in the room preclude looking through – controversially – that teleological functions are

219 Jonathan Cohen assigned to individual signals, rather than the the flies-or-bee-bees are flies, and so on. Then whole cognitive systems sustaining the signal- state S covaries not only with the presence of world covariations taken as basic by informational flies, but also with the presence of small moving theories. Unfortunately, I cannot examine the black objects in E, frog-food in E, flies-or-bee- alternative construal here for reasons of space.) bees in E, and so on. Indeed, state S carries Proponents of teleological accounts need to information about all of these; so which (if any) explain the notion of teleological function natur- is the content of S? According to teleological alistically. Of course, they cannot appeal to the theories, the information that p is the content of well-understood example of the teleological func- S iff the carrying of the information that p by tion of artifact symbols in this context (e.g., the other tokens of the same type increased the fitness symbols on the face of a pressure gauge have the of the frog’s ancestors. But a token’s carrying teleological function of representing pressure), information about any of the candidates con- since these instances of teleological function are sidered above would have an equal effect on the presumably constituted in terms of content: the fitness of ancestral frogs: snapping at flies, small gauge has the function of measuring pressure moving black objects in E, frog-food in E, and because that’s what its makers intended it to flies-or-bee-bees in E would all get exactly the do (for similar reasons, appeals to God’s inten- same things into the frog’s belly so long as all tions to fix teleological functions are off limits to the small moving black objects (/frog-food/ would-be naturalists as well). Instead, these theor- flies-or-bee-bees) in E are flies. As Fodor puts ists typically propose understanding teleological the point, function in terms of natural selection (see Wright 1973, Millikan 1984, Neander 1991). On this it’s equally OK with Darwin which way you view, a signal S of type S* has the teleological describe the intentional objects of y snaps, so function of carrying the information that p (in long as it’s reliable (say, nomologically neces- an organism o) just in case earlier tokens of type sary; anyhow, counterfactual supporting) that S* were selected (in o’s species) by natural selec- all the local flies-or-bee-bees are flies. The point tion because they carried the information that is, of course, that if all the local flies-or-bee- p – that is, S has the function of carrying the bees are flies, then it is reliable that the frog information that p in o just in case the carrying that snaps at one does neither better nor worse of the information that p by earlier tokens of selection-wise than the frog that snaps at the type S* increased the fitness of o’s ancestors. other. (Fodor 1990d: 73) To be sure, questions remain about the natur- alistic bona fides of the account just sketched – This problem is even more pressing when the for example, some object that this formulation covariation between candidate contents is not buys its naturalism at the price of an implausibly merely nomologically necessary, but metaphysic- robust conception of natural selection. More- ally necessary: it is extremely difficult to see how over, the success of this account depends on the natural selection could favor a signal with the possibility of a naturalistic explanation of how content aardvark there over one with the con- tokens are assigned to signal types. However, the tent undetached aardvark part there. most important objections against teleological The problem of misrepresentation for tele- accounts allege that they cannot accommodate ological theories is a consequence of the prob- the desiderata of grain and misrepresentation. lem of grain. As we have seen, the teleologist’s The problem of grain for teleological theories appeal to natural selection leaves it open that S is almost invariably presented in connection with could mean fly-or-bee-bee rather than y. But if so, the frog’s capacity to snap at flies. Consider an then nothing could make it the case that bee- internal state S in the frog that covaries with the bee-caused S-tokens are erroneous rather than presence of flies and mediates his snapping beha- veridical. More generally, if this is right, it is vior. It may be that the frog’s environment E is difficult to see what could make it the case that such that all the local small moving black objects any content-assignment by a teleological theory are flies, all the items of frog-food are flies, all is erroneous.

220 Information and Content

vark as its content.) First, Fodor understands 5 Asymmetric Dependence the laws appealed to in this account as ceteris paribus generalizations – generalizations that are Fodor has proposed another informational ac- true (and counterfactual-supporting) when other count as a way of giving a semantics for expres- things are held equal. As a result, the proposal sions in a language of thought (henceforth, is not vulnerable to potential counterexamples Mentalese). (Fodor has presented significantly involving the failure of aardvarks to cause different versions of the theory over the years; AARDVARKs in dark rooms, for inattentive sub- see Fodor 1987, 1990, 1990, 1994. For reasons jects, and so on: these would be brushed aside of space, I confine myself here to the formula- as cases where the relevant paria aren’t cetera, tion in Fodor 1990e, his most complete state- and therefore as falling outside the intended scope ment of the view.) The intuition behind this of the laws. Second, Fodor insists that the sub- view is that, while a Mentalese symbol can carry junctive conditionals in terms of which asym- information about a number of distinct proper- metric dependence is characterized should be read ties, only one of them is its content – namely, its synchronically rather than diachronically (p. 134 content is that single property on which its carry- n. 18). That is, it might be that aardvark- ing information about all the other candidate pictures played a diachronically ineliminable role properties depends. For example, the intuition in the ontogeny of my thought (say, because I runs, while my mental state can carry informa- was taught to token AARDVARK exclusively by a tion about both aardvark and armadillo on a process involving exposure to aardvark-pictures); dark night, its content is aardvark because it in this case, aardvarks wouldn’t cause AARDVARKs would not carry information about armadillo on in me were it not that aardvark-pictures had a dark night (or anything else) unless it carried first caused AARDVARKs in me. Still, according to information about aardvark. This intuition is Fodor, the aardvark-picture to AARDVARK nomic fleshed out in Fodor 1990e in terms of a relation link is asymmetrically dependent on the aardvark of asymmetric dependence between laws, which to AARDVARK nomic link because, he thinks, the is itself specified in terms of a pair of subjunctive synchronic dependence goes in the opposite dir- conditionals: a law L 1 is said to depend asym- ection from the diachronic dependence: once the metrically on a law L 2 just in case (i) if L 2 did capacity to token AARDVARK is in place, aardvark- not hold, then L1 would not hold either, and pictures wouldn’t cause AARDVARKs unless aard-

(ii) if L 1 did not hold, then L 2 would still hold. varks caused AARDVARKs. (That said, Fodor sometimes suggests that the Asymmetric dependence is designed to resolve asymmetric dependence relations are metaphys- the problems of grain and misrepresentation. ically basic and not in need of cashing in terms Take the problem of grain first: why is it that of subjunctive conditionals; see Fodor 1990e: AARDVARK means aardvark rather than nocturnal 93, 95.) termite-consuming African burrowing mammal? Using this apparatus, Fodor writes (p. 121) (For the purposes of illustration, assume it is that a Mentalese expression S has property p as nomically necessary that these properties covary.) its content if: Answer: even if every AARDVARK I token is caused by something that is an instance of both propert- 1. “ps cause Ss” is a law. ies, instances of nocturnal termite-consuming 2. Some Ss are actually caused by ps. African burrowing mammal would not cause 3. For all q ≠ p, if qs qua qs actually cause Ss, AARDVARKs unless instances of aardvark caused then qs causing Ss is asymmetrically depend- AARDVARKs, while the reverse dependency does not ent on ps causing Ss. hold. Take the problem of misrepresentation next: given that some AARDVARK tokens are caused This formulation deserves supplementation. (In by armadillos on dark nights rather than aard- what follows I shall use capital letters to indicate varks, why do those tokens have the erroneous Mentalese expressions: AARDVARK is a Mentalese content aardvark rather than the veridical con- expression that, we hope, has the property aard- tent armadillo on a dark night (or the veridical

221 Jonathan Cohen disjunctive content aardvark or armadillo on a First, it is unclear that Fodor’s account is dark night)? Once again, asymmetric depend- acceptably naturalistic. As noted, Fodor under- ence supplies an answer: armadillos on dark nights stands the laws on which his account rests as cause AARDVARKs only because the former are mis- true, counterfactual-supporting, ceteris paribus takenly identified as aardvarks, so the nomic link generalizations. However, it seems that the between armadillo on dark night and AARDVARK ceteris paribus conditions governing the aardvark– is asymmetrically dependent on the nomic link AARDVARK link cannot avoid adverting to con- between aardvark and AARDVARK (similarly, the tent for two reasons (this mirrors an objection link between the disjunctive property and discussed in section 3). For one thing, just AARDVARKs is asymmetrically dependent on the which cetera must remain paria to sustain the link between aardvark and AARDVARKs). token–world links the theory requires depends Several objections have been leveled against on what that symbol’s content is: instances of Fodor’s theory. Many of these take the form of aardvark can’t cause AARDVARKs in me if I am counterexamples designed to show that Fodor’s looking through a microscope, but instances conditions do not assign the correct content to of paramecium can’t cause PARAMECIUMs in me Mentalese expressions (Cummins 1989: ch. 5, unless I am looking through a microscope. For Godfrey-Smith 1989, Baker 1991, Boghossian another, just which cetera must remain paria to 1991, Manfredi & Summerfield 1992, Adams & sustain the token–world links the theory requires Aizawa 1994). Fodor has responded to many of depends on what other contents the subject these criticisms (see Fodor 1990e, the replies of believes: I won’t reliably token AARDVARK in the Loewer & Rey 1991, and Fodor 1994). Unfor- presence of aardvarks if I believe that aardvarks tunately, these controversies are often extremely are not macroscopically observable. For both difficult to assess because it is unclear how we these reasons, it seems that the ceteris paribus should understand the relevant subjunctive clauses appearing in the laws required by the conditionals. (The difficulty in assessing alleged theory ultimately must be cashed in terms of counterexamples to Fodor’s theory is even more content, and consequently that the theory can- severe if asymmetric dependence is taken as meta- not be stated without adverting to content. If physically basic – see above.) so, then the view does not abide by the natural- A further complication is that Fodor intends ism constraint. his theory to provide only sufficient, and not Finally, while the asymmetric dependence necessary, conditions for a state’s having content. account is arguably successful in dispatching As Fodor construes it, his task is only to show many instances of the grain problem, it remains how Mentalese expressions could have content – helpless to mark distinctions in content between not to show how Mentalese expressions do have properties whose covariation is metaphysically content. By reducing his aspirations in this way, necessary. Because it is metaphysically neces- Fodor forestalls objections about whether his sary that instances of aardvark covary with conditions are necessary for a state’s having instances of undetached aardvark part, there is a content, and in particular about whether the counterfactual-supporting generalization linking subjunctive conditionals he invokes are true: instances of aardvark to AARDVARK tokens iff there is a counterfactual-supporting generalization It’s enough if I can make good the claim that linking instances of undetached aardvark part to “X” would mean such and such if so and so AARDVARK tokens. Moreover, since the covariation were to be the case. It’s not also incumbent between the two properties is metaphysically upon me to argue that since “X” does mean necessary, this will be so in every metaphysically such and such, so and so is the case. (Fodor possible world, so there can be no asymmetric 1990e: 96) dependence between the two laws. (Presumably Fodor would not respond by appealing to asym- Whether or not the points considered so far metric dependencies in metaphysically impos- are damaging, the asymmetric dependence theory sible worlds, since those worlds will include many faces other important obstacles. where the dependency goes in the opposite

222 Information and Content direction from that required by the theory.) counterfactual support: a generalization is said Consequently, the theory predicts that AARDVARK to be counterfactual-supporting if it is not only (i) has the content aardvark if and only if it true of its instances, but also would be true has the content undetached aardvark part, or of relevant noninstances. For example, the (ii) has the disjunctive content aardvark or generalization “metals expand on heating” is undetached aardvark part. (In Fodor 1994: ch. counterfactual-supporting because it not only 3 he recognizes this problem, and suggests that says something true about what happens to a theory of content can settle on a determinate heated metals, but also says something true about and nondisjunctive content for a Mentalese what would happen to unheated metals were expression by appeal to the inferential relations they heated (contrary to actual fact). In contrast, of thoughts containing that expression. This solu- “everything in Nelson Goodman’s pocket on VE tion is criticized in Gates 1996 and Ray 1997.) Day was silver” is not counterfactual-supporting because, even if all the things in Nelson Goodman’s pocket on VE Day were in fact silver, there are many nonsilver objects that might have 6 Conclusion been in Nelson Goodman’s pocket on VE Day. It is widely thought to be a requirement on Many contemporary philosophers believe that nomic generalizations (as opposed to mere accid- informational theories are the most promising entally true generalizations) that they support proposals for reconciling naturalism with inten- counterfactuals. tional realism. However, it remains to be shown holism of belief fixation: the claim that it is impos- that there is an informational theory of content sible to fix, or come to, a given belief without that satisfies the constraints of section 1 above. holding in place a number of other beliefs at Of course, this does not mean that no informa- the same time. According to this view, an experi- tional theory can succeed. It does mean that, so mental datum confirms (i.e. verifies, gives us far, appeals to information have not resolved the some reason to believe) a given statement only problem of naturalizing content. in conjunction with one’s other theoretical com- mitments, background assumptions about the experiment, and assumptions about the logical Glossary of Key Technical Terms and mathematical apparatus connecting the datum to these other beliefs. analytic necessity: a sentence is said to be ana- intentional: about something. Things that are lytically necessary if it is necessary in virtue of about other things (e.g., mental states, words) its meaning. For example, many believe that are said to have intentional properties. Not to the sentence “bachelors are unmarried” is ana- be confused with a different, ordinary usage of lytically necessary. Compare with metaphysical “intentional” to mean on purpose. and nomic necessity. Mentalese: language of thought. According to carry (information): a signal r carries the informa- Fodor (1975), thinking should be understood tion that p just in case the conditional probability as manipulation of Mentalese symbols with both of p, given r (and k, the knowledge of the receiver syntactic and semantic (intentional) properties. of r), is 1 (but given k alone, less than 1) (Dretske metaphysical necessity: a proposition (or a sentence 1981: 65). expressing a proposition) is said to be metaphys- ceteris paribus: Latin for “other things equal.” ically necessary just in case it is necessary by Ceteris paribus generalizations are nonstrict virtue of metaphysical truths. For example, if the generalizations – generalizations that hold when correct metaphysics of the constitution of water other things are equal. says that water is H2O, then it is metaphysically covariation (properties): properties P and Q covary necessary that non-H2O stuff – even if clear, just in case P is instantiated if and only if Q is potable, odorless, tasteless, etc. – is not water. instantiated. Compare with analytic and nomic necessity.

223 Jonathan Cohen naturalism: the view that the natural properties, Baker, L. R. 1991. “Has content been naturalized?” events, and individuals are the only properties, In Loewer & Rey 1991: 17–32. [Baker argues events, and individuals that exist. that Fodor vacillates between different (and in- consistent) versions of his theory as he attempts nesting: the information that p is nested in to handle different problem cases, but that no the information that q just in case q carries the version is adequate as a general solution.] information that p (Dretske 1981: 71). Specific- Boghossian, P. A. 1991. “Naturalizing content.” ally, p is nomically nested in q just in case it is In Loewer & Rey 1991: 65–86. [Boghossian nomically necessary (necessary by virtue of natural argues that Fodor’s asymmetric dependence view laws) that p is nested in q. Similarly, p is analytic- collapses onto an optimal conditions theory, and ally nested in q just in case the sentence “p is that it is therefore susceptible to the kinds of nested in q” is analytically necessary (necessary objections that Fodor himself raises against the in virtue of its meaning). latter view.] nomic: law-governed. Churchland, P. M. 1981. “Eliminative materialism and the propositional attitudes.” The Journal of nomic necessity: a proposition (or a sentence Philosophy 78: 67–90. [Churchland argues that expressing a proposition) is said to be nomically “folk psychology” – the commonsense predict- necessary just in case it is necessary by virtue of ive/explanatory understanding of the mental that natural laws. For example, the proposition that centrally involves the attribution of intentional metals expand when heated is nomically neces- states – is a radically and systematically erroneous sary. Compare with analytic and metaphysical theory in need of replacement by a scientific psy- necessity. chology pitched at a neural level.] subjunctive conditional: a conditional is a hypo- Cummins, R. 1989. Meaning and Mental Repres- thetical statement of the form “if p then q”; the entation. Cambridge, MA: MIT Press. [Cummins component p is called the antecedent of the discusses a number of (informational and non- conditional, while component q is called the con- informational) approaches to content in this slim sequent. Conditionals whose antecedents are in book, and presents his own “interpretational” the grammatical subjunctive mood neither pre- semantics.] suppose that their antecedents are true nor that Dretske, F. I. 1981. Knowledge and the Flow of they are false; these are called subjunctive condi- Information. Cambridge, MA: MIT Press. [This tionals. For example, “if I were to eat a bagel, is Dretske’s first major defense of an informa- tional theory of content (see section 2 above). then I would be full” is a subjunctive conditional: This book was instrumental in bringing informa- its antecedent is in the subjunctive mood, and it tional approaches to the attention of mainstream presupposes neither that I do eat a bagel, nor philosophers of mind.] that I do not. ——. 1983. “The epistemology of belief.” Synthese 55: 3–19. [Dretske argues that if a system can- not process information to the effect that some- References thing is F, then it cannot represent something as F, a fortiori cannot believe that something is F. Adams, F. and Aizawa, K. 1994. “Fodorian He also endorses an epistemic optimality account.] semantics.” In S. P. Stich and T. A. Warfield, ——. 1988. Explaining Behavior: Reasons in a World eds., : A Reader. Oxford: of Causes. Cambridge, MA: MIT Press. [Dretske Blackwell, pp. 223–42. [Adams and Aizawa argue aims to explain how reasons, citing psychological that there is no tenable reading of Fodor’s states, can figure in the explanation of actions, asymmetric dependence account by proposing given the (in principle) availability of complete counterexamples (cases where the theory assigns explanations in terms of physical causes. As part the wrong content, or assigns content where it of this project, Dretske defends a historical/ should not). They also argue that Fodor gets the teleological information-theoretic account of the order of explanation wrong – that asymmetric intentional.] dependence is a consequence of intentional rela- Fodor, J. A. 1975. The Language of Thought. Cam- tions, not the ground of those relations.] bridge, MA: Harvard University Press. [Fodor

224 Information and Content

argues that our best psychological theories pre- explains and discusses his approach in some detail, suppose the existence of a language of thought and responds to a large number of potential (Mentalese) – a language-like system of symbols objections.] with syntactic and semantic properties.] ——. 1994. The Elm and the Expert: Mentalese and ——. 1984. “Semantics, Wisconsin style.” Synthese, its Semantics. Cambridge, MA: MIT Press. [Fodor 59: 231–50, Repr. in Fodor 1990c: 31–49. engages in a number of projects in this book, [Fodor criticizes views associated with Dretske derived from his Nicod Lectures of 1994. Among and Stampe for failing to solve the disjunction these, he offers an answer to Quinean inscrutab- problem.] ility problems in chapter 3 and some substant- ——. 1987. Psychosemantics: The Problem of Mean- ive revisions and extensions of his asymmetric ing in the Philosophy of Mind. Cambridge, MA: dependence theory in two appendices.] MIT Press. This book contains Fodor’s first pub- Frege, G. 1892. “On sense and meaning.” In B. lished statement of his asymmetrical dependence McGuinness, ed., Gottlob Frege: Collected Papers theory.] on Mathematics, Logic, and Philosophy. Oxford: ——. 1990a. “Information and representation.” Blackwell, pp. 157–77. [In the opening pages of In P. Hanson, ed., Information, Language, and this paper, Frege notes that co-referential expres- Cognition. Vancouver: University of British sions are not intersubstitutable salva veritate in Columbia Press, pp. 175–90. [Fodor marks a contexts. This point is at distinction between “labeling” uses (e.g., “that the heart of the grain constraint discussed in is an aardvark”) and “representational” uses (e.g., section 1 above.] “an aardvark is a beautiful animal”) of the same Gates, G. 1996. “The price of information.” Synthese symbol: representational uses are often not in- 107(3): 325–47. [Gates argues that informational tended to indicate that the cause of the symbol theories of content presuppose a solution to has the property expressed by the symbol. He Quinean inscrutability, and that no such solution insists that the asymmetric dependence account is available.] is intended only to apply to labeling uses of Godfrey-Smith, P. 1989. “Misinformation.” symbols.] Canadian Journal of Philosophy 19: 533–50. ——. 1990b. “Psychosemantics or: Where do [Godfrey-Smith claims that Fodor fails to solve truth conditions come from?” In W. G. Lycan, the disjunction problem – that he has no way to ed., Mind and Cognition: A Reader. Malden, ensure that the content of wild tokens will depend MA: Blackwell, pp. 312–38. [Widely circulated asymmetrically on that of veridical tokens, as re- as a manuscript circa 1980. Fodor defends an quired by the asymmetric dependence theory.] optimality/teleology informational account. He Grice, H. P. 1957. “Meaning.” The Philosophical repudiates this view in Fodor 1984, 1990d, and Review 66: 337–88. [Grice distinguishes between 1987.] “natural meaning” – the kind of meaning that ——. 1990c. A Theory of Content and Other Essays. we are speaking of when we say something like, Cambridge, MA: MIT Press. [This collection “Those spots mean measles” – and “non-natural contains, among a number of previously pub- meaning” – the kind of meaning we speak of lished essays, two previously unpublished papers when we say “Those three rings on the bell mean on content that comprise the most worked-out that he made it safely to shore.” He proposes statement of Fodor’s asymmetric dependence that the non-natural meaning of symbols can be view.] understood in terms of the intentions of the pro- ——. 1990d. “A theory of content, I: The prob- ducers of those symbols. Informational theories lem.” In Fodor 1990c: 51–87. [Here Fodor dis- of content are proposed to explain (as a kind of cusses the problems that an informational theory natural meaning) the content of the intentions of content must solve, and argues that going constituting non-natural meanings.] versions of such theories are unsuccessful at solv- Loewer, B. 1983. “Information and belief.” ing them.] Behavioral and Brain Sciences 6: 75–6. [Among ——. 1990e. “A theory of content, II: The the- other complaints, Loewer’s brief comment on ory.” In Fodor 1990c: 89–136. [This is Fodor’s Dretske objects that the probabilities to which most complete presentation of his asymmetric Dretske appeals cannot be understood in any of dependence theory. Fodor motivates the view, the standard ways.]

225 Jonathan Cohen

——. 1987. “From information to intentionality.” ——. 1989. “Biosemantics.” The Journal of Philo- Synthese 70: 287–317. [Loewer argues that extant sophy 86: 281–97. Repr. in Stich & Warfield 1994. informational theories are unsuccessful in pro- [Millikan clarifies, elaborates, and defends the tele- viding naturalistically acceptable understandings ological theory of content presented in Millikan of content.] 1984.] ——. 1997. “A guide to naturalizing semantics.” Neander, K. 1991. “Functions as selected effects: In B. Hale and C. Wright, eds., A Companion the conceptual analyst’s defense.” Philosophy to the Philosophy of Language. Oxford: Blackwell, of Science 58: 168–84. [Neander defends an pp. 108–26. [Loewer extends the arguments of etiological theory of biological function – a Loewer 1987 to the effect that extant informa- theory on which the function of a trait is the tional theories of content cannot meet reason- effect for which it was selected by natural able desiderata.] selection. This view gives one way of under- —— and Rey, G., eds. 1991. Meaning in Mind: standing the notion of function that figures Fodor and His Critics. Oxford: Blackwell. [This centrally in teleological informational theories book contains a short summary of Fodor’s work of content.] by the editors, 14 critical essays on various aspects Papineau, D. 1993. Philosophical Naturalism. of his oeuvre (including several on his asymmetric Oxford: Blackwell. [This book is a very general dependence theory), Fodor’s replies to all of defense of naturalism in metaphysics and epi- these, and an extensive bibliography of Fodor’s stemology. Part II spells out a teleological con- writings.] strual of intentional properties that is intended Manfredi, P. and Summerfield, D. 1992. “Robust- to respect Papineau’s naturalist aims.] ness without asymmetry: A flaw in Fodor’s theory Peirce, C. S. 1931. Collected Papers. Cambridge, of content.” Philosophical Studies 66: 261–83. MA: Harvard University Press. [Volume 2 of this [Manfredi and Summerfield offer another pur- work contains a very early formulation of an ported counterexample to Fodor’s account. They informational theory of content.] conclude that asymmetric dependence cannot be Quine, W. V. O. 1964. Word and Object. Cam- a general solution to robustness.] bridge, MA: MIT Press. [This is the source of McDowell, J. 1994. Mind and World. Cambridge, the “gavagai” problem that, according to Gates MA: Harvard University Press. [This book is of 1996 plagues contemporary informational theories interest in the present context principally because of content.] it rejects the naturalism constraint I set out in Ray, G. 1997. “Fodor and the inscrutability prob- section 1. McDowell believes that the normative lem.” Mind and Language 12(3/4): 475–89. dimension of content cannot be captured natur- [Ray attempts to undermine Fodor’s answer to alistically, and therefore is unsympathetic to the the inscrutability problem in Fodor 1994. He goal of naturalizing content (and other forms of gives several examples to which, he claims, Fodor’s “bald naturalism”).] answer is inapplicable, and concludes that Fodor’s Millikan, R. G. 1984. Language, Thought, and Other solution cannot secure the scrutability of refer- Biological Categories: New Foundations for Real- ence in the general case.] ism. Cambridge, MA: MIT Press. [Millikan pre- Stampe, D. 1975. “Show and tell.” In B. Freed, sents a general naturalistic theory of teleological A. Marras, and P. Maynard, eds., Forms of Rep- function (one that is intended to apply to body resentation: Proceedings of the 1972 Philosophy organs, behaviors, and mental representations, Colloquium of the University of Western Ontario. inter alia), and defends a teleological account of Amsterdam: North-Holland. [This is one of the content.] earliest formulations of an informational theory ——. 1986. “Thought without laws: cognitive sci- of content, aimed primarily at the understanding ence with content.” The Philosophical Review 95: of linguistic content.] 47–80. [Millikan holds that references to the ——. 1977. “Toward a causal theory of linguistic content of psychological states are ineliminable representation.” In P. French, T. Uehling, and in both folk psychology and cognitive science. H. Wettstein eds., Midwest Studies in Philosophy, However, she thinks, psychology is naturalistic vol. 2. Minneapolis: University of Minnesota because content can be unpacked by appeal to Press. [Here Stampe develops and defends the the evolutionary history of systems and states.] view of Stampe 1975. This paper is widely cited

226 Information and Content

as a primary source of contemporary interest in analysis of biological functions must make refer- informational approaches to content.] ence to the causal background of the bearers Stich, S. P. and Warfield, T. A., eds. 1994. Mental of such functions. He argues that this sort Representation: A Reader. Oxford: Blackwell. [A of “etiological” account avoids difficulties that fine collection of previously published papers on plague competing analyses and secures several mental content. The papers cover a wide range of important desiderata for the notion of function. (informational and non-informational) approaches.] Proponents of teleological accounts of content Wright, L. 1973. “Functions.” The Philosophical have often relied on accounts of function in this Review 82: 139–68. [Wright proposes that the tradition.]

227 Fred Adams

Chapter 17 Knowledge Fred Adams

can certainly believe it. And one cannot know Introduction that the US Supreme Court stopped the state- wide Florida hand recount, if one has no beliefs In business and finance, “managing informa- about it whatsoever, though it can be true. So tion” and “managing knowledge” practically are truth and belief, though independent, are both synonymous. Surfing internet search engines pro- required for knowledge. Knowledge requires that vides confirmation. What is it about knowledge these two things, otherwise independent, come and information that makes the two suited to together. one another? At the level of common sense, if Nozick (1981) claims that knowledge tracks one possesses the information that p, one is in a the truth, such that the beliefs of a knower would position to know that p, all else being equal. And co-vary with the truth – when one applies the if one is uninformed, one is not in a position to right belief-forming methods. Figuratively, a know that p. Similarly, if one knows that p, one knower’s cognitive light of belief would illumin- has more than a justified true belief (Gettier ate, were p true, and not illuminate, were p false. 1963) – justified beliefs can be true accidentally. Attention turns to methods. What sorts of I see Ken and believe I do (someone I know methods would provide the mechanisms to imple- well). But my belief that it is Ken is accidentally ment the sort of tracking of truth required by true and not knowledge because, although I see knowledge on a theory like Nozick’s? For know- Ken, I cannot distinguish Ken from his identical ledge of empirical propositions, one would expect twin Ben, who is standing right behind Ken. to need empirical methods: perceptual observa- Information that it is Ken may secure the needed tion, instrumentation, testimony, or perhaps an connection between belief and truth to take me entire empirical theory. Non-empirical truths of beyond a merely true (even justified) belief that logic or mathematics may require mathematical I see Ken. Whether this is true beyond the level intuition, reason (a form of cognitive computa- of common sense depends upon a deeper under- tion), formal systems, and techniques of proof. standing of knowledge and information. Justification is also a notion that is traditionally Let’s begin with knowledge. It is uncontro- associated with knowledge, for it is often thought versial that knowledge requires truth and belief. to be a necessary ingredient in getting beliefs to One cannot know that Bush got more votes in track the truth in ways capable of generating Florida than Gore in the 2000 American Pres- knowledge. There is a wide variety of theories of idential election, if it is not true, though one justification, and each emphasizes different aspects

228 Knowledge or factors relevant to knowing. Externalists stress information about who pulled the trigger before factors that are possibly outside the mind or one lifts the prints. Thus, information appears to awareness of the believer. Internalists stress factors be mind-independent (and, thereby, language- of which a believer must be aware. Foundational- independent too). ists attend to the origin, grounds, and structure Information must also be capable of having a of support for beliefs that are not inferred from very special relationship to the truth. Since one other beliefs. Coherentists stress the interrela- cannot know what is false, if information is going tion of beliefs and their relations of mutual con- to bring one to know that p, then information firmation. And reliabilists attend to the likelihood must also be tied to the state of affairs that makes that a set of cognitive mechanisms and their con- p true. Otherwise, it is hard to see the value of ditions of operation tend to produce true beliefs information over belief and truth itself. On at more often than false beliefs. Despite their dif- least some accounts, information has this con- ferences, all these theories agree that a belief’s nection to truth (Dretske 1981, Floridi forth- being justified contributes to minimizing that a coming b). One can be misinformed. One can belief is merely true, but not known. The exact be informed that q, when one needs to know relation between a belief’s being justified and its that p, but one cannot be misinformed that p. being true continues to be difficult to specify with For something can only carry the information precision, partly because there can be justified that p, if p. Indeed, if we think of information false beliefs. Still, it is undeniable that epistemic as being contained or carried in one event (set justification is prized because it either increases of events) and as being about another event (set the likelihood that one non-accidentally has a true of events), then the transmission of information belief, or, having that, it satisfies an additional is the product of a correlation and dependency cognitive requirement of knowing (coherence, between the two events (sets). To see this in more evidence, or rationality). detail, let’s consider Dretske’s (1981) attempt Some have argued that if justification is some- to explicate an account of information that may thing other than information (Adams 1986) or be useful in understanding knowledge. other than what, when added to true belief, guar- We will first look at Dretske’s account and see antees that the belief is non-accidentally true how he uses information to explicate knowledge. (Lewis 1996), then justification is neither neces- We will also look at some interesting objections sary nor sufficient for knowledge. Still, a theory of to his account. This will give us a good idea of knowledge that incorporates information might the usefulness of information in understanding tack one or other account of justified belief onto what knowledge is. We will then consider some its conditions of knowing. Alternatively, one interesting open questions about knowledge might attempt to substitute information for justi- and information. And we will close with a survey fied belief into one’s account of knowledge. In of some current philosophical debates about what follows, we will consider an information- knowledge. theoretic attempt to do an “end run” around giving a theory of justified belief. Now let’s turn to information. To be of value Dretske’s Adaptation of to a would-be knower, information must be an Information Theory to Knowledge objective, mind-independent commodity. In prin- ciple, it should be possible for someone to be the first person to learn that p. If S were the first To adapt information theory to a format friendly person brought to know that p by the informa- to a theory of knowledge, several matters need tion that p, then the information that p would to be resolved. For example, to know that Bush appear to have objective properties. The follow- was elected president involves information being ing examples suggest that this is so. Waves of generated by the event of his election. It also radiation traveling through space may contain involves transmission of that information to a information about the Big Bang before anyone prospective knower S. S must detect physical detects it. Fingerprints on the gun may contain events that carry that transmitted information,

229 Fred Adams and those events must cause or sustain S’s belief the information that Susan was selected. We want that Bush was elected. the information at r to be about the events that

Let’s begin with generation of information. transpired at s. Letting “Is(r)” represent this = − An event’s occurrence generates information. information, Is(r) [I(r) noise]. Noise is the How much is generated is a function of how amount of information generated at r that is likely was the event’s occurrence. The more likely independent of what happens at s (not about s), an event, the less information it generates – while and when “Tony” is placed on the forms, but the less likely the event, the more information Susan was selected, the noise = 3 bits. Thus, no it generates. Different ways of classifying events information about s arrives at r. may result in different amounts of information Now for our purposes, the import of these generated. And there are many different ways formulae for calculating amounts of information of trying to measure or quantify amounts of is not so much the absolute values of informa- information. Dretske follows the communication tion generated or transmitted by an event, but industry standard (Weaver & Shannon 1949) of the conditions necessary for transmission. For measuring information in bits (binary digits), rep- most events it would be difficult or impossible resenting the number of binary partitions neces- to determine the exact probabilities and ranges sary to reduce a collection of equally probable of possibilities closed off by an event’s occur- outcomes to one (e.g., beginning with 8, a three- rence. What is important is whether one receives step reduction to 4, to 2, to 1 = 3 bits). The at r as much information as is necessary to know amount of information generated at a source s what happened at s (under a relevant specifica- by the reduction of n equally likely possibilities to tion). For a signal or message to carry the one is represented: I(s) = log n (base 2). Here information that Bush was elected, it must carry I(s) represents the average amount of informa- as much information as was generated by Bush’s tion generated at a source by a reduction of election. We know this is more information than equally likely events. If the range of possible that a Republican ran for office, and more than events at the source s1, s2,...sn, are not all equally that someone was elected. Calculating exactly likely, then the amount of information gener- how much information is generated by Bush’s = ated by the occurrence of si is: I(si) log 1/p(si) election is not as important as determining under (where p = probability). So, for example, sup- what conditions the information that does arrive pose 10 persons apply for a job and nine are carries the information that Bush was elected. from inside the company, one from outside. If s1 This is what Dretske calls the informational con- is the selection for the job of someone outside tent of a signal. = = the company, then I(s1) log 1/.1 3.33 bits Informational content: A signal r carries the of information. For contrast, selection of some- information that s is F = The conditional prob- one from inside the company, s2 would generate ability of s’s being F, given r (and k), is 1 (but, 1/0.9 = 0.15 bits of information. given k alone, less than 1). Next, let’s consider information flow or trans- k is a variable that takes into account how what mission. For information at a receiving point r one already knows may influence the informa- to be about a sending point s, there must be tional value of a signal. If one knew nothing, k dependence between the events at r upon those would go to zero. If I know that Vice President at s. Suppose at s there are 8 candidates equally Cheney is from Texas or Wyoming, and I learn likely to be selected. A selection of Susan gen- that he is not from Texas, I thereby have the erates 3 bits of information. Suppose at r there information that he is from Wyoming. If you are eight equally likely names that may be put hear that he is not from Texas, but don’t already on the employment forms in the employment know Wyoming is the only other possibility, you office. A selection of “Susan” generates 3 bits do not thereby receive the information that he is of information. But there would also be 3 bits from Wyoming. generated if, mistakenly, the name “Tony” were This account of the informational content of placed on the employment forms. Clearly, though a signal has important virtues. If a signal carries this amount of information is the same it is not the information that Bush was elected, then since

230 Knowledge the conditional probability that Bush was elected, and, according to ITK, do not know that I am. given the signal, is 1, then Bush was elected. Some people do hallucinate from tumors. Hence, the account gives information a connec- The possibility of the brain tumor may seem tion to truth. Clearly it will also be the case that extremely remote, because it would have to cause the signal carries as much information about s, such sophisticated hallucinations. But other pos-

Is(r), as was generated by the fact that Bush sibilities are less remote. Do you know where was elected. Noise about the fact that Bush was your car is? Is the probability that it is where elected is zero. Hence, the account gives us a you parked it, given your memory impressions way to understand transmission or flow of of having parked it there, 1? Are cars stolen, information of a specific propositional (factual) where you live? Or, do you know the money in content from source to receiver – not just your pocket is not counterfeit? Is the probability amounts of information. that it is counterfeit zero? Finally, we can give an information-theoretic Suppose I see a person that I take to be Bill account of knowledge (ITK): K knows that s is Clinton. I believe it is Clinton because the per- F = K’s belief that s is F is caused (or causally son looks like Clinton (I’ve seen him on televi- sustained) by the information that s is F. Dretske sion many times). Do I know that it is Clinton? says that this is intended to account for percep- I do only if the probability that someone in my tual knowledge only, that is, perceptually know- vicinity would look just like that (where the ing of something s that it is F. Knowing, by my “that” refers to the way he looks) and not be current visual experiences of my computer, that Bill Clinton is zero. Or, to put it another way, it is on, would count as perceptual knowledge. the probability that it is Clinton, given that it And so would my knowing that the coast is clear, looks like him, must be 1. Now suppose that I by hearing three knocks, on a prearranged sig- get no confirmation – no follow-up information nal. However, here my knowledge also involves or observation. But suppose that it was Clinton. knowing the prearranged signal. I know that the Do I know this? I do only if there are no “dead coast is clear by hearing the three knocks. Clearly ringers” for Clinton in my vicinity – no one who there are other forms of knowledge and other might look like him from an angle or side view, ways of knowing and other kinds of things that or even straight on. Is the world ever that stable, can be known. Whether an ITK can be adapted when it comes to the way people look? If it is to all cases of empirical knowledge is an inter- not, I could never know it was Clinton from esting question still open. It is not the only such a chance meeting, because I would never one. receive the information that it was Clinton. So it is an open question whether the world is stable enough to generate this type of information, and if it is, how often – and about how many things: Interesting Open Questions Clinton, parked cars, legal currency? One open question was whether the world Perhaps the first interesting and open question provides determinate information that p (con- for an ITK is: how many conditional probabil- ditional probabilities of 1). Another is whether ities of 1 exist? If they are required for know- we can know that it does. To know that the ledge and the world does not provide them, then world provides conditional probabilities of 1 is skepticism rules the day. Knowledge, even if it to know that we have the information that p. To exists, may be scarce. Famously, Descartes may know that we have the information that p is to have known he existed because the probability know that we know, on ITK. Now many would that he did, given that he was conscious of doubt- maintain that we do not know that we have ing (thinking), was 1. But is the probability knowledge, even if we do. And if that is correct, zero that I might have a brain tumor causing me then we do not know that we have conditional to hallucinate typing this passage? Unless it is probabilities of 1 in the evidence causing or sus- zero, I am not receiving (by my experiences) the taining our beliefs. So unless we know that we information that I am typing this passage now, have knowledge, we will not know that there are

231 Fred Adams the conditional probabilities of 1 that know- can thereby come to know the thing known by ledge requires. Thus, it also would be an open the speaker. The above example calls this into question whether we know that there is know- question. For, as we have supposed, the person ledge. These skeptical challenges are faced by on the phone hears Al assert that he is working this (or any) theory of knowledge. Here they on an American motorcycle, but does not receive must be left as open questions. ITK tells us what the information that Al is working on an Amer- it takes to have knowledge, but not whether we ican motorcycle. For Al need not be a reliable have it. conduit of information. Al’s utterance may con- Next, on an ITK, since one’s belief that p tain only the information that the motorcycle is contains the information that p (Barwise & Perry American made or German made (but not which). 1983), should not one necessarily be able to Therefore, Al seems to know something that he convey one’s knowledge that p to others? There cannot transmit or pass on – at least if all that he are reasons (Dretske 1982) to say no. Suppose tells the hearer is “it is American made.” that Al repairs motorcycles for a living, and that It is another interesting and important he would never mistake a BMW for a Harley consequence of this type of ITK that logical, Davidson (only BMWs have horizontal pistons). mathematical, or analytic truths generate zero However, Al mistakenly believes both are Amer- information. Metaphysical necessities (water’s ican made. When Al is working on a Harley, will being H2O or Tully’s being Cicero) also gener- = he know that it is American made? Al knows ate zero information. Since I(si) log 1/p(si), that it is a Harley by the label on the motorcycle and since any of these things have a probability and the repair manual. The conditional probab- of 1, their informational value is zero. Hence, ility that it is American made, given that it is a we have a feature that puzzled Frege, viz. both Harley, is 1, and Al believes it is American made that a = a, and that a = b generate no informa- because he believes it is a Harley. So Al’s belief tion. Yet the latter seems informative, while the that the Harley he is repairing is American made former does not. Now it may seem strange indeed is caused and sustained by items carrying the to say that it takes no information to know these. information that it is American made. But in And it certainly seems strange to treat nomic general, when Al thinks something is American necessities (water’s expanding when freezing) in made, the probability that it is, given his belief, the same way one treats mathematical or logical is not 1 (for he thinks BMWs are American truths. made). Thus, Al’s beliefs of the form [motorcycle Perhaps just as puzzling are nomological x is American made] do not contain the informa- impossibilities and logical falsehoods. Since tion that something is an American motorcycle. they have a probability of zero, the amount of Now, if Al is on the phone with someone and says information carried by sentences about them is only that he is repairing an American motorcycle, infinite (or undefined). There are interesting ways the hearer will not receive the information that out of this conclusion (Floridi forthcoming a), Al is working on an American motorcycle. For but we will not have space to pursue them here. Al would have said this even if he had been It is certainly an important and open question working on a BMW. whether an ITK can be developed to explain Immediately following this discussion, we must how we know mathematical or logical truths, ask whether it implies that Al may know that the since these truths generate no information. It is motorcycle he is working on is American made, possible that, while no information is generated but be unable to communicate that knowledge? by these truths, because of their impossibility (This will depend on what the other person on of being false, they are known by means other the phone knows about Al. For our purposes, than receiving information. Several possible ways let’s suppose that the person on the phone knows this might go are available, but we will not very little about Al or motorcycles.) Normally we have time to pursue them here. It will have to would expect that if someone knows something, suffice to say that it might be possible to know and sincerely asserts the truth that is known, then things about necessary truths by getting elements one who hears the assertion and understands it of one’s beliefs and representational structures

232 Knowledge

(mathematical and logical representations and arrived, given the knocks, is 1, but the condi- proofs) to mirror or be isomorphic to the ele- tional probability that the courier has arrived, ments of the domain to be known (Katz 1999). given that the events in the jarred brain have Whether this is possible is still an important and occurred, is not 1. This is because the spy’s brain, fascinating open question in philosophy of math- since jarred sufficiently hard by the box, might ematics and theory of knowledge. be in that state (seeming to hear three knocks) even if the courier had not arrived and there had not been three knocks at the door. Let me explain. Current Philosophical Debates For the purpose of addressing Foley’s example and issues to follow, it is important to discuss the Is an ITK able to withstand scrutiny? Let’s look matter of an information channel. An informa- at some objections. Foley (1987) says that any tion channel condition is any fixed condition informational account that relies upon causation (other than conditions existing at the source s or of a belief by information will be susceptible to receiver r) which, by variation of its value, would well-known problems of causal deviance. Ironic- be able to introduce noise between source and ally, this is exactly one of the things an informa- receiver. In order for the information that a is F tional account was designed to avoid with the to flow from source s to receiver r, there must notion of information, because an information be as much information about a’s being F arriv- channel must screen off causal deviance. So ing at r as is generated by its occurrence at s. what is Foley’s example and does it work? Foley That will be possible only if the channel condi- focuses on Dretske’s example of three quick tions that permit information to flow remain knocks at the door causing a spy to believe that fixed. They must not themselves vary, thereby gen- the courier has arrived and carrying that informa- erating information or nonredundant informa- tion, as well. Foley modifies the case so that tion. Provided that the channel conditions remain it involves a wayward causal chain, as follows. fixed, they do not generate information in the The spy suddenly goes deaf. Then come three form of noise on the channel, with respect to the knocks carrying the information about the information that a is F. courier’s arrival. The knocks cause the spy’s part- Let’s consider an example. Suppose Al has a ner to trip, causing a box to fall on the spy’s metal detector that emits a tone when metal is head, in turn jarring the spy’s brain in such a within 10 inches of its detection surface. For its way that he suddenly comes to believe that the tone to carry the information that there is metal courier has arrived. present, the detector depends on several channel As we’ve seen, ITKs admit that a belief that p conditions. The power supply must be adequate can be knowledge, even if the belief that p does and charged. The magnetic field that detects the not contain the information that p, as long as metal must be in place. The wires that activate it was caused by the information that p. This the tone must be well functioning (no shorts invites the kind of example Foley gives. Still, it is or breaks in the wire). And so on. Were any of quite clear that an information-theoretic account these conditions not to remain fixed (a short needs the proximate cause of the belief (or sus- circuit, say), the detector may emit a tone when taining cause) to carry the information that p. In there is no metal being detected (even if there is Foley’s example, that is not the case. Even if the metal present). Therefore, the detector’s tone three knocks contained the information that p will carry the information that there is metal and for some strange reason the spy’s partner present only when all channel conditions are fixed. would not have tripped unless the knocks did In virtue of these fixed conditions, information contain that information, the rest of the story can flow from the source (detection of metal in doesn’t preserve information. The communica- the magnetic field of the surface) to the receiver tion channel has been broken by the time the (Al hears the emitted tone). box falls on the spy’s head and his brain is jarred. Notice that the tone carries information about The conditional probability that the courier has the presence of the metal, but it does not also

233 Fred Adams carry information about the channel conditions is no signal that K perceives and which carries the of the metal detector. We can use the tone to information that he has a brain lesion. But even tell us about the channel conditions, as we will setting this aside, it is clear that the brain lesion now see. But we cannot use it to tell us about its is not a fixed-channel condition. It introduces channel conditions and about presence of metal noise to K’s cognitive system. Indeed, Plantinga at the same time. If we know in advance that grants that the lesion causes K to believe mostly there is metal (or is none), we may check to see false propositions (a sure sign of noise). There- whether the detector emits a tone (or does not). fore, despite Plantinga’s claim, it simply is false Thereby, we can use old information about what that the conditional probability that K has a brain we already know (there is/is not metal present) lesion, given that K believes he has a lesion, is 1. to gain information about the channel condi- K’s beliefs do not track the truth, as Nozick tions of the detector. Now, because we already would say. K’s suffering from a lesion guarantees know whether or not there is metal present, we that he has a lesion, but not his belief that he has can test to see whether or not the detector is a lesion, not even if his belief is caused by the working properly. We can then tell that it is lesion (as for a tone from a metal detector with working properly and its channel conditions a short caused by its detection of metal). There- indeed are fixed (or tell that it is broken because fore, the example fails. its channel conditions are variable). But if we do However, Plantinga could fix the example not already know in advance about the presence so that there is no noise on the channel. So we of the metal, the tone carries no (new) informa- should consider what ITKs should say to such a tion about the channel conditions themselves. case. Indeed, there is a case by Lehrer that will This is not to say that channel conditions last do (1990: 163–4), his Truetemp case. Suppose forever. Metal detectors break or wear out. But that quite without his being aware, Mr. Truetemp the point is that for the detector to be a source has a high-tech, belief-producing thermometeric of information about the presence or absence of chip implanted in his brain. The chip measures metal, its channel conditions must be fixed. When the surrounding temperature and then directly are they fixed? When they generate no (or no enters a belief that it is that temperature into new) information. Truetemp’s belief box, if you will. Lehrer and It is now easy to see that Foley’s example is Plantinga would agree that Truetemp does not one where noise is introduced into the commun- know the temperature is 56 degrees Fahrenheit, ication channel because the channel conditions of when it is 56 degrees Fahrenheit, and Truetemp the man’s jarred brain are not fixed. The internal believes that it is. As before, there is no signal workings of such a brain introduce variability that Truetemp perceives and which carries the and noise to the system. The same would be information. The information is mainlined into true if there were a short in our metal detector Truetemp’s belief box, as it were. Suppose (when there was metal present). In neither case Truetemp is cognitively unable to withhold belief (hearing a tone or seeming to hear three knocks) despite having no evidence that his beliefs are would the relevant piece of information be carried correct. by the tone (or the auditory experience as of Now, unlike in Plantinga’s example, there do three knocks). seem to be fixed-channel conditions at work here. Similar remarks apply to an example by The thermometric-cognition mechanism is a fixed Plantinga. Suppose K suffers from a brain lesion condition (let’s suppose that it has been in place that causes K to believe a variety of mostly false for years). Since fixed, that it is working reliably propositions. It also causes K to believe that he generates no (or no new) information. It takes has a brain lesion, but K has no evidence for variation in the source (ambient temperature) this belief. Nonetheless, referring to Dretske’s and faithfully delivers information about it to ITK analysis above, Plantinga maintains that the the receiver (the belief box). It is an information- “probability on k & K is suffering from a brain delivery system no less faithful than the delivery lesion is 1” (Plantinga 1993: 195). Notice that systems of our senses. The probability that this is not a case of perceptual knowledge. There ambient temperature is 56 degrees Fahrenheit,

234 Knowledge given Truetemp’s belief that it is, is 1. The main fakes look like barns). So, it seems Al knows difference from our senses is that it skips the P&Q but not Q – intolerable. step of causing a conscious sensation that then But is this nasty consequence really forced causes a reliable belief about what is sensed. The upon an ITK theorist by this type of example? device in Truetemp skips the conscious sensory It is not. First, note that this example has the step, but its delivery of information is, we are same structure as the case above where Al thinks presuming, as stable and faithful as any normal both BMWs and Harleys are American made. sensory perception by standard perceivers under Al knows that the motorcycle he is working on standard conditions. is a Harley (P) and American made (Q). Does It would certainly be possible for an ITK to he not know that it is American made (Q)? It bite the bullet and insist that this is a case of might be thought that he doesn’t because he knowledge, albeit a very unusual kind of case mistakenly believes some BMWs to be American indeed. All that Plantinga or Lehrer have to fight made. But since he confuses no Harleys with this stand is a conflicting intuition and a con- BMWs, he does not believe the Harley is Amer- flicting theory. Since one probably should place ican made because he confuses Harleys with little stock in the conflict of intuition, in the end BMWs. He believes it is American made in part all they may have is a conflicting theory. Which BECAUSE he believes it is a Harley. So ITK would theory is correct will have to be determined by say that Al knows it is American because the looking at overall consistency and explanatory information that it is American is contained in power. Saying that this is indeed a case of know- the information that it is a Harley and Al knows ledge may be true and prove not to be so strange, he is working on a Harley. in the long run. It may come to grow on one. The Kripke example is no different. If Al Only time will tell. believes “this is a barn” (while standing before a As with other externalistic accounts of know- red barn illuminated by white light), does he ledge (Nozick 1981), ITK also rejects the fol- not know that it is a barn? ITKs submit that he lowing closure principle for knowledge: [(K(P) does because he is receiving the information that & K(P → Q)) → K(Q)] (To be read: knowing it is a barn (contained in the information that it that p and knowing that p implies q, entails know- is a red barn). This information is contained in ing that q.) Many find this rejection intolerable. his visual percept. Nothing anywhere near Al I will not here go into the arguments for or looks like a red barn, unless it is both red and a against the rejection of this principle (De Rose barn. So his percept carries the information that & Warfield 1999). However, some have pointed it is a red barn and this, in part, causes his belief out that theories such as ITK will also have to that it is a barn (and his belief that it is red). accept a particularly nasty consequence of the Thus, this particular example is not a case where rejection of closure, a consequence that is even conjunctive closure of knowledge fails. more intolerable. Following an unpublished Of course, whether it is possible for conjunctive example by Kripke, Lehrer (1990) offers the fol- closure of knowledge to fail may still be open. lowing example designed to show that one might Such failure is certainly consistent with ITKs know P&Q, but not know Q (call this a failure rejection of closure, generally. There may well of conjunctive closure). Suppose that a Holly- be cases of the following sort. One might believe wood movie-set has brought fake barns into your that something is a red barn because of one countryside. They have erected fake barns of piece of evidence and continue to believe it is a several colors, but not red ones. Indeed suppose barn by another piece of evidence. The first piece it is not possible to fake a red barn. Now if Al of evidence might contain the information that believes that something is a red barn (P&Q), he this thing looks red and looks to be a barn. The should know because the probability that it is a latter piece of evidence may contain only the red barn given that it looks like a red barn is 1. information that it looks to be a barn. The first But Al should not know that it is a barn (Q) contains the information that it is a barn. The because the probability that it is a barn, given latter carries only the information that it is a that it looks like a barn, is not 1 (because the barn or a fake (but not which). So conjunctive

235 Fred Adams closure might fail under this type of condition, Machines. [Explores avenues to escape the but not under the type of conditions given in the Bar-Hillel/Carnap paradox that contradictions or examples by Kripke or Lehrer. Therefore, ITKs logical impossibilities contain more information are not forced to accept the particularly nasty than the most important scientific discoveries. Not consequences of denial of closure charged by for the beginner.] Kripke and Lehrer. ——. Forthcoming b. “Is information meaningful Interestingly, if it is true that Al knows the data?” Philosophy and Phenomenological Research. structure is both red and a barn (and that the [Revises the definition of semantic information motorcycle is both a Harley and American), then by showing that false information is not a type of information, and that information is well- such cases would demonstrate the falsity of theo- formed, meaningful and truthful data. Not for ries like Lehrer’s (1990: 184). For Lehrer argues the beginner.] convincingly that his theory has as a consequence Foley, R. 1987. “Dretske’s ‘information-theoretic’ that Al does not know it is a red barn or that it is account of knowledge.” Synthese 70: 150–84. a barn. If Al knows both, as I believe to be the [Commentary on Dretske 1981. Accessible.] case, then Lehrer’s account of knowledge would Gettier, E. 1963. “Is justified true belief know- be false. As philosophers are fond of saying, one ledge?” Analysis 23: 121–3. [Easily accessible. person uses modus ponens where another uses Perhaps the most widely read paper on knowledge modus tollens. of the twentieth century. The answer is “no,” by the way.] Katz, J. 1998. Realistic Rationalism. Cambridge, References MA: MIT Press. [Excellent book that attempts to demonstrate the possibility to know mathemat- Adams, F. 1986. “The function of epistemic ical and logical truths despite not coming into justification.” Canadian Journal of Philosophy 16: causal contact with the objects known. Pitched 465–492. [This paper argues that if epistemic at the mature student.] justification permits justified false belief, then it Lehrer, K. 1990. Theory of Knowledge. Boulder, CO: is neither sufficient nor necessary for knowledge. Westview Press. [Overview of the problems and For mature students.] topics of knowledge, and presents Lehrer’s own Barwise, J. and Perry, J. 1983. Situations and systematic view. Accessible, but rigorous.] Attitudes. Cambridge, MA: MIT/Bradford. [An Lewis, D. 1996. “Elusive knowledge.” Australasian important systematic treatment of the logic and Journal of Philosophy 74: 549–67. [Presents and semantics of sentences about language, meaning, defends Lewis’s own view of knowledge, in light belief, and knowledge. Not for the beginner.] of the relevant literature since 1963. Not for the DeRose, K. and Warfield, T., eds. 1999. Skepticism: beginner in philosophy.] A Contemporary Reader. Oxford: Oxford Univer- Nozick, R. 1981. Philosophical Explanations. Cam- sity Press. [Excellent collection of 15 papers on bridge, MA: Harvard University Press. [First-rate skepticism, knowledge and closure, and know- book on many problems of philosophy, including ledge and . Accessible.] Nozick’s own very influential theory of knowledge. Dretske, F. 1981. Knowledge and the Flow of Accessible to the student with some appreciation Information. Cambridge, MA: MIT/Bradford. [A of issues in symbolic logic and counterfactuals.] seminal attempt to apply developments of the Plantinga, A. 1993. Warrant: The Current Debate. mathematical theory of communication to philo- Oxford: Oxford University Press. [Plantinga’s sophical issues of knowledge, belief, concepts, history of post-Gettier epistemology about war- and meaning. Accessible to a beginner.] rant – i.e., what it is that when added to justified ——. 1982. A Cognitive Cul-de-sac, Mind, 91, true belief yields knowledge.] 109–111. [Argues that avenues of knowledge may Weaver, W. and Shannon, C. 1949. The Mathemat- dead-end in some knowers – s may knows that p, ical Theory of Communication. Urbana: The Uni- but be unable to transmit that knowledge to versity of Illinois Press. [A classic mathematical another. Accessible to the beginner.] treatment of information and communication that Floridi, L. Forthcoming a. “Outline of a theory has been widely influential in philosophy and of strongly semantic information.” Minds & cognitive science.]

236 The Philosophy of Computer Languages

Chapter 18 The Philosophy of Computer Languages Graham White

tion did what it was claimed to do. However, 1 Introduction: one might also want to answer more general Two Semantic Projects questions: one might want to design a computer language, and an intended semantics for that Consider a theory whose aim is to say, for a language is often a good place to start. Or one given language, what each of its expressions might want to classify existing computer lan- means. Call it a semantics of that language. guages: it is clear that some of them are more What might be required of such a theory before similar to each other than to others, and that it was allowed that it accomplished its aim or some languages are merely notational variants of did so in an optimal way? (Travis 1986) each other, but how do we put such observa- tions on a more formal footing? And, finally, This is a question which belongs to a well- we might be tempted to say something about established philosophical program: the Davidson– the nature of computation on the basis of the Dummett program of using a formal semantics semantics of the languages in which we express – a “theory of meaning,” as these philosophers computations. call it – in order to attack philosophical questions I would claim that the semantics of computer about the relation between language and reality, language has considerable philosophical interest: or between mind and language (see, for example, it has a different motivation to the usual philo- Wiggins 1997). sophical approach to computation via Turing There is also, among theoretical computer sci- machines, and, correspondingly, it yields differ- entists, a similar area of study: it is usually called ent insights. One of the main difference is that “the semantics of computer languages,” or often the semantics of programming languages is con- simply “semantics.” Its aim is, likewise, to develop cerned with the languages in which people actu- a formal account of the meaning of a given com- ally program, and, particularly, with the languages puter language, and to use that account to answer which have been found to be good to program interesting questions. in; it is thus inescapably connected with the prac- These questions might be severely practical – tice of programming. It is emphatically not a one might, for example, want to formally verify, discipline which is developed out of some a priori using such a semantics, that the language in ques- notion of computation. And, in fact, it has

237 Graham White connections with areas of philosophy which may utterly mysterious. The latter was an extremely seem surprising; Quine’s concept of referential serious problem: without some sort of semantics, transparency, for example, is an important part it was hard to say what counted as a correct of the programming-language enterprise. Some implementation of these languages. Although of this material is quite technical, and is also gen- Lisp eventually achieved a precise semantics, it erally unfamiliar to philosophers (even to those was designed by starting from the implementa- who know the usual technical repertoire of philo- tion and then attempting to find mathematical sophical logic). I have segregated many of the structure in the resulting language: several of more technical details into sections of their own, the Lisp primitives were called after hardware which can, at first reading, be omitted. features of the machine that it was originally implemented on (McCarthy 1981: 175), whereas its original semantics was “ramshackle” (Landin 2000). Nevertheless, it was also possible to see 2 History that the rewards for a precise syntax – or, more ambitiously, a precise semantics – were extremely The semantics of programming languages grew high: Algol “proved to be an object of stunning up in a particular historical context, and it is beauty” (Perlis 1981: 88). worth spending some time describing it: it was developed by a group of philosophically literate mathematicians and computer scientists, and the 2.2 Algol-like languages philosophical influences are quite evident. They are also little known: the history of computing What, then, do these programming languages has tended to focus very much on the very early look like? We will describe a generic language, days, and, at that, mostly on the history of hard- quite similar to Algol; since Algol has had an ware, so that the history of these topics is doubly enormous influence on language design its fea- neglected. tures can be found in many others. These lan- guages have some similarity to formal languages like first-order logic: like these languages, they 2.1 The first programming languages have variables, to which values can be assigned, and they have both predicates and functions. In the early days of computers, programming And many of the basic operations of program- was done by directly writing machine instructions: ming can be viewed as the assignment of values this was difficult and error-prone. Programming to variables, so one might think that these opera- languages were invented to allow programmers tions, too, could be viewed in this way. to write in a more comprehensible form: However, programming languages also have Fortran dates from 1954, and made it possible features which are strikingly different from the to write programs in a notation very like that of sort of logical languages familiar to philosophers. standard mathematics (Backus 1981). Although In large part, these other features come from the Fortran designers paid very little attention to a need to control the structure of programs: theory – Backus, the leader of the project, says programs are extraordinary large entities, and “we simply made up the language as we went programmers can only keep control of this com- along” (1981: 30) – both syntax and semantics plexity by making programs out of smaller com- soon became important factors in the design of ponents, which can be individually constructed programming languages. Algol was designed over and tested and, if possible, reused in many dif- the period 1958–60 (Naur 1981; Perlis 1981), ferent programs. Programs, then, tend to be Lisp from 1958 to 1962 (McCarthy 1981), and made of hierarchically nested components; there many of the difficulties of developing these lan- are various names for these components, but we guages were due to two factors: it was difficult to can – following Algol usage – call them blocks. define the syntax of a language at all precisely, Furthermore, unlike the logical languages and the semantics of these languages seemed which philosophers are familiar with – namely,

238 The Philosophy of Computer Languages variants of untyped first-order logic – modern This requirement means that the semantic programming languages are typed: variables, and values assigned to these entities must belong to a the values that they take on, have types. This is quite intricate system. To see this, consider a par- partly for practical reasons: many programming ticular case of programming entities: namely, those errors can be detected automatically, simply by that are sometimes called subroutines. These are checking the types of the entities involved. But pieces of code that have parameters: they are there are also rather deeper reasons: program- invoked with particular values of their parameters, ming does not fit very well into a set-theoretic and they then perform various actions on them. view of things, since sets – even finite sets – are A subroutine, then, can be thought of as a sort of collections without any extra structure, and, how- function: its arguments are the semantic values ever we may choose to store data in a computer, of its parameters, and its value is the semantic we always do so in some structured way (the value of the expression that it returns. So the data may be ordered, or arranged on the leaves semantic value of a subroutine must be a function of a tree, or the items may be mapped to integers, type: it maps its argument types to its return type. and so on). But now consider a subroutine which takes There is a final difference, which is extremely a subroutine as a parameter. Such things occur far-reaching. Most programming languages allow frequently in the normal practice of program- programs to perform actions that change the ming: for example, we might want to write a values of variables, or which have other irreversible subroutine which constructed a button in a user effects (input or output, for example); we say that interface. Buttons can perform various actions, these actions have side-effects. These features and – because we are writing a subroutine which add further complications to the task of giving we can use for constructing all sorts of buttons – semantics to these programming languages. we want to be able to give the code for the action to the button code as a parameter. The “code for the action,” of course, is itself a subroutine: 2.3 The development of semantics so the code for the button is a subroutine which has a subroutine as one of its parameters. The The first steps towards the semantics of these need for “higher-order” entities of this sort seems languages were taken, in the 1960s, by a group to be natural and pervasive in programming: our – Peter Landin, Dana Scott, and others – example is from user-interface programming, associated with Christopher Strachey (Scott but there is a differently motivated example in 1977; Landin 2000). They provided what is Abelson et al. 1996: 21–31. called a denotational semantics: that is, rather A subroutine with a subroutine as argument, than describe the operations that pieces of code then, can be considered as a piece of code which perform, they associated mathematical objects – takes the subroutine and returns a value: so its denotations, or semantic values – to the syn- semantic value will map the semantic value of tactic entities of a programming language. Values its subroutine parameter to the semantic value are assigned to entities on all scales: the variables of its result. The semantic value of subroutine- (and constants) of the language get values, of calling code, then, is a function which takes course, but so do assignment statements – the another function as an argument: in technical parts of programs which give values to variables terms, it is a functional. – as do blocks and subroutines and, finally, the We must also remember that the functions entire program. This requirement – that pro- corresponding to subroutines cannot, in general, grams should have semantic values on all scales be everywhere defined. We know, from the the- – is part of the basic program of denotational ory of recursive functions, that any reasonably semantics: it is motivated by the view that pro- expressive programming language must have gramming constructs which have well-defined programming constructs – looping, recursion, semantic values will be easy to reason about, and or both – which cannot be guaranteed to give it has, over the years, shown itself to be quite a result in all cases. This must be accommodated justified. in the semantic values of such subroutines.

239 Graham White

2.3.1 Technical interlude: semantic function f(x: int): int values in detail begin y := x; The initial stages of this accommodation are quite f := y + 1 easy to see: we can deal with partial functions end from, let us say, the integers to the integers by regarding them as functions from int – the usual which first assigns the value of its argument to a integers – to intЌ – the integers together with global variable y, and then returns y + 1. Con- an extra element, Ќ, which is the value of the sider also the simpler subroutine function when the computation fails to termin- ate. But now a subroutine which takes such a function g(x: int): int subroutine as argument must itself have a type begin which is not merely the type of functions (int → g := x + 1; int) → int, but rather the type of functions (int end → intЌ) → intЌ (which is, one should point out, much more complex than ((int → int) → f and g yield the same values for the same int)Ќ). The moral is clear: although we only have arguments, but they are not substitutable, one to add a single extra element to our base types, for the other: g changes the values of a global the changes required at higher types become variable, which f does not. (We say that f has progressively more complex. side-effects.) Recursion, also, needs a suitable treatment. Accommodating this sort of behavior, and Consider a recursively defined subroutine, for still preserving the compositional nature of our example: semantics, makes the type system for our semantic values somewhat complex and intricate – see function f(x:int): int Tennent 1994: 250ff for details. begin The development of semantics, then, is a pro- if (x = 0) then f := 1 else cess of progressive elaboration of semantic values. (f := x * f(x − 1)) We might think of it like this: originally, we end have a straightforward conception of what the values of programming entities are: variables stand We can regard this as saying that the subroutine for their values, subroutines stand for functions f is a fixed point of a certain operation, namely from parameters to return values, and so on. We the operation which takes a subroutine F as may call this original conception the intended input and returns, as output, the subroutine G, semantics. However, it proves impossible to pre- defined by serve substitutivity with this intended semantics, so we have to progressively elaborate the semantic function G(x:int): int values that we assign to programs and their parts; begin in the process, these semantic values become if (x = 0) then G :=1 else further and further removed from the original, (G := x * F(x − 1)) intended semantics. end The divergences are caused by phenomena that can be viewed, when measured against the So the semantic value of f must be fixed under intended values, as a lack of referential trans- parency. The intended values of subroutines such ۋ the semantic counterpart of the operation F G; and thus, to handle recursion, the appro- as these ought to be given by the maps, from priate semantic domains must be closed under arguments to return values, that they induce, certain fixed-point operations. Finally, the need but, as we have seen, subroutines with the same to accommodate assignment statements brings intended values might not be intersubstitutable: another complication. Consider the following and such a failure of substitutivity is, in Quinean subroutine: terms, described as referential opacity.

240 The Philosophy of Computer Languages

we define a subroutine – call it S, and that we 3 The Uses of Semantics later invoke it. Suppose also that, in the code defining S, there is a global variable x. Suppose, A working semantics on these lines can, indeed, finally, that we change the definition of x be- be achieved (Stoy 1977; Tennent 1994), and tween the time that S is defined and the time such semantic accounts of programming language that it is invoked. Which value do we use for x? have been widely used. There are two obvious choices: However, the uses are not as direct as one might imagine. It is rarely expedient, for exam- 1. the value it had when S was defined: this is ple, to establish correctness for a particular pro- called lexical binding, and gram by examining the semantic values of it and 2. the value it had when S was invoked: this is its components: these semantic values are usu- called dynamic binding. ally extremely complex. They must necessarily be complex: since it is possible to decide whether It turns out (Stoy 1977: 46ff) that lexical bind- a given program terminates or not, purely on ing gives a language much better substitution the basis of its semantic value, there must be properties, and, in fact, languages with dynamic facts about the semantic values of programs that binding – the typesetting language TeX, for are as difficult to establish as the halting prob- example – are terribly difficult to program with. lem (i.e. undecidable). More generally, it seems to be the case that, if a On the other hand, semantics has a large language has clean, elegant semantic properties, number of metatheoretical uses. One can, for then it will be easy to program in. example, establish equivalences between pro- grams, and, more generally, one can develop, and semantically justify, logics (the so-called 3.1 Identity of programs Floyd–Hoare logics) for reasoning about pro- grams (Tennent 1994: 196ff; Jones 1992); and, We use semantics, then, to conclude facts about unlike direct reasoning with semantic values, these the behavior of programs on the basis of math- logics are, for typical problems, easy to work ematical properties of their semantic values. We with. could, for example, observe that, if the semantic There is another, less formal but very per- values of programs P and Q were different, then vasive, use of semantics. The practice of pro- the programs themselves must be different. gramming involves a great deal of substitution: This is more subtle than it might seem. What replacement of one subroutine by another (or do we mean by identity and difference between one object by another, one library by another, programs? A trivial answer would be that it and so on). We like to have languages in which simply consisted in the identity or difference of substitutions like this are easy to justify: if we their source code: but this is rarely of any inter- can be sure that, if two items “behave the same” est. Programs can vary a good deal, in a merely (in some suitably informal sense) they can safely notational way, and still remain “essentially the be substituted for each other. In the develop- same” (whatever that might mean). ment of semantics, it very soon became apparent A better criterion for the identity of programs that the semantic properties of languages were is that of observational equivalence; philosoph- decisive for this question: that certain semantic ically, it can be regarded as a sort of functionalism properties made substitution behave well. (see Lycan 1995, Block 1995). One definition is as follows: Two programs, P and Q , are observa- tionally equivalent if and only if, whenever the 3.0.1 Technical interlude: an issue in inputs of P and Q are the same, then so are their programming-language design outputs. We can define this also for constituents of Here is an example of the sort of guidance that programs (subroutines, statements, blocks, and semantics can give in language design. Suppose the like: these are generically called program

241 Graham White phrases). We define equivalence by observing what way – for example, in terms of games (Abramsky happens when phrases are substituted for each et al. 1994; Hyland & Ong 2000) – we have a other in programs (here a program with a phrase mathematical object which tells us a great deal deleted is called a program context): about the deep structure of a particular program- ming language. Two program phrases, f and g, are observa- tionally equivalent if and only if, for any pro- gram context P(·), the two programs P(f ) 3.2 Functional programming and P(g) are observationally equivalent. We have been describing an approach to the Observational equivalence is an extremely theory of programming languages which simply versatile property. If we think of anything which seeks to analyze the usual languages that people might be (in the nontechnical sense) an “observa- program in. We might, though, take a different tion” of the behavior of a program – stimulating approach to language design: we might want it with certain input, checking the output, and the theory to be more prescriptive, and design so on – we can automate this observation by programming languages so that they had a good, writing another program to perform it. This other perspicuous, metatheory. program will provide a program context with One of the features which give languages a which we can test the program that we are inter- good metatheory is referential transparency: the ested in: and thus our definition of observational property that terms of the language, which stand equivalence can be regarded as an automated for the same entities, can always be substituted version of the everyday concept of observation. for each other. Languages with side-effects – such However, observational equivalence is a diffi- as statements that change the values of variables cult property to establish, since it talks about – do not have referential transparency. A term of what happens when a program fragment is sub- such a language might stand for, let us say, a stituted into any program context at all, and in number, but might also, in the course of evalu- most cases we have no grasp of this totality of ating that number, change the values of a par- program concepts. ticular variable; another term might evaluate to If our semantics respects observational equi- the same number, and might change the values valence, then we call it fully abstract: of other variables; and it is easy to see that these two terms, even though they evaluated to the A semantic valuation υ(·) is fully abstract if, same numbers, could not be substituted for each whenever program phrases f and g are obser- other. vationally equivalent, υ(f ) and υ( g) are the So we might consider designing a program- same. ming language in which we could not change the values of variables. What would such a language Full abstraction is, of course, an important con- look like? There could be variables, and we could cept, because we are interested in observational have definitions that assigned values to variables: equivalence. But its interest is rather wider than however, once we have let a variable have a cer- that: a fully abstract semantics will, in some way, tain value, we could not subsequently change reflect the essential structure of programs, ab- it. We could also have subroutines: subroutines stracting away from notational or implementa- would take parameters and return values. Because tional details (Tennent 1994: 242). Of course, we have no assignment statements, the result we can – rather fraudulently – define fully abstract returned by a subroutine on particular arguments semantics by starting with a non-fully abstract depends only on the values of its arguments: the semantics and imposing equivalence relations on same subroutine, evaluated on the same argu- it; but unless we have independent access to the ment, always yields the same result. Subroutines, model thus constructed, it would do us no good. then, are extensional: they give the same results In the case when we can find a fully abstract on arguments with the same referents, and in this model, and characterize it in some meaningful respect they behave like mathematical functions.

242 The Philosophy of Computer Languages

Following Quine, it is usual to refer to languages We can go on from here. It is known that with this property as referentially transparent. terms in a suitable λ-calculus can be used to As we have seen, programming needs higher- encode proofs in higher-order intuitionistic logic order constructs. These languages are no excep- (Lambek & Scott 1986); this is called the Curry– tion: we can let higher-order entities (in this Howard correspondence. This suggests that func- case, functions and higher-order functionals) be tional programs can also be considered to be the values of variables, we can pass them as para- proofs in that logic: and such in fact is the case. meters to subroutines, and so on. Following The correspondence between programs and Quine’s slogan “to be is to be the value of a proofs is illuminating in its own right. Consider variable” – a slogan explicitly used by the pioneers a subroutine which takes a parameter – say x – of programming-language semantics – we can, and which computes, for example, 3x + 2. This and do, give a rough ontology to our language: corresponds to the lambda-term the entities that can be the values of variables, that are passed to and returned by subroutines, λx :int.3x + 2, are usually known as “first-class citizens,” and they will figure largely in any account of the which corresponds to a proof of the proposition semantics of the language. These languages – known as functional languages – generally have ∀x :int∃y :int a large array of such higher-order constructs. Lisp is such a language, but is semantically some- but not, of course, just any proof: it is the proof what impure: modern, more principled versions which takes the integer x introduced by ∀, com- are untyped languages such as Scheme, and typed putes 3x + 2, and then uses that integer as a languages such as ML. premise in ∃-introduction.

3.2.1 Technical interlude: 4 Conclusions the lambda calculus

There is an alternative description of these We have surveyed a rather large area of theoret- languages. Consider a subroutine, such as the ical computer science; we now have to consider following: its philosophical relevance. There are, of course, several points of direct relevance: programming- begin function f(x, y): language semantics tells us a great deal about return 3 * x + 2 * y the processes of abstraction involved in program- end ming computers, and also about the nature of algorithms (for which Moschovakis 2001 is a good comparison). However, there are less direct This is the subroutine which takes two para- points of interest. The overall goal of program- meters, x and y, and returns 3x + 2y. We can give ming semantics is quite similar to the philosoph- an alternative description of this – in more math- ical project of developing a theory of meaning: ematical notation – as a term in the λ-calculus: however, the methods and results are strikingly difficult. To a large extent this is because the λ λ x. y.(3x + 2y) philosophical project has been developed in isolation, with unsophisticated technical tools, This rough analogy can be made precise: we can and with the aid of a very small number of set up a correspondence between functional pro- examples, none of them either large or com- grams and λ-calculus terms, which is composi- plex. The practical necessities of producing a tional and extensional, in such a way that we can useful body of theory have made it impossible obtain a semantics for our programs from a for programming-language semantics to indulge semantics for the λ-calculus. in any of these luxuries. So, the comparative use

243 Graham White of programming-language semantics is, perhaps, expressions of our programming languages result more interesting than its direct use. in, when they are suitably compiled and run. However, we are very rarely interested: looking at computers on that sort of level would sub- 4.1 What aren’t we interested in merge the interesting features in an ocean of low-level detail. We would be unable to dis- Programming-language semantics was developed tinguish, on that level, between operations which in order to address certain specific needs, and were performed by the operating system and one must understand the biases resulting from those which were performed by programs; of those needs to be able to understand the theory the programs running on a real computer, by far and its place in the world. the majority would be concerned with trivial housekeeping tasks, rather than anything we were We know the mechanism The first is obvious: interested in: of the instructions which execute we know all about the mechanisms of computers, in an interesting program, by far the majority because, after all, we made them. This contrasts of them would be concerned with rather dull very strongly with the situation in the philosophy tasks such as redrawing windows on the screen, of language and in linguistics, where we have very interacting with the operating system, and so little information about underlying neurophysio- on. On the level of machine-level instructions, logical mechanisms. none of these processes would be distinguishable from each other in any tractable way. We design the systems We also design computers, their operating systems, and the programming lan- guages that we use on them. Many of the choices 4.2 What are we interested in? that we make when we do this are reflections of our needs, rather than of the nature of computa- So what are the interesting problems? tion as such; for example, operating systems are extremely modular, and most programming lan- Making languages different There is a huge guages have a great deal of support for modular- number of different programming languages, ity, simply because modularity makes computers and there are also genuine differences between much easier to program. Modularity does not them. If we were to analyze these languages using seem to be entailed by the nature of computation the methods of recursive function theory, or as such. By contrast, Fodor’s work (Fodor 1983) by representing programs written in them as deploys much more transcendental premises: he is Turing-machine programs, we would find (since attempting to establish that any minds such as programming languages are generally Turing ours must be modular, whatever their mechan- complete) that they were indistinguishable from isms and however those mechanisms might have each other. So we want a theory that is finely arisen. The semantics of programming languages grained enough to be able to represent the can, of course, neither prove nor disprove the genuine differences between languages. On the validity of a program like Fodor’s – though, of other hand, we do not want to differentiate lan- course, it might provide illuminating results on the guages that are merely typographical variants of nature of modularity and on its formal analysis. each other, or which differ simply by trivial defini- tional extensions: we want, that is, a theory that Reference is not problematic If we are using is sensitive to genuine differences between lan- computers to solve a problem in the real world, guages, and only to those. We would also like to we generally know what the expressions of our go on and construct a taxonomy of languages: programming language stand for: we have, if we that is, we would like to arrange languages in are sensible, set things up that way. some sort of formal scheme in which we could describe how to get from one language to Foundationalism is not interesting We may, in another by regularly varying theoretical para- principle, know the physical processes that the meters of some sort.

244 The Philosophy of Computer Languages

Attaining abstraction There is a common we want to make programs correspond to proofs, theme running through all of these considera- there must be a large number of essentially dif- tions. When we are designing, or using, a high- ferent proofs of the same proposition. We can level programming language, we are concerned rephrase this in terms of equivalence of proofs: about attaining a sufficient degree of abstraction. we should be able to define a notion of proof Low-level, detailed, grounded descriptions of our equivalence which disregards merely notational systems are unproblematic, but we do not want variation, but which is not so coarse that the set these: we want to be able to forget about such of equivalence classes becomes trivial. It turns merely implementational details in order to pro- out that, for technical reasons, we can do this gram, and reason about programs, at the level for intuitionist logic, but not for classical logic we are interested in. In order to do this, we need (Girard 1991). to be able to design languages to do it; and in However, this use of intuitionist logic is less order to do that, we need some sort of theoret- ideological than it might seem. We are using it ical conception of what these languages should because we need a finely-grained proof theory, look like. Thus, our semantics should give us a and it would be perfectly possible for an ideo- view of computational processes which is equally logically classical logician to use intuitionist logic as abstract as the languages that we want to for these purposes, only because, for computa- design. Abstraction, then, is an achievement. tional purposes, one needed a fine-grained proof This contrasts sharply with the traditional task theory. of the philosophy of language. Here we start In a similar way, we use higher-order with a high-level view – a speaker’s intuitions logic: higher-order constructions are pervasive about language – and we attempt to find a suit- in programming, and it is appropriate to have a ably grounded account of this high-level view metatheory which reflects that. However, this (Dummett 1991: 13). Attaining the abstract view preference, again, is not straightforwardly ideo- is not a problem: grounding it is. In the semantics logical: these higher-order entities are, after all, of computation, on the other hand, we already algorithms, and carry no taint of the infinite. have a grounded view of our subject-matter: it is Correspondingly, there are constructive models the construction of a suitably abstract view that of set theory in which sets are modeled by equi- is the major difficulty. valence relations on subsets of the integers: we can, in such models, carry out all of the construc- tions needed to develop programming-language 4.3 The technical tools semantics, although we cannot quite accommod- ate all of traditional higher-order logic (Robinson The technical tools used also differ strongly from 1989; Hyland 1982). There is very little that a those current in the philosophy of language com- constructivist can find about such models to munity. Programming-language semanticists use object to. higher-order logic and the mathematical theory Category theory is also part of the semantic of categories: linguistic philosophers use first- toolkit (see Tennent 1994: 290ff for some ex- order logic and set theory. This is a fairly pro- amples). But this is hardly any surprise: category found difference in mathematical cultures, but theory has found wide application in areas it also has to do with the difference between the of mathematics – algebraic topology, algebraic problems that these communities are addressing. geometry, and proof theory – where one wants to disregard “implementational” (or merely nota- Intuitionistic logic Semantics uses intuitionist tional) detail, and concentrate on the essential logic a great deal: we have seen, above, that the features of a situation. semantics of functional programming looks very This can be rephrased in more traditional like the proof theory of higher-order intuitionist philosophical terms as follows. There is a well- logic, because programs correspond to proofs known example, due to Benacerraf (1965), which of certain propositions. For this, we must use is (slightly rephrased) as follows: consider two intuitionist, rather than classical, logic: because mathematicians (A and B) who both talk about

245 Graham White ordered pairs, except that A encodes the ordered Truth is not particularly salient in the semantics pair as {x, {x, y}}, whereas B encodes it as of programming languages, but we do have an {y, {x, y}}. Now – though A and B clearly each important central notion: that of observational have accounts of ordered pairs which are math- equivalence. So, if we do have a fully abstract ematically adequate – they do not seem to be semantics, then it can (after suitable manipulation talking about the same objects (and, in fact, they to express it in philosopher-friendly terms) be can be made to disagree by asking them stupid construed as a sort of counterpart of a theory of questions of the form (“is x a member of ?”)). One approach to this would be to invoke we can derive a great number of conclusions about a difference between specification and imple- observational equivalence of computer languages. mentation, and to say that A and B were simply And there are such fully abstract semantics. using different implementations of a single speci- However, there are one or two caveats to be fication. Of course, to do that we need to have made. There is a presumption that, when one some way of making these specifications explicit: had achieved a theory of meaning, one could and category theory gives us that. We can, given simply examine to see what its “central notion” two sets X and Y, specify their Cartesian product was (Dummett 1991: 34). But these semantic X × Y (the set of ordered pairs with members in theories are possibly a little more recalcitrant: each set) in terms of the two maps X × Y → X they assign mathematical objects – semantic values and X × Y → Y, and of the properties of these – to program phrases, but these mathematical two maps. And this characterization turns out objects do not wear their meanings on their to characterize the construction exactly, without sleeve: there is still room for considerable argu- involving any purely implementational decisions. ment about what they mean. We may, it is true, We can think of category theory as ruling out present mathematical objects using vocabulary the stupid questions which differentiated between which is sufficiently tendentious to make one A’s and B’s mathematics: that is, of giving us a think that they have a clear and obvious mean- distinction between observable and unobservable ing: but this would be merely tendentious, hav- properties of mathematical constructions. The ing to do with a particular presentation of those observable properties are those which can be ex- objects. pressed in terms of mappings (“morphisms,” in The other caveat is this. The semantics of pro- category-theoretic terms) between mathematical gramming languages has paid particular attention objects, and in terms of identities between those to the question of full abstraction: this concept morphisms; the unobservable ones need identities has been somewhat neglected in the philosophy between objects (Bénabou 1985). It is no surprise, of language, where the problems have seemed then, that programming-language semantics, to be those of finding rich enough semantic which is intimately tied to the observable prop- values to hold all of the components of meaning erties of computer programs, also uses category that we want. However, full abstraction ought theory to express that notion of observability. to play a role in the philosophy of language as well: as Quine said, there should be no iden- tity without identity (Quine 1969: 23), and the 4.4 Theories of meaning semantic values that we assign to sentence frag- ments should, in some way, respect the identities Finally, we should compare these semantic the- of the meanings that we are trying to model. ories with the philosophical project of a theory of meaning. We should recall that the goal of Davidson’s program was to develop an axiomatic References theory which would yield, for each sentence of a natural language, a suitable instance of the Abelson, Harold, Sussman, Gerald Jay, and Sussman, schema (Dummett 1991: 63) Julie. 1996. Structure and Interpretation of Com- puter Programs, 2nd ed. Cambridge, MA: MIT S is true if and only if A. Press. [An outstanding, and semantically aware,

246 The Philosophy of Computer Languages

programming textbook, using the functional Landin, Peter J. 2000. “My years with Strachey.” language Scheme.] Higher Order and Symbolic Computation 13: 75–6. Abramsky, Samson, Jagadeesan, Radha, and Lycan, William G. 1995. “Functionalism (1).” In Malacaria, Pasquale. 1994. “Full abstraction for Guttenplan 1995. PCF (extended abstract).” In M. Hagiya and McCarthy, John. 1981. “History of Lisp.” In R. L. J. C. Mitchell, eds., Theoretical Aspects of Com- Wexelblat, ed., History of Programming Lan- puter Software. Lecture Notes in Computer guages. New York: Academic Press. From the Science no. 789. New York: Springer-Verlag. ACM SIGPLAN History of Programming Lan- Backus, John. 1981. “The history of Fortran I, guages Conference, June 1–3, 1978. II and III.” In R. L. Wexelblat, ed., History of Moschovakis, Yiannis N. 2001. What is an algo- Programming Languages. New York: Academic rithm? Pages 919–936 of : Engquist, Björn, & Press. From the ACM SIGPLAN History of Schmid, Wilfried (eds), Mathematics unlimited – Programming Languages Conference, June 1–3, 2001 and Beyond. Berlin: Springer. 1978. Naur, Peter. 1981. “The European side of the last Bénabou, Jean. 1985. “Fibred categories and the phase of the development of Algol.” In R. L. foundations of naïve category theory.” Journal of Wexelblat, ed., History of Programming Lan- Symbolic Logic 50: 10–37. guages. New York: Academic Press. From the Benacerraf, Paul. 1965. “What numbers cannot be.” ACM SIGPLAN History of Programming Lan- Philosophical Review 1974. guages Conference, June 1–3, 1978. Block, Ned. 1995. “Functionalism (2).” In Perlis, Alan J. 1981. “The American side of the Guttenplan 1995. development of Algol.” In R. L. Wexelblat, ed., Fodor, Jerry A. 1983. The Modularity of Mind: An History of Programming Languages. New York: Essay on Faculty Psychology. Cambridge, MA: MIT Academic Press. From the ACM SIGPLAN Press. History of Programming Languages Conference, Dummett, Michael. 1991. The Logical Basis of Meta- June 1–3, 1978. physics. London: Duckworth. Quine, Willard van Ormond. 1969. “Speaking Girard, Jean-Yves. 1991. “A new constructive logic: of objects.” In Ontological Relativity and Other classical logic.” Mathematical Structures In Com- Essays. New York: Columbia University Press. puter Science 1(3): 255–96. Robinson, E. 1989. “How Complete is PER?” In Pro- Guttenplan, Samuel, ed. 1995. A Companion to the ceedings of the Fourth Annual Symposium on Logic Philosophy of Mind. Oxford: Blackwell. in Computer Science. IEEE Press: Washington. Hyland, J. M. E. and Ong, C.-H. L. 2000. “On Scott, Dana S. 1977. “Introduction.” In Stoy 1977. full abstraction for PCF.” Information and Com- Stoy, Joseph E. 1977. Denotational Semantics: The putation 163: 285–408. Scott–Strachey Approach to Programming Lan- Hyland, Martin. 1982. “The effective topos.” In guage Theory. Cambridge, MA: MIT Press. A. S. Troelstra and D. van Dalen, eds., L. E. J. Tennent, R. D. 1994. “Denotational semantics.” Brouwer Centenary Symposium. Studies in Logic S. Abramsky, D. M. Gabbay, and T. S. E. and the Foundations of Mathematics no. 110. Maibaum, eds., Handbook of Logic in Computer Amsterdam: North-Holland. Science, vol. 3. Oxford: Oxford University Press. Jones, C. B. 1992. The Search for Tractable Ways of Travis, Charles. 1986. “Introduction.” In C. Travis, Reasoning About Programs. Tech. rept. UMCS- ed., Meaning and Interpretation. Oxford: 92-4-4. Dept. of Computer Science, University Blackwell. of Manchester. Wiggins, David. 1997. “Meaning and theories of Lambek, J. and Scott, P. J. 1986. Introduction to meaning: meaning and truth conditions from Higher Order Categorical Logic. Cambridge Stud- Frege’s grand design to Davidson’s.” In B. Hale ies in Advanced Mathematics no. 7. Cambridge: and C. Wright, eds., A Companion to the Philo- Cambridge University Press. sophy of Language. Oxford: Blackwell.

247 Thierry Bardini

Chapter 19 Hypertext Thierry Bardini

• Hypertext is: (1) a form of electronic docu- Introduction: Defining Hypertext ment; (2) an approach to information man- agement in which data is stored in a network Defining hypertext can be confusing. Kathleen of nodes and links. It is viewed through inter- Gygi (1990: 282) categorized available defini- active browsers and manipulated through a tions into two types, “broad-spectrum” (Group structure editor. I) and the “more clinical variety” (Group II). • Hypertext connotes a technique for organiz- She found Group I definitions in the popular ing textual information in a complex, non- press and in advertising and marketing literature, linear way to facilitate the rapid exploration and Group II definitions in technical journals of large bodies of knowledge. Conceptually, and research efforts at developing computer- a hypertext database may be thought of as a supported hypertext systems. She gave the fol- directed graph, where each node of the graph lowing examples: is a (usually short) chunk of text, and where the edges of the graph connect each text Group I chunk to other related text chunks. An inter- • Hypertext works by association rather than face is provided to permit the user to view indexing. the text in such a database, traversing links • Hypertext is a format for nonsequential rep- as desired to explore new areas of interest as resentation of ideas. they arise, check background information, and • Hypertext is the abolition of the traditional, so forth. linear approach to information display and • Windows on the screen are associated with processing. objects in a database, and links are provided • Hypertext is nonlinear and dynamic. between these objects, both graphically (as • In hypertext, content is not bound by struc- labeled tokens) and in the database (as ture and organization. pointers).

Group II More recently, Luciano Floridi (1999) • Hypermedia is a style of building systems has synthesized these Group II definitions to for information representation and manage- put the emphasis on the three necessary ele- ment around a network of nodes connected ments that make a hypertext. He described together by typed links. them as:

248 Hypertext

1 a discrete set of semantic units (nodes). fact that we are still talking about text, authors, . . . These units, defined by Roland Barthes and readers, writing and reading, when we talk as lexia . . . , can be about hypertext. It is quite a different position 1.1 alphanumeric documents (pure hypertext) that we will argue for in this chapter, where we 1.2 multimedia documents (hypermedia) will wonder how exactly a set of electronic links 1.3 functional units . . . , in which case we between lexia can actually embody anything. have the multifunctional hypertext or One should also note that Floridi’s standard hypermedia definition of hypertext takes for granted the 2 a set of associations – links or hyperlinks equation of interface with electronic interface: the embedded in nodes by means of special for- proviso that the interface should be both inter- matted areas, known as source and destina- active and dynamic makes it almost necessarily tion anchors – connecting the nodes. These electronic, unless one holds a very loose notion are stable, active cross-references which allow of interactivity. Floridi’s insistence that hypertext the reader to move immediately to other parts is not uniquely a computer-based concept is of a hypertext correct (the electronic fallacy), but his very own 3 an interactive and dynamic interface. standard definition somehow implies that hyper- text has been recently redefined to be quasi- The emphasis on this third and crucial com- uniquely a computer-based implementation. ponent is worth noting since it is its presence This chapter starts from this premise, that we that dispels two of the six fallacies that Floridi should consider as an historical fact. Actually, associates with hypertext: the literary fallacy the first section will describe how the develop- according to which “hypertext began primarily ment of computer-implemented hypertext since as a narrative technique and hence it is essen- the early 1960s has stemmed from a dual origin tially a new form of literary style” and the expres- that somehow reconciles the literary and the com- sionist fallacy that holds that “hypertext has arisen puterized nature of the field. From then, we will as and should be considered primarily a writing- proceed with a general discussion on the ques- pushed phenomenon” (ibid.). Both fallacies, tion of the hypertextual interface, only to get however, are quite common and in fact define a back later to some elements of literary and, bet- whole subfield of hypertext theory. George P. ter yet, dramatic theory, applied to computer- Landow might be the most famous represent- based hypertext. ative of such a theoretical insight on hypertext, that focuses on the first two elements of Floridi’s standard model . . . and risks forgetting the third: Association vs. Connection: The “Hypertext . . . denotes text composed of blocks Dual Origins of Hypertext of text – what Barthes terms a lexia – and the electronic links that join them” (Landow 1992: 4). In fact, Landow even claimed that The standard definition of hypertext is indeed the result of an historical process, in which the The many parallels between computer hyper- organization of its three basic components – a text and critical theory have many points of discrete set of lexias, a set of links, and an inter- interest, the most important of which, perhaps, face – have been progressively stabilized through lies in the fact that critical theory promises to negotiations among actors in the field. theorize hypertext and hypertext promises to The term “hypertext” is usually credited to embody and thereby test aspects of theory, Ted Nelson, who coined it in 1962 with the particularly those concerning textuality, narrat- idea of hyperspace in his mind. According to ive, and the roles and functions of reader and him, his influence was mainly found in the writer. (ibid.: 3) vocabulary of mathematics, where the prefix hyper means “extended and generalized” (Nelson, per- Thus, for Landow, the proviso that the links sonal interview, 3/17/93). To Nelson, hypertext may be “electronic” does not change the basic was a necessary tool for his work as an author,

249 Thierry Bardini a tool that “allows you to see alternative ver- evolution by the means by which individuals sions on the same screen on parallel windows control the external manipulation of symbols” and mark side by side what the differences are” (ibid.: 24). (ibid.). From then on, hyperspace slowly became Engelbart thus postulated a dialectical rela- cyberspace. tionship between the two sublevels of natural At the same time that Ted Nelson coined the language, a relationship in which the symbolic term hypertext, Douglas Engelbart was beginning representation of concepts can affect the way to implement his framework for the Augmenta- these concepts structure the world. The com- tion of Human Intellect at Stanford Research puter could become an open medium that could Institute (SRI, in Menlo Park, CA). Altough his be used to “make sense of the world,” to map framework itself did not explicitly mention hyper- the structure of the world as information flows text, the core of Engelbart’s vision was based on in order to exploit the potential of natural lan- a very similar premise, “this extremely flexible guage to reconfigure our concepts and change way in which computers can represent modules our world. The key to this reconfiguration lay of symbols and can tie them together with any not in any single concept itself, but in their structuring relationship we can conceive of” being already configured – already given in (Engelbart, personal interview, 12/15/92). The nonlinear relationships that could be identified, introduction of an hypertext-like capability in mapped, and changed. When one stretches Engelbart’s framework responded, however, to the notion of technology to include the way a very different motivation than Nelson’s. It was humans use language – as Engelbart realized very based on the premise that computers should be early, according to his own account – it becomes able to perform as a powerful auxiliary to human clearer how it was the influence of Whorf that communication and collaboration if they were was central to the development of hypertext. to manipulate the symbolic language that human Because of how he conceived of the way that beings manipulate. natural language could function in the human– For Engelbart, what seemed promising about computer interface, Douglas Engelbart, along computers was that the processes that match with Ted Nelson, often is credited for pioneer- human and machine to the outside world and to ing work in the field of hypertext or hypermedia. each other could be found in natural language, Many, however, trace the genealogy of hyper- understood in both its physiological and social text not to Engelbart’s extension of the Sapir– dimensions; a language available to all. Of all Whorf Hypothesis, but to the work of Vannevar the tools humans use, language clearly seemed Bush. the metatool, the one that made all the others In a famous article entitled “As We May possible. What Engelbart meant by “language” Think,” Vannevar Bush proposed a new kind of was “the way in which the individual parcels electro-optical device, the memex, “a enlarged out the picture of his world into the concepts intimate supplement of an individual’s memory.” that his mind uses to model the world, and the The result of “utopian fiction and speculative symbols that he attaches to those concepts and engineering,” the memex was an imaginary uses in consciously manipulating the concepts” machine that existed entirely on paper and that (Engelbart 1962: 9). He thus conceived language never was constructed (Nyce & Kahn 1991: 45). as operating at two levels: concept structuring Bush conceived his memex on the basis of but also symbol structuring, in order to model analogies between brain and machine, between and at the same time to represent “a picture of electricity and information: the world.” As Engelbart himself pointed out, this understanding of what language is and does The human mind operates by association. With derives from the work of Benjamin Lee Whorf, one item in its grasp, it snaps instantly to the and it both mirrors and extends the famous Sapir– next that is suggested by the association of Whorf Hypothesis: “Both the language used by thoughts, in accordance with some intricate a culture, and the capability for effective intel- web of trails carried by the cells of the brain lectual activity, are directly affected during the . . . Man cannot hope fully to duplicate this

250 Hypertext

mental process artificially, but he certainly ought What actually differentiates hypertext systems to be able to learn from it. (Bush 1945: 101) from a information-retrieval systems is not the process of “association,” the term Bush proposed Nyce and Kahn (1991: 60) argue that since as analogous to the way the individual mind Bush’s work, “whether access and use of the works. Instead, what constitutes a hypertext sys- records should be based on abstract general prin- tem is the presence of links between lexias on ciples or on personal, i.e., individual associations a human–computer interface. And a process of has been the major issue separating information association analogous to the way the individual retrieval systems from hypertext.” Hypertext sys- mind works is not the only way of establishing tems in this formulation thus rely on the indi- such links. The most important ones already are vidual process of “association” as envisioned by established in natural language. Bush himself Bush, rather than on “abstract general prin- stated that “The process of tying two items ciples,” and “for Bush, and later for Nelson and together is the important thing” (Bush 1945: others engaged in hypertext research, memex 103). Thus, what actually defines hypertext is represented a very personal tool” (ibid.). The the existence of links organizing the information term “hypertext” wasn’t coined until the early in such paths, regardless of the process by which 1960s, however, and Bush himself never used these links were created. the term to describe his work. The reigning term To put it another way, “association” is only at the time for what Bush was proposing was one kind of “connection,” and is in fact the least indeed “information retrieval.” It seems difficult desirable kind, where communication is the goal, to dispute, therefore, that the memex was not precisely because it is the way an individual mind conceived as a medium, only as a personal “tool” works. This distinction too was pointed out by for information retrieval. Personal access to Benjamin Lee Whorf: information was emphasized over communica- tion. The later research of Ted Nelson on Connection is important from a linguistic hypertext is indeed very representative of that standpoint because it is bound up with the emphasis. communication of ideas. One of the necessary It is problematic, however, to grant Bush the criteria of a connection is that it be intelligible status of the “unique forefather” of computerized to others, and therefore the individuality of hypertext systems. For the development of the subject cannot enter to the extent that hypertext, the important distinction is not it does in free association, while a corres- between personal access to information and com- ponding greater part is played by the stock munication, but between different conceptions of conceptions common to people. (Whorf of what communication could mean, and there 1927: 36) were in fact two different approaches to com- munication at the origin of current hypertext Associations can be individual and open-ended. and hypermedia systems. The first is represented “A common stock of conceptions,” by contrast, by Ted Nelson and his Xanadu Project, which tends to be limited, and often is structured, even was aiming at facilitating individual literary cre- hierarchical, and susceptible to rearrangement. ativity. The second is represented by Douglas Ted Nelson, who like Engelbart was very famil- Engelbart and his NLS, as his oN-Line System iar with Whorf’s writings, stressed that the main was called, which was conceived as a way to difference between his views and Engelbart’s support group collaboration. The difference in view indeed concerned the role of structure and objectives signals the difference in means that hierarchy: “To me hierarchy is a special case. I characterized the two approaches. The first don’t say that hierarchies are always invalid, it’s revolved around the “association” of ideas on the just that because they’re so convenient they’ve model of how the individual mind is supposed been used too much. And they represent many to work. The second revolved around the inter- things very badly” (Nelson, interview, 1993). subjective “connection” of words in the systems For Engelbart, with his concern for communica- of natural languages. tion, the opposite was the case.

251 Thierry Bardini

The greatest philosophical achievement of the The Language Machine AI research program might very well be that and the Body it provides an invaluable source of insight into the effect of the formal, conventional nature of A common ground to all hypertext systems, language on efforts to think about the nature of regardless of their emphasis on hierarchy, is the the boundary between humans and machines. issue of electronic display/access to linked lexias, There is yet another metaphor to describe the tra- often dubbed nonlinear (e.g. Landow 1992: 101– ditional research program in AI: “The computer 19). But the emphasis on the notion of linearity, is the physical embodiment of the symbolic cal- again, often reflects back to the model of the culations envisaged by Hobbes and Leibniz. As text, and might miss the essential third compon- such, it is really not a thinking machine but a ent of the standard model of hypertext, inter- language machine” (Winograd 1991: 216). When face. In this sense the interface is indeed the the AI project is understood in this way, the crucial element of an hypertext system, because computer-as-mind metaphor points to the level it is at the interface that the linked lexias are of information-processing and symbolic mani- displayed/accessed. Now, what exactly does pulation, not to the more general concept of interface mean? “thinking.” The point of contact between human and The metaphor of the computer as a language computer, the boundary that separates and joins machine makes sense of the boundary metaphor them, usually has been located only via metaphor. by locating the boundary more accurately, within To say that “the mind is a meat machine,” or, the realm of “verbal agreement.” One can still ask more accurately, that “the mind is a computer,” whether this claim does not also beg the ques- is to use a metaphor: the statement relies on an tion of material differences in the manner of the analogy that “invites the listener to find within Turing test, however. There is more to natural the metaphor those aspects that apply, leaving language than the processing of symbols, more the rest as the false residual, necessary to the than conventional “rules and propositions” that essence of the metaphor” (Newell 1991: 160). lead to “verbal agreement.” If the notion of With the artificial intelligence mind-as-computer “symbol system” is indeed “inherently linguistic” metaphor (see Chapter 9, THE PHILOSOPHY OF (ibid.), everyday natural human language, on the AI AND ITS CRITIQUE), the greatest source of this other hand, cannot simply be reduced to the false residual lies in the human’s direct percep- conventional manipulation of symbols. tion of the computer. Physically, materially, minds Hubert L. Dreyfus has stated this objection and machines or computers are fundamentally regularly since 1972: there are things that com- different things, however much there may be re- puters (still) can’t do because they function in a semblances that permit metaphorical comparisons. binary logic at odds with human reasoning, and When one considers the mind-as-computer binary translations into machine logic of symbols metaphor as a means to make sense of the are far from enough to mimic human thinking. “boundary,” the obvious conclusion is that the AI has been at the same time overly ambitious compared materiality of human beings and in its claim to model human intelligence and computers is the false residual of the mind-as- insufficiently ambitious in trying to understand computer metaphor. One should conclude that the linguistic phenomenon and the path it opens there is no “natural” way to locate the bound- to the body. Engelbart, following Whorf, how- ary that distinguishes and joins them. There is ever, was able to see the ways in which the ana- no ontological connection, that is, between our logous character of natural language, thought, materiality – our bodies – and the material mani- and the human body meant that as a “language festation of the computer. But the ultimate goal machine,” the computer could serve as a genuine of the project to create artificial intelligence was to boundary-spanning object. In this perspective, achieve the material realization of the metaphor the materiality of humans and computers takes on of the computer as a “colleague,” and therefore as a different meaning than that of a “false residual” a mind, a machine that can pass the Turing Test. in a metaphor: both language and technology

252 Hypertext are inherently tied to the body on the human what the two parties think about what the other side of the interface and to circuits on the elec- is thinking about them must perforce be included tronic side. in the model” (ibid.: 14). Facing this “nightmare,” the seminar turned its attention to “more manageable concepts.” They settled on a simpler concept of the user The Computer as a Medium and interface: “the interface is that which joins hu- the Question of the Interface man and computer, conforming to the needs of each.” Laurel concluded that this viewpoint The introduction of visual metaphors has certainly “avoids the central issue of what this all means been the most important aspect of the efforts in terms of reality and representation,” and that to design adequate user interfaces, understood “when we have such trouble defining a concept, here as adequate patterns of interaction between it usually means that we are barking up the wrong the user and the computer. The opening of the tree” (ibid.) But it could also merely require a graphic dimension of the computer as a com- translation of a more complex process of “suspen- munication medium is often thought to be one of sion,” or more precisely, of Hegelian Aufhebung the major contributions of Alan Kay and his team that “combines the idea of suspension, with its at Xerox Palo Alto Research Center (PARC) in connotation of temporary cessation; transcend- the 1970s. A major contribution of this outstand- ence, which suggests a going beyond; and a kind ing set of computer scientists is the “desktop of preservation” (Ashmore 1989: 111). There is metaphor,” the leading metaphor for the per- even a sign of such a process when Brenda Laurel sonal computing interface, as a visual space popu- says that “you can demonstrate Zeno’s paradox lated with iconic representations and organized on the user’s side of the barrier until you’re blue in “windows.” Let us see now how the notion in the face, but it’s only when you traverse it of the hypertextual interface unfolds in more spe- that things get real.” cific models of the features of this symbolic space. Who is that “you”? It applies to the user as In her remarkable Computer as Theater, Brenda well as to the designer of the interface: the prob- Laurel (1991: 12–14), narrates how she and the lem is to enable both of them to act within a participants of a seminar at the Atari Company representation (Laurel 1991: 21). The problem (where she was then working) attempted to in defining the user interface arose in Laurel’s define the user interface. They rapidly dismissed narrative with the cognitive-science assumption the simplest model of the interface represented that representations are part of an interface: the as the space between the person (user) and the user’s representation of the computer, on one computer, that “encompasses what appears on the hand, and the “computer’s understanding of the screen, hardware input/output devices, and their person” on the other. What do you think the drivers.” They dismissed “this over-simplistic computer understands of the person? Not much model” because of its lack of consideration for in itself, unless “you” anthropomorphize the the “person’s ‘mental model’ of the computer computer. You understand a lot more, still, if and the computer’s ‘understanding’ of the per- you understand that the computer can represent son [which] are just as much a part of the inter- the designer. In other words, the computer might face as its physical and sensory manifestations” be able to learn about the user from the repres- (ibid.: 12–13). entation of the user that the designer of the Once this “precognitive-science” model of the interface embodies in his/her design. interface is dismissed, the “conceptual interface” Following Laurel, we thus realize that the is part of the interface. They called the main interface is the representational space where user problem that they encountered then with this and designer meet, act, and communicate. The updated model of the interface the “horrible re- anthropomorphization of the computer translates cursion”: “If you are going to admit that what the user’s desire for the computer’s responsive- the two parties ‘think’ about each other is part ness and capacity to perform action, two major of what is going on, you will have to agree that anthropomorphic qualities of the interface, which

253 Thierry Bardini

“comprise the metaphor of agency” (Laurel In this sense, the degree of interactivity of the 1990: 358). In her defense of anthropomorph- interface can be seen as the relative opportunity ism, she stressed later (p. 143) that the quest for both user and designer to take part in the for these qualities does not lead necessarily to two dimensions of the representation process. the personification of the computer, but to its The joint construction of the plot is the consen- invisibility. This invisibility must be the result of sus reached on a set of agents whose competences a negotiation between user and designer over are negotiated between user and designer. Set- the competence of the interface agents. The first ting the negotiation at the level of the entire set step in this direction is to realize that human of agents allows one to focus on the represent- and nonhuman agents in the interface, like char- ativeness of the interface as a whole, that is, as a acters in a play, cannot be separated from the socially constructed narrative between user and plot itself. designer. Characters such as desktop icons or interface agents in general similarly define, and are defined by, the theatrical frame of the interface as a whole. The Designer as the Third Man The efficacy of the computer interface depends on developing convincing “characters” in the “nar- rative” of the user interface. If their negotiation There were three main shifts in our (occidental) is successful, user and designer reach a consensus conception of the dialogue. The origin-point is on the competence of the agent to perform a of course the Platonic dialogue. The first shift task, and the medium disappears in the process: occurred at the dawn of experimental science, user and designer agree on the “truth” of the and has been beautifully described and explained representation embodied in the agent, and, con- by Shapin and Shaffer (1985) as the “virtual sequently, his/hers/its action appears as “real.” witnessing.” Dialogue then occurs between the The object of the negotiation is the plot, and in scientist and the world, and is inscribed for the the present case, the alternative representations readers in the text: it is therefore, (1) a dialogue of the user and the designer of the task to be between the scientist (author) and the world performed with the help of the computer. (nature) by means of experiments, virtually wit- Being the result of a consensus between nessed by the readers (other scientists, peers). designer and user, the interface agent combines The second shift in our conception of dia- the two orthogonal dimensions of a representa- logue occurred with Cybernetics, the science of tion, delegation and inscription. Delegation is communication and control, and more precisely the process by which the agent is granted the with its theory of information. An alternative right to represent action in the interface, and dialogue emerged, a dialogue where, according inscription is the process that enables the agent to Serres (1982: 66), authors and readers could to perform the action being represented. These trade places, and unite against the noise, “the set two processes conjointly define the competence of these phenomena of interference that become of the agent as the embodiment of the dialogical obstacles to communication.” In this third con- consensus reached by users and designers. Michel ception of the dialogue, the first candidate for Serres enables us to shade a new light on this exclusion was the World: “in order for dialogue process when he states that “To hold a dialogue to be possible, one must close one’s eyes and is to suppose a third man and to seek to exclude cover one’s ears to the song and the beauty of the him” (1982: 66–7, emphasis in original). Follow- sirens” (ibid.: 70). The dialogue hence becomes ing Serres it is not the medium that disappears (2) a dialogue between the author and his or her in the successful negotiation of the competence readers, excluding the world, that is, the object of of the interface agents, but the “third man,” the his or her representative practice. source of noise. In the case of a successful nego- Here comes up against reflexive problem: when tiation between users and designers, the interface (1) is deconstructed by means of (2) one faces agents are under control, they will not create noise the specter of the infinite regress (Ashmore but docilely perform what they are expected to. 1989), since in principle there could always be a

254 Hypertext possible of (2) itself by the same or her dialogue with the world?” Observe that process that deconstructed (1) in the first place. it is through this dialogue that the user engages The trick is here, of course, all included in “∞,” in the “always unfinished business of human the sign standing for the “infinite” and/or its becoming,” and therefore that his or her “ability alternative expression “in principle . . . always . . . to act as a person” should be considered both as possible.” In practice now, this iteration of the a process and the result of this process. Under- same basic operation takes time. Or, as Rotman stood as such, wrighting the interface refers to a (1993: 147, emphasis in original) puts it, “the process of personalization. The designer can thus response to ‘infinity’ developed here can be sum- be conceived as the agent enabling this person- marized as hinging on the issue of embodiment.” alization: his or her action personalizes the inter- If there is a dialogue then, it takes place between face, that is, it creates the conditions for the user the readers and the World. The limit of the to become a person, to act as such. As we will see, embodiment of the author is the first interface this conception could help us understand a crucial in the set-up that unites author and reader in problem in the evolution of the user interface the same fiction, because and against the virtual- from today’s direct manipulation interfaces to ity of the physics (not the metaphysics) of their the possible intelligent user interfaces of tomor- ever-allusive co-presence. The third shift in our row, and, of more concern to our topic, to the conception of dialogue, maybe yet to come, hypertextual interface. would lead us then to (3) a dialogue between the readers and the World, partly witnessed and then wholly imagined by the author, in the practice The Distribution of Intelligence that Malcolm Ashmore called “wrighting.” For at the Interface and the Future Ashmore, “celebratory practical reflexive inquiry of the Person is wrighting beyond the tu quoque.” This celebra- tion, indeed, “is not a matter of authorial pre- sentation . . . not a matter of being correct . . . not An important debate in today’s negotiations a matter of meta-analysis . . . not a matter of solv- about the future of the computer still concerns ing a problem” (ibid.: 110). the question of the distribution of intelligence This scheme can be applied to the idea of at the interface, what most analyses have trans- a dialogue between designer (author) and users lated into the question of the personification of (readers) mediated by the hyper(textual) interface. the computer. But today’s personification of the Each of the three conceptions of the dialogue computer still requires a “momentary suspension previously introduced corresponds to a specific of disbelief” (Laurel 1991: 113). Why should the way to characterize the interaction taking place users temporarily suspend their disbelief? In the at the interface. The first conception considers the case of theater or fiction, Coleridge argued that user as an agent engaged in a human–computer it is for the sake of experiencing other emotional interaction, the second conception considers the responses, in Aristotelian terms catharsis, the user as a subject in a computer mediated com- pleasurable release of emotion. After Coleridge, municative act against the noise of the world, Laurel (ibid.: 120–2) argues that the same process and the third conception considers the user as a should occur in interacting with a computer, person engaged in his or her own dialogue with where catharsis stems from achieving a given task. the world. In this third conception of the inter- When a play fails, when the spectators do not face as the space of the dialogue between the feel the release of emotion associated with the user and the world, the designer becomes a wit- characters’ experiences, they usually blame the ness who imagines a space for this dialogue, and director or the actors. When the interface fails, enables both the world and the user to act within most of the users of current personal computers this space: the designer wrights the interface. blame themselves. This might be the clearest mar- In this respect, the fundamental question of keting success of the personal computing indus- interface design becomes: “how to wright in try, but it is still an obstacle to the diffusion of order to enable the user to act as a person in his its products. Proponents of the personification

255 Thierry Bardini of the computer, however, hold that it should not our definition of what a person is?’” According be the case. In the best tradition of the Artificial to , the process of per- Intelligence program, they ask the users to grant sonalization is the temporal and continuous pro- a momentary suspension of disbelief to the inter- cess which constitutes the unity of the subject. face agent in the narrative space of the interface, The continuous character of the process must and, at the same time, to consider the agent as a be emphasized because for Whitehead, “a nexus person enabling them to carry out a task in the enjoys ‘personal order’ when (a) it is a ‘society,’ “real, situated space” of their work practices. Both and (b) when the genetic relatedness of its mem- spaces merge in the “personal space” of the user, bers orders these members ‘serially’” (1929: 34). and this merger creates a fundamental problem This “genetic relatedness” is the engine of the for the request to suspend disbelief: it is the process. source of breakdowns. Whitehead’s conceptions had a profound This fundamental problem sometimes leads the impact on ’s conceptualization proponents of the personification of the computer of an “ecology of mind,” his framework for the to make ambiguous claims. On the one hand, the exploration of the “natural history of the rela- personification of the computer, via the interface tionships between explicit, implicit and embodied agent, is supposed to lead to its disappearance: ideas in the world of living things.” Whitehead’s the medium is supposed to vanish, be “ready at influence is especially apparent in Bateson’s notion hand,” and be an extension of the body. On the of “socialization,” a keystone in his early work other hand, the personification of the agent in anthropology. Bateson holds that “socialization makes it a part of the “world of people and (by definition) requires interaction, usually of information” and creates identification for the two or more organisms” and that its goal is “sup- user: in this second sense, the medium becomes posedly” this “set of appearances” that we call a “present at hand.” It draws attention to itself. “person,” since a “person” is necessarily a soci- Hence, the interface agent is necessarily a source alized individual (Bateson 1975: 75). Bateson of breakdowns, of never-ending shifts of the made this point clearer in a footnote where he attention of the user. Thus, the personification notes that “the ‘person’ after all, is the mask. It of the interface does not solve the problem of is what is perceivable of a human organism. It is breakdowns; rather, it welds an essential source a unilateral view of the interface between one of breakdowns into the interface. The necessity organism and another” (ibid.). of the user’s “suspension of disbelief” becomes In this respect, Bateson’s notion of the reiterated with each breakdown. Because of its “person” actually dates back to the original Latin hybrid nature as a symbolic and material space, notion of persona, in the sense of the mask of the user interface is a space of breakdowns, and tragedy, whose meaning was reconstructed by the personification of the computer cannot pre- the Latin etymologists as pers/sonare, the mask vent such breakdowns by design: it still requires through (per) which sounds (sonare) the actor’s learning and adaptation on the part of the user. voice (Mauss, [1938]: 350); but it is also the Let us see now what an alternative notion of mask of wax cast on the face of the dead ancestor “personalization” could bring to this question. (imago), the prosopon of the Greek tradition, So far, we have mostly limited ourselves to a which also meant the mask of the ancestor or his narrative conception of the interface. We have statue kept in the wings of the family house seen the drawbacks of current alternatives to (ibid.: 352). Marcel Mauss thus insisted on the reshape the interface, and insisted on the limits original double meaning of the persona, that both of a metaphorical conception. It is now time to hides and reveals the true nature of the individual, complete our review of the question of the future in respect to his origins, his genetic relatedness of the interface by insisting on a phenomenolo- to his ancestors, and his singularity as an actual gical perspective on the “person.” entity. Like Jaron Lanier (1995), “I have long Hence, personalization and socialization are believed that the most important question about the two sides of the same process of human information technology is ‘How does it effect becoming. Personalization puts the emphasis on

256 Hypertext the appearance of the unity of the subject, while The answers Whorf gave to these questions are socialization stresses the relatedness of members considerably more nuanced than the bold formu- in their expression of a common form. The are lations, such as “language is culture,” that some- both basically grounded in two dynamics: (1) a times are attributed to him: “I should be the last continuous dynamic of genetic relatedness stem- to pretend that there is anything so definite as a ming from an origin, and (2) a dynamic of feeling ‘correlation’ between culture and language . . . which constitutes a common sense for this origin. We have plenty of evidence that this is not the Bertrand Russell, Whitehead’s former pupil case” (ibid.). Whorf’s answer to the question of and collaborator, held that there were two ways whether our concepts of time, space, and matter to define a person: (1) in derivation of memory, are universal and unconditioned or “in part con- since “each person’s experience is private to him- ditioned by the structure of particular languages” self, and when one experience consists of recol- was that both propositions are true: space may lecting another, the two are said to belong to indeed be perceived in a similar fashion by every the same ‘person’”; and (2) in derivation of the individual, and therefore be common to all body, since “we can then define a ‘person’ as human beings as a result of the basic conditions the series of mental occurrences connected with of human physiology; while at the same time the a given body” (Russell 1935: 140). These two concept of space is also a linguistic construction derivations are the two projections which anchor and therefore varies with the different human Rotman’s tripartite scheme: the Agent is active groups singularized by their language. is the field of memory, the Subject is active in a Thus, for Whorf, the connections between lan- field of mental occurrences connected to a given guage, cultural norms, and behavior are to be body, and the Person transcends and unites both found at the level of observation and representa- in the field of “consciousness.” “Consciousness” tion, not the level of perception. However uncon- is an afterthought: it is the result of an articula- scious the part that language plays in this process tion of the two previous elementary processes, may be, Whorf postulated that it always plays experience and order, it is the name human a central role in constituting the “real world” beings give to the realization of their dual nature through the process of sharing meaning: “Con- as persons. Now, how on earth could hypertext cepts of time and matter are not given in sub- affect this process of human becoming? stantially the same form by experience to all men but depend upon the nature of language or lan- guages through the use of which they have been developed” (ibid.: 159). Hypertext, Cybernetics, and The relativity of time and space, of course, Space-time was not a notion limited to Whorf’s work on comparative linguistics. It was fundamental to In an essay written in 1939 and entitled “The cybernetics too (see Chapter 14, CYBERNETICS). Relation of Habitual Thought and Behavior to Norbert Wiener likewise insisted on alternative Language,” Benjamin Lee Whorf introduces his conceptions of time, although not exactly in the inquiry as follows: perspective presented here by Whorf. He wrote (1948: 37) that “it is thus not too much to say That portion of the whole investigation here that not only the Newtonian astronomy but even to be reported may be summed up in two the Newtonian physics has become a picture of questions: (1) Are our own concepts of time, the average result of a statistical situation, and space, and matter given in substantially the hence an account of an evolutionary process.” same form by experience to all men, or are This point is crucial to cybernetics because it they in part conditioned by the structure of eventually justified the functional analogy be- particular languages? (2) Are there traceable tween living organisms and machines, and hence affinities between (a) cultural and behavioral between brains and computers. Wiener reinforced norms and (b) large-scale linguistic patterns? this idea when he stated that “the great contribu- (Whorf 1941) tion of Heisenberg to physics was the replacement

257 Thierry Bardini of this still quasi-Newtonian world of Gibbs by science side of cybernetics appears in the synthesis one in which time series can in no way be reduced provided by Gregory Bateson’s writings. The core to an assembly of determinate threads of devel- of this synthesis revisited Korzybski’s notion of opment in time” (ibid.: 92). the connection between “map” and “territory,” The relativist revolution in how time and and applied it at a basic level: space are conceived helped make possible insights such as Whorf’s, most notably via the “gen- The bridge between map and territory is dif- eral semantics” program epitomized by Alfred ference. It is only news of difference that can Korzybski’s writings. Korzybski established a get from the territory to the map, and this fact connection between the changing worldview in is the basic epistemological statement about physics and an overall framework for social sci- the relationship between all reality out there ences that granted a new epistemological status and all perception in here: that the bridge must to meaning and language. In fact, Whorf’s always be in the form of difference. (Bateson “Relation of Habitual Thought and Behavior to 1979: 240) Language” actually appeared in ETC: A Review of General Semantics, the General Semantics Bateson also redefined the basic cybernetic International Society’s journal. For Korzybski notion of “information” as “any difference that (1926), “mankind” is not bound by time, but makes a difference” (1979: 228). Defined in these instead a “time-binding” class of life that has sur- terms, for Whorf, what matters is not simply vived in evolution by its ability to learn from past that language determines our concepts and that experiences and to pass this knowledge on from the relational network of our concepts (links, generation to generation through language. For connections) determines our view of the world. Korzybski and his followers, it was necessary to What matters is the difference between percep- change the linguistic conception of the relation- tion on the physiological level and mental con- ship between the “word” and the “thing-in-the- cepts on the level of language, a difference that world.” Korzybski (1933) best expressed this new itself has different implications, depending on perspective in his famous analogy between maps which side of it is emphasized. For Korzybski, and language: “a map is not the territory” to what matters is the difference between the map which it corresponds, “words are not the things and the territory – not the map by itself, or the they represent.” This central premise led him to territory. Difference is thus by definition the site question the fundamental basis of the Newtonian of an interface. In the formulations of Whorf, worldview and the Aristotelian system of logic, Korzybski, and Bateson, the interface is this dif- and to propose, instead, a “relativist” reformula- ference that makes a difference, at the boundary tion of the law of identity: “identity is a relative between the physical and the conceptual realms. matter: relative to the history of the things con- sidered, relative to the environment the thing is in, relative to our own practical purposes, relat- Conclusion ive to the frame of reference from which it is viewed” (Reiser 1989: 85–6). As in the shifting worldview in physics, the The computer will be “personal” when it allows extension of the Aristotelian logic and its applica- its users to act as persons, to experience the tion to language in the General Semantics pro- world as both embodied subject and ideal agent, gram relied heavily on a quasi-statistical approach, united in actual socialized and enduring persons. or more accurately, on a theory of classes, since This could happen if and only if the computer words were not considered any more as “ident- could affect, in terms of signification, the way ical” with what they represented, but rather as a humans express their genetic relatedness through class of things that they could represent. The their “mutual prehensions” of each other. These social scientists of the cybernetics group were of prehensions are the inscribing and incorporating course aware of this line of thought. The con- practices which characterize the human experi- nection between Korsybski’s ideas and the social ence in the fields of virtual and actual entities.

258 Hypertext

However, the computer “memory” is organized Acknowledgments quite differently from the human memory: computer “time” is discreet, and corresponds to This chapter is a revised version of a paper an event-oriented worldview, anchored in the entitled “Bridging the gulfs: from hypertext to renewed notion of probabilistic time brought in cyberspace” that appeared in the Journal of by the relativist tradition. This worldview seems Computer-mediated Communication in september at first to contradict the requisite of continuity. 1997 (). Elements of the final seriality: there are the discreet traces of previous sections also appeared in Boot-strapping: Douglas interactions, and because of that they refer only Engelbart, Coevolution and the Origins of Per- to virtual entities. The fact that a human being sonal Computing (2000). I thank Luciano Floridi could imagine a continuous temporality in his for his interest in these previous works and for hermeneutical relation to the computer does not his help in editing the present chapter. change the fact that the computer does not share it as an embodiment relation: this continuity is only an interpretation of the human user and of References his interaction with the computer. Could this interpretation be embodied in Ashmore, M. 1989. The Reflexive Thesis: Wrighting the computer in the same the way that it is in Sociology of Scientific Knowledge. Chicago: Univer- human beings? A starting-point would be the sity of Chicago Press. [Advanced undergraduates, ability of the computer to allow humans to make graduates.] full use of their body, to act inside the space Bateson, G. 1979. Mind and Nature: A Necessary that the interface presents and represents, to feel Unity. New York: Dutton. [Advanced under- and experience tomorrow’s worlds as bodies and graduates, graduates.] minds, through the symbolic and material space ——. 1991 [1975]. “Some components of that an interface affords for exploration, play, socialization for trance.” In R. E. Donaldson, and work. The social and cultural construction ed., Sacred Unity: Further Steps to an Ecology of the personal computer user, so far, has led to of Mind. New York: HarperCollins, pp. 73–88. an overwhelming hegemony of the visual sense [Advanced undergraduates, graduates.] and symbolic coding, following in that a larger Bush, V. 1991 [1945]. “As we may think.” In trend in modernity. Whether the computer could J. M. Nyce and P. Kahn, eds., From Memex to be interpreted as a person or as slave, whether it Hypertext: Vannevar Bush and the Mind’s Machine. could be metaphorically considered as a con- San Diego, CA: Academic Press, pp. 83–107. scious entity or not, is finally of little interest in (Originally published in Atlantic Monthly 176[1]: regard to what the computer does. Nothing pre- 641–9.) [Undergraduates, graduates.] vents a priori the computer from participating Engelbart, D. C. 1962. “Augmenting human intellect: a conceptual framework.” Report to the in the ongoing evolution of human beings, in Director of Information Sciences, Air Force Office the way that human tools have done for a long of Scientific Research, Menlo Park: Stanford time. Conscious efforts to design personal com- Research Institute, Oct. [Advanced undergraduates, puters should take this into account and strive graduates.] for a harmonious human experience, in its full- Floridi, L. 1999. Philosophy and Computing: An est expression. So far, the personal interface has Introduction. London: Routledge. [Undergradu- remained a marking interface. Typing and click- ates, graduates.] ing is marking, indexing, punching a hole. But Gygi, K. 1990. “Recognizing the symptoms of this despairing notion might hide the great poten- hypertext . . . and what to do about it.” In tial of the hypertextual interface: a system that B. Laurel, ed., The Art of Human Computer could allow human action in cybernetic space- Interface Design. Reading, MA: Addison-Wesley, time, through the linked events of human becom- pp. 279–87. [Undergraduates, graduates.] ing, in language and gestures, in thought and Korzybski, A. 1926. Time-binding: The General reality. Theory. Washington, DC: J. C. Wood. [Graduates.]

259 Thierry Bardini

——. 1933. Science and Sanity: An Introduction to Semantics, pp. 89–97. [Advanced undergraduates, Non-Aristotelian Systems and General Semantics. graduates.] Lancaster, PA: Science Press. [Advanced under- Rotman, B. 1993. Ad Infinitum: The Ghost in the graduates, graduates.] Turing Machine: Taking God out of Mathematics Landow, G. P. 1992. Hypertext: The Convergence of and Putting the Body Back In. Stanford: Stanford Contemporary Critical Theory and Technology. 2nd University Press. [Graduates.] ed. 1997. Baltimore: Johns Hopkins University Russell, B. 1997 [1935]. Religion and Science. Press. [Undergraduates, graduates.] Oxford: Oxford University Press. [Undergradu- Lanier, J. 1995. “Agents of alienation.” Journal of ates, graduates.] Consciousness Studies 2: 76–81. [Undergraduates, Serres, M. 1982. “Platonic dialogue.” In J. V. Harari graduates.] and D. F. Bell, eds., Hermes: Literature, Science, Laurel, B. 1990. “Interface agents: metaphors with Philosophy. Baltimore, MD: Johns Hopkins Uni- character.” In B. Laurel, ed., The Art of Human versity Press, pp. 65–70. [Advanced undergradu- Computer Interface Design. Reading, MA: ates, graduates.] Addison-Wesley, pp. 355–65. [Undergraduates, Shapin, S. and Schaffer, S. 1985. Leviathan and the graduates.] Air-pump: Hobbes, Boyle, and the Experimental ——. 1991. Computers as Theater. Reading, MA: Life. Princeton, NJ: Princeton University Press. Addison-Wesley. [Undergraduates, graduates.] [Advanced undergraduates, graduates.] Mauss, M. 1950. [1938] “Une catégorie de l’esprit Whitehead, A. N. 1978 [1929]. Process and Real- humain: la notion de personne celle de ‘moi.’” ity. New York: Free Press. [Graduates.] In Sociologie et anthropologie. Paris: Presses Univer- Whorf, B. L. 1956 [1927] “On the connection of sitaires de France, pp. 331–62. [Advanced under- ideas.” In J. B. Carroll, ed., Language, Thought, graduates, graduates.] and Reality: Selected Writings of Benjamin Lee Newell, A. 1991. “Metaphors for mind, theories of Whorf. Cambridge, MA: MIT Press, pp. 35–9. mind: should the humanities mind?” In J. J. [Undergraduates, graduates.] Sheehan and M. Sosna, eds., The Boundaries of ——. 1956 [1941] “The relation of habitual Humanity: Humans, Animals, Machines. Berkeley, thought and behavior to language.” In J. B. CA: University of California Press, pp. 158–97. Carroll, ed., Language, Thought, and Reality: [Advanced undergraduates, graduates.] Selected Writings of Benjamin Lee Whorf. Cam- Nyce, J. M. and Kahn, P. 1991. “A machine for the bridge, MA: MIT Press, pp. 134–59. [Advanced mind: Vannevar Bush’s memex.” In J. M. Nyce undergraduates, graduates.] and P. Kahn, eds., From Memex to Hypertext: Wiener, N. 1961 [1948]. Cybernetics or Control Vannevar Bush and the Mind’s Machine. San and Communication in the Animal and the Diego, CA: Academic Press, pp. 39–66. [Under- Machine, 2nd ed. Cambridge, MA: MIT Press. graduates, graduates.] [Advanced undergraduates, graduates.] Reiser, O. L. 1989. “Historical-cultural signific- Winograd, T. 1991. “Thinking machines: can there ance of non-Aristotelian movement and the be? Are we?” In J. J. Sheehan and M. Sosna, eds., methodological contribution of Korzybski.” In The Boundaries of Humanity: Humans, Animals, S. I. Berman, ed., Logic and General Semantics: Machines. Berkeley, CA: University of California Writings of Oliver L. Reiser and Others. San Press, pp. 198–223. [Advanced undergraduates, Francisco: International Society for General graduates.]

260 Part VI Logic and Probability Logic

Chapter 20 Logic G. Aldo Antonelli

concerned with the identification and justification Origins of the Modern of valid inference forms. Conception of Logic The articulation of logic as a prescriptive dis- cipline is, ideally, a two-fold task. The first task Logic is an ancient discipline that, ever since its requires the identification of a class of valid argu- inception some 2500 years ago, has been con- ments. The class thus identified must have cer- cerned with the analysis of patterns of valid tain features: not just any class of arguments reasoning. Aristotle first developed the theory of will do. For instance, it is reasonable to require the syllogism (a valid argument form involving that the class of valid argument be closed under predicates and quantifiers), and later the Stoics the relation “having the same logical form as,” singled out patterns of propositional argumenta- in that if an argument is classified as valid, then tion (involving sentential connectives). The study so is any other argument of the same logical of logic flourished in ancient times and during the form. It is clear, then, that such an identification middle ages, when logic was regarded, together presupposes, and rests on, a notion of logical with grammar and rhetoric (the other two dis- form. ciplines of the trivium), as the foundation of The question of what constitutes a good humanistic education. theory of logical form exceeds the boundary of Throughout its history, logic has always had a the present contribution, and hence we will not prescriptive as well as a descriptive component. be concerned with it. We shall limit ourselves to As a descriptive discipline, logic aims to capture the observation that one can achieve the desired the arguments accepted as valid in everyday lin- closure conditions by requiring that the class guistic practice. But this aspect, although present of valid arguments be generated in some uni- throughout the history of the field, has since the form way from some restricted set of principles. inception of the modern conception of logic, For instance, Aristotle’s theory of the syllogism some 100 or 150 years ago, taken up a position accomplishes this in a characteristically elegant more in the background, and in fact some have fashion: subject–predicate propositions are clas- argued that it is no longer part of logic pro- sified on the basis of their forms into a small per, but belongs to other disciplines (linguistics, number of classes, and syllogisms are then gen- psychology, or what have you). Nowadays logic erated by allowing the two premises and the is, first and foremost, a prescriptive discipline, conclusion to take all possible forms.

263 G. Aldo Antonelli

The second task, however, is much harder. nothing but a stunning success. Its mathematical Once a class of arguments is identified, one natur- properties provide a crucial benchmark for the ally wants to know what it is that makes these assessment of alternative logical frameworks. We arguments valid. In other words, in order to are not going, in this chapter, to provide an accomplish this second task, one needs a general introduction to the nuts and bolts of FOL: the theory of logical consequence, and such a theory interested reader can consult any one of the many was not only unavailable to the ancients, it would excellent introductory texts that are available, not be available until the appearance of mod- such as, for example, Enderton 1972. ern symbolic logic in the late 1800s, when an FOL provides an implementation of the effort was undertaken to formalize and represent so-called “no-counterexample” consequence rela- mathematical reasoning, and it would not be tion: a sentence φ is a consequence of a set Γ of completely developed until the middle of the sentences if and only if one cannot reintepret twentieth century. the language in which Γ and φ are formulated in It is only with the development of the first gen- such a way as to make all sentences in Γ true and φ ψ ψ eral accounts of the notion of logical consequence false. An inference from premises 1,..., k that modern symbolic logic was born. Modern to a conclusion φ is valid if φ is a consequence ψ ψ symbolic logic is only a little over 100 years old, of { 1,..., k}, i.e., if the inference has no dating back to the end of the nineteenth century, counterexample. and in particular to Gottlob Frege’s Begriffsschrift For this to be a rigorous account of logical (1879) and Richard Dedekind’s Was sind und consequence, the underlying notion of interpreta- was sollen die Zahlen? (1888). With these two tion needs to be made precise. This was accom- works we have the beginning of a rigorous plished by Alfred Tarski in 1935, who defined the account of logical consequence, an account that notion of truth on an interpretation (see Tarski will be perfected by Alfred Tarski in the early 1956 for a collection of his technical papers). In 1930s. so doing, Tarski overcame both a technical and This chapter focuses on the development of a philosophical problem. The technical problem modern symbolic logic from the point of view has to do with the fact that in FOL quantified of the notion of logical consequence. After pre- sentences are obtained from components that are senting a streamlined account of what is regarded not, in turn, sentences, so that a direct recursive as the crowning achievement of modern symbolic definition of truth for sentences breaks down logic, i.e., the systematization of first-order at the quantifier case. In order to overcome this logic, we consider consequence relations from an problem Tarski introduced the auxiliary notion abstract point of view. In the next section we look of satisfaction. The philosophical obstacle had at consequence relations that are of particular to do with the fact that the notion of truth conceptual interest, in that they aim to capture was considered suspiciously metaphysical among patterns of defeasible reasoning in which con- logicians trained in the environment of the clusions are drawn tentatively, subject to being . This was a factor, for instance, in retracted in the light of additional evidence. Fin- Gödel’s reluctance to formulate his famous ally, we look at nonmonotonic logics devised to undecidability results in terms of truth. capture such defeasible inference. Tarski’s analysis yielded a mathematically precise definition for the “no-counterexample” consequence relation 3 of FOL: we say that φ is a consequence of a set Γ of sentences, written First-order Logic Γ 3 φ, if and only if φ is true on every inter- pretation on which every sentence in Γ is true. First-order logic (henceforth: FOL) was origin- At first glance, there would appear to be some- ally developed (through the work of Frege, thing intrinsically infinitary about 3. Regardless Dedekind, Russell & Whitehead, Hilbert, Gödel, of whether Γ is finite or infinite, to check whether and Tarski) for the representation of mathemat- Γ 3 φ one has to “survey” infinitely many pos- ical reasoning. As such, FOL turned out to be sible interpretations, and check whether any one

264 Logic of them is a counterexample to the entailment of one of the most successful branches of mod- claim, i.e., whether any one of them is such that ern symbolic logic: model theory. Compactness all sentences in Γ are true on it while φ is false. and the Löwenheim–Skolem theorem character- However, surprisingly, in FOL the infinitary ize FOL; as shown by Per Lindström in 1969, nature of 3 is only apparent. As Gödel showed any logical system (meeting certain “regularity” in 1929, the relation 3, although defined by conditions) for which both compactness and universally quantifying over all possible inter- Löwenheim–Skolem hold, is no more expressive pretations, can be analyzed in terms of the exist- than FOL (see Ebbinghaus et al. 1994: ch. 13 ence of finite objects of a certain kind, viz., formal for an accessible treatment). proofs. A formal proof is a finite sequence of Another important consequence of Gödel’s sentences, each one of which is either an axiom, completeness theorem has to do with the ques- or an assumption, or is obtained from previous tion of whether and to what extent one can ones by means of one of a finite number of devise an effective procedure to determine if a inference rules, such as modus ponens. Many dif- sentence φ is valid, or, more generally, whether ferent axiomatizations of FOL exist, and a par- Γ 3 φ for given Γ and φ. First, some terminology. ticularly simple and elegant one can be found in We say that a set Γ of sentences is decidable if Enderton 1972. If a sentence φ occurs as the last there is an effective procedure, i.e., a mechanic- line of a proof, then we say that the proof is a ally executable set of instructions, that deter- proof of φ; and we say that φ is provable from Γ, mines, for each sentence φ, whether φ belongs written Γ 2 φ, if and only if there is a proof of φ to Γ or not. Notice that such a procedure gives all of whose assumptions are drawn from Γ. both a positive and a negative test for member- Gödel’s famous completeness theorem of 1929 ship of a sentence φ in Γ. A set of sentences is states that the two relations 3 and 2 are exten- semidecidable if there is an effective procedure sionally equivalent: for any φ and Γ, Γ 3 φ if and that determines if a sentence φ is a member of Γ, only if Γ 2 φ. This is a remarkable feature of but might not provide an answer if φ is not a FOL, which has a number of consequences. One member of Γ. In other words, Γ is semidecidable of the deepest consequences follows from the if there is a positive, but not necessarily a negat- fact that proofs are finite objects, and hence that ive test for membership in Γ. Equivalently, Γ is Γ 2 φ Γ if and only if there is a finite subset 0 semidecidable if it can be given an effective list- Γ Γ 2 φ of such that 0 . This, together with the ing, i.e., if it can be mechanically generated. completeness theorem, gives us the compactness These notions can be generalized to relations theorem: Γ 3 φ if and only if there is a finite among sentences of any number of arguments. Γ Γ Γ 3 φ subset 0 of such that 0 . There are many For instance, it is an important feature of the interesting equivalent formulations of the theo- axiomatizations of FOL, such as that of Enderton rem, but the following is perhaps the most often 1972, that both the set of axioms and the rela- φ φ ψ ψ cited. Say that a set of sentences is consistent if tion that holds among 1,..., k and when φ φ they can all be made simultaneously true on some can be inferred from 1,..., k by one of the interpretation; then the compactness theorem says rules, are decidable. As a result, the relation that Γ φ φ φ φ that a set is consistent if and only if each of its holds among 1,..., k and whenever 1,..., φ φ finite subsets is by itself consistent. k is a proof of is also decidable. (See Chap- Another important consequence of Gödel’s ter 2, COMPUTATION, for further details on these completeness theorem is the following form of notions.) the Löwenheim–Skolem theorem: if all the sen- The import of Gödel’s completeness theorem tences in Γ can be made simultaneously true in is that if the set Γ is decidable (or even only some interpretation, then they can also be made semidecidable), then the set of all sentences φ simultaneously true in some (other) interpreta- such that Γ 3 φ is semidecidable. Indeed, one tion whose universe is no larger than the set ގ can obtain an effective listing for such a set by of the natural numbers. systematically generating all proofs from Γ. The Together, the compactness and the question arises of whether, beside this positive Löwenheim–Skolem theorem are the beginning test, there might not also be a negative test for a

265 G. Aldo Antonelli sentence φ being a consequence of Γ. This sense a relation that fails to satisfy this require- “decision problem” (Entscheidungsproblem) was ment can be called a consequence relation. originally proposed by David Hilbert in 1900, Cut, a form of transitivity, is another crucial and it was solved in 1936 independently by feature of consequence relations. Cut is a con- Alonzo Church and Alan Turing. The Church– servativity principle: if φ is a consequence of Γ, Turing theorem states that, in general, it is not then ψ is a consequence of Γ together with φ decidable whether Γ 3 φ, or even if φ is valid. (It only if it is already a consequence of Γ alone. In is important to know that for many, even quite other words, by adjoining to Γ something which expressive, fragments of first-order logic the is already a consequence of Γ does not lead to decision problem is solvable; see Börger et al. any increase in inferential power. Cut is best 1997 for details.) We should also notice the fol- regarded as the statement that the “length” of a lowing fact that will be relevant later in our dis- proof does not affect the degree to which the cussion: say that a sentence φ is consistent if {φ} assumptions support the conclusion. Where φ is is consistent, i.e., if its negation ¬φ is not valid. already a consequence of Γ, if ψ can be inferred Then the set of all sentences φ such that φ is from Γ together with φ, then ψ can also be consistent is not even semidecidable, for a positive obtained via a longer “proof” that proceeds indir- test for such a set would yield a negative test for ectly by first inferring φ. It is immediate to check the set of all valid sentences, which would thus be that FOL satisfies Cut. decidable, against the Church–Turing theorem. It is worth noting that many forms of prob- abilistic reasoning fail to satisfy Cut, precisely because the degree to which the premises sup- port the conclusion is inversely correlated to the Consequence relations length of the proof. To see this, we adapt a well- known example. Let Ax abbreviate “x was born In the previous section, we defined a consequence in Pennsylvania Dutch country,” Bx abbreviate “x relation 3 by saying that Γ 3 φ if and only if φ is is a native speaker of German,” and Cx abbreviate true on every interpretation on which every “x was born in Germany.” Further, let Γ com- sentence in Γ is true. In general, it is possible to prise the statements “Most A’s are B’s,” “Most consider what abstract properties a relation of B’s are C’s,” and Ax. Then Γ supports Bx, and Γ consequence between sets of sentences and together with Bx supports Cx, but Γ by itself single sentences could have. Let 1 be any such does not support Cx. Statements of the form relation. Consider the following properties, all of “Most A’s are B’s” are interpreted probabilistic- which are satisfied by the consequence relation 3 ally, as saying that the conditional probability of of FOL: B given A is, say, greater that 50 percent; like- wise, we say that Γ supports a statement φ if Γ Supraclassicality: if Γ 3 φ then Γ 1 φ. assigns φ a probability p > 50 percent. Reflexivity: if φ∈Γ then Γ 1 φ; Since Γ contains “Most A’s are B’s” and Ax, Cut: If Γ 1 φ and Γ, φ 1 ψ then Γ 1 ψ; it supports Bx (in the sense that the probability Monotony: If Γ 1 φ and Γ⊆∆ then ∆ 1 φ. of Bx is greater that 50 percent); similarly, Γ together with Bx supports Cx; but Γ by itself The first property is supraclassicality, which cannot support Cx. Indeed, the probability of states that if φ follows from Γ in FOL, then someone who was born in Pennsylvania Dutch it also follows according to 1; i.e., 1 extends 3 country being born in Germany is arbitrarily close (the relation 3 is trivially supraclassical). Of the to zero. Examples of inductive reasoning such as remaining conditions, the most straightforward the one just given cast some doubt on the possib- is reflexivity: it says that if φ belongs to the set Γ, ility of coming up with a well-behaved relation then φ is a consequence of Γ. This is a very of probabilistic consequence (see Chapter 21, minimal requirement on a relation of logical con- PROBABILITY IN ). sequence. We certainly would like all sentences Special considerations apply to monotony. in Γ to be inferable from Γ. It’s not clear in what Monotony states that if φ is a consequence of Γ

266 Logic then it is also a consequence of any set contain- Rational Monotony: If Γ 6 ¬φand Γ 1 ψ, then ing Γ (as a subset). The import of monotony is Γ, φ 1 ψ. that one cannot preempt conclusions by adding new premises to the inference. It is clear why Both Cautious Monotony and the stronger prin- FOL satisifies monotony: semantically, if φ is true ciple of Rational Monotony are special cases of on every interpretation on which all sentences of Monotony, and are therefore not in the fore- Γ are true, then φ is also true on every interpreta- ground as long as we restrict ourselves to the tion on which all sentences in a larger set ∆ are classical consequence relation 3 of FOL. true (similarly, proof-theoretically, if there is a Although superficially similar, these principles proof of φ all of whose assumptions are drawn are quite different. Cautious Monotony is the from Γ, then there is also a proof of φ – indeed, converse of Cut: it states that adding a con- the same proof – all of whose assumptions are sequence φ back into the premise-set Γ does not drawn from ∆). lead to any decrease in inferential power. Cautious Many consider this feature of FOL as in- Monotony tells us that inference is a cumulative adequate to capture a whole class of inferences enterprise: we can keep drawing consequences typical of everyday (as opposed to mathematical that can in turn be used as additional premises, or formal) reasoning, and therefore question without affecting the set of conclusions. Together the descriptive adequacy of FOL, when it comes with Cut, Cautious Monotony says that if φ is a to representing commonsense inferences. In consequence of Γ then for any proposition ψ, ψ everyday life, we quite often reach conclusions is a consequence of Γ if and only if it is a con- tentatively, only to retract them in the light of sequence of Γ together with φ. It has been often further information. For instance, when told that pointed out by Dov Gabbay that Reflexivity, Cut Stellaluna is a mammal, we infer that she does and Cautious Monotony are critical properties not fly, because mammals, by and large, don’t for any well-behaved nonmonotonic consequence fly. But the conclusion that Stellaluna doesn’t fly relation (see Gabbay et al. 1994, Stalnaker 1994). can be retracted when we learn that Stellaluna is The status of Rational Monotony is much more a bat, because bats are a specific kind of mammals, problematic. As we observed, Rational Monotony and they do fly. So we infer that Stellaluna does can be regarded as a strengthening of Cautious fly after all. This process can be further iterated. Monotony, and like the latter it is a special case We can learn, for instance, that Stellaluna is a of Monotony. However, there are reason to think baby bat, and that therefore she does not know that Rational Monotony might not be a correct how to fly yet. Such complex patterns of defeasible feature of a nonmonotonic consequence rela- reasoning are beyond the reach of FOL, which tion. A counterexample due to Stalnaker (1994: is, by its very nature, monotonic. 19) involves three composers: Verdi, Bizet, and For these and similar reasons, people have Satie. Suppose that we initially accept (correctly striven, over the last 20 years or so, to devise but defeasibly) that Verdi is Italian, while nonmonotonic formalisms capable of represent- Bizet and Satie are French. Suppose now that ing defeasible inference. We will take a closer we are told by a reliable source of information look at these formalisms below, but for now we that Verdi and Bizet are compatriots. This leads want to consider the issue from a more abstract us no longer to endorse the propositions that point of view. Verdi is Italian (because he could be French), When one gives up monotony in favor of and that Bizet is French (because he could be descriptive adequacy, the question arises of what Italian); but we would still draw the defeasible formal properties of the consequence relation to consequence that Satie is French, since nothing put in its place. Two such properties have been that we have learned conflicts with it. By letting considered in the literature, for an arbitrary con- I(v), F(b), and F(s) represent our initial beliefs sequence relation 1: about the nationality of the three composers, and C(v, b) represent that Verdi and Bizet are Cautious Monotony: If Γ 1 φ and Γ 1 ψ, then Γ, compatriots, the situation could be represented φ 1 ψ. as follows:

267 G. Aldo Antonelli

C(v, b) 1 F(s). very essence of defeasible reasoning that conclu- sions can be retracted when new facts are learned. Now consider the proposition C(v, s) that Verdi But conflicts of the second kind can be handled and Satie are compatriots. Before learning that in two different ways: one can draw inferences C(v, b) we would be inclined to reject the pro- either in a “cautious” or “bold” fashion (also position C(v, s) because we endorse and I(v)and known as “skeptical” or, respectively, “credu- F(s), but after learning that Verdi and Bizet are lous”). These two options correspond to widely compatriots, we can no longer endorse I(v), and different ways to construe a given body of defeas- therefore no longer reject C(v, s). The situation ible knowledge, and yield different results as to then is as follows: what defeasible conclusions are warranted on the basis of such a knowledge base. C(v, b) 6 ¬ C(v, s). The difference between these basic attitudes comes to this. In the presence of potentially con- However, if we added C(v, s) to our stock of flicting defeasible inferences (and in the absence beliefs, we would lose the inference to F(s): in of further considerations such as specificity – see the context of C(v, b), the proposition C(v, s) is below), the credulous reasoner always commits equivalent to the statement that all three com- to as many defeasible conclusions as possible, posers have the same nationality, and this leads us subject to a consistency requirement, whereas the to suspend our assent to the proposition F(s). In skeptical reasoner withholds assent from potenti- other words, and contrary to Rational Monotony: ally conflicted defeasible conclusions. A famous example from the literature, the C(v, b), C(v, s) 6 F(s). so-called “Nixon diamond,” will help make the distinction clear. Suppose our knowledge base The previous discussion gives a rather clear contains (defeasible) information to the effect picture of the desirable features a nonmonotonic that a given individual, Nixon, is both a Quaker consequence relation. Such a relation should and a Republican. Quakers, by and large, are satisfy Supraclassicality, Reflexivity, Cut, and pacifists, whereas Republicans by and large are Cautious Monotony. not. The question is, what defeasible conclusions are warranted on the basis of this body of know- ledge, and in particular whether we should infer that Nixon is a pacifist or that he is not pacifist. Varieties of Defeasible Reasoning Figure 20.1 provides a schematic representation of this state of affairs in the form of a (defeasible) A separate issue from the formal properties of a network. nonmonotonic consequence relation, although The credulous reasoner has no reason to prefer one that is strictly intertwined with it, is the either conclusion (“Nixon is a pacifist”; “Nixon issue of how conflicts between potential defeasible is not a pacifist”) to the other one, but will de- conclusions are to be handled. finitely commit to one or the other. The skeptical There are two different kinds of conflicts that reasoner recognizes that this is a conflict not can arise within a given nonmonotonic frame- between hard facts and defeasible inferences, but work: (i) conflicts between defeasible conclusions between two different defeasible inferences. Since and “hard facts”; and (ii) conflicts between one the two possible inferences in some sense “can- potential defeasible conclusion and another (many cel out,” the skeptical reasoner will refrain from formalisms, for instance, provide some form of drawing either one. defeasible inference rules, and such rules might Whereas many of the early formulations of have conflicting conclusions). When a conflict defeasible reasoning have been credulous, skept- (of either kind) arises, steps have to be taken to icism has gradually emerged as a viable altern- preserve or restore consistency. ative, which can, at times, be better behaved. All defeasible formalisms handle conflicts of Arguments have been given in favor of both the first kind in the same way: indeed, it is the skeptical and credulous inference. Some have

268 Logic

equate to represent defeasible reasoning was for Pacifist the first time accompanied by several proposals of formal frameworks within which one could at least begin to talk about defeasible inferences in a precise way, with the long-term goal of provid- ing for defeasible reasoning an account that could Republican Quaker at least approximate the degree of success achieved by FOL in the formalization of mathematical reasoning. The publication of a monographic issue of the Artificial Intelligence Journal in 1980 can be regarded as the “coming of age” of defeas- Nixon ible formalisms. The development of nonmonotonic (or Figure 20.1: The Nixon diamond defeasible) logics has been guided all along by a rich supply of examples. One of the early sources of motivation for the development of non- argued that credulity seems to better capture a monotonic logic comes from database theory. certain class of intuitions, while others have Consider for instance the closed world assump- objected that although a certain degree of “jump- tion: suppose that you need to travel from ing to conclusions” is by definition built into Oshkosh to Minsk, so you consult your travel any nonmonotonic formalism, such jumping to agent, who, not surprisingly, tells you that there conclusions needs to be regimented, and that are no direct flights. How does the travel agent skepticism provides precisely the required regi- know? In a sense, he doesn’t: his database does mentation. (A further issue in the skeptical/ not list any direct flights between Oshkosh and credulous debate is the question of whether so- Minsk, and he assumes that the database is com- called “floating conclusions” should be allowed; plete. In other words, what we have in this see Horty 2002 for a review of the literature and example is an attempt to minimize the extension a substantial argument that they should not.) of a given predicate (“flight-between” in this case). Moreover, such a minimization needs to take place not with respect to what the database explicitly contains but with respect to what it Nonmonotonic Logics implies. The idea of minimization is at the basis of As we have mentioned, over the last twenty years one of the earliest nonmonotonic formalisms, or so, a number of so-called “nonmonotonic” McCarthy’s circumscription. Circumscription logical frameworks have emerged, expressly makes explicit the intuition that, all other things devised for the purpose of representing defeasible being equal, extensions of predicates should be reasoning. The development of such frameworks minimal. Again, consider principles such as “all represents one of the most significant develop- normal birds fly.” Here we are trying to minimize ments both in logic and artificial intelligence, the extension of the abnormality predicate, and and has wide-ranging consequences for our philo- assume that a given bird is normal unless we have sophical understanding of argumentation and positive information to the contrary. Formally, inference. this can be represented using second-order logic. Pioneering work in the field of nonmonotonic In second-order logic, in contrast to FOL, one logics was carried out beginning in the late 1970s is allowed to explicitly quantify over predicates, by (among others) J. McCarthy, D. McDermott, forming sentences such as ∃P∀xPx (“there is a & J. Doyle, and R. Reiter (see Ginsberg 1987 universal predicate”) or ∀P(Pa ↔ Pb) (“a and for a collection of early papers in the field). With b are indiscernible”). In circumscription, given these efforts, the realization (which was hardly predicates P and Q, we abbreviate ∀x(Px → Qx) new) that ordinary first-order logic was inad- as P ≤ Q, and P ≤ Q ∧ Q 8 P as P < Q. If A(P)

269 G. Aldo Antonelli

Flies

Mammal

Bat

Baby Bat

Stellaluna

Figure 20.2: An Inheritance network; links of the form A → B represent the fact that typical A’s are B’s, and links A →/ B represent the fact that typical A’s are not B’s is a formula containing occurrences of a predic- taxonomically organized body of knowledge, we ate P, then the circumscription of P in A is the presuppose that subclasses inherit properties from second-order sentence A*(P): their superclasses: dogs have lungs because they are mammals, and mammals have lungs. How- A(P) ∧ 5∃Q [A(Q) ∧ Q < P]. ever, there can be exceptions, which can interact in complex ways. To use an example already in- A*(P) says that P satisfies A, and that no smaller troduced, mammals, by and large, don’t fly; since predicate does. Let Px be the predicate “x is bats are mammals, in the absence of any informa- abnormal,” and let A(P) be the sentence “All tion to the contrary, we are justified in inferring normal birds fly.” Then the sentence “Tweety is that bats do not fly. But then we learn that bats a bird,” together with A*(P) implies “Tweety are exceptional mammals, in that they do fly: the flies,” for the circumscription axiom forces the conclusion that they don’t fly is retracted, and extension of P to be empty, so that “Tweety is the conclusion that they fly is drawn instead. normal” is automatically true. In terms of con- Things can be more complicated still, for in turn, sequence relations, circumscription allows us to as we have seen, baby bats are exceptional bats, define, for each predicate P, a nonmonotonic in that they do not fly (does that make them relation A(P) 1 φ that holds precisely when A*(P) unexceptional mammals?). Here we have poten- 3 φ. (This basic form of circumscription has been tially conflicting inferences. When we infer that generalized, for in practice one needs to minimize Stellaluna, being a baby bat, does not fly, we are the extension of a predicate, while allowing the resolving all these potential conflicts based on a extension of certain other predicates to vary.) specificity principle: more specific information over- From the point of view of applications, however, rides more generic information. Nonmonotonic circumscription has a major shortcoming, namely inheritance networks were developed for the the absence of a complete inference procedure, purpose of capturing taxonomic examples such due to the fact that, in general, second-order logic as the above. Such networks are collections of lacks such a procedure. The price one pays for nodes and directed (“is a”) links representing the greater expressive power of second-order logic taxonomic information. When exceptions are is that there are no complete axiomatizations, as allowed, the network is interpreted defeasibly. we have for FOL. Figure 20.2 gives a network representing this Another nonmonotonic formalism inspired by state of affairs. In such a network, if there is a the intuition of minimization of abnormalities is link of the form A → B, then information about nonmonotonic inheritance. Whenever we have a A’s is more specific than information about B’s,

270 Logic and hence should override it. Research on non- and there is no evidence that θ might be false, monotonic inheritance focuses on the different then the rule allows the inference of ξ. As is ways in which one can make this idea precise. clear, application of the rule requires that a con- The main issue in defeasible inheritance is to sistency condition be satisfied, and rules can inter- characterize the set of assertions that are sup- act in complex ways. In particular it is possible ported by a given network. It is of course not that application of a rule might cause the con- enough to devise a representational formalism, sistency condition to fail (as when θ is ¬ξ). one also needs to specify how the formalism is Reiter’s default logic uses the notion of an exten- to be interpreted, and this is precisely the focus sion to make precise the idea that the consist- of much work in nonmonotonic inheritance. Such ency condition has to be met both before and a characterization is accomplished through the after the rule is applied. Given a set Γ of defaults, notion of extension of a given network. There an extension for Γ is, roughly, a set of defaults are two competing characterizations of exten- whose consistency condition is met both before sion for this kind of networks, one that follows and after their being triggered; an extension the credulous strategy and one that follows the therefore represents a set of inferences that can skeptical one. Both proceed by first defining be reasonably and consistently drawn using the degree of a path through the network as the defaults from Γ. More in particular (and in typical length of the longest sequence of links connect- circular fashion), an extension for Γ is a maximal ing its endpoints, and then building extensions subset ∆ of Γ the conclusions of whose defaults by considering paths in ascending order of their both imply all the prerequisites of defaults in ∆ degrees. We are not going to review the details, and are consistent with all the justifications of since many of the same issues arise in connec- defaults in ∆. tion with default logic (which is treated to greater This definition can be made precise as follows. length below), but Horty 1994 provides an ex- By a default theory we mean a pair (W, ∆), where tensive survey. It is worth mentioning that since ∆ is a (finite) set of defaults, and W is a set of the notion of degree makes sense only in the sentences (a world description). The idea is that case of acyclic networks, special issues arise when W represents the strict or background informa- networks contain cycles (see Antonelli 1997 for tion, whereas ∆ specifies the defeasible informa- a treatment of inheritance on cyclic networks). tion. Given a pair (T1, T2) of sets of sentences, a Although the language of nonmonotonic net- default such as the equation above is triggered 3 η 7 ¬θ works is expressively limited by design (in that by (T1, T2) if and only if T1 and T2 (i.e., θ only links of the form “is a” can be represented is consistent with T2). Notice how this defini- in a natural fashion), such networks represent an tion is built “on top” of 3: we could, conceiv- extremely useful setting in which to test and ably, employ a different relation here. Finally we hone one’s intuitions and methods for handling say that a set of sentences E is an extension for a defeasible information, which are then extended default theory (W; ∆) if and only if to more expressive formalisms. Among the latter = is Reiter’s “Default Logic,” which is perhaps the E E0 E1 ... En ..., most flexible among nonmonotonic frameworks. = In Default Logic, the main representational tool where: E0 W, and is that of a default rule, or simply a default. A default is a defeasible inference rule of the form η : θ EE+ = {ξ : ∈ ∆ nn1 ξ ηθ : is triggered by (EE, )} , n ξ (notice the occurrence of the limit E in the defini- η θ ξ (where , , are sentences in a given language, tion of En+1). There is an alternative characteriza- respectively called the prerequisite, the justifica- tion of extensions: given a default theory, let tion, and the conclusion of the default). The ᑭ be an operator defined on sets of sentences interpretation of the default is that if η is known, such that for any set S of sentences, ᑭ(S) is the

271 G. Aldo Antonelli smallest set containing W, deductively closed (i.e., One option is to take the union of the exten- such that if ᑭ(S) 3 φ then φ∈ᑭ(S)), and such sions of the theory, and consider φ a consequence that if a default with consequent ξ is triggered of a default theory (W, ∆) if and only if φ∈ E , by (S, S) then ξ ∈ ᑭ(S). Then one can show for some extension E. But this option is im- that E is an extension for (W, ∆) if and only if E mediately ruled out, in that it leads to endorsing is a fixed point of ᑭ, i.e., if ᑭ(E) = E. contradictory conclusion, as in the second ex- For any given default theory, extensions need ample above. It is widely believed that any viable not exist, and even when they exist, they need not notion of defeasible consequence for default logic be unique. Let us consider a couple of examples must have the property that the set {φ : (W, ∆) of these phenomena. Our first example is a de- 1 φ} must be consistent whenever W is. Once fault theory that has no extension: let W contain this option is ruled out, only two alternatives are the sentence η, and let ∆ comprise the single left. default The first alternative, known as the “credulous” or “bold” strategy, is to pick an extension E for ηθ : the theory, and say that φ is a defeasible con- . φ ∈ 5θ sequence if and only if E. The second altern- ative, known as the “skeptical” or “cautious” If E were an extension, then the default above strategy, is to endorse a conclusion φ if and only would have to be either triggered or not triggered if φ is contained in every extension of the theory. by it, and either case is impossible. Both the credulous and the skeptical strategy Let us now consider an example of a default have problems. The problem with the credulous theory with multiple extensions. Like before, let strategy is that the choice of E is arbitrary: with W contain the sentence η, and suppose ∆ com- the notion of extension introduced by Reiter, prises the two defaults extensions are orthogonal: of any two distinct ex- tensions, neither one contains the other. Hence, ηθ : η : ξ there seems to be no principled way to pick an , and . 5ξ 5θ extension over any other one. This has led a num- ber of researcher to endorse the skeptical strategy This theory has exactly two extensions, one in as a viable approach to the problem of defeas- which the first default is triggered and one in ible consequence. But as shown by Makinson, which the second one is. It is easy to see that at skeptical consequence, as based on Reiter’s notion least a default has to be triggered in any exten- of extension, fails to be cautiously monotonic. sion, and that both defaults cannot be triggered To see this, consider the default theory (W, ∆), by the same extension. where W is empty, and ∆ comprises the two These examples are enough to bring out a defaults: number of features. First, it should be noted that neither one of the two characterizations of : θ θη ∨ : 5 θ , and . default logic given above gives us a way to “con- θ 5θ struct” extension by means of anything resemb- ling an iterative process. Essentially, one has to This theory has only one extension, coinciding “guess” a set of sentences E, and then verify that with the deductive closure of {θ}. hence, if we it satisfies the definition of an extension. put (W, ∆) 1 φ if and only if φ belongs to every Further, the fact that default theories can have extension of (W, ∆), we have (W, ∆) 1 θ, as well zero, one, or more extensions raises the issue of as (W, ∆) 1 θ∨η(by the deductive closure of what inferences one is warranted in drawing from extensions). Now consider the theory with ∆ as a given default theory. The problem can be pre- before, but with W containing the sentence θ∨ sented as follows: given a default theory (W; ∆), η. This theory has two extensions: one the same what sentences φ can be regarded as defeasible as before, but also another one coinciding with consequences of the theory? On the face of it, the deductive closure of {¬θ}, and hence not there are several options available. containing θ. It follows that the intersection of

272 Logic the extensions no longer contains θ, so that ({θ ing, which belong to the family of autoepistemic ∨η}, ∆) 6 θ, against cautious monotony. (Notice logics. Modal logics in general have proved to be that the same example establishes a counter- one of the most flexible tools for modeling all example for Cut for the credulous strategy, when sorts of dynamic processes and their complex we pick the extension of ({θ∨η}, ∆) that con- interactions. Beside the applications in knowledge tains ¬θ.) representation, which we are going to treat It is clear that the issue of how to define a below, there are modal frameworks, known as nonmonotonic consequence relation for default dynamic logics, that play a crucial role, for in- logic is intertwined with the way that conflicts stance, in the modeling of serial or parallel com- are handled. The problem of course is that in putation. The basic idea of modal logic is that this case neither the skeptical nor the credulous the language is interpreted with respect to a give strategy yields an adequate relation of defeas- set of states, and that sentences are evaluated ible consequence. In Antonelli 1999 a notion of relative to one of these states. What these states general extension for default logic is introduced, are taken to represent depends on the particular showing that this notion yields a well-behaved application under consideration (they could be relation of defeasible consequence that satisfies epistemic states, or states in the evolution of a all four requirements of Supraclassicality, Reflex- dynamical system, etc.), but the important thing ivity, Cut, and Cautious Monotony. is that there are transitions (of one or more dif- A different set of issues arises in connection ferent kinds) between states. In the case of one with the behavior of default logic from the point transition that is both transitive (i.e., such that if of view of computation. As we have seen for a a → b and b → c then a → c) and euclidean (if a given semidecidable set Γ of sentences, the set → b and a → c then b → c), the resulting modal of all Γ that are a consequence of Γ in FOL is system is referred to as K45. Associated with itself semidecidable. In the case of default logic, each kind of state transition there is a corres- to formulate the corresponding problem one ponding modality in the language, usually rep- extends (in the obvious way) the notion of resented as a box ٗ. A sentence of the form ٗA (semi)decidability given above to sets of defaults. is true at a state s if and only if A is true at every The problem, then, is to decide, given a default state s ′ reachable from s by the kind of transition -theory (W, ∆) and a sentence φ whether (W, ∆) associated with ٗ (see Chellas 1980 for a com 1 φ, where 1 is defined, say, skeptically (it doesn’t prehensive introduction to modal logic). really make a difference computationally whether In autoepistemic logic, the states involved are 1 is defined skeptically or credulously). Such a epistemic states of the agent (or agents). The problem is not even semidecidable, the essential intuition underlying autoepistemic logic is that reason being that in general, in order to deter- we can sometimes draw inferences concerning the mine whether a default is triggered by a pair of state of the world using information concerning sets of sentences, one has to perform a consist- our own knowledge or ignorance. For instance, ency check. But the consistency checks are not the I can conclude that I do not have a sister given only source of complexity in default logic. For that if I did I would probably know about it, and instance, we could restrict our language to con- nothing to that effect is present in my “know- junctions of atomic sentences and their negations ledge base.” But such a conclusion is defeasible, (making consistency checks feasible). Even so, since there is always the possibility of learning the problem of determining whether a given new facts. default theory has an extension would still be In order to make these intuitions precise, con- highly intractable (NP-complete, to be precise, sider a modal language in which the necessity ”.as shown by Kautz & Selman 1991), seemingly operator ٗ is interpreted as “it is known that because the problem requires checking all pos- As in default logic or defeasible inheritance, the sible sequences of firings of defaults (see Chap- central notion in autoepistemic logic is that of ter 2, COMPLEXITY, for these and related notions). an extension of a theory S, i.e., a consistent and Default logic is intimately connected with cer- self-supporting sets of beliefs that can reasonably tain modal approaches to nonmonotonic reason- be entertained on the basis of S. Given a set S of

273 G. Aldo Antonelli

sentences, let S0 be the subset of S composed of with computational complexity of the most basic .those sentences containing no occurrences of ٗ; questions concerning the framework i further, let the introspective closure S 0 of S0 be There is a potential tension between (i) and the set (ii): the desire to capture a broad range of intui- tions can lead to ad hoc solutions that can some- ∋ φ φٗ { : S0}, times undermine the desirable formal properties of the framework. In general, the development n and the negative introspective closure S 0 of S0 the of nonmonotonic logics and related formalisms set has been driven, since its inception, by considera- tion (i) and has relied on a rich and well-chosen ∌ 5ٗφ φ { : S0}. array of examples. Of course, there is some ques- tion as to whether any single framework can i The set S 0 is called the introspective closure aspire to be universal in this respect. because it explicitly contains positive information More recently, researchers have started pay- i about the agent’s epistemic status: S 0 expresses ing attention to consideration (ii), looking at n what is known (similarly, S 0 contains negative the extent to which nonmonotonic logics have information about the agent’s epistemic status, generated well-behaved relations of logical con- stating explicitly what is not known). With these sequence. As Makinson (1994) points out, prac- notions in place, we define an extension for S to titioners of the field have encountered mixed be a set T of sentences such that: success. In particular, one abstract property, Cau- tious Monotony, appears at the same time to be = φ φ i n T { : follows from S T 0 T 0 in crucial and elusive for many of the frameworks K45}. to be found in the literature. This is a fact that is perhaps to be traced back, at least in part, to the Autoepistemic logic provides a rich language, above-mentioned tension between the require- with interesting mathematical properties and con- ment of material adequacy and the need to gen- nections to other nonmonotonic formalisms. It is erate a well-behaved consequence relation. faithfully intertranslatable with Reiter’s version The complexity issue appears to be the most of default logic, and provides a defeasible frame- difficult among the ones that have been singled work with well-understood modal properties. out. Nonmonotonic logics appear to be stub- bornly intractable with respect to the correspond- ing problem for classical logic. This is clear in the case of default logic, given the ubiquitous Conclusion consistency checks. But beside consistency checks, there are other, often overlooked, sources of com- There are three major issues connected with plexity that are purely combinatorial. Other forms the development of logical frameworks that can of nonmonotonic reasoning, beside default logic, adequately represent defeasible reasoning: (i) are far from immune from these combinatorial material adequacy; (ii) formal properties; and roots of intractability. Although some import- (iii) complexity. Material adequacy concerns the ant work has been done trying to make various question of how broad a range of examples is nonmonotonic formalisms more tractable, this is captured by the framework, and the extent to perhaps the problem on which progress has been which the framework can do justice to our intui- slowest in coming. tions on the subject (at least the most entrenched ones). The question of formal properties has to do with the degree to which the framework References allows for a relation of logical consequence that satisfies the above-mentioned conditions of Antonelli, G. Aldo. 1997. “Defeasible inheritance Supraclassicality, Reflexivity, Cut, and Cautious over cyclic networks.” Artificial Intelligence 92(1): Monotony. The third set of issues has to do 1–23. [A treatment of nonmonotonic inheritance

274 Logic

networks that include cycles, inspired by Kripke’s for Gentzen systems, resolution, and many-sorted 3-valued approach to truth theories.] first-order logic.] ——. 1999. “A directly cautious theory of defeasible Galton, Antony. 1990. Logic for Information Tech- consequence for default logic via the notion of nology. Chichester and New York: John Wiley & general extension.” Artificial Intelligence 109 Sons. [A thorough introduction to proof systems (1–2): 71–109. [Introduces a well-behaved con- for first-order logic and their metatheory.] sequence relation for default logic, based on Ginsberg, M. L., ed. 1987. Readings in Non- a generalization of Reiter’s original notion of monotonic Reasoning. Los Altos, CA: Morgan extension.] Kauffman. [A collection of early papers in Börger, Egon, Grädel, Erich, and Gurevich, Yuri. nonmonotonic logic. Somewhat hard to find, but 1997. The Classical Decision Problem. Berlin and invaluable.] New York: Springer-Verlag. [A standard refer- Gödel, Kurt. 1930. “Die Vollständigkeit der Axiome ence on the complexity of the decision problem des logischen Funktionenkalküls.” Monatshefte für for FOL and many of its fragments.] Mathematik und Physik 37. [A classic.] Chellas, Brian F. 1980. Modal Logic: An Introduc- Horty, John F. 1994. “Some direct theories of non- tion. Cambridge: Cambridge University Press. monotonic inheritance.” In Gabbay et al. 1994: [The standard introduction to modal logic. Cover 111–87. [A very clear presentation of all the major standard possible-world semantics as well as other issues in defeasible inheritance.] variants.] ——. 2002. “Skepticism and floating conclusions.” Ebbinghaus, H.-D., Flum, J., and Thomas, W. Artificial Intelligence Journal 135(1–2): 55–72. 1994. Mathematical Logic, 2nd ed. New York [This article addresses some issues in the founda- and Berlin: Springer-Verlag. [A graduate-level tions of defeasible reasoning. Makes the case that, introduction to symbolic logic. Notable for its contrary to what many have argued, floating con- treatment of Lindström’s results.] clusions are not always warranted.] Enderton, Herbert. 1972. A Mathematical Intro- Kautz, H. and Selman, B. 1991. “Hard problems duction to Logic, 2nd ed. New York: Academic for simple default logic.” Artificial Intelligence Press. New York: Harcourt/Academic Press. [An Journal 49: 243–79. [A seminal work addressing introduction to logic aimed at advanced under- complexity issues in nonmonotonic inference.] graduates. Enderton’s axiomatization of FOL is Lindström, Per. 1969. “On extensions of element- widely used.] ary logic.” Theoria 35. [The original reference Fitting, Melvin. 1990. First-order Logic and Auto- for Lindström’s theorems.] mated Theorem Proving. New York and Berlin: Makinson, David. 1994. “General patterns in non- Springer-Verlag. [An introduction to formal logic monotonic reasoning.” In Gabbay et al. 1994: emphasizing the mechanization of theorem- 35–110. [Best general treatment of the issues proving. Notable for its treatment of both resolu- concerning abstract consequence relations for vari- tion and tableaux methods.] ous defeasible formalisms.] Gabbay, D. M., Hogger, C. J., and Robinson, J. A., Stalnaker, Robert. 1994. “Nonmonotonic con- eds. 1994. Handbook of Logic in Artificial Intel- sequence relations.” Fundamenta Informaticæ 21: ligence and Logic Programming, vol. 3. Oxford: 7–21. [A deep paper assessing several abstract pro- Oxford University Press. [Contains excellent sur- perties of nonmonotonic consequence relations.] veys of all the major nonmonotonic formalisms Tarski, Alfred. 1935. “Der Wahrheitsbegriff in den as well as articles addressing the foundations of formalisierten Sprachen.” Studia Logica, pp. 261– nonmonotonic logic.] 405: English tr. in Tarski 1956: 152–278. [The Gallier, Jean H. 1986. Logic for Computer Science: classic, original treatment of semantics for FOL.] Foundations of Automated Theorem Proving. New ——. 1956. Logic, Semantics, and Metamathematics. York: Harper & Row. [A technically sophistic- Oxford: Oxford University Press, ed. and tr. ated treatment of first-order logic using mainly J. H. Woodger. [A collection of Tarski’s mostly proof-theoretic methods. Covers Cut-elimination technical papers.]

275 Donald Gillies

Chapter 21 Probability in Artificial Intelligence Donald Gillies

out a specialized task, and then code this know- Introduction ledge into the computer. In this way they were able to produce “expert systems” which per- The first two sections of this chapter provide an formed specific tasks at the level of human account of how the development of artificial intel- experts. One of the most important of these early ligence (AI) led to some remarkable innovations expert systems (MYCIN) was concerned with in probability theory – most notably to the new the diagnosis of blood infections. This system theory of Bayesian networks. The final section will now be briefly described, and it will then be presents some discussion of the philosophical shown that its implementation led to the prob- problems posed by these new ideas. As we shall lem of how to introduce probability into AI. see, they have a considerable impact on quite a MYCIN was developed in the 1970s by number of traditional questions in the philosophy Edward Shortliffe and his colleagues in collabora- of science. tion with the infectious diseases group at the Stanford medical school. The medical knowledge in the area was codified into rules of the form: IF such and such symptoms are observed, THEN The Breakthrough with Expert likely conclusion is such and such. MYCIN’s Systems in the 1970s knowledge base comprised over 400 such rules which were obtained from medical experts. An Research in AI began in the 1950s and many example of such a rule will be given in a moment, important ideas were developed by the pioneers but first it would be as well to present some (see Chapter 9, THE PHILOSOPHY OF AI AND ITS evidence of MYCIN’s success. CRITIQUE). Then in the 1970s a breakthrough To test MYCIN’s effectiveness a compar- was produced by the creation of expert systems. ison was made in 1979 of its performance with The lead here was taken by the Stanford heur- that of 9 human doctors. The program’s final istic programming group, particularly Buchanan, conclusions on 10 real cases were compared Feigenbaum, and Shortliffe. What they discovered with those of the human doctors, including the was that the key to success was to extract from actual therapy administered. Eight other experts an expert the knowledge he or she used to carry were then asked to rate the 10 therapy recom-

276 Probability in Artificial Intelligence mendations and award a mark, without knowing question, and they answer (1975: 352–5) that which, if any, came from a computer. They were in typical medical applications there is not enough requested to give 1 for a therapy which they data to obtain the requisite objective probabil- regarded as acceptable and 0 for an unacceptable ities. This in turn is because of the inadequacy of therapy. Since there were 8 experts and 10 cases, hospital records, and the changes which are con- the maximum possible mark was 80. The results tinually occurring in disease categories. It is inter- were as follows (cf. Jackson 1986: 106): esting to note that only three years previously, another group working on computer diagnosis MYCIN 52 Actual therapy 46 had reached exactly the opposite conclusion. This Faculty-1 50 Faculty-4 44 research group, working in Leeds, was headed Faculty-2 48 Resident 36 by de Dombal. Their results are contained in de Inf dis fellow 48 Faculty-5 34 Dombal et al. 1972, and Leaper et al. 1972. Faculty-3 46 Student 24 They will now be briefly summarised. De Dombal’s group created a computer diag- So MYCIN came first in the exam, though the nosis system for acute abdominal pain based on difference between it and the top human experts Bayesian reasoning (see appendix), but using was not significant. objective probabilities obtained from a sample Let us now examine one of MYCIN’s rules. of 600 patients. In a test involving 472 patients, The following rule is given by Shortliffe and this system was correct in 91.1 percent of cases, Buchanan in their 1975, p. 357. while the clinical team was correct in 79.7 per- cent of cases. The system was then changed by If: (1) the stain of the organism is gram using instead of these objective probabilities, positive (S ), and 1 subjective probabilities obtained from the clini- (2) the morphology of the organism is cians. Its performance for the same 472 patients, coccus (S ), and 2 but using these subjective probabilities, dropped (3) the growth conformation of the from 91.1 to 82.2 percent. Moreover the per- organism is chains (S ) 3 formance of the computer system with subject- Then: there is suggestive evidence (0.7) that ive probabilities was actually worse than that the identity of the organism is streptococcus of the clinicians in the case of diseases which

(H1). occurred relatively rarely. The conclusion seemed inescapable that human doctors are bad at estim-

In symbols this could be written: If S1 & S2 & ating probabilities, especially in the case of dis-

S3, then there is suggestive evidence p that H1, eases which occur infrequently, and that therefore = where p 0.7. Here S1, S2, S3 are the observa- the use of objective probabilities obtained from tions/symptoms, which support hypothesis H1 data is preferable. to a particular degree. These rules were obtained Despite the strong evidence for this con- from the medical experts. The numbers they clusion, it was ignored for the next 20 years, contain such as 0.7 look like probabilities, and and nearly every researcher in the field made they too were obtained from the experts. The use of subjective probabilities. There are two expert was in fact asked: “On a scale of 1 to 10, possible reasons for this. First of all it may, in how much certainty do you affix to this conclu- many cases, have been difficult to obtain object- sion?” The answer was then divided by 10. ive probabilities from data. Secondly the general It looks as if Shortliffe and Buchanan are using methodology of expert systems research, since it probability in the subjective sense to measure involved obtaining knowledge from the experts, the degree of personal belief held by an expert. may have encouraged the idea of obtaining prob- This at once raises the question of why subject- abilities as the degrees of belief of these experts. ive probabilities (see the appendix to this chapter) I will discuss further the question of objective obtained from experts are preferred to object- versus subjective probabilities in the final sec- ive probabilities (see appendix) obtained from tion of this chapter, which deals with the related data. Shortliffe and Buchanan do consider this philosophical problems. Let us now return to a

277 Donald Gillies consideration of MYCIN, and the sample rule given earlier. H1 So far we have rather assumed that the number 0.7 in the rule from MYCIN is an ordinary prob- ability, but this is not the case, as Shortliffe and Buchanan make clear in the following passage H2 (1975: 358):

this rule at first seems to say P(H1 | S1 & S2 H3

& S3) = 0.7, ....Questioning of the expert gradually reveals, however, that despite the apparent similarity to a statement regarding E1 E2 E3 a conditional probability, the number 0.7 dif- fers significantly from a probability. The expert may well agree that P(H | S & S & S ) = Figure 21.1: Part of PROSPECTOR’s inference 1 1 2 3 = 0.7, but he becomes uneasy when he attempts network. H1 There are massive sulfide deposits. = = to follow the logical conclusion that therefore H2 There are clay minerals. H3 There is a = P(not.H | S & S & S ) = 0.3. The three reduction process. E1 Barite is overlying sulfide. 1 1 2 3 = observations are evidence (to degree 0.7) in E2 Galena, sphalerite, or chalcopyrite fill cracks = favor of the conclusion that the organism is a in rhyolite or dacite. E3 There are bleached streptococcus and should not be construed as rocks. evidence (to degree 0.3) against streptococcus.

Shortliffe and Buchanan used this observation to pieces of evidence imply the same hypothesis. motivate the introduction of a nonprobabilistic Furthermore, there are often chains of evid- model of evidential strength. Their measure of ences and hypotheses. For these reasons it is evidential strength was called a certainty factor, natural to represent a collection of rules as a and certainty factors neither obeyed the standard graph structure or inference net. axioms of probability theory, the Kolmogorov axioms (see appendix), nor combined like A part of PROSPECTOR’s inference network probabilities. is shown in figure 21.1. Evidence E1 is taken as Certainty factors were criticized by those who supporting hypothesis H1, and this is indicated favored a probabilistic approach – cf. Adams 1976 by the arrow joining them in the inference net- and Heckerman 1986 – and in fact the next work. Similarly E2 supports hypothesis H1, while expert system we will consider (PROSPECTOR) E3 supports H3 which supports H2 which sup- did move more in the direction of standard ports H1. Note how these rather complicated rela- probability. tions are simply and elegantly represented by the PROSPECTOR, an expert system for mineral arrows of the network. Each inference arrow has exploration, was developed in the second half of a strength associated with it, and this is obtained the 1970s at the Stanford Research Institute. A from the expert as in the case of MYCIN. good general account of the system is given by PROSPECTOR, however, differs from MYCIN Gaschnig in his 1982. PROSPECTOR’s most in using subjective Bayesianism (see appendix) important innovation was to represent know- rather than certainty factors. This subjective ledge by an inference network (or net). This is Bayesianism is not entirely pure, since, as in the motivated by Duda et al in their 1976 as follows case of MYCIN, it is combined with the use of (p. 1076): fuzzy logic formulae (see appendix). This use of fuzzy logic tended to disappear in further A collection of rules about some specific sub- developments. ject area invariably uses the same pieces of evid- In PROSPECTOR, Bayesianism is formulated ence to imply several different hypotheses. It using odds rather than probabilities. The odds also frequently happens that several alternative on a hypothesis H[O(H)] are defined as follows:

278 Probability in Artificial Intelligence

O(H) = P(H)/P(¬H) (21.3) are plausible. The search for a justifica- tion of these assumptions led, as we shall see in Writing down Bayes theorem (see appendix) first the next section, to the modification of the con- for H and then for ¬H, we get cept of inference network, and the emergence of the concept of Bayesian network. P(H |E) = P(E |H)P(H)/P(E) P(¬H | E) = P(E | ¬H)P(¬H)/P(E) The Emergence of Bayesian So dividing gives Networks in the 1980s O(H | E) =λ(E)O(H) (21.1) The concept of Bayesian networks was intro- where λ(E) is the likelihood ratio P(E | H)/P(E duced and developed by Pearl in a series of | ¬H). (21.1) is the odds and likelihood form of papers: Pearl 1982, 1985a, 1985b, 1986, Kim & Bayes theorem, and it is used in PROSPECTOR Pearl 1983, and a book: Pearl 1988. An import- to change the prior odds on H to the posterior ant extension of the theory was carried out by odds given evidence E. Lauritzen and Spiegelhalter (1988), while Nea- Let us now consider the problems which arise politan’s 1990 book gave a clear account of these new ideas and helped to promote the use of if we have several different pieces of evidence E1, Bayesian networks in the AI community. In what E2,..., En say. We might in practice have to update using any subset of these pieces of evid- follows, I will comment on a few salient features of Bayesian networks which will be important ence Ei, Ej,..., Ek say, where (i, j,...k) is any subset of (1, 2, . . . , n). If we use (21.1), this when we consider their philosophical implica- λ tions in the next section. would involve having values of (Ei & Ej & The actual term Bayesian (or Bayes) network ...& Ek) for all subsets of (1, 2, . ..n). When we remember that, on this approach the values was introduced in Pearl’s 1985b, where it is of λ are obtained from the domain experts, we defined as follows (1985b: 330): can see that obtaining the requisite values of λ is scarcely possible. Clearly some simplifying Bayes Networks are directed acyclic graphs in assumptions are necessary to produce a workable which the nodes represent propositions (or variables), the arcs signify the existence of dir- system, and the designers of PROSPECTOR ect causal influences between the linked pro- therefore made the following two conditional positions, and the strengths of these influences independence (see appendix) assumptions: are quantified by conditional probabilities. = P(E1,..., En | H) P(E1 | H)...P(En | H) This verbal account is illustrated by a diagram (21.2) which is reproduced in figure 21.2. If we compare the network of figure 21.2 ¬ = ¬ P(E1,..., En | H) P(E1 | H)... with that of figure 21.1, two differences should ¬ P(En | H) (21.3) be noted immediately. First of all, the arrows in the inference network of figure 21.1 represent a

Given these assumptions, the whole problem of relation of support holding between e.g. E3 and updating with many pieces of evidence becomes H3, while the arrows in the Bayesian network of simple, and, in fact, figure 21.2 represent causal influences, so that,

e.g. the arrow joining X1 to X2 means that X1 =λ λ λ O(H |E1 &...& En) 1, 2 ... n O(H) causes X2. Secondly, corresponding to the first λ =λ where i (Ei) difference, we can say that, in a certain sense, the arrows of a Bayesian network run in the opposite The only remaining problem was whether the direction to those of an inference network. Pearl conditional independence assumptions (21.2) and puts this point as follows (1986: 253–4):

279 Donald Gillies

X 1 D

A

X2 X3

C B

X5 X4 F E

X6 Figure 21.4: Pearl’s (1982) tree example Figure 21.2: A Bayes network and says that one factor that led him to the idea was his consideration of the concept of influence (a) Inference Network E H 3 3 diagrams introduced by Howard and Matheson (1984). Pearl decided to limit the influences specifically to causal influences. But what is the advantage of this reversal (b) Bayesian Network E3 H3 of arrows and introduction of causal links? To answer this question, we must return to the ques- Figure 21.3: A set of nodes from figure 21.1 tion of the conditional independence assump- tions which are needed in order to make Bayesian in many expert systems (e.g. MYCIN), . . . rules updating (see appendix) feasible. Before doing point from evidence to hypothesis (e.g. if do, however, I would like to make one further symptom, the disease), thus denoting a flow of point about Bayesian networks. If, in such a net- mental inference. By contrast, the arrows in work, an arrow runs from node A to node B, Bayes’ networks point from causes to effects or then A is said to be a parent of B, and B a child from conditions to consequence, thus denot- of A. If a node has no parents, it is called a root,

ing a flow of constraints in the physical world. so that in figure 21.2, X1 is a root. In a Bayesian network, it is possible for a child to have several

This reversal of arrows from inference networks parents. Thus in figure 21.2, X5 has parents X2 to Bayesian networks is illustrated in figure 21.3, and X3. If, however, every child has at most one which shows one pair of nodes taken from the parent, the network is called a tree. Pearl started portion of PROSPECTOR’s inference network his investigation of networks with trees since they shown in figure 21.1. are mathematically simpler. Let us similarly begin = In figure 21.3, E3 There are bleached rocks, our account of the conditional independence = while H3 There is a reduction process. From assumptions in the case of trees by considering the point of view of an inference network (a), Pearl’s first paper on the subject, his 1982. This we regard the evidence of bleached rocks as paper is particularly helpful in showing how the supporting the hypothesis that there is a reduc- method of Bayesian networks developed from tion process, while, from the point of view of PROSPECTOR’s method of inference networks. a Bayesian network (b), we regard there being In his 1982, Pearl considers the example of a reduction process as the cause of there being a tree illustrated in figure 21.4. Here A, B, C, bleached rocks. In his 1993, Pearl gives an . . . are variables standing for hypotheses or account of his discovery of Bayesian networks, observations. A takes the values A1, A2, . . . , and 280 Probability in Artificial Intelligence similarly for the other variables. Pearl now intro- unexplicated by the network below. This u duces the conditional independence assumptions interpretation renders P(Bi |D (B)) identical which he is going to make as follows (1982: 134): to the classical notion of subjective prior probability.

let Dd(B) stand for the data obtained from the tree rooted at B, and Du(B) for the data These analogies can be further emphasized by obtained from the network above B. The the notation presence of only one link connecting Du(B) and (B) implies: λ = (Bi) defP(Dd(B) | Bi) π = u u = (Bi) defP(Bi |D (B)) P(Bj | Ai, D (B)) P(Bj | Ai)...

In other words, a node is conditionally inde- which allows one to write (21.4) in the form pendent given its parent of the rest of the net- =λα π work except its descendents. If we here substitute P(Bi | E) (Bi) (Bi). (21.5) “parents” for “parent,” we get a statement of the conditional independence assumptions which The analogy between (21.1) and (21.5) is clear. define a Bayesian network. On the basis of (21.5), Pearl develops an In fact Bayes’s theorem, together with the algorithm which allows Bayesian updating to take above conditional independence assumptions, place. If one of the variables which represents yields the product rule: an observation is set to a particular value, the changes brought about by this new information u P(Bi |D (B), Dd(B)) in all the probabilities throughout the tree can =α u P(Dd(B) | Bi) · P(Bi |D (B)) (21.4) be computed in an efficient manner. In sub- sequent work he extends this updating algorithm where α is a normalization constant. Pearl now to more complicated networks. Kim & Pearl makes a most interesting comparison between 1983 generalizes from trees to Bayesian networks (21.4) and the updating formula of PROSPEC- which are singly connected, i.e. there exists only TOR given as (21.1) above. one (undirected) path between any pair of nodes. Pearl in his 1986 tackled the further extension O(H | E) =λ(E)O(H) (21.1) to Bayesian networks which are multiply con- nected. This problem was also investigated by u Lauritzen & Spiegelhalter who in their 1988 Since D (B), Dd(B) represents the total data, solved it using the idea of reducing a multiply or evidence E bearing on our hypothesis Bi, u connected network to a tree of cliques. Their P(Bi |D (B), Dd(B)) corresponds to O(H |E). algorithm has been generally adopted by the AI P(Dd(B) | Bi) has the form of a likelihood (see λ u community. appendix) as does (E); while P(Bi | D (B)) cor- responds to the prior probability of traditional Let us now turn from these powerful math- Bayesianism in the following sense (Pearl 1982: ematical developments to the consideration of a 134): conceptual point. It will have been noted that two rather different definitions of Bayesian net- the multiplicative role of the prior probability work have been given. The first definition is in . . . is taken over by the conditional probabil- terms of causes. Thus in figure 21.2 the arrows ity of a variable based only on the evidence are taken as denoting a causal link between the gathered by the network above it, excluding two nodes which they join. The second defini- the data collected from below. . . . The root tion is by contrast purely probabilistic. In figure is the only node which requires a prior prob- 21.2 the variables X1, X2,...X6 are taken to be ability estimation, since it has no network random variables with a joint probability distri- above. Du(B) should be interpreted as the bution (see appendix), and the network becomes available background knowledge which remains a Bayesian network if the relevant conditional

281 Donald Gillies independence assumptions are satisfied. I will from the other C. Reichenbach’s causal fork is henceforth use the term “Bayesian network” for just a simple case of a Bayesian network. We can networks defined purely probabilistically in the indeed apply his term “screening off” to Bayesian manner just explained, and call the networks networks by saying that in such networks, the defined in terms of causes: “causal networks.” parents of a node screen it off from all the other Pearl tends, however, to use the terms “Bayesian nodes in the network except its descendents. network” and “causal network” interchangeably, We are now in a position to summarize the because he believes the two notions to be closely ingenious way in which Bayesian networks solved connected. More specifically, his idea is that if in the problem of handling uncertainty in expert a network the parents of every node represent systems. In most of the domains considered, e.g. the direct causes of that node, then the relevant medical diagnosis, a domain expert is very fam- conditional independence assumptions will auto- iliar with the various causal factors operating. It matically be satisfied. As he says (1993: 52): should therefore be an easy matter to get him or her to provide a causal network. By the addition Causal utterances such as “X is a direct cause of probabilities this can be turned into a Bayesian of Y” were given a probabilistic interpreta- network. In earlier systems such as MYCIN tion as distinctive patterns of conditional inde- or PROSPECTOR, conditional independence pendence relationships that can be verified assumptions were made for the purely ad hoc empirically. and pragmatic reason of allowing the updating to become possible. For Bayesian networks, how- A suggested link between causality and conditional ever, the causal information obtained from the independence in fact goes back to Reichenbach expert provides a justification for making a set 1956. Reichenbach considers two events, B and of conditional independence assumptions in the C say, which are correlated. For example, in a manner first suggested by Reichenbach. Moreover traveling troupe of actors, B = the leading lady as Pearl, Lauritzen, and Spiegelhalter have shown, has a stomach upset, and C = the leading man this set of conditional independence assumptions has a stomach upset. We can explain such correla- is sufficient to allow Bayesian updating to become tions, according to Reichenbach, by finding a computationally feasible. Everything fits together common cause, namely that the leading lady in a most satisfying manner. There is only one and the leading man always have dinner together. weak link in the chain. It turns out, as we shall The common stomach upsets occur when the see in the next section, that it is possible to have food in the local restaurant has gone off. Denote a bona fide causal network in which the requisite “dining together” by A. We then have the causal conditional independence assumptions are not graph shown in figure 21.5. satisfied. Reichenbach then claimed that, conditional on A, B, and C were no longer correlated but independent, i.e. P(B&C |A) = P(B |A)P(C | A). He also expressed this idea by saying that a Philosophical Problems Connected common cause A screens one of its effects B off with Probability in AI

The preceding sections have outlined some re- B markable developments in AI. Let us now turn to a consideration of the philosophical implica- tions of these developments. In fact they have a A profound impact on at least three central prob- lems in the philosophy of science. The first of these is the Bayesianism versus non-Bayesianism C debate. This has continued among philosophers of science for the last 50 years with no signs of Figure 21.5: Dining together abating. In the 1950s the major contenders were

282 Probability in Artificial Intelligence

Carnap (in favor of Bayesianism) versus Popper Let us start therefore with the long-running (against Bayesianism). In the late 1980s and controversy about Bayesianism. As so often hap- 1990s, we have had Howson & Urbach (1989) in pens in such controversies, it turns out that the favor of Bayesianism, and Miller (1994) against, definition of Bayesianism is not entirely clear, while the most recent developments in the debate and I believe that the controversy involves two as seen by leading experts in the area are to rather different issues. The first of these issues be found in Corfield & Williamson 2001. The is the question of whether we should use the new results in AI are clearly relevant to this standard mathematical calculus of probability in controversy, and indeed would seem to favor the handling uncertainty, or whether some other cal- Bayesian camp, though, as we shall see, this sup- culus might be appropriate. Here, of course, the port is more qualified than it might at first appear. Bayesians favor the use of the standard calculus. Secondly, and closely connected, there is the As an example of a non-Bayesian position we problem of the interpretation of probability. Here can take the view of Popper (see his 1934, and, the main division is between those who favor an for a discussion, Gillies 1998a) that the corrob- objective interpretation, such as frequency or oration of universal laws of science [C(H, E)] is propensity, and those who favor a subjective, or not a probability function, i.e. does not satisfy degree of belief, interpretation. Of course pluralist the standard axioms of probability. In symbols positions are also possible in which different inter- the claim is that C(H, E) ≠ P(H | E). pretations are considered appropriate in different As we have seen, this debate occurred also in contexts. One such pluralist position is defended the AI context. MYCIN used a nonprobabilistic in Gillies 2000, chs. 8 and 9. Once again the measure of evidential strength, and several other developments in AI just described are clearly rel- nonprobabilistic approaches were proposed and evant to this issue, and indeed would seem to developed by AI workers. (For some details, see favor the subjective interpretation of probability. Ng & Abramson 1990.) However, the develop- Thirdly there is the problem of the relation ment of AI has given a relatively unequivocal of causality and probability. Of course the prob- verdict. Probabilistic measures have proved much lem of causality is one of the oldest and most more successful in practice than nonprobabilistic central questions in . Aristotle, measures, and the latter have tended to disappear. Hume, and Kant all made fundamental contri- AI has thus supported Bayesianism in this first butions to the analysis of causality. In the last 50 sense. It should be added, however, that this years, however, the debate has taken a new turn does not give a decisive verdict against Popper’s through the emergence of a notion of indeter- ideas on corroboration. Popper was considering minate causality (see appendix), and the corres- the corroboration of hypotheses which were uni- ponding investigation of the relations between versal scientific laws. Most AI systems, however, causality and probability. Many leading philo- have as hypotheses singular statements, such as sophers of science, including Cartwright, Fetzer, “this patient’s infection is caused by streptococci” Popper, Reichenbach, Salmon, and Suppes, have or “that mountain range contains massive sul- written on this issue, and a good recent survey of fide deposits.” It is possible that Bayesianism is these philosophical developments is to be found appropriate for singular statements, while a non- in Salmon 1998. Now clearly the new theory of Bayesian approach is appropriate for universal Bayesian networks, involving as it does a novel hypotheses (see Gillies 1998a: 154–5 for argu- combination of causality and probability, is highly ments in favor of this position). relevant to this debate. These implications of AI Let us now turn to the second and rather dif- for the philosophy of science are so important ferent issue involved in the Bayesian controversy. that considerations of AI are bound to play a role It can be most easily approached by considering of increasing importance in philosophy of science the form that the debate has taken within stat- in the coming years. In this short chapter, there istics. Classical statisticians such as Neyman were is room only for a few preliminary observations, strongly opposed to Bayesianism. Yet Neyman and I will confine myself to discussing just the never used any formal system other than the first two of the above problems. standard mathematical theory of probability.

283 Donald Gillies

Neyman was clearly not an anti-Bayesian in the Lauritzen and Spiegelhalter were also work- sense we have just considered. In what sense, ing in the tradition of subjective Bayesianism, but then, was he against Bayesianism? The answer is they seem less definitely committed to this view not immediately clear, since, because he accepted than Pearl. This is what they say (1988: 159): standard probability theory, Neyman a fortiori accepted Bayes’ theorem. The answer to this con- Our interpretation of probabilities is that of a undrum is that the second issue in the Bayesian subjectivist Bayesian . . . This seems a conveni- controversy is really about the interpretation of ent and appropriate view in an area concerned probability. Neyman, following von Mises, re- with the rational structuring and manipulation garded the objective interpretation of probabil- of opinion, and the subjectivist objectives of a ity as the only valid one (cf. Gillies 1998b: 6–9). coherent system of probabilities representing This meant that some applications of Bayes’s belief in verifiable propositions, successively theorem were illegitimate in his eyes because updated on the basis of available evidence, ap- they necessitated giving a degree of belief inter- pears to fit remarkably the objectives of expert pretation to some of the probabilities used. This systems research. However, many of the tech- applied particularly to the case of giving an a niques presented here are appropriate in dis- priori distribution to a fixed, but unknown, ciplines where graphical structures are used and parameter θ. Since θ is fixed and does not vary a frequentist interpretation is more appropriate, randomly, it does not make sense to assign it an such as complex pedigree analysis in genetics. objective probability distribution, but, since it makes perfect sense for someone to have differ- So Lauritzen and Spiegelhalter think that in some ent degrees of belief in different possible values cases at least the probabilities in Bayesian net- of θ, θ can easily be given a subjective probabil- works might be given an objective interpretation. ity distribution. Many Bayesian analyses involve Neapolitan (1990) is also favorable to objective giving a priori distributions to parameters probabilities in Bayesian networks. So although such as θ, and so become illegitimate to a strict Bayesian networks were created within the tradi- objectivist such as Neyman. To sum up: the sec- tion of subjective Bayesianism, it might nonethe- ond issue involved in the Bayesian controversy is less be possible to interpret the probabilities they really the same as our second general philosoph- contain objectively. Arguably this is likely to be ical question concerning the interpretation of a good strategy in many cases. probability. A first argument in favor of an objective What have the AI developments given above interpretation is an appeal to the results, given shown as regards this controversy? It is immedi- earlier, of de Dombal’s group at Leeds. Their ately clear that they have lent support to the test showed that a diagnostic computer system subjective interpretation of probability. Pearl has performed far better when using objective prob- always argued for a subjective degree of belief abilities derived from data than when it used sub- interpretation of the probabilities in Bayesian net- jective probabilities obtained from the clinicians. works, and this remains true of his latest, highly These results are very striking, but the issue interesting, paper on the foundations of the is not simply one of a choice between different subject (Pearl 2001). In this paper he describes ways of interpreting probabilities. It should now himself as “only a half-Bayesian.” However, his be pointed out that this choice carries with it departure from standard Bayesianism arises be- methodological implications. If we are inter- cause he thinks that prior probability distributions preting the probabilities as objective, then any are inadequate to express background knowledge, proposed value of a probability must be seen as and that one needs also to use causal judgments a conjecture which could be right or wrong, and which cannot be expressed in probabilistic terms. may therefore need testing. Thus objective prob- As far as the interpretation of probability is abilities lead to a Popperian methodology of con- concerned he remains faithful to the subjective, jectures and refutations in which testing plays a degree-of-belief, view which he says he adopted central role. This is indeed the methodology of in 1971 after reading Savage. classical statistics.

284 Probability in Artificial Intelligence

Let us next contrast this with the use of subjective probabilities. Any such probability L expresses the degree of belief of an individual at a particular moment. Further evidence does not refute the claim that that individual held that degree of belief at that time. It may however LDR lead the individual to change his or her degree of belief in the light of the new evidence. In the Bayesian approach, the belief change takes place through Bayesian conditionalization or updating S M V (see appendix), i.e. through the change from a prior probability P(H) to a posterior probability Figure 21.6: Endoscopic Bayesian network P(H | E). To sum up then. The use of objective probabilities goes with the Popperian methodo- logy of statistical testing; while the use of sub- In every Bayesian network certain assumptions jective probabilities goes with the methodology of independence or conditional independence are of Bayesian conditionalization. There have been made. In this case, S, M, V must be condition- examples which show that the use of a testing ally independent of L and mutually condition- methodology can be advantageous in the con- ally independent, given LDR. Using a Popperian struction of Bayesian networks. testing methodology, these assumptions were One such example (see Sucar 1991, and Sucar, considered as conjectures which needed to be Gillies, & Gillies 1993) concerned a medical checked by statistical tests. These tests showed, instrument called an “endoscope.” This allowed a however, that the conditional independence doctor to put into the colon of a patient a small assumptions were not satisfied. In fact it turned camera which transmitted an image of the interior out that, given LDR, M and V were strongly of the colon to a television screen. In this image correlated rather than independent. an expert could recognize various things in the The response to this situation was to elimin- interior of the colon. Let us take two such things ate one of the two parameters M and V on the as examples. One is called the “lumen,” which is grounds that, since they were correlated, only one the opening of the colon. Despite its name, it could give almost as much information as both. generally appeared as a large dark region; but The results of this elimination were tested using sometimes it was smaller and surrounded by con- a random sample of more than 130 images of centric rings. Another is called a “diverticulum” the colon. It turned out that the elimination of and is a small malformation in the wall of the one of the parameters gave better results than colon, which can cause some illnesses. A divertic- those obtained using all three parameters. For ulum generally appeared as a dark region, smaller example, using all three parameters (S, M, V), than the lumen, and often circular. It was a prob- the system recognized the lumen correctly in 89 lem then to program a computer to recognize percent of cases, while, if it was modified by from the image the lumen or a diverticulum. This eliminating M, and using only (S, V) it recog- is a typical problem of computer vision. To solve nized the lumen correctly in 97 percent of cases it, an attempt was made to construct a Bayesian (for further details, see Sucar, Gillies, & Gillies network with the help of an expert in medical 1993: 206). At first sight this seems a paradox, endoscopy. because these better results were obtained using Figure 21.6 shows only a small part of this less information. The explanation is simple, how- network, but it is sufficient to illustrate the points ever. Undoubtedly there is more information in which are to be made. L stands for the lumen which all three parameters (S, M, V) than in only two causes a large dark region (LDR) to appear on the (S, V). But the greater amount of information screen. This in turn produces values for the vari- in the three parameters was used with mathem- ables S (size of the region), and M and V (mean atical assumptions of conditional independence and variance of the light intensity of the region). which were not correct. The lesser amount of

285 Donald Gillies information in the two parameters S, V was, by or ¬A. In standard or classical logic, it is contrast, used with true mathematical assump- assumed that either A is true or ¬A is true, but tions. So less information in a correct model not both. For vague predicates such as bald, this worked better than more information in a mis- “two-valued” assumption is obviously not wholly taken model. Moreover, since the modified accurate. As Jones gradually loses hair, it may be Bayesian network was simpler, the calculations difficult to say at a certain stage whether he is bald using it were carried out more quickly. So, to or not bald. Fuzzy logic attempts to deal with conclude, the modified Bayesian network was this problem by allowing us to say that Jones is more efficient, and gave better results. This shows bald to some degree, where these degrees run the value of using objective probabilities, and a from 1 (= completely hairy) to 0 (= completely Popperian methodology of statistical testing. bald). The example also shows that it is possible to Probability theory originated from the study obtain a causal network from an expert for which of games of chance, and these still afford a good the assumptions of conditional independence are illustration of some of the basic concepts of the not satisfied. Thus, although causal networks are theory. If we roll a fair die, the probability of useful heuristic guides for the construction of getting 5 is 1/6. This is written P(5) = 1/6. A Bayesian networks, they are not infallible guides. conditional probability is the probability of a This raises the third of the philosophical ques- result given that something else has happened. tions mentioned above, namely the question of For example, the probability of 5 given that the how causality is related to probability. However, result was odd, is no longer 1/6, but 1/3; while space unfortunately does not permit a discussion the probability of 5 given that the result was of this question here. even, is no longer 1/6, but 0. A conditional I conclude by briefly mentioning some fur- probability is written P(A | B). So we have ther investigations into the use of probability P(5 | odd) = 1/3, and P(5 | even) = 0. A related in AI, which raise important philosophical ques- concept is independence. Two events A and B tions. This chapter has focused on the so-called are said to be independent if the conditional symbolic approach to AI, but there is in addition probability of A given B is the same as the prob- the approach using neural networks which also ability of A, or, in symbols, if P(A | B) = P(A). makes extensive use of probability. An account Successive rolls of a die are normally assumed to of some of the problems here is to be found in be independent, that is to say, the probability Williams 2001. Another area of AI involving of getting a 5 is always the same, namely 1/6, probability is machine learning, that is to say, regardless of what results have appeared so far. the attempt to program computers to induce An important concept for probability in AI is laws from data. An account of how new results conditional independence. A and B are said to be in machine learning impinge upon longstanding conditionally independent given C, if P(A | B philosophical discussions of induction is to be & C) = P(A | C). found in Gillies 1996. Probability theory of course is not just applied to games of chance. Another typical problem is sampling from a particular population. Suppose Appendix we select a man at random from a population of men, and measure his height. If we write the In the chapter some technical terms from logic, result as X, X is said to be a random variable, probability, and causality are used. The meaning because the value of X varies randomly from of these terms is explained in what follows. The one individual to another. These values are, how- terms defined are underlined, and the text con- ever, distributed in a particular fashion. So X is tains a reference to this appendix when the term said to have a probability distribution. In the is first introduced. example of male heights, this will be the familiar Logic is concerned with propositions such as bell-shaped curve, or normal distribution. A set A = Jones is bald. The negation of a proposi- of random variables is said to have a joint prob- tion, e.g. Jones is not bald, is written as not.A, ability distribution.

286 Probability in Artificial Intelligence

Probabilities satisfy a standard set of axioms, is 2/3, this means that I believe to degree known as the Kolmogorov axioms, after the 2/3 that it will rain in London tomorrow. A mathematician Kolmogorov who first produced woman’s degree of belief can be measured by the formulation of these axioms which has come the rate at which she is prepared to bet, or her to be generally accepted by mathematicians. A betting quotient. It can be shown that, starting number of basic results follow from these axioms, from this way of measuring belief, the standard for example P(A) + P(¬A) = 1. One of the axioms of probability can be derived. most famous theorems to follow from the An application of the subjective theory of prob- axioms is Bayes Theorem. We will consider a ability to Bayesianism produces what is known particular case of this theorem dealing with a as subjective Bayesianism. Here P(H) is taken to hypothesis H, and some evidence E. Bayes represent the prior degree of belief of Mr. R, Theorem states that say, that H is true, while P(H | E) represents his posterior degree of belief in H after he has come P(H | E) = P(E | H)P(H)/P(E) to know evidence E. A rational man on this approach changes his degree of belief in the light of The components of this formula have the fol- new evidence E from P(H) to P(H | E), where lowing names: the value of P(H | E) is calculated using Bayes Theorem. These basic concepts of probability P(H) is called the prior or a priori probability of have been treated here very briefly, and a much H fuller account is to be found in Gillies 2000. P(E) is called the prior or a priori probability of Turning finally to causality, the traditional view E of causality, to be found in e.g. Kant, was that, P(E | H) is called the likelihood if A causes B, then B follows necessarily from A. P(H |E) is called the posterior or a posteriori An example would be: decapitation causes death. probability of H. In the twentieth century, however, a weaker notion of causality has developed which could The use of the formula to go from P(H) to be called indeterminate causality. A familiar P(H | E) is known as Bayesian conditionalization, example would be: smoking causes lung cancer. or Bayesian reasoning or Bayesian updating. This is held to be true even though many people Bayesianism is, roughly speaking, the view that smoke all their lives and never contract lung the problem of relating hypotheses to evidence cancer. The sense of causality here is something can be solved by Bayesian reasoning. like: smoking is an important factor in pro- There are a number of different interpreta- ducing lung cancer; or perhaps: smokers have tions of probability, and these can be classified a higher probability of contracting lung cancer as subjective and objective. An objective prob- than nonsmokers. This notion of indeterminate ability is one which is supposed to be a feature causality naturally raises the question of how of the objective world, such as mass or electrical causality is related to probability. charge. A well-known objective interpretation of probability is the frequency interpretation. For example, to say that the probability of 5 is 1/6 References on this interpretation is taken to mean that, in a long series of rolls of the die, the result 5 will Adams, J. B. 1976. “A probability model of medical appear with a frequency of approximately 1/6. reasoning and the MYCIN model.” Mathemat- Those who adopt this interpretation estimate ical Biosciences 32: 177–86. their probabilities from frequency data. Corfield, D. and Williamson, J., eds. 2001. Founda- A subjective probability, by contrast, is taken tions of Bayesianism. Dordrecht: Kluwer. to be the measure of the degree of belief de Dombal, F. T., Leaper, D. J., Staniland, J. R., of a particular individual that some event will McCann, A. P., and Horrocks, J. C. 1972. occur. For example, if I say that my subjective “Computer-aided diagnosis of acute abdominal probability that it will rain in London tomorrow pain.” British Medical Journal 2: 9–13.

287 Donald Gillies

Duda, R. O., Hart, P. E., and Nilsson, N. J. 1976. Miller, D. 1994. . Chicago and “Subjective Bayesian methods for rule-based La Salle, IL: Open Court. inference systems.” Proceedings of the National Neapolitan, R. E. 1990. Probabilistic Reasoning in Computer Conference (AFIPS) 45: 1075–82. Expert Systems: Theory and Algorithms. New York: Gaschnig, J. 1982. “Prospector: an expert system for John Wiley. mineral exploration.” In D. Michie, ed., 1982: Ng, K. and Abramson, B. 1990. “Uncertainty man- 47–64. agement in expert systems.” IEEE Expert 5: 29– Gillies, D. A. 1996. Artificial Intelligence and 48. Scientific Method. Oxford: Oxford University Press. Pearl, J. 1982. “Reverend Bayes on inference ——. 1998a. “Confirmation theory.” In D. M. engines: a distributed hierarchical approach.” Gabbay and P. Smets, eds., Handbook of Defeasible Proceedings of the National conference on AI Reasoning and Uncertainty Management Systems, (ASSI-82), pp. 133–6. vol. 1. Dordrecht: Kluwer, pp. 135–67. ——. 1985a. “How to do with probabilities what ——. 1998b. “Debates on Bayesianism and the people say you can’t.” Proceedings of the Second theory of Bayesian networks.” Theoria 64: 1–22. IEEE Conference on AI Applications, Miami, FL, ——. 2000. Philosophical Theories of Probability. pp. 6–12. London and New York: Routledge. ——. 1985b. “Bayesian networks: a model of self- Heckerman, D. 1986. “Probabilistic interpretations activated memory for evidential reasoning.” Pro- for MYCIN’s certainty factors.” In L. N. Kanal ceedings of the Cognitive Science Society. New York: and J. F. Lemmer, eds., Uncertainty in Arti- Ablex (Elsevier), pp. 329–34. ficial Intelligence. Amsterdam: North-Holland, ——. 1986. “Fusion, propagation and structuring pp. 167–96. in belief networks.” Artificial Intelligence 29: Howard, R. A. and Matheson, J. E. 1984. “Influ- 241–88. ence diagrams.” In R. A. Howard and J. E. ——. 1988. Probabilistic Reasoning in Intelligent Matheson, eds., The Principles and Applications Systems: Networks of Plausible Inference. San of Decision Analysis, vol. 2. Menlo Park, CA: Mateo, CA: Morgan Kaufmann. Strategic Decisions Group, pp. 721–62. ——. 1993. “Belief networks revisited.” Artificial Howson, C. and Urbach, P. 1989. Scientific Intelligence 59: 49–56. Reasoning: The Bayesian Approach. La Salle, IL: ——. 2001. “Bayesianism and causality, or, why Open Court. I am only a half-Bayesian.” In Corfield & Jackson, P. 1986. Introduction to Expert Systems. Williamson, eds., 2001. Wokingham, UK: Addison-Wesley. Popper, K. R. 1972 [1934]. The Logic of Scientific Kim, J. H. and Pearl, J. 1983. “A computational Discovery. London: Hutchinson. model for combined causal and diagnostic reas- Reichenbach, H. 1956. The Direction of Time. oning in inference systems.” In Proceedings of Berkeley: University of California Press. the 8th International Joint Conference on AI Salmon, W. C. 1998. Causality and Explana- (IJCAI-85), pp. 190–3. tion. New York, Oxford: Oxford University Lauritzen, S. L. and Spiegelhalter, D. J. 1988. Press. “Local computations with probabilities on graph- Shortliffe, E. H. and Buchanan, B. G. 1975. ical structures and their application to expert “A model of inexact reasoning in medicine.” systems (with discussion).” Journal of the Royal Mathematical Biosciences 23: 351–79. Statistical Society B 50: 157–224. Sucar, L. E. 1991. Probabilistic Reasoning in Leaper, D. J., Horrocks, J. C., Staniland, J. R., and Knowledge-based Vision Systems. Ph.D. disserta- de Dombal, F. T. 1972. “Computer-assisted tion, Imperial College, University of London. diagnosis of abdominal pain using ‘estimates’ ——, Gillies, D. F., and Gillies, D. A. 1993. provided by clinicians.” British Medical Journal “Objective probabilities in expert systems.” Arti- 4: 350– 4. ficial Intelligence 61: 187–208. Michie, D., ed. 1982. Introductory Readings in Williams, P. 2001. “Probabilistic learning models.” Expert Systems. New York: Gordon and Breach. In Corfield & Williamson, eds., 2001.

288 Game Theory: Nash Equilibrium

Chapter 22 Game Theory: Nash Equilibrium Cristina Bicchieri

Game theory aims to understand situations curious reader will find them interesting enough in which decision-makers interact. Chess is an to feel stimulated to go much further. example, as are firms competing for business, politicians competing for votes, jury members deciding on a verdict, animals fighting over prey, Strategic Interaction bidders competing in auctions, threats and pun- ishments in long-term relationships, and so on. What all these situations have in common is that In a strategic interaction, the outcome of an the outcome of the interaction depends on what action depends, among other things, upon the the parties jointly do. Decision-makers may be actions of other agents. Other agents have plans, people, organizations, animals, robots, or even preferences, and beliefs and, unless one is cer- genes. Game theory is a vast field, and some of tain of which action will be chosen by another its branches have seen a rapid development in agent, one will have to form beliefs about other the last few years, especially evolutionary game agents’ possible choices, and even beliefs about theory and experimental game theory, two areas the expectations that may guide another agent where some of the most exciting research is in choosing a particular action. Whereas rational being developed. For reasons of space, this chap- choice is relatively straightforward in individual ter is limited to an assessment of the fundamen- decision-making, it becomes more complicated tal concept of noncooperative game theory, that in a strategic decision context. of a Nash equilibrium. If we take Nash equilib- A game is just the abstract, formal description rium to be a predictive tool, we run into prob- of a strategic interaction. Any strategic inter- lems both at the normative and descriptive levels. action involves two or more decision-makers Many theorists have expressed misgivings about (players), each with two or more ways of acting players’ ability to infer an equilibrium from (strategies), such that the outcome depends on rationality principles alone, as well as their will- the strategy choices of all the players. Each player ingness to play equilibrium strategies in real life. has well-defined preferences among all the Evolutionary game theory and experimental game possible outcomes, enabling corresponding von theory have tried to respond to the normative Neumann–Morgenstern utilities (payoffs) to be and descriptive challenges, respectively. Though assigned.1 A game makes explicit the rules gov- in what follows I shall focus solely on the erning players’ interaction, the players’ feasible normative challenges, it is to be hoped that the strategies, and their preferences over outcomes.

289 Cristina Bicchieri

A possible representation of a game is in Nash Equilibrium normal form. A normal-form game is completely defined by three elements: a list of players i = 1, . . . , n; for each player i, a finite set of Expected utility maximization has always been a pure strategies Si; a payoff function ui that gives building-block of game theory, but for many player i’s payoff ui(s) for each n-tuple of strat- decades game theorists have paid little attention egies (s1,..., sn), where to the link between rational choice and strategic interaction, or how the outcome of strategic n → interaction can be derived from rational choices. ui: XSj R. j =1 In a well-known passage of their book, Theory of Games and Economic Behavior, von Neumann All players other than some given player i are and Morgenstern said that rational players who customarily denoted as “− i”. A player may know (i) all there is to know about the structure choose to play a pure strategy, or instead he may of the game they are playing, (ii) all there is to choose to randomize over his pure strategies; a know about the beliefs and motives of the other probability distribution over pure strategies is players, (iii) that every player is rational, (iv) that σ called a mixed strategy and is denoted by i. The every player knows (i) to (iii), (v) that every pure strategies over which a player randomizes players knows (i) to (iv), and so on, will be able are called the support of the resulting mixed strat- to infer the optimal strategy for every player. In egy. Each player’s randomization is assumed to that case, each player will behave rationally by be statistically independent of that of his oppon- maximizing his expected utility conditional on ents, and the payoffs to a mixed strategy are what he expects the others to do. the expected values of the corresponding pure The above-quoted passage is important, strategy payoffs. because it states what could be rightly called the The 2 × 2 matrix in figure 22.1 depicts a two- “central dogma” of game theory: that rational player normal-form game: each player picks a players will always jointly maximize their expected strategy independently, and the outcome, rep- utilities, or play a Nash equilibrium. Nash equi- resented in terms of players’ payoffs, is the joint librium (Nash 1951) is the standard solution product of these two strategies. The game in concept for noncooperative games. Informally, a figure 22.1 is one of complete information, in Nash equilibrium specifies players’ actions and that the players are assumed to know the rules beliefs such that (i) each player’s action is optimal of the game (which include players’ strategies) given his beliefs about other players’ choices; (ii) and other players’ payoffs. If players are allowed players’ beliefs are correct. Thus an outcome that to enter into binding agreements before the game is not a Nash equilibrium requires either that a is played, we say that the game is cooperative. player chooses a suboptimal strategy, or that some Noncooperative games instead make no allowance players “misperceive” the situation. for the existence of an enforcement mechanism More formally, a Nash equilibrium is a vector that would make the terms of the agreement σ σ of strategies ( *,1 . . . , *n), one for each of the n binding on the players. σ players in the game, such that each *i is optimal σ given (or is a best reply to) *−i. Note that 2 σ optimality is only conditional on a fixed −i, not CD σ on all possible −i. A strategy that is a best reply to a given combination of the opponents’ strat- C 3, 3 0, 4 egies may fare poorly vis-à-vis another strategy combination. 1 A common interpretation of Nash equilib- D 4, 0 1, 1 rium is that of a self-enforcing agreement. Were players to agree in pre-play negotiation to play a particular strategy combination, they would have Figure 22.1: A two-player normal-form game an incentive to stick to the agreement only in

290 Game Theory: Nash Equilibrium

7), any of his strategies (pure or mixed) would 2 be a best reply to 2’s choice and conversely, were cd 2 to know that 1 randomizes with probabilities (4/9, 5/9), any of her strategies, pure or mixed, a 9, 4 4, 9 would be a best reply. Paradoxically, if players 1 agree to play a mixed-strategy equilibrium, they have no incentive to play their part in the equi- b 6, 7 8, 3 librium. A mixed-strategy equilibrium is a self- enforcing agreement only in the weak sense that – given the other players’ equilibrium behavior – Figure 22.2: A mixed-strategy Nash equilibrium each player is indifferent between all the strategies (and lotteries over these strategies) in the support case the agreed-upon combination is a Nash equi- of her equilibrium mixed strategy. librium. In the case of a strict Nash equilibrium, There are, however, more serious questions any deviation from the equilibrium strategy nets raised by the Nash equilibrium concept. Ken a player an inferior payoff. If the equilibrium Binmore (1987, 1988) has argued that there are is not strict, however, a deviation from equilib- two possible interpretations of Nash equilibrium. rium play may earn a player the same payoff as According to the evolutive interpretation, a Nash the equilibrium strategy. In the latter case, the equilibrium is an observed regularity. Players incentive to follow the Nash equilibrium is less know the equilibrium, and test the rationality of strong. The lack of a strong incentive to play their behavior given this knowledge acquired one’s part in a Nash equilibrium is particularly from experience. The players (and the game obvious in the case of mixed-strategy equilibria. theorist) can accordingly predict that a given Consider the game in figure 22.2. equilibrium will be played, since they are accus- This game has no Nash equilibrium in pure tomed to coordinate upon that equilibrium and strategies but Nash proved that – provided cer- expect (correctly) others to do the same. Accord- tain restrictions are imposed on strategy sets and ing to the more commonly adopted eductive payoff functions – a game has at least an equilib- interpretation instead, a game is a unique event. rium in mixed strategies. Nash’s result general- In this case it makes sense to ask whether players izes von Neumann’s theorem (1928) that every can deduce what others will do from the informa- game with finitely many strategies has an equi- tion available to them. The players (and the game librium in mixed strategies. theorist) can predict that an equilibrium will be Suppose 1 plays (4/9 a, 5/9 b). Then if 2 played just in case they have enough informa- chooses c, her expected utility is 4(4/9) + 7(5/ tion to infer players’ choices. The standard as- 9) = 17/3. If 2 chooses d, she nets 9(4/9) + sumptions game theorists make about players’ 3(5/9) = 17/3. So if 1 randomizes between a rationality and knowledge should in principle be and b with probabilities (4/9, 5/9), 2 is indif- sufficient to guarantee that an equilibrium will ferent between c, d, or a lottery in which she obtain. The following assumptions are standard: chooses c with probability p and d with prob- ability (1 − p). Suppose 2 chooses (4/7 c, 3/7 CK1. The structure of the game, including d). In this case 1 nets 48/7 if he plays a, and players’ strategy sets and payoff functions, 48/7 if he plays b. Hence 1 is indifferent is common knowledge among players. between a, b, and any lottery (ap, b(1 − p)). The CK2. The players are rational (i.e., they are combination (4/9 a, 5/9 b), (4/7 c, 3/7 d) expected-utility maximizers) and this is is a mixed-strategy Nash equilibrium. common knowledge. In a mixed-strategy equilibrium, the equilib- rium strategy of each player makes the other The concept of common knowledge was intro- indifferent between the strategies on which he is duced by Lewis (1969), and later formalized by randomizing. For example, if 1 were to know Aumann (1976). Simply stated, common know- that 2 randomizes with probabilities (4/7, 3/ ledge of p among a group G means that each

291 Cristina Bicchieri

will be attained. Bernheim (1984) and Pearce 2 (1984) have argued that assuming players’ ration- abc ality (and common knowledge thereof) can only guarantee that a strategy will be rationalizable, A 2, –1 2, 1 0, 2 in the sense of being supported by internally 1 consistent beliefs about other players’ choices and beliefs. But a combination of rationalizable B 3, 0 1, 0 4, –1 strategies may not constitute a Nash equilibrium. In the game depicted in figure 22.3, all six com- binations of strategies are rationalizable, yet only Figure 22.3: Unique Nash equilibrium in pure one of them is an equilibrium. The fact that a strategies Nash equilibrium is always a combination of rationalizable strategies is of no help in predict- member of G knows p, and each knows that ing it will be played. each knows p, and so on ad infinitum. Com- Note that there are strategies that are not mon knowledge of rationality, preferences, and rationalizable, in the sense of not being sup- strategies may facilitate the task of predicting an ported by coherent beliefs. Consider the game opponent’s strategy but, as I have argued else- in figure 22.1. Suppose, again, that player 1 is where (Bicchieri 1993), it does not guarantee deciding what to do. A quick assessment of the that the resulting prediction will be correct. game will tell him that, whatever player 2 does, Consider a game that has a unique Nash equi- he is better off by choosing strategy D. And he librium in pure strategies (figure 22.3). Can the must also be able to see that player 2, if rational, players infer what other players will do from CK1 will never choose strategy C, since she, too, will and CK2? always do better by playing D. In this case, CK1 Here player 1 has two pure strategies, A and and CK2 will lead the players to accurately pre- B, and player 2 has three pure strategies, a, b, dict the outcome of the game. and c. There is a unique Nash equilibrium in In a Nash equilibrium, the optimality of a pure strategies, (B,a), but it is not evident that strategy is only conditional on a fixed strategy σ players can infer that it will be played by reason- combination −i, not on all possible combina- σ ing from CK1 and CK2. As an example of how tions −i. In the game of figure 22.1 instead, players may reach a conclusion on how to play, the Nash equilibrium strategies (D, D) are also consider the following argument by player 1. “If optimal with respect to any strategy choice of player 2 believes that I will play A, then it is the opponent. Whatever player 1 does, player 2 optimal for her to pick c. And why would she is better off by choosing D, and the same is true think I play A? Well, she must believe that I of player 1. We say that a strategy si is strictly expect her to play b, to which A is a best reply. dominated by another strategy ti if, for every And why would I expect her to play b? I would choice of strategies of the other players, i’s pay-

(she will think), if I were to believe she expects off from choosing ti is strictly greater than his me to play B . . .” It is easy to verify that such a payoff from choosing si. In our example, C is chain of reasoning can justify the choice of any strictly dominated by D for both players. A strictly strategy for both players. dominated strategy is never rationalizable, since The concept of Nash equilibrium embodies a the belief that a player plays it is inconsistent notion of individual rationality, since each play- with common knowledge of rationality. We say er’s equilibrium strategy is a best reply to the that si is weakly dominated by ti if, for every choice opponents’ strategies, but unfortunately it does of strategies of the other players, i’s payoff from not specify how players come to form the beliefs choosing ti is at least as great as i’s payoff from about each other’s strategies that support equi- choosing si. Note that weakly dominated strat- librium play. Beliefs, that is, can be internally egies are rationalizable, since there always exists consistent but fail to achieve the interpersonal an opponents’ strategy combination to which a consistency that guarantees that an equilibrium player’s weakly dominated strategy is a best reply.

292 Game Theory: Nash Equilibrium

question is not whether players are fully rational 2 or believe each other to be. Rather, we want to LR know how far they can go in inferring a Nash solution from CK1 and CK2. As we have seen, U 8, 10 –100, 9 in most cases the answer is “not far.” 1 Predictability is hampered by another common problem encountered in game theory: Multiple D 7, 6 6, 5 Nash equilibria. Suppose two players have to divide $100 among themselves. They must re- strict their proposals to integers, and each has to Figure 22.4: Iterated dominance independently propose a way to split. If the total proposed by both is equal to or less than $100, When a game has a unique Nash equilibrium, each gets what she proposed, otherwise they get we can predict that it will be played if we are nothing. This game has 101 Nash equilibria. Is able to show that players, armed with common there a way to predict which one will be chosen? knowledge of rationality and of the structure of Alternatively, is there a way a player can infer the game, will infer the Nash solution. If players what the other will do, and thus adjust her pro- have dominated strategies, CK2 entails that they posal accordingly? In real life, many people would will eliminate them, and this is common know- go for the 50/50 split. It is simple, and seems ledge (we assume that the consequences of CK1 equitable. In Schelling’s words, it is a focal point and CK2 are common knowledge, too). Often (Schelling 1960). A focal point equilibrium has after we have eliminated strictly dominated strat- some property that makes it salient: A solution egies for one player, we may find that there are may be salient because of historical precedent, now strictly dominated strategies for another or because it embodies cultural norms we share player, which will be eliminated as well. This (Lewis 1969). Unfortunately, mere salience is process of successive elimination can continue not enough to provide a player with a reason for until there are no more strictly dominated strat- choice. In our example, only if it is common egies left. If a unique strategy remains for each knowledge that the 50/50 split is the salient player, we say the game has been solved by iter- outcome it becomes rational to propose $50. ated dominance. It is easy to prove that a strat- Game theory, however, filters out any social or egy profile thus obtained is a Nash equilibrium cultural information regarding strategies, leav- (Bicchieri 1993). ing players with the task of coordinating their Consider for example the game in figure 22.4. actions on the sole basis of common knowledge R is a strictly dominated strategy for player 2, of rationality (and of the structure of the game). and since rationality is common knowledge, 2 is Consider now another game that many readers expected to eliminate R as a possible choice. would intuitively know how to solve: figure 22.5. Player 1 will now expect L to be played, in which The game of figure 22.5 has two Nash equilibria case U dominates D. (U, L) is the solution to in pure strategies: (a, c) and (b, d), but in the the game, and it is inferrable from CK1 and CK2. Note that assuming common knowledge of rationality (or at least some level of mutual 2 knowledge of rationality) is crucial to obtaining cd the (U, L) solution. If there were some doubt about a player’s rationality, the solution would a 2, 2 1, 1 unravel. For example, if 1 were to think there is 1 a 0.01 chance that R is chosen, then he would be better off by choosing D. In real life this is b 1, 1 1, 1 likely to occur. That is, in real life a player may “play safe” and prudently choose D, but we are now discussing a completely different point. The Figure 22.5: Focal-point equilibrium

293 Cristina Bicchieri

(b, d) equilibrium each player plays a weakly predictions using rationality arguments alone. In dominated strategy. (a, c) is a Pareto-dominant most cases, however, the set of possible outcomes equilibrium point, since it gives both players a is still too large. higher payoff than any other equilibrium in the A very different approach to the problem of game. For this very reason, it should be a nat- indeterminacy is to start by considering the set ural focal point for both players. Should we of Nash equilibria, and ask whether some of them confidently predict that (a, c) will be the solu- should be eliminated because they are in some tion of the game? We have seen that focal points sense “unreasonable.” This is the approach taken must be common knowledge among the players by the refinement program (Kohlberg 1990, van before it becomes rational for them to play Damme 1991). the focal-point equilibrium. If no such common knowledge is present, rationality alone is not a reliable guide. Could elimination of weakly dom- Normal-form Refinements inated strategies do the trick? We know that when a player has a strictly dominated strategy, ration- ality dictates eliminating it, hence predicting Consider again the game in figure 22.5. How behavior is a (relatively) simple matter. The case reasonable is the equilibrium (b, d)? Under what of weakly dominated strategies is not that circumstances would players agree to play it, and straightforward. For one, a weakly dominated then stand by the agreement? The equilibrium strategy is still a best reply to some opponent’s strategies (b, d) are weakly dominated but – as strategy. Putting it differently, weak dominance I have already argued – common knowledge of means that there is at least one choice on the rationality does not force players to eliminate part of an opponent that makes one indifferent them. Prudence, however, may suggest that one between the weakly dominated strategy and some should never be too sure of the opponents’ other strategy. In our example, were player 2 to choices. Even if players have agreed to play a believe that 1 plays b, she would be indifferent given equilibrium, some uncertainty remains. between c and d, since both c and d are best If so, we should try to model this uncertainty replies to b; and conversely, were player 1 to in the game. Selten’s insight was to treat perfect expect 2 to play d, he would be indifferent rationality as a limit case (Selten 1965). His between a and b, since both strategies are best “trembling hand” metaphor presupposes that replies to d. deciding and acting are two separate processes, One possible solution is to introduce a rule in that even if one decides to take a particular according to which also weakly dominated strat- action, one may end up doing something else by egies should be eliminated by a rational player. mistake. An equilibrium strategy should be opti- Eliminating weakly dominated strategies is an mal not only against the opponents’ strategies, example of an “eductive” procedure. When ask- but also against some very small probability ε> ing how the players’ deductive processes might 0 that the opponents make “mistakes.” Such an unfold, one must specify some basic principles equilibrium is trembling-hand perfect. Is the equi- of rationality, and then examine which choices librium (b, d) perfect? If so, b must be optimal are consistent with common knowledge of the against c being played with probability ε and d specified principles. Such choices may or may being played with probability 1 −ε for some not result in an equilibrium, but at least the link small ε> 0. But in this case the payoff to a is 2ε, between rational choice and equilibrium (when whereas the payoff to b is ε. Hence for all ε> 0, there is such link) is made clear. The advantage a is a better strategy choice. The equilibrium (b, of this approach is that it is possible to refine d) is not perfect, but (a, c) is. A prudent player our predictions about how players might choose therefore would discard (b, d). In this simple without assuming that they will coordinate on a game, checking perfection is easy, since only one particular equilibrium. Principles such as iterated mistake is possible. With many strategies, there strict dominance and rationalizability are ex- are many more possible mistakes to take into amples of how it is possible to restrict the set of account. Similarly, with many players we may

294 Game Theory: Nash Equilibrium need to worry about who is likely to make a Whereas normal-form games are represented mistake. by matrices, extensive-form games are represented Note that the starting-point of this approach by trees. A matrix description shows the out- is the set of Nash equilibria of the game. It is comes, represented in terms of players’ payoffs, assumed that players can calculate them, and for every possible combination of strategies the agree to play one. The goal now is to rule out players might choose. A tree representation is all those Nash equilibria that are not reasonable sequential, because it shows the order in which agreements. In principle, an equilibrium that is actions are taken by the players. It is quite nat- reasonable under a given criterion of reasonable- ural to think of sequential-move games as being ness might cease to be such under another, more ones in which players choose their strategies one restrictive criterion. Specifying why an equilib- after the other, and of simultaneous-move games rium might be unacceptable is made easier by as ones in which players choose their strategies taking into account what would happen if one or at the same time. What is important, however, is more players were to “deviate” from the agreed- not the temporal order of events per se, but upon solution. The reason is intuitive: a player whether players know about other players’ ac- should not agree to play his part in an equilib- tions when they have to choose their own. In the rium if – were the unexpected to happen – he normal-form representation, players’ information would have been better off by playing another about other players’ choices is not represented. strategy. Therefore we may say that a crucial This is the reason why a normal-form game could property required of an equilibrium is that it is represent any one of several extensive-form stable to players’ deviating from it. To reason games. When the order of play is irrelevant to a about “deviations” from equilibrium it is helpful game’s outcome, then restricting oneself to the to have a richer description of the game. This is normal form is justifiable. When the order of the reason why most of the refinement literature play is relevant, however, the extensive form must refers to games in extensive form, where the be specified. order in which players move and the informa- In an extensive-form game, the information a tion they have when making a choice are made player has when she is choosing an action is explicit. explicitly represented using information sets, which partition the nodes of the tree. If an information set contains more than one node, the player who has to make a choice at that information set will Games in Extensive Form be uncertain as to which node she is at. Not knowing at which node one is means that the The extensive form of a game specifies the fol- player does not know which action was chosen lowing information: a finite set of players i = 1, by the preceding player. If a game contains in- ...n, one of which might be nature (N); the formation sets that are not singletons, the game order of moves; the players’ choices at each move is one of imperfect information. It may also be and what each player knows when she has to the case that a player does not remember what choose; the players’ payoffs as a function of their she previously did. In this case, the game is one moves; finally, moves by nature correspond to of imperfect recall. All the games we consider probability distributions over exogenous events. here, however, will be ones of perfect recall, in The order of play is represented by a game tree that players will be assumed to remember what T, which is a finite set of partially ordered nodes they did and knew previously. t ∈ T that satisfy a precedence relation denoted A strategy for player i is a complete plan of by “ ”. A subgame is a collection of branches action that specifies an action at every node at of a game such that they start from the same which it is i’s turn to move. Note that a strategy node and the branches and the node together specifies actions even at nodes that will never be form a game tree by itself. In figure 22.6a, for reached if that strategy is played. Consider the example, player 2’s decision node as well as her game in figure 22.6a. It is a finite game of per- moves form a subgame of the original game. fect information in which player 1 moves first. If

295 Cristina Bicchieri

Player 2 da

Dx 1, 0 1, 0

Dy 1, 0 1, 0

112 Player 1 Aa y 2 Ax 0, 2 4, 1 3 Ddx

Ay 0, 2 2, 3 1 0 4 0 2 1 (a) (b)

Figure 22.6a: Extensive-form game Figure 22.6b: Normal-form game he chooses D at his first node, the game ends Representing the sequential version of the and player 1 nets a payoff of 1, whereas player 2 game as one of perfect information (figure 22.6a) gets 0. But choosing D at the first node is only helps to solve it. Suppose player 1 were to reach part of a strategy for player 1. For example, it his last node. Since he is by assumption rational, can be part of a strategy that recommends “play he will choose x, which guarantees him a payoff D at your first node, and x at your last node”. of 4. Knowing (by assumption) that 1 is rational, Another strategy may instead recommend play- player 2 – if she were to reach her decision node ing D at his first node, and y at his last decision – would play d, since by playing a she would net node. Though it may seem surprising that a strat- a lower payoff. Finally, since (by assumption) egy specifies actions even at nodes that will not player 1 knows that 2 is rational and that she be reached if that strategy is played, we must knows that 1 is rational, he will choose D at his remember that a strategy is a full contingent plan first decision node. The equilibrium (Dy, d) of action. For example, the strategy Dx recom- should therefore be ruled out, since it recom- mends playing D at the first node, thus effect- mends an irrational move at the last node. In ively ending the game. It is important, however, the normal form, both equilibria survive. The to be able to have a plan of action in case D is reason is simple: Nash equilibrium does not con- not played. Player 1 may, after all, make a mistake strain behavior out of equilibrium. In our exam- and, because of 2’s response, find himself called ple, if 1 plans to choose D and 2 plans to choose to play at his very last node. In that case, having d, it does not matter what player 1 would do at a plan helps. Note that a strategy cannot be his last node, since that node will never be changed during the course of the game. Though reached. a player may conjecture about several scenarios The sequential procedure we have used to of moves and countermoves before playing the conclude that only (Dx, d) is a reasonable solu- game, at the end of deliberation a strategy must tion is known as backward induction (Zermelo be chosen and followed through the game. 1913). In finite games of perfect information The game in figure 22.6a has two Nash with no ties in payoffs, backward induction always equilibria in pure strategies: (Dx, d) and (Dy, identifies a unique equilibrium. The premise of d). This is easy to verify by looking at figure the backward-induction argument is that mutual 22.6b, the normal-form representation of the rationality and the structure of the game are com- game. Is there a way to solve the indeterminacy? mon knowledge among the players (CK1 and

296 Game Theory: Nash Equilibrium

CK2). It has been argued by Binmore (1987), 2’s choice at an off-equilibrium node. His con- Bicchieri (1989, 1993), and Reny (1992) that clusion about 2’s choice will affect his own under certain conditions common knowledge of choice. But player 2’s choice will depend upon rationality leads to inconsistencies. For example, how she interprets 1’s off-equilibrium move. if player 2 were to reach her decision node, would Player 1, in turn, must be able to anticipate 2’s she keep thinking that player 1 is rational? How interpretation of his deviating from the equilib- would she explain 1’s move? If 1’s move is in- rium path. For example, if 2 were to interpret consistent with CK2, player 2 will be unable to the deviation as a mistake, would she still play predict future play; as a corollary, what constitutes her part in the equilibrium (Dy, d), and choose an optimal choice at her node remains undefined. d? If she expects y to be played at the last node, As a consequence of the above criticisms, the a is a best reply. But, at his last node, why would usual premises of backward-induction arguments rational player 1 choose y? Is the agreement to have come to be questioned (Pettit & Sugden play (Dy, d) reasonable? 1989; Basu 1990; Bonanno 1991). The earliest refinement proposed to rule out implausible equilibria in extensive games of perfect information is subgame perfection (Selten 1965). A Nash equilibrium is subgame perfect Extensive-form Refinements if its component strategies – when restricted to any subgame – remain a Nash equilibrium of The goal of the refinement program, however, the subgame. The equilibrium (Dy, d) is not has not been the formalization of players’ subgame perfect: in the subgame starting at the reasoning. The arguments proposed have been last node, y is a dominated strategy. Note that informal, their purpose being the elimination the backward induction equilibrium is always of implausible equilibria. In the normal form, subgame perfect. Subgame perfection, however, Selten’s trembling-hand perfection requires only applies (nontrivially) to games that have players to check how a strategy will perform were proper subgames. Any Nash equilibrium of a another player to take an action that has zero game without proper subgames is trivially sub- probability in equilibrium. In the extensive-form game perfect (since the whole game can be con- representation, players ask what would happen sidered a subgame), but in this case the criterion off-equilibrium, at points in the game tree that does not help in resolving the indeterminacy. In will never be reached if the equilibrium is played. figure 22.7, for example, both (c, L) and (a, R) In both cases, the starting-point is an equilibrium, are (trivially) subgame-perfect equilibria. which is checked for stability against possible The game of figure 22.7 is a case in which we deviations. would still like to eliminate equilibria that require By its nature, the Nash equilibrium concept players to behave suboptimally in parts of the does not restrict action choices off the equilib- game that are reached with zero probability if a rium path, because those choices do not affect given equilibrium is played, but cannot be con- the payoff of the player who moves there. For sidered subgames. Kreps and Wilson’s (1982b) example, the equilibrium (Dy, d) in the game of sequential equilibrium is an answer to this prob- figure 22.6 lets player 1 make an irrational choice lem. A sequential equilibrium is a combination at the last node, since that choice is not going of strategies and beliefs such that each player has to affect his payoff (which is determined by his a belief (a probability assessment) over the nodes choosing D at the beginning of the game). How- at each of his information sets. At any informa- ever, the strategy of a player at an off-equilibrium tion set x where i has to play – given player i’s information set can affect what other players beliefs at x and the equilibrium strategies of the choose in equilibrium. Suppose the players con- other players – i’s strategy for the rest of the sider agreeing to play (Dy, d). In order to choose game must still maximize his expected payoff. D, player 1 must decide what would happen As players move through the game tree, they were he to play A instead. To decide whether D rationally update their beliefs using Bayes’s rule is a rational move, 1 has to think about player (for more on this, see Chapter 21, PROBABILITY

297 Cristina Bicchieri

probability (1 − p − q) very close to one, but can play b by mistake with higher probability (p = X 2/100) than playing a by mistake (q = 1/100). If this is what 2 believes, she should choose c L. Given her beliefs, L has an expected utility a of 0.04, whereas R has an expected utility b of 0.01. Both equilibria therefore survive some 3 perturbations. 3 Note, however, that strategy b is strictly dom- inated by c for player 1, therefore it is highly y 2 unlikely that 1 would play b. This example is meant to stress the importance of restricting off- LR LRequilibrium beliefs. Such beliefs, too, should be rationally justified. The equilibrium (c, L) must be eliminated because it is supported by the belief that a dominated strategy will be played 2 0 0 5 2 0 0 1 off-equilibrium, and this belief is inconsistent with common knowledge of rationality. Figure 22.7: Subgame-perfect, sequential, and The need to add a condition of plausibility perfect equilibria for off-equilibrium beliefs motivated the forward induction refinement (Kohlberg & Mertens IN ARTIFICIAL INTELLIGENCE). The problem with 1986). Off-equilibrium beliefs, for example, the notion of sequential equilibrium is that – should be consistent with common knowledge provided beliefs are revised according to Bayes’s of rationality and any inference one may draw rule – no further restriction is imposed upon from it. A deviation from equilibrium should them. The consequence is that far too many therefore be interpreted, whenever possible, as a Nash equilibria are still considered admissible or rational move. In our example, player 1’s devi- “reasonable.” At player 2’s information set, if ation from the equilibrium (c, L) should not be for some reason she assigns a higher probability interpreted as a mistake, but rather as a signal to node y than y′, then her optimal choice is L. that he intends to play a (and get a higher pay- If instead she judges y and y′ to be equiprobable, off). In this case player 2 would respond with R. she will choose R. Thus both (c, L) and (a, R) The conclusion is that equilibrium (c, L) is not survive as sequential equilibria, since each is sup- robust to deviations, and should be eliminated. ported by some acceptable belief. As I mentioned at the outset, the refinement Another common refinement is Selten’s perfect program attempts to establish stability criteria equilibrium (Selten 1975). In this case players for Nash equilibrium. It presupposes that players are explicitly assumed to interpret deviations from will choose to play a Nash equilibrium after equilibrium play as “mistakes,” and respond ac- having eliminated several alternative equilibria cordingly. A perfect equilibrium must be robust on the ground that they are unreasonable. What to small perturbations of players’ equilibrium counts as a reasonable equilibrium, however, strategies. Selten’s notion, however – by not depends upon how off-equilibrium behavior is imposing restrictions upon players’ beliefs – lays interpreted. This, in turn, hinges on players out- itself open to the same criticism addressed to of-equilibrium beliefs. So far we have developed Kreps and Wilson’s refinement. If there are sev- no comprehensive theory of out-of-equilibrium eral possible “mistakes” a player can make, and behavior that indicates, for example, when a devi- beliefs are unrestricted, some equilibria cannot ation should be interpreted as a signal and when be ruled out simply because they are supported as a mistake. Such theory would supply substant- by beliefs that make some mistakes more likely ive (as opposed to merely formal) rationality than others. Suppose that in our example player criteria for players’ beliefs, and would thus expand 2 believes that player 1 intends to play c with the traditional notion of practical rationality to

298 Game Theory: Nash Equilibrium include an epistemic component. This theoret- HD ical inadequacy undermines the eductive goal of inferring (and predicting) equilibrium play from B – C B – C H , B, 0 rationality principles alone. 2 2

B B D 0, B , Selection by Evolution 2 2

Figure 22.8: Hawk and dove game A Nash equilibrium need not be interpreted as a unique event. If we think of it as an observed regularity, we want to know by what process Suppose injury has a payoff in terms of loss of such equilibrium is reached and what accounts fitness equal to C, and victory corresponds to a for its stability. When multiple equilibria are pos- gain in fitness B. If hawk meets hawk, or dove sible, we want to know why players converged meets dove, each has a 50 percent chance of to one in particular and then stayed there. An victory. If a dove meets another dove, the winner alternative way of dealing with multiple equilibria gets B and the loser gets nothing, so the average is to suppose that the selection process is made increase in fitness for a dove meeting another by nature. dove is B/2. A dove meeting a hawk retreats, Evolutionary theories are inspired by popula- so her fitness is unchanged, whereas the hawk tion biology (e.g. Maynard Smith & Price 1973). gets a gain in fitness B. If a hawk meets another These theories dispense with the notion of the hawk, they escalate until one wins. The winner decision-maker, as well as with best responses/ has a fitness gain B, the loser a fitness loss C. So optimization, and use in their place a natural the average increase in fitness is (B − C)/2. The selection, “survival-of-the-fittest” process together latter payoff is negative, since we assume the with mutations to model the frequencies with cost of injury is greater than the gain in fitness which various strategies are represented in the obtained by winning the contest. We assume population over time. In a typical evolutionary that players will be randomly paired in repeated model, players are preprogrammed for certain encounters, and in each encounter they will play strategies, and are randomly matched with other the stage game of figure 22.8. players in pair-wise repeated encounters. The If the population were to consist predomin- relative frequency of a strategy in a population is antly of hawks, selection would favor the few simply the proportion of players in that popula- doves, since hawks would meet mostly hawks tion who adopt it. The theory focuses on how and end up fighting with an average loss in fitness the strategy profiles of populations of such agents of (B − C)/2, and 0 > (B − C/2). In a popula- evolve over time, given that the outcomes of tion dominated by doves, hawks would spread, current games determine the frequency of dif- since every time they meet a dove (which would ferent strategies in the future. be most of the time) they would have a fitness As an example, consider the symmetric game gain of B, whereas doves on average would only in figure 22.8 and suppose that there are only two get B/2. possible behavioral types: “hawk” and “dove.”2 Maynard Smith interpreted evolutionary games A hawk always fights and escalates contests as something that goes on at the phenotypic until it wins or is badly hurt. A dove sticks to level. The fitness of a phenotype depends on its displays and retreats if the opponent escalates frequency in the population. A strategy is a pheno- the conflict; if it fights with another dove, they type, and a player is just an instance of such a will settle the contest after a long time. Payoffs behavioral phenotype. In our example, we have are expected changes in fitness due to the out- only two behavioral phenotypes: “hawk” and come of the game. Fitness here means just “dove.” Evolutionary game theory wants to know reproductive success (e.g., the expected number how strategies do on average when games are of offspring per time unit). played repeatedly between individuals who are

299 Cristina Bicchieri randomly drawn from a large population. The opposite would occur in a population composed average payoff to a strategy depends on the com- mainly of hawks. So for example if “hawks” domin- position of the population, so a strategy may do ate the population, a mutant displaying “dove” very well (in terms of fitness) in an environment behavior can invade the population, since indi- and poorly in another. If the frequency of hawks viduals bearing the “dove” trait will do better in the population is q and that of doves corres- than hawks. The main solution concept used in pondingly (1 − q), the average increase in fitness evolutionary game theory is the evolutionarily for the hawks will be q(B − C)/2 + (1 − q)B, stable strategy (ESS) introduced by Maynard and (1 − q)B/2 for the doves. The average pay- Smith and Price (1973). A strategy or behavioral off of a strategy in a given environment deter- trait is evolutionarily stable if, once it dominates mines its future frequency in the population. in the population, it does strictly better than any Strategies that, on average, earn high payoffs in mutant strategy, hence it cannot be invaded. To the current environment are assumed to increase formalize this concept, let me first make a brief in frequency, and strategies that, on average, earn digression. In a symmetric game like hawk/dove, lower payoffs are assumed to decrease in fre- we have a finite set of pure strategies S and a quency. If the average payoffs of the different corresponding set ∆ of mixed strategies. A popu- strategies are the same, then the composition of lation state is equivalent to a mixed-strategy the population is stable. In our example, the x ∈∆. Note that the evolutionary model gives a average increase in fitness for the hawks will be natural interpretation to mixed strategies as the equal to that for the doves when the frequency proportions of certain strategies (or traits) in a of hawks in the population is q = B/C. At that population. A state in which each individual plays frequency, the proportion of hawks and doves is a pure strategy and the proportion of different stable. If the frequency of hawks is less that B/ strategies correspond to x is called a polymorphic C, then they do better than doves, and will con- state. Alternatively, we may interpret the popu- sequently spread; if their frequency is larger than lation state x as monomorphic, in the sense that B/C, they will do worse than doves and will each player plays the mixed-strategy x. In a two- shrink. player game, being matched against a randomly Note that if C > B, then (B − C)/2 < 0, so drawn individual in population state x is equi- the game in figure 22.8 has two pure-strategy valent to being matched against an individual Nash equilibria: (H, D) and (D, H). There is who plays the mixed strategy x. Hence the aver- also a mixed-strategy equilibrium in which Hawk age payoff of playing strategy y in population is played with probability q = B/C and Dove is state x is equal to the expected payoff to y when played with probability (1 − q) = C − B/C. If played against the mixed strategy x, i.e. u(y, x). the game of figure 22.8 were played by rational The population average in this case is equal to agents who chose which behavior to display, we the expected payoff of the mixed strategy x when would be at a loss in predicting their choices. matched against itself, i.e. u(x, x). From CK1 and CK2 the players cannot infer that In a symmetric, two-player game, x is an ESS a particular equilibrium will be played; moreover, if and only if, for all y ∈∆ such that y ≠ x, since there are no dominated strategies, all pos- sible outcomes are rationalizable. In the hawk/ (1) u(x, x) > u(y, x) dove example, however, players are not rational and do not choose their strategies. So if an equi- or librium is attained it must be the outcome of some process very different from rational delib- (2) u(x, x) = u(y, x), and u(x, y) > u(y, y). eration. The process at work is natural selection: high-performing strategies increase in frequency Condition (1) tells us that strategy x is a unique whereas low-performing strategies’ frequency best reply against itself. If the bulk of the popu- diminishes and eventually goes to zero. lation consists of type x and a small number of We have seen that in a population composed mutants of type y enters the population, if x mostly of doves, hawks will thrive, and the does better against x than y does against x, y will

300 Game Theory: Nash Equilibrium be less fit and disappear. However, if x is a mixed In the hawk/dove example, we have assumed strategy, we know (1) does not hold. In this that the success of a strategy depends on the case, for x to be an ESS, (2) must hold. If both outcome of pairwise random matches. It is often x and y perform equally well against x, then y the case that a strategy’s success depends not will be less fit than x if x does better against y on the strategy played by a particular opponent, that y does against y. but on the population-wide frequencies of strat- In the hawk/dove game, neither of the two egies. When examining behavior in a population pure behavioral types is evolutionarily stable, since game, we adopt the concept of an evolutionarily each can be invaded by the other. We know, stable state (also ESS) (Hofbauer & Sigmund however, that a population in which there is a 1998). proportion q = B/C of hawks and (1 − q) = C − Suppose the game has N pure strategies, with B/C of doves is stable. This means that the type an N × N symmetric expected payoff matrix = of behavior that consists in escalating fights with A (aij). There is an infinite number of players, probability q = B/C cannot be invaded by any and each player initially commits to playing other type, hence it is an ESS. To show that the exactly one of the N pure strategies. Let p be the mixed strategy x = (B/C, C − B/C) is an ESS, N × 1vector denoting the population-wide pro- we have to show that condition (2) is satisfied. portion of each of the N strategies (player types) Indeed, u(x, y) − u(y, y) = 1/2C(B − Cq)2 is in the population at a given time. Let greater than zero for all q ≠ B/C. An ESS is a strategy that, when it dominates == fpii() ∑ apjji Ap the population, is a best reply against itself. There- j fore an evolutionarily stable strategy such as (B/ C, C − B/C) is a Nash equilibrium. Though every denote the fitness of strategy i and let ESS is a Nash equilibrium, the reverse does not = hold; in our stage game, there are three Nash ∑ fpi () Ap equilibria, but only the mixed-strategy equilib- i rium (B/C, C − B/C) is an ESS. However, when a strategy is a unique best reply to itself, it is denote the population-wide payoff. The both an ESS and a strict Nash equilibrium. In population-wide weighted average fitness value T P this special case the reverse also holds: every strict is p Ap. We say that is an evolutionarily stable ≠ P P Nash equilibrium is an ESS. In a strict equilib- state if for any p in the neighborhood of , rium, there exists no other strategy which is an we have: alternative best reply to the equilibrium strategy, and this guarantees non-invadability. As an P T Ap > p T Ap example, consider the game in figure 22.5. There are two pure-strategy Nash equilibria, (a, c) and This captures the idea that the population-wide (b, d), but only (a, c) is strict. It is easy to verify payoff under P is higher (locally) than for any that (a, c) consists of ESS satisfying condition other vector p.3 (1). The definitions of evolutionarily stable strat- The prior examples show how evolution egies or states are static. To describe the dynamic can at least partially solve the problem of equilib- process that leads to a certain distribution of rium selection without imposing heroic cognit- strategies in a population, we have models of ive requirements on players. An ESS is, in fact, the selection dynamics that express the growth not just a Nash equilibrium but also a perfect rate of a strategy i in population state p as a and proper equilibrium (van Damme 1987). function of i’s average payoff in p relative to the Furthermore, an evolutionary account of how a average payoff to other strategies in p. ESS do Nash equilibrium is achieved provides an explana- not refer to a specific dynamic, but biologists tion of the dynamics of the selection process, and evolutionary game theorists frequently use something which the refinement program can- deterministic replicator dynamics (Taylor & not do. Jonker 1978) of the form:

301 Cristina Bicchieri

ptApt() () When we think of strategies, however, we usu- (*) pt( += 1) ii, i pT () t Ap () t ally refer to behaviors that are not genetically inherited. In economic and political applications where p(t) denotes the population-wide propor- of game theory, actors can be firms, political tions at time t, the denominator is a measure of parties, nations. Even when actors are indivi- average strategy fitness in the population at t, duals, their strategies have a strong cultural com- and the numerator measures the fitness of strat- ponent. Evolutionary models can still be applied egy i at time t. Strategies with above-average to explain how Nash equilibria are attained and fitness see their proportions increase, and those whether they are stable, but selection mechanisms with below average fitness see their proportions in this case work through processes of cultural decrease. Note that (*) is a deterministic system transmission such as learning and imitation. which allows some strategies to become extinct, Learning and imitation are subject to mistakes, = in the sense that pi(t) 0 for some i, t. To and new strategies may enter the population prevent extinction, mutations are added, but either by random mistake or by purposeful a discussion of how to modify (*) to include innovation. Payoffs in this case cannot represent mutations and how to interpret the latter would fitness changes, but if we give them a utility take us too far from the present topic. For an interpretation, we must provide for interpersonal analysis of stochastic models, see Foster and comparisons of utilities. Indeed, to imitate a more Young (1990). successful individual, one must be able to com- ESS are asymptotically stable fixed points of pare one’s payoffs with the payoffs of others, this replicator dynamic, though the converse need but traditional von Neumann–Morgenstern util- not be true (see e.g. Samuelson 1997). A similar ities do not allow for such comparisons. relationship holds between the replicator dynamic Evolutionary games provide us with a way of and Nash equilibria: if P is a Nash equilibrium of explaining how agents that may or may not be the symmetric N × N game with expected pay- rational and – if so – subject to severe information off matrix A, then P is a stationary state of the and calculation restrictions, achieve and sustain replicator dynamic. a Nash equilibrium. When there exist evolution- In evolutionary theory replication, variation arily stable strategies (or states), we know which and heredity are the basic assumptions. Any en- equilibrium will obtain, without the need to tity capable of replicating itself with differential postulate refinements in the way players inter- success will be subject to an evolutionary process. pret off-equilibrium moves. Yet we need to know Differential success, in turn, is related to heredit- much more about processes of cultural transmis- ary variations. In biology, replicators are genes sion, and to develop adequate ways to represent and in genetic evolution, variation is provided payoffs, so that the promise of evolutionary games by random mutations and recombinations of gene is actually fulfilled. sequences. Behavioral patterns can be replicators, too, in the sense that behavioral trait x is replic- ated when a gene x that predisposes its carriers Notes to behave according to this pattern replicates itself. This means that bearers of gene x will 1 In a game, a player’s action may have one of behave in ways that make them reproductively several possible consequences, depending on the successful, so that in the next generation there other players’ choices. It is usually assumed that will be more copies of x. To the extent that the probabilities with which the consequences behavior x promotes the replication of its pre- occur are objective and known to the decision- disposing gene, we are correct in saying that the maker. Suppose action a has two possible behavior is replicating itself. Individuals are just consequences, x and x′, which occur with prob- bearers of such genetic material, hence they are ability p and (1 − p), respectively. Choosing born with fixed behavioral traits. Variation of action a is like choosing a lottery that gives competing strategies is provided by random prize x with probability p, and prize x′ with mutations and recombinations of gene sequences. probability (1 − p). Agents are assumed to have

302 Game Theory: Nash Equilibrium

preferences over such lotteries. If preferences **Foster, D. and Young, P. 1990. “Stochastic are complete, transitive, and satisfy a number of evolutionary game dynamics.” Theor. Pop. Biol. other conditions (von Neumann & Morgenstern 38: 219–32. 1944), they can be represented by the expecta- *Harsanyi, J. and Selten, R. 1987. A General Theory tion of a real-valued utility function U: C → R of Equilibrium Selection in Games. Cambridge, (unique up to a positive linear transformation) MA: MIT Press. such that, for any two lotteries a and b, a b **Hofbauer, J. and Sigmund, K. 1998. Evolution- Σ >Σ iff x∈a p(x)U(x) y∈bp(y)U(y). A rational agent ary Games and Population Dynamics. Cambridge: will choose an action (lottery) a* that maxim- Cambridge University Press. izes the expected value of a von Neumann– *Kohlberg, E. 1990. “Refinement of Nash equilib- Morgenstern utility function. rium: the main ideas.” In T. Ichiishi, A. Neyman,

2 A two-player game is symmetric if (a) S1 and S2 and Y. Tauman, eds., Game Theory and Applica- have the same, finite number of elements, and tions. San Diego: Academic Press. (b) the payoff matrix is symmetric, i.e. for all i **—— and Mertens, J.-F. 1986. “On the strategic ∈ = and j S, u1(i, j) u2(j, i). stability of equilibria.” Econometrica 54: 1003– 3 By contrast, P is a symmetric Nash equilibrium 37. if P T AP ≥ pT AP for all feasible p. **Kreps, D. and Wilson, R. 1982. “Sequential Equilibria.” Econometrica 50: 863–94. *Lewis, D. 1969. Convention. Cambridge: Cam- References bridge University Press. *Maynard Smith, J. 1982. Evolution and the Theory (Note: references marked with one asterisk are of Games. Cambridge: Cambridge University readable by an undergraduate with some basic Press. mathematical knowledge. Two asterisks mean the *—— and Price, G. 1973. “The logic of animal reader must have greater mathematical sophistica- conflicts.” Nature 246: 15–18. tion and/or some background knowledge in game **Myerson, R. 1978. “Refinements of the Nash theory.) equilibrium concept.” International Journal of Game Theory 7: 73–80. **Aumann, R. 1976. “Agreeing to disagree.” *Nash, J. 1951. “Noncooperative games.” Annals Annals of Statistics 4: 1236–9. of Mathematics 54: 289–95. *Basu, K. 1990. “On the non-existence of a **Pearce, D. G. 1984. “Rationalizable strategic rationality definition for extensive games.” Inter- behavior and the problem of perfection.” national Journal of Game Theory 19: 33–44. Econometrica 52: 1029–50. **Bernheim, B. D. 1984. “Rationalizable strategic *Pettit, P. and Sugden, R. 1989. “The backward behavior.” Econometrica 52: 1007–28. induction paradox.” The Journal of Philosophy 4: *Bicchieri, C. 1988. “Common knowledge and 1–14. backward induction: A solution to the paradox.” *Reny, P. 1992. “Rationality in extensive form In M. Vardi, ed., Theoretical Aspects of Reasoning games.” Journal of Economic Perspectives 6: 92– about Knowledge. Los Altos, CA: Morgan 100. Kaufmann Publishers. **Samuelson, L. 1997. Evolutionary Games and *——. 1989. “Self-refuting theories of strategic Equilibrium Selection. Cambridge, MA: MIT interaction: A paradox of common knowledge.” Press. Erkenntnis 30: 69–85. *Schelling, T. 1960. The Strategy of Conflict. Cam- *——. 1993. Rationality and Coordination. Cam- bridge, MA: Harvard University Press. bridge: Cambridge University Press. **Selten, R. 1965. “Spieltheorethische behandlung *Binmore, K. 1987. “Modeling rational players I.” eines oligopolmodels mit nachfragetragheit.” Economics and Philosophy 3, 179–214. Zietschrift fur die gesante Staatwissenschaft 121: *——. 1988. “Modeling rational players II.” Eco- 301–24. nomics and Philosophy 4: 9–55. **——. 1975. “Re-examination of the perfectness *Bonanno, G. 1991. “The Logic of Rational Play concept for equilibrium points in extensive in Games of Perfect Information.” Economics and games.” International Journal of Game Theory 4: Philosophy 7: 37–61. 22–55.

303 Cristina Bicchieri

**Taylor, P. D. and Jonker, L. 1978. “Evolution- *—— and Morgenstern, O. 1944. Theory of Games ary stable strategies and game dynamics.” Math. and Economic Behavior. Princeton: Princeton Biosci. 40: 145–56. University Press. **van Damme, E. 1987. Stability and Perfection of *Zermelo, E. 1913. “Uber eine anwendung der Nash Equilibria. Berlin: Springer Verlag. megenlehre auf die theorie des schachspiels.” In *von Neumann, J. 1928. “Zur theorie der Proceedings of the Fifth International Congress of gesellschaftspiele.” Mathematische Annalen 100: Mathematicians, vol. 2. Cambridge: Cambridge 295–320. University Press.

304 Part VII Science and Technology Computing in the Philosophy of Science

Chapter 23 Computing in the Philosophy of Science Paul Thagard

the inferential procedures by which knowledge 1 Introduction can be increased. Although epistemologists often focus on mundane knowledge, the most impress- What do philosophers do? Twenty or thirty years ive knowledge gained by human beings comes ago, one might have been told “analyze con- through the operation of science: experimenta- cepts” or “evaluate arguments.” The answer tion, systematic observation, and theorizing “write computer programs” would have inspired concerning the experimental and observational a blank stare and computational philosophy of results. In attempting to understand the struc- science might have sounded like the most self- ture and development of scientific knowledge, contradictory enterprise in philosophy since busi- philosophers of science have traditionally em- ness ethics (Thagard 1988). But computer use ployed a number of approaches, such as logical has since become much more common in philo- analysis and historical case studies. Computational sophy, and computational modeling can be seen modeling provides an additional method that as a useful addition to philosophical method, has already advanced understanding of such not as the abandonment of it (see Chapter 26, traditional problems as theory evaluation and COMPUTATIONAL MODELING AS A PHILOSOPHICAL scientific discovery. METHODOLOGY). This chapter concerns how computational If philosophy consisted primarily of conceptual models are making substantial contributions to analysis, or mental self-examination, or investiga- the philosophy of science. It reviews the progress tion of a priori truths, then computer modeling made by three distinct computational approaches: would indeed be alien to the enterprise. However, cognitive modeling, engineering artificial intelli- one may argue in favor of a different picture of gence, and theory of computation. The aim of philosophy, as primarily concerned with produc- cognitive modeling is to simulate aspects of ing and evaluating theories, for example theories human thinking; for philosophy of science, this of knowledge (epistemology), reality (meta- becomes the aim to simulate the thinking that physics), and right and wrong (ethics). One of scientists use in the construction and evaluation of the main functions of a theory of knowledge is hypotheses. Much artificial intelligence research, to explain how knowledge grows. This requires however, is not concerned with modeling human describing both the structure of knowledge and thinking, but with constructing algorithms that

307 Paul Thagard perform well on difficult tasks independently of 1980s, these debates were superseded by com- whether the algorithms imitate human thinking. putational research on discovery that showed how Similarly, the engineering AI approach to philo- actual cases of scientific discovery can be modeled sophy of science seeks to develop computational algorithmically. Although the models that have models of discovery and evaluation independently been produced to date clearly fall well short of of questions of human psychology. Computa- simulating all the thought processes of creat- tional philosophy of science has thus developed ive scientists, they provide substantial insights two approaches that reflect the two trends in AI into how scientific thinking can be viewed research, one concerned with modeling human computationally. performance and the other with machine intelli- Because of the enormous number of possible gence. A third approach uses abstract mathemat- solutions for any scientific problem, the algo- ical analysis and applies the theory of computation rithms involved in scientific discovery cannot to problems in the philosophy of science. guarantee that optimal discoveries will be made from input provided. Instead, computer models of discovery employ heuristics, approximate methods for attempting to cut through data 2 Cognitive Modeling complexity and find patterns. The pioneering step in this direction was the BACON project of Cognitive science is the interdisciplinary study Pat Langley, Herbert Simon, and their colleagues of mind, embracing philosophy, psychology, arti- (Langley et al. 1987). BACON is a program ficial intelligence, neuroscience, linguistics, and that uses heuristics to discover mathematical laws anthropology. From its modern origins in the from quantitative data, for example discovering 1950s, cognitive science has primarily worked Kepler’s third law of planetary motion. Although with the computational-representational under- BACON has been criticized for assuming an standing of mind: we can understand human oversimple account of human thinking, Qin and thinking by postulating mental representations Simon (1990) found that human subjects could akin to computational data structures and mental generate laws from numerical data in ways quite procedures akin to algorithms (Thagard 1996). similar to BACON. The cognitive-modeling approach in computa- Scientific discovery produces qualitative as well tional philosophy of science views topics such as as quantitative laws. Kulkarni and Simon (1988) discovery and evaluation as open to investigation produced a computational model of Krebs’ using the same techniques employed in cognitive discovery of the urea cycle. Their program, science. To understand how scientists discover KEKADA, reacts to anomalies, formulates and evaluate hypotheses, we can develop com- explanations, and carries out simulated experi- puter models that employ data structures and ments in much the way described in Hans Krebs’ algorithms intended to be analogous to human laboratory notebooks. mental representations and procedures. This Not all scientific discoveries are as data-driven approach can be viewed as part of naturalistic as the ones discussed so far. They often involve epistemology, which sees the study of know- the generation of new concepts and hypotheses ledge as closely tied to human psychology, not that are intended to refer to non-observable as an abstract logical exercise. entities. Thagard (1988) developed computa- tional models of conceptual combination, in which new theoretical concepts such as sound 2.1 Discovery wave are generated, and of abduction, in which new hypotheses are generated to explain puzzling In the 1960s and 1970s, philosophers of science phenomena. Magnani (2001) has also discussed discussed whether there is a “logic of discovery” how abductive inference can produce discoveries and whether discovery (as opposed to evalu- in science and mathematics. Darden (1990, 1998) ation) is a legitimate topic of philosophical (as has investigated computationally how theories opposed to psychological) investigation. In the that have empirical problems can be repaired.

308 Computing in the Philosophy of Science

One of the most important cognitive mech- Coherence theories of knowledge, ethics, and anisms for discovery is analogy, since scientists even truth have been popular among philo- often make discoveries by adapting previous sophers, but the notion of coherence is usually knowledge to a new problem. Analogy played a left rather vague. Hence coherence theories do role in some of the most important discoveries not appear sufficiently rigorous when compared ever made, such as Darwin’s theory of evolution to theories couched more formally using deduct- and Maxwell’s theory of electromagnetism. ive logic or probability theory. But connectionist During the 1980s, the study of analogy went models show how coherence ideas can be pre- well beyond previous philosophical accounts cisely and efficiently implemented. Since the mid- through the development of powerful com- 1980s, connectionist (neural network, PDP) putational models of how analogs are retrieved models have been very influential in cognitive from memory and mapped to current problems science. Only loosely analogous to the operation to provide solutions. Falkenhainer, Forbus, and of the brain, such models have numerous units Gentner (1989) produced SME, the Structure that are roughly like neurons, connected to each Mapping Engine, and this program was used to other by excitatory and inhibitory links of vary- model analogical explanations of evaporation ing strengths. Each unit has an activation value and osmosis (Falkenhainer 1990). Holyoak and that is affected by the activations of the units to Thagard (1989) used different computational which it is linked, and learning algorithms are methods to produce ACME, the Analogical Con- available for adjusting the strengths on links in straint Mapping Engine, which was generalized response to experience. into a theory of analogical thinking that applies In ECHO, a connectionist computational to scientific as well as everyday thinking (Holyoak model of explanatory coherence developed in & Thagard 1995). Thagard (1992), units are used to represent The above research projects illustrate how propositions that can be hypotheses or descrip- thought processes, such as those involved in tions of evidence, and links between units to numerical law generation, theoretical concept represent coherence relations. For example, if a formation, abduction, and analogy, can be under- hypothesis explains a piece of evidence, then stood computationally. Examples of nonpsycho- ECHO places an excitatory link between the unit logical investigations of scientific discovery are representing the hypothesis and the unit repres- described in the sections on engineering AI and enting the evidence. If two hypotheses are con- theory of computation. tradictory or competing, then ECHO places an inhibitory link between the units representing the two hypotheses. Repeatedly adjusting the 2.2 Evaluation activations of the units based on their links with other units results in a resting state in which How scientific hypotheses are evaluated has been some units are on (hypotheses accepted) and a central problem in philosophy of science since other units are off (hypotheses rejected). ECHO the nineteenth-century debates between John has been used to model many important cases in Stuart Mill and . Work in the the history of science (Nowak & Thagard 1992a, logical positivist tradition has centered on the 1992b; Thagard 1991, 1992, 1999). Eliasmith concept of confirmation, asking what it is for and Thagard (1997) argued that ECHO pro- hypotheses to be confirmed by observations. More vides a better account of hypothesis evaluation recently, various philosophers of science have than available Bayesian accounts, and challenged taken a Bayesian approach to hypothesis evalu- proponents of Bayesian models to produce ation, using probability theory to analyze scientific simulations of that are as detailed reasoning. In contrast, it is possible to develop and historically accurate as existing ECHO simu- an approach to hypothesis evaluation that lations. ECHO has also been used in a com- combines philosophical ideas about explanatory putational model of scientific consensus in which coherence with a connectionist (neural network) a group of scientists reach agreement about computational model (Thagard 1992, 2000). what theory to adopt by exchanging information

309 Paul Thagard about data, hypotheses, and explanations (Thagard trained as a philosopher before moving into AI 1999). research, reviewed over a dozen AI programs A different connectionist account of inference that formulate hypotheses to explain empirical to best explanation is given by Churchland data (Buchanan 1983). One of the earliest and (1989). He conjectures that abductive discovery most impressive programs was DENDRAL, and inference to the best explanation can both which performed chemical analysis. Given be understood in terms of prototype activation spectroscopic data from an unknown organic in distributed connectionist models, i.e. ones in chemical sample, it determined the molecular which concepts and hypotheses are not repres- structure of the sample (Lindsay et al. 1980). ented by individual units but by patterns of The program META-DENDRAL pushed the activation across multiple units. There is con- discovery task one step farther back: given a col- siderable psychological evidence that distributed lection of analytic data from a mass spectrometer, representations and prototypes are important in it discovered rules explaining the fragmentation human cognition, but no one has yet produced behavior of chemical samples. Buchanan has a running computational model of hypothesis continued to work on computational models of evaluation using these ideas. Nonconnectionist hypothesis formation in science (Buchanan & models of hypothesis evaluation, including pro- Philips 2001). babilistic ones, are discussed in the next section. Ideally, discovery programs should be capable of advancing science by producing novel results. One of the most successful in this regard is a program for chemical discovery, MECHEM, 3 Engineering AI which automates the task of finding mechan- ism for chemical reactions. Given experimental The cognitive-modeling approach to computa- evidence about a reaction, the program searches tional philosophy of science allies philosophy of for the simplest mechanism consistent with theory science with cognitive science and naturalistic and experiment (Valdés-Pérez 1994, 1995). epistemology, and it can be very fruitful, as the Valdés-Pérez has also written programs that previous section shows. However, much valu- have contributed to discoveries in biology and able work in AI and philosophy has been done physics. Kocabas and Langley (2000) have that makes no claims to psychological plausibil- developed a computational aid for generating ity. One can set out to build a scientist without process explanations in nuclear astrophysics. trying to reverse-engineer a human scientist. The In order to model biologists’ discoveries engineering AI approach to computational philo- concerning gene regulation in bacteria, Karp sophy of science is allied, not with naturalistic, (1990) wrote a pair of programs, GENSIM psychologistic epistemology, but with what has and HYPGENE. GENSIM was used to repres- been called “android epistemology,” the epi- ent a theory of bacterial gene regulation, and stemology of machines that may or may not HYPGENE formulates hypotheses that improve be built like humans (Ford, Glymour, & Hayes the predictive power of GENSIM theories given 1995). This approach is particularly useful when experimental data. More recently, Karp has it exploits such differences between digital com- shifted from modeling historical discoveries to puters and humans as the capacity for very fast the attempt to write programs that make original searches to perform tasks that human scientists discoveries from large scientific databases, such cannot do very well. as ones containing information about enzymes, proteins, and metabolic pathways (Karp & Mavrovouniotis 1994). A 1997 special issue of 3.1 Discovery the journal Artificial Intelligence (vol. 91) con- tains several examples of recent work on com- One goal of engineering AI is to produce pro- putational discovery. grams that can make discoveries that have eluded Cheeseman (1990) used a program that humans. Bruce Buchanan, who was originally applied Bayesian probability theory to discover

310 Computing in the Philosophy of Science previously unsuspected fine structure in the by probabilistic philosophers of science, who have infrared spectra of stars. Machine learning tech- tended to ignore the substantial problem of the niques are also relevant to social science research, intractability of Bayesian algorithms. particularly the problem of inferring causal models Theory evaluation in the context of medical from social data. The TETRAD program looks reasoning has been investigated by a group of at statistical data in fields such as industrial artificial intelligence researchers at Ohio State development and voting behavior and builds University (Josephson & Josephson 1994). They causal models in the form of a directed graph developed a knowledge-based system called RED of hypothetical causal relationships (Glymour, that uses data concerning a patient’s blood sam- Scheines, Spirtes, & Kelly 1987; Spirtes, Glymour, ple to infer what red-cell antibodies are present & Scheines 1993). in the patient. RED performs an automated ver- One of the fastest-growing areas of artificial sion of inference to the best explanation, using intelligence is “data mining,” in which machine heuristics to form a composite hypothesis con- learning techniques are used to discover regu- cerning what antibodies are present in a sample. larities in large computer data bases such as the Interestingly, Johnson and Chen (1996) com- terabytes of image data collected by astronom- pared the performance of RED with the per- ical surveys (Fayyad, Piatetsky-Shapiro, & Smyth formance of the explanatory coherence program 1996). Data mining is being applied with com- ECHO on a set of 48 cases interpreted by clin- mercial success by companies that wish to learn ical experts. Whereas RED produced the experts’ more about their operations, and similar machine judgments in 58 percent of the cases, ECHO learning techniques may have applications to large was successful in 73 percent of the cases. Hence, scientific data bases such as those being produced although the engineering AI approach to scient- by the human genome project. ific discovery has some evident advantages over the cognitive-modeling approach in dealing with some problems, such as mining hypotheses from 3.2 Evaluation large data bases, the cognitive-modeling approach exemplified by ECHO has not yet been surpassed The topic of how scientific theories can be evalu- by a probabilistic or other program that ignores ated can also be discussed from a computational human performance. perspective. Many philosophers of science (e.g. Howson & Urbach 1989) adopt a Bayesian approach to questions of hypothesis evaluation, 4 Theory of Computation attempting to use probability theory to describe and prescribe how scientific theories are assessed. But computational investigations of probabilistic Both the cognitive-modeling and engineering AI reasoning must deal with important problems approaches to philosophy of science involve writ- involving tractability that are usually ignored by ing and experimenting with running computer philosophers. A full-blown probabilistic approach programs. But it is also possible to take a more to a problem of scientific inference would need theoretical approach to computational issues to establish a full joint distribution of probabilit- in the philosophy of science, exploiting results in ies for all propositions representing hypotheses the theory of computation to reach conclusions and evidence, which would require 2n probabilit- about processes of discovery and evaluation. ies for n hypotheses, quickly exhausting the stor- age and processing capacities of any computer. Ingenious methods have been developed by com- 4.1 Discovery puter scientists to avoid this problem by using causal networks to restrict the number of prob- Scientific discovery can be viewed as a problem abilities required and to simplify the processing in formal-learning theory, in which the goal is involved (Pearl 1988, Neapolitain 1990). Sur- to identify a language given a string of inputs prisingly, such methods have not been explored (Gold 1968). Analogously, a scientist can be

311 Paul Thagard thought of as a function that takes as input a of hypotheses to explain them. They then showed sequence of formulas representing observations that the problem of picking the best explanation of the environment and produces as output a set is NP-hard, i.e. it belongs to a class of problems of formulas that represent the structure of the that are generally agreed by computational theor- world (Jain, Osherson, Royer, & Sharma 1999; ists to be intractable in that the amount of time Kelly & Glymour 1989; Martin & Osherson to compute them increases exponentially as the 1998; Osherson & Weinstein 1989). Although problems grow in size (see Chapter 2, COMPLEX- formal-learning theory has produced some ITY). Similarly, Thagard and Verbeurgt (1998) interesting theorems, they are limited in their and Thagard (2000) generalized explanatory relevance to the philosophy of science in several coherence into a mathematical coherence prob- respects. Formal-learning theory assumes a fixed lem that is NP-hard. What these results show is language and therefore ignores the conceptual that theory evaluation, whether it is conceived and terminological creativity that is important in terms of Bayesian probabilities, heuristic as- to scientific development. In addition, formal- sembly of hypotheses, or explanatory coherence, learning theory tends to view hypotheses pro- must be handled by computational approxima- duced as a function of input data, rather than as tion, not through exhaustive algorithms. So far, a much more complex function of the data and the theoretical results concerning coherence and the background concepts and theories possessed scientific evaluation have been largely negative, by a scientist. Formal-learning theory also over- but they serve to outline the limits within which emphasizes the goal of science to produce true computational modeling must work. descriptions, neglecting the important role of Kelly (1996, 2001) has approached the prob- explanatory theories and hypothetical entities in lem of evaluation from a different epistemological scientific progress. and formal perspective. In contrast to the view Nevertheless, ongoing work in formal- that scientific inference adopts theories on the learning theory may shed light on more realistic basis of their coherence with data and each other, kinds of scientific discovery. Schulte (2000) has he suggests that philosophy of science should provided a formal account of a problem in aim to specify reliable procedures that are guar- physics that starts with observed reactions and anteed to converge to correct outputs. For Kelly, infers conservation principles that govern all reac- it is not enough to be able to say when one tions among elementary particles. He shows that theory is better than another; rather, a formal there is a reliable inference procedure that is specification of scientific methods should pro- guaranteed to arrive at an empirically adequate vide algorithms that come with a guarantee that set of conservation principles as more and more they will converge on the right answers, as in evidence is obtained. Kelly (1996) also discusses Schulte’s (2000) method of inferring con- more realistic methods for modeling science, servation laws. Although Kelly has proved some which are reviewed in the next section. interesting theorems concerning the conditions under which such guarantees might be available, it is not clear whether scientific inference in gen- 4.2 Evaluation eral is open to this kind of analysis. Inference from numerical data to mathematical laws that The theory of computational complexity has describe them may use methods that can be provided some interesting results concerning proven to be reliable, but inference to theories hypothesis evaluation. Suppose you have n hypo- that postulate such theoretical entities as quarks, theses and you wish to evaluate all the ways in genes, and mental representations seems inescap- which combinations of them can be accepted ably risky. There is good empirical evidence, and rejected: you then have to consider 2n pos- particularly in the enormously successful techno- sibilities, an impossibly large number for even logical applications of theories in the natural not very large n. Bylander et al. (1991) gave a sciences, that scientific method does sometimes formal definition of an abduction problem con- converge on an approximation to truth. But it sisting of a set of data to be explained and a set is unlikely that guarantees of reliability and

312 Computing in the Philosophy of Science convergence to the truth will turn out to be ing of computational models. This methodology available for complex, theory-oriented kinds of typically has numerous advantages over pencil- scientific inquiry. and-paper constructions. First, it requires con- siderable precision, in that to produce a running program the structures and algorithms postulated as part of scientific cognition need to be explicitly 5 What Computing Adds to and carefully specified. Second, getting a program Philosophy of Science to run provides a test of the feasibility of its assumptions about the structure and processes More than twenty years ago, Aaron Sloman of scientific development. Contrary to the pop- (1978) published an audacious book, The Com- ular view that clever programmers can get a puter Revolution in Philosophy, which predicted program to do whatever they want, producing a that within a few years any philosopher not program that mimics aspects of scientific cogni- familiar with the main developments of artificial tion is often very challenging, and production of intelligence could fairly be accused of profes- a program provides a minimal test of computa- sional incompetence. Since then, computational tional feasibility. Moreover, the program can then ideas have had a substantial impact on the phi- be used for testing the underlying theoretical losophy of mind, but a much smaller impact on ideas by examining how well the program works epistemology and philosophy of science. Why? on numerous examples of different kinds. Com- One possible reason is the kind of training that parative evaluation becomes possible when dif- most philosophers have, which includes little ferent programs accomplish a task in different preparation for actually doing computational ways: running the programs on the same data work. Philosophers of mind have often been able allows evaluation of their computational models to learn enough about artificial intelligence to and background theoretical ideas. Third, if the discuss it, but for epistemology and philosophy program is intended as part of a cognitive model, of science it is much more useful to perform it can be assessed in terms of how well it models computations rather than just to talk about them. human thinking. Thus this chapter can end with a summary of The assessment of cognitive models can what is gained by adding computational modeling address questions such as the following: to the philosophical tool kit. Bringing artificial intelligence into philosophy Genuineness: is the model a genuine instantia- of science introduces new conceptual resources for tion of the theoretical ideas about the structure dealing with the structure and growth of scientific and growth of scientific knowledge, and is the knowledge. Instead of being restricted to the usual program a genuine implementation of the model? representational schemes based on formal logic Breadth of application: does the model apply to and ordinary language, computational approaches lots of different examples, not just a few that to the structure of scientific knowledge can in- have been cooked up to make the program work? clude many useful representations such as proto- Scaling: does the model scale up to examples typical concepts, concept hierarchies, production that are considerably larger and more complex rules, causal networks, mental images, dynamic than the ones to which it has been applied? models, and so on. Philosophers concerned with Qualitative fit: does the computational model the growth of scientific knowledge from a com- perform the same kinds of tasks that people do putational perspective can go beyond the narrow in approximately the same way? resources of inductive logic to consider algorithms Quantitative fit: can the computational model for generating numerical laws, discovering causal simulate quantitative aspects of psychological networks, forming concepts and hypotheses, and experiments, e.g. ease of recall and mapping in evaluating competing explanatory theories. analogy problems? AI not only provides new conceptual resources Compatibility: does the computational model to philosophy of science, it also brings a new simulate representations and processes that are methodology involving the construction and test- compatible with those found in theoretical

313 Paul Thagard accounts and computational models of other What role does visual imagery play in the struc- kinds of cognition? ture and growth of scientific knowledge? Although various philosophers, historians, and psycholo- Computational models of the thought pro- gists have documented the importance of visual cesses of scientists that satisfy these criteria have representations in scientific thought, existing the potential to increase our understanding of computational techniques have not been well- the scientific mind enormously. Engineering AI suited for providing detailed models of the cog- need not address questions of qualitative and nitive role of pictorial mental images (see e.g. quantitative fit with the results of psychological Shelley 1996). Computational models of high- experiments, but should employ the other four level visual cognition are beginning to be devel- standards of assessment. oped (e.g. Davies & Goel 2000; Croft & Thagard There are numerous issues connecting com- 2002), but they have not yet been applied to putation and the philosophy of science that have scientific discovery and evaluation. not been touched on in this review. Computer Perhaps problems such as these will, like other science can itself be a subject of philosophical issues concerning discovery and evaluation, yield investigation, and some work has been done to computational approaches that involve cog- discussing epistemological issues that arise in nitive modeling, engineering AI, and the theory computer research (see e.g. Colburn 2000; Fetzer of computation. 1998; Floridi 1999; Thagard 1993). In particu- lar, the philosophy of artificial intelligence and cognitive science are fertile areas of philosophy Acknowledgments of science. Computer modeling can also be useful in developing fields of interest to both This chapter is an updated and expanded version philosophy and science, such as artificial life (see of Thagard 1998. I am grateful to Luciano Floridi Chapter 15, ARTIFICIAL LIFE, and Chapter 26, and Kevin Kelly for comments on earlier versions, COMPUTATIONAL MODELING AS A PHILOSOPHICAL and to the Natural Sciences and Engineering METHODOLOGY). The focus of this chapter has Research Council of Canada for grant support. been more narrow, on how computational models can contribute to philosophy of science. By way of conclusion, here is a list of some References open problems that seem amenable to computa- tional/philosophical investigation: Buchanan, B. 1983. “Mechanizing the search for In scientific discovery, how are new questions explanatory hypotheses.” PSA 1982, vol. 2. East generated? Formulating a useful question such Lansing: Philosophy of Science Association. [An as “How might species evolve?” or “Why do the early paper on the prospects for mechanizing planets revolve around the sun?” is often a pre- discovery.] Buchanan, B. G. and Phillips, J. P. 2001. “Toward requisite to more data-driven and focused pro- a computational model of hypothesis formation cesses of scientific discovery, but no computational and model building in science.” Unpublished account of scientific question generation has yet manuscript, University of Pittsburgh. [Describes been given. recent computational work on discovery.] What is the role of emotions in scientific think- Colburn, T. R. 2000. Philosophy and Computer ing? Scientists often generate questions in part Science. Armonk: M. E. Sharpe. [Introduction to through emotional stimuli such as curiosity and philosophical issues in computer science.] surprise, but no computer simulations of scientific Bylander, T., Allemang, D., Tanner, M., and thinking have yet taken emotions into account. Josephson, J. 1991. “The computational com- Emotions are also outputs from scientific evalu- plexity of abduction.” Artificial Intelligence, ation, as when scientists praise a theory as elegant 49: 25–60. [Proves that abductive inference is or beautiful or exciting. Thagard 2002 discusses NP-hard.] emotions and inputs and outputs in scientific Cheeseman, P. 1990. “On finding the most prob- discovery and evaluation. able model.” In J. Shrager and P. Langley, eds.,

314 Computing in the Philosophy of Science

Computational Models of Scientific Discovery and debate.” In T. W. Bynum and J. H. Moor, eds., Theory Formation. San Mateo, CA: Morgan The Digital Phoenix: How Computers Are Chang- Kaufmann, pp. 73–96. [Applies a Bayesian ing Philosophy. Oxford: Blackwell, pp. 253–73. approach to machine learning.] [Philosophical reflection on a live issue in com- Churchland, P. 1989. A Neurocomputational Per- puter science.] spective. Cambridge, MA: MIT Press. [Takes a Floridi, L. 1999. Philosophy and Computing: An neural-network approach to the philosophy of Introduction. London: Routledge. [Discusses science.] philosophical issues about computers.] Croft, D. and Thagard, P. 2002. “Dynamic Ford, K. M., Glymour, C., and Hayes, P. J., eds. imagery: a computational model of motion and 1995. Android Epistemology. Menlo Park, CA: visual imagery.” In L. Magnani, ed., Model-based AAAI Press. [Collection of articles on non- Reasoning: Scientific, Technology, Values. New psychological approaches to AI.] York: Kluwer/Plenum. [Develops algorithms for Glymour, C., Scheines, R., Spirtes, P., and Kelly, K. modeling visual reasoning.] 1987. Discovering Causal Structure. Orlando: Darden, L. 1990. “Diagnosing and fixing fault in Academic Press. [Provides computational theories.” In J. Shrager and P. Langley, eds., methods for discovering causal structure from Computational Models of Discovery and Theory numerical data.] Formation. San Mateo, CA: Morgan Kaufman, Gold, E. 1968. “Language identification in the pp. 219–246. [Discusses how anomalies in scient- limit.” Information and Control 10: 447–74. ific theories can be resolved.] [Early seminal paper on formal-learning theory.] ——. 1998. “Anomaly-driven redesign: computa- Holyoak, K. J. and Thagard, P. 1989. “Analogical tional philosophy of science experiments.” In mapping by constraint satisfaction.” Cognitive T. W. Bynum and J. H. Moor, eds., The Digital Science 13: 295–355. [Describes a connectionist Phoenix: How Computers Are Changing Philo- model of analogical reasoning.] sophy. Oxford: Blackwell, pp. 62–78. [Describes —— and Thagard, P. 1995. Mental Leaps: Analogy a computer model of theory revision.] in Creative Thought. Cambridge, MA: MIT Davies, J. R. and Goel, A. K. 2000. “A com- Press/Bradford Books. [Surveys cognitive science putational theory of visual analogical transfer.” research on analogy from the perspective of the Technical report GIT-COGSCI-2000/3, Georgia authors’ multiconstraint theory.] Institute of Technology. [Provides a computa- Howson, C. and Urbach, P. 1989. Scientific tional account of visual analogy.] Reasoning: The Bayesian Tradition. Lasalle, IL: Eliasmith, C. and Thagard, P. 1997. “Waves, par- Open Court. [Defends a Bayesian approach to ticles, and explanatory coherence.” British Journal the philosophy of science.] for the Philosophy of Science 48: 1–19. [Uses the Jain, S., Osherson, D., Royer, J. S., and Sharma, A. computer program ECHO to simulate reasoning 1999. Systems That Learn, 2nd ed. Cambridge, in physics.] MA: MIT Press. [An up-to-date treatment of Falkenhainer, B. 1990. “A unified approach to formal-learning theory.] explanation and theory formation.” In J. Shrager Johnson, T. R. and Chen, M. 1995. “Comparison and P. Langley, eds., Computational Models of of symbolic and connectionist approaches for Discovery and Theory Formation. San Mateo, CA: multiple disorder diagnosis: heuristic search vs. Morgan Kaufman pp. 157–196. [Describes a rich explanatory coherence.” Unpublished manuscript, simulation of scientific reasoning.] Ohio State University. [A computational evalu- ——, Forbus, K. D., and Gentner, D. 1989. “The ation of different approaches to diagnosis.] structure-mapping engine: algorithms and exam- Josephson, J. R. and Josephson, S. G., eds. 1994. ples.” Artificial Intelligence 41: 1–63. [Presents Abductive Inference: Computation, Philosophy, a powerful computational approach to analogy.] Technology. Cambridge: Cambridge University Fayyad, U., Piatetsky-Shapiro, G., and Smyth, P. Press. [Presents results of an extensive research 1996. “From data mining to knowledge discovery project on abduction.] in databases.” AI Magazine 17(3): 37–54. [A Karp, P. and Mavrovouniotis, M. 1994. “Repres- survey of applied machine learning.] enting, analyzing, and synthesizing biochemical Fetzer, J. H. 1998. “Philosophy and computer pathways.” IEEE Expert 9(2): 11–21. [Describes science: feflections on the program verification a computational model of biological reasoning.]

315 Paul Thagard

Kelly, K. 1996. The Logic of Reliable Inquiry. New Osherson, D. and Weinstein, S. 1989. “Identifiable York: Oxford University Press. [Uses formal collections of countable structures.” Philosophy of methods to analyze how scientists can achieve Science 56: 94–105. [Applies formal-learning true beliefs.] theory to the philosophy of science.] ——. 2001. “The logic of success.” British Journal Pearl, J. 1988. Probabilistic Reasoning in Intelli- for the Philosophy of Science, 51: 639–66. [Uses gent Systems. San Mateo, CA: Morgan Kaufman. formal techniques to analyze the success of [Classic work on Bayesian networks.] scientific inference.] Qin, Y. and Simon, H. 1990. “Laboratory replica- —— and Glymour, C. 1989. “Convergence to the tion of scientific discovery processes.” Cognitive truth and nothing but the truth.” Philosophy of Science 14: 281–312. [Experimental confirmation Science 56: 185–220. [Applies formal-learning of computational theory of discovery.] theory to the philosophy of science.] Schulte, O. 2000. “Inferring conservation laws in Kocabas, S. and Langley, P. 2000. “Computer : a case study in the problem of generation of process explanations in nuclear induction.” British Journal for the Philosophy of astrophysics.” International Journal of Human- Science 51: 771–806. [Provides a formal model Computer Studies 53: 377–92. [Use mathemat- of scientific induction.] ical techniques to make astrophysical discoveries.] Shelley, C. P. 1996. “Visual abductive reasoning in Kulkarni, D. and Simon, H. 1988. “The processes archaeology.” Philosophy of Science 63: 278–301. of scientific discovery: the strategy of experimenta- [Analyzes interesting examples of hypothesis tion.” Cognitive Science 12: 139–75. [Simulates formation using visual imagery.] Krebs’ chemical discoveries.] Sloman, A. 1978. The Computer Revolution in Philo- Langley, P., Simon, H., Bradshaw, G., and Zytkow, sophy. Atlantic Highlands, NJ: Humanities Press. J. 1987. Scientific Discovery. Cambridge, MA: [An early work linking philosophy with AI.] MIT Press/Bradford Books. [The pioneering Spirtes, P., Glymour, C., and Scheines, R. 1993. work on cognitive modeling of discovery.] Causation, Prediction, and Search. New York: Lindsay, R., Buchanan, B., Feigenbaum, E., and Springer-Verlag. 2nd ed. 2000, published by MIT Lederberg, J. 1980. Applications of Organic Press. [Describes formal methods for deriving Chemistry for Organic Chemistry: The DENDRAL causal models from data.] Project. New York: McGraw Hill. [Classic early Thagard, P. 1988. Computational Philosophy of AI work on scientific reasoning.] Science. Cambridge, MA: MIT Press/Bradford Magnani, L. 2001. Abduction, Reason, and Science: Books. [Develops computational methods in the Processes of Discovery and Explanation. New York: philosophy of science for both discovery and Kluwer/Plenum. [A philosophical discussion of evaluation.] inference to explanatory hypotheses.] ——. 1991. “The dinosaur debate: explanatory Martin, E. and Osherson, D. 1998. Elements of coherence and the problem of competing hypo- Scientific Inquiry. Cambridge, MA: MIT Press. theses.” In J. Pollock and R. Cummins, eds., [Basic work in formal-learning theory.] Philosophy and AI: Essays at the Interface. Neapolitain, R. 1990. Probabilistic Reasoning in Cambridge, MA: MIT Press/Bradford Books, Expert Systems. New York: John Wiley. [An pp. 279–300. [Simulates different views on why introduction to Bayesian networks.] dinosaurs became extinct.] Nowak, G. and Thagard, P. 1992a. “Copernicus, ——. 1992. Conceptual Revolutions. Princeton: Ptolemy, and explanatory coherence.” In R. Giere, Princeton University Press. [Provides a ed., Cognitive Models of Science, vol. 15. Min- cognitive-computational account of scientific neapolis: University of Minnesota Press, pp. 274– revolutions.] 309. [Simulates the acceptance of Copernicus over ——. 1993. “Computational tractability and con- Ptolemy.] ceptual coherence: why do computer scientists —— and ——. 1992b. “Newton, Descartes, believe that P ≠ NP?” Canadian Journal of Philo- and explanatory coherence.” In R. Duschl and sophy 23: 349–64. [Discusses grounds for current R. Hamilton, eds., Philosophy of Science, Cognitive beliefs about computational complexity.] Psychology and Educational Theory and Practice. ——. 1996. Mind: Introduction to Cognitive Science. Albany: SUNY Press, pp. 69–115. [Simulates the Cambridge, MA: MIT Press. [A concise interdis- acceptance of Newton’s physics over Descartes’s.] ciplinary textbook.]

316 Computing in the Philosophy of Science

——. 1998. “Computation and the philosophy of Science. Cambridge: Cambridge University Press. science.” In T. W. Bynum and J. H. Moor, eds., [Describes the role of emotion in scientific The Digital Phoenix: How Computers Are Chang- thinking.] ing Philosophy. Oxford: Blackwell, pp. 48–61. —— and Verbeurgt, K. 1998. “Coherence as [Earlier version of the present article.] constraint satisfaction.” Cognitive Science 22: 1– ——. 1999. How Scientists Explain Disease. 24. [A mathematical/compuational analysis of Princeton: Princeton University Press. [Discusses coherence.] the development and acceptance of the bacterial Valdés-Pérez, R. E. 1994. “Conjecturing hidden theory of ulcers in cognitive/computational entities via simplicity and conservation laws: terms.] Machine discovery in chemistry.” Artificial ——. 2000. Coherence in Thought and Action. Cam- Intelligence 65: 247–80. [Presents powerful com- bridge, MA: MIT Press. [Applies a computational putational techniques for making original scient- model of coherence to scientific and other kinds ific discoveries.] of thinking.] ——. 1995. “Machine discovery in chemistry: ——. 2002. “The passionate scientist: emotion in new results.” Artificial Intelligence 74: 191–201. scientific cognition.” In P. Carruthers, S. Stich, [Describes additional computational discoveries and M. Siegal, eds., The Cognitive Basis of in chemistry.]

317 Timothy Colburn

Chapter 24 Methodology of Computer Science Timothy Colburn

is a science concerned with the study of com- Introduction putational processes. A computational process is distinguished from, say, a chemical or electrical Science and philosophy are often distinguished process, in that it is studied “in ways that ignore by pointing out that science seeks explanation its physical nature” (Hailperin et al. 1999: 3). while philosophy seeks justification. To ask what For example, the process by which a card player accounts for the neuronal firing of synapses in the arranges cards in her hand, and the process by brain, for example, is a scientific question, while which a computer sorts names in a customer list, to ask what would constitute adequate grounds though they share nothing in common physic- for believing that an artificially constructed neural ally, may nevertheless embody the same com- network is conscious is a philosophical one. So putational process. They may, for example, both philosophy has been characterized as the critical proceed by scanning the items to be arranged one evaluation of beliefs through the analysis of con- by one, determining the proper place of each cepts in a given area of inquiry. Of course, sci- scanned item relative to the items already scanned, ence is also concerned with critically evaluating and inserting it into that place, perhaps necessit- beliefs and analyzing concepts. However, philo- ating the moving of previously scanned items to sophy is a non-empirical, or a priori, discipline, make room. This process (known as an insertion in distinct contrast with science. sort in computer science terms) can be precisely Computer science would seem to be distin- described in a formal language without talking guished from philosophy just as any other science. about playing cards or semiconducting elements. But computer science is unique among the sci- When so described, one has a computational ences in the types of models it creates. In seek- model of the process in the form of a computer ing explanations, science often constructs models program. This model can be tested, in a way to test hypotheses for explaining phenomena. analogous to how a hypothesis is tested in the These models, in the form of experimental appar- natural sciences, by executing the program and atus, are of course physical objects. The models observing its behavior. It can also be reasoned built and manipulated in computer science, about abstractly, so that questions can be an- however, are not physical at all. Computer science swered about it, such as whether there are other

318 Methodology of Computer Science processes which will have the same effect but science and the nature of abstraction in com- achieve it more efficiently. Building computational puter science. The remainder of this chapter turns models and answering these kinds of questions its attention to these issues. form a large part of what computer scientists do. The explosive growth in the number of com- puter applications in the last several decades has Computer Science and shown that there is no limit to how many real- Mathematics world processes are amenable to modeling by computer. Not only have traditional activities, like record-keeping, investing, publishing, and Philosophical contributions to the foundations of banking, been simply converted to control by scientific disciplines often center around “pivotal computational models, but whole new kinds of questions” regarding reductionist attempts. In activity have been created that would not be the philosophy of biology, for example, the ques- possible without such models. These are the by- tion is whether the science of the organic can now-familiar “virtual” activities described in the be reduced to the science of the inorganic (the language of cyberspace: e-mail, chatrooms, web reduction of biological to physical laws). In math- surfing, online shopping, internet gaming, and ematics, logicism claims that all of mathematics so on. can be reduced to logic. For science in general, The role of philosophy in that subfield of logical advocated the reduction of computer science known as artificial intelligence theoretical vocabulary to observational vocabu- (AI) has long been recognized, given the roles lary. An early “pivotal question” in the philosophy of knowledge and reasoning in AI. And the of computer science is whether computer science reverse, the role of AI in philosophy, has even can be reduced to a branch of mathematics. How been highlighted by some, albeit controversially. a computer scientist answers this question can (See Chapter 9, THE PHILOSOPHY OF AI AND ITS greatly influence his or her methodology. CRITIQUE, and Chapter 10, COMPUTATIONALISM, The range of perspectives from which the CONNECTIONISM, AND THE PHILOSOPHY OF MIND.) reductionist issue can be addressed is wide. Con- But apart from considerations arising from the sider the following view, expressed by C. A. R. modeling of knowledge and reason, computer Hoare: “Computer programs are mathematical science is ripe for the good old-fashioned ana- expressions. They describe, with unprecedented lysis that philosophy can provide for any science. precision and in the most minute detail, the Thus, a perfectly reasonable role of philosophy behavior, intended or unintended, of the com- is to attempt to place computer science within puter on which they are executed” (Hoare 1986: the broad spectrum of inquiry that constitutes 115). And this alternative, offered by C. Floyd: science. The concern here is to deal with the “Programs are tools or working environments inevitable identity crises that crop up in the self- for people. [They] are designed in processes of image of any adolescent, which computer science learning and communication so as to fit human certainly is. Philosophy should address questions needs” (Floyd 1987: 196). The view expressed like: What is the relation between mathematics by Hoare is unequivocal: computer programs are and computer science? Is there a sense in which mathematical expressions. The quote by Floyd computer science is experimental science? Is a is less precise, but expresses a view on the func- computer programmer merely a data wizard, or tion of programs for humans in decidedly non- can she also engage in information modeling? mathematical terms. While these views do not What is the nature of abstraction in computer necessarily contradict one another, they can most science? What are the ontological implications definitely signal contrasting interpretations as to of computer science concepts? From the point how computer programs ought to be designed, of view of computer science methodology, the built, and used. most probing of these questions concerns the While these quotations express views on the relation between mathematics and computer function and status of computer programs, the

319 Timothy Colburn differences of opinion extend to the broader The Formal Verification Debate notion of computing as a science, in which the task of actually creating a program is but one aspect. Of related concern, for example, are the The use of formal verification in computer sci- activities of program specification and verification. ence has generated debate since the appearance A program specification is a detailed description of a paper on verification and social processes by of a program’s input and output, ignoring the R. DeMillo, R. Lipton, and A. Perlis in 1979. details of how the program actually accomplishes But it was not until 1988 that these questions its task. Program verification is the process of drew the attention of a “pure” philosopher, when determining whether a program actually conforms J. Fetzer resurrected the program verification/ to its specification. These activities are just as social process debate of a decade earlier and germane to the software process as writing the subjected it to genuine . program itself, and there is no agreement on Before this time, debate on the issues was evid- whether or not program specifications should be enced mainly by differences in visionary accounts mathematical entities and whether or not pro- of how the young discipline of computer science gram verification can be a purely mathematical ought to proceed, given not by philosophers but activity. by computer science practitioners and teachers. There is agreement, however, on the possibil- One of the early proponents of formal program ity of mathematically reasoning about programs verification was John McCarthy, who is also given as the abstract representations of algorithms, as credit for coining the term “artificial intelligence” opposed to programs as the causal manipula- in the 1950s. McCarthy was originally motivated tions of bits. For example, given a program P by a theory of computation that would allow, consisting of the statements S1;...; Sn, it is pos- among other advantages, the automatic transla- sible to construct statements like “Let S1, S2,..., tion from one linguistic paradigm to another. One and Sn be an abstract representation of program can, in fact, look back now after nearly 30 years P. Then P has property R,” where R describes and confirm that automatic program translation, some aspect of the execution of P in the abstract with the help of precise language specification, has sense. For example, R might describe limits on been accomplished in the case of language com- the time it would take P to run, or the amount pilers. These are programming tools that translate of memory P would require to execute. By giv- programs written in familiar human languages ++ ing precise interpretations to the Si in a pertinent like Basic, C , and Java, into the machine lan- language and appropriately choosing R, it may guage of computers, which is composed only of be possible that the statement above is a theorem zeroes and ones. However, as every language in a formal language. This is in fact the approach reference manual’s warranty disclaimer demon- taken by modern researchers in computer sci- strates, no automatic compiler in all cases cor- ence who are concerned with reasoning about rectly translates programs into machine language. algorithms and data structures. Thus, there is the distinction between (1) using While this is a justification of how reasoning mathematical methods during language transla- about programs can be regarded as mathemat- tion to produce highly reliable machine language ical, it is yet a much broader claim to say that code, and (2) using mathematical methods to computer science is, or ought to aspire to be, a prove that a source program would behave, in branch of mathematics. For there are still the an abstract sense, exactly as its specification im- issues of whether the specification, generation, plies. McCarthy, seeing no obstacle to (2), wrote: or maintenance of programs (apart from reason- ing about completed ones) is or ought to be like It should be possible almost to eliminate a mathematical activity. The issue which motiv- debugging. Debugging is the testing of a pro- ates and underlies much of the tension in philo- gram on cases one hopes are typical, until it sophical discussion of computer science is formal seems to work. This hope is frequently vain. verification, or mathematically reasoning about a Instead of debugging a program, one should program’s outcome. prove that it meets its specifications, and this

320 Methodology of Computer Science

proof should be checked by a computer pro- germane to neither the truth of mathematical gram. (McCarthy 1962: 22) theorems nor program verifications:

While McCarthy was one of the first to express Indeed, while social processes are crucial in this opinion, it came to be shared by others, determining what theorems the mathematical who strove to describe what such a proof would community takes to be true and what proofs it be like. P. Naur recognized that one way to talk takes to be valid, they do not thereby make about a program both as a static, textual entity them true or valid. The absence of similar and as a dynamic, executing entity, was to con- social processes in determining which programs ceive of the program as executing, but from time are correct, accordingly, does not affect which to time to conceptually “halt” it and make state- programs are correct. (Fetzer 1988: 1049) ments about the state of its abstract machine at the time of halting (Naur 1966). By making a DeMillo, Lipton, and Perlis hit upon, for exam- number of these “snapshots” of a conceptually ple, the boredom, tedium, and lack of glamor executing program, and by providing justifica- involved in reviewing proofs produced by tions for each on the basis of the previous one, a mechanical verifiers. But for Fetzer, if this is all proof about the state of the abstract machine there is to their criticism of formal verification, it upon termination could be constructed. is not substantial. As Fetzer pointed out, social Though this idea held much promise for processes are characterized by transitory patterns believers in the mathematical paradigm (i.e., of human behavior which, one could imagine, that computer science is a branch of formal in different circumstances would reserve for pro- mathematics), it came under attack in the above- gram verification the same sort of excitement mentioned essay by DeMillo, Lipton, and Perlis. and collegial collaboration that marks the best They argued that mechanically produced pro- mathematical research. Thus DeMillo, Lipton, gram verifications, which are long chains of and Perlis have identified a difference in practice dense logical formulas, are not what constitute between mathematical research and formal pro- mathematical proofs. In coming to be accepted, gram verification, but not in principle. a mathematical proof undergoes social processes Fetzer believes that formal program verification in its communication and peer scrutiny, pro- cannot fulfill the role that some of its advocates cesses that cannot be applied to unfathomable would assign to it within software engineering, pages of logic. While DeMillo, Lipton, and Perlis but he attacks it from a nonsocial, more strictly did not subscribe to the mathematical paradigm, philosophical perspective. This has to do with they also did not deny that programming is like the relationship between mathematical models mathematics. An analogy can be drawn between and the causal systems they are intended to de- mathematics and programming, but “the same scribe. Close scrutiny of this relationship reveals, social processes that work in mathematical proofs for Fetzer, the relative, rather than absolute, doom verifications” (DeMillo et al. 1979: 275). nature of the program correctness guarantee that Social processes, they argued, are critical: formal verification can provide. It is only pos- sible to prove formally something about a formal No matter how high the payoff, no one will model, that is, a formal program model rendered ever be able to force himself to read the in formal text. It is not possible to prove formally incredibly long, tedious verifications of real- something about a causal model, that is, an actual, life systems, and unless they can be read, executing program represented in a physical, elec- understood, and refined, the verifications are tronic substrate of bistable processor and memory worthless. (DeMillo et al. 1979: 276) elements. “[I]t should be apparent that the very idea of the mathematical paradigm for computer Although Fetzer was also a critic of the science trades on ambiguity” (Fetzer 1991: 209). mathematical paradigm for computer science, Strong claims of formal verificationists are it was for different reasons. He argued that victim to this ambiguity because they ignore the presence or absence of social processes is several distinctions: between programs running

321 Timothy Colburn on abstract machines with no physical counter- of bedrooms, whether the bathrooms are full or part and programs running on abstract machines half, and so on. The realtor’s description leaves with a physical counterpart; between “programs- out architectural detail but describes the same as-texts” and “programs-as-causes”; and between entity at a more general level, and so it is an pure and applied mathematics. Recognizing these abstraction of the architect’s description. But distinctions, for Fetzer, reveals that the claim abstraction is relative. For example, the archi- that it is possible to reason in a purely a priori tect’s description is itself an abstraction when manner about the behavior of a program is true compared to a metallurgist’s description of the if the behavior is merely abstract; false, and dan- nails, screws, and other fasteners making up the gerously misleading otherwise. This guarantees house, and the botanist’s description of the vari- the indispensability of empirical methods in the ous cellular properties of the wood it contains. software development process, for example, the The computer scientist’s world is a world of use of program testing in an effort to eliminate nothing but abstractions. It would not be possible program bugs. to create the complex virtual worlds described above if the only things computer scientists could talk about were bits, bytes, and microcircuits. One can give an accurate idea of what computer Abstraction in Computer Science scientists do by describing the abstraction tools they use. Now to characterize computer science Computer scientists are often thought to labor as involved with abstractions seems to claim for exclusively in a world of bits, logic circuits, and it a place alongside mathematics as a purely for- microprocessors. Indeed, the foundational con- mal endeavor. But the general trends in all pro- cepts of computer science are described in the gramming are toward higher-quality software by language of binary arithmetic and logic gates, abstracting away from the lower-level concepts but it is a fascinating aspect of the discipline that in computer science and toward the objects and the levels of abstraction that one can lay upon information that make up the real world. This is this foundational layer are limitless, and make a kind of abstraction that is fundamentally dif- possible to model familiar objects and processes ferent from that which takes place in mathematics. of everyday life entirely within a digital world. Understanding the difference is crucial in avoid- When digital models are sufficiently realistic, ing the persistent misconception by some that the environments they inhabit are called virtual computer science is just a branch of pure math- worlds. So today, of course, there are virtual ematics. Both mathematics and computer sci- libraries, virtual shopping malls, virtual commu- ence are marked by the introduction of abstract nities, and even virtual persons, like the digital objects into the realm of discourse, but they differ version of actor Alan Alda created in an episode fundamentally in the nature of these objects. The of PBS’s Scientific American Frontiers. difference has to do with the abstraction of form Complex virtual worlds such as these are made versus the abstraction of content. possible by computer scientists’ ability to dis- Traditionally, mathematics, as a formal science, tance themselves from the mundane and tedious has been contrasted with the factual sciences such level of bits and processors through tools of as physics or biology. As natural sciences, the abstraction. To abstract is to describe something latter are not concerned with abstraction beyond at a more general level than the level of detail that offered by mathematics as an analytical tool. seen from another point of view. For example, The literature is full of strict bifurcations between an architect may describe a house by specifying the nature of formal and factual science in terms the height of the basement foundation, the loca- of the meanings of the statements involved tion of load-bearing walls and partitions, the R- in them. R. Carnap, for example, employs the factor of the insulation, the size of the window analytic/synthetic distinction in claiming that the and door rough openings, and so on. A realtor, formal sciences contain only analytic statements. however, may describe the same house as having Since analytic statements are true only by virtue a certain number of square feet, a certain number of the transformation rules of the language in

322 Methodology of Computer Science which they are made, Carnap is led to the view The propositions of mathematics have, there- that “[t]he formal sciences do not have any objects fore, the same unquestionable certainty which at all; they are systems of auxiliary statements is typical of such propositions as “All bachelors without objects and without content” (Carnap are unmarried,” but they also share the com- 1953: 128). Thus, according to Carnap the ab- plete lack of empirical content which is associ- straction involved in mathematics is one totally ated with that certainty: The propositions of away from content and toward the pure form of mathematics are devoid of all factual content; linguistic transformations. they convey no information whatever on any Not all philosophers of mathematics agree empirical subject matter. (Hempel 1953: 159) with Carnap that mathematics has only lin- guistic utility for scientists, but there is agreement In each of these accounts of mathematics, all on the nature of mathematical abstraction being concern for the content or subject-matter of to remove the meanings of specific terms. M. specific terms is abandoned in favor of the form of Cohen and E. Nagel, for example, present a set the deductive system. So the abstraction involved of axioms for plane geometry; remove all refer- results in essentially the elimination of content. ences to points, lines, and planes; and replace In computer science, content is not totally them with symbols used merely as variables. They abstracted away in this sense. Rather, abstraction then proceed to demonstrate a number of theo- in computer science consists in the enlargement rems as consequences of these new axioms, show- of content. For computer scientists, this allows ing that pure deduction in mathematics proceeds programs and machines to be reasoned about, with terms that have no observational or sensory analyzed, and ultimately efficiently implemented meaning. An axiom system may just happen to in physical systems. For computer users, this describe physical reality, but that is for experi- allows useful objects, such as documents, shop- mentation in science to decide. Thus, again, a ping malls, and chatrooms, to exist virtually in a mathematical or deductive system is abstract by purely electronic space. virtue of a complete stepping away from the con- Understanding abstraction in computer sci- tent of scientific terms: ence requires understanding some of the history of software engineering and hardware develop- Every [deductive] system is of necessity ab- ment, for it tells a story of an increasing distance stract: it is the structure of certain selected rela- between programmers and the machine-oriented tions, and must consequently omit the structure entities that provide the foundation of their work. of other relations. Thus the systems studied in This increasing distance corresponds to a con- physics do not include the systems explored in comitant increase in the reliance on abstract biology. Furthermore, as previously shown, a views of the entities with which the discipline is system is deductive not in virtue of the special fundamentally concerned. These entities include meanings of its terms, but in virtue of the machine instructions, machine-oriented processes, universal relations between them. The specific and machine-oriented data types. The remainder quality of the things which the terms denote of this chapter will explain the role of abstrac- do not, as such, play any part in the system. Thus the theory of heat takes no account tion with regard to these kinds of entities. of the unique sensory qualities which heat Language abstraction. At the grossest physical phenomena display. A deductive system is level, a computer process is a series of changes therefore doubly abstract: it abstracts from the in the state of a machine, where each state is specific qualities of a subject matter, and it described by the presence or absence of elec- selects some relations and neglects others. trical charges in memory and processor elements. (Cohen & Nagel 1953: 138–9) But programmers need not be directly concerned with machine states so described, because they As a final example, consider C. Hempel’s can make use of software development tools assessment of the nature of mathematics while that allow them to think in other terms. For arguing for the thesis of logicism, or the view example, with the move from assembly to high- that mathematics is a branch of logic: level language, computer scientists can abandon

323 Timothy Colburn

for i ← 1 to n do for j ← 1 to m do ReadMatrix(A,n,m) read(A[i,j]) ReadMatrix(B,m,p) for j ← 1 to m do MultiplyMatrices(A,B,C,n,m,p) for k ← 1 to p do read(B[j,k]) for i ← 1 to n do Figure 24.2: Multiplying matrices with procedural for k ← 1 to p do begin abstraction C[i,k] ← 0 for j ← 1 to m do ← + C[i,k] C[i,k] A[i,j] * B[j,k] of view of the higher-level process, these details end are irrelevant; all that is really necessary to invoke the process is the names of the input and output Figure 24.1: Multiplying matrices matrices and their dimensions, given as para- meters to the lower-level procedures. Of course, talk about particular machine-oriented entities the details of how the lower-level procedures like instructions, registers, and word integers in perform their actions must be given in their favor of more abstract statements and variables. definitions, but the point is that these defini- High-level language programs allow machine tions can be strictly separated from the processes processes to be described without reference to that call upon them. What remains, then, is the any particular machine. Thus, specific language total abstraction of a procedure’s use from its content has not been eliminated, as in mathemat- definition. Whereas the language example had ical or deductive systems, but replaced by abstract the abstraction of the content of computer in- descriptions with more expressive power. structions, here there is the abstraction of the Procedural Abstraction. Abstraction of lan- content of whole computational procedures. And guage is but one example of what can be con- again, the abstraction step does not eliminate sidered the attempt to enlarge the content of content in favor of form as in mathematics; it what is programmed about. Consider also the renders the content more expressive. practice of procedural abstraction that arose with Data Abstraction. As a last example, consider the introduction of high-level languages. Along the programmer’s practice of data abstraction. with the ability to speak about abstract entities Machine-oriented data types, such as integers, like statements and variables, high-level languages arrays, floating point numbers, and characters, introduced the idea of modularity, according to are, of course, themselves abstractions placed on which arbitrary blocks of statements gathered the physical states of memory elements inter- into procedures could assume the status of state- preted as binary numbers. They are, however, ments themselves. For example, consider the intimately tied to particular machine architectures high-level language statements given in figure in that there are machine instructions specifically 24.1. It would take a studied eye to recognize designed to operate on them. (For example, that these statements describe a process of filling integer instructions on one machine may operate an n × m matrix A and an m × p matrix B with on 32 bits while similar operations on another numbers and multiplying them, putting the machine may operate on 64 bits.) They are also × Σm result in an n p matrix C such that Ci,k = j=1 built into the terminology of all useful high-

Ai,jBj,k. But by abstracting out the three major level languages. But this terminology turns out operations in this process and giving them pro- to be extremely impoverished if the kinds of cedure names, the program can be written at a things in the world being programmed about higher, and more readable, level as in figure 24.2. include, as most current software applications These three statements convey the same informa- do, objects like customers, recipes, flight plans, tion about the overall process, but with less or chat rooms. machine detail. No mention is made, say, of the The practice of data abstraction is the speci- order in which matrix elements are filled, or fication of objects such as these and all opera- indeed of matrix subscripts at all. From the point tions that can be performed on them, without

324 Methodology of Computer Science reference to the details of their implementation as they analyze their domains for logics, rules, in terms of other data types. Such objects, called classifications, hierarchies, and other convenient abstract data types (ADTs), once they become abstractions. Much of the computational model- implemented, assume their place among integers, ing process will be characterized by activity more arrays, and so on as legitimate objects in the akin to logic and ontology than programming computational world, with their representation per se. (See the chapters in this volume by Smith, details, which are necessarily more machine- Antonelli, Gillies, and Bicchieri on ontology, logic, oriented, being invisible to their users. The result and probability.) Now more than ever, there is is that programs that are about customers, recipes, room for much philosophical research in the flight plans, and so on are written in terms that development of future computational modeling are natural to these contexts, and not in the languages. One might even venture the predic- inflexible terms of the underlying machine. The tion that philosophy will come to be an influen- programs are therefore easier to write, read, and tial tool in the analysis and practice of computer modify. The specification and construction of science methodology. abstract data types are primary topics in under- graduate computer science curricula, as evidenced by the many textbooks devoted to these topics. References But this again is a type of abstraction that does not eliminate empirical content, as in mathem- Carnap, R. 1953. “Formal and factual science.” In atics, but rather enlarges the content of terms H. Feigl and M. Brodbeck, eds., Readings in the by bringing them to bear directly on things in Philosophy of Science. New York: Appleton- a non-machine-oriented world. Century-Crofts, pp. 123–8. [Carnap’s paper is one of a group of seminal papers on the philo- sophy of mathematics collected in this anthology. Conclusion They are central to understanding the distinction between mathematics and science, and the role mathematics plays in science. For the advanced Computer science will always be built on a sci- philosophy student.] entific and engineering foundation requiring spe- Cohen, M. and Nagel, E. 1953. “The nature of a cialists with the most acutely analytic, creative, logical or mathematical system.” In H. Feigl and and technological minds. But the modeling and M. Brodbeck, eds., Readings in the Philosophy abstraction abilities that this foundation provides of Science. New York: Appleton-Century-Crofts, opens the field to virtually anyone willing to learn pp. 129–47. [This paper is from the same group its languages. As computer science grows as a as Carnap 1953.] discipline, its methodology will be less depend- DeMillo, R., Lipton, R., and Perlis, A. 1979. ent on the specialists maintaining the founda- “Social processes and proofs of theorems and tion and more dependent on those able to programs.” Communications of the ACM 22: 271–80. [This paper ignited a debate within the develop, implement, and use high-level languages computer science community regarding the role for describing computational processes. What is of formal methods in computer science meth- meant by a “computational process” has diverged odology, specifically program verification. For the so much from the notion of a machine process intermediate computer science student.] that a currently popular language paradigm, Fetzer, J. 1988. “Program verification: the very namely object-oriented design and programming, idea.” Communications of the ACM 31: 1048– de-emphasizes traditional algorithmic forms of 63. [This paper rekindled the program verification program control in favor of the notions of classes, debate from a philosopher’s point of view and objects, and methods. (See, for example, the caused impassioned responses from the com- current widespread interest in Java.) puter science community. For the intermediate As programming languages evolve, it will philosophy student.] be necessary for software developers to be con- ——. 1991. “Philosophical aspects of program veri- versant in the analytical tools of philosophers fication.” Minds and Machines 1: 197–216. [This

325 Timothy Colburn

paper summarized the wreckage of the program Hempel, C. 1953. “On the nature of mathemat- verification debate, again from the philosophical ical truth.” In H. Feigl and M. Brodbeck, eds., point of view, arguing against the mathematical Readings in the Philosophy of Science. New York: paradigm. For the intermediate philosophy Appleton-Century-Crofts, pp. 148–62. [This student.] paper is from the same group as Carnap 1953.] Floyd, C. 1987. “Outline of a paradigm change Hoare, C. 1986. “Mathematics of programming.” in software engineering.” In G. Bjerknes et al., BYTE, Aug.: 115–49. [A clear statement of the eds., Computers and Democracy: A Scandinavian view that computer science is a species of pure Challenge. Hants., England: Gower, pp. 191– mathematics. For the advanced computer science 210. [This paper was one of the first to come student.] out of the computer science community advoc- McCarthy, J. 1962. “Towards a mathematical ating a view of software as process rather than science of computation.” Proceedings of the IFIP product. For the introductory student in any Congress 62: 21–8. [Another statement of the discipline.] mathematical paradigm for computer science, with Hailperin, M., Kaiser, B., and Knight, K. 1999. an emphasis on the role of recursion in computer Concrete Abstractions: An Introduction to Com- science methodology. For the advanced com- puter Science. Pacific Grove, CA: PWS Publish- puter science student.] ing. [An excellent introduction to computer Naur, P. 1966. “Proof of algorithms by general science that emphasizes the role of abstractions snapshots.” BIT 6: 310–16. [An example of math- in computer science methodology. For the first ematical reasoning about programs in pursuit of time computer science student.] program verification. For the advanced computer science student.]

326 Philosophy of Information Technology

Chapter 25 Philosophy of Information Technology Carl Mitcham

Philosophy of information technology may be of information technology, like the philosophy seen as a special case of the philosophy of techno- of computer science, is properly initiated by the logy. Philosophical reflection on technology aims effort to define that on which it seeks to reflect. in general to comprehend the nature and mean- Once preliminary definitions are negotiated, ing of the making and using, especially of things philosophies of x, often against a historico- made and used. Such reflection nevertheless philosophical background, recapitulate in differ- exhibits a tension between two major traditions: entially weighted forms the main branches of one arising within engineering, another in the philosophy tout court – metaphysics, epistemo- humanities (Mitcham 1994). For the former or logy, and ethics – with particular emphases expansionist view, technology is deeply and com- reflecting both the unique philosophical chal- prehensively human, and thus properly extended lenges of x and the context of presentation. In into all areas of life; according to the latter or the present case, for example, although ethical limitationist perspective, technology is a restricted and political issues play a prominent role in the and properly circumscribed dimension of the philosophy of information technology, they are human. This distinction and corresponding ten- treated lightly here because of the more ex- sions may also be seen at play in the philosophy tensive coverage provided by Chapters 5 and 6 of information technology (IT), between those (COMPUTER ETHICS and COMPUTER-MEDIATED who would critically celebrate and extend IT and COMMUNICATION AND HUMAN–COMPUTER INTER- those who would cautiously subordinate and ACTION). Here the stress is on theoretical issues delimit it. Diverse metaphysical, epistemological, concerning especially metaphysical assessments of and ethical arguments are marshaled to defend information technology. one position over the other, as well as to build bridges between these two philosophical poles. Philosophies of x commonly begin with What Is Information Technology? attempts to define x. Philosophy of science, for instance, logically opens with the , by considering various proposals for Information technology – or such closely related distinguishing science from other forms of terms as “information systems” and “media knowledge or human activity. The philosophy technology” – is commonly described as that

327 Carl Mitcham technology constituted by the merging of data- context, a signal is not a message. Kenneth Sayre processing and telecommunications (with diverse (1976) and Fred I. Dretske (1981) are neverthe- input devices, processing programs, commun- less two important attempts to develop semantic ications systems, storage formats, and output theories of information grounded in the tech- displays). It arose from earlier forms of electronic nical concept of information (see Chapter 17, communications technology (telegraph, tele- KNOWLEDGE). phone, phonograph, radio, motion pictures, “Technology” too is a term with narrow and television) by way of computers and cybernetics broad definitions. In the narrow or engineering (see Chapter 14), an earlier term that still casts sense, technology is constituted by the system- its shadow over IT, as in such coinages as atic study and practice of the making and using “cyberspace” and other cognates. It may never- of artifacts (cf. the curricula of technological theless be useful to begin by attempting to universities), and to some extent by the physical rethink what is perhaps too facile in such a artifacts themselves (from hammers to cars and description. computers). Indeed, a distinction is often drawn The terms “information” and “technology” between premodern techne or technics and mod- are both subject to narrow and broad, not to say ern technology. For thousands of years human engineering and humanities, definitions. Devel- making and using proceeded by intuitive, trial- oped by Claude Shannon (Shannon & Weaver and-error methods, remained mostly small-scale 1949), the technical concept of information is and dwarfed by natural phenomena. With the defined as the probability of a signal being trans- rise of modern methods for making and using, mitted from device A to device B, which can be these activities became systematically pursued mathematically quantified (see Chapters 4 and (often on the basis of scientific theories) and 13, treating INFORMATION and THE PHYSICS OF their products began to rival natural phenomena INFORMATION, respectively). The theory of infor- in scale and scope. In a broader humanities mation opened up by this distinct conceptual parlance, technology covers both intuitive, small- analysis has become the basis for constructing scale and scientific, large-scale making and using and analyzing digital computational devices and in all its modes – as knowledge, as artifact, as a whole range of information (also called com- activity, and even as volition. munication) technologies, from telephones to Given these narrow and broad definitions televisions and the internet. for each element in the compound term, one In contrast to information (and information may postulate a two by two matrix and imagine technologies) in the technical sense is the con- four different information technology exemplars cept of information in a broader or semantic (figure 25.1). In what follows, a significant sample sense. Semantic information is not a two-term from among these possible information technolo- relation – that is, a signal being transmitted from gies will be analyzed in order to illustrate diverse A to B – but a three-term relation: a signal trans- facets of a potentially comprehensive philosophy mitted from one device to another, which is then of information technology. understood as saying something to a person C. Although information technologies in the tech- nical sense readily become information techno- Information Technology in logies in the semantic sense, there is no precise Historico-philosophical relation between technical and semantic informa- Perspective tion. Independent of its probability as a signal, some particular transmission may possess any number of different semantic meanings. A signal Philosophy is not coeval with human thought, in the form of two short clicks or light flashes but emerges from and against prephilosophical (Morse Code for the letter “i”), could be a self- reflection that it nevertheless continues or mir- referential pronoun, part of the word “in,” a rors. Prior to the rise of philosophy, mythological notation in Latin numerals of the number one – and poetic narratives often expressed the ambiv- or any number of other possibilities. Absent the alence of the human experience of tool making

328 Philosophy of Information Technology

Premodern technology Scientific technology

Technical sense of Alphabetic writing Electronic and source code information signal transmission

Semantic sense of Books and related texts Works of high information representational electronic communications media (movies, TV programs, hypertexts, etc.)

Figure 25.1: Information technology exemplars and using. Stories of the conflict between Cain (Phaedrus 274d ff.). Socrates himself comments (builder of cities) and Abel (pastoral shepherd), on the silence of written words, and Plato fam- of Prometheus (who stole fire for humans from ously remarked on the limitations of writing the gods), of Hephaestus (the deformed god of even in his own works (Letter VII, 341b–e and the forge), and of Icarus (the inventor who went 344c–d). The Politicus (300c ff.), however, con- too far) all attest to the problematic character of cludes with a modest defense of written laws, human engagement with what has come to be and the Ion presents the poet as one inspired by called technology. The story of the Tower of the gods. Babel (Genesis 11) even suggests the destructive Aristotle, in an analysis that echoes Plato’s linguistic repercussions of an excessive pursuit of assessment in The Republic of artifice and poetry technological prowess. as thrice-removed from being itself, notes the By contrast, when the prophet Ezekiel learns inability of techne to effect a substantial unity of in the desert to infuse dry bones (alphabetic con- form and matter. “If a bed were to sprout,” says sonants) with the breath of the spirit (unwritten Aristotle, “not a bed would come up but a tree” vowels), it is as if God were speaking directly (Physics II, i, 193a12–16). In a parallel analysis through him (Ezekiel 37). Indeed, God himself of the relation between experiences, spoken creates through speech or logos (Genesis 1), and words, and writing at the beginning of On Inter- writes the law both in stone and in the hearts of pretation, Aristotle places the written word at a people. Thus, information technologies in their two removes from experience and three removes earliest forms – speech and writing – manifest at from the things experienced, thus implying a least two fundamental experiences of the human dilution of contact with reality as one moves condition: sin or hubris and transcendence, the from the information technology of speech to demonic and the divine. that of writing. Spoken words refer to experi- Greek philosophical reflection on techne like- ence; written words to spoken words. wise noted the two-fold tendency of human skill In contrast to Aristotle’s characterization of in the making and using of artifacts to be pur- words in strictly human terms, Christianity re- sued in isolation from the good and to particip- affirms the divine character of the transcendent ate in the divine. This is as true of information word incarnate (John 1) and of the transmission technai, such as oratory and writing, as it is of of the gospel through that preaching which rep- the mechanical and military arts. In Plato’s resents the word (Romans 10:17). Indeed, ac- Gorgias, for instance, Socrates challenges the cording to Augustine, Christian preaching unites sophist to reintegrate the techniques of rhetoric truth and language with an efficacy that the with the pursuit of truth, to eschew the tricks of Platonists could not imagine (De vera religione i, gaining power divorced from knowledge of the 1 ff.). This is an argument that has been revived good. In the Phaedrus, Socrates tells the story of in Catherine Pickstock’s theological interpretation how King Thamus rejected the Egyptian god of that information technology known as liturgy Theuth’s invention of writing on the grounds that (Pickstock 1998). At the same time, the meaning it would replace real with merely virtual memory of the words of revelation in Scripture is not

329 Carl Mitcham always obvious, thus requiring the development associated with a manifold of social transforma- of principles of interpretation (see Augustine’s tions: religious, political, economic, and cultural. De doctrina christiana). Faith in the Scriptures as The philosophical influences of such changes have the word of God solves, as it were, the technical been legion. To cite but one example, as the question concerning the extent to which the world was increasingly filled with texts, and texts signal has been accurately transmitted from A themselves were severed from stable lifeworlds to B (God to humans), but not the semantic of interpreters, philosophy became increasingly question of what this signal means (to whom it linguistic philosophy, in two forms. In contin- speaks and about what). The meaning of rev- ental Europe, hermeneutics was redefined by elation requires a science of interpretation or Friedrich Schleiermacher as the interpretation hermeneutics if its information (from the Latin of all (not just sacred) texts, by Wilhelm Dilthey informare, to give form) is truly to convert those as the foundation of the Geistwissenschaften or who receive it. humanistic sciences, and by Martin Heidegger This dedication to the development of tech- as the essence of Dasein or human being. In this niques of interpretation led to a unique medieval same milieu, Ferdinand de Saussure invented the flowering of logical, rhetorical, and hermeneutic science of linguistics, focusing neither on effi- prowess. Reflecting the effulgence of poetic cient signal transmission nor on multiple levels exegesis of sacred texts, Thomas Aquinas defends of external reference but on language as a sys- the metaphorical “hiding of truth in figures” as tem of words that mutually define one another fitting to the word of God, and argues the power through their internal relations. In the Anglo- of Scripture to signify by way of multiple refer- American world, especially under the influence ences: historical or literal, allegorical, tropolo- of , philosophy became lin- gical or moral, and anagogical or eschatological guistic philosophy, which takes the meaning of (Summa theologiae I, q.1, art.9–10). What is words to be constituted by their uses, thus call- equally remarkable is that – no doubt stimulated ing attention to multiple contexts of use, what by the literal and spiritual interpretations of rev- Wittgenstein called ways of life. Indeed, in some elation as granting the world a certain autonomy forms the resultant philosophy of language turns and calling upon human beings to exercise posit- into a kind of behaviorism or is able to make ive mastery over it – the flowering of semantic common cause with . studies was paralleled by an equally unpreced- In another instance, theories were posited ented blossoming of physical technologies. about the relation between changes in informa- Examples include the waterwheel and windmill, tion technologies and cultural orders. The con- the moldboard plow, the horse collar, the lateen trast between orality and literacy has been sail, and the mechanical clock. elaborated by a series of scholars – from Albert The modern world opens, paradoxically, by Lord and Milman Parry to Marshall McLuhan, pitting the second form of technological progress Walter Ong, and Ivan Illich – who have posited (physical ) against the first (poetic creat- complementary theories about relations between ivity). Metaphorical words are to be rejected in information technology transformations and the pursuit of real things and ever more power- cultures. With McLuhan, for example, there is ful technologies (see especially the arguments of a turn not just from technical signal to semantic and René Descartes). The histor- message, but an attempt to look at the whole ical result was to turn exegesis into criticism and new electronic signal transmitting and receiving semantic analysis into a drive for conceptual technology (never mind any specific semantic clarity, in a reform of the techniques of commun- content) as itself a message. In his own con- ication that became most manifest in the new densed formulation: the medium (or particular rhetoric of modern natural science – as well as in form of information technology) is the message the invention of a whole new information tech- (McLuhan 1964). nology known as moveable type. Stimulated especially by McLuhan, reviews of The invention of the printing press and the the historical influences between philosophy and consequent democratization of reading can be IT begin to mesh into a

330 Philosophy of Information Technology that privileges IT experience the way G. W. F. to the nature and meaning for information Hegel privileged politics and ideas. Here Paul technology. Levinson’s “natural history of information tech- The expansionist approach has its roots in tech- nology” (1997) is a worthy illustration. nical thinking about IT, first in terms of physical entities. At least since Norbert Wiener (1948) effectively posited that, along with matter and energy, information is a fundamental constituent Information Technology and of reality, questions have been raised about the Metaphysics metaphysical status of information. Building on Wiener’s own analysis, distinctions may be drawn Although the historico-philosophical background between three fundamentally different kinds of points to an emergence, in conjunction with technology: those which transform matter (ham- information technology, of new cultural constel- mers and assembly lines), those which produce lations in human affairs, pointing alone is insuf- and transform energy (power plants and motors), ficient to constitute philosophy. Popular attempts and those which transform information (com- to think the new IT lifeworld have emphasized munication systems and computers). economics and politics, in which issues are A related phenomenology of human engage- decided about e-banking and e-commerce on the ment would observe how the being of IT differs basis of market forces and political power. The from tools and machines. Unlike tools (which ethics of information technology, as an initiation do not function without human energy input into philosophical reflection – that is, into and guidance) or machines (which derive energy thought in which issues are assessed on the basis from nonhuman sources but still require human of argument and insight rather than money and guidance), information technologies are in dis- votes – has highlighted issues of privacy, equity, tinctive ways independent of the human with and accountability. Yet given that the funda- regard to energy and immediate guidance; they mental question for ethics concerns how to act are self-regulating (cybernetic). In this sense, in accord with what really is, there are reasons steam engines with mechanical governors on to inquire into the kind of reality disclosed by them or thermostatically controlled heating IT – that is, to raise metaphysical (beyond the systems are examples of information machines. physical or empirical) and ontological (from ontos, Insofar as the operation of more electronically the Greek word for “being”) questions. advanced IT is subject to human guidance, What are the fundamental structures of the guidance ceases to be direct or mechanical and IT phenomenon? What is real and what is is mediated by humanly constructed programs appearance with regard to IT? Richard Coyne (electronically coded plans). What is the onto- (1995), for instance, argues that it is illusory to logical status of programs? What are their rela- view IT as simply a novel instrument available tions to intentions? Indeed, in IT, operation and for the effective realization of traditional projects use appear to have become distinguishable. IT is for conserving and manipulating data. Albert a new species of artifact, a hybrid that is part Borgmann (1999) insightfully distinguishes machine running on its own and part utility struc- between information about reality (science), ture like a road waiting to be driven on – hence information for reality (engineering design), and the term “new media” (as both means and envir- information as reality (the high-definition rep- onment). The static availability of such struc- resentations and creations emerging from IT) – tures is contingent on their semi-autonomous and further the increasing prominence, glamor, dynamic functioning. and malleability of information as reality is hav- Second, in terms of the cognitive capabilities ing the effect of diminishing human engagement of IT, transempirical questions arise about the with more fundamental realities. With regard to extent to which computers (as pervasive elements the kinds of metaphysical issues raised by Coyne, in IT) imitate human cognitive processes. Do Borgmann, and others, it is useful to distinguish computers think? What kind of intelligence is again expansionist and limitationist approaches artificial intelligence (AI)? Are the different kinds

331 Carl Mitcham of AI – algorithmic, heuristic, connectionist, From the Hegelian perspective there is some- embodied, etc. – different forms of intelligence? thing both adolescent and irresponsible about Such ontological questions now blend into an ongoing Socratic negativity that refuses to others, concerning the extent to which high-tech take responsibility for world creation. Indeed, artifacts are different from living organisms. Bio- Socratic negativity easily becomes a philosophic- technology has breached Aristotle’s distinction ally clichéd substitute for true thinking. From between natural tree and artificial bed, growth the Socratic perspective, however, the expansion- and construction, the born and the made. Soon ist approach comes on the scene as a court philo- computer programs may also be able not just to sophy, especially insofar as it flatters the king mimic patterns of growth on the screen (arti- and counsels expanding an already popular and ficial life, see Chapter 15), but autonomous, widely affirmed domain of influence. In a state artificial agents that are able to reproduce them- already dominated by information technology, selves. At the nano-scale, robotic design will the Socratic tradition thus finds expression in hardly be distinct from genetic engineering. Will repeatedly questioning the nature and meaning any differences in being remain? of IT – a questioning that must ultimately go From the technical perspective, information is metaphysical. ubiquitous in both the organic and the artificial At a first level, however, the questioning of worlds. The wall between the two is vanishing, IT will be, as already suggested, ethical. For in- although, insofar as the technical concept of stance, does IT not threaten privacy? Even more information becomes a category of explanation in profoundly, does the IT mediation of human biology, it has also been argued to have distinct action in complex software programs, which are ideological roots (see Kay 2000 on this point). created by multiple technicians and are not even The cyborg (cybernetic organism) is a living in principle able to be fully tested (Zimmerli machine, not a goddess (Haraway 1991). Within 1986), not challenge the very notion of moral such a reality, the ethical imperative becomes ex- accountability? At a second level are political perimenting with ourselves, what Coyne (1995) questions: Is the internet structured so as to calls a pragmatic interaction with advancing promote social justice through equity of access? IT. This is an attitude widely present among Is it compatible with democracy? Furthermore, leading IT designers such as Mark Weiser at IT exists on the back of a substantial industrial the famous Xerox Palo Alto Research Center base, whose environmental sustainability is at (PARC), the ethos of which is commonly celeb- least debatable. Insofar as IT depends on an rated in Wired magazine. It has also have been unsustainable base, might not its own justice and given philosophical articulation by media philo- goodness be compromised? At still a third level sopher Wolfgang Schirmacher. For Schirmacher are psychological questions, blending into epi- (1994), IT is a kind of artificial nature, a post- stemological ones. Does the exponential growth technology in which we are free (and obligated, of information availability not challenge the if we would act in harmony with the new way of human ability to make sense of it? Information being in the world) to live without predetermina- overload or information anxiety (see Wurman tions, playfully and aesthetically. 2001) is one of the most widely cited para- The limitationist approach originates in a dif- doxes of IT life. Finally, at a fourth level are ferent, more skeptical stance. Issues are no doubt psychological-anthropological questions about oversimplified by characterizing one approach as the social implications of the new “mode of pro-IT and another as con-IT – although such a information” (Poster 1990), what it means to contrast captures some measure of real differ- live a “virtual life” (Brook & Boal 1995) and ence (but see Gordon Graham, 1999, for a down- “life on the screen” (Turkle 1995). to-earth philosophical utilization of this contrast The third and fourth dimensions of limitation- using the terms technophiles and neoluddites). ist, Socratic questioning – that is, the epistemo- Perhaps a better contrast would be that of Hegel logical and anthropological levels – hint at the versus Socrates: the comprehensive critical affir- metaphysical. Information technology may hide mation as opposed to the argumentative gadfly. reality from us in a much more fundamental way

332 Philosophy of Information Technology than simply by means of information overload. drive), distributed (by internet), and switched It may deform our being at deeper levels than about (forwarded, reprocessed, data mined, etc.). the psychological. To develop this possibility it Indeed, in another text Heidegger makes the is useful to refer at some length to Martin reference to IT explicit, although under the name Heidegger, the most influential exponent of this of cybernetics. “Cybernetics,” he writes, “trans- position. forms language into an exchange of news. The According to Heidegger’s highly influential arts become regulated-regulating instruments of argument in “The Question Concerning Tech- information” (Heidegger 1977 [1966]: 376). nology” (1977 [1954]), technology is constituted Modern information technology thus does to lan- not so much by machines or even instrumental guage what modern non-information technology means in general as by its disclosure of reality, does to the material world: turns it into Bestand, its unhiding, its truth. Premodern technology in that is, a resource for human manipulation. the form of poiesis functioned as a bringing or What is wrong with this? The basic answer leading forth that worked with nature, and as is that modern technology, including modern such revealed Being as alive with its own bringing information technology, conceals as well as it forth, the way a seed blossoms into a flower or reveals. Insofar as we persist in emphasizing the an acorn grows into an oak tree. Modern tech- revealing and ignore the concealing, concealing nology, by contrast, is not so much a bringing will actually dominate. We will not be fully aware forth as a challenging-forth that reveals the world of what is going on. To develop this point re- as Bestand or manipulatable resource. quires a brief elaboration of Heidegger’s theory In reading Heidegger it is crucial to recognize of hermeneutics. In his version of hermeneutics, that he felt it necessary to couch his insights in a which argues interpretation (more than rational- special vocabulary (“bringing forth,” “challenging ity) as the defining characteristic of the human, forth,” “Bestand”), because of the way ordinary Heidegger makes two basic claims. concepts are sedimented with assumptions that The first is that no revealing (the acquisition themselves help conceal the dimensions of real- of information in the semantic sense) is ever sim- ity to which he invites attention. In Heidegger’s ple; it always involves the process of interpreta- own words: tion. Interpretation itself proceeds in texts, in perception, in thinking, and in life by means of a The revealing that rules throughout modern dialectic between part and whole, what is called technology has the character of a setting-upon, the hermeneutic circle. The part is only revealed in the sense of a challenging-forth. That chal- in terms of the whole, and the whole in terms of lenging happens in that the energy concealed the parts. As a result, Heideggerian hermeneutics in nature is unlocked, what is unlocked is trans- postulates a pregivenness in all revealing or, as formed, what is transformed is stored up, what he also likes to say, unconcealing. Our minds is stored up is, in turn, distributed, and what and our lives open not as with a , but is distributed is switched about ever anew. with an immanent reality (both part and whole) Unlocking, transforming, storing, distributing, waiting to be brought forth into the light of and switching about are ways of revealing. appearances. Understanding proceeds by means (Heidegger 1977 [1954]: 297–8) of a process of moving from part to whole and vice versa, repeatedly to make the implicit expli- To this distinctive way of revealing Heidegger cit, to reveal the concealed, analogous especially also gives a special name: Gestell or enframing. to the ways that premodern technology also Although Heidegger seems to be thinking here worked to till the fields and to fashion hand- of electric power generation, the same description crafted artifacts. The upshot is that not only is would in many ways be applicable to information all information subject to interpretation, but that technology. There is a challenging that happens all information technology is part of a larger when digitally concealed information is unlocked lifeworld and cannot be understood apart from (from, say, a computer disk), transformed (by such an implicit whole. To think otherwise is a some software program), stored up (on a hard metaphysical mistake.

333 Carl Mitcham

The second claim is that any unconcealing is 1966 [1955]L 46). Meditative thinking, pre- at one and the very same time a concealing. This modern and even preclassical Greek philosophical second claim has even more profound implica- thinking, which was once in touch with the root tions for information technology, which through of human existence, and out of which by means its expanding realms makes information more of a narrowing and intensified calculative think- and more omnipresent. Information technology ing has emerged, has been replaced by calculative appears to reveal with a vengeance. According thinking in the form of “all that with which to Heidegger, however, this is ultimately an modern techniques of communication stimulate, illusion – and dangerous to what it means to be assail, and drive human beings” (Heidegger 1966 human. The problem is not just one of sensory [1955]: 48). Technology, especially information or information overload, but of information as technology, conceals this meditative thinking, a concealing of Being itself, the fundamental which Heidegger terms Gelassenheit, releasement nature of reality, of the distinctly human relation or detachment. “Releasement toward things and to such reality. openness to the mystery . . . promise us a new For Heidegger the rise of modern technology, ground and foundation upon which we can stand and its culmination in cybernetics or informa- and endure the world of technology without tion technology, is the culmination of a historico- being imperiled by it” (Heidegger 1966 [1955]: philosophical trajectory of thinking that began 55). The fundamental threat in information tech- with the Greeks. With Plato and Aristotle, Being nology is thus a threat to the human being’s was first revealed, however tentatively and min- “essential nature” and the “issue of keeping imally, as a presence that could be re-presented meditative thinking alive” (ibid.: 56). in thought or rationalized. Over the course of its 2,500-year history, philosophy has success- ively spun off the various scientific disciplines Current Research and Open Issues as specialized ways to re-present the world: in mathematics, in logic, in astronomy, in physics, in chemistry, in biology, in , and now What is most remarkable is the fact that in the interdisciplinary fields of molecular bio- Heidegger’s radical critique of technology in logy, cognitive science, and more. This con- general and information technology in par- tinuing development is the end of philosophy ticular has been subject to significant practical in two senses: its perfection and its termination. appropriations by IT users and designers, thus The very success of scientific revealing grew building bridges between the engineering and out of a specialization of thinking as philo- humanities, the expansionist and limitationist, sophy that entailed leaving behind or concealing traditions in the philosophy of information tech- thinking in a more fundamental sense, some- nology. Raphael Capurro (1986), for instance, thing that Heidegger refers to as Lichtung, brings Heidegger to bear on the field of library translated variously as “lighting” or “opening.” and information science. Hubert Dreyfus (2001) “Perhaps there is a thinking,” Heidegger writes, examines the Internet from a philosophical per- “which is more sober-minded than the incessant spective indebted to Heidegger. With slightly frenzy of rationalization and the intoxicating more expansion, one may also reference two quality of cybernetics” (Heidegger 1977 [1966]: other leading examples: Terry Winograd and 391). Fernado Flores, and Richard Coyne. At the In another text, Heidegger describes this “new same time serious challenges have been raised by task of thinking” at “the end of philosophy” by Mark Poster to the adequacy of a Heideggerian means of a comparison between what he calls approach to IT. calculative and meditative thinking. “Calculative In the mid-1980s, computer scientists thinking never stops, never collects itself. Winograd and Flores argued at length that Calculative thinking is not meditative thinking, Heideggerian analyses could disclose the reasons not thinking which contemplates the meaning behind the failures of information technologies which reigns in everything that is” (Heidegger to function as well in the office as computer

334 Philosophy of Information Technology scientists predicted. In Winograd and Flores References (1987) they argue that Heideggerian insights can thus be a stimulus for redesigning computer Borgmann, A. 1999. Holding On To Reality: The systems. Nature of Information at the Turn of the Millen- A decade later architectural theorist Coyne nium. Chicago: University of Chicago Press. (1995) goes even further, arguing that not just [Distinguishes natural information (about reality), Heidegger but the post-Heideggerian thought cultural information (for constructing reality), and of Jacques Derrida provides a philosophical ac- technological information (information becom- count of what is going on among leading-edge ing a reality in its own right). Seeks to establish information technology designers. Building on guidelines for assessing and limiting information Heidegger’s notion that all revealing involves a as reality.] simultaneous concealing, Derrida proposes to Brook, J. and Boal, I. A. 1995. Resisting the Virtual deconstruct specific concepts, methods, and dis- Life: The Culture and Politics of Information. San ciplinary formations precisely to bring to light Francisco: City Lights. [Twenty-one critical essays on IT inequities, impacts on the body, the degrad- their hidden aspects, that on which they depend ing of the workplace, and cultural deformations.] without knowing or acknowledging it. For Coyne Capurro, R. 1986. Hermeneutik der Fachinforma- this opens the way for and justifies the turn from tion. Freiburg: Alber. [For a short English a commitment to rational method in informa- paper that reviews the thesis of this book, see tion technology design to the renewed reliance R. Capurro, “Hermeneutics and the phenom- on metaphor. enon of information,” Research in Philosophy and Heidegger and Derrida thus revalidate the Technology 19: 79–85.] creative significance of metaphor – of thinking of Coyne, R. 1995. Designing Information Technology a computer operating system as “windows,” of a in the Postmodern Age: From Method to Meta- screen “desktop” with “icons,” even of the mind phor. Cambridge, MA: MIT Press. [Advances in as a computer. It is precisely a play with such information technology are examined from the “irrational” connections that facilitates advances diverse philosophical perspectives of analytic in information technology design. With Aquinas, philosophy, pragmatism, phenomenology, crit- Coyne seeks to defend the metaphorical “hiding ical theory, and hermeneutics, in order to reveal of truth in figures” as functional not just in their different implications for the design and theology but also in technology. Whether either development of new electronic communications Aquinas or Heidegger would counsel such ap- media.] propriation of their philosophies of information Dretske, F. 1983. Knowledge and the Flow of technology is, of course, seriously in doubt. Information. Cambridge, MA: MIT Press. [The As if to reinforce such doubt about such most well-developed theory of perception and creative appropriations, Poster argues at length empirical knowledge based on mathematical that Heidegger “captures the revealing of mod- information theory.] Dreyfus, H. L. 2001. On the Internet. New York: ern technology only, not postmodern techno- Routledge. [A phenomenologically influenced but logy.” Indeed, “some information technologies, interdisciplinary critique.] in their complex assemblages, partake not only Graham, G. 1999. The Internet: A Philosophical of [Gestell] but also of forms of revealing that Inquiry. New York: Routledge. [Questions and do not conceal but solicit participants to a rela- criticizes both neoluddite and technophile claims tion to Being as freedom” (Poster 2001: 32–3). about the dangers and implications of the internet.] For Poster a more adequate approach to the Haraway, D. 1991. Simians, Cyborgs, and Women: philosophical understanding of IT is through The Reinvention of Nature. New York: Routledge. Felix Guattari’s image of the rhizome and a [See especially the “Manifesto for Cyborgs” phenomenology of the enunciative properties of included in this book.] specific technologies. A potentially comprehens- Heidegger, M. 1966 [1955]. Discourse on Thinking, ive philosophy of IT thus remains, not unlike all tr. J. M. Anderson and E. H. Freund. New York: philosophy, suspended in and energized by its Harper and Row. [This includes translation of fundamental alternatives. Heidegger’s essay, “Gelassenheit.”]

335 Carl Mitcham

——. 1977 [1954]. “The question concerning tech- Poster, Mark. 2001. What’s the Matter with the nology,” tr. W. Lovitt. In M. Heidegger, Basic Internet. Minneapolis, MI: University of Minne- Writings. New York: Harper and Row, pp. 287– sota Press. [A critique of applying Heidegger’s 317. [Heidegger’s most important critique of analysis of technology to information technology, technology.] with special reference to postmodern thinkers such ——. 1977 [1966]. “The end of philosophy as Felix Guattari.] and the task of thinking,” tr. J. Stambaugh. In Sayre, K. M. 1976. Cybernetics and the Philosophy of M. Heidegger, Basic Writings. New York: Harper Mind. London: Routledge & Kegan Paul. [Argues and Row, pp. 373–92. [A brief statement a naturalist theory of mind based on mathematical of Heidegger’s philosophy of the history of information theory.] philosophy.] Schirmacher, W. 1994. “Media and postmodern Kay, L. E. 2000. Who Wrote the Book of Life: A technology.” In G. Bender and T. Druckrey, eds., History of the Genetic Code. Stanford: Stanford Culture on the Brink: Ideologies of Technologies. University Press. [A critical assessment of informa- Seattle: Bay Press. [The other contributions to tion as a metaphor in biology.] this book are useful as well.] Levinson, P. 1997. The Soft Edge: A Natural History Shannon, C. and Weaver, W. 1949. The Mathemat- and Future of the Information Revolution. New ical Theory of Communication. Urbana: University York: Routledge. [A new information medium of Illinois Press. [Contains two classic papers: (such as computers) does not so much replace Shannon’s, from the Bell System Technical Journal an old medium (such as the telephone) as (1948); and Weaver’s, from Scientific American complement it.] (1949).] McLuhan, M. 1964. Understanding Media: The Turkle, Sherry. 1995. Life on the Screen: Identity Extensions of Man. New York: McGraw-Hill. [It in the Age of the Internet. New York: Simon & is not the information content of a medium (such Schuster. [A psychologist’s analysis of emerging as speech or television) that is most influential forms of self-definition unique to the internet on a culture, but the character or structure of experience.] the medium itself. Electronic media are structur- Wiener, N. 1948. Cybernetics: Or, Control and Com- ally distinct from, say, books. “The medium is munication in the Animal and the Machine. Cam- the message.”] bridge, MA: MIT Press. [The classic statement Mitcham, C. 1994. Thinking through Technology: of the engineering theory of cybernetics. In other The Path between Engineering and Philosophy. works Wiener also examined the social and ethical Chicago: University of Chicago Press. [A general implications of his theories.] introduction to the philosophy of technology that Winograd, T. and Flores, F. 1987. Understanding distinguishes two major traditions: engineering Computers and Cognition: A New Foundation and humanities philosophy of technology. The for Design. Reading, MA: Addison-Wesley. [A former argues for the expansion, the latter for Heideggerian analysis by two computer scientists.] the delimitation of technology as object, know- Wurman, R. S. 2001. Information Anxiety 2. ledge, activity, and volition.] Indianapolis, IN: Que. [This is the second Pickstock, C. 1998. After Writing: On the Litur- edition of a widely cited critique of information gical Consummation of Philosophy. Oxford, UK: overload by a well-known architect and student Blackwell. [A critique of Derrida and defense of the work of Louis Kahn.] of information as subordinate to the context Zimmerli, W. 1986. “Who is to blame for data created by linguistic and bodily performance in a pollution?” In C. Mitcham and A. Huning, eds., historical tradition.] Philosophy and Technology II: Information Tech- Poster, Mark. 1990. The Mode of Information: nology and Computers in Theory and Practice. Poststructuralism and Social Context. Chicago: Boston: Reidel. [One of 20 original papers from a University of Chicago Press. [Argues that four conference, introduced by an overview of “In- new modes of information – TV ads, data bases, formation technology and computers as themes electronic writing, and computer science – create in the philosophy of technology,” and followed by a world in which humans are socially constituted an annotated bibliography on philosophical stud- differently than in pre-electronic IT history.] ies of information technology and computers.]

336 Computational Modeling

Chapter 26 Computational Modeling as a Philosophical Methodology Patrick Grim

Since the sixties, computational modeling has models that describe the behavior of . . . become increasingly important in both the phys- systems over extended periods of time. (p. 3) ical and the social sciences, particularly in physics, theoretical biology, sociology, and economics. On the other hand, Fritz Rohrlich (1991) and Since the eighties, philosophers too have begun Paul Humphreys (1991), among others, em- to apply computational modeling to questions in phasize the importance of tractability as a logic, epistemology, philosophy of science, philo- motivation for computer modeling. Humphreys sophy of mind, philosophy of language, philo- builds that feature into his working definition: sophy of biology, ethics, and social and political philosophy. This chapter analyzes a selection of interesting examples in some of these areas. A computer simulation is any computer- implemented method for exploring the prop- erties of mathematical models where analytic methods are unavailable. (1991: 501) Computer Models in the Sciences and Philosophy: Benefits and Most authors have not attempted strict definition, Limitations conceding that the notion of a “model” is vague and may even have several distinct senses (Fetzer What qualifies as a computer model or a com- 1999). Still, several important features of mod- puter simulation has itself been subject to philo- els are repeatedly emphasized in the literature: sophical scrutiny, but without clear consensus. (1) models occupy a conceptual role somewhere In a classic statement, T. Naylor (1966) defines between empirical data and traditional theory; computer simulation as (2) modeling represents a wide variety of tech- niques, rather than a single tool; and (3) model a numerical technique for conducting experi- construction itself is part of the “art” of science ments on a digital computer which involves (see especially Rohrlich 1991). There is also certain types of mathematical and logical general agreement on reasons for welcoming

337 Patrick Grim computer modeling in particular: (1) increased The development of complex systems over time mathematical tractability, particularly in under- could hardly have been envisaged at all before standing complex and dynamic processes over the computer, but has now become a topic of time; (2) a methodologically important vividness philoso-phical thought experiment in a wide or graphic immediacy that is often characteristic range of areas. of computer models; and (3) the possibility of Another benefit of computer modeling is that computational “experiment.” This last feature is its methodological demands work as a counter- clear to anyone who has worked with computer force against philosophical vices of imprecision, models, and is noted in almost every outline of vagueness, and obscurity. The environment of computational modeling since Naylor (1966), computer modeling enforces “unflinchingly and quoted above. without compromise, the central philosophical A number of authors portray computer experi- desideratum of clarity: one is forced to construct mentation in general as a technological extension theory in the form of fully explicit models, so of an ancient tradition of thought experiment. detailed and complete that they can be pro- It is this “experimental” aspect of computational grammed ” (Grim, Mar, & St. Denis 1998: 10). modeling that has been seen as a particularly John Pollock has emphasized that one constraint important addition to philosophical methodo- imposed by computer modeling is simply that logy. Kyburg (1998: 37) speaks of a “kind of the theory at issue must actually work the way it philosophical laboratory or testing ground.” is supposed to. “As mundane as this constraint Grim, Mar, and St. Denis (1998: 10) speak of may seem, I am convinced that most epistemo- “an important new environment for philosoph- logical theories fail to satisfy it.” The fact that ical research,” and Bynum and Moor (1998: 6) computer modeling imposes demands of preci- speak of computing as “a medium in which to sion and detail “can have a very therapeutic effect model philosophical theories and positions”: on a profession that is overly fond of hand- waving” (Pollock 1998: 34). Computing and related concepts significantly Another benefit of a computational environ- enhance philosophy by providing a kind of ment is the prospect of exploring possible vari- intellectual clay that philosophers can mold and ations on theory. With a computer model in place, shape and study. Through computing, abstract variations on the model are generally easy. One ideas – which philosophers like to manipulate can explore consequences of epistemological, – can be instantiated and investigated. There biological, or social theories in slightly different is nothing wrong with good armchair reflection environments or with slightly different para- . . . But armchair reflection has its limitations. meters. The result is that the theory with which (1998: 2–3) one begins may be replaced by a variation that appears more promising in action. The exploration of abstract philosophical ideas As has always been true of the interplay by means of computer models offers a number between technology and pure science, it is also of major benefits. One benefit is an astounding possible for the application of philosophical ideas increase in manageable complexity. Although in computational models at the bottom to sug- philosophers have long appealed to thought gest new and intriguing philosophical ideas at experiments, practical necessity has limited these the top. It was computer work in graphing the to our individual human powers of calculation. semantics of infinite-valued paradox, for example, As Bynum and Moor note, “armchair recursion that suggested the proof for a theorem on formal doesn’t recur very many times.” With computer undefinability of chaos (Grim, Mar, & St. Denis models, on the other hand, the computational 1998). Models developed in order to tackle old ceiling is lifted on philosophical imagination. problems may open up new territory for philo- Complex interactions that previously could only sophical exploration as well. A central question be vaguely guessed at can now be calculated with in Hobbes is how cooperation can emerge in a ease, and consequences of such interactions can society of self-serving egoists. Game-theoretic be revealed with a depth previously impossible. attempts to answer that question have been

338 Computational Modeling developed and expanded in computer models That is part of what makes them models. Models, (see Chapter 22, GAME THEORY). Those models like abstract laws, prove useful in both explana- have in turn raised further questions regarding tion and exploration precisely because they are evolution and ethics, rationality and justice, and simpler than the full phenomenon, and thus easier the unpredictability of social behavior in even our to handle and track. We need simple models simplest models (Axelrod 1984, Danielson 1992, because we need a simpler way of understanding Skyrms 1996, Grim, Mar, & St. Denis 1998). complex phenomena, and because we need to Some benefits of computational modeling are separate what is important in what happens from evident even in aspects for which it is occasionally the distracting but unimportant details. criticized. Formal models in general and philo- One can then argue that neither the abstract sophical computer models in particular are bound level nor the simplicity of models requires fur- to be abstract. The high level of abstraction, ther defense. With computational modeling as however, can be seen not as a weakness but as with any methodology, however, there are some an indication of potential power and promise. real intellectual dangers. New methodologies Distinct phenomena that appear in quite dif- always offer new ways of approaching particular ferent contexts – biology and economics, for kinds of questions. There is thus always a tempta- example, or logic and physics – may nonetheless tion to phrase questions in only those ways that have a similar structure or exhibit a similar dy- the new method can handle, or to ask only those namics. The search for patterns that hold across questions that the new method can easily address. different disciplines has characterized new fields We may end up considering only those versions such as chaos theory and artificial life (see Chap- of a theory that can be readily modeled, for ter 15), and promises to be an area in which example, or attending only to those types of philosophical computer modeling could flourish theory that can be modeled at all. The only as well. As Bedau (1998: 135) remarks, known cure for such a danger is to be aware of it. Although we do not want to ignore promis- By abstracting away from the details of chaotic ing new techniques, we must be aware that they systems (such as ecologies, turbulent fluid flow, will inevitably come with limitations. For any and economic markets), chaos science seeks promising new tool, there will always be further fundamental properties that unify and explain questions, equally deep and serious, for which it a diverse range of chaotic systems. Similarly, may not be the best approach. by abstracting away from the details of life-like Computational models carry a more specific systems (such as ecologies, immune systems, danger as well. As models increase in sophistica- and autonomously evolving social groups) and tion in a particular tradition of model-building, synthesizing these processes in artificial media, they are inevitably built out of their simpler pre- typically computers, the field of artificial life decessors. Computational models often incorpor- seeks to understand the essential processes ate earlier passages of code wholesale. Thus, if early shared by broad classes of life-like systems. models in a tradition carry an inessential feature or an unexamined assumption, that feature or One promise of philosophical computer modeling assumption is likely to remain, unquestioned and is highly abstract crossdisciplinary work of pre- uncriticized, throughout later work as well. One cisely this type. of the earliest models to be applied to questions Models in the physical and particularly the in economics, for example, was built using a social sciences are also sometimes met with the quite particular balance of potential gains and objection that they are mere models: that losses: the gains and losses characteristic of the the phenomena being modeled are complex in Prisoner’s Dilemma, discussed further below. ways to which a simple model could not pos- There is now extensive research in theoretical bio- sibly do justice. The same is to be expected as an logy, evolutionary psychology, and philosophy occasional response to philosophical computer using the descendants of that original model. modeling. In reply, it must simply be admitted But it is almost never asked whether the particu- that all models have major limitations built in. lar gains and losses built into the model reflect

339 Patrick Grim realistic assumptions for the specific applications Logic at issue. It must finally be admitted that individual models can certainly fail. Within both philosophy At a fundamental level, computers are logical and the sciences a simple model may turn out to machines: their basic operations can be outlined be too simple, or may be simple in the wrong in terms of standard logical connectives. This ways. It may abstract not from accidental fea- immediately suggests the possibility of turning tures but from fundamentally essential aspects of to computers as tools for extending work in tra- what is being modeled. The possibility always ditional logic. One might expect intensive work remains that a model captures too few aspects of on theorem-proving programs, for example, and the phenomenon, or the wrong ones. That models indeed research of this type has an impressive can go wrong in these ways is grounds for criti- history (see especially the bibliography on logic cism of individual models, of course, but it con- and automated theorem proving at ). What is interesting, however, is that com- short it quite generally suggests a better model. puters have also played a key role in the devel- The following sections emphasize several areas opment of nontraditional logical models. of current exploration in philosophical computer Suppose we start from a set of premises P1, and modeling. Evident in much of this work is a add a few more to create a larger set of premises strong interdisciplinary or cross-disciplinary tend- P2. In traditional logic, anything we could have ency. Modeling techniques developed primarily deduced from premises P1 will also be deducible within physics have been applied within logic from the inclusive set P2: classical logics are (Grim, Mar, & St. Denis 1998); techniques monotonic. Much of everyday reasoning, however, developed within computer science have been seems to be nonmonotonic. If I am told that brought together with traditions in economics Tweety is a bird, I conclude that Tweety can fly. and sociology and applied to questions of ethics If given more information – that Tweety is a and social-political philosophy (Skyrms 1996, bird and a penguin, for example – I may with- Danielson 1992, Grim, Mar, & St. Denis 1998). draw that commitment. The need to handle The comments of Clark Glymour and his collab- “defeasible” or nonmonotonic reasoning of this orators with regard to “android epistemology” kind quickly became evident in attempts at can be applied to philosophical computer model- modeling patterns of reasoning in artificial intel- ing more generally: ligence, and the development of rival approaches is active and ongoing (Pollock 1998, Kyburg The force of this idea can be seen in the way 1998, Gabbay, Hogger, & Robinson 1994). in which it violates all kinds of traditional The fact that computer models can offer vivid disciplinary boundaries in science, bringing images of abstract phenomena is exploited for together engineering and the life sciences, logical purposes in Grim, Mar, and St. Denis placing mathematical linguistics in the heart 1998. Here simple considerations from truth- of electrical engineering and requiring moral table semantics are extended to construct “value philosophers to understand computation the- solids,” which portray in spatial terms the com- ory. University deans, forced to work within binatory operation of connectives within particu- the old hierarchies, weep with frustration, and lar systems (figure 26.1). Spatial representation the work often has to be done in new “inter- disciplinary” – and often undisciplined – of logical properties can make formal relations research centers and institutes which escape immediately apparent: in figure 26.1, for example, the old categories. We live in interesting times. the duality of conjunction and disjunction is (Ford, Glymour, & Hayes 1995: xii) reflected in the fact that the value solid for “and” could be inverted and inserted into that for The rest of this chapter is devoted to some key “or.” Modeling of this sort has also produced examples of philosophical modeling being done some surprises, such as the persistent reappear- in these interesting times. ance of the fractal Sierpinski gasket in a range of

340 Computational Modeling

Figure 26.1: Value solids for AND (left) and OR (right) value solids and the possibility of generating value sentential (figure 26.3). Their purposes go far solids by cellular-automata-like rules. beyond the pedagogical. Traditional logic, as they The fact that computer models can capture portray it, has concentrated on only a “narrow complex dynamics is used by Grim (1993) and slice” of a broader realm of valid information- Grim, Mar, and St. Denis (1998) in work on self- extraction. “In the long run, logic must come to reference and paradox in infinite-valued logics. grips with how people use a multitude of rep- Informally presented, the Liar sentence (“This resentations in rigorous ways. This will force us sentence is false”) seems to produce an alternation to extend and enrich the traditional notions of between truth and falsity: “if it’s true, since it says syntax, semantics, logical consequence and proof it’s false, it must be false . . . but if it’s false, and it . . . In the process, what seemed like a finished says it’s false, it must be true . . .” That dynamics success story in philosophical and mathematical is modeled in the first frame of figure 26.2. analysis will be refashioned in exciting new ways” The authors consider relatives of the Liar within (Barwise & Etchemendy 1998: 112). infinite-valued logics, such as the Chaotic Liar – “this sentence is as true as you think it is false” – which has a dynamics fully chaotic in the math- Epistemology ematical sense, illustrated for slight differences in initial estimated value in the second frame of figure 26.2. One of the goals of epistemology is to under- Computer-modeling prospects for new ap- stand how we come to know. It is related forms proaches to logic are developed in a different way of engineering – epistemology “from the design in John Barwise and Jon Etchemendy’s Hyperproof stance,” “epistemological engineering,” or (1994). In previous work, Barwise and Etche- “android epistemology” – that have produced mendy had developed Tarski’s World as a visual exciting research programs in philosophical com- aid for teaching quantificational logic, with puter modeling. It is clear that all of these will Turing’s World as a similarly visual introduction overlap with the tasks of artificial intelligence to Turing machines. Sensitized from that experi- more generally (see Chapter 9, THE PHILOSOPHY ence to the power and ubiquity of visualization OF AI AND ITS CRITIQUE). Here as elsewhere in in reasoning, Barwise and Etchemendy’s goal in computer modeling, interdisciplinary collabora- Hyperproof is to expand logic beyond its current tion is the rule rather than the exception. ties to sentential syntax to a formalization of John Pollock’s OSCAR Project takes as its information-processing that can exploit various objective “the construction of a general theory forms of representation, diagrammatic as well as of rationality and its computer-implementation

341 Patrick Grim

True

The Liar

False Iterations

True

The Chaotic Liar

False Iterations

Figure 26.2: Dynamics of the Liar and the Chaotic Liar over progressive iterations. The Chaotic Liar is shown for small differences in initial values. in an artificial rational agent” (Pollock 1998: 17). Paul Thagard has used computational The general epistemic task is taken to be that of modeling in pursuing a wide range of issues in systematically updating a set of beliefs towards philosophy of science and epistemology more those which are warranted, and OSCAR is generally (see Chapter 23, COMPUTING IN THE designed to exploit techniques of both defeasible PHILOSOPHY OF SCIENCE). ECHO was developed and deductive logic toward that end. Pollock places as a connectionist computational model of particular emphasis on control of an epistemic explanatory coherence, and Thagard has applied system in terms of interests and goals. In epi- it to a number of examples from the history of stemology, as elsewhere in philosophy, one of the science and in critique of other approaches benefits of computer modeling is the necessity (Thagard 1992, Eliasmith & Thagard 1997). He of making assumptions explicit in a way that also has also attempted to model major aspects of exposes them to fruitful criticism. The fact that analogical thinking using the programs ARCS the program avoids Bayesian probability theory and ACME (Holyoak & Thagard, 1997). The is taken as a point in favor of OSCAR’s rational- deeply interdisciplinary character of much of ity by Pollock, but forms the basis for a number his research is particularly clear in the work on of criticisms in Kyburg (1998). The project is induction (Holland, Holyoak, Nisbett, & described in Pollock (1989) and (1995), and a Thagard 1987), in which Thagard works with current version of OSCAR is downloadable from two psychologists and a computer scientist in . attempting to construct a computational model

342 Computational Modeling

Figure 26.3: A diagrammatic proof from Barwise and Etchemendy’s Hyperproof of scientific reasoning which avoids traditional innatism and empiricism, the character and limits philosophical problems. of the human mind, and even freedom and con- The TETRAD project should also be men- sciousness are now debated with illustrations tioned as a rigorous computational attempt to drawn from competing computational architec- treat epistemological issues regarding causal tures. Work by John Searle, Daniel Dennett, Jerry inference and probability with an eye to issues in Fodor, and David Chalmers features prominently experimental methodology (Spirtes, Glymour, & in the philosophical debate, if not so prominently Scheines 2001). A very different computational in the details of computational modeling. Re- approach explores the complex epistemological search by the Churchlands is noteworthy for dynamics of competing information. Some of the framing contemporary debates in terms of cur- information received by an epistemic agent is rent resources in neurophysiology and computer about the accuracy or reliability of information science, with an emphasis on neural networks sources. In simple models, this can both pro- as models for both the power and the peculiar duce a variety of epistemic chaos and suggest inscrutability of the workings of the human mind maps for the management of epistemic chaos (P. S. Churchland & Sejnowski 1993, P. M. (Grim, Mar, & St. Denis 1998). Churchland 1995). A variety of attempts to approach issues in philosophy of mind using the tools of dynamical systems theory are represented in Robert Port and Timothy van Gelder (1997). Philosophy of Mind, Philosophy Some attempts have recently been made to of Language, and Philosophy apply tools of computational modeling to issues of Biology in the philosophy of language. Structural mapping approaches to analogy and ambiguity appear in It could be argued that the entire field of AI Holyoak and Thagard 1997. They form the core qualifies as computational modeling in philo- of a program designed to generate and interpret sophy of mind. Traditional debates regarding metaphors in Steinhart and Kittay 1994 and

343 Patrick Grim

Steinhart 1995. Computer modeling can also be In 1980, political scientist Robert Axelrod expected to play an important role in the inevit- invited experts in game theory from various fields able conflict between Chomskian models of lin- to submit programs for a Computer Prisoner’s guistic representation and language learning and Dilemma Tournament (Axelrod 1984). Sub- alternative connectionist proposals (McClelland, mitted strategies played 200 games against all St. John, & Taraban 1992). A game-theoretic other strategies, themselves, and a strategy that attempt to understand linguistic convention chose responses at random. The winner of that initiated in Lewis’s Convention (1969) is further tournament was a strategy called “Tit for Tat” developed with the tools of replicator dynamics (TFT). Cooperate with TFT and it will co- in Skyrms’s Evolution of the Social Contract operate with you on the next round. Defect (1996). In this same tradition, simple computa- against TFT and it will defect against you. The tional environments which show emergence of fact that such a cooperative strategy triumphed coordinated signaling behavior are offered as in the first tournament was a surprise. Its con- models for a theory of meaning as use in Grim, tinued success in later tournaments, where its Kokalis, Tafti, & Kilb 2000 and 2002, and Grim, reputation clearly made it the strategy to beat, St. Denis, & Kokalis 2003. was a further surprise. Axelrod and Hamilton Much as AI can be seen as a computational (1981) replaced the tournament competition version of philosophy of mind, Artificial Life with an “ecological model,” which employs the (ALife) can be seen as a computational version replicator dynamics of theoretical biology: more of philosophy of biology (see Chapter 15). Mark successful strategies “breed” to occupy a larger Bedau uses ALife to illustrate both his theory percentage of the population. Here too TFT of life as supple adaptation (1996) and his con- triumphs. The affinity of these results with the sideration of emergent phenomena in biology Hobbesian question of how social cooperation (1997). In time, it seems inevitable that tools can grow in a community of self-serving egoists from AI and ALife will be brought together to is striking, and the contemporary models inevit- answer questions in philosophy of mind and ably drew the attention of philosophers. Could philosophy of biology. Some simple mathematical the triumph of something that looked like altru- models in this direction are offered in Peter ism in this formal model be telling us something Godfrey-Smith 1996. about the dynamics or nature of social coopera- tion and ethics? Brian Skyrms has pursued this cluster of philosophical questions using the tools of computer modeling, adding also techniques Ethics, Social and Political drawn from dynamical systems or chaos theory. Philosophy “Using these tools of evolutionary dynamics, we can now study aspects of the social contract from Computational modeling and ethics might seem a fresh perspective” (1996: p. x). In Evolution an unlikely combination, but these are linked and the Social Contract, Skyrms shows that an in an intriguing interdisciplinary history. Game evolutionary model using replicator dynamics goes theory (see Chapter 22) was developed by von beyond rational decision theory in producing Neumann and Morgenstern (1944) as an attempt particular principles of fair division and a “law of at a mathematical theory applicable to economics mutual aid.” The potentially chaotic dynamics and political strategy. The Prisoner’s Dilemma, of a bargaining game is illustrated in figure 26.4. a two-person game that seems to capture a basic Grim, Mar, & St. Denis (1998) emphasize tension between individual and collective advant- spatialized models of emerging cooperation. age, quickly became a paradigm for work in Figure 26.5, for example, shows a spatialized aspects of economics, theoretical sociology, and conquest by TFT in a field of 8 simple strategies. eventually theoretical biology. Played in terms Nowak and Sigmund (1992) showed that a of “cooperations” and “defections” on each side, greater level of cooperation (“generous TFT,” the Prisoner’s Dilemma has been referred to as which forgives defection against it in 1/3 of the e. coli of social psychology. all cases) triumphs in versions of Axelrod and

344 Computational Modeling

S3

S1 S2

Figure 26.4: A chaotic attractor in game-theoretical dynamics (from Skyrms 1997)

Hamilton’s model that are arguably more real- 1991: 506). The use of computational models istic in incorporating stochastic or probabilistic to demonstrate naturalistic conclusions of this imperfections. Grim, Mar and St. Denis show that sort is contested in Grim, Mar, and St. Denis spatialization of such models results in an even 1998. They offer as a counterexample the suc- higher level of cooperation. They also consider a cess of certain discriminatory strategies, which version of the spatialized model with “fuzzy” play TFT with others of their own color but values of intermediate cooperation and defec- always defect against outsiders. Strategic success tion, show a formal undecidability result for the cannot simply be identified with morality, they Spatialized Prisoner’s Dilemma, and take some argue, since discriminatory strategies are clearly first steps toward applying the model to ques- successful in such models, but it is clear that tions of racial discrimination. analogous racial discrimination is morally wrong. Peter Danielson characterizes his work in this How social strategies may develop or propagate tradition as “artificial morality,” intended to com- is one question; whether they should be judged bine game theory and artificial intelligence in the as genuinely ethical is another. development of “instrumental contractarianism” Powerful new tools are now available for as an ethical theory. Building on the work of further research in this general tradition. TIERRA David Gauthier’s Morals by Agreement (1986) and the later AVIDA are ALife platforms that and constructing a range of PROLOG models, may be customized to pursue questions in both Danielson’s attempt is to show at least that it philosophy of biology and social and political can be rational to be moral. Danielson seems philosophy. TIERRA is available by anonymous more willing than other researchers in the area, ftp from . A good intro- however, to use modeling as an argument for duction to AVIDA, packaged with the software, something more: for a version of ethical natural- is Adami, Introduction to Artificial Life (1998). ism in which morality simply is that strategy that SWARM is a powerful general platform for agent- proves successful. In some forms of such a view, based modeling, developed by members of the such as Michael Ruse’s Darwinian naturalism, Santa Fe Institute as a way of offering researchers the conclusion is that morality is something other in various fields a powerful “lingua franca” for than what we have thought it to be: “Morality computational experimentation. The program is no more than a collective illusion fobbed off can be downloaded at . A on us by our genes for reproductive ends” (Ruse good introduction to this platform is Luna and

345 Patrick Grim

Figure 26.5: Progressive conquest by Tit for Tat, shown in black, in an array of 8 simple strategies. TFT eventually conquers the entire field.

Stefansson (2000), which offers a SWARM in- been active model-borrowing between different stantiation for the Spatialized Prisoner’s Dilemma research programs and a flourishing interdiscip- as one of its opening examples. linary awareness. The application of computa- tional modeling to philosophical questions has just begun, and first returns are promising in a number of areas. Conclusion It must be recognized that novel techniques carry some intellectual risks. Powerful new Computational modeling offers not a single sys- methods inevitably draw attention to those ques- tematized method but an enormous toolbox of tions, or those forms of questions, for which the potential models and techniques. A number of methods hold out the most promise. In philo- basic modeling tools – the tools of dynamical sophy as elsewhere it must be remembered that systems, neural nets, and cellular automata, for the new techniques should take their place among example – have found application across the a range of traditional tools for approaching a physical and social sciences. The result has range of perennial questions. It is also important

346 Computational Modeling to beware of fixing on a few simple models too Bedau, M. 1996. “The nature of life.” In M. Boden, early. Continuing development of new variations ed., The Philosophy of Artificial Life. Oxford: is important in order to avoid narrow assump- Oxford University Press, pp. 331–57. [Bedau’s tions and to facilitate wide exploration. theory of life as supple adaptation illustrated in Progress in philosophical computer modeling terms of artificial life. Suitable for any interested is already proceeding so swiftly that any over- reader.] view is bound to be partial and incomplete. In ——. 1997. “Weak emergence.” In J. Tomberlin, this chapter, the attempt has been to emphasize ed., Philosophical Perspectives: Mind, Causation, the general promise of such an approach by high- and World (vol. 11). Oxford: Blackwell, pp. 375– 99. [One of the best philosophical treatments of lighting a few examples of intriguing and note- the notion of emergent phenomena. Requires worthy current work. some philosophical background.] ——. 1998. “Philosophical content and method of artificial life.” In Bynum & Moor 1998: 37–47. Acknowledgments [A good general discussion of artificial life as computational philosophy of biology. Under- I am grateful to Jason Alexander, Mark Bedau, graduate and above.] Terrell Bynum, John Etchemendy, James Fetzer, ——. 1999. “Can unrealistic computer models Henry Kyburg, Paul Humphreys, James Moor, illuminate theoretical biology?” In A. Wu, ed., Brian Skyrms, and Eric Steinhart for helpful Proceedings of the 1999 Genetic and Evolutionary input, and to members of the Group for Logic Computation Conference Workshop Program, & Formal Semantics and Luciano Floridi for pp. 20–3. Available on-line at . [An argument for the vir- tues of abstract modeling. Suitable for any inter- ested reader.] Burkholder, L. 1992. Philosophy and the Computer. References Oxford: Westview Press. [A valuable anthology of computer-assisted philosophical work earlier Adami, C. 1998. Introduction to Artificial Life. New than Bynum & Moor 1998. Different contribu- York: Springer-Verlag. [A good introduction to tions at different levels of difficulty.] the AVIDA program for artificial life. Advanced Bynum, T. and Moor, J., eds. 1998. The Digital undergraduate and above.] Phoenix: How Computers are Changing Philosophy. Axelrod, R. 1984. The Evolution of Cooperation. Oxford: Blackwell. [An important anthology of New York: Basic Books. [A major source for work in computer modeling, from which several game-theoretic modeling regarding competition other references are drawn. A good introduction and cooperation, and a good introduction for for anyone.] any interested reader.] Churchland, P. M. 1995. The Engine of Reason, the —— and Hamilton, W. 1981. “The evolution of Seat of the Soul. Cambridge, MA: MIT Press. [A cooperation.” Science 211: 1390–6. [Application philosophically informed review of current work of replicator dynamics to earlier game-theoretic in neural nets, including also neurophysiology. work regarding competition and cooperation. Suitable for any interested reader.] Technical but clear.] Churchland, P. S. and Sejnowski, T. 1993. The Barwise, J. and Etchemendy, J. 1994. Hyperproof. Computational Brain. Cambridge, MA: MIT Stanford, CA: CSLI and Cambridge University Press. [A philosophically informed review of Press. [A system of computer-enhanced logic. current work in neurophysiology, including also Advanced undergraduate and above.] work on neural nets. Occasionally technical but —— and ——. 1998. “Computers, visualization, clearly written.] and the nature of reasoning.” In T. Bynum and Danielson, P. 1992. Artificial Morality: Virtuous J. Moor, eds., The Digital Phoenix. Oxford: Robots for Virtual Games. New York: Routledge. Blackwell, pp. 93–116. [Working reflections on [Uses formal game-theoretic modeling to argue computer prospects for new approaches to logic. for a form of ethical naturalism. Undergraduate Undergraduate and above.] and above.]

347 Patrick Grim

Eliasmith, C. and Thagard, P. 1997. “Waves, par- ical Computer Modeling. Cambridge, MA: MIT ticles, and explanatory coherence.” British Journal Press. [A rich sampler of philosophical computer for the Philosophy of Science 48: 1–19. [An applica- modeling by a single research group, including a tion of ECHO to the history of the wave- CD-ROM with animations and all source code. particle debate in physics. Some background in Undergraduate and above.] physics required.] ——, St. Denis, P. and Kokalis, T. 2003. “Learn- Fetzer, J. 1999. “The role of models in com- ing to communicate: the emergence of signaling puter science.” The Monist 82: 20–36. [A critical in spatialized arrays of neural nets.” Group for examination of various approaches to models. Logic & Formal Semantics, Dept. of Philosophy, Advanced undergraduate and above.] SUNY Stony Brook, Research Report no. 01-01, Ford, K., Glymour, C., and Hayes, P. 1995. forthcoming in Adaptive Behavior. [A neural net Android Epistemology. Menlo Park, CA: AAAI model showing emergence of simple patterns of Press/MIT Press. [An anthology of philosoph- communication in order to capture food and ical reflections on epistemological modeling. avoid predation. Advanced undergraduate level Undergraduate and above.] and above.] Gabbay, D., Hogger, C., and Robinson, J. 1994. Holland, J., Holyoak, K., Nisbett, R., and Thagard, Handbook of Logic in Artificial Intelligence P. 1987. Induction: Processes of Inference, Learn- and Logic Programming, vol. 3: Nonmonotonic ing, and Discovery. Cambridge, MA: MIT Press. Reasoning and Uncertain Reasoning. Oxford: [A better philosophy of science through com- Clarendon Press. [An overview of formal work in puter modeling. Graduate and above.] defeasible and nonmonotonic logics. Very tech- Holyoak, K. and Thagard, P. 1997. “The analo- nical, graduate and above.] gical mind.” American Psychologist 52: 35–44. Gauthier, D. 1986. Morals by Agreement. Oxford: [An attempt to model reasoning by analogy. Ad- Oxford University Press. [A contractarian model vanced undergraduate and above.] important for Danielson’s “artificial morality” Humphreys, P. 1991. “Computer simulations.” models. Undergraduate and above.] In A. Fine, M. Forbes, and L. Wessels, eds., PSA Godfrey-Smith, P. 1996. Complexity and the Func- 1990, vol. 2: 497–506. East Lansing, MI: Philo- tion of Mind in Nature. Cambridge: Cambridge sophy of Science Association. [A philosophical University Press. [Mathematical models relevant examination of computer models in the sciences. to both philosophy of mind and philosophy of Occasionally technical, advanced undergraduate biology. Advanced undergraduate and above.] and above.] Grim, P. 1993. “Self-reference and chaos in fuzzy Kyburg, H. 1998. “Epistemology and computing.” logic.” IEEE Transactions on Fuzzy Systems 1: In Bynum & Moor 1998: 37–47. [Includes 237–53. [Dynamical models for the behavior of criticism of Pollock’s OSCAR project. Suitable self-referential sentences in fuzzy logic. Some- for any interested reader.] times technical, graduate and above.] Luna, F. and Stefansson, B., eds. 2000. Economic ——, Kokalis, T., Tafti, A., and Kilb, N. 2000. Simulations in Swarm: Agent-based Modeling and “Evolution of communication in perfect and Object Oriented Programming. Boston: Kluwer imperfect worlds.” World Futures 56: 179–97. Academic. [A good introduction to a new plat- [A spatialized model in which simple patterns of form for social, and biological experimentation communication emerge as behavioral coordina- developed by the Santa Fe Institute and tion within a community. Advanced undergradu- downloadable at . Different ate and above.] elements accessible at advanced undergraduate ——, ——, ——, and ——. 2002. “Evolution of and graduate level.] communication with a spatialized genetic algo- McClelland, J. L., St. John, M., and Taraban, R. rithm.” Evolution of Communication 3: 105–34. 1992. “Sentence comprehension: a parallel dis- [A genetic algorithm model showing emergence tributed processing approach.” In L. Burkholder, of simple patterns of communication in order to ed., Philosophy and the Computer. Oxford: capture food and avoid predation. Advanced Westview, pp. 34–56. [An opening salvo in the undergraduate and above.] growing controversy between Chomskian lin- ——, Mar, G., and St. Denis, P. 1998. The Philo- guistics and connectionist models. Suitable for sophical Computer: Exploratory Essays in Philosoph- any interested reader.]

348 Computational Modeling

Naylor, T. 1966. Computer Simulation Techniques. Skyrms, B. 1996. Evolution of the Social Contract. New York: Wiley. [An early text, now of mostly Cambridge: Cambridge University Press. [A historical interest. Advanced undergraduate.] seminal source of formal game-theoretic model- Nowak, M. and Sigmund, K. 1992. “Tit for tat ing relevant to questions in ethics and social and in heterogeneous populations.” Nature 355: political philosophy. Undergraduate and above.] 250–52. [Introducing stochastic imperfection into ——. 1997. “Chaos and the explanatory significance game-theoretic models of cooperation. The result: of equilibrium: strange attractors in evolution- greater generosity. Advanced undergraduate and ary game dynamics.” In C. Bicchieri, R. Jeffrey, above. and B. Skyrms, eds., The Dynamics of Norms. Pollock, J. 1989. How to Build a Person: A Prolego- Cambridge: Cambridge University Press. [Game- menon. Cambridge, MA: MIT Press. [Problems theoretical dynamics with an emphasis on chaos. faced and lessons drawn from the OSCAR project. Sometimes technical, graduate and above.] Undergraduate and above.] Spirtes, P., Glymour, C., and Scheines, R. 2001. ——. 1995. Cognitive Carpentry. Cambridge, MA: Causation, Prediction, and Search, 2nd ed. Cam- MIT Press. [Problems faced and lessons drawn from bridge, MA: MIT Press. [A sophisticated attempt the OSCAR project. Undergraduate and above.] to model causation and prediction well enough ——. 1998. “Procedural epistemology.” In Bynum to improve research design. Technical, graduate & Moor 1998: 17–36. [An outline of the and above.] OSCAR project. Undergraduate and above.] Steinhart, E. 1995. “NETMET: a program for Port, R. and van Gelder, T., eds. 1997. Mind as generating and interpreting metaphors.” Com- Motion: Explorations in the Dynamics ff Cogni- puters and Humanities 28: 383–92. [An attempt tion. Cambridge, MA: MIT Press. [An anthology at programming semantic “fields.” Undergradu- of work modeling psychological processes in terms ate and above.] of dynamical systems theory. Different contribu- —— and Kittay, E. 1994 “Generating metaphors tions at different levels.] from networks.” In J. Hintikka, ed., Approaches Rohrlich, P. 1991. “Computer simulations in the to Metaphor. Dordrecht: Kluwer Academic, physical sciences.” In A. Fine, M. Forbes, and pp. 41–94. [An attempt at programming semantic L. Wessels, eds., PSA 1990, vol. 2: 507–18. East “fields.” Undergraduate and above.] Lansing, MI: Philosophy of Science Association. Thagard, P. 1992. Conceptual Revolutions. [A philosophical examination of computer models Princeton: Princeton University Press. [The in the sciences. Graduate and above.] ECHO program applied to examples from the Ruse, M. 1991. “The significance of evolution.” In history of science. Undergraduate and above.] P. Singer, ed., A Companion to Ethics. Oxford: Von Neumann, J. and Morgenstern, O. 1944. Theory Blackwell, pp. 500–510. [A particularly strong of Games and Economic Behavior. Princeton, NJ: version of ethical naturalism. Suitable for any Princeton University Press. [The seminal work interested reader.] on game theory. Graduate and above.]

349 Index

Index

Note: Abbreviations used in the index: AI = artificial intelligence; ALife = artificial life; CMC = computer- mediated communication; GDI = General Definition of Information; HCI = human–computer interaction; IT = information technology; MTC = mathematical theory of communication; SDI = Special Definition of Information; VR = virtual reality.

AARON, digital art, 114 android epistemology, 310, 340 abductive inference, 308, 310 anthropology, IT, 332–3 abstract data types (ADTs), 325 antireductionist theories, information, 40, 41 abstraction, 242, 245, 246, 322–5, 339 Aquinas, Saint Thomas, 330, 335 accelerating Turing machines (ATMs), 183–4 Aristotle ACME, 309, 342 hypertext, 258 action theory, ethics, 67–8 internet culture, 100–1 adaptive evolution, 202–5, 207 IT, 329 adequacy conditions, content theory, 216, ontology, 155, 156 217–18, 219–20, 221–2 syllogism, 263 adequatist ontology, 155, 156 Arnold, V. I., 31 agent-based models, ALife, 198–9 art agonistic work, internet culture, 103, 104 automatic, 114 alethic neutrality (AN), 45–6, 54 digital see digital art Algol, 238 artificial intelligence (AI), 119–33 Algol-like languages, 238–9 ALife, 198–9 algorithms Chinese Room, 26, 122 ALife models, 199 closed systems, 127–8 cognitive science, 191 communication, 131–2 computational complexity, 18–19 complexity–philosophy relation, 26 computational theory of mind, 136–8, 140 computer modeling, 343 computer science methodology, 320 connectionism, 128–9, 133 modeling scientific discovery, 308 see also networks, connectionist analog computers, 178, 184 and cybernetics, 190–3 Analogical Constraint Mapping Engine, 309, data mining, 311 342 eliminative materialism, 124 analogy, cognitive modeling, 309, 342 engineering, 307–8, 310–11, 313, 314

350 Index folk psychology, 123–4 symbol systems, 14–15, 121, 125, 130, 133, formal systems, 125, 126, 127 191, 252 frame problem, 128 syntax processing, 124–6, 127 HCI, 78 teleological activity, 191 hermeneutic critique, 131 Turing machines, 14–15, 119, 120–1 history, 119 Turing test, 120, 122, 139, 209, 252 Holy Grail of, 138–9 weak, 122–3 hypertext, 252–3 artificial life (ALife), 197–210 intelligent machines, 133 adaptationism, 202–5 language machine metaphor, 252–3 computational methodology, 198–201, 210 language of thought, 126–7 computer modeling, 344, 345 mentalese, 126–7 emergence, 201–2 and the mind see mind, and AI; mind, evolution, 202–6, 207, 208–9 computational theory of nature of life, 206–7 modeling strategies, 198–9 philosophical methodology, 209–10 ontology, 160 roots, 197–8 open systems, 128 strong, 207–8 other minds, 132–3 weak, 208 parallel processing, 26, 129, 192 artificial neural networks, 142, 189 philosophy of science, 307–8, 310–11, 313, Ashby, W. R., 190 314 Ashmore, M., 254–5 physical machines, 120–1 attractors probability in, 128, 276–87 ALife models, 204 Bayes theorem, 279, 287 systems science, 31, 32–3, 34 Bayesian networks, 279–82, 284, 285–6 autoepistemic logics, 273–4 Bayesianism, 278–9, 282–5, 287 automata causality, 282, 283, 286, 287 cellular, 182, 183, 197, 198, 200 certainty factors, 278 cybernetics, 187, 188 conditional independence assumptions, 279, automatic art, 114 280–1, 282, 285–6 automatic control devices, 187 expert systems, 276–9, 282 automatic formal system, mind as, 139–40 fuzzy logic, 278, 286 automatons, finite state, 4, 138 inference networks, 278, 279–80, 311 autonomy, cybernetics, 194 interpretation problem, 283–6 AVIDA, 345 machine learning, 286, 311 Axelrod, R., 344 MYCIN, 276–7, 278 neural networks, 286 back-propagation objective probabilities, 277, 284, 285, 287 ALife models, 199 PROSPECTOR, 278–9 network training, 144–5, 192, 199 scientific discovery programs, 310–11 backward induction, 296–7 scientific theory evaluation, 311 BACON, 308 subjective probabilities, 277, 278, 284, 285, balance dystopianism, 94–5, 103 287 balance utopianism, 96–9, 103 terminology, 286–7 Bar-Hillel, Y., 53–4 quantum processes, 127 Barwise, J., 341 robotics, 192–3 Bateson, G., 44, 256, 258 scientific discovery programs, 310–11 Baudrillard, J., 173 scientific theory evaluation, 311 Bayes theorem, 279, 287 semantic engines, 125–6, 129, 133 game theory, 297–8 semiotic systems, 129–33 Bayesian models serial processing, 26 scientific discovery, 310–11 strong, 122, 123 scientific theory evaluation, 309, 311

351 Index

Bayesian networks, 279–82, 284, 285–6 connectionism, 129, 133, 142, 189 Bayesianism, 278–9, 282–5, 287 cybernetics, 189, 192 Bedau, M., 339, 344 folk psychology, 124 being, nature of, 167 inherence-utopianism, 96 IT, 329–30, 331–4 systems science, 35, 36 see also ontology; virtual reality Brey, P., 69 belief fixation, holism of, 219, 223 Brooks, R., 192 Benacerraf, P., 245–6 Buchanan, B. G., 277, 278, 310 Bénard experiment, 34 Bush, V., 250–1 Bernstein, R., 100–2 butterfly effect, systems, 30–1, 35, 36 Bertalanffy, L. von, 194 Bylander, T., 312 best-match problems, 145–6 Bynum, T., 338 Bigelow, J., 187–8 Bigg, S., 110–11 calculation, culture of, 98 binding problem, perception, 35 capitalism, VR, 175–6 Binkley, T., 111–12 Capurro, R., 334 Binmore, K., 291 car drivers, 36 biology cardiology, 35 ALife, 198, 199–201 Carnap, R., 53–4, 157, 322–3 adaptationism, 202–3, 204–5 Cartesian dualism computer modeling, 344, 345 and AI, 119 cybernetics, 187, 193 Chinese Room, 122 discovery programs, 308, 310 connectionism, 129 evolutionary game theory, 299, 302 functionalism, 192 information, 57 hermeneutic critique, 131 systems science, 35, 36, 37 semiotic systems, 129 theory evaluation, 311 Turing test, 120 Birkerts, S., 94–5 CMC, 78, 79–80, 83–4 Boden, M., 115 cybernetics, 188 bodies, recursive physics, 184–5 VR, 173 body, the category theory, 245–6 cybernetics, 187, 188 causality, AI, 282, 283, 286, 287 hypertext, 252–3, 255, 259 cautious monotony, 267, 268, 272, 274 mind–body dualism see Cartesian dualism cellular automata (CAs), 182, 183 Boltzmann machine, 146 ALife, 182, 197, 198, 200 Boolean logic, infomania, 94 certainty factors, 278 Boolean networks ceteris paribus generalizations, 221, 222, 223 ALife, 198, 204–5 Chalmers, D., 132–3, 343 complexity theory, 22, 25 chaotic dynamics, modeling, 344 Boolos, G. S., 12 Chaotic Liar dynamics, 341 Borgmann, A., 100, 331 chaotic systems, 30, 31, 32–3, 34, 36, 37, brain 128 and AI Charniak, E., 123, 128 connectionism, 129, 133 Cheeseman, P., 310–11 eliminative materialism, 124 chemistry, discovery programs, 310 folk psychology, 124 Chen, M., 311 parallel processing, 26, 129 chess problem, complexity theory, 20, 21 semiotic systems, 129 Chinese Room, 26, 27, 122, 141 Church–Turing thesis, 13–14, 138 Chomsky, N., 127 complexity–philosophy relation, 26–7 Christianity, 329–30 computational powers, 178, 184–5 Church, A., 9–10, 266 computational theory of mind, 135 see also Church–Turing thesis

352 Index

Church–Turing (Turing) thesis, 7, 119 complexity, 18–27 computational theory of mind, 138 Boolean networks, 22, 25 first-order logic, 266 Chinese Room, 26, 27 misunderstandings, 10–15 classification of problems, 20 Churchland, P. M. cryptosystems, 23 Church–Turing thesis, 10, 14 decision problems, 19–20 computer modeling, 343 hierarchies of problems, 21–2 cosmic network, 26 NP-completeness, 23–4 folk psychology, 124 parallel computation, 24–5 intentional irrealism, 215 definitions, 18, 19–20 scientific theory evaluation, 310 efficient reduction technique, 21 Churchland, P. S., 10, 14, 26, 343 exponential space algorithms, 20, 21 circumscription, nonmonotonic logic, exponential time algorithms, 19, 20, 21 269–70 generalized chess problem, 20, 21 closed systems, AI, 127–8 goal, 18 closed world assumption, 269 hierarchies of problems, 21–2 coding theory, MTC, 50–1 hypothesis evaluation, 312 cognition imitation game, 25–6 complexity–philosophy relation, 26–7 informational, 57 connectionist modeling, 146–7 Kolmogorov, 19, 57 systems science, 36 ladder of, 205–6 see also artificial intelligence linear time algorithms, 18, 19 cognitive functions, 136–8 map coloring problem, 19, 23 cognitive processes, simulation, 191 nonclassical propositional logics, 22 cognitive science, 308 nonmonotonic logics, 274 algorithms, 191 NP-complete problems, 19, 22–4 cognitive modeling, 307, 308–10, 311, oracle machines, 23–4, 25 313–14 palindrome problem, 20 inherence dystopianism, 99 parallel computation, 24–5, 26 Cohen, H., 114 and philosophy, 25–7 Cohen, M., 323 polynomial size circuits, 25 coherence theories, 309–10, 312, 342 polynomial space algorithms, 20, 24 colonialism, CMC, 80 polynomial time algorithms, 18–19, 20, 22, colonization, internet culture, 97 23–4 color perception, 146–7 quadratic time algorithms, 18, 19 commercialization, 82–3, 176 quantified propositional logic, 24 commodification, 82–3, 176 reducibility of problems, 21 common knowledge, game theory, 291–2, 293, satisfiability problem of propositional logic, 19, 294, 296–7 20, 22–3 communication serial computation, 26–7 AI, 131–2 space (computer storage), 18, 19–20, 21, 23–4 computer-mediated, 76–84 space constructible functions, 21 hypertext, 81, 251, 252–9 space hierarchy theorem, 21 internet, ethics, 72, 73 of systems see systems science see also communication, computer-mediated time (computation steps), 18–20, 21, 22–4 mathematical theory of, 46–53 true sentences of number theory problem, 21–2 telepresence, 169 Turing machines, 19–20, 23–5 communication technologies see information weak second-order theory of one successor, 22 technology word problem for commutative semigroups, 21 compactness theorem, 265 computable numbers complete information, game theory, 290 digital physics, 184 completeness theorem, 265–6 Turing machines, 4, 8

353 Index computation, 3 personhood, 78–80, 81–2, 83, 256–9 ALife, 209 worldview, 76, 77–84 Church–Turing thesis, 7, 10–15, 138 commercialization, 82–3 complex systems, 36–8 commodification, 82–3 complexity theory, 18–27 education, 84 human, 6–7, 11, 13–14, 138 epistemology, 78–82, 83–4 physics of, 183–4 globalization, 82–3 Turing machines, 3–15, 138 hypertext, 81, 257–8, 259 computational modeling logic, 81–2 computer science methodology, 318–25 ontology, 78–80, 83–4 abstraction, 322–5 personhood, 78–80, 81–2, 83 debugging, 320–1 politics, 80, 82, 84 formal verification, 320–2 semiotics, 80–1 insertion sort, 318 uncovering, 83–4 mathematics, 319–23 computer programs see programs social processes, 321 computer science methodology, 318–25 specification, 320 computer simulation see computational as philosophical methodology, 337–47 modeling abstraction, 339 computer-supported cooperative work (CSCW), benefits, 337–9, 346 77–8, 79, 80, 83, 84 biology, 344, 345 computing in philosophy of science, 307–14 definitions, 337–8 cognitive modeling, 307, 308–10, 311, epistemology, 341–3 313–14 ethics, 344–5 engineering AI, 307–8, 310–11, 313, 314 language, 343–4 open problems, 314 limitations, 339–40, 346–7 theoretical approaches, 307, 311–13 logic, 340–1 conceptualizations, ontology, 161–3 mind, 343 conditional independence, AI, 279, 280–1, 282, political, 345 285–6 social, 344–5 connectionism philosophy of science, 307–14 AI, 128–9, 133 cognitive modeling, 307, 308–10, 311, ALife models, 199 313–14 computational theory of mind, 129, 133, 135, engineering AI, 307, 310–11, 313, 314 137, 142–8 open problems, 314 cybernetics, 189–90 theoretical approaches, 307, 311–13 connectionist networks, 142–7, 189–90 computational systems, 36–8 hypothesis evaluation, 309–10, 342 computational theory of mind see mind, consciousness computational theory of Church–Turing thesis, 13, 14 computer art see digital art human becoming, 257 computer ethics see ethics, computer inherence dystopianism, 99 computer languages, 139–40 consequence relations, logic, 264–5, 266–8, abstraction, 242, 245, 246, 323–4 272–3 computer science methodology, 320, 323–4, conservative (Hamiltonian) systems, 29–30, 31, 325 35–6 semantics see semantics, computer languages consistency, logic, 266 computer-mediated communication (CMC), content and information, 42–5, 215–23 76–84 adequacy conditions, 216, 217–18, 219–20, definition, 77 221–2 future research uncovering, 83–4 asymmetric dependence, 221–3 hypertext, 81, 251, 252–9 epistemic optimality, 218–19 interdisciplinary dialogue, 83–4 grain, 216, 217–18, 219–20, 221

354 Index

information flow, 216–18 Danielson, P., 345 knowledge, 230–1 Darden, L., 308 misrepresentation, 216, 218, 219–20, 221–2 data abstraction, 324–5 naturalism, 215, 216, 217, 218–19, 220, 222, data coding, MTC, 46–52 223 data and information ontological naturalism, 215 GDI, 42–6 teleology, 219–20 MTC, 46–53 Conway, J., 182, 200, 208 SDI, 46, 54 cooperative games, 290, 344–5 semantic approach, 54–5 Cope, D., 114 data mining, 311 Copeland, B. J., 183 data privacy, 70–1 counterfactual-supporting generalizations, 221, data production, world rate, x–xi 222, 223 data transmission, MTC, 46–52 Coyne, R., 331, 332, 335 Davidson–Dummett program, 237, 246 crackers, 71–2 Dawkins, R., 203 Craik, K., 189 de Dombal, F. T., 277, 284 creativity, digital art, 113–15 debugging, 320–1 credulity, logic, 268–9, 272, 273 decidability, logic, 265, 273 cryptography decision-making, game theory see game complexity theory, 23 theory MTC, 50 decision problems culture, CMC, 80, 83, 84 complexity theory, 19–20 Cunningham, D., 80 hierarchies of problems, 21–2 Curry–Howard correspondence, 243 NP-completeness, 23–4 Cut, logic, 266, 267, 273 parallel computation, 24–5 cybernetics, 186–95 first-order logic, 266 AI, 190–3 Turing machines, 4, 9–10, 19–20 ALife, 197–8 decision trees, 145 CMC, 79, 80 Dedekind, R., 264 computer programs, 191 default logic, 271–3, 274 definitions, 186, 187 defeasible reasoning, 267–74, 340 ethics, 71–2, 332 Delta rule, 144 feedback systems, 187, 191, 193, 194 DeMillo, R., 320, 321 goals, 186 democracy, CMC, 82, 84 Heidegger, 333–4 DENDRAL, 310 human sciences, 187–90, 193–4 Dennett, D. C., 10, 140, 205, 343 hypertext, 257–8 derivative data, in GDI, 43 ideas behind classical, 186–90 Derrida, J., 173, 335 interdisciplinary nature, 186–90, 193–4 Descartes, R., 119 biology, 187, 193 cybernetics, 188 engineering, 187 dream problem, 171–2 logic, 189–90 VR, 171–2, 173, 174 neurosciences, 189–90, 193 see also Cartesian dualism philosophy, 187–9 deterministic chaos, 30, 128 psychology, 187–9, 193 deterministic computers, 36 limits of classical, 190–3 deterministic processes, 29–30, 128 neural networks, 189, 191–2 Deutsch, D., 11 robotics, 190, 192–3 dialogue, and hypertext, 254–5 second-order, 194–5 Dibbell, J., 68 systems theory, 194–5 differential equations teleological machines, 191 finite digital physics, 180–1 CYC ontology, 160 systems science, 29, 34–5

355 Index digital art, 106–15 dystopianism, internet culture, 94–5, 97, 99–102, AARON, 114 103, 104 automatic art, 114 computers as tools in, 107–8, 115 ECHO, 309–10, 311, 342 creativity, 113–15 ecological systems, 35, 37 defining, 108–13 ALife models, 200, 208, 209 drawing, 114 economics EMI, 114 CMC, 82–3 evaluation, 107–8 data privacy, 71 interactivity, 110–11, 112–13 internet culture, 103–4 interface artworks, 113 systems science, 35–6, 38 literature, 107–8, 109–11 VR, 175–6 music, 107, 111–12, 114 education, philosophical, 84 ontology, 111–12 eductive interpretation, game theory, 291, 294 painting, 111 effective methods, Turing machines, 6, 7, 11, 12, performance art, 112 138 photography, 109, 113 elements, systems science, 28–9 physical embodiment, 111–12 Eliasmith, C., 309, 342 poetry, 114–15 eliminative materialism, 124 VR, 110 emergence, ALife, 201–2 digital universes, 180–5 EMI, digital art, 114 Dilthey, W., 330 emotions, role in science, 314 discovery, scientific energy minimization problem, 146 cognitive modeling, 308–9, 310 Engelbart, D., 78, 250, 251, 252 engineering AI, 310–11 engineering, cybernetics, 187 open questions, 314 engineering AI, 307–8, 310–11, 313, 314 theoretical approaches, 311–12 entropy, MTC, 52–3 dissipative systems, 29, 31, 33, 34 Entscheidungsproblem, 4, 9–10, 266 dominated strategies, game theory, 292, 293 see also decision problems drawing, digital art, 114 environment, internet culture, 93–4, 95–6 dreams, VR, 171–2 environmental information, 45, 57 Dretske, F., information, 328 epidemic processes, 37 knowledge, 229–31 epistemic optimality, 218–19 probabilistic approach, 53–4, 216–18, epistemically oriented semantic information, 41, 230–1 42, 45–6, 54–5 Dreyfus, H. L., 13, 131, 252, 334 epistemology dualism android, 310, 340 and AI, 119 CMC, 78–82, 83–4 Chinese Room, 122 computer modeling, 341–3 connectionism, 129 computing in philosophy of science, 210, 307, functionalism, 192 308, 313, 314 hermeneutic critique, 131 cybernetics, 194–5 semiotic systems, 129 IT, 332–3 Turing test, 120 Etchemendy, J., 341 CMC, 78, 79–80, 83–4 ethics, computer, 65–74 and cybernetics, 188 action theory, 67–8 recursive physics, 178, 184–5 applied, 68–9 VR, 173 approaches to, 65–6 Dummett, M., 246 codes of, 70 dynamic binding, 241 computational modeling, 344–5 dynamic logics, 273 for computer professionals, 69–70 dynamical systems, 28–38, 182, 194 crackers, 71–2

356 Index

cybernetics, 71–2, 332 finite state automatons, 4, 138 disclosive, 69 first-order logic (FOL), 264–7, 269 hackers, 71–2 fixed points, systems science, 31, 33, 34 internet, 68, 69, 72–3 Flores, F., 78, 334–5 IT, 332 Floridi, L., 248–9 metaphysical foundations, 66 Floyd, C., 319 metatheoretical, 66–8 Floyd–Hoare logics, 241 methodological, 66–8 fluctuation terms, systems science, 29 panopticon effect, 70–1 Fodor, J. policy vacuums, 66 computational theory of mind, 140–1, 147 privacy, 70–1 computer modeling, 343 search engines, 69 information and content, 215, 220, 221–3 social contract theory, 69–70 language of thought, 126–7, 140–1, 221–3 software ownership, 72 syntax processing, 125, 127 synthetic, 68 Foley, R., 233–4 technology, 66–8 folk psychology, 123–4 VR, 68, 73–4 Forbus, K. D., 309 evaluation of hypotheses Ford, K. M., 310, 340 cognitive modeling, 309–10, 311, 342 formal-learning theory, 311–12 engineering AI, 311 formal properties, logic, 274 theoretical approaches, 312–13 formal systems Evans, F., 99–100 AI, 125, 126, 127 evolution interpreted automatic, 126, 139–40 ALife, 202–6, 207, 208–9 see also semantic engines; Turing machines computer modeling, 344 formal verification, 320–2 ladder-of-complexity hypothesis, 205–6 Fortran, semantics, 238 life as, 206, 207 forward induction, game theory, 298 systems science, 35 foundationalism, 244 evolutionarily stable strategy (ESS), 300–2 fractal features, 37 evolutionary game theory, 289, 299–302, 344 Frame Fallacy, 99 evolutive interpretation, game theory, 291 frame problem, 128 expansionist approach, IT, 327, 331–2 Frege, G., 264 experimental game theory, 289 frogs, fly-snapping capacity, 220 expert systems, 276–9, 282 functional programming, 242–3 extensions, nonmonotonic logic, 271–4 functionalism, 191–2 extensive-form games, 295–9 ALife, 209 fuzzy logic, 278, 286 factual information, 41, 42, 45–6, 54–5 Falkenhainer, B., 309 Gabbay, D., 267 false information, 45–6 Gabriel, P., 111 Fano Code, MTC, 50, 51 Gaia hypothesis, 95–6 Fayyad, U., 311 Game of Life, 182, 200, 208 feedback systems, cybernetics, 187, 191, 193, game theory, 289–302 194 backward induction, 296–7 feedforward networks, 143–4 common knowledge, 291–2, 293, 294, 296–7 ALife models, 199 complete information, 290 Feigl, H., 188 computer modeling, 344 feminism, CMC, 79, 82 cooperative games, 290, 344–5 ferromagnetism, 33–4 dominated strategies, 292, 293 Fetzer, J., 320, 321–2, 337 eductive interpretation, 291, 294 finite digital physics, 180–2, 184 evolutionarily stable strategy, 300–2 finite recursion, 179, 180 evolutionary, 289, 299–302, 344

357 Index game theory (cont’d) Guarino, N., 160, 162 evolutive interpretation, 291 Guttenplan, S., 13 experimental, 289 Gygi, K., 248 extensive-form games, 295–7 extensive-form refinements, 297–9 Habermas, J., 82 focal point equilibrium, 293 hackers, 71–2 forward induction, 298 halting function, 8–9 hawk and dove game, 299–301 halting problem, 8 information sets, 295 halting theorem, 9 mixed strategies, 290, 291 Hamelink, C., 84 Nash equilibria, 289, 290–4, 296, 297, 298, Hamilton, W., 344 299, 301, 302 Hamiltonian (conservative) systems, 29–30, 31, noncooperative games, 290 35–6 normal-form games, 290, 295 Haraway, D., 79 normal-form refinements, 294–5 Hardy, G. H., 10 Pareto-dominant equilibrium, 294 harmonic oscillators, 29, 30, 31 perfect equilibrium, 298 Harnad, S., 209 Prisoner’s Dilemma, 339–40, 344–5, 346 Hartle, J. B., 11 pure strategies, 290 Haugeland, J., 126, 139, 140 rationality, 290, 291, 292–4, 296–9 hawk and dove game, 299–301 replicator dynamics, 301–2, 344 Hayes, P. J., 160, 310, 340 sequential equilibrium, 297–8 Hayles, K., 79–80 strategic interaction, 289–90 Hebb, D., Hebb rule, 144, 189 strategy, 295–6 Hegel, G. W. F., 332 subgame perfection, 297 Heidegger, M. subgames, 295, 297 internet culture, 94, 95, 100–1, 102 TFT strategy, 344–5 IT, 330, 333–5 trembling-hand perfection, 294–5, 297 Heidt, S., 173 “gavagai” puzzle, 217 Heim, M., 94, 95, 168 gender issues, HCI, 79 Hempel, C., 323 gene regulation, 310 Henry, G. C., 11 Genesereth, M. R., 161 hermeneutics genetic evolution, game theory, 302 AI, 131 genetic neutrality (GeN), 43, 45 CMC, 78, 84 genetic systems, 35 cybernetics, 194–5 GENSIM, 310 IT, 330, 333 Gentner, D., 309 heuristics, cognitive modeling, 308 Geroch, R., 11 higher-order logic, 245 global mind, 96 Hilbert, D., 9–10 globalization Hilbert program, 9–10 CMC, 82–3 Hinton, G. E., 145–6 VR, 175–6 Hoare, C. A. R., 319 Glymour, C., 310, 311, 340, 343 Hodges, A., 10 Gnosticism, 79 holism of belief fixation, 219, 223 Gödel, K., 9, 265–6 Holocaust, 100, 101, 102 Gotterbarn, D., 70 Holyoak, K. J., 309, 342, 343 Gould, S. J., 203, 205 homeostatic systems grain, problem of, 216, 217–18, 219–20, 221 ALife, 197–8 graphical user interfaces (GUIs), 98, 113 cybernetics, 187, 190, 193–4 Gregory, R. L., 10 Hopfield, J., 143, 145, 146, 191–2 Grim, P., 338, 340–1, 344, 345 Hrachovec, H., 81 Gruber, T. R., 161 Hull, C. L., 188–9

358 Index human–computer interaction (HCI), 76–80 postmodernism, 93, 99 AI, 78 relativity, 258 cybernetics, 79, 80 VR, 174–5 definition, 77 imitation game, 25–6 gender issues, 79 independence, conditional, 279, 280–1, 282, hypertext, 252–9 285–6 ontology, 78–80 individual-based models, ALife, 198–9 personhood, 78–80 induction, problem of, 128 politics, 80 inference humanness, personhood, 78–80, 81–2, 83, 93, hypothesis generation, 308 256–9 scientific theory evaluation, 310, 311, 312 Hume, D., 128 inference networks, AI, 278, 279–80, 311 Humphreys, P., 337 infomania, 94, 95 hyperdigital computers, 183–4 information, 40–58 hyperdigital physics, 182, 184 alethic neutrality, 45–6, 54 hyperdigital universes, 182–4 antireductionist theories of, 40, 41 Hyperproof, 341 biological, 57 hypertext, 81, 248–59 see also biology, ALife associations, 251 centralized approaches, 41 connections, 251 see also information, epistemically oriented cybernetics, 257–8 semantic definitions, 248–9 channel conditions, 233–5 dialogue, 254–5 complexity, 57 electronic fallacy, 249 and content, 42–5, 215–23 expressionist fallacy, 249 adequacy conditions, 216, 217–18, 219–20, interface, 249 221–2 and the body, 252–3, 255, 259 asymmetric dependence, 221–3 computer as language machine, 252–3 epistemic optimality, 218–19 computer as medium, 253–4 flow of information, 216–18 design, 254–5 grain, 216, 217–18, 219–20, 221 intelligence distribution, 255–7, 259 and knowledge, 230–1 and the person, 255–7, 259 misrepresentation, 216, 218, 219–20, 221–2 wrighting, 255 naturalism, 215, 216, 217, 218–19, 220, language, 250, 251, 252–3, 257, 258 222, 223 literary fallacy, 249 ontological naturalism, 215 origins, 249–51 teleology, 219–20 space-time, 257–8 decentralized approaches, 41 HYPGENE, 310 definitions, 328 hypothesis evaluation environmental, 45, 57 cognitive modeling, 309–10, 311, 342 epistemically oriented semantic, 41, 42, 45–6, engineering AI, 311 54–5 theoretical approaches, 312–13 factual, 41, 42, 45–6, 54–5 hypothesis generation false, 45–6 cognitive modeling, 308–9, 310 General Definition of, 42–6 engineering AI, 310–11 genetic neutrality, 43, 45 open questions, 314 inherence dystopianism, 100 theoretical approaches, 311–12 instructional, 57 internet, integrity, 73 identity and knowledge, 228–36 CMC, 79, 83–4, 258 information theory adapted to, 229–31 internet culture, 96–8, 99 information-theoretic accounts (ITKs), modernism, 99 231–6

359 Index information limitationist approach, 327, 331–3 and knowledge (cont’d) linguistic philosophy, 330 objections to ITKs, 233–6 metaphor in, 335 open questions, 231–3 metaphysics, 331–4 MTC, 46–53 open questions, 335 entropy, 52–3 information theory, use of term, 46, 328 implications, 51–3 see also information, MTC Kolmogorov complexity, 57 information-theoretic accounts of knowledge noise, 50–1 (ITKs), 231–6 raw information quantification, 47–53 inherence dystopianism, 94–5, 97, 99–102, 104 redundancy, 50–1 inherence instrumentalism, 102–3, 104 semantic information, 52–3 inherence utopianism, 95–6, 97, 100, 103–4 multicentered approaches, 41 inheritance, nonmonotonic logic, 270–1 nonreductionist theories of, 40, 41, 42 instructional information, 57 see also information, epistemically oriented instrumentalism, internet culture, 94, 102–3, semantic 104 ontological neutrality, 43–5 Integrated Connectionist Symbol (ICS) physics of, 57, 178–85 architecture, 147–8 pragmatic, 57 intentional content, information see information, see also game theory and content and reality, 100 interactivity reductionist theories of, 40–1 digital art, 110–11, 112–13 semantic, 42–6, 53–6 VR, 168–9 as content, 42–5 interface see also information, and content digital art, 113 definition, 328 hypertextual, 249, 252–9 degree of informativeness, 54–5 internet culture, 98 as factual information, 45–6 internet inferential approach, 54 communication, 72, 73 modal approach, 54 see also computer-mediated communication and MTC, 52–3 culture see internet culture probabilistic approach, 53–4, 216–18, 230–3 detachment, 93 systemic approach, 54 ethics, 68, 69, 72–3 see also information, and knowledge anonymity, 72, 73 Special Definition of, 46, 54 crime, 71–2 taxonomic neutrality, 43, 45 global scope, 72, 73 technological, 100 reproducibility, 72–3 typological neutrality, 43, 45 search engines, 69 unified theory of (UTI), 40 metatool account, 102, 104 useful, 57 personality change, 93–4 see also game theory systems science, 36–7 valuable, 57, 70–1 VR, 176 information science, and ontology, 158–64 internet culture, 84, 92–104 information sets, game theory, 295 cybercrime, 71–2 information society, ethics, 69–74 dystopianism, 94, 103 information systems, systems science, 36–7 balance, 94–5, 103 information technology, 327–35 inherence, 94–5, 97, 99–102, 104 definitions, 327–8 instrumentalism, 94 expansionist approach, 327, 331–2 inherence, 102–3, 104 Heideggerian analyses, 330, 333–5 utopianism, 94, 102, 103 hermeneutics, 330, 333 balance, 96–9, 103 historico-philosophical perspective, 328–31 inherence, 95–6, 97, 100, 103–4

360 Index interpretation Landau, L. D., 33–4 IT, 333 Landow, G. P., 249 logic, 264 Langley, P., 308, 310 probability in AI, 283–6 Langton, C., 197 Introna, L. D., 69 language introspective closure, 274 and AI intuitionistic logic, 245 communication, 131–2 Ising model, 198 folk psychology, 124 formal systems, 127 Jeffrey, R. C., 12 mentalese, 126–7 Johnson, T. R., 311 syntax processing, 125, 127 Johnson-Laird, P., 13–14 computer, 139–40 Josephson, J. R., 311 abstraction, 242, 245, 246, 323–4 Josephson, S. G., 311 computer science methodology, 320, 323–4, Joyce, M., 110 325 justification theories, 228–9 semantics see semantics, computer languages computer modeling, 343–4 Kahn, P., 250, 251 formal-learning theory, 311–12 KAM theorem, 31 hypertext, 250, 251, 252–3, 257, 258 Kant, I., 157, 172–3 modal logic, 273 Karp, P., 310 ontology, 157–8 Kauffman, S., 204–5 of thought (mentalese), 126–7, 140–1, 221–3 Kay, A., 253 and worldview, 78, 258 KEKADA, 308 language machine, computer as, 252–3 Kelly, K., 95–6, 311, 312 Lansdown, J., 108 Kilb, N., 344 Laplace, P.-S., 29, 30 Kittay, E., 343–4 Laplacean spirit, 29 Kleene, S. C., 7 Laurel, B., 253–4, 255 knowledge, 228–6 Lauritzen, S. L., 281, 284 closure principle, 235–6 learning information theory adapted to, 229–31 ALife models, 199 information-theoretic accounts (ITKs), 231–6 connectionist networks, 144, 189–90, 192, 199 objections to, 233–6 intentional contents of signals, 218 open questions for, 231–3 probability in AI, 286, 311 justification theories, 228–9 legislation theories of see epistemology cybercrime, 72 Kocabas, S., 310 data privacy, 71 Kokalis, T., 344 Lehrer, K., 234–6 Kolb, D., 81 leisure, internet culture, 103–4 Kolmogorov–Arnold–Moser (KAM) theorem, 31 Lewis, G., 111 Kolmogorov axioms, 287 Lewontin, R., 203 Kolmogorov complexity, 19, 57 lexical binding, 241 Korzybski, A., 258 Liar dynamics, 341 Kreps, D., 297–8 liberation, through CMC, 79, 82 Kripke, S., 141, 235–6 libertarianism, 102 Kulkarni, D., 308 life Kurzweil, R., 115 artificial, 197–210 Kyburg, H., 338 as cluster of properties, 206 as evolution, 206, 207 Lacan, J., 174 Game of, 182, 200, 208 ladder-of-complexity hypothesis, 205–6 as metabolization, 206–7 lambda calculus, 243 test for, 209

361 Index limit cycles, 31, 33, 34 modal, 273 limitationist approach, IT, 327, 331–3 model theory, 265 Lindsay, R., 310 monotony, 266–8, 272, 274, 340 linear systems, 29 Nixon diamond, 268–9 linguistic philosophy, 330 nonclassical propositional, 22 computer modeling, 343–4 nonmonotonic, 267–74 Lipton, R., 320, 321 autoepistemic, 273–4 Lisp, semantics, 238, 243 circumscription, 269–70 literature, digital art, 107–8, 109–11 closed world assumption, 269 living systems computer modeling, 340 ALife, 197–201 default, 271–3, 274 adaptationism, 202–3, 204–5 defeasible consequences, 272–3 emergence, 201, 202 defeasible inheritance, 270–1 evolutionary progress, 205 degree notion, 271 nature of life, 206–7 extensions, 271–4 strong, 209 inheritance, 270–1 cybernetics, 186, 187–90, 193–4, 257–8 minimization, 269 evolutionary game theory, 299, 302, 344, 345 origins of modern, 263–4 Gaia hypothesis, 95–6 quantified propositional, 24 systems science, 35, 36, 37 reflexivity, 266, 267, 268 Loewer, B., 218 satisfaction, 264 logic, 263–74 satisfiability problem of propositional, 19, 20, autoepistemic, 273–4 22–3 CMC, 81–2 second-order, 269–70 compactness theorem, 265 semantics of computer languages, 245–6 completeness theorem, 265–6 skepticism, 268–9, 272–3 complexity, 274 states, 273 computer modeling, 340–1 supraclassicality, 266, 268 consequence relations, 264–5, 266–8, 272–3 symbolic, 264 consistency, 266 transitions, 273 credulity, 268–9, 272, 273 truth, 264 Cut, 266, 267, 273 value solids, 340–1 cybernetics, 189–90 logical consequence, 264 decidability, 265, 273 logical form, 263 default, 271–3, 274 logicism, 319, 323 defeasible reasoning, 267–74, 340 Lorenz system, 31 diagrammatic representation, 341 Lotka–Volterra equations, 35, 180–1 dynamic, 273 Lovelock, J., 95 first-order, 264–7, 269 Löwenheim–Skolem theorem, 265 Floyd–Hoare, 241 Luna, F., 345–6 formal properties, 274 Lyapunov exponents, 33 fuzzy, 278, 286 Hyperproof, 341 McCarthy, J., 320–1 hypertext, 258 McClelland, J. L., 344 infomania, 94 McCulloch, W., 189 interpretation, 264 McDermott, D., 123, 128 introspective closure, 274 McDowell, J., 215 logical consequence, 264 machine learning, AI, 286, 311 logical form, 263 McLuhan, M., 330 Löwenheim–Skolem theorem, 265 Magnani, L., 308 material adequacy, 274 Makinson, D., 274 mathematics as branch of, 319, 323 map coloring problem, 19, 23

362 Index

Mar, G., 338, 340–1, 344, 345 Turing test, 120, 139, 252 Marr, D., 137–8 weak, 122 Marx, K., 103–4, 176 see also mind, computational theory of material adequacy, logic, 274 as automatic formal system, 139–40 mathematical theory of communication (MTC), Church–Turing thesis, 13–14, 138 46–53 cognitive modeling, 307, 308–10, 311, entropy, 52–3 313–14 implications, 51–3 complexity–philosophy relation, 26–7 Kolmogorov complexity, 57 computational theory of, 120, 135–48 noise, 50–1 algorithms, 136 – 8, 140 raw information quantification, 47–53 cognitive abilities, 146–7 redundancy, 50–1 cognitive functions, 136–8 semantic information, 52–3 connectionism, 135, 137, 142–8 mathematics, computer science, 319–23 hypertext, 252 Mauss, M., 256 ICS architecture, 147–8 maximality thesis (thesis M), 11–13, 14 language of thought, 140–1 Maynard Smith, J., 207, 299, 300 networks, 142–7, 192 Mayr, E., 206, 207 nonsymbolic algorithms, 137 meaning, theory of see semantics psychosemantics, 141 mechanical methods, Turing machines, 7, 12 subsymbolic, 147 mechanism, and cybernetics, 188–9 symbol system, 135, 137, 139–41, 147–8, MECHEM, 310 252 memex, 250–1 symbolic algorithms, 136, 137 Mendelson, E., 12 tasks of, 135 mental states, intentional content, 215 Turing machines, 138 see also information, and content see also mind, and AI mentalese, 126–7, 140–1, 221–3 connectionism, 129, 133, 135, 137, 142–8 metabolization, life as, 206–7 functional architecture see mind, computational metadata, in GDI, 43 theory of META-DENDRAL, 310 global, 96 metaphysical necessity, 222–3 language of thought, 126–7, 140–1, 221–3 metaphysics, 155 mentalese, 126–7, 140–1, 221–3 IT, 331–4 parallel processing, 26–7, 129, 142 ontology see ontology as semiotic system, 129–33 VR, 167–8, 170–3, 175–6 serial processing, 26–7 metatool account, internet, 102, 104 symbol systems, 130, 133, 135, 139–41, mind 147–8, 252 and AI, 119, 138–9 as syntactic engine, 140 Chinese Room, 26, 122 syntactic theory of, 125 computer modeling, 343 systems science, 36 eliminative materialism, 124 mind–body dualism see Cartesian dualism folk psychology, 124 minimization, nonmonotonic logic, 269 frame problem, 128 mirror images, VR, 170–1, 174, 176 functionalism, 192 misrepresentation, content theory, 216, 218, hermeneutic critique, 131 219–20, 221–2 hypertext, 252 modal logics, 273 mentalese, 126–7 model theory, 265 parallel processing, 26, 129 modeling, computational see computational semantic engines, 126, 129, 133 modeling semiotic systems, 129–33 modernism symbol systems, 130, 133, 252 CMC, 79, 81, 82 syntax processing, 125–6, 127 internet culture, 98–9

363 Index money, virtual existence, 176 and cybernetics, 189, 191–2 monotony, logic, 266–8, 272, 274, 340 probability in AI, 286 Moor, J., 66, 68, 338 scientific theory evaluation, 309–10, 342 moral issues see ethics, computer systems science, 36, 37, 38 Morgenstern, O., 290, 344 see also networks, connectionist Moser, J. K., 31 neurosciences, cybernetics, 189–90, 193 Multi-User Dungeons (MUDs), 97–8 neutral shadows, adaptationism, 203–4 music, digital art, 107, 111–12, 114 Newell, A., 11, 14–15, 121, 125, 130 MYCIN, 276–7, 278 Newman, M., 3 Neyman, J., 283–4 Nagel, E., 323 Nilsson, L., 161 Nash equilibria, 289, 290–4, 296, 297, 298, Nissenbaum, H., 69 299, 301, 302 Nixon diamond, 268–9 natural selection noise, MTC, 50–1 ALife, 203–5, 207 noncooperative games, 290 evolutionary game theory, 299–302 non-equilibrium transitions, 34, 36 naturalism nonlinear interactions, 33, 36 definition, 224 nonlinear models, 34–5, 37–8 informational theory of content, 215, 216, nonmonotonic logics, 267–74, 340 217, 218–19, 220, 222, 223 nonrecursive physics, 180, 184 nature, internet culture, 93–4, 95–6 nonreductionist theories, information, 40, 41, 42 see also living systems normal-form games, 290, 295 Naur, P., 321 Nowak, M., 344–5 Naylor, T., 337 Nozick, R., 228 Nazism, 100, 101 NP-complete problems, 19, 20, 22–4 negative feedback systems, 187, 191, 193 Nyce, J. M., 250, 251 Nelson, T., 249–50, 251 Net see internet objective probabilities, AI, 277, 284, 285, 287 NETtalk, 144–5 Odifreddi, P., 11 networks Ontek, 160 Bayesian, 279–82, 284, 285–6 ontological commitments, 156–7, 158, 163 Boolean ontological naturalism, 215 ALife, 198, 204–5 ontological neutrality (ON), 43–5 complexity theory, 22, 25 ontology, 155–64 connectionist, 142–7 adequatist, 155, 156 ALife models, 199 alternative possible worlds, 161 best-match problems, 145–6 applied, 164 connected problem, 145 CMC, 78–80, 83–4 cybernetics, 189–90 conceptualizations, 161–3 learning, 144, 189–90, 192, 199 CYC project, 160 multiple soft constraint satisfaction, 145–6, definitions, 155, 161–2 147 digital art, 111–12 pattern recognition, 144–5 external metaphysics, 157–8 scientific theory evaluation, 309–10, 342 fluxist, 155 training, 144–5, 192, 199 future directions, 162–4 see also neural networks and information science, 158–64 inference, 278, 279–80, 311 internal metaphysics, 157–8 nonmonotonic, 271 IT, 331–2 neural networks methods, 156 artificial, 142, 189 Ontek, 160 complexity–philosophy relation, 26 ontological commitments, 156–7, 158, 163 connectionist computational paradigm, 142 PACIS, 160

364 Index

philosophical, 155–8, 162–4 physical symbol system, 14–15, 121, 125, 130, reductionist, 155 133 in scientific theories, 156–7, 158, 163 physics software, 178 ALife, 198 substantialist, 155 of computation, 183–4 top-level, 159, 162 discovery programs, 308, 310–11 uses, 159–61 finite digital, 180–1, 184 operational data, in GDI, 43 hyperdigital, 182, 184 opponent processing theory, 146–7 of information, 57, 178–85 oracle machines, 23–4, 25 programs, 179–80 order parameters, systems science, 33–5, 36, 37, recursive, 178–85 38 systems science, 33–5 OSCAR, 341–2 theories, 180 transfinite digital, 182–3, 184 PACIS, 160 Piatetsky-Shapiro, G., 311 painting, digital art, 111 Pickstock, C., 329 palindrome problem, 20 Pitts, W., 189 panopticon effect, 70–1 Plantinga, A., 234, 235 parallel distributed processing (PDP), 26, 142, , 173, 329 192, 199 Plotinus, 170–1, 172, 176 parallel processing poetry, digital art, 114–15 AI, 26, 129, 192 poiesis, 101, 333 complexity theory, 24–5, 26 Poincaré, H., 30 computational theory of mind, 142 political systems, 35, 37–8 parallel random access machine (PRAM), 24 politics PCs see personal computers CMC, 80, 82, 84 Pearl, J., 279–82, 284 data privacy, 71 Peirce, C. S., 80–1, 129, 164 information technology, 332 Penrose, R., 127 internet culture, 103–4 perception virtual reality, 176 binding problem, 35 Pollock, J., 338, 341–2 network models, 146–7 Popper, K., 104, 283, 285 VR, 171–3 Port, R., 343 perceptrons, 143, 189–90, 191, 192 Poster, M., 175, 335 performance art, 112 Postman, N., 100 periodic limit cycles, 31, 33, 34 postmodern culture, 93, 97, 98 Perlis, A., 320, 321 postmodernism personal computers, xi CMC, 79, 80, 81–2 user interface, 255–6, 258–9 internet culture, 93, 94–5, 98–9 personalization, hypertext, 256–7, 258–9 pragmatic information, 57 personhood see also game theory CMC, 78–80, 81–2, 83, 256–9 praxis, internet culture, 100–1 postmodern culture, 93 Price, G., 299, 300 personification, computers, 255–6, 258–9 printing problem, Turing machines, 8 phase spaces, systems science, 28, 29–31, 32–3 Prisoner’s Dilemma, 339–40, 344–5, 346 phase transitions, systems science, 33–6, 37 privacy, computer ethics, 70–1 phenomenology, CMC, 78, 84, 256–7 probabilistic reasoning, Cut, 266 Philips, J. P., 310 see also probability philosophy of computing and information (PCI), probability xii in artificial intelligence, 128, 276–87 photography, digital art, 109, 113 Bayes theorem, 279, 287 phronesis, 100–1 Bayesian networks, 279–82, 284, 285–6

365 Index probability Putnam, H., 12–13, 138, 157, 191 in artificial intelligence (cont’d) Pylyshyn, Z., 127, 140–1, 147 Bayesianism, 278–9, 282–5, 287 causality, 282, 283, 286, 287 Qin, Y., 308 certainty factors, 278 quantum computers, 36 conditional independence assumptions, 279, quantum information, 36 280–1, 282, 285–6 quantum physics, 181 expert systems, 276–9, 282 quantum processes, 127 fuzzy logic, 278, 286 quantum systems, 29 inference networks, 278, 279–80, 311 quasi-periodic limit cycles, 31, 33, 34 interpretation problem, 283–6 Quine, W. V. O. machine learning, 286, 311 Democritean physics, 184 MYCIN, 276–7, 278 “gavagai” puzzle, 217 neural networks, 286 ontology, 156–7 objective probabilities, 277, 284, 285, 287 programming-language semantics, 238, 240, PROSPECTOR, 278–9 243, 246 scientific discovery programs, 310–11 scientific theory evaluation, 311 racial discrimination, 345 subjective probabilities, 277, 278, 284, 285, 287 random events, systems science, 29 terminology, 286–7 random variables, 286 semantic information, 53–4, 216–18, 230–3 Rational Monotony, 267–8 probability distribution, 286 rationality probability theory, 286 CMC, 78, 81, 82 MTC see mathematical theory of communication computer modeling, 341–2 problem-solving resource analysis see complexity game theory, 290, 291, 292–4, 296–9 procedural abstraction, 324 Ray, T. S., 200, 208 Production Systems, 140 reality professional ethics, 69–70 ALife, 197 programming languages, 139–40 cybernetics, 194 abstraction, 242, 245, 246, 323–4 internet culture, 97, 100 computer science methodology, 320, 323–4, 325 ontology see ontology semantics see semantics, computer languages recursive physics, 178–85 programs virtual see virtual reality computational philosophy of science, reason see rationality 308–11, 313–14 recursive generability principle, 127 computer science methodology, 318–20 recursive physics, 178–85 abstraction, 323–5 RED, engineering AI, 311 debugging, 320–1 reduction, efficient, 21 social processes, 321 reductionism specification, 320 computer science, 319 verification, 320–2 information, 40–1 and cybernetics, 191 reductionist ontology, 155 physics of information, 179–80 redundancy, MTC, 50–1 PROSPECTOR, 278–9 referential transparency, 240, 242–3 psychology Reffen Smith, B., 108, 113 Church–Turing thesis, 10–11, 13 reflexivity, logic, 266, 267, 268 cognitive modeling, 308–10 Reichenbach, H., 282 cybernetics, 187–9, 193 Reiser, O. L., 258 folk, 123–4 Reiter, R., 271, 272 inherence dystopianism, 99 religion, CMC, 79 IT, 332 replicator dynamics, 301–2, 344 public sphere, CMC, 82 representation, hypertext, 253–4

366 Index resources, problem-solving analysis see complexity second-order cybernetics, 194–5 Rheingold, H., 168 second-order logic, 269–70 robotics, 190, 192–3, 200–1 Sejnowski, T., 144–5 Rohrlich, F., 337 self Rosenberg, C. R., 144–5 internet culture, 96–8, 99 Rosenblatt, T., 189–90 personhood Rosenblueth, A., 187–8 CMC, 78–80, 81–2, 83, 256–9 Ross, T., 190 postmodern culture, 93 Rotman, B., 255, 257 VR, 174–5 Rowe, R., 111 self-organization Rumelhart, D., 192 ALife, 199–200, 204–5, 208 Ruse, M., 345 cybernetics, 190, 191, 194 Russell, B., 257 systems science, 33–6, 37, 38, 194 self-regulation St. Denis, P., 338, 340–1, 344, 345 ALife, 197–8 St. John, M., 344 cybernetics, 187 Saltz, D., 112 self-similarity, systems science, 36 Sandbothe, M., 80, 84 Selten, T., 294, 297, 298 satisfaction, logic, 264 semantic engines, AI, 125–6, 129, 133 satisfiability problem of propositional logic, 19, semantic information, 42–6, 53–6 20, 22–3 as content, 42–5 Saussure, F. de, 330 see also information, and content Sayre, K., 328 definition, 328 Scheines, R., 311, 343 degree of informativeness, 54–5 Schelling, T., 293 as factual information, 45–6 Schirmacher, W., 332 inferential approach, 54 Schleiermacher, F., 330 modal approach, 54 Schrödinger, E., 206 and MTC, 52–3 Schulte, O., 312 probabilistic approach, 53–4, 216–18, 230–3 science systemic approach, 54 ALife, 198 semantics computational philosophy of, 307–14 compositionality of meaning, 127 analogy, 309, 342 computational theory of mind, 141 cognitive modeling, 307, 308–10, 311, computer languages, 237–46 313–14 abstraction, 242, 245, 246 discovery, 308–9, 310–12, 314 Algol-like, 238–9 emotions, 314 blocks, 238 engineering AI, 307–8, 310–11, 313, 314 category theory, 245–6 formal-learning theory, 311–12 Curry–Howard correspondence, 243 function of philosophy, 307 denotational semantics, 239 hypothesis evaluation, 309–10, 311, 312–13, differentiating, 244 342 dynamic binding, 241 open problems, 314 Floyd–Hoare logics, 241 theoretical approaches, 307, 311–13 foundationalism, 244 visual imagery, 314 full abstraction, 242, 246 see also computational modeling, as functional programming, 242–3 philosophical methodology higher-order logic, 245 computer, methodology of, 318–25 history, 238–40 ontological commitments, 156–7, 158, 163 identity of programs, 241–2 philosophy compared, 318 intuitionistic logic, 245 search engines, ethics, 69 lambda calculus, 243 Searle, J., 13, 26, 27, 122, 141, 343 lexical binding, 241

367 Index semantics software computer languages (cont’d) ownership ethics, 72 modularity, 244 physics of information, 178 observational equivalence, 241–2, 246 see also programs program phrases, 242 solid-state lasers, 34 recursion, 240 soul referential transparency, 240, 242–3 immortality, 185 subroutines, 239, 240, 241, 242–3 internet culture, 100–1 substitutions, 241 space taxonomy, 244 CMC, 80, 84 technical tools, 245–6 digital universe, 181, 182–3 theories of meaning, 246 hypertext, 257–8 uses, 241–5 VR, 172, 173 values, 239, 240, 241, 243, 246 space complexity, 18, 19–20 formal systems meaning, 127 NP completeness, 23–4 and hypertext, 258 space constructible functions, 21 model-theoretic, 157 space hierarchy theorem, 21 see also semantic information Spiegelhalter, D. J., 281, 284 semiotic systems, AI, 129–33 Spirtes, P., 311, 343 semiotics, CMC, 80–1 Stalnaker, R., 267–8 serial processing, 26–7 state space, systems science, 30 Serres, M., 254 states Shank, G., 80 modal logic, 273 Shannon, C., 40–1, 46, 47, 49, 51, 57, 328 systems science, 28–30 Shavlik, J. W., 145 Stefansson, B., 346 Shaw, J., 111 Steinart, E., 343–4 Shortliffe, E. H., 277, 278 Stich, S., 125 Sigmund, K., 344–5 stochastic computers, 36 sign-using (semiotic) systems, 129–33 stochastic nonlinear differential equations, 34–5 Simon, H., 121, 125, 130, 308 stochastic processes, 29 simulation Strawsonian descriptive metaphysics, 157 AI, 122 Structure Mapping Engine (SME), 309 ALife, 208 subject, the, VR, 174–5 computer see computational modeling subjective probabilities, AI, 277, 278, 284, 285, culture of, 98–9 287 Skyrms, B., 344 supraclassicality, logic, 266, 268 Sloman, A., 313 SWARM, 345–6 Smith, A., 36 symbol systems, 135 Smolensky, P., 11, 147–8 AI, 14–15, 121, 125, 130, 133, 191 Smyth, P., 311 computational theory of mind, 130, 133, 135, social contract theory, 69–70 139–41, 147–8 , computer modeling, 344–5 hypertext, 252 social processes, computer science, 321 symbol-using semiotic systems, 129–33 social relations, virtual, 176 symbolic logic, 264–74 social systems, 33, 35–6, 37–8 symbols, MTC, 50, 52 socialization, human becoming, 256–7 syntactic engine, mind as, 140 society, computers in syntactic theory of mind (STM), 125 CMC, 76–84 syntax processing, AI, 124–6, 127 digital art, 106–15 systems ethics, 65–74 AI HCI, 76–84 chaotic, 128 internet culture, 71–2, 84, 92–104 closed, 127–8

368 Index

formal, 125, 126, 127 phase spaces, 28, 29–31, 32–3 open, 127–8 phase transitions, 33–6, 37 semiotic, 129–33 physical systems, 33–5 computational, 36–8 political systems, 35, 37–8 dynamical, 28–38, 182, 194 probabilistic states, 29 formal, 125, 126, 127 quantum computers, 36 interpreted automatic, 126, 139–40 quantum information, 36 see also semantic engines; Turing machines quantum systems, 29 information, 36–7 quasi-periodic limit cycles, 31, 33, 34 systems science, 28–38 random events, 29 ALife, 198, 202, 204 self-organization, 33–6, 37, 38, 194, attractors, 31, 32–3, 34, 204 204–5 automobile traffic, 36 self-similarity, 36 basins of attraction, 31 social systems, 33, 35–6, 37–8 Bénard experiment, 34 state space, 30 biological systems, 35, 36, 37 stochastic computers, 36 butterfly effect, 30–1, 35, 36 stochastic nonlinear differential equations, chaos, 30, 31, 32–3, 34, 36, 37 34–5 classical stochastic processes, 29 stochastic processes, 29 classification of systems, 31 Takens’ theorem, 32–3 computational systems, 36–8 time-depending states, 28–30 conservative systems, 29–30, 31, 35–6 time series, 28, 29–30 continuity of processes, 29 time-series analysis, 32–3 cybernetics, 194–5 traffic systems, 36–7 deterministic chaos, 30 trajectories, 29, 30–1, 33 deterministic computers, 36 deterministic processes, 29–30 Tafti, A., 344 difference equations, 29 Takens’ theorem, 32–3 differential equations, 29, 34–5 Taraban, R., 344 dissipative systems, 29, 31, 33, 34 Tarski, A., 264 economic systems, 35–6, 38 taxonomic neutrality (TaN), 43, 45 elements, 28–9 techne, 101, 328, 329 emergence, 202 technocratic rationality, 99–100 fixed points, 31, 33, 34 technology, 328–9 fluctuation terms, 29 information see information technology fluctuations, 34, 37–8 technopolies, 100 fractal features, 37 teleology Hamiltonian systems, 29–30, 31, 35–6 cybernetics, 187–8, 191 harmonic oscillators, 29, 30, 31 information and content, 219–20 information systems, 36–7 telepresence, 169–70 internet, 36–7 TETRAD, 311, 343 limit cycles, 31, 33, 34 TFT strategy, 344–5 linear systems, 29 Thagard, P., 342–3 Lorenz system, 31 theology, IT, 329–30 Lyapunov exponents, 33 theory, definition, 180 natural systems, 33–5, 36, 37, 38 theory evaluation see hypothesis evaluation neural networks, 36, 37, 38 theory generation see hypothesis generation non-equilibrium transitions, 34, 36 thesis M, 11–13, 14 nonlinear interactions, 33, 36 thesis S, 14 nonlinear models, 34–5, 37–8 Thorndike, E., 188 order parameters, 33–5, 36, 37, 38 thought, language of (mentalese), 126–7, 140–1, periodic limit cycles, 31, 33, 34 221–3

369 Index thought experiments maximality thesis, 11–13, 14 ALife, 205, 210 mechanical methods, use of term, 7, 12 computer modeling, 338 misunderstandings about, 10–15 TIERRA, 200, 208, 345 and physical machines, 121 time physics of information, 178, 183–4 CMC, 80, 84 printing problem, 8 digital universe, 181, 182–3 simulation fallacy, 13–14 hypertext, 257–8 thesis M, 11–13, 14 VR, 170, 172, 173 thesis S, 14 time complexity, 18–20 uncomputable numbers, 4, 8 hierarchies, 21 universal, 6 NP-completeness, 22–4 Turing test, 120, 122, 139, 209, 252 reducibility, 21 Turing thesis see Church–Turing thesis time series, systems science, 28 Turkle, S., 96–7, 175 time-series analysis, 32–3 typological neutrality (TyN), 43, 45 traffic systems, 36–7 trajectories, systems science, 29, 30–1, 33 ultrastability, cybernetics, 190 transfinite digital physics, 182–3, 184 uncomputable numbers, 4, 8 transfinite recursion, 179–80, 182 unified theory of information (UTI), 40 transitions, logic, 273 universal Turing machine (UTM), 6 traveling salesman problem, 145 see also Turing machines Travis, C., 237 universe, as computer, 178–85 trembling-hand perfection, 294–5, 297 useful information, 57 Turing, A. M., 3, 4, 119 see also game theory imitation game, 25–6 user interface see interface see also Turing machines utopianism, internet culture, 94, 95–9, 100, 102, Turing machines, 3–15 103–4 accelerating (ATMs), 183–4 AI, 14–15, 119, 120–1 Valdés-Pérez, R. E., 310 atomic operations, 4–5 valuable information, 57, 70–1 basic operations, 4–5 value solids, 340–1 Church–Turing fallacy, 13 van Gelder, T., 343 Church–Turing thesis, 7, 10–15, 119, verification, computer science, 320–2 138 virtual metaphysics, 168, 170–3 complexity theory, 19–20 virtual reality (VR), 167–76 NP-completeness, 23–4 defining criteria, 168–70 oracle machines, 23–4, 25 digital art, 110 parallel computation, 24–5 economics, 175–6 computable numbers, 4, 8 ethics, 68, 73–4 computational theory of mind, 138 globalization, 175–6 decision problems, 4, 9–10, 266 internet culture, 96, 97 complexity theory, 19–20, 23–5 intersubjectivity, 169, 172 effective methods, 6, 7, 11, 12, 138 metaphysics, 167–8, 170–3, 175–6 Entscheidungsproblem, 4, 9–10, 266 personal identity, 174–5 equivalence fallacy, 14–15 politics, 176 example, 5–6 space, 172, 173 finite state automatons, 4, 138 the subject, 174–5 halting function, 8–9 technologies, 168–70 halting problem, 8 telepresence, 169–70 halting theorem, 9 time, 170, 172, 173 human computation, 6–7, 11, 13–14, 138 virtual software objects, 178 inherence dystopianism, 99–100 virtual space, 173

370 Index virtual time, 170, 173 Weaver, W., 41, 52 virtual worlds, 322 web see internet visual imagery, 314 Welty, C., 162 visual perception, 146–7 Wheeler, J., 44, 181 vivisystems, 95–6 Whitehead, A. N., 256 von Neumann, J. Whorf, B. L., 250, 251, 257, 258, 259 ALife, 197 Wiener, N. computational theory of mind, 139 cybernetics, 186, 187–8, 193–4, 197 cybernetics, 189 information, 44, 331 game theory, 290, 291, 344 time, 257–8 Turing machines, 3, 10 Wilson, R., 297–8 von Neumann machines, 139 Winograd, T., 78, 252, 334–5 Wittgenstein, L., 330 Walser, R., 169 work, internet culture, 103–4 Walter, W. G., 190 worldview, CMC, 76, 77–84, 257–8, 259 Walton, K., 108–9 weak second-order theory of one successor Zeilinger, A., 181 (WS1S), 22 Zeno point, 180

371