The Brain As Artificial Intelligence : Prospecting the Frontiers of Neuroscience
Total Page:16
File Type:pdf, Size:1020Kb
Original citation: Fuller, Steve (2018) The brain as artificial intelligence : prospecting the frontiers of neuroscience. AI & SOCIETY .doi:10.1007/s00146-018-0820-1 Permanent WRAP URL: http://wrap.warwick.ac.uk/101005 Copyright and reuse: The Warwick Research Archive Portal (WRAP) makes this work of researchers of the University of Warwick available open access under the following conditions. This article is made available under the Creative Commons Attribution 4.0 International license (CC BY 4.0) and may be reused according to the conditions of the license. For more details see: http://creativecommons.org/licenses/by/4.0/ A note on versions: The version presented in WRAP is the published version, or, version of record, and may be cited as it appears here. For more information, please contact the WRAP Team at: [email protected] warwick.ac.uk/lib-publications AI & SOCIETY https://doi.org/10.1007/s00146-018-0820-1 ORIGINAL ARTICLE The brain as artificial intelligence: prospecting the frontiers of neuroscience Steve Fuller1 Received: 12 March 2017 / Accepted: 24 January 2018 © The Author(s) 2018. This article is an open access publication Abstract This article explores the proposition that the brain, normally seen as an organ of the human body, should be understood as a biologically based form of artificial intelligence, in the course of which the case is made for a new kind of ‘brain exception- alism’. After noting that such a view was generally assumed by the founders of AI in the 1950s, the argument proceeds by drawing on the distinction between science—in this case neuroscience—adopting a ‘telescopic’ or a ‘microscopic’ orientation to reality, depending on how it regards its characteristic investigative technologies. The paper concludes by recommending a ‘microscopic’ yet non-reductionist research agenda for neuroscience, in which the brain is seen as an underutilised organ whose energy efficiency is likely to outstrip that of the most powerful supercomputers for the foreseeable future. Keywords Brain exceptionalism · Cognitive agriculture · Microscope · Neural nets · Neuroscience · Reductionism · Supercomputer · Telescope In what follows, I explore the proposition that the brain, nor- long term may see organically grown neural networks—if mally seen as an organ of the human body, should be under- not full-fledged brains—carrying out many if not most of stood as a biologically based form of artificial intelligence the functions that nowadays are taken to be the purview of (AI). As I observe in the “Introduction”, this proposition silicon-based computers. In this respect, I am arguing for was assumed by the founders of AI in the 1950s, though it a new kind of ‘brain exceptionalism’, one based not on the has been generally side-lined over the course of AI’s history. brain’s natural mysteries but on its relative energy efficiency However, advances in both neuroscience and more conven- vis-à-vis competing (silicon) technologies. tional AI make it interesting to consider the issue anew. The main body of the paper approaches the matter from a dis- tinction that the philosopher Bas van Fraassen has drawn in 1 Introduction: the brain as a work terms of science adopting a ‘telescopic’ or a ‘microscopic’ in progress orientation to reality. The history of neuroscience has exhib- ited both tendencies from its inception, not least in terms of The history of artificial intelligence (AI) is normally told in the alternative functions performed by the field’s character- terms of three research strategies that coalesced in the wake istic technologies. Appreciating the full implications of this of the Macy Conferences, which were first held at Dartmouth distinction requires escaping from the ‘reductionist’ prob- College in 1956: one focused on neural networks, one on lematic that continues to haunt philosophical discussions of cybernetic systems and one simply on computational power neuroscience’s aspirations as a mode of inquiry. As becomes (Gardner 1985: Chap. 6). The third, which carried the least clear by the Conclusion, my own preference is for an ambi- ontological baggage, became the dominant understanding of tious ‘microscopic’ agenda for neuroscience, which in the AI. A striking feature of all three strategies—which perhaps helps to explain the dominance of the third—is the failure to clearly distinguish modelling and making the target phe- * Steve Fuller [email protected] nomenon, ‘intelligence’ (Dupuy 2000: Chap. 2). This may have been to do with the preponderance of mathematicians 1 Auguste Comte Chair in Social Epistemology, Department among the AI founders, who privileged the search for ana- of Sociology, University of Warwick, Coventry CV4 7AL, lytic clarity over the need to provide concrete instantiation, UK Vol.:(0123456789)1 3 AI & SOCIETY sometimes to the exasperation of interloping physicists assumption that the way in which biological brains instanti- (Malapi-Nelson 2017: Chaps. 7–8). ate intelligence captures something deep about intelligence To be sure, this persistent ambiguity has contributed to itself, which remains untapped if we only focus on the the Turing Test’s iconic status as popular culture’s portal brain’s known functionalities, as these may simply reflect the into the world of AI. After all, the test’s fascination rests on specific pathways through which brains have so far evolved. the difficulty that humans normally have in distinguishing a In other words, what McCulloch and Pitts had captured ‘natural’ from an ‘artificial’ intelligence (i.e. a fellow human was the idea that the organic is organized—not only that from an android) simply based on the candidate being’s per- brains perform as they do because of the environments in formance in response to their questions. It led one of the AI which they have been so far placed, but also that in differ- pioneers, psychiatrist Ross Ashby, to quip that passing the ent environments brains would behave in (potentially better) Turing test simply meant that questioner and respondent had ways precisely because of their material composition. The established a common standard of something they agree to latter point, which also informs the thinking behind contem- call ‘intelligence’ without saying what it is. For Ashby, the porary ‘synthetic biology’, envisages the genes and cells that ‘brain’ was whatever could reliably enable an entity to pass compose ‘natural’ living beings as literal building blocks— such a test, regardless of its material composition (Malapi- self-contained energy modules, if you will—that could be Nelson 2017: Chap. 6). This orientation to intelligence is ‘organized’ to produce new (and improved) bio-based edi- rather like the economist’s view that a good’s value is simply fices by drawing on the expertise of both biologists and engi- the price that it fetches in a free market or, perhaps more to neers (Church and Regis 2012; cf.; Fuller 2016a). The most the point, the Popperian view that what both scientists and obvious precedent for this line of thought is capitalism’s lay people mean by ‘true’ is simply whatever passes critical default understanding of labour as an underutilised produc- scrutiny. In each of these cases, the normative standard— tive force or unexploited potential that may be improved by intelligence, value and truth—functions as a reversible con- ‘better organisation’. Indeed, one of the British AI pioneers, vention rather than as a fixed essence. Stafford Beer, picked up on this point to become one of the However, the blurring of the natural/artificial distinction early exemplars of the ‘management guru’ (Pickering 2010: cannot simply be reduced to the abstractness—or ontologi- Chap. 6). The final section of this paper revisits this sensibil- cal neutrality—of the AI research agenda itself. After all, ity via ‘cognitive agriculture’. the one arguably ‘natural’ focus of AI research, the brain, To be sure, more than a half-century after AI’s original has itself always been regarded within this tradition a work encounters with the brain, biochemistry continues to offer in progress, an artefact in the making. Indeed, Ashby pub- much the same appraisal of the organ’s self-creative capac- lished a book in 1952 with the bold title, Design for a Brain. ity, even in its natural carbon-centred terms: human genes Moreover, the original AI researchers specifically focused produce 100 billion neurons that are only locally organised on the brain—Warren McCulloch and Walter Pitts—did not at an individual’s birth but become more globally integrated try to model the entire organ in terms of its multiple known through repeated experience and feedback over the individ- functionalities, which one might think would have been the ual’s lifetime (Williams and Frausto da Silva 2007: 483). most logical way to go and is probably the dominant sense Observations of this sort have helped to fuel the recent inter- in which AI researchers today think about the prospect of est in ‘epigenetics’, with its promise of training up an always computers possessing ‘brains’. Instead they constructed an only partially formed brain via a cocktail of chemicals and idealised version of the biological brain’s functional equiva- external stimuli, even to the point of allowing for an ‘adult lent to an atom, the neuron. Vast combinations of these neu- neurogenesis’ (Rubin 2009). rons were portrayed as engaged in the parallel processing of But more directly relevant for our purposes, this general data from a variety of sensory sources, resulting in emergent ‘neuroplastic’ understanding of the brain casts doubt on the patterns of co-activity which over time become so integrated popular but crude distinction that the brain is ‘hardware’ to that they then—and only then—constitute a ‘brain’. the mind’s ‘software’. The metaphor jars because it suggests In effect, McCulloch and Pitts had proposed to grow a too strong an association between computer hardware and brain in a silicon setting.