Larger GPU-Accelerated Brain Simulations with Procedural Connectivity

Total Page:16

File Type:pdf, Size:1020Kb

Larger GPU-Accelerated Brain Simulations with Procedural Connectivity bioRxiv preprint doi: https://doi.org/10.1101/2020.04.27.063693; this version posted April 28, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. Larger GPU-accelerated brain simulations with procedural connectivity James C Knighta,1 and Thomas Nowotnya aCentre for Computational Neuroscience and Robotics, School of Engineering and Informatics, University of Sussex, Brighton, United Kingdom This manuscript was compiled on April 27, 2020 1 Large-scale simulations of spiking neural network models are an im- large proportion of each chip to memory. However, while such 25 2 portant tool for improving our understanding of the dynamics and ul- on-chip memory is fast, it can only be fabricated at relatively 26 3 timately the function of brains. However, even small mammals such low density so that many of these systems economize – either 27 12 4 as mice have on the order of 1 × 10 synaptic connections which, in by reducing the maximum number of synapses per neuron to as 28 5 simulations, are each typically charaterized by at least one floating- few as 256 or by reducing the precision of the synaptic weights 29 6 point value. This amounts to several terabytes of data – an unreal- to 6 (9), 4 (5) or even 1 bit (7). This allows some classes 30 7 istic memory requirement for a single desktop machine. Large mod- of spiking neural networks to be simulated very efficiently, 31 8 els are therefore typically simulated on distributed supercomputers but reducing the degree of connectivity to fit within the con- 32 9 which is costly and limits large-scale modelling to a few privileged straints of current neuromorphic systems inevitably changes 33 10 research groups. In this work, we describe extensions to GeNN – our the dynamics of brain simulations (10). Unlike most other 34 11 Graphical Processing Unit (GPU) accelerated spiking neural network neuromorphic systems, the SpiNNaker (6) neuromorphic su- 35 12 simulator – that enable it to ‘procedurally’ generate connectivity and percomputer is fully programmable and combines large on-chip 36 13 synaptic weights ‘on the go’ as spikes are triggered, instead of stor- and external memories, distributed across the system, which 37 14 ing and retrieving them from memory. We find that GPUs are well- enables real-time simulation of large-scale models (11). This 38 15 suited to this approach because of their raw computational power is promising for the future but, due to its prototype nature, 39 16 which, due to memory bandwidth limitations, is often under-utilised the availability of SpiNNaker hardware is limited and even 40 17 when simulating spiking neural networks. We demonstrate the value moderately-sized simulations still require a physically large 41 3 18 of our approach with a recent model of the Macaque visual cortex system (9 boards for a model with around 10 × 10 neurons 42 6 9 6 19 consisting of 4.13 × 10 neurons and 24.2 × 10 synapses. Using and 300 × 10 synapses (11)). 43 20 our new method, it can be simulated on a single GPU – a signifi- Modern GPUs have relatively little on-chip memory and, 44 21 cant step forward in making large-scale brain modelling accessible instead, dedicate the majority of their silicon area to arith- 45 22 to many more researchers. Our results match those obtained on a metic logic units. GPUs use dedicated hardware to rapidly 46 23 supercomputer and the simulation runs up to 35 % faster on a single switch between tasks so that the latency of accessing exter- 47 24 high-end GPU than previously on over 1000 supercomputer nodes. nal memory can be ‘hidden’ behind computation, as long as 48 there is sufficient computation to be performed. For example, 49 spiking neural networks | GPU | high-performance computing | brain the memory latency of a typical modern GPU can be com- 50 simulation pletely hidden if each CUDA core performs approximately 10 51 52 6 arithmetic operations per byte of data accessed from memory. 1 he brain of a mouse has around 70 × 10 neurons, but 12 2 Tthis number is dwarfed by the 1 × 10 synapses which 3 connect them (1). In computer simulations of spiking neu- 4 ral networks, propagating spikes involves adding the synaptic Significance Statement 5 DRAFT input from each spiking presynaptic neuron to the postsy- Simulations are an important tool for investigating how brains 6 naptic neurons. The information describing which neurons work. However, in order to faithfully reproduce some of the 7 are synaptically connected and with what weight is typically features found in biological brains, large models are required. 8 generated before a simulation is run and stored in large arrays. Simulating such models has, until now, required so much mem- 9 For large-scale brain models this creates high memory require- ory that it could only be done on large, expensive supercom- 10 ments, so that they can typically only be simulated on large puters. In this work, we present a new method for simulating 11 distributed computer systems using software such as NEST (2) large models that significantly reduces memory requirements. 12 or NEURON (3). By careful design, these simulators can keep This method is particularly well-suited for use on Graphical 13 the memory requirements for each node constant, even when a Processing Units (GPUs), which are a common fixture in many 14 simulation is distributed across thousands of nodes (4). How- workstations. We demonstrate that using our new method we 15 ever, high performance computer (HPC) systems are bulky, can not only simulate a very large brain model on a single GPU, 16 expensive and consume a lot of power and are hence typi- but also do so up to 35 % faster than in previous supercomputer 17 cally shared resources, only accessible to a limited number of simulations. 18 researchers and for time-limited investigations. 19 Neuromorphic systems (5–9) take inspiration from the brain J.K. and T.N. wrote the paper. T.N. is the original developer of GeNN. J.K. is currently the primary 20 and have been developed specifically for simulating large spik- GeNN developer and was responsible for extending the code generation approach to the proce- dural simulation of synaptic connectivity. J.K. performed the experiments and the analysis of the 21 ing neural networks more efficiently. One particular relevant results that are presented in this work. 22 feature of the brain is that its memory elements – the synapses The authors declare no conflict of interest. 23 – are co-located with the computing elements – the neurons. 1 24 In neuromorphic systems, this often translates to dedicating a To whom correspondence should be addressed. E-mail: [email protected] www.pnas.org/cgi/doi/10.1073/pnas.XXXXXXXXXX PNAS | April 27, 2020 | vol. XXX | no. XX | 1–6 bioRxiv preprint doi: https://doi.org/10.1101/2020.04.27.063693; this version posted April 28, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. 53 Unfortunately, propagating a spike in a spiking neural network We implemented procedural connectivity in GeNN by repur- 112 54 simulation is likely to require accessing around 8 B of mem- posing our previously developed parallel initialisation methods. 113 55 ory but perform many fewer than the required 80 instructions. Instead of running them once for all synapses at the beginning 114 56 This makes spike propagation highly memory bound. Nonethe- of the simulation, we rerun the methods during the simula- 115 57 less, we have shown in previous work (12) that, as GPUs have tion to regenerate the outgoing synapses of each neuron that 116 58 significantly higher total memory bandwidth than even the fires a spike and immediately use the identified connections 117 3 59 fastest CPU, moderately sized models of around 10 × 10 neu- and weights to run the post-synaptic code which calculates 118 9 60 rons and 1 × 10 synapses can be simulated on a single GPU the effect of the spike onto other neurons. This is possible 119 61 with competitive speed and energy consumption. However, because the outgoing synaptic connections from each neuron 120 62 individual GPUs do not have enough memory to simulate are typically largely independent from those of other neurons 121 63 larger brain models and, although small numbers of GPUs as we shall see from typical examples below. 122 64 can be connected using the high-speed NVLink (13) intercon- In the absence of knowledge of the exact microscopic connec- 123 65 nect, larger GPU clusters suffer from the same communication tivity in the brain, there are a number of typical connectivity 124 66 overheads as any other distributed HPC system. schemes that are used in brain models. We will now discuss 125 67 In this work, we present a novel approach that uses the two typical examples and how they can be implemented ef- 126 68 large amount of computational power available on a GPU to ficiently on a GPU. One very common connectivity scheme 127 69 reduce both memory and memory bandwidth requirements is the ‘fixed probability connector’ for which each neuron in 128 70 and enable large-scale brain simulations on a single GPU the presynaptic population is connected to each neuron in 129 71 workstation.
Recommended publications
  • Reconstruction and Simulation of the Cerebellar Microcircuit: a Scaffold Strategy to Embed Different Levels of Neuronal Details
    Reconstruction and simulation of the cerebellar microcircuit: a scaffold strategy to embed different levels of neuronal details Claudia Casellato, Elisa Marenzi, Stefano Casali, Chaitanya Medini, Alice Geminiani, Alessandra Pedrocchi, Egidio D’Angelo Department of Brain and Behavioral Sciences, University of Pavia, Italy Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milan, Italy Computational models allow propagating microscopic phenomena into large‐scale networks and inferencing causal relationships across scales. Here we reconstruct the cerebellar circuit by bottom‐up modeling, reproducing the peculiar properties of this structure, which shows a quasi‐crystalline geometrical organization well defined by convergence/divergence ratios of neuronal connections and by the anisotropic 3D orientation of dendritic and axonal processes [1]. Therefore, a cerebellum scaffold model has been developed and tested. It maintains scalability and can be flexibly handled to incorporate neuronal properties on multiple scales of complexity. The cerebellar scaffold includes the canonical neuron types: Granular cell, Golgi cell, Purkinje cell, Stellate and Basket cells, Deep Cerebellar Nuclei cell. Placement was based on density and encumbrance values, connectivity on specific geometry of dendritic and axonal fields, and on distance‐based probability. In the first release, spiking point‐neuron models based on Integrate&Fire dynamics with exponential synapses were used. The network was run in the neural simulator pyNEST. Complex spatiotemporal patterns of activity, similar to those observed in vivo, emerged [2]. For a second release of the microcircuit model, an extension of the generalized Leaky Integrate&Fire model has been developed (E‐GLIF), optimized for each cerebellar neuron type and inserted into the built scaffold [3].
    [Show full text]
  • The Human Brain Project—Synergy Between Neuroscience, Computing, Informatics, and Brain-Inspired Technologies
    COMMUNITY PAGE The Human Brain ProjectÐSynergy between neuroscience, computing, informatics, and brain-inspired technologies 1,2 3 4 5 Katrin AmuntsID *, Alois C. Knoll , Thomas LippertID , Cyriel M. A. PennartzID , 6 7 8 9 10 Philippe RyvlinID , Alain DestexheID , Viktor K. Jirsa , Egidio D'Angelo , Jan G. BjaalieID 1 Institute for Neuroscience and Medicine (INM-1), Forschungszentrum JuÈlich, Germany, 2 C. and O. Vogt Institute for Brain Research, University Hospital DuÈsseldorf, Heinrich Heine University DuÈsseldorf, DuÈsseldorf, Germany, 3 Institut fuÈr Informatik VI, Technische UniversitaÈt MuÈnchen, Garching bei MuÈnchen, a1111111111 Germany, 4 JuÈlich Supercomputing Centre, Institute for Advanced Simulation, Forschungszentrum JuÈlich, a1111111111 Germany, 5 Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, the a1111111111 Netherlands, 6 Department of Clinical Neurosciences, Centre Hospitalo-Universitaire Vaudois (CHUV) and a1111111111 University of Lausanne, Lausanne, Switzerland, 7 Unite de Neurosciences, Information & Complexite a1111111111 (UNIC), Centre National de la Recherche Scientifique (CNRS), Gif-sur-Yvette, France, 8 Institut de Neurosciences des Systèmes, Inserm UMR1106, Aix-Marseille UniversiteÂ, Faculte de MeÂdecine, Marseille, France, 9 Department of Brain and Behavioral Science, Unit of Neurophysiology, University of Pavia, Pavia, Italy, 10 Institute of Basic Medical Sciences, University of Oslo, Oslo, Norway * [email protected] OPEN ACCESS Citation: Amunts K, Knoll AC, Lippert T, Pennartz CMA, Ryvlin P, Destexhe A, et al. (2019) The Abstract Human Brain ProjectÐSynergy between neuroscience, computing, informatics, and brain- The Human Brain Project (HBP) is a European flagship project with a 10-year horizon aim- inspired technologies. PLoS Biol 17(7): e3000344. ing to understand the human brain and to translate neuroscience knowledge into medicine https://doi.org/10.1371/journal.pbio.3000344 and technology.
    [Show full text]
  • NEST Desktop
    NEST Conference 2019 A Forum for Users & Developers Conference Program Norwegian University of Life Sciences 24-25 June 2019 NEST Conference 2019 24–25 June 2019, Ås, Norway Monday, 24th June 11:00 Registration and lunch 12:30 Opening (Dean Anne Cathrine Gjærde, Hans Ekkehard Plesser) 12:50 GeNN: GPU-enhanced neural networks (James Knight) 13:25 Communication sparsity in distributed Spiking Neural Network Simulations to improve scalability (Carlos Fernandez-Musoles) 14:00 Coffee & Posters 15:00 Sleep-like slow oscillations induce hierarchical memory association and synaptic homeostasis in thalamo-cortical simulations (Pier Stanislao Paolucci) 15:20 Implementation of a Frequency-Based Hebbian STDP in NEST (Alberto Antonietti) 15:40 Spike Timing Model of Visual Motion Perception and Decision Making with Reinforcement Learning in NEST (Petia Koprinkova-Hristova) 16:00 What’s new in the NEST user-level documentation (Jessica Mitchell) 16:20 ICEI/Fenix: HPC and Cloud infrastructure for computational neuroscientists (Jochen Eppler) 17:00 NEST Initiative Annual Meeting (members-only) 19:00 Conference Dinner Tuesday, 25th June 09:00 Construction and characterization of a detailed model of mouse primary visual cortex (Stefan Mihalas) 09:45 Large-scale simulation of a spiking neural network model consisting of cortex, thalamus, cerebellum and basal ganglia on K computer (Jun Igarashi) 10:10 Simulations of a multiscale olivocerebellar spiking neural network in NEST: a case study (Alice Geminiani) 10:30 NEST3 Quick Preview (Stine Vennemo & Håkon Mørk)
    [Show full text]
  • An Overview of Mind Uploading
    IOSR Journal of Engineering (IOSR JEN) www.iosrjen.org ISSN (e): 2250-3021, ISSN (p): 2278-8719 PP 18-25 An Overview of Mind Uploading Mrs. Mausami Sawarkar1, Mr. Dhiraj Rane2 Dept of CSE, Priydarshani J L CoE, Nagpur Dept. of CS, GHRIIT, Nagpur Abstract: Mind uploading is an ongoing area of active research, bringing together ideas from neuroscience, computer science, engineering, and philosophy. Realistically, mind uploading likely lies many decades in the future, but the short-term offers the possibility of advanced neural prostheses that may benefit us. Mind uploading is the conceptual futuristic technology of transferring human minds tocomputer hardware using whole-brain emulation process. In this research work, we have discuss the brief review of the technological prospectsfor mind uploading, a range of philosophical and ethical aspects of the technology. We have also summarizes various issues for mind uploading. There are many technologies are working on these issues will be summarize. These include questions about whether uploads will have consciousness and whether uploading willpreserve personal identity, as well as what impact on society a working uploading technology is likelyto have and whether these impacts are desirable. The issue of whether we ought to move forwardstowards uploading technology remains as unclear as ever. Keywords: Mind Uploading, Whole brain simulation, I. Introduction Mind uploading is an ongoing area of active research, bringing together ideas from neuroscience, computer science, engineering, and philosophy. Realistically, mind uploading likely lies many decades in the future, but the short-term offers the possibility of advanced neural prostheses that may benefit us. Mind uploading is a popular term for a process by which the mind, a collection of memories, personality, and attributes of a specific individual, is transferred from its original biological brain to an artificial computational substrate.
    [Show full text]
  • Brain Modelling As a Service: the Virtual Brain on EBRAINS
    Brain Modelling as a Service: The Virtual Brain on EBRAINS Michael Schirnera,b,c,d,e, Lia Domidef, Dionysios Perdikisa,b, Paul Triebkorna,b,g, Leon Stefanovskia,b, Roopa Paia,b, Paula Prodanf, Bogdan Valeanf, Jessica Palmera,b, Chloê Langforda,b, André Blickensdörfera,b, Michiel van der Vlagh, Sandra Diaz-Pierh, Alexander Peyserh, Wouter Klijnh, Dirk Pleiteri,j, Anne Nahmi, Oliver Schmidk, Marmaduke Woodmang, Lyuba Zehll, Jan Fousekg, Spase Petkoskig, Lionel Kuschg, Meysam Hashemig, Daniele Marinazzom,n, Jean-François Mangino, Agnes Flöelp,q, Simisola Akintoyer, Bernd Carsten Stahls, Michael Cepict, Emily Johnsont, Gustavo Decou,v,w,x, Anthony R. McIntoshy, Claus C. Hilgetagz, a, Marc Morgank, Bernd Schulleri, Alex Uptonb, Colin McMurtrieb, Timo Dickscheidl, Jan G. Bjaaliec, Katrin Amuntsl,d, Jochen Mersmanne, Viktor Jirsag, Petra Rittera,b,c,d,e,* aBerlin Institute of Health at Charité – Universitätsmedizin Berlin, Charitéplatz 1, 10117 Berlin, Germany bCharité – Universitätsmedizin Berlin, corporate memBer of Freie Universität Berlin and HumBoldt Universität zu Berlin, Department of Neurology with Experimental Neurology, Charitéplatz 1, 10117 Berlin, Germany cBernstein Focus State Dependencies of Learning & Bernstein Center for Computational Neuroscience, Berlin, Germany dEinstein Center for Neuroscience Berlin, Charitéplatz 1, 10117 Berlin, Germany eEinstein Center Digital Future, Wilhelmstraße 67, 10117 Berlin, Germany fCodemart, Str. Petofi Sandor, Cluj Napoca, Romania gInstitut de Neurosciences des Systèmes, Aix Marseille Université,
    [Show full text]
  • Computational Tools to Investigate Neurological Disorders
    International Journal of Molecular Sciences Review Modeling Neurotransmission: Computational Tools to Investigate Neurological Disorders Daniela Gandolfi 1, Giulia Maria Boiani 1, Albertino Bigiani 1,2 and Jonathan Mapelli 1,2,* 1 Department of Biomedical, Metabolic and Neural Sciences, University of Modena and Reggio Emilia, Via Campi 287, 41125 Modena, Italy; daniela.gandolfi@unimore.it (D.G.); [email protected] (G.M.B.); [email protected] (A.B.) 2 Center for Neuroscience and Neurotechnology, University of Modena and Reggio Emilia, Via Campi 287, 41125 Modena, Italy * Correspondence: [email protected]; Tel.: +39-059-205-5459 Abstract: The investigation of synaptic functions remains one of the most fascinating challenges in the field of neuroscience and a large number of experimental methods have been tuned to dissect the mechanisms taking part in the neurotransmission process. Furthermore, the understanding of the insights of neurological disorders originating from alterations in neurotransmission often requires the development of (i) animal models of pathologies, (ii) invasive tools and (iii) targeted pharmacological approaches. In the last decades, additional tools to explore neurological diseases have been provided to the scientific community. A wide range of computational models in fact have been developed to explore the alterations of the mechanisms involved in neurotransmission following the emergence of neurological pathologies. Here, we review some of the advancements in the development of computational methods employed to investigate neuronal circuits with a particular focus on the application to the most diffuse neurological disorders. Citation: Gandolfi, D.; Boiani, G.M.; Bigiani, A.; Mapelli, J. Modeling Keywords: synaptic transmission; computational modeling; synaptic plasticity; neurological Neurotransmission: Computational disorders Tools to Investigate Neurological Disorders.
    [Show full text]
  • The Grateful Un-Dead? Philosophical and Social Implications of Mind-Uploading
    1 July 23, 2020 The grateful Un-dead? Philosophical and Social Implications of Mind-Uploading Ivan William Kelly Professor Emeritus, University of Saskatchewan, Saskatoon, SK, Canada Email address: [email protected] Abstract: The popular belief that our mind either depends on or (in stronger terms) is identical with brain functions and processes, along with the belief that advances in technological in virtual reality and computability will continue, has contributed to the contention that one-day (perhaps this century) it may be possible to transfer one’s mind (or a simulated copy) into another body (physical or virtual). This is called mind-uploading or whole brain emulation. This paper serves as an introduction to the area and some of the major issues that arise from the notion of mind- uploading. The topic of mind/brain uploading is, at present, both a philosophical and scientific minefield. This paper ties together several thoughts on the topic in the popular, philosophical, and scientific literature, and brings up some related issues that might be worth exploring. A challenge is made that uploading a mind from a brain may have intrinsic limitations on what can be uploaded: whatever is uploaded may not be recognizably human, or will be a very limited version. Keywords: mind uploading, brain emulation, mind transfer, virtual reality, robots, android, the mind-body problem, mind transfer, trans-human, 1.Introduction What if (a big if) you actually could upload or transfer your mind into a computer network or a robot? That is, you would now have your mind in a different body or ‘live in’ a virtual reality (or virtual realities).
    [Show full text]
  • The Whole Brain Architecture Approach: Accelerating the Development of Artificial General Intelligence by Referring to the Brain
    The whole brain architecture approach: Accelerating the development of artificial general intelligence by referring to the brain Hiroshi Yamakawaa,b,c aThe Whole Brain Architecture Initiative, Nishikoiwa 2-19-21, Edogawa-ku, Tokyo, 133-0057, Japan bThe University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-0033, Japan cRIKEN, 6-2-3, Furuedai, Suita, Osaka 565-0874, Japan Abstract The vastness of the design space created by the combination of a large num- ber of computational mechanisms, including machine learning, is an obstacle to creating an artificial general intelligence (AGI). Brain-inspired AGI devel- opment, in other words, cutting down the design space to look more like a biological brain, which is an existing model of a general intelligence, is a promis- ing plan for solving this problem. However, it is difficult for an individual to design a software program that corresponds to the entire brain because the neuroscientific data required to understand the architecture of the brain are extensive and complicated. The whole-brain architecture approach divides the brain-inspired AGI development process into the task of designing the brain reference architecture (BRA)|the flow of information and the diagram of cor- responding components|and the task of developing each component using the BRA. This is called BRA-driven development. Another difficulty lies in the extraction of the operating principles necessary for reproducing the cognitive- behavioral function of the brain from neuroscience data. Therefore, this study proposes the Structure-constrained Interface Decomposition (SCID) method, which is a hypothesis-building method for creating a hypothetical component diagram consistent with neuroscientific findings.
    [Show full text]
  • Why Is There No Successful Whole Brain Simulation (Yet)?
    Biological Theory https://doi.org/10.1007/s13752-019-00319-5 ORIGINAL ARTICLE Why is There No Successful Whole Brain Simulation (Yet)? Klaus M. Stiefel1,3 · Daniel S. Brooks2 Received: 14 June 2018 / Accepted: 1 March 2019 © Konrad Lorenz Institute for Evolution and Cognition Research 2019 Abstract With the advent of powerful parallel computers, efforts have commenced to simulate complete mammalian brains. How- ever, so far none of these efforts has produced outcomes close to explaining even the behavioral complexities of animals. In this article, we suggest four challenges that ground this shortcoming. First, we discuss the connection between hypothesis testing and simulations. Typically, efforts to simulate complete mammalian brains lack a clear hypothesis. Second, we treat complications related to a lack of parameter constraints for large-scale simulations. To demonstrate the severity of this issue, we review work on two small-scale neural systems, the crustacean stomatogastric ganglion and the Caenorhabditis elegans nervous system. Both of these small nervous systems are very thoroughly, but not completely understood, mainly due to issues with variable and plastic parameters. Third, we discuss the hierarchical structure of neural systems as a principled obstacle to whole-brain simulations. Different organizational levels imply qualitative differences not only in structure, but in choice and appropriateness of investigative technique and perspective. The challenge of reconciling different levels also undergirds the challenge of simulating and hypothesis testing, as modeling a system is not the same thing as simulating it. Fourth, we point out that animal brains are information processing systems tailored very specifically for the ecological niches the respective animals live in.
    [Show full text]
  • 2020 Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy
    http://gcrinstitute.org 2020 Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy McKenna Fitzgerald, Aaron Boddy, & Seth D. Baum Global Catastrophic Risk Institute Cite as: McKenna Fitzgerald, Aaron Boddy, and Seth D. Baum, 2020. 2020 Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy. Global Catastrophic Risk Institute Technical Report 20-1. Technical reports are published to share ideas and promote discussion. They have not necessarily gone through peer review. The views therein are the authors’ and are not necessarily the views of the Global Catastrophic Risk Institute. Executive Summary Artificial general intelligence (AGI) is artificial intelligence (AI) that can reason across a wide range of domains. While most AI research and development (R&D) deals with narrow AI, not AGI, there is some dedicated AGI R&D. If AGI is built, its impacts could be profound. Depending on how it is designed and used, it could either help solve the world’s problems or cause catastrophe, possibly even human extinction. This paper presents a survey of AGI R&D projects that are active in 2020 and updates a previous survey of projects active in 2017. Both surveys attempt to identify every active AGI R&D project and characterize them in terms of relevance to ethics, risk, and policy, focusing on seven attributes: The type of institution in which the project is based Whether the project publishes open-source code Whether the project has military connections The nation(s) in which the project is based The project’s goals for its AGI The extent of the project’s engagement with AGI safety issues The overall size of the project The surveys use openly published information as found in scholarly publications, project websites, popular media articles, and other websites.
    [Show full text]
  • 1 Simulation in Computational Neuroscience Liz Irvine (Cardiff
    Simulation in computational neuroscience Liz Irvine (Cardiff) Word count: 5187 1. Introduction In computational neuroscience the term ‘model’ is often used to capture simulations as well, but simulations have distinct features, and distinct epistemic uses. I follow Parker’s (2009) definition of a simulation as “a time-ordered sequence of states that serves as a representation of some other time-ordered sequence of states [e.g. of the target system]” (p. 486). From a starting state at t0, the program uses an algorithm to calculate the state of the system at t1, and given these values, the program can calculate the next state of the system at t2, and so on. Simulations are used in cognitive neuroscience for a variety of reasons (see Connectionism Chapter for more detailed discussion). One of interest here is that simulation offers a scaffolded way to investigate the internal functioning of a complex system without necessarily needing a lot of experimental data about that internal functioning. That is, unlike some standard approaches to simulation in other disciplines (Winsberg, 2010), simulations in neuroscience are often built without much background knowledge about the system they are (sometimes not very closely) trying to simulate. A description of this is Corrado and Doya’s ‘model-in-the-middle’ approach (Corrado & Doya, 2007; Corrado et al. 2009). They argue that investigating the process of decision making using standard methods in systems neuroscience is massively underconstrained. The standard method involves manipulating some variable of interest (e.g. visual input) and seeing what effect this has on other variables (e.g. motor action), and from there, trying to abstract away an account of decision making.
    [Show full text]
  • BRAIN in the SHELL Assessing the Stakes and the Transformative Potential of the Human Brain Project
    7 BRAIN IN THE SHELL Assessing the Stakes and the Transformative Potential of the Human Brain Project Philipp Haueis and Jan Slaby Introduction The Human Brain Project (HBP) is a large-scale European neuroscience and com- puting project that is one of the biggest funding initiatives in the history of brain research.1 With a planned budget of 1.2 billion € over the next decade and building on the prior Blue Brain Project, the project initiated by Henry Markram pursues the ambitious goal of simulating the entire human brain – all the way from genes to cognition – with the help of exascale information and communication technology (ICT). The HBP hopes to thereby produce new, brain-like computing technologies, so-called neuromorphic computers, which would be both highly energy-efficient and usable by the general public. Despite the potentially enormous significance for both brain research and com- puter technology – as well as society and culture – the ambition and approach of the HBP has been a matter of substantial controversy from the beginning, both in the scientific community and the general public. Critics have claimed that the model-based bottom-up approach of the project is scientifically wrong-headed, have accused the project of not being managed transparently, and have attested that the aims of the HBP are too ambitious, such that the project is likely to waste valuable resources for research and infrastructure in Europe. It is currently – in mid-2015 – still a debated question whether the HBP in its current format and direction should be pursued at all (Bartlett, 2015). The controversy was fuelled in early 2014 by an open letter to the European Commission (EC) signed by over 750 European neuroscientists urging a reform of the HBP even before its operational phase was set to begin.
    [Show full text]