Deep Active Inference Agents Using Monte-Carlo Methods
Total Page:16
File Type:pdf, Size:1020Kb

Load more
Recommended publications
-
Deep Active Inference for Autonomous Robot Navigation
Published as a workshop paper at “Bridging AI and Cognitive Science” (ICLR 2020) DEEP ACTIVE INFERENCE FOR AUTONOMOUS ROBOT NAVIGATION Ozan C¸atal, Samuel Wauthier, Tim Verbelen, Cedric De Boom, & Bart Dhoedt IDLab, Department of Information Technology Ghent University –imec Ghent, Belgium [email protected] ABSTRACT Active inference is a theory that underpins the way biological agent’s perceive and act in the real world. At its core, active inference is based on the principle that the brain is an approximate Bayesian inference engine, building an internal gen- erative model to drive agents towards minimal surprise. Although this theory has shown interesting results with grounding in cognitive neuroscience, its application remains limited to simulations with small, predefined sensor and state spaces. In this paper, we leverage recent advances in deep learning to build more complex generative models that can work without a predefined states space. State repre- sentations are learned end-to-end from real-world, high-dimensional sensory data such as camera frames. We also show that these generative models can be used to engage in active inference. To the best of our knowledge this is the first application of deep active inference for a real-world robot navigation task. 1 INTRODUCTION Active inference and the free energy principle underpins the way our brain – and natural agents in general – work. The core idea is that the brain entertains a (generative) model of the world which allows it to learn cause and effect and to predict future sensory observations. It does so by constantly minimising its prediction error or “surprise”, either by updating the generative model, or by inferring actions that will lead to less surprising states. -
Allostasis, Interoception, and the Free Energy Principle
1 Allostasis, interoception, and the free energy principle: 2 Feeling our way forward 3 Andrew W. Corcoran & Jakob Hohwy 4 Cognition & Philosophy Laboratory 5 Monash University 6 Melbourne, Australia 7 8 This material has been accepted for publication in the forthcoming volume entitled The 9 interoceptive mind: From homeostasis to awareness, edited by Manos Tsakiris and Helena De 10 Preester. It has been reprinted by permission of Oxford University Press: 11 https://global.oup.com/academic/product/the-interoceptive-mind-9780198811930?q= 12 interoception&lang=en&cc=gb. For permission to reuse this material, please visit 13 http://www.oup.co.uk/academic/rights/permissions. 14 Abstract 15 Interoceptive processing is commonly understood in terms of the monitoring and representation of the body’s 16 current physiological (i.e. homeostatic) status, with aversive sensory experiences encoding some impending 17 threat to tissue viability. However, claims that homeostasis fails to fully account for the sophisticated regulatory 18 dynamics observed in complex organisms have led some theorists to incorporate predictive (i.e. allostatic) 19 regulatory mechanisms within broader accounts of interoceptive processing. Critically, these frameworks invoke 20 diverse – and potentially mutually inconsistent – interpretations of the role allostasis plays in the scheme of 21 biological regulation. This chapter argues in favour of a moderate, reconciliatory position in which homeostasis 22 and allostasis are conceived as equally vital (but functionally distinct) modes of physiological control. It 23 explores the implications of this interpretation for free energy-based accounts of interoceptive inference, 24 advocating a similarly complementary (and hierarchical) view of homeostatic and allostatic processing. -
Machine Learning for Neuroscience Geoffrey E Hinton
Hinton Neural Systems & Circuits 2011, 1:12 http://www.neuralsystemsandcircuits.com/content/1/1/12 QUESTIONS&ANSWERS Open Access Machine learning for neuroscience Geoffrey E Hinton What is machine learning? Could you provide a brief description of the Machine learning is a type of statistics that places parti- methods of machine learning? cular emphasis on the use of advanced computational Machine learning can be divided into three parts: 1) in algorithms. As computers become more powerful, and supervised learning, the aim is to predict a class label or modern experimental methods in areas such as imaging a real value from an input (classifying objects in images generate vast bodies of data, machine learning is becom- or predicting the future value of a stock are examples of ing ever more important for extracting reliable and this type of learning); 2) in unsupervised learning, the meaningful relationships and for making accurate pre- aim is to discover good features for representing the dictions. Key strands of modern machine learning grew input data; and 3) in reinforcement learning, the aim is out of attempts to understand how large numbers of to discover what action should be performed next in interconnected, more or less neuron-like elements could order to maximize the eventual payoff. learn to achieve behaviourally meaningful computations and to extract useful features from images or sound What does machine learning have to do with waves. neuroscience? By the 1990s, key approaches had converged on an Machine learning has two very different relationships to elegant framework called ‘graphical models’, explained neuroscience. As with any other science, modern in Koller and Friedman, in which the nodes of a graph machine-learning methods can be very helpful in analys- represent variables such as edges and corners in an ing the data. -
Neural Dynamics Under Active Inference: Plausibility and Efficiency
entropy Article Neural Dynamics under Active Inference: Plausibility and Efficiency of Information Processing Lancelot Da Costa 1,2,* , Thomas Parr 2 , Biswa Sengupta 2,3,4 and Karl Friston 2 1 Department of Mathematics, Imperial College London, London SW7 2AZ, UK 2 Wellcome Centre for Human Neuroimaging, Institute of Neurology, University College London, London WC1N 3BG, UK; [email protected] (T.P.); [email protected] (B.S.); [email protected] (K.F.) 3 Core Machine Learning Group, Zebra AI, London WC2H 8TJ, UK 4 Department of Bioengineering, Imperial College London, London SW7 2AZ, UK * Correspondence: [email protected] Abstract: Active inference is a normative framework for explaining behaviour under the free energy principle—a theory of self-organisation originating in neuroscience. It specifies neuronal dynamics for state-estimation in terms of a descent on (variational) free energy—a measure of the fit between an internal (generative) model and sensory observations. The free energy gradient is a prediction error—plausibly encoded in the average membrane potentials of neuronal populations. Conversely, the expected probability of a state can be expressed in terms of neuronal firing rates. We show that this is consistent with current models of neuronal dynamics and establish face validity by synthesising plausible electrophysiological responses. We then show that these neuronal dynamics approximate natural gradient descent, a well-known optimisation algorithm from information geometry that follows the steepest descent of the objective in information space. We compare the information length of belief updating in both schemes, a measure of the distance travelled in information space that has a Citation: Da Costa, L.; Parr, T.; Sengupta, B.; Friston, K. -
Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning
Deep Neuroevolution: Genetic Algorithms are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning Felipe Petroski Such Vashisht Madhavan Edoardo Conti Joel Lehman Kenneth O. Stanley Jeff Clune Uber AI Labs ffelipe.such, [email protected] Abstract 1. Introduction Deep artificial neural networks (DNNs) are typ- A recent trend in machine learning and AI research is that ically trained via gradient-based learning al- old algorithms work remarkably well when combined with gorithms, namely backpropagation. Evolution sufficient computing resources and data. That has been strategies (ES) can rival backprop-based algo- the story for (1) backpropagation applied to deep neu- rithms such as Q-learning and policy gradients ral networks in supervised learning tasks such as com- on challenging deep reinforcement learning (RL) puter vision (Krizhevsky et al., 2012) and voice recog- problems. However, ES can be considered a nition (Seide et al., 2011), (2) backpropagation for deep gradient-based algorithm because it performs neural networks combined with traditional reinforcement stochastic gradient descent via an operation sim- learning algorithms, such as Q-learning (Watkins & Dayan, ilar to a finite-difference approximation of the 1992; Mnih et al., 2015) or policy gradient (PG) methods gradient. That raises the question of whether (Sehnke et al., 2010; Mnih et al., 2016), and (3) evolution non-gradient-based evolutionary algorithms can strategies (ES) applied to reinforcement learning bench- work at DNN scales. Here we demonstrate they marks (Salimans et al., 2017). One common theme is can: we evolve the weights of a DNN with a sim- that all of these methods are gradient-based, including ES, ple, gradient-free, population-based genetic al- which involves a gradient approximation similar to finite gorithm (GA) and it performs well on hard deep differences (Williams, 1992; Wierstra et al., 2008; Sali- RL problems, including Atari and humanoid lo- mans et al., 2017). -
15 | Variational Inference
15 j Variational inference 15.1 Foundations Variational inference is a statistical inference framework for probabilistic models that comprise unobserv- able random variables. Its general starting point is a joint probability density function over observable random variables y and unobservable random variables #, p(y; #) = p(#)p(yj#); (15.1) where p(#) is usually referred to as the prior density and p(yj#) as the likelihood. Given an observed value of y, the first aim of variational inference is to determine the conditional density of # given the observed value of y, referred to as the posterior density. The second aim of variational inference is to evaluate the marginal density of the observed data, or, equivalently, its logarithm Z ln p(y) = ln p(y; #) d#: (15.2) Eq. (15.2) is commonly referred to as the log marginal likelihood or log model evidence. The log model evidence allows for comparing different models in their plausibility to explain observed data. In the variational inference framework it is not the log model evidence itself which is evaluated, but rather a lower bound approximation to it. This is due to the fact that if a model comprises many unobservable variables # the integration of the right-hand side of (15.2) can become analytically burdensome or even intractable. To nevertheless achieve its two aims, variational inference in effect replaces an integration problem with an optimization problem. To this end, variational inference exploits a set of information theoretic quantities as introduced in Chapter 11 and below. Specifically, the following log model evidence decomposition forms the core of the variational inference approach (Figure 15.1): ln p(y) = F (q(#)) + KL(q(#)kp(#jy)): (15.3) In eq. -
Acetylcholine in Cortical Inference
Neural Networks 15 (2002) 719–730 www.elsevier.com/locate/neunet 2002 Special issue Acetylcholine in cortical inference Angela J. Yu*, Peter Dayan Gatsby Computational Neuroscience Unit, University College London, 17 Queen Square, London WC1N 3AR, UK Received 5 October 2001; accepted 2 April 2002 Abstract Acetylcholine (ACh) plays an important role in a wide variety of cognitive tasks, such as perception, selective attention, associative learning, and memory. Extensive experimental and theoretical work in tasks involving learning and memory has suggested that ACh reports on unfamiliarity and controls plasticity and effective network connectivity. Based on these computational and implementational insights, we develop a theory of cholinergic modulation in perceptual inference. We propose that ACh levels reflect the uncertainty associated with top- down information, and have the effect of modulating the interaction between top-down and bottom-up processing in determining the appropriate neural representations for inputs. We illustrate our proposal by means of an hierarchical hidden Markov model, showing that cholinergic modulation of contextual information leads to appropriate perceptual inference. q 2002 Elsevier Science Ltd. All rights reserved. Keywords: Acetylcholine; Perception; Neuromodulation; Representational inference; Hidden Markov model; Attention 1. Introduction medial septum (MS), diagonal band of Broca (DBB), and nucleus basalis (NBM). Physiological studies on ACh Neuromodulators such as acetylcholine (ACh), sero- indicate that its neuromodulatory effects at the cellular tonine, dopamine, norepinephrine, and histamine play two level are diverse, causing synaptic facilitation and suppres- characteristic roles. One, most studied in vertebrate systems, sion as well as direct hyperpolarization and depolarization, concerns the control of plasticity. The other, most studied in all within the same cortical area (Kimura, Fukuda, & invertebrate systems, concerns the control of network Tsumoto, 1999). -
Model-Free Episodic Control
Model-Free Episodic Control Charles Blundell Benigno Uria Alexander Pritzel Google DeepMind Google DeepMind Google DeepMind [email protected] [email protected] [email protected] Yazhe Li Avraham Ruderman Joel Z Leibo Jack Rae Google DeepMind Google DeepMind Google DeepMind Google DeepMind [email protected] [email protected] [email protected] [email protected] Daan Wierstra Demis Hassabis Google DeepMind Google DeepMind [email protected] [email protected] Abstract State of the art deep reinforcement learning algorithms take many millions of inter- actions to attain human-level performance. Humans, on the other hand, can very quickly exploit highly rewarding nuances of an environment upon first discovery. In the brain, such rapid learning is thought to depend on the hippocampus and its capacity for episodic memory. Here we investigate whether a simple model of hippocampal episodic control can learn to solve difficult sequential decision- making tasks. We demonstrate that it not only attains a highly rewarding strategy significantly faster than state-of-the-art deep reinforcement learning algorithms, but also achieves a higher overall reward on some of the more challenging domains. 1 Introduction Deep reinforcement learning has recently achieved notable successes in a variety of domains [23, 32]. However, it is very data inefficient. For example, in the domain of Atari games [2], deep Reinforcement Learning (RL) systems typically require tens of millions of interactions with the game emulator, amounting to hundreds of hours of game play, to achieve human-level performance. As pointed out by [13], humans learn to play these games much faster. This paper addresses the question of how to emulate such fast learning abilities in a machine—without any domain-specific prior knowledge. -
Respiration, Neural Oscillations, and the Free Energy Principle
fnins-15-647579 June 23, 2021 Time: 17:40 # 1 REVIEW published: 29 June 2021 doi: 10.3389/fnins.2021.647579 Keeping the Breath in Mind: Respiration, Neural Oscillations, and the Free Energy Principle Asena Boyadzhieva1*† and Ezgi Kayhan2,3† 1 Department of Philosophy, University of Vienna, Vienna, Austria, 2 Department of Developmental Psychology, University of Potsdam, Potsdam, Germany, 3 Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany Scientific interest in the brain and body interactions has been surging in recent years. One fundamental yet underexplored aspect of brain and body interactions is the link between the respiratory and the nervous systems. In this article, we give an overview of the emerging literature on how respiration modulates neural, cognitive and emotional processes. Moreover, we present a perspective linking respiration to the free-energy principle. We frame volitional modulation of the breath as an active inference mechanism Edited by: in which sensory evidence is recontextualized to alter interoceptive models. We further Yoko Nagai, propose that respiration-entrained gamma oscillations may reflect the propagation of Brighton and Sussex Medical School, United Kingdom prediction errors from the sensory level up to cortical regions in order to alter higher level Reviewed by: predictions. Accordingly, controlled breathing emerges as an easily accessible tool for Rishi R. Dhingra, emotional, cognitive, and physiological regulation. University of Melbourne, Australia Georgios D. Mitsis, Keywords: interoception, respiration-entrained neural oscillations, controlled breathing, free-energy principle, McGill University, Canada self-regulation *Correspondence: Asena Boyadzhieva [email protected] INTRODUCTION † These authors have contributed “The “I think” which Kant said must be able to accompany all my objects, is the “I breathe” which equally to this work actually does accompany them. -
Active Inference Body Perception and Action for Humanoid Robots
Active inference body perception and action for humanoid robots Guillermo Oliver1, Pablo Lanillos1;2, Gordon Cheng1 Abstract—Providing artificial agents with the same computa- Body perception tional models of biological systems is a way to understand how intelligent behaviours may emerge. We present an active inference body perception and action model working for the first time in a humanoid robot. The model relies on the free energy principle proposed for the brain, where both perception and action goal is to minimise the prediction error through gradient descent on the variational free energy bound. The body state (latent variable) is inferred by minimising the difference between the observed (visual and proprioceptive) sensor values and the predicted ones. Simultaneously, the action makes sensory data sampling to better Action correspond to the prediction made by the inner model. We formalised and implemented the algorithm on the iCub robot and tested in 2D and 3D visual spaces for online adaptation to visual changes, sensory noise and discrepancies between the Fig. 1. Body perception and action via active inference. The robot infers model and the real robot. We also compared our approach with its body (e.g., joint angles) by minimising the prediction error: discrepancy classical inverse kinematics in a reaching task, analysing the between the sensors (visual sv and joint sp) and their expected values (gv(µ) suitability of such a neuroscience-inspired approach for real- and gp(µ)). In the presence of error it changes the perception of its body µ world interaction. The algorithm gave the robot adaptive body and generates an action a to reduce this discrepancy. -
When Planning to Survive Goes Wrong: Predicting the Future and Replaying the Past in Anxiety and PTSD
Available online at www.sciencedirect.com ScienceDirect When planning to survive goes wrong: predicting the future and replaying the past in anxiety and PTSD 1 3 1,2 Christopher Gagne , Peter Dayan and Sonia J Bishop We increase our probability of survival and wellbeing by lay out this computational problem as a form of approxi- minimizing our exposure to rare, extremely negative events. In mate Bayesian decision theory (BDT) [1] and consider this article, we examine the computations used to predict and how miscalibrated attempts to solve it might contribute to avoid such events and to update our models of the world and anxiety and stress disorders. action policies after their occurrence. We also consider how these computations might go wrong in anxiety disorders and According to BDT, we should combine a probability Post Traumatic Stress Disorder (PTSD). We review evidence distribution over all relevant states of the world with that anxiety is linked to increased simulations of the future estimates of the benefits or costs of outcomes associated occurrence of high cost negative events and to elevated with each state. We must then calculate the course of estimates of the probability of occurrence of such events. We action that delivers the largest long-run expected value. also review psychological theories of PTSD in the light of newer, Individuals can only possibly approximately solve this computational models of updating through replay and problem. To do so, they bring to bear different sources of simulation. We consider whether pathological levels of re- information (e.g. priors, evidence, models of the world) experiencing symptomatology might reflect problems and apply different methods to calculate the expected reconciling the traumatic outcome with overly optimistic priors long-run values of alternate courses of action. -
Pnas11052ackreviewers 5098..5136
Acknowledgment of Reviewers, 2013 The PNAS editors would like to thank all the individuals who dedicated their considerable time and expertise to the journal by serving as reviewers in 2013. Their generous contribution is deeply appreciated. A Harald Ade Takaaki Akaike Heather Allen Ariel Amir Scott Aaronson Karen Adelman Katerina Akassoglou Icarus Allen Ido Amit Stuart Aaronson Zach Adelman Arne Akbar John Allen Angelika Amon Adam Abate Pia Adelroth Erol Akcay Karen Allen Hubert Amrein Abul Abbas David Adelson Mark Akeson Lisa Allen Serge Amselem Tarek Abbas Alan Aderem Anna Akhmanova Nicola Allen Derk Amsen Jonathan Abbatt Neil Adger Shizuo Akira Paul Allen Esther Amstad Shahal Abbo Noam Adir Ramesh Akkina Philip Allen I. Jonathan Amster Patrick Abbot Jess Adkins Klaus Aktories Toby Allen Ronald Amundson Albert Abbott Elizabeth Adkins-Regan Muhammad Alam James Allison Katrin Amunts Geoff Abbott Roee Admon Eric Alani Mead Allison Myron Amusia Larry Abbott Walter Adriani Pietro Alano Isabel Allona Gynheung An Nicholas Abbott Ruedi Aebersold Cedric Alaux Robin Allshire Zhiqiang An Rasha Abdel Rahman Ueli Aebi Maher Alayyoubi Abigail Allwood Ranjit Anand Zalfa Abdel-Malek Martin Aeschlimann Richard Alba Julian Allwood Beau Ances Minori Abe Ruslan Afasizhev Salim Al-Babili Eric Alm David Andelman Kathryn Abel Markus Affolter Salvatore Albani Benjamin Alman John Anderies Asa Abeliovich Dritan Agalliu Silas Alben Steven Almo Gregor Anderluh John Aber David Agard Mark Alber Douglas Almond Bogi Andersen Geoff Abers Aneel Aggarwal Reka Albert Genevieve Almouzni George Andersen Rohan Abeyaratne Anurag Agrawal R. Craig Albertson Noga Alon Gregers Andersen Susan Abmayr Arun Agrawal Roy Alcalay Uri Alon Ken Andersen Ehab Abouheif Paul Agris Antonio Alcami Claudio Alonso Olaf Andersen Soman Abraham H.