Charles Boutens Computing Optimal Input Encoding for Memristor Based Reservoir

Total Page:16

File Type:pdf, Size:1020Kb

Charles Boutens Computing Optimal Input Encoding for Memristor Based Reservoir Optimal Input Encoding For Memristor Based Reservoir Computing Charles Boutens Supervisor: Prof. dr. ir. Joni Dambre Counsellor: Prof. dr. ir. Joni Dambre Master's dissertation submitted in order to obtain the academic degree of Master of Science in Engineering Physics Department of Electronics and Information Systems Chair: Prof. dr. ir. Rik Van de Walle Faculty of Engineering and Architecture Academic year 2016-2017 Optimal Input Encoding For Memristor Based Reservoir Computing Charles Boutens Supervisor(s): Prof. dr. ir. Joni Dambre1 Abstract| One of the most promising fields of unconventional ap- With Moore's law reaching its end, alternatives to the proaches to computation might be the brain-inspired field digital computing paradigm are being investigated. These of analogue, neuromorphic computing [4]. Here, the ulti- unconventional approaches range from quantum- and opti- cal computing to promising analogue, neuromorphic imple- mate vision is to use self-organised neural networks consist- mentations. A recent approach towards computation with ing of nano-scale components with variable properties and analogue, physical systems is the emerging field of physi- erroneous behaviour, inevitable at the nano-scale. These cal reservoir computing (PRC). By considering the physical system as an excitable, dynamical medium called a reser- neuromorphic approaches require highly connected com- voir, this approach enables the exploitation of the intrinsic plex neural networks with adaptive synapse-like connec- computational power available in all physical systems. tions. Recently [5] [6] [7] interest has arisen in the function- In this work the RC approach towards computation is ap- alities of locally connected switching networks. These net- plied to networks consisting of resistive switches (RS). A densely connected atomic switch network (ASN) and a net- works consist of switches and memory components such as work build up from T iO2 memristors are used as reservoirs molecular transistors, negative differential resistances [8], in simulation. Both simulation models rely on volatile ex- memristors [6] or atomic switches [9]. tensions of Strukov's widely spread current controlled mem- ristor (CCMR) model. The dynamics of the ASN in simu- Contrary to the digital framework, in analogue com- lation are thoroughly characterized and based on these ob- puting the state of the system can take up any contin- servations two encoding schemes are proposed in order to uous value. Hence processes cannot be reliably repeated solve two reservoir benchmark tasks, memory capacity and NARMA-10 task. Experiments executed with the reservoirs with exact equivalence. Also, in the fabrication process consisting of T iO2 memristors lead to the observation of the of these analogue computation blocks, device variability failure of the RC approach in the absence of volatility of the is inevitable and thus a computational framework aiming used devices. at using these systems should account for both inter- and By comparing the simulation with the actual dynamics of the ASN, it is concluded that the CCMR model, used intra-device variability. to describe the individual atomic switches, fails at captur- This is where the RC approach towards computing with ing some of the fundamental characteristics of the device. By looking at the performance of both the ASN and the recurrent neural networks (RNN) comes into play [10] [11] T iO2 reservoir, and at similar research found in the litera- [12]. Here an artificial, randomly assembled RNN is consid- ture, it is argued that the CCMR models are unsuited for ered as a dynamical, excitable medium called the reservoir, the RC approach. Nevertheless, the real ASN devices do whose internal dynamics are untrained. The RC concept show promising reservoir properties. A voltage controlled memristor model (VCMR), already used in a memristor- relies on two key points. Firstly, the detailed nature of the based RC approach, is suggested to continue further re- reservoir is unimportant, only its overall dynamics play a search. Also, some interesting new architectures are briefly role. Secondly, RC allows for a simple training approach discussed that could increase the performance of these RS reservoirs. where only the readout layer is trained. This flexibility en- Index terms { memristor based reservoir computing, ables the RC approach to be applied to a large variety of neuromorphic computation, physical reservoir computing, dynamical systems. Swapping the RNN reservoir with a memristor models, memristor models for reservoir comput- physical system, leads to what is called physical reservoir ing. computing (PRC). PRC presents the right framework and tools to exploit the analogue computational power, intrin- I. Introduction sic to physical dynamical systems, in a robust way: i.e. Digital computation has been the standard for over half without the need to control the randomness and fluctua- a century, mainly thanks to its extreme robustness against tions of the system. The physical system is considered as a noise and variability. However, with the continuous de- reservoir that maps the input via its own intrinsic dynam- mand for more computational power and the slowing down ics onto a high dimensional feature space. These features of Moores law [1] [2], alternatives to the silicon based com- are read out by measuring the system's state and are sub- putational methods are being developed and researchers sequently combined in a trained way (using any standard are reconsidering the field of unconventional computing regression technique) to produce the desired output. PRC which stayed on the sideline of their Turing machine broth- has been demonstrated in various physical systems, both ers for many years [3]. experimentally and in simulations, ranging from a bucket of water in [13], morphological [14], to very promising pho- 1Department of Electronics and Information Systems, Ghent Uni- tonics implementations [15] [16] [17] [18] [19]. In the light of versity neuromorphic computing, the RC approach has also been applied to networks consisting of nano-scale switching ele- in Figure 1. Here w corresponds to the width of the doped ments based on memristive junctions [20] [21] [22] [23] [24]. region. The total resistance of the memristor device is de- With the advancements in material sciences, self- termined by two variable resistors connected in series. RON assembled networks of memristive devices at the nano-scale and ROF F denote the resistance of the memristor in both have been produced, forming very promising candidates limit cases where w = D and w = 0 respectively. Ohm's in the search for synthetic synapses for the fabrication law leads to the following port equation (1): of neuromorphic systems. These devices exhibit similar properties present in biological neurons, such as hystere- v(t) = RON x + ROF F (1 − x) i(t): (1) sis, short-term and long-term-placticity (STP, LTP), long- where x = w 2 [0; 1]. Applying a bias v across the device term depression etc [6] [5]. One very promising implemen- D will cause the charged dopants to drift towards the undoped tation is a highly interconnected atomic switch network region and hence move the boundary between both regions (ASN) [7] [5] [22] [9]. In the ASN, the individual atomic [27]. Here, the simplest case of ohmic electronic conduc- switches are composed of a AgjAg SjAg metal-insulator- 2 tion and linear ionic drift in a uniform field is considered, metal (MIM) interface, where the switching process is gov- leading to the following state equation: erned mostly by a combination of two phenomena: the formation/annihilation of a conductive metal filament to- dx µR = ON i(t); (2) gether with a bias induced phase transition in the insulator dt D2 layer of the MIM junction. A multi-electrode-array (MEA) for the state variable x. Here µ denotes the average ion is being overgrown by a complex structure of densely con- mobility. The coupled equations 1, 2 take the normal form nected silver nanowires, the ASN. The electrodes on the for a current controlled memristor [28]. This CCMR model MEA can serve both as a source or readout. This densely is used as basis for both the T iO memristor and the ASN connected network structure gives rise to new interesting 2 models in [26] and [22] respectively. properties unseen in single AgjAg SjAg atomic switches. It 2 A simulation model for the ASN has been developed in has been reported[5] that the ASNs show signs of an oper- MATLAB [22]. The CCMR model, introduced above, is ational regime near the "edge of chaos", where the network used as a starting point to describe the behavior of a single exhibits avalanche dynamics, criticality and power law scal- atomic switch. The port equation 1 remains unchanged ing of temporal metastability. More over, the ASN shows with w 2 [0;D], now representing the length of the Ag distributed, continuous network activity caused by the in- nano filament. If w = 0 there is no conducting filament, terplay between filament formation and dissolution across corresponding with a resistance R . For w = D the the whole network. A voltage drop due to the formation of OF F filament is fully formed, the switch is in the ON state, a filament in the network instigates the filament's thermo- R . Next a new state equation is introduced for w: dynamical instability and hence its corresponding decay. ON The second memristive device under investigation is the T iO2 memristor created by HP [25]. This device consists dw(t) Ron = [µv i(t)] Ω(w(t)) − τ (w(t) − D) + η(t); (3) of a thin T iO2 film sandwiched between two platinum elec- dt D trodes. The T iO2 layer consists of a highly conductive + channel (T iO2−x), due to positively charged oxygen va- where µv represents the ionic mobility of Ag and Ω(w) = 2 cancies, and a very narrow insulating T iO2 barrier near [w (D−w)=D ] is a window function, modelling the nonlin- the positive electrode.
Recommended publications
  • Reservoir Transformers
    Reservoir Transformers Sheng Sheny, Alexei Baevskiz, Ari S. Morcosz, Kurt Keutzery, Michael Auliz, Douwe Kielaz yUC Berkeley; zFacebook AI Research [email protected], [email protected] Abstract is more, we find that freezing layers may actually improve performance. We demonstrate that transformers obtain im- Beyond desirable efficiency gains, random lay- pressive performance even when some of the ers are interesting for several additional reasons. layers are randomly initialized and never up- dated. Inspired by old and well-established Fixed randomly initialized networks (Gallicchio ideas in machine learning, we explore a variety and Scardapane, 2020) converge to Gaussian pro- of non-linear “reservoir” layers interspersed cesses in the limit of infinite width (Daniely et al., with regular transformer layers, and show im- 2016), have intriguing interpretations in metric provements in wall-clock compute time until learning (Rosenfeld and Tsotsos, 2019; Giryes convergence, as well as overall performance, et al., 2016), and have been shown to provide on various machine translation and (masked) excellent “priors” either for subsequent learn- language modelling tasks. ing (Ulyanov et al., 2018) or pruning (Frankle and 1 Introduction Carbin, 2018). Fixed layers allow for efficient low-cost hardware implementations (Schrauwen Transformers (Vaswani et al., 2017) have dom- et al., 2007) and can be characterized using only a inated natural language processing (NLP) in re- random number generator and its seed. This could cent years, from large scale machine transla- facilitate distributed training and enables highly tion (Ott et al., 2018) to pre-trained (masked) efficient deployment to edge devices, since it only language modeling (Devlin et al., 2018; Rad- requires transmission of a single number.
    [Show full text]
  • The Promise of Spintronics for Unconventional Computing
    The promise of spintronics for unconventional computing Giovanni Finocchio,1,* Massimiliano Di Ventra,2 Kerem Y. Camsari,3 Karin Everschor‐Sitte,4 Pedram Khalili Amiri,5 and Zhongming Zeng6 1Department of Mathematical and Computer Sciences, Physical Sciences and Earth Sciences, University of Messina, Messina, 98166, Italy ‐ Tel. +39.090.3975555 2Department of Physics, University of California, San Diego, La Jolla, California 92093, USA, ‐ Tel. +1.858.8226447 3School of Electrical and Computer Engineering, Purdue University, West Lafayette, Indiana 47907, USA ‐ Tel. +1.765.4946442 4Institut für Physik, Johannes Gutenberg‐Universität Mainz, DE‐55099 Mainz, Germany ‐ Tel. +49.61313923645 5Department of Electrical and Computer Engineering, Northwestern University, Evanston, Illinois 60208, USA ‐ Tel. +1.847.4671035 6 Key Laboratory of Multifunctional Nanomaterials and Smart Systems,, Suzhou Institute of Nano‐tech and Nano‐bionics, Chinese Academy of Sciences, Ruoshui Road 398, Suzhou 215123, P. R. China ‐ Tel. +86.0512.62872759 1 Abstract Novel computational paradigms may provide the blueprint to help solving the time and energy limitations that we face with our modern computers, and provide solutions to complex problems more efficiently (with reduced time, power consumption and/or less device footprint) than is currently possible with standard approaches. Spintronics offers a promising basis for the development of efficient devices and unconventional operations for at least three main reasons: (i) the low-power requirements of spin-based devices, i.e., requiring no standby power for operation and the possibility to write information with small dynamic energy dissipation, (ii) the strong nonlinearity, time nonlocality, and/or stochasticity that spintronic devices can exhibit, and (iii) their compatibility with CMOS logic manufacturing processes.
    [Show full text]
  • Quantum Reservoir Computing: a Reservoir Approach Toward Quantum Machine Learning on Near-Term Quantum Devices
    Quantum reservoir computing: a reservoir approach toward quantum machine learning on near-term quantum devices Keisuke Fujii1, ∗ and Kohei Nakajima2, y 1Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyama, Toyonaka, Osaka 560-8531, Japan. 2Graduate School of Information Science and Technology, The University of Tokyo, Bunkyo-ku, 113-8656 Tokyo, Japan (Dated: November 11, 2020) Quantum systems have an exponentially large degree of freedom in the number of particles and hence provide a rich dynamics that could not be simulated on conventional computers. Quantum reservoir computing is an approach to use such a complex and rich dynamics on the quantum systems as it is for temporal machine learning. In this chapter, we explain quantum reservoir computing and related approaches, quantum extreme learning machine and quantum circuit learning, starting from a pedagogical introduction to quantum mechanics and machine learning. All these quantum machine learning approaches are experimentally feasible and effective on the state-of-the-art quantum devices. I. INTRODUCTION a useful task like machine learning, now applications of such a near-term quantum device for useful tasks includ- Over the past several decades, we have enjoyed expo- ing machine leanring has been widely explored. On the nential growth of computational power, namely, Moore's other hand, quantum simulators are thought to be much law. Nowadays even smart phone or tablet PC is much easier to implement than a full-fledged universal quan- more powerful than super computers in 1980s. Even tum computer. In this regard, existing quantum sim- though, people are still seeking more computational ulators have already shed new light on the physics of power, especially for artificial intelligence (machine learn- complex many-body quantum systems [9{11], and a re- ing), chemical and material simulations, and forecasting stricted class of quantum dynamics, known as adiabatic complex phenomena like economics, weather and climate.
    [Show full text]
  • Arxiv:1808.04962V3 [Cs.ET] 15 Apr 2019 Dynamical Systems, Neuromorphic Device 2010 MSC: 68Txx, 37N20
    Recent Advances in Physical Reservoir Computing: A Review Gouhei Tanakaa;b;1, Toshiyuki Yamanec, Jean Benoit H´erouxc, Ryosho Nakanea;b, Naoki Kanazawac, Seiji Takedac, Hidetoshi Numatac, Daiju Nakanoc, and Akira Hirosea;b aInstitute for Innovation in International Engineering Education, Graduate School of Engineering, The University of Tokyo, Tokyo 113-8656, Japan bDepartment of Electrical Engineering and Information Systems, Graduate School of Engineering, The University of Tokyo, Tokyo 113-8656, Japan cIBM Research { Tokyo, Kanagawa 212-0032, Japan Abstract Reservoir computing is a computational framework suited for temporal/sequential data processing. It is derived from several recurrent neural network models, including echo state networks and liquid state machines. A reservoir comput- ing system consists of a reservoir for mapping inputs into a high-dimensional space and a readout for pattern analysis from the high-dimensional states in the reservoir. The reservoir is fixed and only the readout is trained with a simple method such as linear regression and classification. Thus, the major advan- tage of reservoir computing compared to other recurrent neural networks is fast learning, resulting in low training cost. Another advantage is that the reser- voir without adaptive updating is amenable to hardware implementation using a variety of physical systems, substrates, and devices. In fact, such physical reservoir computing has attracted increasing attention in diverse fields of re- search. The purpose of this review is to provide an overview of recent advances in physical reservoir computing by classifying them according to the type of the reservoir. We discuss the current issues and perspectives related to physical reservoir computing, in order to further expand its practical applications and develop next-generation machine learning systems.
    [Show full text]