Efficient Neurocomputing with Optical Neural Networks

Total Page:16

File Type:pdf, Size:1020Kb

Efficient Neurocomputing with Optical Neural Networks ICCAD: G: Light in Artificial Intelligence: Efficient Neurocomputing with Optical Neural Networks Jiaqi Gu, [email protected], ACM ID: 7981158 ECE Department, University of Texas at Austin, Austin, TX USA 78712 1 PROBLEM AND MOTIVATION SOFTWARE Deep neural networks have received an explosion of interest for their ONN Design Stack Challenges Proposed Methods superior performance in various intelligent tasks and high impacts on Optical Neural Scalability FFT-ONN: Efficient ONN our lives. The computing capacity is in an arms race with the rapidly Architecture Design Architecture [18-20] escalating model size and data amount for intelligent information processing. Practical application scenarios, e.g., autonomous vehicles, data centers, and edge devices, have strict energy efficiency, latency, ONN Robustness ROQ: Noise-Aware and bandwidth constraints, raising a surging need to develop more effi- Model Training Quantization Flow [11,12] cient computing solutions. However, as Moore’s law is winding down, it becomes increasingly challenging for conventional electrical proces- sors to support such massively parallel and energy-hungry artificial Deployment & Learnability FLOPS: Efficient On-Chip intelligence (AI) workloads. Limited clock frequency, millisecond-level On-Chip Training Learning Protocol [21,22] latency, high heat density, and large energy consumption of CPUs, FPGAs, and GPUs motivate us to seek an alternative solution using HARDWARE silicon photonics. Silicon photonics is a promising hardware platform that could represent a paradigm shift in efficient AI acceleration with Figure 1: Challenges in current ONN design stack and the pro- its CMOS-compatibility, intrinsic parallelism of optics, and near-zero posed hardware/software co-design solutions. power consumption. With potentially petaFLOPS per mm2 execu- tion speed and attojoule/MAC computational efficiency, fully-optical neural networks (ONNs) demonstrate orders-of-magnitude higher performance than their electrical counterparts [1–6]. However, previ- ous ONN designs have a large footprint and noise robustness issues, which prevent practical applications of photonic accelerators. In this work, we propose to explore efficient neuromorphic com- puting solutions with optical neural networks. Various photonic inte- grated circuit designs and software-hardware co-optimization meth- ods are explored and presented here to enable high-performance photonic accelerators with lower area cost, better energy efficiency, higher variation-robustness, and more on-device learnability. (a) (b) Figure 2: (a) Matrix multiplication achieved by cascaded MZI 2 BACKGROUND AND RELATED WORK arrays [1]. (b) Micro-ring resonator (MRR) based ONN [13]. Optical computing has been demonstrated as a promising substitution for electronics in efficient artificial intelligence due to its ultra-high order-of-magnitude higher inference throughput and comparable ac- bandwidth, sub-nanosecond latency, attojoule/MAC energy efficiency. curacy on the vowel recognition task compared with GPUs [1]. Zhao The early research efforts focus on diffractive free-space optical com- et al.[10] proposed a slimmed ONN architecture to cut down the puting, optical reservoir computing [7], and spike processing [8, 9] area cost by 30-50% through a software-hardware co-design method- to achieve optical multi-layer perceptrons (MLPs). Recently, the in- ology. These architectures generally have a large area cost and low tegrated optical neural networks (ONNs) have attracted extensive robustness due to phase accumulation error in the cascaded MZI research interests given their compactness, energy efficiency, and meshes [11, 10, 12]. Micro-ring resonator (MRR) based incoherent electronics-compatibility [1,3,4,6]. Figure1 shows the ONN design ONNs have been proposed to build MRR weight banks for matrix stack, including architecture and circuit design, model optimization, multiplication [13, 14], shown in Fig. 2(b) the MRR-ONN has a small and final deployment and on-chip training. Due to the complexity footprint, but it suffers from robustness issues due to MRR sensitivity. of photonic analog circuit design and non-ideality in physical chip To enable scalable optical AI accelerators, novel photonic structures deployment, the power of ONNs will not be fully unleashed without are in high demand to construct compact, efficient, and expressive careful optimization on scalability, robustness, and learnability. optical neural architectures. In the first design stage, neural architectures and their correspond- Given an ONN architecture, the second stage is to perform ONN ing photonic circuits will be jointly designed to map neurocomput- model optimization. Specifically, we need to determine the optical ing to optical components. Previously, Shen et al.[1] successfully component configurations that can achieve the AI tasks with highper- demonstrates a singular-value-decomposition-based coherent ONN formance and fidelity. Previously, hardware-unaware model training constructed by cascaded Mach-Zehnder interferometer (MZI) arrays, is adopted to obtain a theoretically trained ONN model. However, the shown in Fig. 2(a). Their photonic tensor core prototype demonstrates trained ONN weights are not necessarily implementable given limited device control precision and non-ideal process variations [11, 12, 15]. quantization scheme ROQ for robust ONN design, and an efficient ONN Prior work exists to show that the error accumulation effects cause un- on-chip learning framework FLOPS leveraging stochastic zeroth- desired sensitivity of analog ONN chips to noises, which generally lead order optimization for in situ ONN training. to unacceptable accuracy drop and even complete malfunction [11, 15]. However, existing training methods lack practical estimation of de- 3.1 FFT-ONN: Hardware-Efficient ONN vice non-ideality, which leaves a large room for hardware-software Architecture co-design to bridge the performance gap between theoretical models Though ONNs have ultra-fast execution speed, they typically have and the physically deployed ONN chips. large area cost due to the physical limits of the optical devices. Previ- The third stage happens after ONN chip manufacturing. Once there ous MZI-based ONNs [1, 10] consume a large number of MZIs, thus is any change in the environment, tasks, or data distribution, auto- fail to provide efficient and scalable photonic solutions. We focuson matic tuning and on-chip training will be performed in situ to calibrate answering the following critical questions: 1) how to construct a new the circuit states and quickly adapt the ONN chip accordingly. Thanks ONN architecture using much fewer optical components without sac- to the ultra-fast execution speed and reconfigurability of silicon pho- rificing its expressiveness, 2) how to efficiently map high-dimensional tonics chip, ONNs are also perfect self-learning platforms. Such self- neural networks to 2-dimensional (2D) photonic integrated chips learnability is especially beneficial to offload centralized cloud-based (PICs), and 3) how to further cut down the power consumption of training to resource-limited edge devices, which not only saves ex- optical devices with minimum performance degradation [18–20]. pensive communication cost but also boosts the intelligence of edge Proposed Frequency-Domain Optical MLP: To remedy the area cost computing units. To enable such on-device learnability, prior work at- issue, we propose a novel optical multi-layer perceptron architecture tempts to perform in situ ONN training to boost the training efficiency. based on fast Fourier transform, as shown in Fig.3 Instead of imple- Brute-force device tuning [1] and evolutionary algorithms [16] are menting general matrix multiplication, we adopt a block-circulant proposed to search for an optimal device configuration with a large weight matrix as an efficient substitution. This block-circulant matrix number of ONN queries. Adjoint variable methods [17] are applied to can efficiently realize restricted linear projection via an FFT-based fast directly generate and read out the gradients w.r.t. device configura- multiplication algorithm. After transformed into the Fourier domain, tions in-situ to achieve parallel on-chip backpropagation. Though the matrix multiplication can be cast to lightweight element-wise multi- training speed is already orders-of-magnitude higher than software plication between two Fourier-domain signals. A photonic butterfly training, their scalability and robustness are inadequate for practi- network is designed using compact photonic devices to achieve on- cal on-chip learning due to algorithmic inefficiency or prohibitive chip optical FFT/IFFT. Frequency-domain weights are implemented in hardware overhead. the element-wise multiplication (EM) stage to achieve complex-valued Overall, existing studies still fail to provide hardware-efficient, ro- multiplication by leveraging the polarization feature of light. Opti- bust, and self-learnable ONN designs. Hence, better ONN architecture cal splitter and combiner trees are designed to perform light-speed designs and advanced circuit-algorithm co-optimization are still in fan-out and partial product accumulation, respectively. This frame- great demand. Therefore, we propose a holistic ONN design solution work has provably comparable model expressivity with classical NNs. to help build scalable, reliable, and adaptive photonic accelerators Without sacrificing any classification accuracy, this frequency-domain with the following
Recommended publications
  • Data-Driven and Model-Based Design
    Data-Driven and Model-Based Design Stavros Tripakis Aalto University and University of California, Berkeley Abstract—This paper explores novel research directions arising the AI system, we must also trust its LECs. However, today’s from the revolutions in artificial intelligence and the related fields machine learning systems are highly unpredictable [66]. Pro- of machine learning, data science, etc. We identify opportunities totyping such systems may seem quick and easy, but it incurs a for system design to leverage the advances in these disciplines, as well as to identify and study new problems. Specifically, we high cost, called technical debt [58]. These problems are not propose Data-driven and Model-based Design (DMD) as a new theoretical. Numerous traffic violations, accidents, and even system design paradigm, which combines model-based design human casualties have resulted as the fault of LECs. It is with classic and novel techniques to learn models from data. therefore correctly and widely recognized that AI systems are not dependable [23]. Index Terms—System design, formal methods, machine learn- ing, verification, synthesis How to achieve dependability of LECs? “Manual” reason- ing about these systems is prohibitive. Systems like neural I. INTRODUCTION networks become incomprehensible as their size grows [59]. Moreover, using a trial-and-error approach (testing) does not A. Artificial Intelligence scale [38]. Computers and an abundance of data from all sorts of We argue that we need rigorous, formal modeling and sources are revolutionizing many fields of science and tech- verification techniques, in order to nology, from biology and medicine, to astronomy and the 1) model AI systems, and in particular their LECs; social sciences.
    [Show full text]
  • Requirements Engineering Challenges in Building AI-Based Complex Systems
    Requirements Engineering Challenges in Building AI-Based Complex Systems Hrvoje Belani Marin Vuković, Željka Car IT Division Faculty of Electrical Engineering and Computing Ministry of Health University of Zagreb Zagreb, Croatia Zagreb, Croatia [email protected] {marin.vukovic; zeljka.car}@fer.hr Abstract—This paper identifies and tackles the challenges of Based on the discussion, the RE4AI (“RE for AI”) taxono- the requirements engineering discipline when applied to devel- my has been outlined, considering to be useful to inform the opment of AI-based complex systems. Due to their complex be- tailoring of AI development process. The elements of the tax- haviour, there is an immanent need for a tailored development onomy have been recognized from the AI field – data, model process for such systems. However, there is still no widely used and system, and aligned with the typical RE activities. The and specifically tailored process in place to effectively and effi- ciently deal with requirements suitable for specifying a software purpose of the proposed taxonomy is to help practitioners and solution that uses machine learning. By analysing the related researchers to focus on certain challenges upfront and tailor work from software engineering and artificial intelligence fields, their AI research approaches and development processes ac- potential contributions have been recognized from agent-based cordingly. software engineering and goal-oriented requirements engineering This work has been partly motivated by [3] as well as the research, as well as examples from large product development first author’s participation in [4]. Method and settings are pre- companies. The challenges have been discussed, with proposals sented in section II, while section III provides an overview of given how and when to tackle them.
    [Show full text]
  • Building Agents with Agents and Patterns
    Building Agents with Agents and Patterns L. Sabatucci (1), M. Cossentino (2,3), S. Gaglio (1,2) (1) DINFO - Dipartimento di Ingegneria Informatica, Università degli Studi di Palermo - Viale delle Scienze, 90128 Palermo, Italy (2) Istituto di Calcolo delle Reti ad Alte Prestazioni, Consiglio Nazionale delle Ricerche; ( 3 ) SET - Université de Technologie Belfort-Montbéliard - 90010 Belfort cedex, France [email protected]; [email protected]; [email protected] are strategies helping people to find their way through Abstract—The use of design patterns proved successful in complex situations by applying ready solution to solve lowering the development time and number of errors when difficult problems. Also they help in diagnosing, revising, producing software with the object-oriented paradigm. Now the and improving a group's work [11][14]. need for a reuse technique is occurring for the emergent agent paradigm, for which a great effort is currently spending in x Patterns improve software maintenance: a project methodology definitions. In this work we present our experiences obtained with patterns reuse is robust and simpler to in the identification, description, production and use of agents modify with respect to traditional projects [19]. patterns. A repository of patterns was enriched during these years so to request a classification criteria and a documentation Our definition of pattern come from traditional object- template useful to help user during the selection. oriented design patterns, revised for the agent paradigm. In particular we use an ontological approach, strongly influenced Index Terms—Multiagent systems, patterns, reuse models and tools. by the study of multi-agent system (MAS) meta-models.
    [Show full text]
  • Information Systems Foundations: the Role of Design Science
    Information Systems Foundations: The role of design science Information Systems Foundations: The role of design science Dennis N. Hart and Shirley D. Gregor (Editors) THE AUSTRALIAN NATIONAL UNIVERSITY E P R E S S E P R E S S Published by ANU E Press The Australian National University Canberra ACT 0200, Australia Email: [email protected] This title is also available online at: http://epress.anu.edu.au/is_foundations_citation.html National Library of Australia Cataloguing-in-Publication entry Author: Information Systems Foundations (‘The role of design science’) Workshop (2008 : Canberra, A.C.T.) Title: Information systems foundations : the role of design science / edited by Shirley D. Gregor and Dennis N. Hart. ISBN: 9781921666346 (pbk.) ISBN: 9781921666353 (eBook) Notes: Workshop held at the Australian National University in Canberra from 2-3 October, 2008. Includes bibliographical references. Subjects: Management information systems--Congresses. Information resources management--Congresses. System design--Congresses. Other Authors/Contributors: Gregor, Shirley Diane. Hart, Dennis N. Dewey Number: 658.4038 All rights reserved. No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying or otherwise, without the prior permission of the publisher. Cover design by Teresa Prowse Cover ilustration by Jackson Gable Printed by Griffin Press This edition © 2010 ANU E Press Contents Preface . ix Philosophical Foundations 1 . Identification-interaction-innovation: a phenomenological basis for an information services view . 3 Dirk Hovorka, Matt Germonprez 2 . How critical realism clarifies validity issues in theory- testing research: analysis and case . 21 Robert B. Johnston, Stephen P.
    [Show full text]
  • Values in Architecture a New Design Paradigm: Monist Submission 8Th April 2008 Ian Ritchie
    VALUES IN ARCHITECTURE A NEW DESIGN PARADIGM: MONIST SUBMISSION 8TH APRIL 2008 IAN RITCHIE INTRODUCTION I will try to explain why I believe we are at the beginning of new design paradigm – an important new movement in urbanism and architecture whose visual aesthetic will be multifarious – yet derived from a creative synthesis of science, ecology and ethics. This opposes recent architectural ’isms’, especially superficial and selfish architecture that gratifies itself on hyperbole to gain media self-aggrandisements. Intelligent, social and selfless architectural expression capable of the most marvellous and spiritually uplifting engineered structures must challenge turn-of-the century stunt-making architectural gymnastics. I will address five questions, 1. How does our intellectual heritage shape our actions? 2. What are we thinking about today? 3. How are we behaving as designers? 4. How should we design today? 5. How should we make things? HOW DOES OUR INTELLECTUAL HERITAGE SHAPE OUR ACTIONS? I want to step back 2,500 years to hint at the origins of our present discomfort of having to live with apparent contradictions, and in particular how homo sapiens sapiens has to face up to and find ways of taming a rampant homo faber and homo consumeris. The Greeks sought to reconcile the idea of ‘perpetual change and eternal becoming’ put forward by Heraclitus with that of the ‘unchangeable being’ of Parmenides. The outcomes were to have a profound impact upon the development of our western society. The paradox was resolved when the Greeks thought of the atom as the inert fundamental unchangeable ‘being’, yet which, moved by undefined forces (spirits) could combine with other atoms to generate change.
    [Show full text]
  • Meta-Design Paradigm Based Approach for Iterative Rapid Development of Enterprise Web Applications
    META-DESIGN PARADIGM BASED APPROACH FOR ITERATIVE RAPID DEVELOPMENT OF ENTERPRISE WEB APPLICATIONS Athula Ginige AeIMS Research Group, University of Western Sydney, Sydney, Australia Keywords: Enterprise web application development, Meta-design paradigm, Iterative development, Auto generation of web applications. Abstract: Developing enterprise software or web applications that meet user requirements within time and budget still remains a challenge. The success of these applications mostly depends on how well the user requirements have been captured. The literature shows progress has been made on two fronts; improving ways requirements are captured and increasing interaction between users and developers to detect gaps or miscommunication of requirements early in the lifecycle by using iterative rapid development approaches. This paper presents a Meta-Design paradigm based approach that builds on work already done in the area of Model Driven Web Engineering to address this issue. It includes a Meta-Model of an enterprise web application to capture the requirements and an effective way of generating the application. 1 INTRODUCTION (Standish_Group 1995; Sauer, Gemino et al. 2007; Escalona and Arago´n 2008). Today enterprises are depending heavily on web A good approach to find whether the developed based business applications to stay competitive. This system meets the user requirements is by giving the class of applications are broadly categorised as system to the users as quickly as possible and to get Enterprise Web Applications. Still there are many the users to use the system. If there was loss of challenges to developing Enterprise Web information in capturing the requirements or if the Applications within time and budget that meet user requirements have evolved since these were initially requirements (Krigsman 2008).
    [Show full text]
  • Architect and the Paradigms of Sustainable Development: a Review of Dilemmas
    Preprints (www.preprints.org) | NOT PEER-REVIEWED | Posted: 20 December 2017 doi:10.20944/preprints201712.0145.v1 Peer-reviewed version available at Sustainability 2018, 10, 100; doi:10.3390/su10010100 1 Review 2 Architect and the paradigms of sustainable 3 development: a review of dilemmas 4 Wojciech Bonenberg 1 and Oleg Kapliński 1* 5 1 Faculty of Architecture, Poznań University of Technology, 60-965 Poznań, Nieszawska Str. 13C, Poland 6 * Correspondence: [email protected]; Tel.: +48-61-6653260 7 Abstract: The article presents the architect's attitude towards the paradigms of sustainable 8 development. The place and role of the architect in the implementation of the multidimensional 9 process of sustainable design has been presented. Basic dilemmas and antinomies have been 10 presented. The analysis of architect's attitudes towards these problems was performed in various 11 contexts, examining the architect's awareness and his environment in view of changes under way. 12 The article draws attention to the status of knowledge, changes in design paradigms, legislative and 13 organizational requirements. The importance of architectural culture level, the need for training, 14 ways to support the implementation of new design paradigms through integrated activities have 15 been indicated. The research results regarding public awareness of architecture and sustainable 16 development are illustrated with examples from Poland. 17 Keywords: architect, sustainable architecture, paradigms of design, knowledge, society, Poland 18 19 1. Introduction 20 The results of developing sustainable architecture are founded on the symbiosis of ecologists 21 and architects. It began with these two professional groups proposing a change in functioning of the 22 building, i.e., a transition from a linear approach to a closed circulation plan.
    [Show full text]
  • Large-Scale Optical Neural Networks Based on Photoelectric Multiplication
    PHYSICAL REVIEW X 9, 021032 (2019) Large-Scale Optical Neural Networks Based on Photoelectric Multiplication Ryan Hamerly,* Liane Bernstein, Alexander Sludds, Marin Soljačić, and Dirk Englund Research Laboratory of Electronics, MIT, 50 Vassar Street, Cambridge, Massachusetts 02139, USA (Received 12 November 2018; revised manuscript received 21 February 2019; published 16 May 2019) Recent success in deep neural networks has generated strong interest in hardware accelerators to improve speed and energy consumption. This paper presents a new type of photonic accelerator based on coherent detection that is scalable to large (N ≳ 106) networks and can be operated at high (gigahertz) speeds and very low (subattojoule) energies per multiply and accumulate (MAC), using the massive spatial multiplexing enabled by standard free-space optical components. In contrast to previous approaches, both weights and inputs are optically encoded so that the network can be reprogrammed and trained on the fly. Simulations of the network using models for digit and image classification reveal a “standard quantum limit” for optical neural networks, set by photodetector shot noise. This bound, which can be as low as 50 zJ=MAC, suggests that performance below the thermodynamic (Landauer) limit for digital irreversible computation is theoretically possible in this device. The proposed accelerator can implement both fully connected and convolutional networks. We also present a scheme for backpropagation and training that can be performed in the same hardware. This architecture will enable a new class of ultralow-energy processors for deep learning. DOI: 10.1103/PhysRevX.9.021032 Subject Areas: Interdisciplinary Physics, Optoelectronics, Photonics I. INTRODUCTION ideally compatible with training as well as inference.
    [Show full text]
  • The Impact of Introducing Artificial Intelligence Technology to Architecture and Its Leverage on the Concept of Design Automation
    THE IMPACT OF INTRODUCING ARTIFICIAL INTELLIGENCE TECHNOLOGY TO ARCHITECTURE AND ITS LEVERAGE ON THE CONCEPT OF DESIGN AUTOMATION by YASSER M. EL-QUESSNY B.Arch. Cairo University 1984 SUBMITTED TO THE DEPARTMENT OF ARCHITECTURE IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF MASTER OF SCIENCE IN ARCHITECTURE STUDIES AT THE MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 1987 Copyright @ 1987 Yasser M. El-Quessny The author hereby grants to M.I.T. permission to reproduce and to distribute publicaly copies of this thesis document in whole or in part. ., / ! Signature o FAuthor -, qI -7r- Yasser M. El-Quessny Department of Architecture May 6, 1987 Certified by Eric Dluhosch Associate Professor of Building Technology Thesis Supervisor Accepted b Julian Beinart Chairman epartmental Committee for Graduate Students MASSACHUSETTS INSTITUTE OF TECHNOLOGY JUN 0 8 1987 LIBRARIES RotaD jh~jj I q .P I AUi IP~d~ * A. L11 * A1~ dAl&LUph- P.&6iJ j i~j~i ,i *d ii THE IMPACT OF INTRODUCING ARTIFICIAL INTELLIGENCE TECHNOLOGY TO ARCHITECTURE AND ITS LEVERAGE ON THE CONCEPT OF DESIGN AUTOMATION by YASSER M. EL-QUESSNY Submitted to the Department of Architecture on May 6th. 1987 in Partial Fulfillment of the Requirements for the Degree of Master's of Science in Architecture Studies ABSTRACT The problem addressed in this research is understanding the nature evolving and new technologies into the domain of architectural design and building technology. This thesis, essentially, is an exploration of the ways that "Artificial Intelligence" techniques may support systematic and rational architectural design and, by extension, the "Building Systems" process. The motivation for working in this area of research stems from the serious need to develop a new methodological design approach for architects.
    [Show full text]
  • Self-Learning Photonic Signal Processor with an Optical Neural Network Chip
    Self-learning photonic signal processor with an optical neural network chip Hailong Zhou1Ϯ, Yuhe Zhao1Ϯ, Xu Wang1, Dingshan Gao1, Jianji Dong1 and Xinliang Zhang1 1Wuhan National Laboratory for Optoelectronics, School of Optical and Electronic Information, Huazhong University of Science Ϯ and Technology, Wuhan 430074, China; These authors contributed equally to this work. e-mail: [email protected]; [email protected]. Photonic signal processing is essential in the optical communication and optical computing. Numerous photonic signal processors have been proposed, but most of them exhibit limited reconfigurability and automaticity. A feature of fully automatic implementation and intelligent response is highly desirable for the multipurpose photonic signal processors. Here, we report and experimentally demonstrate a fully self-learning and reconfigurable photonic signal processor based on an optical neural network chip. The proposed photonic signal processor is capable of performing various functions including multichannel optical switching, optical multiple-input- multiple-output descrambler and tunable optical filter. All the functions are achieved by complete self-learning. Our demonstration suggests great potential for chip-scale fully programmable optical signal processing with artificial intelligence. Photonic signal processing is widespread both in the optical communication and optical computing. Almost all aspects of signal processing functions were developed, such as multichannel optical switching1-7, mode (de)multiplexer8,9, optical filter10-14, polarization management15-17, waveguide crossing18,19 and logic gates20-22. Plentiful photonic signal processing devices have been reported based on either discrete components or photonic integrated circuits1,9,15,20-24. Photonic signal processing devices based on discrete components are usually bulkier and less power efficient, whereas a photonic integrated signal processing chip has a much smaller footprint and a higher power efficiency.
    [Show full text]
  • An Optical Neural Chip for Implementing Complex-Valued
    ARTICLE https://doi.org/10.1038/s41467-020-20719-7 OPEN An optical neural chip for implementing complex- valued neural network ✉ ✉ H. Zhang1,M.Gu 2,3 , X. D. Jiang 1 , J. Thompson 3, H. Cai4, S. Paesani5, R. Santagati 5, A. Laing 5, Y. Zhang 1,6, M. H. Yung7,8, Y. Z. Shi 1, F. K. Muhammad1,G.Q.Lo9, X. S. Luo 9, B. Dong9, D. L. Kwong4, ✉ ✉ L. C. Kwek 1,3,10 & A. Q. Liu 1 Complex-valued neural networks have many advantages over their real-valued counterparts. 1234567890():,; Conventional digital electronic computing platforms are incapable of executing truly complex- valued representations and operations. In contrast, optical computing platforms that encode information in both phase and magnitude can execute complex arithmetic by optical inter- ference, offering significantly enhanced computational speed and energy efficiency. However, to date, most demonstrations of optical neural networks still only utilize conventional real- valued frameworks that are designed for digital computers, forfeiting many of the advantages of optical computing such as efficient complex-valued operations. In this article, we highlight an optical neural chip (ONC) that implements truly complex-valued neural networks. We benchmark the performance of our complex-valued ONC in four settings: simple Boolean tasks, species classification of an Iris dataset, classifying nonlinear datasets (Circle and Spiral), and handwriting recognition. Strong learning capabilities (i.e., high accuracy, fast convergence and the capability to construct nonlinear decision boundaries) are achieved by our complex-valued ONC compared to its real-valued counterpart. 1 Quantum Science and Engineering Centre (QSec), Nanyang Technological University, 50 Nanyang Ave, 639798 Singapore, Singapore.
    [Show full text]
  • Any Progress in Systematic Design?
    CAAD FUTURES DIGITAL PROCEEDINGS 1986 5 Chapter 1 Any progress in systematic design? A. H. Bridges 1.1 Introduction In order to discuss this question it is necessary to reflect awhile on design methods in general. The usual categorization discusses 'generations' of design methods, but Levy (1981) proposes an alternative approach. He identifies five paradigm shifts during the course of the twentieth century which have influenced design methods debate. The first paradigm shift was achieved by 1920, when concern with industrial arts could be seen to have replaced concern with craftsmanship. The second shift, occurring in the early 1930s, resulted in the conception of a design profession. The third happened in the 1950s, when the design methods debate emerged; the fourth took place around 1970 and saw the establishment of 'design research'. Now, in the 1980s, we are going through the fifth paradigm shift, associated with the adoption of a holistic approach to design theory and with the emergence of the concept of design ideology. A major point in Levy's paper was the observation that most of these paradigm shifts were associated with radical social reforms or political upheavals. For instance, we may associate concern about public participa- tion with the 1970s shift and the possible use (or misuse) of knowledge, information and power with the 1980s shift. What has emerged, however, from the work of colleagues engaged since the 1970s in attempting to underpin the practice of design with a coherent body of design theory is increasing evidence of the fundamental nature of a person's engagement with the design activity.
    [Show full text]