Open Stevenpetronethesis.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

Open Stevenpetronethesis.Pdf THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE SCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE NONLINEAR NEURON DISCRIMINANT FUNCTIONS FOR ALTERNATE DEEP LEARNING TRAINING ALGORITHMS STEVEN PETRONE SPRING 2020 A thesis submitted in partial fulfillment of the requirements for baccalaureate degrees in Computer Engineering with honors in Computer Engineering Reviewed and approved* by the following: Christopher Griffin Associate Research Professor Applied Research Laboratory & Department of Mathematics Thesis Supervisor Vijay Narayanan Professor of Computer Science and Engineering Honors Adviser *Signatures are on file in the Schreyer Honors College. i Abstract In this work we present a novel neuron discriminant function that allows for alternate training algorithms for deep learning. The new neuron type, which we call a posynomial neuron, can be combined with linear neurons to represent functions that are exponentials when inferencing new data, but are only polynomials of the network weights. We show that the properties of these net- works can be resistant to the vanishing gradient problem. We also formulate training the network as a geometric programming problem and discuss the interesting benefits this can have over training a network with gradient descent, such as data set analysis and network interpretability. We provide a C++ library that implements both posynomial and sigmoidal networks but provides flexibility for additional novel layer types. We also provide a tensor library that has applications beyond deep learning. ii Table of Contents List of Figures iv List of Tables v Acknowledgements vi 1 Introduction 1 1.1 Deep Learning . .2 1.2 Contribution . .2 1.3 Organization . .2 2 Literature Review 3 2.1 Neural Networks . .4 2.1.1 Artificial Neurons from Biology . .4 2.1.2 Logistic Regression and Nonlinearities . .4 2.1.3 Hidden Layers and XOR problem . .5 2.1.4 Derivative Backpropagation as an Engine for Gradient Descent . .6 2.1.5 Survey of Advancements in Neural Networks . .7 2.2 Posynomials . 10 2.2.1 Monomials, Polynomials, Posynomials . 10 2.2.2 Geometric Programming . 10 3 Theory 13 3.1 Posynomial Neural Network . 14 3.1.1 Monomial Based Discriminant Function . 14 3.1.2 Transforming Input Data Before a Posynomial Network . 14 3.1.3 Derivatives for gradient descent . 18 3.2 Geometric and Convex Programming . 19 3.2.1 Formulation . 19 4 Implementation 21 4.1 Data Structures . 22 4.1.1 Tree . 22 4.1.2 List . 22 4.1.3 Tensor . 22 iii 4.1.4 Layer . 23 4.1.5 Network . 24 4.2 Algorithms . 26 4.2.1 Inference . 26 4.2.2 Train/Learn . 27 4.3 Verification and Bug Finding . 27 4.3.1 Eradicating Memory Leaks with Valgrind . 27 4.3.2 Input Fuzzing with American Fuzzy Lop . 28 5 Experimental Results 29 5.1 XOR With 3 Layer Network . 30 5.2 XOR with 2 Layer Network . 34 5.3 Training XOR using an objective function and a nonlinear solver. 38 5.4 Run time Analysis . 39 5.4.1 Run time of Dot Product . 39 6 Conclusion 41 6.1 Discussion . 42 6.2 Future Directions . 42 6.2.1 Alternative to Gradient Descent . 42 6.2.2 Develop theory and training algorithm for complex value neural networks . 42 6.2.3 Layer analogous to the convolutional layer . 42 6.2.4 Library improvements . 43 Bibliography 44 iv List of Figures 2.1 Diagram of single neuron . .5 3.1 Rotation of XOR data points . 15 3.2 Posynomial learnable saddle point . 16 3.3 Posynomial learnable saddle point contour plot . 16 3.4 Saddle point that computes XOR . 17 3.5 Saddle point contour plot that computes XOR . 17 4.1 UML class diagram . 25 4.2 Valgrind verifying absence of memory leaks . 28 4.3 AFL Fuzzing to identify bugs . 28 5.1 Diagram of 3 layer XOR network . 30 5.2 Initial weights of 3 layer XOR network . 30 5.3 3 layer XOR network showing correct outputs . 31 5.4 Near perfect outputs for 3 layer XOR network . 31 5.5 Transformation of inputs by first layer in 3 layer XOR network . 32 5.6 Posynomial layers of 3 layer XOR network . 32 5.7 Contour plot of posynomial layers of 3 layer XOR network . 33 5.8 Plot of 3 layer XOR network . 33 5.9 Contour plot of 3 layer XOR network . 34 5.10 Diagram of 2 layer XOR network . 35 5.11 Initial weights of 2 layer XOR network . 35 5.12 Correct outputs for 2 layer XOR network . 36 5.13 Perfect outputs for 2 layer XOR network . 36 5.14 Plot of 2 layer XOR network . 37 5.15 Contour plot of 2 layer XOR network . 37 5.16 Posynomnial XOR trained with nonlinear solver . 38 5.17 Contour plot of posynomial XOR trained with nonlinear solver . 39 5.18 Run time of dot product implementations for tensor library . 40 v List of Tables 2.1 Definition of XOR . .6 vi Acknowledgements I would like to thank Christopher Griffin for his guidance with this work. 1 Chapter 1 Introduction 2 1.1 Deep Learning Deep learning is a common cross disciplinary research topic. While commonly thought of as a subfield of computer science and engineering, deep learning and the study of neural networks was inspired by attempts to study the mammalian brain. Now that it is a well established and popular field, deep learning has found applications in a surprisingly diverse range of disciplines. Deep learning attempts to solve problems such as classification, regression, and even content generation by training a network of artificial neurons using large data sets. Individual neurons ‘fire’ based on a computation between their inputs, the weighted connections and an internal bias. It is the process of tuning these weights that motivates the term ‘learning’ and the multilayered structure of the neurons that motivates use of the term ‘deep’. 1.2 Contribution The main contribution of this work is introducing a novel neural network layer type that is multiplicative in nature rather than linear. We show that this layer is a log transform, different from a linear layer with no activation function. When combined with a linear layer, a two layer network can resemble functions that compute multiplicative exponentials of the input when infer- encing. However, when the output is considered a function of the network weights it is a simple polynomial objective function, which allows for many novel optimization algorithms in addition to traditional gradient descent. We introduce some of these algorithms and their possible benefits to training speed, network interpretability, and security. We also discuss interesting properties of posynomial neural networks such as their resistance to the vanishing gradient problem. To train these networks we created a lightweight C++ library. The library includes many specialized data structures, including a generalized Tensor library that has applications beyond deep learning. Also included are traditional neural network data structures such as layers and networks and the neces- sary algorithms to train them. 1.3 Organization In Chapter 2 we will review the literature to provide an overview of the history and develop- ment of deep learning and using neural networks to learn from data, as well as a survey of more recent advancements in the field. We will also discuss posynomials and signomials as classes of functions, and Geometric Programming, a specialized optimization technique for these classes. In Chapter 3 we will introduce the theory of our contribution by describing our novel neuron types and the required calculus to train them with gradient descent. Additionally, we will formulate an objective function to train a neural network with Geometric Programming. In Chapter 4 we will discuss the implementation details of the libraries we created as part of this work and the steps we took to validate their functionality, stability, and security. Chapter 5 provides our experimental results where we discuss training multiple posynomial networks on the classic XOR problem and provide graphical representations of the network that show an intuitive way of understanding how posynomial networks work. We conclude our work in Chapter 6 where we provide a discussion along with future directions. 3 Chapter 2 Literature Review 4 2.1 Neural Networks Deep learning has had several periods of prolific research since its inception, with the most recent resurgence being in recent years. With recent advancements and discoveries, deep learn- ing is likely to forever change academic, industrial, government, and consumer technologies [1]. Deep learning is being implemented in IoT devices [2, 3] for computer vision, speech recogni- tion, and language processing from embedded devices to in ultra large scale data centers [4, 5] for training models from terabytes or more of data. Countless fields have discovered applications for deep learning including medical image processing [6], time series forecasting [7], autonomous vehicle driving [8], art generation [9], drug discovery [10, 11], language processing [12] and translation[13, 14]. Neural networks and deep learning have even been used to beat Chess and Go champions. Deep learning involves training a network of artificial neurons with large data sets in an attempt to mimic how biological organisms learn from experience. In fact, much of the early research that led to the field of deep learning was an attempt to understand the brain. 2.1.1 Artificial Neurons from Biology Many ideas that led to the creation of artificial neural networks came from attempts to under- stand the brain. The 1943 paper by Walter Pitts and Warren McCulloch [15] was one of the first attempts at describing a brain’s mental activity in terms of neural networks that can be quanti- fied with propositional logic. A neural network can be thought of as a mathematical graph where vertices emulate neurons.
Recommended publications
  • ABSTRACT XUE, XIANGZHONG. Electronic
    ABSTRACT XUE, XIANGZHONG. Electronic System Optimization Design via GP-Based Surrogate Modeling. (Under the direction of Dr. Paul D. Franzon.) For an electronic system with a given circuit topology, the designer’s goal is usually to automatically size the device and components to achieve globally optimal performance while at the same satisfying the predefined specifications. This goal is motivated by a human desire for optimality and perfection. This research project improves upon current optimization strategies. Many types of convex programs and convex fitting techniques are introduced and compared, and the evolution of Geometric Program (GP)-based optimization approaches is investigated through a literature review. By these means, it is shown that a monomial-based GP can achieve optimal performance and accuracy only for long-channel devices, and that a piecewise linear (PWL)-based GP works well only for short-channel, narrow devices without many data fitted. Based on known GP optimizer and convex PWL fitting techniques, an innovative surrogate modeling and optimization algorithm is proposed to further improve performance accuracy iteratively for a wide transistor with a short channel. The new surrogate strategy, which comprises a fine model and a coarse model, can automatically size the device to create a reusable system model for designing electronic systems and noticeably improving prediction accuracy, particularly when compared to the pure, GP-based optimization method. To verify the effectiveness and viability of the proposed surrogate strategy, a widely used two-stage optimization design is employed, entailing an operational amplifier (op-amp) and LC-tuned oscillator. In addition, an involved analysis and simulation demonstrate that the optimal results of both coarse and fine models in the proposed surrogate strategy may gradually converge to each other iteratively while achieving over 10% improvement in performance accuracy compared to the previous, PWL-based GP algorithm.
    [Show full text]
  • Mathematicalsciences News
    mathematical sciences news 2015 contents 2015 Letter from Department 03 Head, Tom Bohman Math News Shorts 04 Faculty Notes 06 Editor-in-Chief Tom Bohman Writer Feature 08 Amy Pavlak Laird Alumnus and Nobel Laureate Contributing Writers Bill Hrusa John Nash Wins Abel Prize David Kinderlehrer Dejan Slep ´cev Photography Democracy 2.1 16 Carnegie Mellon University photographers Ken Andreyo and Tim Kaulen Graphic Design Carnegie Mellon University Marketing & Communications Po-Shen Loh Receives 20 NSF CAREER Award Carnegie Mellon University Department of Mathematical Sciences Wean Hall 6113 Pittsburgh, PA 15213 Boris Bukh Wins Sloan 22 math.cmu.edu Research Fellowship Carnegie Mellon University does not discriminate in admission, employment, or administration of its programs or activities on the basis of race, color, national origin, sex, handicap or disability, age, sexual orientation, gender identity, religion, Undergraduate creed, ancestry, belief, veteran status or genetic 24 information. Furthermore, Carnegie Mellon University does not discriminate and is required Research not to discriminate in violation of federal, state or local laws or executive orders. Inquiries concerning the application of and compliance with this statement should be directed to the Vice President for Campus Affairs, Carnegie Mellon University, 5000 Forbes Avenue, Class of 2015 Pittsburgh, PA 15213, telephone 412-268-2056. 26 Obtain general information about Carnegie Mellon University by calling 412-268-2000. Produced for the Department of Mathematical Sciences by the Marketing & Communications, November, 2015, 16-177. ©2015 Carnegie Mellon University, All rights reserved. No part of this publication may be reproduced in any form without written permission from Carnegie Mellon Unversity’s Department of Mathematical Sciences.
    [Show full text]
  • Notices of the American Mathematical Society
    OF THE AMERICAN MATHEMATICAL SOCIETY VOLUME 16, NUMBER 3 ISSUE NO. 113 APRIL, 1969 OF THE AMERICAN MATHEMATICAL SOCIETY Edited by Everett Pitcher and Gordon L. Walker CONTENTS MEETINGS Calendar of Meetings ..................................... 454 Program for the April Meeting in New York ..................... 455 Abstracts for the Meeting - Pages 500-531 Program for the April Meeting in Cincinnati, Ohio ................. 466 Abstracts for the Meeting -Pages 532-550 Program for the April Meeting in Santa Cruz . ......... 4 73 Abstracts for the Meeting- Pages 551-559 PRELIMINARY ANNOUNCEMENT OF MEETING ....................•.. 477 NATIONAL REGISTER REPORT .............. .. 478 SPECIAL REPORT ON THE BUSINESS MEETING AT THE ANNUAL MEETING IN NEW ORLEANS ............•................... 480 INTERNATIONAL CONGRESS OF MATHEMATICIANS ................... 482 LETTERS TO THE EDITOR ..................................... 483 APRIL MEETING IN THE WEST: Some Reactions of the Membership to the Change in Location ....................................... 485 PERSONAL ITEMS ........................................... 488 MEMORANDA TO MEMBERS Memoirs ............................................. 489 Seminar of Mathematical Problems in the Geographical Sciences ....... 489 ACTIVITIES OF OTHER ASSOCIATIONS . 490 SUMMER INSTITUTES AND GRADUATE COURSES ..................... 491 NEWS ITEMS AND ANNOUNCEMENTS ..... 496 ABSTRACTS OF CONTRIBUTED PAPERS .................•... 472, 481, 500 ERRATA . • . 589 INDEX TO ADVERTISERS . 608 MEETINGS Calendar of Meetings NOn:: This Calendar lists all of the meetings which have been approved by the Council up to the date at which this issue of the c;Noticei) was sent to press. The summer and annual meetings are joint meetings of the Mathematical Association of America and the American Mathematical Society. The meeting dates which fall rather far in the future are subject to change. This is particularly true of the meetings to which no numbers have yet been assigned.
    [Show full text]
  • Promoting OR
    Volume 9 | Number 3 | September 2015 | ISSN 2223-4373 International Federation of Operational Research Societies NEWS What’s Inside Editorial Promoting OR From the IFORS AC 1 Promoting OR Sue Merchant, IFORS VP at large From the Editor At EURO this summer I was struck by the enthusiasm, amongst 2 Of Contributions and Transitions developing country delegates in particular, for undertaking the type Regional Conferences of practical OR which makes a real difference in industry, commerce 2 EURO 2015: Scottish Kilts and and government. However the enthusiasts reported a problem Research Clans with Impact which many of us have faced over the years – how to get business 3 OR People Make EURO: Glasgow and government to recognise that OR makes a difference and hence Hosts 27th EURO Conference recruit OR people. Indeed John Ranyard and Robert Fildes’ recent 4 Environment Takes Center Stage international study, with responses from 28 countries, confirmed that the lack of client at APORS 2015 awareness and understanding was a continuing and widespread problem. They stress the OR for Development importance of practitioner communities and OR societies in helping to drive awareness of OR 5 ICORD 2015 in countries across the world. 6 A Perfect Setting for a Workshop on Uplifting Living Conditions 6 Mexico Sets Stage for ICORD 2016 Various methods have been tried over the years to demonstrate how powerful OR is: for example, Obituaries INFORMS’ major campaign ‘The Science of 7 Heiner Müller-Merbach Better’ was initiated a few years ago and the UK 8 John Forbes Nash Jr. also adopted this approach, producing masses 9 Philippe Van Asbroek of promotional material, a DVD of OR activity Conferences in successful companies (which is available 9 IFORS 2020 Hosting Bids Shortlisted from the UK ORS) and setting up websites 10 13th EUROPT: Auld Reekie Blows ‘Science of Better.co.uk’ and ‘LearnaboutOR.
    [Show full text]
  • Regents Communication
    THE UNIVERSITY OF MICHIGAN Regents Communication Recommendations for approval of new appointments and promotions for regular associate and full professor ranks, with tenure PRO~~OTIONRECOMMENDATION The University of Michigan College of Literature, Science, and the Arts ACTIOS REQUEST: Faculty Promotion Approval NAME: Fernando Coronil CURRENT TITLES: Associate Professor of Anthropology, with tenure, and Associate Professor of History, with tenure, College of Literature, Science, and the Arts RECOMMENDED TITLES: Professor of Anthropology, with tenure, and Professor of History, with tenure, College of Literature, Science, and the Arts EFFECTIVE DATE: September 1, 2008 It is recommended that Fernando Coronil, associate professor of anthropology, with tenure, and associate professor of history, with tenure, College of Literature, Science, and the Arts, be promoted to professor of anthropology, w~thtenure, and professor of history, with tenure, College oCL,iterature, Science, and the Arts. effective September 1, 2008. Academic Degrees: Ph.D. 1987 University of Chicago M.A. 1970 University of Chicago B.A. 1967 Stanford University Professional Record: 2007 -2008 Visiting Presidential Professor, 711e Graduate Center, City University of New York 2004 - 2005 Fellow, David Rockefeller Center of Latin American St~tdies,and Visiting Associate Professor, Department of History, Harvard University 1997 - present Associate Professor, Department of Anthropology and Department of History, University of Michigan 1991 - 1997 Assistant Professor, Department of Anthropology and Department of History, University of Michigan 1988 - 1991 Assistant Professor, Michigan Society of Fellows, University of Michigan 1987 Visiting Faculty Fellow, Helen Kellogg Institute, ll~liversityof Notre Dame Summary of Evaluatio~~: Teaching - Professor Coronil's pedagogical work has been spread across four units.
    [Show full text]
  • Decisions Under Risk Programming with Economic Applications José Hugo Portillo-Campbell Iowa State University
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1969 Decisions under risk programming with economic applications José Hugo Portillo-Campbell Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Economics Commons Recommended Citation Portillo-Campbell, José Hugo, "Decisions under risk programming with economic applications " (1969). Retrospective Theses and Dissertations. 4141. https://lib.dr.iastate.edu/rtd/4141 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. 70-13,521 PORTILLO-CAMPBELL, José Hugo, 1941- DECISIONS UNDER RISK PROGRAMMING WITH ECONOMIC APPLICATIONS. Iowa State University, Ph.D., 1969 Economics, general University Microfilms, Inc., Ann Arbor, Michigan (£) Copyright by JOSÉ HUGO PORTILLO-CAMPBELL 1970 THIS DISSERTATION HAS BEEN MICROFILMED EXACTLY AS RECEIVED DECISIONS UNDER RISK PROGRAMMING WITH ECONOMIC APPLICATIONS by José Hugo Portillo-Campbell A Dissertation Submitted to the Graduate Faculty in Partial Fulfillmen The Requirements for the Degree of DOCTOR OF PHILOSOPHY Major Subject: Economics Approved: Signature was redacted for privacy. In Charge ot major worK Signature was redacted for privacy. Head ot Major Department Signature was redacted for privacy. Iowa State University Ames, Iowa 1969 PLEASE NOTE: Some pag--s have small and indistinct type. Filmed as received. University Microfilms ii TABLE OF CONTENTS Page 1. INTRODUCTION 1 2. REVIEW OF LITERATURE IN RISK PROGRAMMING 3 2.1.
    [Show full text]
  • Also Inside: • O.R
    INNOVATIVE EDUCATION: Back-to-school special issue August 2015 Volume 42 • Number 4 ormstoday.informs.org Great Expectations The sky’s the limit for quant grads, but will their education match their lofty goals? • Analytics and the MBA curriculum • Teaching professional practitioners • Competitive simulations: game on • Introducing analytics to adolescents • Profiles of 2015 UPS Prize finalists Also Inside: • O.R. takes on ‘Grand Challenges’ • Roundtable profile: Analytics at IDA Shell unlocks the benefits of optimization with AIMMS The team at Shell Global Solutions Their AIMMS-based decision-support tool, GMOS/ has driven optimization solutions with NetSim has contributed to millions of dollars in AIMMS since 2003. cost savings. This valuable tool has helped Shell define an optimal asset structure, and within that achieve: “GMOS has helped us to optimize our supply chain and select the feedstock with • Optimal sourcing of raw materials, the best commercial terms at any given time. • Optimized manufacturing volumes by plant, driven The folks in the petrochemicals business through tiered pricing really trust this tool to help them make • Enhanced decision support for manufacturing the right decisions.” Jim Nastoff, Strategy locations and logistics distribution costs Manager at Shell Experience the power and ease of use of AIMMS for yourself. Contact [email protected] to get a FREE 30-day trial. Award winning software company AIMMS works with half of the Fortune 500 companies to optimize their businesses, making them even more successful. AIMMS_June2015.indd 1 5/26/15 2:38 PM Lindo Systems_August2015.indd 1 7/24/15 10:47 AM Contents August 2015 | Volume 42, No.
    [Show full text]
  • Notices of the American Mathematical Society
    Calendar This Calendar lists all of the meetings which have been approved by the Council up to the date this issue of the cNotictiJ was sent to press. The summer and annual meetings are joint meetings of the Mathematical Association of America and the American Mathematical Society. The meeting dates which fall rather far in the future are subject to change; this is particularly true of meetings to which no numbers have yet been assigned. Abstracts should be submitted on special forms which are available in most departments of mathematics; forms can also be obtained by writing to the headquarters of the Society. Abstracts to be presented at the meeting in person must be received at the headquarters of the Society in Providence, Rhode Island, on or before the deadline for the meeting. Meeting Deadline for Abstracts* Number Date Place and News Items 731 January 22-26, 1976 San Antonio, Texas Nov, 5, 1975 (82nd Annual Meeting) 732 March 4-5, 1976 Tallahassee, Florida Jan. 8, 1976 733 March 15-20, 1976 Urbana, illinois Jan. 8, 1976 734 April 11-14, 1976 New York, New York Feb. 17, 1976 735 April 23-24, 1976 Reno, Nevada Feb. 17, 1976 736 June 18-19, 1976 Portland, Oregon April 27, 1976 737 August 24-28, 1976 Toronto, Canada June 15, 1976 (Both Summer Meeting) November 19-20, 1976 Columbia, South Carolina November 26-27, 1976 Albuquerque, New Mexico January 27-31, 1977 St. Louis, Missouri (83rd Annual Meeting) *Deadline for abstracts not presented at a meeting (by title) February 1976 issue: January 6 April 1976 issue: February 10 OTHER EVENTS February 24, 1976 Symposium on Some Mathematical Questions in Biology, Boston, Massachusetts March 15-18, 1976 Symposium on Probability, Urbana, lllinois Please alf"JX the peel-off label on these c/'(oticti) to correspondence with the Society concerning fiscal matters, changes of address, promotions, or when placing orders for hooks and journals.
    [Show full text]
  • Notices of the American Mathematical Society
    Calendar This Calendar lists all of the meetings which have been approved by the Council up to the date this issue of the cJ{oliiriJ was sent to press. The summer and annual meetings are joint meetings of the Mathematical Association of America and the American Mathematical Society. The meeting dates which fall rather far in the future are subject to change; this is particularly true of meetings to which no numbers have yet been assigned. Abstracts should be submitted on special forms which are available in most departments of mathematics; forms can also be obtained by writing to the headquarters of the Society. Abstracts to be presented at the meeting in person must be received at the headquarters of the Society in Providence, Rhode Island, on or before the deadline for the meeting. Meeting Deadline for Abstracts* Number Date Place and News Items 716 October 26, 1974 Middletown, Connecticut Sept. 3, 1974 717 November 8-9, 1974 Nashville, Tennessee Sept, 25, 1974 718 November 23, 1974 Los Angeles, California Sept. 25, 1974 719 November 23, 1974 Houston, Texas Sept. 25, 1974 720 January 23-27, 1975 Washington, D. C. Nov, 6, 1974 (81st Annual Meeting) March 20-21, 1975 Mobile, Alabama March 23-26, 1975 New York, New York April 18-19, 1975 Monterey, California August 18-22, 1975 Kalamazoo, Michigan November 7-8, 1975 Blacksburg, Virginia November 15, 1975 Los Angeles, California January 22-26, 1976 San Antonio, Texas (82nd Annual Meeting) *Deadline for abstracts not presented at a meeting (by title). October 1974 issue: August 29 OTHER EVENTS January 1975 Symposium on Some Mathematical Questions in Biology New York, New York November 6, 1974 January 1975 Symposium on Theory vs.
    [Show full text]
  • A SHORT HISTORY of CIRCUITS and SYSTEMS CIRCUITS a SHORT HISTORYA SHORT of CIRCUITS and SYSTEMS CIRCUITS and Franco Maloberti and Anthony C
    A SHORT HISTORY OF A SHORT HISTORY OF CIRCUITS AND SYSTEMS A SHORT HISTORY OF A SHORT HISTORY CIRCUITS AND SYSTEMS CIRCUITS AND Franco Maloberti and Anthony C. Davies (Editors) SYSTEMS After an overview of major scientific discoveries of the 18th and 19th Franco Maloberti and Anthony C. Davies (Editors) centuries, which created electrical science as we know and understand it and led to its useful applications in energy conversion, transmission, manufacturing industry and communications, this Circuits and Systems History book fills a gap in published literature by providing a record of the many outstanding scientists, mathematicians and engineers who laid the foundations of Circuit Theory and Filter Design from the mid-20th Century. Additionally, the book records the history of the IEEE Circuits and Systems Society from its origins as the small Circuit Theory Group of the Institute of Radio Engineers (IRE), which merged with the American Institute of Electrical Engineers (AIEE) to form IEEE in 1963, to the large and broad-coverage worldwide IEEE Society which it is today. Many authors from many countries contributed to the creation of this book, working to a very tight time-schedule. The result is a substantial contribution to their enthusiasm and expertise which it is hoped that readers will find both interesting and useful. It is sure that in such a book omissions will be found and in the space and time available, much valuable material had to be left out. It is hoped that this book Anthony C. Davies (Editors) Franco Maloberti and will stimulate an interest in the marvellous heritage and contributions that have come from the many outstanding people who worked in the Circuits and Systems area.
    [Show full text]
  • Mathematical Sciences Meetings and Conferences Section
    OTIC-ES OF THE AMERICAN MATHEMATICAL SOCIETY 1990 Steele Prizes page 801 1990 Norbert Wiener Prize page 808 Renewing U.S. Mathematics A Plan for the 1990s page 813 Applying for NSF Support page 838 Mathematical Education page 844 Serving at the NSF page 851 SEPTEMBER 1990, VOLUME 37, NUMBER 7 Providence, Rhode Island, USA ISSN 0002-9920 ' Calendar of AMS Meetings and Conferences This calendar lists all meetings which have been approved prior to Mathematical Society in the issue corresponding to that of the Notices the date this issue of Notices was sent to the press. The summer which contains the program of the meeting, insofar as is possible. and annual meetings are joint meetings of the Mathematical Associ­ Abstracts should be submitted on special forms which are available in ation of America and the American Mathematical Society. The meet­ many departments of mathematics and from the headquarters office ing dates which fall rather far in the future are subject to change; this of the Society. Abstracts of papers to be presented at the meeting is particularly true of meetings to which no numbers have been as­ must be received at the headquarters of the Society in Providence, signed. Programs of the meetings will appear in the issues indicated Rhode Island, on or before the deadline given below for the meet­ below. First and supplementary announcements of the meetings will ing. Note that the deadline for abstracts for consideration for pre­ have appeared in earlier issues. sentation at special sessions is usually three weeks earlier than that Abstracts of papers presented at a meeting of the Society are pub­ specified below.
    [Show full text]
  • A Biographical Note on John Forbes Nash, Jr. (1928- Prepared by L
    A Biographical Note on John Forbes Nash, Jr. (1928- Prepared by L. Karstensson 1/4/2002 1. General Comment Mathematician; Awarded The Central Bank ofi Sweden Prize in Economic Science in Memory of Alfred Nobel in 1994 for work in the theory of games. Biographer, Sylvia Nasar, summarizes her work, Nash's life, as "a story about the mystery of the human mind, in three acts: genius, madness, reawakening" (1, p. 22]. 2. Chronology 1928 Birth: Born on June 13 in the Bluefield Sanitarium in Bluefield,. West Virginia. He was the older of two children of John Forbes Nash, Sr. and Margaret Virginia Martin. The father was a Texas A&M trained electrical engineer; worked for the Appalachian Power Company. The mother studied languages at Martha Washington College and University of West Virginia; taught school for ten years before her marriage. Nash's younger sister is Martha Nash Legg. 1928-1945 Childhood and youth: Attended Bluefield public schools. Read E. T. Bell, Men of Mathematics. Took some mathematics and science courses at Bluefield College beginning in 1941 at age 13. 1945-1952 University education: 1945-1948 Undergraduate studies at Carnegie Institute of Technology: Began studying engineering, moved to chemistry, and finally completed a major in mathematics. Took only one course in economics, an.elective course in international economics. Studied mathematics under John Synge and Richard Duffin. Some of the faculty referred to him as "a young Gauss" [l, pp. 42]. Received both a BA and MA in mathematics in 1948. Gained admission to the top mathematics .graduate programs in the country: Chicago, Harvard, Michigan, and Princeton.
    [Show full text]