Neural Networks

Total Page:16

File Type:pdf, Size:1020Kb

Neural Networks SLAC-TN-03-028 Neural Networks Patrick Smith Office of Science, Student UndergraduateLaboratory Internship (SULI) Stanford University Stanford Linear Accelerator Center Menlo Park, California August 14,2003 Preparedin partial fulfillment of the requirementsof the Office of Science,Department of Energy’s ScienceUndergraduate Laboratory Internship under the direction of Tony Johnson. Participant: Signature ResearchAdvisor: Signature Work supported in part by the Department of Energy contract DE-AC03-76SF00515. INTRODUCTION Physicistsuse large detectorsto measureparticles createdin high-energy collisions at particle accelerators.These detectors typically producesignals indicating either where ionization occurs along the path of the particle, or where energyis depositedby the.particle.The data producedby thesesignals is fed into pattern recognition programsto try to identify what particles were produced,and to measurethe energy and direction of theseparticles. Ideally, there are many techniquesused in this pattern recognition software. One technique,neural networks, is particularly suitable for identifying what type of particle causedby a set of energy deposits. Neural networks can derive meaningfrom complicatedor imprecisedata, extract patterns,and detect trendsthat are too complex to be noticed by either humansor other computer related processes. To assistin the advancementof this technology,Physicists use a tool kit to experimentwith severalneural network techniques.The goal of this researchis interface a neural network tool kit into JavaAnalysis Studio (JAS3), an applicationthat allows data to be analyzedfrom any experiment. As the final result, a physicist will have the ability to train, test, and implement a neural network with the desiredoutput while using JAS3 to analyzethe results or output. Before an implementationof a neural network can take place, a firm understandingof what a neural network is and how it works is beneficial. A neural network is an artificial representation of the humanbrain that tries to simulatethe learning process[5]. It is also important to think of the word artificial in that definition as computerprograms that use calculationsduring the learning process. In short, a neural network learnsby representativeexamples. Perhapsthe easiestway to describethe way neural networks learn is to explain how the humanbrain functions. The humanbrain containsbillions of neural cells that are responsiblefor processing information [2]. Each one of thesecells acts as a simple processor.When individual cells interact with one another, the complex abilities of the brain are made possible. In neural networks, the input or data are processedby a propagationfunction that adds up the values of all the incoming data. The ending value is then comparedwith a threshold or specific value. The resulting value must exceedthe activation function value in order to becomeoutput. The activation function is a mathematicalfunction that a neuron usesto produce an output referring to its input value. [8] Figure 1 depicts this process.Neural networks usually have three componentsan input, a hidden, and an output. These Iayerscreate the end result of the neural network. A real world example is a child associatingthe word dog with a picture. The child saysdog and simultaneouslylooks a picture of a dog. The input is the spoken word “dog”, the hidden is the brain processing,and the output will be the category of the word dog basedon the picture. This illustration describeshow a neural network functions. MATERIALS AND METHODS Before the implementation of a neural network tool kit into JAS3, JAS3 and all the componentshad to be installed. An editor that was capableof the processeshad to be obtained and downloadedsuccessfully on the workstation. The neural network that we originally used was Java Object Oriented Neural Engine (JOONB). This editor would eventually have to be replacedas you will find out why in the conclusion of this paper. There were severalfactors that were involved in choosing the editor. It had to be user- friendly, free software, flexible and capableof handling the amount of data for our purpose. The tool-kit will be used by physicists who are analyzing their specific data, so the application must be easy to use and understand. An equally important factor is that the editor is written in Java 2 programminglanguage because the editor will be interfacedwith JAS3, a Javabased application. Many companiessell neural network software. Thesesoftware packages are very expensive. For the purposeof our researchthere was no needfor one of those software packages.A more generaleditor would be beneficial. An editor that was flexible was also important. An editor with functionality that includes graphsand many other easyto use featureswould improve usability. There are many editors via the World Wide Web that met our needsand were free to download along with open sourcecode. Finally, the editor had to be able to handle the specified amount of data. Searchingfor a neural network editor involved the exploration of the intemet. Many networks were available,however; five met most of the qualifications. The Stuttgart Neural Network simulator met the qualifications by a researcherat SLAC on a previous project. NeuroSolutionsmet all the requirementsexcept that it was not free; in fact it was quite expensive. Cortex also met all the requirementsexcept that it was written in C++. JOONE satisfiedall requirements,so originally JOONE was chosen. Now that the editor was chosen,understanding the componentsof it was important. JOONE a GUI editor used a graphical user interface that allows you to create,modify, and train a neural network. JOONE for the most part is very self explanatoryand there is a tutorial that is somewhathelpful. In addition the creator of JOONE, Paolo Marrone, respondedto questions concerninghis editor that were unclear to us. Figure 2 showsthe JOONE editor. For this study we have useda neural network to identify calorimeter clustersin a possible linear collinear collider detector. In such a detectorhundreds of particles are createdin each collision, eachin which depositsits energyinto a set of calorimeter cells. Software groups the energy depositsinto clusters,representing the energydeposited by a single particle. By measuringvarious propertiesof the cluster and feeding thesemeasurements into a neural network we can attempt to estimatewhat type of particle createdeach cluster. To train the network we use simulatedevents, in which we alreadyknow which type of particle createdeach energy deposit. The data that we usedin the neural net camefrom Gary Bower a SLAC physicist who createda neural network using a different editor. His training and validation data was theoretically valuable in a couple of instances.When training a neural network, the data should be trained on more than one net to contrastand compareresults. Bower’s data gave us that elementfor comparison. Also, we could use it to train in different modesjust to seethe varying outputs. We constructeda neural network that included 3 layers. The input layer consistedof 15, the hidden 2.5,and the output 6. That meansthat the data set, the first 15 numbersin a row would be the input and the last 6 would be the desiredoutput accordingto the data combination of the input. The hidden layer varies. According to Kevin Swinger, the hidden layer value should be no more than twice the input. It is practical to experimentwith the hidden layer value. When the layers where finished, anothercomponent was usedto read the data into the network referred to as the file input. The componentsimply read the first 15 columns into the network. Another componentcalled the teacherread the last 6 columns or desiredresults into the network. The 21 rows of data that we usedto train our network was the form of binary numbers. The first 15 inputs were named: NE [0], NE [I], NE [2], NE [O]/NE [l], NE [l]/NE [2], firs& IastL, length, firstDdiff, aveLE5, angsep*lOOOO,aveLHits2, nhits, ClusEtot, and CE[2] was the quantitiesthat describedthe clusters. The last 6 desiredoutputs were nngamma,nnchhad, nneuhad,nngammafrag, nnhadfrag, and nnnone, Thesequantities identified what type of cluster it was. Example, if the first 15 binary numbersdescribe the last 6 rows to be “1 0 0 0 0 0" it would be considereda nngammacluster. This is because“ 1 0 0 0 0 0” = nngammain binary, as “0 0 1 0 0 0” = to a nneuhadcluster in binary. Seetable 1 for completetranslations. We set the initial learning rate q (eta) and momentumterms LZ(alpha). Eta dictatesthe proportion of the calculatederror that contributesto the weight change. Alpha relatesto the size of the previous updatefor eachweight. [8] Although we left the valuesto the editor default, the valueswere important to monitor and prevent over fitting the data. We then set the number of cycles (epochs)for the network to train the data to 10,000. The network doesthe rest. The purposeof training the net is to minimize the error which the network makesat each output unit over the entire data set. [S] So while training we looked for our error to move closer and closer to zero. It will never actually reachzero. If it did the network simply memorizedthe training data. So if the training error reacheda predeterminedtarget value, flattened out, or start to rise, training should be stopped. The last stepwas to read the output into a file for comparison. The output was read into AIDA. AIDA histograminterfaces allowed for a simple and easyto interpret implementationof the output. RESULTS
Recommended publications
  • Reverse Image Search for Scientific Data Within and Beyond the Visible
    Journal, Vol. XXI, No. 1, 1-5, 2013 Additional note Reverse Image Search for Scientific Data within and beyond the Visible Spectrum Flavio H. D. Araujo1,2,3,4,a, Romuere R. V. Silva1,2,3,4,a, Fatima N. S. Medeiros3,Dilworth D. Parkinson2, Alexander Hexemer2, Claudia M. Carneiro5, Daniela M. Ushizima1,2,* Abstract The explosion in the rate, quality and diversity of image acquisition instruments has propelled the development of expert systems to organize and query image collections more efficiently. Recommendation systems that handle scientific images are rare, particularly if records lack metadata. This paper introduces new strategies to enable fast searches and image ranking from large pictorial datasets with or without labels. The main contribution is the development of pyCBIR, a deep neural network software to search scientific images by content. This tool exploits convolutional layers with locality sensitivity hashing for querying images across domains through a user-friendly interface. Our results report image searches over databases ranging from thousands to millions of samples. We test pyCBIR search capabilities using three convNets against four scientific datasets, including samples from cell microscopy, microtomography, atomic diffraction patterns, and materials photographs to demonstrate 95% accurate recommendations in most cases. Furthermore, all scientific data collections are released. Keywords Reverse image search — Content-based image retrieval — Scientific image recommendation — Convolutional neural network 1University of California, Berkeley, CA, USA 2Lawrence Berkeley National Laboratory, Berkeley, CA, USA 3Federal University of Ceara,´ Fortaleza, CE, Brazil 4Federal University of Piau´ı, Picos, PI, Brazil 5Federal University of Ouro Preto, Ouro Preto, MG, Brazil *Corresponding author: [email protected] aContributed equally to this work.
    [Show full text]
  • Neural Network FAQ, Part 1 of 7
    Neural Network FAQ, part 1 of 7: Introduction Archive-name: ai-faq/neural-nets/part1 Last-modified: 2002-05-17 URL: ftp://ftp.sas.com/pub/neural/FAQ.html Maintainer: [email protected] (Warren S. Sarle) Copyright 1997, 1998, 1999, 2000, 2001, 2002 by Warren S. Sarle, Cary, NC, USA. --------------------------------------------------------------- Additions, corrections, or improvements are always welcome. Anybody who is willing to contribute any information, please email me; if it is relevant, I will incorporate it. The monthly posting departs around the 28th of every month. --------------------------------------------------------------- This is the first of seven parts of a monthly posting to the Usenet newsgroup comp.ai.neural-nets (as well as comp.answers and news.answers, where it should be findable at any time). Its purpose is to provide basic information for individuals who are new to the field of neural networks or who are just beginning to read this group. It will help to avoid lengthy discussion of questions that often arise for beginners. SO, PLEASE, SEARCH THIS POSTING FIRST IF YOU HAVE A QUESTION and DON'T POST ANSWERS TO FAQs: POINT THE ASKER TO THIS POSTING The latest version of the FAQ is available as a hypertext document, readable by any WWW (World Wide Web) browser such as Netscape, under the URL: ftp://ftp.sas.com/pub/neural/FAQ.html. If you are reading the version of the FAQ posted in comp.ai.neural-nets, be sure to view it with a monospace font such as Courier. If you view it with a proportional font, tables and formulas will be mangled.
    [Show full text]
  • From Digital Computing to Brain-Like Neuromorphic Accelerator
    Mahyar SHAHSAVARI Lille University of Science and Technology Doctoral Thesis Unconventional Computing Using Memristive Nanodevices: From Digital Computing to Brain-like Neuromorphic Accelerator Supervisor: Pierre BOULET Thèse en vue de l’obtention du titre de Docteur en INFORMATIQUE DE L’UNIVERSITÉ DE LILLE École Doctorale des Sciences Pour l’Ingénieur de Lille - Nord De France 14-Dec-2016 Jury: Hélène Paugam-Moisy Professor of University of the French West Indies (Université des Antilles), France (Rapporteur) Michel Paindavoine Professor of University of Burgundy (Université de Bourgogne), France (Rapporteur) Virginie Hoel Professor of Lille University of Science and Technology, France(Président) Said Hamdioui Professor of Delft University of Technology, The Netherlands Pierre Boulet Professor of Lille University of Science and Technology, France Acknowledgment Firstly, I would like to express my sincere gratitude to my advisor Prof. Pierre Boulet for all kind of supports during my PhD. I never forget his supports specially during the last step of my PhD that I was very busy due to teaching duties. Pierre was really punctual and during three years of different meetings and discussions, he was always available on time and as a record he never canceled any meeting. I learned a lot form you Pierre, specially the way of writing, communicating and presenting. you were an ideal supervisor for me. Secondly, I am grateful of my co-supervisor Dr Philippe Devienne. As I did not knew any French at the beginning of my study, it was not that easy to settle down and with Philippe supports the life was more comfortable and enjoyable for me and my family here in Lille.
    [Show full text]
  • Unconventional Computing Using Memristive Nanodevices: from Digital Computing to Brain-Like Neuromorphic Accelerator
    Mahyar SHAHSAVARI Lille University of Science and Technology Doctoral Thesis Unconventional Computing Using Memristive Nanodevices: From Digital Computing to Brain-like Neuromorphic Accelerator Supervisor: Pierre BOULET Thèse en vue de l’obtention du titre de Docteur en INFORMATIQUE DE L’UNIVERSITÉ DE LILLE École Doctorale des Sciences Pour l’Ingénieur de Lille - Nord De France 14-Dec-2016 Jury: Hélène Paugam-Moisy Professor of University of the French West Indies (Université des Antilles), France (Rapporteur) Michel Paindavoine Professor of University of Burgundy (Université de Bourgogne), France (Rapporteur) Virginie Hoel Professor of Lille University of Science and Technology, France(Président) Said Hamdioui Professor of Delft University of Technology, The Netherlands Pierre Boulet Professor of Lille University of Science and Technology, France Acknowledgment Firstly, I would like to express my sincere gratitude to my advisor Prof. Pierre Boulet for all kind of supports during my PhD. I never forget his supports specially during the last step of my PhD that I was very busy due to teaching duties. Pierre was really punctual and during three years of different meetings and discussions, he was always available on time and as a record he never canceled any meeting. I learned a lot form you Pierre, specially the way of writing, communicating and presenting. you were an ideal supervisor for me. Secondly, I am grateful of my co-supervisor Dr Philippe Devienne. As I did not knew any French at the beginning of my study, it was not that easy to settle down and with Philippe supports the life was more comfortable and enjoyable for me and my family here in Lille.
    [Show full text]
  • Neural Network FAQ, Part 1 of 7: Introduction
    Neural Network FAQ, part 1 of 7: Introduction Archive-name: ai-faq/neural-nets/part1 Last-modified: 2002-05-17 URL: ftp://ftp.sas.com/pub/neural/FAQ.html Maintainer: [email protected] (Warren S. Sarle) Copyright 1997, 1998, 1999, 2000, 2001, 2002 by Warren S. Sarle, Cary, NC, USA. --------------------------------------------------------------- Additions, corrections, or improvements are always welcome. Anybody who is willing to contribute any information, please email me; if it is relevant, I will incorporate it. The monthly posting departs around the 28th of every month. --------------------------------------------------------------- This is the first of seven parts of a monthly posting to the Usenet newsgroup comp.ai.neural-nets (as well as comp.answers and news.answers, where it should be findable at any time). Its purpose is to provide basic information for individuals who are new to the field of neural networks or who are just beginning to read this group. It will help to avoid lengthy discussion of questions that often arise for beginners. SO, PLEASE, SEARCH THIS POSTING FIRST IF YOU HAVE A QUESTION and DON'T POST ANSWERS TO FAQs: POINT THE ASKER TO THIS POSTING The latest version of the FAQ is available as a hypertext document, readable by any WWW (World Wide Web) browser such as Netscape, under the URL: ftp://ftp.sas.com/pub/neural/FAQ.html. If you are reading the version of the FAQ posted in comp.ai.neural-nets, be sure to view it with a monospace font such as Courier. If you view it with a proportional font, tables and formulas will be mangled.
    [Show full text]
  • European Journal of Computer Science and Information
    European Journal of Computer Science and Information Technology Vol.9, No.2, pp.21-28, 2021 Print ISSN: 2054-0957 (Print), Online ISSN: 2054-0965 (Online) COMPARING THE TECHNOLOGIES USED FOR THE APPLICATION OF ARTIFICIAL NEURAL NETWORKS Elda Xhumari PhD Candidate, University of Tirana, Faculty of Natural Sciences, Department of Informatics ABSTRACT: In the past few years, artificial neural networks (ANNs) have become a practice wherever it is necessary to solve problems of forecasting, classification, or control. ANNs are intuitively attractive because they are based on a primitive biological model of nervous systems. In the future, the development of such neurobiological models may lead to the creation of truly thinking computers. The areas of application of neural networks are very diverse - these are text and speech recognition, semantic search, expert systems, and decision support systems, prediction of stock prices, security systems, text analysis, etc. Based on the wide application of artificial neural networks, the application of numerous and diverse technologies can be inferred. In this research paper, a comparative approach towards the above-mentioned technologies will be undertaken in order to identify the optimal technology for each problem area in accordance with their respective advantages and disadvantages. KEYWORDS: artificial neural networks, applications, modules INTRODUCTION Artificial Neural Network (ANN) is a computational method that integrates the concepts of the human biological brain to computers for machine learning. The main purpose is to enable computers to adapt and learn the exact behaviour of human beings. As such, machine learning facilitates the continued processing of information that leads to the capacity to solve complex problems and equations that are beyond human ability (Goncalves et al., 2013).
    [Show full text]
  • Data Mining Products
    APPENDIX A Data Mining Products A.1 BIBLIOGRAPHIC NOTES The following paragraphs provide a superficial overview of data mining tools on the market. Links are provided to resources where more detailed information can be found. 4Thought • Product: 4Thought • Vendor: Cognos Inc. • URL: www.cognos.com/products/4thought • Functions: Prediction • Techniques: Neural networks (MLP) • Platforms: Windows 4Thought supports decision support applications, including forecasting. A spread- sheet interface is provided. Data can be accessed as text files, Excel, Lotus, dBase, Impromptu, Powerplay, and Scenario files. AC2 • Product: AC2 • Vendor: ISoft • URL: www.alice-soft.com/html/prod ac2.htm • Functions: Clustering, classification, prediction, segmentation • Techniques: Decision trees • Platforms: Unix, Windows AC2 is a complete data mining toolkit designed for knowledgeable users. AC2 has a graphical object-oriented modeling language and C/C++ libraries. Interactive editing of trees is supported. It behaves as a multiplatform library of data mining functions. Affinium Model • Product: Affinium Model • Vendor: Unica Corporation • URL: www.unicacorp.com • Functions: Classification, segmentation • Platforms: Unix, Windows Affinium Model provides CRM applications with segmentation capabilities. 274 Appendix A Data Mining Products 275 AI Trilogy • Product: AI Trilogy • Vendor: Ward Systems Group, Inc. • URL: www.wardsystems.com • Functions: Classification, forecasting, prediction • Techniques: Genetic algorithms, jackknife estimation, neural networks • Platforms:
    [Show full text]
  • Neural Network Tools: an Overview from Work on "Automatic Classification of Alzheimer’S Disease from Structural MRI" Report for Master’S Thesis
    Neural Network Tools: An Overview From work on "Automatic Classification of Alzheimer’s Disease from Structural MRI" Report for Master’s Thesis Eivind Arvesen Østfold University College [email protected] Abstract Machine learning has exploded in popularity during the last decade, and neural networks in particular have seen a resurgence due to the advent of deep learning. In this paper, we examine modern neural network software-libraries and describe the main differences between them, as well as their strengths and weaknesses. 1 Introduction at The University of Montreal, built with re- search in mind. It has a particular focus on The field of machine learning is now more al- deep learning. The library is focused on easy luring than ever, helping researchers and busi- configurability for expert users (i.e. machine nesses to tackle a variety of challenging prob- learning researchers) —unlike "black-box" li- lems such as development of autonomous vehi- braries, which provide good performance with- cles, visual recognition systems and computer- out demanding knowledge about underlying al- aided diagnosis. There is a plethora of different gorithms from users. One way to put it would tools and libraries available that aim to make be that the library values ease of experimenta- work on machine learning problems easier for tion over ease of use. users, and while most of the more general alter- natives are quite similar in terms of functional- Though Pylearn2 assumes some technical ity, they nonetheless differ in terms of approach sophistication on the part of users, the library and design goals. also has a focus on reusability, and the re- This report describes several popular ma- sulting modularity makes it possible to com- chine learning tools, attempts to showcase their bine and adapt several re-usable components strengths and characterizes situations in which to form working models, and to only learn they would be well suited for use.
    [Show full text]
  • Performance of Neural Network Image Classification on Mobile CPU and GPU
    Master’s Programme in Computer, Communication and Information Sciences Performance of Neural Network Image Classification on Mobile CPU and GPU Sipi Seppälä MASTER’S THESIS Aalto University MASTER’S THESIS 2018 Performance of Neural Network Image Classifcation on Mobile CPU and GPU Sipi Seppälä Otaniemi, 2018-04-20 Supervisor: Professor Antti Ylä-Jääski Advisors: M.Sc. Teemu Kämäräinen Dr. Matti Siekkinen Aalto University School of Science Master’s Programme in Computer, Communication and Information Sciences Abstract Aalto University, P.O. Box 11000, FI-00076 Aalto www.aalto.fi Author Sipi Seppälä Title Performance of Neural Network Image Classifcation on Mobile CPU and GPU School School of Science Master’s programme Computer, Communication and Information Sciences Major Computer Science Code SCI3042 Supervisor Professor Antti Ylä-Jääski Advisors M.Sc. Teemu Kämäräinen, Dr. Matti Siekkinen Level Master’s thesis Date 2018-04-20 Pages 86 Language English Abstract Artifcial neural networks are a powerful machine learning method, with impressive results lately in the feld of computer vision. In tasks like image classifcation, which is a well-known problem in computer vision, deep learning convolutional neural networks can even achieve human-level prediction accuracy. Although high-accuracy deep neural networks can be resource-intensive both to train and to deploy for inference, with the advent of lighter mobile-friendly neural network model architectures it is fnally possible to achieve real-time on-device inference without the need for cloud offoading. The inference performance can be further improved by utilizing mobile graphics processing units which are already capable of general-purpose parallel computing. This thesis measures and evaluates the performance aspects – execution latency, through- put, memory footprint, and energy usage – of neural network image classifcation inference on modern smartphone processors, namely CPU and GPU.
    [Show full text]
  • Managing a Real-Time Massively-Parallel Neural Architecture
    MANAGING A REAL-TIME MASSIVELY-PARALLEL NEURAL ARCHITECTURE A THESIS SUBMITTED TO THE UNIVERSITY OF MANCHESTER FOR THE DEGREE OF DOCTOR OF PHILOSOPHY IN THE FACULTY OF ENGINEERING AND PHYSICAL SCIENCES 2012 By James Cameron Patterson School of Computer Science Contents Abstract 11 Declaration 12 Copyright 13 Acknowledgements 14 The Author 15 1 Introduction 16 1.1 Research Motivation .......................... 17 1.1.1 Not Chips,but Cores with Everything ............. 17 1.1.2 Large System Management ................... 20 1.1.3 Neural Networks on Parallel Computers ............ 20 1.1.4 Monitoring the SpiNNaker Neural System ........... 21 1.2 Contributions .............................. 23 1.3 Publications ............................... 24 1.4 Thesis Overview ............................ 26 2 Neural Computing 29 2.1 Biology ................................. 30 2.2 Neural Network Modelling ....................... 33 2.2.1 First Generation Artificial Neural Networks (ANNs) ..... 34 2.2.2 Second Generation ANNs ................... 34 2.2.3 Third Generation ANNs .................... 36 2.3 Systems for Neural Network Modelling ................ 39 2.3.1 General Purpose Hardware ................... 40 2.3.2 Specialised Hardware Platforms ................ 45 2 2.3.3 The Centre Ground ....................... 48 3 The SpiNNaker Neural Architecture 50 3.1 Architectural Requirements ....................... 51 3.2 System Architecture ........................... 51 3.3 SpiNNaker Systems ........................... 53 3.4 SpiNNaker Communications .....................
    [Show full text]
  • Journal of Science Development of Web Based Courseware For
    Research Article GU J Sci 32(4): 1138-1148 (2019) DOI: 10.35378/gujs.473450 Gazi University Journal of Science http://dergipark.gov.tr/gujs Development of Web Based Courseware for Artificial Neural Networks Mehmet BILEN1,* , Ali Hakan ISIK2 , Tuncay YIGIT3 1 Mehmet Akif Ersoy University, Cavdir Vocational High School, 15000 Burdur, Turkey 2 Mehmet Akif Ersoy University, Faculty of Engineering and Architecture, 15000, Burdur, Turkey 3 Suleyman Demirel University, Faculty of Engineering, 32000, Isparta, Turkey Highlights • This paper focuses on developing a web based educational courseware. • Artificial neural network procedures can be performed and results can be traced in the courseware. • Courseware has the ability to change initial parameters and observe outputs in real time. • Students can understand better the working mechanism of ANN with the help of the courseware. Article Info Abstract Artificial Neural Networks (ANN) are important data processing algorithms which are used for Received: 22/10/2018 solving nonlinear problems. Through classical approaches, mathematical infrastructure and Accepted: 05/06/2019 complex equations in ANN are difficult to understand. Interactive and multimedia-based courseware has the potential to overcome these difficulties. In this study, a web based educational courseware for ANN was developed to provide an effective and efficient learning environment Keywords so that the difficulties can be overcome. This interactive courseware was also enriched with Artificial Neural animations and text-based course contents. In addition to this, the effects of ANN parameters’ Networks changes were observed directly through graphical results. In this way, users can easily understand E-Learning the fundamentals and working mechanism of ANN.
    [Show full text]
  • Development of Artificial Neural Network Software And
    DEVELOPMENT OF ARTIFICIAL NEURAL NETWORK SOFTWARE AND MODELS FOR ENGINEERING MATERIALS ABDALLAH F. BSEISO Bachelor of Civil Engineering Palestine Polytechnic University June 2017 submitted in partial fulfillment of requirements for the degree MASTER OF SCIENCE IN CIVIL ENGINEERING at the CLEVELAND STATE UNIVERSITY December 2020 We hereby approve this thesis For ABDALLAH F. BSEISO Candidate for the Master of Civil Engineering degree For the Department of Civil and Environmental Engineering And CLEVELAND STATE UNIVERSITY’S College of Graduate Studies by _____________________________ Dr. Josiah Owusu-Danquah Advisor, Chair _____________________________ Department & Date _____________________________ Dr. Srinivas Allena _____________________________ Department & Date _____________________________ Dr. Stephen F Duffy _____________________________ Department & Date Student’s Date of Defense: 11/24/2020 DEDICATION To my family, who have given all the support through these years, my professors, and friends. DEVELOPMENT OF ARTIFICIAL NEURAL NETWORK SOFTWARE AND MODELS FOR ENGINEERING MATERIALS. ABDALLAH F. BSEISO ABSTRACT Artificial Neural Network (ANN), which is inspired by biological neural networks in the human brain, is one important tool of machine learning that creates artificial intelligence through computational systems. The creation of this intelligence is contingent on learning from available data regarding a specific subject. Although machine learning, in general, has profuse applications in most scientific disciplines, yet few have been developed in civil engineering due to the required time consuming and demanding programming. In order to minimize this, intelligible ANN software has been developed in this research capable of training networks with any number of hidden layers and nodes for each layer. Furthermore, two models have been created to demonstrate the robust applications of ANN.
    [Show full text]