Uhm Ms 3812 R.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

Uhm Ms 3812 R.Pdf UNIVERSITY OF HAWAI'I LIBRARY ADVANCED MARINE VEIDCLE PRODUCTS DATABASE -A PRELIMINARY DESIGN TOOL A THESIS SUBMITTED TO THE GRADUATE DMSION OF THE UNIVERSITY OF HAWAI'I IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF MASTER OF SCIENCE IN OCEAN ENGINEERING AUGUST 2003 By Kristen A.L.G. Woo Thesis Committee: Kwok Fai Cheung, ChaiIperson Hans-Jurgen Krock John C. Wiltshire ACKNOWLEDGEMENT I would like to thank my advisor Prof. Kwok Fai Cheung and the other committee members Prof. Hans-Jiirgen Krock and Dr. John Wiltshire for the time and effort they spent with me on this project. I would also like to thank the MHPCC staff, in particular, Mr. Scott Splean for their advice and comments on the advanced-marine-vehicle products database. Thanks are also due to Drs. Woei-Min Lin and JUll Li of SAIC for their advice on the neural network and preliminary ship design tools. I would also like to thank Mr. Yann Douyere for his help with MatLab. The work described in this thesis is a subset of the project "Environment for Design of Advanced Marine Vehicles and Operations Research" supported by the Office of Naval Research, Grant No. NOOOI4-02-1-0903. iii ABSTRACT The term advanced marine vehicle encompasses a broad category of ship designs typically referring to multihull ships such as catamarans, trimarans and SWATH (small waterplane area twin hull) ships, but also includes hovercrafts, SES (surface effect ships), hydrofoils, and advanced monohulls. This study develops an early stage design tool for advanced marine vehicles that provides principal particulars and additional parameters such as fuel capacity and propulsive power based on input ship requirements. This is accomplished by compiling a product database of existing advanced marine vehicles for the development of relationships between ship characteristic parameters. The relationships are analyzed using both the multiple regression analysis and a neural network. Because of the scatter of the data, the first order or linear multiple regression analysis is adopted. The neural network, on the other hand, is a non-linear learning tool that uses existing ship parameters to predict future ship designs. The results are compared with the actual ship parameters to validate the preliminary design tools. The Maui High Performance Computing Center has developed a geographic information system interface for the statistical tools and databases with the additional capabilities to analyze ocean wave environment for the definition ofship design requirements. IV TABLE OF CONTENTS Acknowledgement iii Abstract iv List ofTables vii List ofFigures viii I Introduction 1 1.1 Advanced Marine Vehicles 1 1.2 Ship Design Database .3 1.3 Objectives and Approach 5 2 Advanced Marine Vehicle Products Database 8 2. I Data Sources 8 2.1.1 Ship Data 8 2.1.2 Wave Data 8 2.2 Database Structure 9 2.2.1 SWATH 10 2.2.2 Catamaran and Trimaran II 2.2.3 Hovercraft and SES II 2.2.4 Hydrofoil I2 2.2.5 Sailing Catamarans and Trimarans .1 3 2.3 GIS Interface I5 2.3.1 Informational Features 15 2.3.2 Feature Controls 16 3 Multiple Regression Analysis I8 3. I Theoretical Background I8 3.2 Program Structure I9 3.3 Analysis Routines .20 4 Neural Network .23 4.1 Theoretical Background 23 4.2 Program Layout 25 4.3 Analysis Routines 27 4.3.1 Setup the Data 27 4.3.2 Design the Neural Network 29 4.3.3 Train the Neural Network .30 4.3.4 Test the Neural Network .3 I 4.3.5 Using the Trained Neural Network .3 I 4.3.6 Setup ofthe Trained Neural Network for the User. 32 4.3.7 New Data 32 v 5 Results and Discussion 33 5.1 Multiple Regression Analysis .33 5.1.1 User Instructions .33 5.1.2 Results .34 5.1.2.1 Catamaran Results 34 5.1.2.2 SWATH Ship Results 36 5.1.2.3 SES Results .37 5.1.2.4 Hydrofoil Results .39 5.1.2.5 Hovercraft Results 41 5.2 Neural Network .43 5.2.1 General Setup .44 5.2.2 General Instructions to Operate the Excel Output File .44 5.2.3 Ship Type Recommendations .45 5.2.3.1 User Instructions .46 5.2.3.2 Results 46 5.2.4 Ship Parameters .47 5.2.4.1 General User Instructions .48 5.2.4.2 Catamaran Results .48 5.2.4.3 Hovercraft Results .49 5.2.4.4 SWATH Ship Results 50 5.2.4.5 Hydrofoil Results 51 5.3 Comparison ofResults .52 6 Conclusions and Recommendations 55 References 88 Vi LIST OF TABLES 1. Multiple Regression Analysis 1: Results for Catamarans with 3 Inputs 35 2. Multiple Regression Analysis 2: Results for Catamarans with 2 Inputs 36 3. Multiple Regression Analysis Results for SWATH ships 37 4. Multiple Regression Analysis 1: Results for SES with 3 inputs 38 5. Multiple Regression Analysis 2: Results for SES with 2 inputs 38 6. Multiple Regression Analysis 3: Results for SES with 2 inputs 39 7. Multiple Regression Analysis I: Results for Hydrofoils with 3 inputs 40 8. Multiple Regression Analysis 2: Results for Hydrofoils with 3 inputs 40 9. Multiple Regression Analysis 3: Results for Hydrofoils with 2 inputs 41 10. Multiple Regression Analysis 1: Results for Hovercrafts with 4 inputs 42 11. Multiple Regression Analysis 2: Results for Hovercrafts with 3 inputs 42 12. Multiple Regression Analysis 3: Results for Hovercrafts with 3 inputs 43 13. Multiple Regression Analysis 4: Results for Hovercrafts with 2 inputs 43 14. Neural Network Results for the Recommendation ofShip Types 47 15. Neural Network Results for the Preliminary Design ofCatamarans 49 16. Neural Network Results for the Preliminary Design ofHydrofoils 50 17. Neural Network Results for the Preliminary Design of SWATH ships 51 18. Neural Network Results for the Preliminary Design ofHydrofoils 52 19. Catamaran: Neural Network verses Multiple Regression 53 20. SWATH: Neural Network verses Multiple Regression 54 21. HYdrofoil: Neural Network verses Multiple Regression 54 22. Hovercraft: Neural Network verses Multiple Regression 54 vii LIST OF FIGURES 1. Types ofAdvanced Marine Vehicles (with permission from Mustafa Insel) 57 2. ENDEAVOR GIS Website: AMV Homeports 57 3. ENDEAVOR GIS Website: AMV Routes 58 4. ENDEAVOR GIS Website: Significant Wave Height 58 5. ENDEAVOR GIS Website: Peak Wave Period 59 6. ENDEAVOR GIS Website: Peak Wave Direction 59 7. ENDEAVOR GIS Website: Wind Speed 60 8. ENDEAVOR GIS Website: Wind Direction 60 9. ENDEAVOR GIS Website: Topographic and Bathymetry Data 61 10. ENDEAVOR GIS Website: Query Setup 61 II. ENDEAVOR GIS Website: Query Output... 62 12. ENDEAVOR GIS Website: Ship Photo from AMV Database 62 13. Analyse-It: Tutorial Example ~ 63 14. Analyse-It: Set Dataset Properties Toolbar 64 15. Analyse-It: Dataset Properties Dialog Box 64 16. Analyse-It: Variable Properties Dialog Box 65 17. Analyse-It: Linear Regression Menu 65 18. Analyse-It: Linear Regression Dialog Box 66 19. Analyse-It: Multiple Regression Analysis Report Table 66 20. Analyse-It: Multiple Regression Analysis Report Graphs 67 21. Biological Neuron (from Anderson and McNeil, 1992) 68 22. Neural Network Processing Element (from Anderson and McNeil, 1992) 68 23. Sigmoid Transfer Function (from Anderson and McNeil, 1992) 69 24. Artificial Neural Network (from Anderson and McNeil, 1992) 69 25. Multilayer Perception Model (from Anderson and McNeil, 1992) 70 26. Back-propagation Learning Process (from Anderson and McNeil, 1992) 70 27. NeuroSolutions NeuralExpert Design Tool 7l viii 28. NeuroSolutions NeuralBuilder Design TooL 71 29. NeuroSolutions Training Report Produced in Excel 72 30. NeuroSolutions Preprocess Data Menu 73 31. NeuroSolutions Tag Data Menu 74 32. NeuroSolutions Create/Open Network Menu 75 33. NeuroSolutions Breadboard 75 34. NeuroSolutions Train Network Menu 76 35. NeuroSolutions Train Menu 76 36. NeuroSolutions Train N Times Function 77 37. NeuroSolutions Vary A Parameter Function 78 38. NeuroSolutions Test Network Menu 79 39. NeuroSolutions Apply Production Dataset Function 79 40. NeuroSolutions Input for the Production Dataset.. 80 41. NeuroSolutions Output for the Production Dataset. 81 42. Excel User Application: Introduction Page 82 43. Excel User Application: Input Page 83 44. Excel User Application: Output Page 84 45. Multiple Regression Analysis Output.. 85 46. Neural Network User Application: Ship Type Recommendation Input.. 86 47. Neural Network User Application: Ship Type Recommendation Output 87 IX 1 INTRODUCTION 1.1 Advanced Marine Vehicles Advanced marine vehicle (AMV) encompasses a broad category of ship designs typically referring to multihull ships, but also includes hovercrafts, SES (surface effect ships), hydrofoils, and advanced monohulls. Figure I shows the different types of AMVs. A subset of AMVs, covering SWATH (small waterplane area twin hull) ships, SES, and hydrofoils are, considered advantageous for a number of naval missions (Lavis et aI., 1990). These vessels sustain most of their weight by dynamic lift or by powered aerostatic lift (Eames, 1985). Each ship design category has its own advantages over the widely used monohull design that is seen in many ships today. The first category of advance marine vehicles is multihull vessels, which consist of catamarans, trimarans, and SWATH ships. Catamarans are vessels with twin hulls, while trimarans have two outriggers attached to a center hull. Skinner (200 I) pointed out that catamaran and trimaran hull designs are faster, more stable in high sea states, have similar internal volumes as single-hull ships, and are more fuel-efficient than conventional hull designs. The layout of a trimaran is advantageous for military applications in that the increased beam, compared to typical monohulls, provides an excellent area for a flight deck; the cross deck also provides additional space in the center ofthe ship. The beam ofa trimaran is typically 45 to 60% greater than a monohull with a similar displacement (Greig et a!., 1995).
Recommended publications
  • Neural Networks
    AIAI andand thethe NetNet •• Buyer’sBuyer’s GuideGuide 19.1 Where Intelligent Technology Meets the Real World www.pcai.com IntelligentIntelligent Applications Applications TThehe ModernModern RoboticRobotic Also:Also: “Movement”“Movement” Agents, Business Applications, TThehe StateState ofof AIAI Today:Today: TheThe Business Intelligence, WWeb:eb: AI’sAI’s NewNew PlaygroundPlayground Computational Intelligence, Data Analysis & Mining, Robotics:Robotics: RobotsRobots ThatThat Intelligent Applications, Intelligent Tools, MimicMimic AnimalsAnimals Intelligent Tutoring, Intelligent Web Searching, TThehe NationalNational ScienceScience Neural Networks, Foundation:Foundation: EncouragingEncouraging Robotics, tthehe ResearchersResearchers ofof Speech Recognition, TTomorrowomorrow Web Based Expert Systems, Plus:Plus: AIAI andand thethe Net,Net, Training, Bookzone,Bookzone, Buyer’sBuyer’s Guide,Guide, AI Conferences, ProductProduct UUpdatespdates and more! PC AI 2 19.1 Quantities Limited Buy PC AI Back Issues 1995 1999 A Great Resource 9 #1 Intelligent Tools 13 #1 Intelligent Tools & Languages 9 #2 Fuzzy Logic / Neural Networks (Knowledge Verification) for AI Research 9 #3 Object Oriented Development 13 #2 Rule and Object Oriented 9 #4 Knowledge-Based Systems Development (Data Mining) $8.00/Issue - US 9 #5 AI Languages 13 #3 Neural Nets & Fuzzy Logic (For Us and Canadian and 9 #6 Business Applications (Searching) Foreign Postage 13 #4 Knowledge-Based Systems contact PC AI or visit the 1996 (Fuzzy Logic) 10 #1 Intelligent Applications PC AI web site) 13 #5 Data Mining (Simulation and 10 #2 Object Oriented Development Modeling) Order online at 10 #3 Neural Networks / Fuzzy Logic 13 #6 Business Applications www.pcai.com 10 #4 Knowledge-Based Systems (Machine Learning) Total amount enclosed 10 #5 Genetic Algorithm & Modeling 10 #6 Business Applications 2000 $____________.
    [Show full text]
  • Kriging Prediction with Isotropic Matérn Correlations: Robustness
    Journal of Machine Learning Research 21 (2020) 1-38 Submitted 12/19; Revised 7/20; Published 9/20 Kriging Prediction with Isotropic Mat´ernCorrelations: Robustness and Experimental Designs Rui Tuo∗y [email protected] Wm Michael Barnes '64 Department of Industrial and Systems Engineering Texas A&M University College Station, TX 77843, USA Wenjia Wang∗ [email protected] The Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong Editor: Philipp Hennig Abstract This work investigates the prediction performance of the kriging predictors. We derive some error bounds for the prediction error in terms of non-asymptotic probability under the uniform metric and Lp metrics when the spectral densities of both the true and the imposed correlation functions decay algebraically. The Mat´ernfamily is a prominent class of correlation functions of this kind. Our analysis shows that, when the smoothness of the imposed correlation function exceeds that of the true correlation function, the prediction error becomes more sensitive to the space-filling property of the design points. In particular, the kriging predictor can still reach the optimal rate of convergence, if the experimental design scheme is quasi-uniform. Lower bounds of the kriging prediction error are also derived under the uniform metric and Lp metrics. An accurate characterization of this error is obtained, when an oversmoothed correlation function and a space-filling design is used. Keywords: Computer Experiments, Uncertainty Quantification, Scattered Data Approx- imation, Space-filling Designs, Bayesian Machine Learning 1. Introduction In contemporary mathematical modeling and data analysis, we often face the challenge of reconstructing smooth functions from scattered observations.
    [Show full text]
  • Artificial Neural Networks
    ARTIFICIAL NEURAL NETWORKS: A REVIEW OF TRAINING TOOLS Darío Baptista, Fernando Morgado-Dias Madeira Interactive Technologies Institute and Centro de Competências de Ciências Exactas e da Engenharia, Universidade da Madeira Campus da Penteada, 9000-039 Funchal, Madeira, Portugal. Tel:+351 291-705150/1, Fax: +351 291-705199 Abstract: Artificial Neural Networks became a common solution for a wide variety of problems in many fields. The most frequent solution for its implementation consists of building and training the Artificial Neural Network within a computer. For implementing a network in an efficient way, the user can access a large choice of software solutions either commercial or prototypes. Choosing the most convenient solution for the application according to the network architecture, training algorithm, operating system and price can be a complex task. This paper helps the Artificial Neural Network user by providing a large list of solution available and explaining their characteristics and terms of use. The paper is confined to reporting the software products that have been developed for Artificial Neural Networks. The features considered important for this kind of software to have in order to accommodate its users effectively are specified. The development of software that implements Artificial Neural is a rapidly growing field driven by strong research interests as well as urgent practical, economical and social needs. Copyright CONTROLO2012 Keywords: Artificial Neural Networks, Training Tools, Training Algorithms, Software. 1. INTRODUCTION commercialization of new ANN tools. With the purpose to inform which tools are available at present Nowadays, in different areas, it is important to and make the choice of which tool to use, this paper analyse nonlinear data to do prediction, classification contains the description of software that has been or to build models.
    [Show full text]
  • On Prediction Properties of Kriging: Uniform Error Bounds and Robustness
    On Prediction Properties of Kriging: Uniform Error Bounds and Robustness Wenjia Wang∗1, Rui Tuoy2 and C. F. Jeff Wuz3 1The Statistical and Applied Mathematical Sciences Institute, Durham, NC 27709, USA 2Department of Industrial and Systems Engineering, Texas A&M University, College Station, TX 77843, USA 3The H. Milton Stewart School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA Abstract Kriging based on Gaussian random fields is widely used in reconstructing unknown functions. The kriging method has pointwise predictive distributions which are com- putationally simple. However, in many applications one would like to predict for a range of untried points simultaneously. In this work we obtain some error bounds for the simple and universal kriging predictor under the uniform metric. It works for a scattered set of input points in an arbitrary dimension, and also covers the case where the covariance function of the Gaussian process is misspecified. These results lead to a better understanding of the rate of convergence of kriging under the Gaussian or the Matérn correlation functions, the relationship between space-filling designs and kriging models, and the robustness of the Matérn correlation functions. Keywords: Gaussian Process modeling; Uniform convergence; Space-filling designs; Radial arXiv:1710.06959v4 [math.ST] 19 Mar 2019 basis functions; Spatial statistics. ∗Wenjia Wang is a postdoctoral fellow in the Statistical and Applied Mathematical Sciences Institute, Durham, NC 27709, USA (Email: [email protected]); yRui Tuo is Assistant Professor in Department of Industrial and Systems Engineering, Texas A&M University, College Station, TX 77843, USA (Email: [email protected]); Tuo’s work is supported by NSF grant DMS 1564438 and NSFC grants 11501551, 11271355 and 11671386.
    [Show full text]
  • Rational Function Approximation
    Rational function approximation Rational function of degree N = n + m is written as p(x) p + p x + + p xn r(x) = = 0 1 ··· n q(x) q + q x + + q xm 0 1 ··· m Now we try to approximate a function f on an interval containing 0 using r(x). WLOG, we set q0 = 1, and will need to determine the N + 1 unknowns p0,..., pn, q1,..., qm. Numerical Analysis I – Xiaojing Ye, Math & Stat, Georgia State University 240 Pad´eapproximation The idea of Pad´eapproximation is to find r(x) such that f (k)(0) = r (k)(0), k = 0, 1,..., N This is an extension of Taylor series but in the rational form. i Denote the Maclaurin series expansion f (x) = i∞=0 ai x . Then i m i n i ∞ a x q x P p x f (x) r(x) = i=0 i i=0 i − i=0 i − q(x) P P P If we want f (k)(0) r (k)(0) = 0 for k = 0,..., N, we need the − numerator to have 0 as a root of multiplicity N + 1. Numerical Analysis I – Xiaojing Ye, Math & Stat, Georgia State University 241 Pad´eapproximation This turns out to be equivalent to k ai qk i = pk , k = 0, 1,..., N − Xi=0 for convenience we used convention p = = p = 0 and n+1 ··· N q = = q = 0. m+1 ··· N From these N + 1 equations, we can determine the N + 1 unknowns: p0, p1,..., pn, q1,..., qm Numerical Analysis I – Xiaojing Ye, Math & Stat, Georgia State University 242 Pad´eapproximation Example x Find the Pad´eapproximation to e− of degree 5 with n = 3 and m = 2.
    [Show full text]
  • Piecewise Linear Approximation of Streaming Time Series Data with Max-Error Guarantees
    Piecewise Linear Approximation of Streaming Time Series Data with Max-error Guarantees Ge Luo Ke Yi Siu-Wing Cheng Zhenguo Li Wei Fan Cheng He Yadong Mu HKUST Huawei Noah’s Ark Lab AT&T Labs Abstract—Given a time series S = ((x1; y1); (x2; y2);::: ) and f(xi) yi " for all i. This is because the `1=`2-error is a prescribed error bound ", the piecewise linear approximation ill-suitedj − forj ≤ online algorithms as it is a sum of errors over (PLA) problem with max-error guarantees is to construct a the entire time series. When the algorithm has no knowledge piecewise linear function f such that jf(xi) − yij ≤ " for all i. In addition, we would like to have an online algorithm that takes about the future, in particular the length of the time series n, the time series as the records arrive in a streaming fashion, and it is impossible to properly allocate the allowed error budget outputs the pieces of f on-the-fly. This problem has applications over time. Another advantage of the ` -error is that it gives wherever time series data is being continuously collected, but us a guarantee on any data record in1 the time series, while the data collection device has limited local buffer space and the ` =` -error only ensures that the “average” error is good communication bandwidth, so that the data has to be compressed 1 2 without a bound on any particular record. Admittedly, the ` - and sent back during the collection process. 1 Prior work addressed two versions of the problem, where error is sensitive to outliers, but one could remove them before either f consists of disjoint segments, or f is required to be feeding the stream to the PLA algorithm, and there is abundant a continuous piecewise linear function.
    [Show full text]
  • The Nnlib2 Library and Nnlib2rcpp R Package for Implementing Neural Networks
    The nnlib2 library and nnlib2Rcpp R package for implementing neural networks Vasilis N Nikolaidis1 1 University of Peloponnese DOI: 10.21105/joss.02876 Software • Review Summary • Repository • Archive Artificial Neural Networks (ANN or NN) are computing models used in various data-driven applications. Such systems typically consist of a large number of processing elements (or Editor: Kakia Chatsiou nodes), usually organized in layers, which exchange data via weighted connections. An ever- increasing number of different neural network models have been proposed and used. Among Reviewers: the several factors differentiating each model are the network topology, the processing and • @schnorr training methods in nodes and connections, and the sequences utilized for transferring data • @MohmedSoudy to, within and from the model etc. The software presented here is a C++ library of classes and templates for implementing neural network components and models and an R package Submitted: 22 October 2020 that allows users to instantiate and use such components from the R programming language. Published: 23 May 2021 License Authors of papers retain copyright and release the work Statement of need under a Creative Commons Attribution 4.0 International A significant number of capable, flexible, high performance tools for NN are available today, License (CC BY 4.0). including frameworks such as Tensorflow (Abadi et al., 2016) and Torch (Collobert et al., 2011), and related high level APIs including Keras (Chollet & others, 2015) and PyTorch (Paszke et al., 2019). Ready-to-use NN models are also provided by various machine learning platforms such as H2O (H2O.ai, 2020) or libraries, SNNS (Zell et al., 1994) and FANN (Nissen, 2003).
    [Show full text]
  • Neural Network FAQ, Part 1 of 7
    Neural Network FAQ, part 1 of 7: Introduction Archive-name: ai-faq/neural-nets/part1 Last-modified: 2002-05-17 URL: ftp://ftp.sas.com/pub/neural/FAQ.html Maintainer: [email protected] (Warren S. Sarle) Copyright 1997, 1998, 1999, 2000, 2001, 2002 by Warren S. Sarle, Cary, NC, USA. --------------------------------------------------------------- Additions, corrections, or improvements are always welcome. Anybody who is willing to contribute any information, please email me; if it is relevant, I will incorporate it. The monthly posting departs around the 28th of every month. --------------------------------------------------------------- This is the first of seven parts of a monthly posting to the Usenet newsgroup comp.ai.neural-nets (as well as comp.answers and news.answers, where it should be findable at any time). Its purpose is to provide basic information for individuals who are new to the field of neural networks or who are just beginning to read this group. It will help to avoid lengthy discussion of questions that often arise for beginners. SO, PLEASE, SEARCH THIS POSTING FIRST IF YOU HAVE A QUESTION and DON'T POST ANSWERS TO FAQs: POINT THE ASKER TO THIS POSTING The latest version of the FAQ is available as a hypertext document, readable by any WWW (World Wide Web) browser such as Netscape, under the URL: ftp://ftp.sas.com/pub/neural/FAQ.html. If you are reading the version of the FAQ posted in comp.ai.neural-nets, be sure to view it with a monospace font such as Courier. If you view it with a proportional font, tables and formulas will be mangled.
    [Show full text]
  • Rethinking Statistical Learning Theory: Learning Using Statistical Invariants
    Machine Learning (2019) 108:381–423 https://doi.org/10.1007/s10994-018-5742-0 Rethinking statistical learning theory: learning using statistical invariants Vladimir Vapnik1,2 · Rauf Izmailov3 Received: 2 April 2018 / Accepted: 25 June 2018 / Published online: 18 July 2018 © The Author(s) 2018 Abstract This paper introduces a new learning paradigm, called Learning Using Statistical Invariants (LUSI), which is different from the classical one. In a classical paradigm, the learning machine constructs a classification rule that minimizes the probability of expected error; it is data- driven model of learning. In the LUSI paradigm, in order to construct the desired classification function, a learning machine computes statistical invariants that are specific for the problem, and then minimizes the expected error in a way that preserves these invariants; it is thus both data- and invariant-driven learning. From a mathematical point of view, methods of the classical paradigm employ mechanisms of strong convergence of approximations to the desired function, whereas methods of the new paradigm employ both strong and weak convergence mechanisms. This can significantly increase the rate of convergence. Keywords Intelligent teacher · Privileged information · Support vector machine · Neural network · Classification · Learning theory · Regression · Conditional probability · Kernel function · Ill-Posed problem · Reproducing Kernel Hilbert space · Weak convergence Mathematics Subject Classification 68Q32 · 68T05 · 68T30 · 83C32 1 Introduction It is known that Teacher–Student interactions play an important role in human learning. An old Japanese proverb says “Better than thousand days of diligent study is one day with a great teacher.” What is it exactly that great Teachers do? This question remains unanswered.
    [Show full text]
  • Function Approximation with Mlps, Radial Basis Functions, and Support Vector Machines
    Table of Contents CHAPTER V- FUNCTION APPROXIMATION WITH MLPS, RADIAL BASIS FUNCTIONS, AND SUPPORT VECTOR MACHINES ..........................................................................................................................................3 1. INTRODUCTION................................................................................................................................4 2. FUNCTION APPROXIMATION ...........................................................................................................7 3. CHOICES FOR THE ELEMENTARY FUNCTIONS...................................................................................12 4. PROBABILISTIC INTERPRETATION OF THE MAPPINGS-NONLINEAR REGRESSION .................................23 5. TRAINING NEURAL NETWORKS FOR FUNCTION APPROXIMATION ......................................................24 6. HOW TO SELECT THE NUMBER OF BASES ........................................................................................28 7. APPLICATIONS OF RADIAL BASIS FUNCTIONS................................................................................38 8. SUPPORT VECTOR MACHINES........................................................................................................42 9. PROJECT: APPLICATIONS OF NEURAL NETWORKS AS FUNCTION APPROXIMATORS ...........................52 10. CONCLUSION ..............................................................................................................................59 CALCULATION OF THE ORTHONORMAL WEIGHTS ..................................................................................63
    [Show full text]
  • Forecasting with Artificial Neural Networks
    Forecasting with Artificial Neural Networks EVIC 2005 Tutorial Santiago de Chile, 15 December 2005 Æ slides on www.neural-forecasting.com Sven F. Crone Centre for Forecasting Department of Management Science Lancaster University Management School email: [email protected] EVIC’05 © Sven F. Crone - www.bis-lab.com Lancaster University Management School? EVIC’05 © Sven F. Crone - www.bis-lab.com What you can expect from this session … Simple back propagation algorithm [Rumelhart et al. 1982] ∂C(t pj ,o pj ) E p = C(t pj , o pj ) o pj = f j (net pj ) Δ p w ji ∝ − ∂w ji ∂C(t ,o ) ∂C(t ,o ) ∂net pj pj = pj pj pj ∂w ji ∂net pj ∂w ji ∂C(t ,o ) Æ „How to …“ on Neural δ = − pj pj pj ∂net pj Network Forecasting ∂C(t ,o ) ∂C(t ,o ) ∂o δ = − pj pj = pj pj pj pj with limited maths! ∂net pj ∂opj ∂net pj ∂o pt ' = f j (net pj ) ∂net pj ∂C(t pj ,opj ) ' δ pj = f j (net pj ) Æ CD-Start-Up Kit for ∂o pj ∂ w o ∂C(t ,o ) ∂net ∂C(t ,o ) ∑ ki pi Neural Net Forecasting pj pj pk = pj pj i ∑ ∂net ∂o ∑ ∂net ∂o k pk pj k pk pj Æ 20+ software simulators ∂C(t pj ,opj ) = ∑ wkj = − ∑δ pj wkj k ∂net pk k Æ datasets ' δ pj = f j (net pj )∑δ pj wkj Æ literature & faq k ⎧∂C(t pj ,opj ) ' ⎪ f j (net pj ) if unit j is in the output layer ⎪ ∂opj Æ slides, data & additional info on δ pj = ⎨ ⎪ ' f j (net pj )∑δ pk wpjk if unit j is in a hidden layer www.neural-forecasting.com ⎩⎪ k EVIC’05 © Sven F.
    [Show full text]
  • Rational Function Approximation
    Approximation Theory Pad´eApproximation Numerical Analysis and Computing Lecture Notes #13 — Approximation Theory — Rational Function Approximation Joe Mahaffy, [email protected] Department of Mathematics Dynamical Systems Group Computational Sciences Research Center San Diego State University San Diego, CA 92182-7720 http://www-rohan.sdsu.edu/∼jmahaffy Spring 2010 Joe Mahaffy, [email protected] Rational Function Approximation — (1/21) Approximation Theory Pad´eApproximation Outline 1 Approximation Theory Pros and Cons of Polynomial Approximation New Bag-of-Tricks: Rational Approximation Pad´eApproximation: Example #1 2 Pad´eApproximation Example #2 Finding the Optimal Pad´eApproximation Joe Mahaffy, [email protected] Rational Function Approximation — (2/21) Pros and Cons of Polynomial Approximation Approximation Theory New Bag-of-Tricks: Rational Approximation Pad´eApproximation Pad´eApproximation: Example #1 Polynomial Approximation: Pros and Cons. Advantages of Polynomial Approximation: [1] We can approximate any continuous function on a closed inter- val to within arbitrary tolerance. (Weierstrass approximation theorem) Joe Mahaffy, [email protected] Rational Function Approximation — (3/21) Pros and Cons of Polynomial Approximation Approximation Theory New Bag-of-Tricks: Rational Approximation Pad´eApproximation Pad´eApproximation: Example #1 Polynomial Approximation: Pros and Cons. Advantages of Polynomial Approximation: [1] We can approximate any continuous function on a closed inter- val to within arbitrary tolerance. (Weierstrass approximation theorem) [2] Easily evaluated at arbitrary values. (e.g. Horner’s method) Joe Mahaffy, [email protected] Rational Function Approximation — (3/21) Pros and Cons of Polynomial Approximation Approximation Theory New Bag-of-Tricks: Rational Approximation Pad´eApproximation Pad´eApproximation: Example #1 Polynomial Approximation: Pros and Cons. Advantages of Polynomial Approximation: [1] We can approximate any continuous function on a closed inter- val to within arbitrary tolerance.
    [Show full text]