Best L1 Approximation of Heaviside-Type Functions in a Weak-Chebyshev Space

Total Page:16

File Type:pdf, Size:1020Kb

Best L1 Approximation of Heaviside-Type Functions in a Weak-Chebyshev Space Best L1 approximation of Heaviside-type functions in Chebyshev and weak-Chebyshev spaces Laurent Gajny,1 Olivier Gibaru1,,2 Eric Nyiri1, Shu-Cherng Fang3 Abstract In this article, we study the problem of best L1 approximation of Heaviside-type functions in Chebyshev and weak-Chebyshev spaces. We extend the Hobby-Rice theorem [2] into an appro- priate framework and prove the unicity of best L1 approximation of Heaviside-type functions in an even-dimensional Chebyshev space under the condition that the dimension of the subspace composed of the even functions is half the dimension of the whole space. We also apply the results to compute best L1 approximations of Heaviside-type functions by polynomials and Her- mite polynomial splines with fixed knots. Keywords : Best approximation, L1 norm, Heaviside function, polynomials, polynomial splines, Chebyshev space, weak-Chebyshev space. Introduction Let [a, b] be a real interval with a < b and ν be a positive measure defined on a σ-field of subsets of [a, b]. The space L1([a, b],ν) of ν-integrable functions with the so-called L1 norm : b k · k1 : f ∈ L1([a, b],ν) 7−→ kfk1 = |f(x)| dν(x) Za forms a Banach space. Let f ∈ L1([a, b],ν) and Y be a subset of L1([a, b],ν) such that f∈ / Y where Y is the closure of Y . The problem of the best L1 approximation of the function f in Y consists in finding a function g∗ ∈ Y such that : ∗ arXiv:1407.0228v3 [math.FA] 26 Oct 2015 kf − g k1 ≤ kf − gk1, ∀g ∈ Y. (1) If Y is a compact set or a finite dimensional subspace of L1([a, b],ν), then such a function exists for any f ∈ L1([a, b],ν). In this article, we focus on the finite-dimensional case. An important characterization theorem when ν is a non-atomic measure (See for example [3, 7] and a development in [15]) claims that finding a best L1 approximation of a function f remains to define a sign-changes function s (i.e. a function for which values alternate between -1 and 1 at certain abscissae) such that : b s(x)g(x) dν(x) = 0, ∀g ∈ Y. Za 1Arts et M´etiers ParisTech, LSIS - UMR CNRS 7296, 8 Boulevard Louis XIV, 59046 Lille Cedex, France 2INRIA Lille-Nord-Europe, NON-A research team, 40, avenue Halley 59650 Villeneuve d’Ascq 3North Carolina State University, Edward P. Fitts Department of Industrial and Systems Engineering, NC 27695- 7906, United States of America. 1 The Hobby-Rice theorem [2] enables to show the existence of such functions. This major result helps characterize and compute solutions of best L1 approximation of a continuous function in Chebyshev and weak-Chebyshev spaces. It has been widely studied in the litterature [4, 8, 12, 16, 20, 21]. We recall that an n-dimensional Chebyshev space (resp. weak-Chebyshev space) on [a, b] is a subspace of C0[a, b] composed of functions which have at most n − 1 distinct zeros on [a, b] (resp. strong changes of sign) [5, 19]. Classical examples of such spaces are respectively polynomial functions and polynomial spline functions with fixed knots. To the best of our knowledge, there are few references in the litterature that deal with the challenging problem of best L1 approximation of discontinuous functions, especially Heaviside-type functions. Definition 1. We say that f is a Heaviside-type function on [a, b] with a jump at δ ∈ (a, b) if f ∈ C0([a, b]\{δ}) and if the limits of the function |f| on both sides of δ exist and are finite. We denote by Jδ[a, b] the vector space of such functions. Existing papers focusing on one or two-sided best L1 approximation deal with specific spaces. Bustamante et al. have shown that best polynomial one-sided L1 approximation of the Heaviside function can be obtained by Hermite interpolation at zeros of some Jacobi polynomials [1]. Moskona et al. have studied the problem of best two-sided Lp (p ≥ 1) approximation of the Heaviside function using trigonometrical polynomials [9]. Saff and Tashev have done a similar work using polygonal lines [18]. We propose in this article a more general framework for best (two-sided) L1 approximation of Heaviside-type functions in Chebyshev and weak-Chebyshev spaces which includes the two previous cases. This problem is presented in Section 1. We evidence the encountered difficulties that lead us to extend the classical Hobby-Rice theorem. This extension is presented in details in Section 2. In the third section, we strongly use this result to characterize best L1 approximations of Heaviside-type functions in a Chebyshev space. We give sufficient conditions on the Chebyshev space to obtain a unique L1 best approximation for every Heaviside-type function. We apply these results to polynomial approximation. In Section 4, we study the problem of best L1 approximation of Heaviside-type functions in a weak-Chebyshev space. In particular, this theory is applied to the Hermite polynomial spline case in Section 4. 1 Best L1 approximation In this section, we recall an important characterization theorem of best L1 approximation (See for example [3, 7]) and the Hobby-Rice theorem. We evidence the reason why the Hobby-Rice theorem fails to give a more precise characterization of best L1 approximations of Heaviside-type functions. For convenience, we define the zero set Z(f)= {x ∈ [a, b], f(x) = 0} and the sign function : 1 if x> 0, sign : x ∈ R 7−→ sign(x)= 0 if x = 0, −1 if x< 0. ∗ Theorem 1.1. Let Y be a subspace of L1([a, b],ν) and f ∈ L1([a, b],ν)\Y . Then g is a best L1 approximation of f in Y if and only if : b sign(f(x) − g∗(x))g(x) dν(x) ≤ |g(x)| dν(x), for all g ∈ Y. (2) Z Z ∗ a Z(f−g ) ∗ ∗ In particular, if ν(Z(f − g )) = 0, then g is a best L1 approximation of f in Y if and only if : b sign(f(x) − g∗(x))g(x) dν(x) = 0, (3) Za 2 for all g ∈ Y . The second part of the latter theorem claims that finding a best L1 approximation of a function f may remain to find a sign-changes function s, that alternates between −1 and 1 at certain abscissae, such that: b s(x)g(x) dν(x) = 0, ∀g ∈ Y. (4) Za If there exists a function g∗ ∈ Y which interpolates f at these abscissae and no others, then g∗ is a best L1 approximation of f in Y . Thus, the non-linear problem of best L1 approximation remains to a simple Lagrange interpolation problem. Actually, based on a result by Phelps (Lemma 2 in [13]), it can be shown that if ν is a non-atomic ∗ measure then g is a best L1 approximation of f in Y if and only if there exists a sign changes function satisfying (4) and which coincides ν-almost everywhere with sgn(f −g∗) on [a, b]\Z(f −g∗) (see also Theorem 2.3 in [15]). From now, we consider in this article that we are in this case. Thus, either if a best L1 approximation coincides or not with its reference function on a set of non-zero measure, the determination of a sign-changes function verifying (4) is fundamental. Indeed, it characterizes a best L1 approximation in both cases. The Hobby-Rice theorem [2] enables to show the existence of such a sign-changes function with the number of sign-changes lower or equal to the dimension of Y . Theorem 1.2 (Hobby, Rice, 1965). Let Y be an n-dimensional subspace of L1([a, b],ν) where ν is a finite, non-atomic measure. Then there exists a sign-changes function : r+1 i s(x)= (−1) 1[αi−1,αi[(x), (5) Xi=1 such that for all g ∈ Y : b s(x)g(x) dν(x) = 0, (6) Za where r ≤ n and : a = α0 < α1 < α2 < · · · < αr < αr+1 = b. Remark. A simple proof of this theorem can be found in [14] using the Borsuk antipodality theorem (See for example [10], page 21). One cannot guarantee in general that a sign function with r ≤ n changes of sign as stated in the Hobby-Rice theorem enables us to define a function g∗ that coincides with the function to approximate f at these abscissae and no others. In Figure 1, we can observe a best L1 approximation of a continuous function, x 7→ sin(3x), in a 4-dimensional space - the space of cubic functions on the interval [−1, 1] - having five intersections with the above function. The following corollary partially explains this observation. Corollary 1. Let Y be an n-dimensional subspace of L1([a, b],ν) where ν is a finite, non-atomic measure and m be an integer such that m ≥ n. Then there exist a sign-changes function : rm+1 i s(x)= (−1) 1[αi−1,αi[(x), (7) Xi=1 3 1.5 1 0.5 0 -0.5 -1 -1.5α α α α α α α 0 1 2 3 4 5 6 Figure 1: Best L1 cubic approximation (solid line) of x 7→ sin(3x) (dashed line) on [−1, 1]. such that for all g ∈ Y : b s(x)g(x) dν(x) = 0, (8) Za where rm ≤ m : a = α0 < α1 < α2 < · · · < αrm < αrm+1 = b. Proof. Complete Y in a m-dimensional subspace of L1([a, b],ν).
Recommended publications
  • Rational Approximation of the Absolute Value Function from Measurements: a Numerical Study of Recent Methods
    Rational approximation of the absolute value function from measurements: a numerical study of recent methods I.V. Gosea∗ and A.C. Antoulasy May 7, 2020 Abstract In this work, we propose an extensive numerical study on approximating the absolute value function. The methods presented in this paper compute approximants in the form of rational functions and have been proposed relatively recently, e.g., the Loewner framework and the AAA algorithm. Moreover, these methods are based on data, i.e., measurements of the original function (hence data-driven nature). We compare numerical results for these two methods with those of a recently-proposed best rational approximation method (the minimax algorithm) and with various classical bounds from the previous century. Finally, we propose an iterative extension of the Loewner framework that can be used to increase the approximation quality of the rational approximant. 1 Introduction Approximation theory is a well-established field of mathematics that splits in a wide variety of subfields and has a multitude of applications in many areas of applied sciences. Some examples of the latter include computational fluid dynamics, solution and control of partial differential equations, data compression, electronic structure calculation, systems and control theory (model order reduction reduction of dynamical systems). In this work we are concerned with rational approximation, i.e, approximation of given functions by means of rational functions. More precisely, the original to be approximated function is the absolute value function (known also as the modulus function). Moreover, the methods under consideration do not necessarily require access to the exact closed-form of the original function, but only to evaluations of it on a particular domain.
    [Show full text]
  • Techniques of Variational Analysis
    J. M. Borwein and Q. J. Zhu Techniques of Variational Analysis An Introduction October 8, 2004 Springer Berlin Heidelberg NewYork Hong Kong London Milan Paris Tokyo To Tova, Naomi, Rachel and Judith. To Charles and Lilly. And in fond and respectful memory of Simon Fitzpatrick (1953-2004). Preface Variational arguments are classical techniques whose use can be traced back to the early development of the calculus of variations and further. Rooted in the physical principle of least action they have wide applications in diverse ¯elds. The discovery of modern variational principles and nonsmooth analysis further expand the range of applications of these techniques. The motivation to write this book came from a desire to share our pleasure in applying such variational techniques and promoting these powerful tools. Potential readers of this book will be researchers and graduate students who might bene¯t from using variational methods. The only broad prerequisite we anticipate is a working knowledge of un- dergraduate analysis and of the basic principles of functional analysis (e.g., those encountered in a typical introductory functional analysis course). We hope to attract researchers from diverse areas { who may fruitfully use varia- tional techniques { by providing them with a relatively systematical account of the principles of variational analysis. We also hope to give further insight to graduate students whose research already concentrates on variational analysis. Keeping these two di®erent reader groups in mind we arrange the material into relatively independent blocks. We discuss various forms of variational princi- ples early in Chapter 2. We then discuss applications of variational techniques in di®erent areas in Chapters 3{7.
    [Show full text]
  • HP 33S Solving Trigonometry Problems the Trigonometric Functions Trigonometric Modes Practice Working Problems Involving Trig Functions
    hp calculators HP 33S Solving Trigonometry Problems The trigonometric functions Trigonometric modes Practice working problems involving trig functions hp calculators HP 33S Solving Trigonometry Problems The trigonometric functions The trigonometric functions, sine, cosine, tangent, and related functions, are used in geometry, surveying, and design. They also occur in solutions to orbital mechanics, integration, and other advanced applications. The HP 33S provides the three basic functions, and their inverse, or “arc” functions. These work in degrees, radians and gradians modes. In addition, π is provided as a function on the right-shifted “cos” key, and the sign function is on the right-shifted “sin” key. The secant, cosecant and cotangent functions are easily calculated using the n, k, and q keys respectively, followed by ,. To help remember whether the secant function corresponds to the inverse sine or cosine, it can be helpful to note that the first letters of “secant” and “cosecant” are inverted in relation to those of “sine” and “cosine”, just as the secant and cosecant are the inverted cosine and sine functions. The two keys ° and ± also use trigonometric functions to convert between rectangular and radial coordinates. They can therefore be useful in some trigonometric calculations. Trigonometric modes The HP 33S can calculate trigonometric functions in any of these three modes: Degrees, Radians or Gradians. Practice working problems involving trig functions Example 1: Select the appropriate angle mode. Solution: Press the Ý key below the screen. Figure 1 Press 1, 2 or 3 to select DEGrees, RADians or GRADians mode, or use the arrow keys Ö, Õ, × and Ø to select the required mode and then press Ï.
    [Show full text]
  • Notes on Distributions
    Notes on Distributions Peter Woit Department of Mathematics, Columbia University [email protected] April 26, 2020 1 Introduction This is a set of notes written for the 2020 spring semester Fourier Analysis class, covering material on distributions. This is an important topic not covered in Stein-Shakarchi. These notes will supplement two textbooks you can consult for more details: A Guide to Distribution Theory and Fourier Transforms [2], by Robert Strichartz. The discussion of distributions in this book is quite compre- hensive, and at roughly the same level of rigor as this course. Much of the motivating material comes from physics. Lectures on the Fourier Transform and its Applications [1], by Brad Os- good, chapters 4 and 5. This book is wordier than Strichartz, has a wealth of pictures and motivational examples from the field of signal processing. The first three chapters provide an excellent review of the material we have covered so far, in a form more accessible than Stein-Shakarchi. One should think of distributions as mathematical objects generalizing the notion of a function (and the term \generalized function" is often used inter- changeably with \distribution"). A function will give one a distribution, but many important examples of distributions do not come from a function in this way. Some of the properties of distributions that make them highly useful are: Distributions are always infinitely differentiable, one never needs to worry about whether derivatives exist. One can look for solutions of differential equations that are distributions, and for many examples of differential equations the simplest solutions are given by distributions.
    [Show full text]
  • Three-Complex Numbers and Related Algebraic Structures
    S S symmetry Article Three-Complex Numbers and Related Algebraic Structures Wolf-Dieter Richter School Information, University of Rostock, Ulmenstraße 69, Haus 3, 18057 Rostock, Germany; [email protected] Abstract: Three-complex numbers are introduced for using a geometric vector product in the three- dimensional Euclidean vector space R3 and proving its equivalence with a spherical coordinate product. Based upon the definitions of the geometric power and geometric exponential functions, some Euler-type trigonometric representations of three-complex numbers are derived. Further, a 3 general l2 −complex algebraic structure together with its matrix, polynomial and variable basis vector 3 representations are considered. Then, the classes of lp-complex numbers are introduced. As an application, Euler-type formulas are used to construct directional probability laws on the Euclidean unit sphere in R3. Keywords: geometric vector product; geometric vector power; geometric exponential function; three-complex algebraic structure; matrix, polynomial and variable basis representations; Euler-type 3 representations; lp-complex numbers; construction of directional probability laws 1. Introduction While a crucial property of complex numbers x + iy is very briefly described by Citation: Richter, W.-D. the equation i2 = −1, the numbers x + iy + jz considered in the present work have the Three-Complex Numbers and fundamental property i2 = j2 = ij = −1. The basic aid for establishing this property Related Algebraic Structures. consists in the use of suitable coordinates and the appropriate handling of rotations on Symmetry 2021, 13, 342. https:// spheres in the space R3. Such rotations are often described in terms of three angles, of doi.org/10.3390/sym13020342 which the best known by name are probably the Euler angles.
    [Show full text]
  • Maple 7 Programming Guide
    Maple 7 Programming Guide M. B. Monagan K. O. Geddes K. M. Heal G. Labahn S. M. Vorkoetter J. McCarron P. DeMarco c 2001 by Waterloo Maple Inc. ii • Waterloo Maple Inc. 57 Erb Street West Waterloo, ON N2L 6C2 Canada Maple and Maple V are registered trademarks of Waterloo Maple Inc. c 2001, 2000, 1998, 1996 by Waterloo Maple Inc. All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the copyright holder, except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use of general descriptive names, trade names, trademarks, etc., in this publication, even if the former are not especially identified, is not to be taken as a sign that such names, as understood by the Trade Marks and Merchandise Marks Act, may accordingly be used freely by anyone. Contents 1 Introduction 1 1.1 Getting Started ........................ 2 Locals and Globals ...................... 7 Inputs, Parameters, Arguments ............... 8 1.2 Basic Programming Constructs ............... 11 The Assignment Statement ................. 11 The for Loop......................... 13 The Conditional Statement ................. 15 The while Loop....................... 19 Modularization ........................ 20 Recursive Procedures ..................... 22 Exercise ............................ 24 1.3 Basic Data Structures .................... 25 Exercise ............................ 26 Exercise ............................ 28 A MEMBER Procedure ..................... 28 Exercise ............................ 29 Binary Search ......................... 29 Exercises ........................... 30 Plotting the Roots of a Polynomial ............. 31 1.4 Computing with Formulæ .................. 33 The Height of a Polynomial ................. 34 Exercise ...........................
    [Show full text]
  • Nested Formulas for Cosine and Inverse Cosine Functions Based on Vi`Ete’Sformula for Π
    Nested formulas for cosine and inverse cosine functions based on Vi`ete'sformula for π Artur Kawalec Abstract In this article, we develop nested representations for cosine and inverse cosine functions, which is a generalization of Vi`ete'sformula for π. We explore a natural inverse relationship between these representations and develop numerical algorithms to compute them. Throughout this article, we perform numerical computation for various test cases, and demonstrate that these nested formulas are valid for complex arguments and a kth branch. We further extend the presented results to hyperbolic cosine and logarithm functions, and using additional trigonometric identities, we explore the sine and tangent functions and their inverses. 1 Introduction In 1593s, Vi`etedeveloped an infinite nested radical product formula for π given as p p p q p p 2 2 2 + 2 2 + 2 + 2 = ··· ; (1) π 2 2 2 which, historically, may have been the first infinite series for π. Vi`etederived this formula geometrically by means of approximating areas of an n-sided polygon inscribed in a circle, and obtaining a successive sequence of ratios of areas of 2n-sided polygon to an n-sided polygon that is converging to an area of a circle, from which 2/π follows as a converging value as n ! 1 [2]. Subsequently, Euler also developed a new proof of Vi`ete'sformula using trigonometric identities and infinite products, namely, Euler found a product identity arXiv:2007.08639v1 [math.GM] 15 Jul 2020 1 sin(x) Y x x x x = cos = cos cos cos :::; (2) x 2n 2 22 23 n=1 where by a repeated application of the half-angle identity for cosine r x 1 − cos(x) cos = ± (3) 2 2 the Vi`ete'sformula for π is obtained by substituting x = π=2.
    [Show full text]
  • IBM ILOG CPLEX Optimization Studio OPL Language Reference Manual
    IBM IBM ILOG CPLEX Optimization Studio OPL Language Reference Manual Version 12 Release 7 Copyright notice Describes general use restrictions and trademarks related to this document and the software described in this document. © Copyright IBM Corp. 1987, 2017 US Government Users Restricted Rights - Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Trademarks IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines Corp., registered in many jurisdictions worldwide. Other product and service names might be trademarks of IBM or other companies. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at www.ibm.com/legal/copytrade.shtml. Adobe, the Adobe logo, PostScript, and the PostScript logo are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States, and/or other countries. Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both. UNIX is a registered trademark of The Open Group in the United States and other countries. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Java and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its affiliates. Other company, product, or service names may be trademarks or service marks of others. © Copyright IBM Corporation 1987, 2017. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Figures ............... v Limitations on constraints ........ 59 Formal parameters ...........
    [Show full text]
  • Lecture 9 Connecting Laplace and Fourier Transforms
    Signals and Systems Lecture 9 DR TANIA STATHAKI READER (ASSOCIATE PROFESSOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON Connecting Laplace and Fourier transform • In this lecture we will start connecting the Laplace and the Fourier transform. • Recall that when we studied the Laplace transform we mentioned the concept of the Region Of Convergence (ROC) of the transform. • This concept did not exist in the Fourier transform. Example: Laplace transform of a causal exponential function • Recall the Laplace transform of the causal exponential function 푥 푡 = 푒−푎푡푢(푡). Note that from this point forward I will interchange 풆−풂풕풖(풕) with 풆풂풕풖(풕). ∞ ∞ 푋 푠 = ℒ 푥 푡 = 푒−푎푡푢(푡)푒−푠푡푑푡 = 푒−(푠+푎)푡푑푡 −∞ 0 1 ∞ −1 −1 1 = 푒−(푠+푎)푡 = [ 푒− 푠+푎 ∙∞ − 푒− 푠+푎 ∙0 ] = (0 − 1) = . −(푠+푎) 0 (푠+푎) 푠+푎 푠+푎 . Note that in order to have 푒−(푠+푎)∙∞ = 0, the real part of 푠 + 푎 must be positive, i.e., Re 푠 + 푎 > 0 ⇒ Re 푠 > −Re 푎 = Re −푎 . [ROC is shown in the figure below right with the shaded area.] . Note that the value s = −푎 makes the denominator of the Laplace transform 0. 퐑퐞{−풂} . Note that for a decaying exponential we want 푎 > 0. Example: Fourier transform of a causal exponential function • Let us find now the Fourier transform of the causal exponential function: 푥 푡 = 푒−푎푡푢(푡). ∞ ∞ 푋 휔 = ℱ 푥 푡 = 푒−푎푡푢(푡)푒−푗휔푡푑푡 = 푒−(푗휔+푎)푡푑푡 −∞ 0 1 ∞ −1 = 푒−(푗휔+푎)푡 = 푒− 푗휔+푎 ∙∞ − 푒− 푗휔+푎 ∙0 −(푗휔 + 푎) 0 (푗휔 + 푎) −1 1 = (0 − 1) = .
    [Show full text]
  • Introduction to Image Processing #5/7
    Outline Introduction to Image Processing #5/7 Thierry Geraud´ EPITA Research and Development Laboratory (LRDE) 2006 Thierry Geraud´ Introduction to Image Processing #5/7 Outline Outline 1 Introduction 2 Distributions About the Dirac Delta Function Some Useful Functions and Distributions 3 Fourier and Convolution 4 Sampling 5 Convolution and Linear Filtering 6 Some 2D Linear Filters Gradients Laplacian Thierry Geraud´ Introduction to Image Processing #5/7 Outline Outline 1 Introduction 2 Distributions About the Dirac Delta Function Some Useful Functions and Distributions 3 Fourier and Convolution 4 Sampling 5 Convolution and Linear Filtering 6 Some 2D Linear Filters Gradients Laplacian Thierry Geraud´ Introduction to Image Processing #5/7 Outline Outline 1 Introduction 2 Distributions About the Dirac Delta Function Some Useful Functions and Distributions 3 Fourier and Convolution 4 Sampling 5 Convolution and Linear Filtering 6 Some 2D Linear Filters Gradients Laplacian Thierry Geraud´ Introduction to Image Processing #5/7 Outline Outline 1 Introduction 2 Distributions About the Dirac Delta Function Some Useful Functions and Distributions 3 Fourier and Convolution 4 Sampling 5 Convolution and Linear Filtering 6 Some 2D Linear Filters Gradients Laplacian Thierry Geraud´ Introduction to Image Processing #5/7 Outline Outline 1 Introduction 2 Distributions About the Dirac Delta Function Some Useful Functions and Distributions 3 Fourier and Convolution 4 Sampling 5 Convolution and Linear Filtering 6 Some 2D Linear Filters Gradients Laplacian
    [Show full text]
  • Hausdorff Approximation of the Sign Function by a Class of Parametric
    Article Biomath Communications 3 (2016) Biomath Communications www.biomathforum.org/biomath/index.php/conference Hausdorff approximation of the sign function by a class of parametric activation functions Nikolay Kyurkchiev, Svetoslav Markov Institute of Mathematics and Informatics, Bulgarian Academy of Sciences Sofia, Bulgaria Abstract In this paper we study the distance between the sign function and a class of parametric activation functions. The distance is measured in Hausdorff sense, which is natural in a situation when a sign function is involved. Precise upper and lower bounds for the Hausdorff distance have been obtained. Key words: Parametric Algebraic Activation function (PAA), Parametric Hyperbolic Tangent Activation function (PHTA), sign func- tion, Hausdorff distance 2010 Mathematics Subject Classification: 41A46, 62P10 Introduction Sigmoidal functions (also known as \activation functions") find multiple applications to neural networks [1],[4]{[8]. We study the distance between the sign function and a special class of sigmoidal Citation: Nikolay Kyurkchiev, Svetoslav Markov, Hausdorff approximation of the sign function by a class of parametric activation functions, http://dx.doi.org/10.11145/bmc.2016.12.217 functions, so-called parametric activation functions. The distance is measured in Hausdorff sense, which is natural in a situation when a sign function is involved. Precise upper and lower bounds for the Hausdorff distance are reported. Parametric Algebraic Activation (PAA) function The following are common examples of activation functions [1]: 1 σ(t) = (the squashing function (logistic)) 1 + e−t 8 0; if t ≤ −1; > <t+1 σ(t) = 2 ; if−1 ≤ t ≤ 1; (the piecewise linear(cut; ramp)Lfunction) :> 1; if t ≥ 1: 1 1 σ(t) = + arctan(t) (the arctan sigmoidal function): 2 π Elliott activation function [2] is defined as x f(x) = : 1 + jxj The Parametric Algebraic Activation (PAA) function [3] is given by x(1 + ajxj) f (x) = ; x 2 ; a ≥ 0: (1) a 1 + jxj(1 + ajxj) R For a = 0 we obtain the Elliott activation function.
    [Show full text]
  • Real Functions
    Real Functions September 25, 2011 1 Introduction A function over D, being D is a rule that for every number x of D ⊂ ℜ associates only one number of . It is represented in the form f : D . ℜ → ℜ y is called image of x and it is written as y = f(x). It is understood by domain of the function f the set: Domf = x : y with y = f(x) { ∈ℜ ∃ ∈ℜ } While the image of the function is defined as the set: Imf = y : x with y = f(x) { ∈ℜ ∃ ∈ℜ } The graphic of the function f is defined as the subset of 2: ℜ G(f)= (x,y) 2 : y = f(x) { ∈ℜ } 2 Operations with functions Given two functions f, g : , we define the following operations: ℜ←ℜ sum of two functions: • (f + g)(x)= f(x)+ g(x) The domain of f + g is the intersection of the domains Dom f Dom g. ∩ product of two functions: • (f g)(x)= f(x) g(x) · · The domain of f + g is the intersection of the domains Dom f Dom g. ∩ Composition of two functions: • (g f)(x)= g(f(x)) ◦ This function exists for the x so that f(x) Dom g, for every x ∈ ℜ ∈ ∈ ℜ so that Imf Dom g. ⊂ 1 2.1 Kinds of functions Given a function f : , it is said that it is an even function when it • ℜ→ℜ verifies that: f( x)= f(x), x Domf − ∀ ∈ . This function would present symmetry with respect to OY axis. and it is odd when it verifies that f( x)= f(x), x Domf − − ∀ ∈ .
    [Show full text]