TOPIC 0 Introduction

Total Page:16

File Type:pdf, Size:1020Kb

TOPIC 0 Introduction TOPIC 0 Introduction 1 Review Course: Math 535 (http://www.math.wisc.edu/~roch/mmids/) - Mathematical Methods in Data Science (MMiDS) Author: Sebastien Roch (http://www.math.wisc.edu/~roch/), Department of Mathematics, University of Wisconsin-Madison Updated: Sep 21, 2020 Copyright: © 2020 Sebastien Roch We first review a few basic concepts. 1.1 Vectors and norms For a vector ⎡ 푥 ⎤ ⎢ 1 ⎥ 푥 ⎢ 2 ⎥ 푑 퐱 = ⎢ ⎥ ∈ ℝ ⎢ ⋮ ⎥ ⎣ 푥푑 ⎦ the Euclidean norm of 퐱 is defined as ‾‾푑‾‾‾‾ ‖퐱‖ = 푥2 = √퐱‾‾푇‾퐱‾ = ‾⟨퐱‾‾, ‾퐱‾⟩ 2 ∑ 푖 √ ⎷푖=1 푇 where 퐱 denotes the transpose of 퐱 (seen as a single-column matrix) and 푑 ⟨퐮, 퐯⟩ = 푢 푣 ∑ 푖 푖 푖=1 is the inner product (https://en.wikipedia.org/wiki/Inner_product_space) of 퐮 and 퐯. This is also known as the 2- norm. More generally, for 푝 ≥ 1, the 푝-norm (https://en.wikipedia.org/wiki/Lp_space#The_p- norm_in_countably_infinite_dimensions_and_ℓ_p_spaces) of 퐱 is given by 푑 1/푝 ‖퐱‖ = |푥 |푝 . 푝 (∑ 푖 ) 푖=1 Here (https://commons.wikimedia.org/wiki/File:Lp_space_animation.gif#/media/File:Lp_space_animation.gif) is a nice visualization of the unit ball, that is, the set {퐱 : ‖푥‖푝 ≤ 1}, under varying 푝. There exist many more norms. Formally: 푑 푑 Definition (Norm): A norm is a function ℓ from ℝ to ℝ+ that satisfies for all 푎 ∈ ℝ, 퐮, 퐯 ∈ ℝ (Homogeneity): ℓ(푎퐮) = |푎|ℓ(퐮) (Triangle inequality): ℓ(퐮 + 퐯) ≤ ℓ(퐮) + ℓ(퐯) (Point-separating): ℓ(푢) = 0 implies 퐮 = 0. ⊲ The triangle inequality for the 2-norm follows (https://en.wikipedia.org/wiki/Cauchy– Schwarz_inequality#Analysis) from the Cauchy–Schwarz inequality (https://en.wikipedia.org/wiki/Cauchy– Schwarz_inequality). 푑 Theorem (Cauchy–Schwarz): For all 퐮, 퐯 ∈ ℝ |⟨퐮, 퐯⟩| ≤ ‖퐮‖2‖퐯‖2. 푑 The Euclidean distance (https://en.wikipedia.org/wiki/Euclidean_distance) between two vectors 퐮 and 퐯 in ℝ is the 2-norm of their difference 푑(퐮, 퐯) = ‖퐮 − 퐯‖2. Throughout we use the notation ‖퐱‖ = ‖퐱‖2 to indicate the 2-norm of 퐱 unless specified otherwise. 푑 We will often work with collections of 푛 vectors 퐱1 , … , 퐱푛 in ℝ and it will be convenient to stack them up into a matrix ⎡ 퐱푇 ⎤ ⎡ 푥 푥 ⋯ 푥 ⎤ ⎢ 1 ⎥ 11 12 1푑 푇 ⎢ ⎥ ⎢ 퐱2 ⎥ ⎢ 푥21 푥22 ⋯ 푥2푑 ⎥ 푋 = ⎢ ⎥ = ⎢ ⎥ , ⎢ ⋮ ⎥ ⎢ ⋮ ⋮ ⋱ ⋮ ⎥ 푇 ⎣ ⎦ ⎣ 퐱푛 ⎦ 푥푛1 푥푛2 ⋯ 푥푛푑 where 푇 indicates the transpose (https://en.wikipedia.org/wiki/Transpose): (Source (https://commons.wikimedia.org/wiki/File:Matrix_transpose.gif)) NUMERICAL CORNER In Julia, a vector can be obtained in different ways. The following method gives a row vector as a two-dimensional array. In [1]: # Julia version: 1.5.1 using LinearAlgebra, Statistics, Plots In [2]: t = [1. 3. 5.] Out[2]: 1×3 Array{Float64,2}: 1.0 3.0 5.0 To turn it into a one-dimensional array, use vec (https://docs.julialang.org/en/v1/base/arrays/#Base.vec). In [3]: vec(t) Out[3]: 3-element Array{Float64,1}: 1.0 3.0 5.0 To construct a one-dimensional array directly, use commas to separate the entries. In [4]: u = [1., 3., 5., 7.] Out[4]: 4-element Array{Float64,1}: 1.0 3.0 5.0 7.0 To obtain the norm of a vector, we can use the function norm (https://docs.julialang.org/en/v1/stdlib/LinearAlgebra/#LinearAlgebra.norm) (which requires the LinearAlgebra package): In [5]: norm(u) Out[5]: 9.16515138991168 which we can check "by hand" In [6]: sqrt(sum(u.^2)) Out[6]: 9.16515138991168 The . above is called broadcasting (https://docs.julialang.org/en/v1.2/manual/mathematical-operations/#man- dot-operators-1). It applies the operator following it (in this case taking a square) element-wise. Exercise: Compute the inner product of 푢 = (1, 2, 3, 4) and 푣 = (5, 4, 3, 2) without using the function dot (https://docs.julialang.org/en/v1/stdlib/LinearAlgebra/#LinearAlgebra.dot). Hint: The product of two real numbers 푎 and 푏 is a * b . In [7]: # Try it! u = [1., 2., 3., 4.]; # EDIT THIS LINE: define v # EDIT THIS LINE: compute the inner product between u and v ⊲ To create a matrix out of two vectors, we use the function hcat (https://docs.julialang.org/en/v1.2/base/arrays/#Base.hcat) and transpose. In [8]: u = [1., 3., 5., 7.]; v = [2., 4., 6., 8.]; X = hcat(u,v)' Out[8]: 2×4 Adjoint{Float64,Array{Float64,2}}: 1.0 3.0 5.0 7.0 2.0 4.0 6.0 8.0 With more than two vectors, we can use the reduce (https://docs.julialang.org/en/v1/base/collections/#Base.reduce-Tuple{Any,Any}) function. In [9]: u = [1., 3., 5., 7.]; v = [2., 4., 6., 8.]; w = [9., 8., 7., 6.]; X = reduce(hcat, [u, v, w])' Out[9]: 3×4 Adjoint{Float64,Array{Float64,2}}: 1.0 3.0 5.0 7.0 2.0 4.0 6.0 8.0 9.0 8.0 7.0 6.0 1.2 Multivariable calculus 1.2.1 Limits and continuity ‾‾‾푑‾‾‾‾‾2 푇 푑 Throughout this section, we use the Euclidean norm ‖퐱‖ = √∑푖=1 푥푖 for 퐱 = (푥1, … , 푥푑 ) ∈ ℝ . 푑 The open 푟-ball around 퐱 ∈ ℝ is the set of points within Euclidean distance 푟 of 퐱, that is, 푑 퐵푟(퐱) = {퐲 ∈ ℝ : ‖퐲 − 퐱‖ < 푟}. 푑 푑 A point 퐱 ∈ ℝ is a limit point (or accumulation point) of a set 퐴 ⊆ ℝ if every open ball around 퐱 contains an element 퐚 of 퐴 such that 퐚 ≠ 퐱. A set 퐴 is closed if every limit point of 퐴 belongs to 퐴. (Source (https://www.math212.com/2017/12/limit-point-closure.html)) 푑 푑 A point 퐱 ∈ ℝ is an interior point of a set 퐴 ⊆ ℝ if there exists an 푟 > 0 such that 퐵푟(퐱) ⊆ 퐴. A set 퐴 is open if it consists entirely of interior points. (Source (https://commons.wikimedia.org/wiki/File:Open_set_-_example.png)) 푑 푇 A set 퐴 ⊆ ℝ is bounded if there exists an 푟 > 0 such that 퐴 ⊆ 퐵푟(0), where 0 = (0, … , 0) . 푑 Definition (Limits of a Function): Let 푓 : 퐷 → ℝ be a real-valued function on 퐷 ⊆ ℝ . Then 푓 is said to have a limit 퐿 ∈ ℝ as 퐱 approaches 퐚 if: for any 휖 > 0, there exists a 훿 > 0 such that |푓(퐱) − 퐿| < 휖 for all 퐱 ∈ 퐷 ∩ 퐵훿(퐚) ∖ {퐚}. This is written as lim 푓(퐱) = 퐿. 퐱→퐚 ⊲ Note that we explicitly exclude 퐚 itself from having to satisfy the condition |푓(퐱) − 퐿| < 휖. In particular, we may have 푓(퐚) ≠ 퐿. We also do not restrict 퐚 to be in 퐷. 푑 Definition (Continuous Function): Let 푓 : 퐷 → ℝ be a real-valued function on 퐷 ⊆ ℝ . Then 푓 is said to be continuous at 퐚 ∈ 퐷 if lim 푓(퐱) = 푓(퐚). 퐱→퐚 ⊲ (Source (https://commons.wikimedia.org/wiki/File:Example_of_continuous_function.svg)) We will not prove the following fundamental analysis result, which will be useful below. See e.g. Wikipedia 푑 (https://en.wikipedia.org/wiki/Extreme_value_theorem). Suppose 푓 : 퐷 → ℝ is defined on a set 퐷 ⊆ ℝ . We ∗ ∗ say that 푓 attains a maximum value 푀 at 퐳 if 푓(퐳 ) = 푀 and 푀 ≥ 푓(퐱) for all 퐱 ∈ 퐷. Similarly, we say 푓 attains a minimum value 푚 at 퐳∗ if 푓(퐳∗ ) = 푚 and 푚 ≥ 푓(퐱) for all 퐱 ∈ 퐷. Theorem (Extreme Value): Let 푓 : 퐷 → ℝ be a real-valued, continuous function on a nonempty, closed, 푑 bounded set 퐷 ⊆ ℝ . Then 푓 attains a maximum and a minimum on 퐷. 1.2.2 Derivatives We begin by reviewing the single-variable case. Recall that the derivative of a function of a real variable is the rate of change of the function with respect to the change in the variable. Formally: Definition (Derivative): Let 푓 : 퐷 → ℝ where 퐷 ⊆ ℝ and let 푥0 ∈ 퐷 be an interior point of 퐷. The derivative of 푓 at 푥0 is ′ d푓(푥0) 푓(푥0 + ℎ) − 푓(푥0) 푓 (푥0) = = lim d푥 ℎ→0 ℎ provided the limit exists. ⊲ (Source (https://commons.wikimedia.org/wiki/File:Tangent_to_a_curve.svg)) Exercise: Let 푓 and 푔 have derivatives at 푥 and let 훼 and 훽 be constants. Show that [훼푓(푥) + 훽푔(푥)]′ = 훼푓 ′ (푥) + 훽푔′ (푥). ⊲ The following lemma encapsulates a key insight about the derivative of 푓 at 푥0 : it tells us where to find smaller values. Lemma (Descent Direction): Let 푓 : 퐷 → ℝ with 퐷 ⊆ ℝ and let 푥0 ∈ 퐷 be an interior point of 퐷 where ′ ′ 푓 (푥0) exists. If 푓 (푥0) > 0, then there is an open ball 퐵훿(푥0) ⊆ 퐷 around 푥0 such that for each 푥 in 퐵훿(푥0): (a) 푓(푥) > 푓(푥0) if 푥 > 푥0 , (b) 푓(푥) < 푓(푥0) if 푥 < 푥0 . ′ If instead 푓 (푥0) < 0, the opposite holds. ′ Proof idea: Follows from the definition of the derivative by taking 휖 small enough that 푓 (푥0) − 휖 > 0. ′ Proof: Take 휖 = 푓 (푥0)/2. By definition of the derivative, there is 훿 > 0 such that 푓(푥 + ℎ) − 푓(푥 ) 푓 ′ (푥 ) − 0 0 < 휖 0 ℎ for all 0 < ℎ < 훿. Rearranging gives ′ 푓(푥0 + ℎ) > 푓(푥0) + [푓 (푥0) − 휖]ℎ > 푓(푥0) by our choice of 휖. The other direction is similar. ◻ 푑 For functions of several variables, we have the following generalization. As before, we let 퐞푖 ∈ ℝ be the 푖-th standard basis vector. 푑 Definition (Partial Derivative): Let 푓 : 퐷 → ℝ where 퐷 ⊆ ℝ and let 퐱0 ∈ 퐷 be an interior point of 퐷.
Recommended publications
  • Three Problems of the Theory of Choice on Random Sets
    DIVISION OF THE HUMANITIES AND SOCIAL SCIENCES CALIFORNIA INSTITUTE OF TECHNOLOGY PASADENA, CALIFORNIA 91125 THREE PROBLEMS OF THE THEORY OF CHOICE ON RANDOM SETS B. A. Berezovskiy,* Yu. M. Baryshnikov, A. Gnedin V. Institute of Control Sciences Academy of Sciences of the U.S.S.R, Moscow *Guest of the California Institute of Technology SOCIAL SCIENCE WORKING PAPER 661 December 1987 THREE PROBLEMS OF THE THEORY OF CHOICE ON RANDOM SETS B. A. Berezovskiy, Yu. M. Baryshnikov, A. V. Gnedin Institute of Control Sciences Academy of Sciences of the U.S.S.R., Moscow ABSTRACT This paper discusses three problems which are united not only by the common topic of research stated in thetitle, but also by a somewhat surprising interlacing of the methods and techniques used. In the first problem, an attempt is made to resolve a very unpleasant metaproblem arising in general choice theory: why theconditions of rationality are not really necessary or, in other words, why in every-day life we are quite satisfied with choice methods which are far from being ideal. The answer, substantiated by a number of results, is as follows: situations in which the choice function "misbehaves" are very seldom met in large presentations. the second problem, an overview of our studies is given on the problem of statistical propertiesIn of choice. One of themost astonishing phenomenon found when we deviate from scalar­ extremal choice functions is in stable multiplicity of choice. If our presentation is random, then a random number of alternatives is chosen in it. But how many? The answer isn't trival, and may be sought in many differentdirections.
    [Show full text]
  • The Halász-Székely Barycenter
    THE HALÁSZ–SZÉKELY BARYCENTER JAIRO BOCHI, GODOFREDO IOMMI, AND MARIO PONCE Abstract. We introduce a notion of barycenter of a probability measure re- lated to the symmetric mean of a collection of nonnegative real numbers. Our definition is inspired by the work of Halász and Székely, who in 1976 proved a law of large numbers for symmetric means. We study analytic properties of this Halász–Székely barycenter. We establish fundamental inequalities that relate the symmetric mean of a list of nonnegative real numbers with the barycenter of the measure uniformly supported on these points. As consequence, we go on to establish an ergodic theorem stating that the symmetric means of a se- quence of dynamical observations converges to the Halász–Székely barycenter of the corresponding distribution. 1. Introduction Means have fascinated man for a long time. Ancient Greeks knew the arith- metic, geometric, and harmonic means of two positive numbers (which they may have learned from the Babylonians); they also studied other types of means that can be defined using proportions: see [He, pp. 85–89]. Newton and Maclaurin en- countered the symmetric means (more about them later). Huygens introduced the notion of expected value and Jacob Bernoulli proved the first rigorous version of the law of large numbers: see [Mai, pp. 51, 73]. Gauss and Lagrange exploited the connection between the arithmetico-geometric mean and elliptic functions: see [BB]. Kolmogorov and other authors considered means from an axiomatic point of view and determined when a mean is arithmetic under a change of coordinates (i.e. quasiarithmetic): see [HLP, p.
    [Show full text]
  • Continuous Probability Distributions Continuous Probability Distributions – a Guide for Teachers (Years 11–12)
    1 Supporting Australian Mathematics Project 2 3 4 5 6 12 11 7 10 8 9 A guide for teachers – Years 11 and 12 Probability and statistics: Module 21 Continuous probability distributions Continuous probability distributions – A guide for teachers (Years 11–12) Professor Ian Gordon, University of Melbourne Editor: Dr Jane Pitkethly, La Trobe University Illustrations and web design: Catherine Tan, Michael Shaw Full bibliographic details are available from Education Services Australia. Published by Education Services Australia PO Box 177 Carlton South Vic 3053 Australia Tel: (03) 9207 9600 Fax: (03) 9910 9800 Email: [email protected] Website: www.esa.edu.au © 2013 Education Services Australia Ltd, except where indicated otherwise. You may copy, distribute and adapt this material free of charge for non-commercial educational purposes, provided you retain all copyright notices and acknowledgements. This publication is funded by the Australian Government Department of Education, Employment and Workplace Relations. Supporting Australian Mathematics Project Australian Mathematical Sciences Institute Building 161 The University of Melbourne VIC 3010 Email: [email protected] Website: www.amsi.org.au Assumed knowledge ..................................... 4 Motivation ........................................... 4 Content ............................................. 5 Continuous random variables: basic ideas ....................... 5 Cumulative distribution functions ............................ 6 Probability density functions ..............................
    [Show full text]
  • Probability with Engineering Applications ECE 313 Course Notes
    Probability with Engineering Applications ECE 313 Course Notes Bruce Hajek Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign January 2017 c 2017 by Bruce Hajek All rights reserved. Permission is hereby given to freely print and circulate copies of these notes so long as the notes are left intact and not reproduced for commercial purposes. Email to [email protected], pointing out errors or hard to understand passages or providing comments, is welcome. Contents 1 Foundations 3 1.1 Embracing uncertainty . .3 1.2 Axioms of probability . .6 1.3 Calculating the size of various sets . 10 1.4 Probability experiments with equally likely outcomes . 13 1.5 Sample spaces with infinite cardinality . 15 1.6 Short Answer Questions . 20 1.7 Problems . 21 2 Discrete-type random variables 25 2.1 Random variables and probability mass functions . 25 2.2 The mean and variance of a random variable . 27 2.3 Conditional probabilities . 32 2.4 Independence and the binomial distribution . 34 2.4.1 Mutually independent events . 34 2.4.2 Independent random variables (of discrete-type) . 36 2.4.3 Bernoulli distribution . 37 2.4.4 Binomial distribution . 38 2.5 Geometric distribution . 41 2.6 Bernoulli process and the negative binomial distribution . 43 2.7 The Poisson distribution{a limit of binomial distributions . 45 2.8 Maximum likelihood parameter estimation . 47 2.9 Markov and Chebychev inequalities and confidence intervals . 50 2.10 The law of total probability, and Bayes formula . 53 2.11 Binary hypothesis testing with discrete-type observations .
    [Show full text]
  • HARMONIC ANALYSIS 1. Maximal Function for a Locally Integrable
    HARMONIC ANALYSIS PIOTR HAJLASZ 1. Maximal function 1 n For a locally integrable function f 2 Lloc(R ) the Hardy-Littlewood max- imal function is defined by Z n Mf(x) = sup jf(y)j dy; x 2 R : r>0 B(x;r) The operator M is not linear but it is subadditive. We say that an operator T from a space of measurable functions into a space of measurable functions is subadditive if jT (f1 + f2)(x)j ≤ jT f1(x)j + jT f2(x)j a.e. and jT (kf)(x)j = jkjjT f(x)j for k 2 C. The following integrability result, known also as the maximal theorem, plays a fundamental role in many areas of mathematical analysis. p n Theorem 1.1 (Hardy-Littlewood-Wiener). If f 2 L (R ), 1 ≤ p ≤ 1, then Mf < 1 a.e. Moreover 1 n (a) For f 2 L (R ) 5n Z (1.1) jfx : Mf(x) > tgj ≤ jfj for all t > 0. t Rn p n p n (b) If f 2 L (R ), 1 < p ≤ 1, then Mf 2 L (R ) and p 1=p kMfk ≤ 2 · 5n=p kfk for 1 < p < 1, p p − 1 p kMfk1 ≤ kfk1 : Date: March 28, 2012. 1 2 PIOTR HAJLASZ The estimate (1.1) is called weak type estimate. 1 n 1 n Note that if f 2 L (R ) is a nonzero function, then Mf 62 L (R ). Indeed, R if λ = B(0;R) jfj > 0, then for jxj > R we have Z λ Mf(x) ≥ jfj ≥ n ; B(x;R+jxj) !n(R + jxj) n and the function on the right hand side is not integrable on R .
    [Show full text]
  • Probabilistic Models for Shapes As Continuous Curves
    J Math Imaging Vis (2009) 33: 39–65 DOI 10.1007/s10851-008-0104-3 Probabilistic Models for Shapes as Continuous Curves Jeong-Gyoo Kim · J. Alison Noble · J. Michael Brady Published online: 23 July 2008 © Springer Science+Business Media, LLC 2008 Abstract We develop new shape models by defining a stan- 1 Introduction dard shape from which we can explain shape deformation and variability. Currently, planar shapes are modelled using Our work contributes to shape analysis, particularly in med- a function space, which is applied to data extracted from ical image analysis, by defining a standard representation images. We regard a shape as a continuous curve and identi- of shape. Specifically we develop new mathematical mod- fied on the Wiener measure space whereas previous methods els of planar shapes. Our aim is to develop a standard rep- have primarily used sparse sets of landmarks expressed in a resentation of anatomical or biological shapes that can be Euclidean space. The average of a sample set of shapes is used to explain shape variations, whether in normal subjects defined using measurable functions which treat the Wiener or abnormalities due, for example, to disease. In addition, measure as varying Gaussians. Various types of invariance we propose a quasi-score that measures global deformation of our formulation of an average are examined in regard to and provides a generic way to compare statistical methods practical applications of it. The average is examined with of shape analysis. relation to a Fréchet mean in order to establish its valid- As D’Arcy Thompson pointed in [50], there is an impor- ity.
    [Show full text]
  • Two Notes on Financial Mathematics
    TWO NOTES ON FINANCIAL MATHEMATICS Daniel Dufresne Centre for Actuarial Studies University of Melbourne The two notes which follow are aimed at students of financial mathematics or actuarial science. The first one, “A note on correlation in mean variance portfolio theory” is an elementary discussion of how negative correlation coefficients can be, given any set of n random variables. The second one, “Two essential formulas in actuarial science and financial mathematics,” gives a more or less self-contained introduction to Jensen’s inequality and to another formula, which I have decided to call the “survival function expectation formula.” The latter allows one to write the expectation of a function of a random variable as an integral involving the complement of its cumulative distribution function. A NOTE ON CORRELATION IN MEAN VARIANCE PORTFOLIO THEORY Daniel Dufresne Centre for Actuarial Studies University of Melbourne Abstract This note is aimed at students and others interested in mean variance portfolio theory. Negative correlations are desirable in portfolio selection, as they decrease risk. It is shown that there is a mathematical limit to how negative correlations can be among a given number of securities. In particular, in an “average correlation model” (where the correlation coefficient between different securities is constant) the correlation has to be at least as large as −1/(n − 1), n being the number of securities. Keywords: Mean-variance portfolio theory; average correlation mod- els 1. INTRODUCTION One of the more technical points encountered when teaching the basics of mean variance portfolio theory is the restrictions which apply on the covariance matrix of the securities’ returns.
    [Show full text]
  • SMALL STRONG SOLUTIONS for TIME-DEPENDENT MEAN FIELD GAMES with LOCAL COUPLING 1. Introduction We Study Solutions of the Mean Fi
    SMALL STRONG SOLUTIONS FOR TIME-DEPENDENT MEAN FIELD GAMES WITH LOCAL COUPLING DAVID M. AMBROSE Abstract. For mean field games with local coupling, existence results are typically for weak solutions rather than strong solutions. We identify conditions on the Hamiltonian and the coupling which allow us to prove the existence of small, locally unique, strong solutions over any finite time interval in the case of local coupling; these conditions place us in the case of superquadratic Hamiltonians. For the regularity of solutions, we find that at each time in the interior of the time interval, the Fourier coefficients of the solutions decay exponentially. The method of proof is inspired by the work of Duchon and Robert on vortex sheets in incompressible fluids. Petites solutions fortes pour jeux `achamp moyens avec une d´ependance temporelle et un couplage local R´esum´e: Pour les jeux `achamp moyens avec couplage local, les r´esultatsd'existence sont typiquement obtenus pour des solutions faibles plut^otque pour des solutions fortes. Nous identifions des conditions sur le Hamiltonien et sur le couplage qui nous permettent de d´emontrer l'existence d'une solution forte, petite et localement unique pour tout intervalle de temps fini dans le cas d'un couplage local; ces conditions nous placent dans une situation de Hamiltonien super-quadratique. Pour la r´egularit´edes solutions, nous trouvons que, pour chaque point dans l'int´erieurde l'intervalle de temps, les coefficients de Fourier des solutions d´ecroissent exponentiellement. La preuve est inspir´eepar les travaux de Duchon et Robert sur les nappes de tourbillons de fluides incompressibles.
    [Show full text]
  • 7 Probability Theory and Statistics
    7 Probability Theory and Statistics • • • In the last chapter we made the transition from discussing information which is considered to be error free to dealing with data that contained intrinsic errors. In the case of the former, uncertainties in the results of our analysis resulted from the failure of the approximation formula to match the given data and from round-off error incurred during calculation. Uncertainties resulting from these sources will always be present, but in addition, the basic data itself may also contain errors. Since all data relating to the real world will have such errors, this is by far the more common situation. In this chapter we will consider the implications of dealing with data from the real world in more detail. 197 Numerical Methods and Data Analysis Philosophers divide data into at least two different categories, observational, historical, or empirical data and experimental data. Observational or historical data is, by its very nature, non-repeatable. Experimental data results from processes that, in principle, can be repeated. Some1 have introduced a third type of data labeled hypothetical-observational data, which is based on a combination of observation and information supplied by theory. An example of such data might be the distance to the Andromeda galaxy since a direct measurement of that quantity has yet to be made and must be deduced from other aspects of the physical world. However, in the last analysis, this is true of all observations of the world. Even the determination of repeatable, experimental data relies on agreed conventions of measurement for its unique interpretation.
    [Show full text]
  • Convergence of Stochastic Processes
    University of Pennsylvania ScholarlyCommons Technical Reports (CIS) Department of Computer & Information Science April 1992 Convergence of Stochastic Processes Robert Mandelbaum University of Pennsylvania Follow this and additional works at: https://repository.upenn.edu/cis_reports Recommended Citation Robert Mandelbaum, "Convergence of Stochastic Processes", . April 1992. University of Pennsylvania Department of Computer and Information Science Technical Report No. MS-CIS-92-30. This paper is posted at ScholarlyCommons. https://repository.upenn.edu/cis_reports/470 For more information, please contact [email protected]. Convergence of Stochastic Processes Abstract Often the best way to adumbrate a dark and dense assemblage of material is to describe the background in contrast to which the edges of the nebulosity may be clearly discerned. Hence, perhaps the most appropriate way to introduce this paper is to describe what it is not. It is not a comprehensive study of stochastic processes, nor an in-depth treatment of convergence. In fact, on the surface, the material covered in this paper is nothing more than a compendium of seemingly loosely-connected and barely- miscible theorems, methods and conclusions from the three main papers surveyed ([VC71], [Pol89] and [DL91]). Comments University of Pennsylvania Department of Computer and Information Science Technical Report No. MS- CIS-92-30. This technical report is available at ScholarlyCommons: https://repository.upenn.edu/cis_reports/470 Convergence of Stochastic Processes MS-CIS-92-30 GRASP LAB 311 Robert Mandelbaum University of Pennsylvania School of Engineering and Applied Science Computer and Information Science Department Philadelphia, PA 19104-6389 April 1992 Convergence of Stochastic Processes Robert Mandelbaum Department of Computer and Information Science University of Pennsylvania February 1992 Special Area Examination Advisor: Max Mintz Contents 1 Introduction 1 2 Revision of Basic Concepts 4 2.1 Linearity.
    [Show full text]
  • Asymptotic Behavior of the Sample Mean of a Function of the Wiener Process and the Macdonald Function
    J. Math. Sci. Univ. Tokyo 16 (2009), 55–93. Asymptotic Behavior of the Sample Mean of a Function of the Wiener Process and the MacDonald Function By Sergio Albeverio,Vadim Fatalov and Vladimir I. Piterbarg Abstract. Explicit asymptotic formulas for large time of expec- tations or distribution functions of exponential functionals of Brownian motion in terms of MacDonald’s function are proven. 1. Introduction. Main Results Many problems in probability and physics concern the asymptotic be- −1 T havior of the distribution function of the sample mean T 0 g(ξ(t))dt for large T , where ξ(t), t ≥ 0, is a random almost surely (a. s.) continuous process taking values in X ⊂ R and g is a continuous function on X, see [29], [37], [30], [7], [38, page 208]. Distributions and means of exponential functionals of Brownian motion and Bessel processes are intensively treated in connection with mathematical problems of financial mathematics, [19], [43]. In many cases exact distributions are evaluated, [33]. Nevertheless, dif- ficult problems remain open of evaluation of the asymptotic behavior of the sample means. As one of few examples of such the solutions, see [21], [22], where asymptotic behaviors of sample means of exponentials of a random walk, with applications to Brownian motion, where evaluated. In order to work these problems out, a fruitful idea is to pass from the nonlinear functional of ξ to a linear functional of the sojourn time Lt(ξ,·) (for definition see (2.1)), by the relation (2.2). For a wide class of homogeneous Markov processes ξ(t), main tools to evaluate such an asymptotic behavior are based on the Donsker-Varadhan large deviation 2000 Mathematics Subject Classification.
    [Show full text]
  • Concentration Inequalities
    A Gentle Introduction to Concentration Inequalities Karthik Sridharan Abstract This notes is ment to be a review of some basic inequalities and bounds on Random variables. A basic understanding of probability theory and set algebra might be required of the reader. This document is aimed to provide clear and complete proof for some inequalities. For readers familiar with the topics, many of the steps might seem trivial. None the less they are provided to simplify the proofs for readers new to the topic. This notes also provides to the best of my knowledge, the most generalized statement and proof for Symmetrization lemma. I also provides the less famous but geralized proof for Jensen’s inequality and logarithmic sobolev inequality. Refer [2] for a more detailed review of many of these inequalities with examples demonstrating their uses. 1 Preliminary Throughout this notes we shall consider a probability space (⌦, ,P) where E ⌦is the sample space, is the event class which is a σ algebra on ⌦and P E − is a probability measure. Further, we shall assume that there exists a borel measurable function mapping every point ! ⌦to a real number uniquely 2 called a random variable. We shall call the space of random variables X (note that X R). Further, we shall assume that all the functions and sets ✓ defined in the notes are measurable under the probability measure P . 2 Chebychev’s Inequality Let us start this notes by proving what is refered to as Chebychev’s Inequal- ity in [3]. Note, often by Chebychev’s inequality an inequality derived from the below proved theorem is used.
    [Show full text]