Calculus of Variations Raju K George, IIST

Total Page:16

File Type:pdf, Size:1020Kb

Calculus of Variations Raju K George, IIST Calculus of Variations Raju K George, IIST Lecture-1 In Calculus of Variations, we will study maximum and minimum of a certain class of functions. We first recall some maxima/minima results from the classical calculus. Maxima and Minima Let X and Y be two arbitrary sets and f : X Y be a well-defined function having domain → X and range Y . The function values f(x) become comparable if Y is IR or a subset of IR. Thus, optimization problem is valid for real valued functions. Let f : X IR be a real valued → function having X as its domain. Now x X is said to be maximum point for the function f if 0 ∈ f(x ) f(x) x X. The value f(x ) is called the maximum value of f. Similarly, x X is 0 ≥ ∀ ∈ 0 0 ∈ said to be a minimum point for the function f if f(x ) f(x) x X and in this case f(x ) is 0 ≤ ∀ ∈ 0 the minimum value of f. Sufficient condition for having maximum and minimum: Theorem (Weierstrass Theorem) Let S IR and f : S IR be a well defined function. Then f will have a maximum/minimum ⊆ → under the following sufficient conditions. 1. f : S IR is a continuous function. → 2. S IR is a bound and closed (compact) subset of IR. ⊂ Note that the above conditions are just sufficient conditions but not necessary. Example 1: Let f : [ 1, 1] IR defined by − → 1 x = 0 f(x)= − x x = 0 | | 6 1 −1 +1 −1 Obviously f(x) is not continuous at x = 0. However the f(x) has a minimum point x0 = 0 and maximum points at x = 1, x = 1. Continuity condition of the Weierstrass theorem is violated − but still the function has maximum and minimum. Example 2: Consider a function f : ( 1, 1) IR defined by − → 2 x< 0 f(x)= − 2 x 0 ≥ 2 () −1 +1 −2 f(x) has a maximum value x = 2 and a minimum value x = 2 even though both the conditions − (a) and (b) of Weierstrass theorem are violated. Example 3: Let f : [ 1, 1] IR be defined by f(x)= x2. − → 2 −1 1 This function satisfies both the conditions of Weierstrass theorem. f(x) has a minimum value 0 at x = 0 and maximum value 1 at x = 1 and x = 1. − Example 4: Let f : ( 1, 1) IR defined by f(x)= x2, f has a minimum at x = 0. − → −1 1 But f has no maximum point as x = 1 and x = 1 are outside the domain of the function. Here − the condition (b) is violated. Necessary condition for Maximum/Minimum when f is differentiable. Theorem Let f : S IR be a differentiable function and let x be an interior point of S and let x is → 0 0 either a maximum point or minimum point of f. Then the first derivative of f vanishes at x0. ie f ′(x0) = 0. This condition is just a necessary condition but not sufficient condition. An interior point x D IR is said to be a stationary point if f (x ) = 0. A stationary point 0 ∈ ⊆ ′ 0 x0 need not be maximum point/minimum point. However, if f ′′(x0) > 0 then x0 is a minimum point and if f ′′(x0) < 0 then x0 is a maximum point. Example 5: Let f : IR IR be defined by → f(x)= x3 3 2 f ′(x) = 3x =0 when x = 0 x3 However, x = 0 is neither a maximum point nor a minimum point of f(x)= x3. Example 6: Let f : IR IR be given by → f(x)= x2 Hence, f ′(x0) = 2x0 =0 when x0 = 0. Obviously x0 = 0 is a stationary point and this stationary point is minimum point of f(x), as f ′′(x0) = 2 > 0. Necessary conditions for Maxima/Minima functions of several variables Multi-variable functions. Let f : IRn IR be a real valued function of n - variables defined on IRn. If f has partial derivatives → at x IRn. If x is a maximum point/minimum point of the function f(x) then 0 ∈ 0 ∂f ∂f ∂f = 0, = 0, = 0 ∂x ∂y · · · ∂x 1 x=x0 2 x=x0 n 1 x=x0 − 4 A stationary point x0 is maximum point if the matrix ∂2f ∂2f ∂2f 2 ∂x ∂x1∂x2 · · · ∂x1∂xn 1 ∂2f ∂2f ∂2f ∂x ∂x ∂x2 · · · ∂x ∂x M = 2 1 2 2 n ··· ··· ··· ··· ∂2f ∂2f ∂2f 2 ∂xn∂x1 ∂xn∂x2 · · · ∂xn x=x0 is negative definite and x0 is minimum point if M is positive definite. Functionals: Let S be a set of functions. Let f : S IR be a real valued function. Such functions are known → as a functionals. In otherwords, a functional is a real valued function whose domain is a set of functions. Example 7: Let C[0, 1] be the set of all continuous functions defined on [0, 1] Let I : C[0, 1] IR be a function defined by → 1 I(y)= y(x) dx Z0 Obviously I(y) is a functional on C[0, 1]. The following table gives the values of I(y) for different functions y(x), listed in the table. y(x) I(y) x 0.5 x2 0.333 x3 0.25 sin x 0.4597 cos x 0.8415 ex 1.7183 1 1 We can find for which function y, the functional I(y) has a maximum value or minimum value. In the above example, I(y) will have minimum value for y(x) = x3 and I(y) will have maximum value for the function y(x)= ex out of the seven functions given here. 1 Let C [x1,x2] denote the set of continuously differentiable functions defined on [x1,x2]. Now x2 1 consider a functional I : C [x ,x ] IR defined by I(y) = F (x,y,y′)dx subject to the end 1 2 → Zx1 conditions y(x1)= y1 and y(x2)= y2. 5 In calculus of variations the basic problem is to find a function y for which the functional I(y) is maximum or minimum. We call such functions as extremizing functions and the value of the functional at the extremizing function as extremum. Consider the extremization problem x2 Extremize y I(y)= F (x,y,y′)dx Zx1 subject to the end conditions y(x1)= y1 and y(x2)= y2, where F is a twice continuously differen- tiable function. Example Find the shortest smooth plane curve joining two distinct points P (x1,y1) and Q(x2,y2). y Q(x2,y2) ds dy dx y(x) P(x1,y1) x There are infinitely many functions y passing through the given two points P (x1,y1) and Q(x2,y2). We are looking for a function which will have minimum arc length. Let ds be a small strip on the curve then we have. ds2 = dx2 + dy2 ds 2 dy 2 = 1+ dx dx dy 2 ds = 1+ dx dx s Q x2 dy 2 Total arc length I(y)= ds = 1+ dx 1 s dx ZP Zx Thus, the problem is to minimize I(y) subject to the end conditions y(x1)= y1 and y(x2)= y2. Lecture-2 Lemma (Fundamental Lemma of Calculus of Variations) b If f(x) is a continuous function defined on [a, b] and if f(x)g(x) dx = 0 for every function a g(x) C(a, b) such that g(a)= g(b) = 0 then f(x) 0 forZ all x [a, b]. ∈ ≡ ∈ 6 Proof: Let f(x) = 0 for some c (a, b). Without loss of generality let us assume that f(c) > 0. Now 6 ∈ because of continuity of f we have f(x) > 0 for some interval [x ,x ] [a, b] that contains the 1 2 ⊂ point c. (x x )(x x) for x [x ,x ] Let g(x)= − 1 2 − ∈ 1 2 0 outside [x ,x ] 1 2 Note that (x x )(x x) is positive for x (x ,x ). − 1 2 − ∈ 1 2 Now consider b x1 ✘✘✿ 0 x2 b ✘✘✿ 0 ✘✘✘ ✘✘✘ f(x)g(x) dx = ✘f(✘x)g(x) dx + f(x)g(x) dx + ✘f(✘x)g(x) dx a a x1 x2 Z Z x2 Z Z = f(x)g(x) dx x1 Z x2 = f(x)(x x )(x x) dx > 0 − 1 2 − Zx1 Thus we get a contradiction to what is given in the Lemma. This implies that f(x) 0 on [a, b]. ≡ Euler-Lagrange Equation (Necessary Condition for Extremum) Theorem: If y(x) is an extremizing function for the problem x2 Minimize/Maximize I(y) = F (x,y,y′) dx (1) Zx1 with end conditions y(x1)= y1 and y(x2)= y2 then y(x) satisfies the BVP. d ∂F ∂F = 0 (2) dx ∂y − ∂y ′ y(x1)= y1 and y(x2)= y2. (3) Equation (2) is known as the Euler-Lagrange equation. Proof: Let y(x) be an extremizing function for the functional I(y) in (1). 7 Y y(x) + η(x) y2 y1 y(x) η(x) O x X 1 x2 Let Y = y(x)+ ǫ η(x) be a variation of y(x), where η(x) is a continuously differentiable function with η(x1)=0= η(x2) and ǫ is a small constant. Hence I along the path Y = y(x)+ ǫ η(x) is given by x2 x2 I(ǫ)= F (x, y(x)+ ǫ η(x), y′(x)+ ǫ η′(x)) dx = F (x,Y (x),Y ′(x)) dx Zx1 Zx1 Since y(x) is an extremizing function, I(ǫ) has extremum when ǫ = 0.
Recommended publications
  • CALCULUS of VARIATIONS and TENSOR CALCULUS
    CALCULUS OF VARIATIONS and TENSOR CALCULUS U. H. Gerlach September 22, 2019 Beta Edition 2 Contents 1 FUNDAMENTAL IDEAS 5 1.1 Multivariable Calculus as a Prelude to the Calculus of Variations. 5 1.2 Some Typical Problems in the Calculus of Variations. ...... 6 1.3 Methods for Solving Problems in Calculus of Variations. ....... 10 1.3.1 MethodofFiniteDifferences. 10 1.4 TheMethodofVariations. 13 1.4.1 Variants and Variations . 14 1.4.2 The Euler-Lagrange Equation . 17 1.4.3 Variational Derivative . 20 1.4.4 Euler’s Differential Equation . 21 1.5 SolvedExample.............................. 24 1.6 Integration of Euler’s Differential Equation. ...... 25 2 GENERALIZATIONS 33 2.1 Functional with Several Unknown Functions . 33 2.2 Extremum Problem with Side Conditions. 38 2.2.1 HeuristicSolution. 40 2.2.2 Solution via Constraint Manifold . 42 2.2.3 Variational Problems with Finite Constraints . 54 2.3 Variable End Point Problem . 55 2.3.1 Extremum Principle at a Moment of Time Symmetry . 57 2.4 Generic Variable Endpoint Problem . 60 2.4.1 General Variations in the Functional . 62 2.4.2 Transversality Conditions . 64 2.4.3 Junction Conditions . 66 2.5 ManyDegreesofFreedom . 68 2.6 Parametrization Invariant Problem . 70 2.6.1 Parametrization Invariance via Homogeneous Function .... 71 2.7 Variational Principle for a Geodesic . 72 2.8 EquationofGeodesicMotion . 76 2.9 Geodesics: TheirParametrization.. 77 3 4 CONTENTS 2.9.1 Parametrization Invariance. 77 2.9.2 Parametrization in Terms of Curve Length . 78 2.10 Physical Significance of the Equation for a Geodesic . ....... 80 2.10.1 Freefloatframe ........................
    [Show full text]
  • FUNCTIONAL ANALYSIS 1. Banach and Hilbert Spaces in What
    FUNCTIONAL ANALYSIS PIOTR HAJLASZ 1. Banach and Hilbert spaces In what follows K will denote R of C. Definition. A normed space is a pair (X, k · k), where X is a linear space over K and k · k : X → [0, ∞) is a function, called a norm, such that (1) kx + yk ≤ kxk + kyk for all x, y ∈ X; (2) kαxk = |α|kxk for all x ∈ X and α ∈ K; (3) kxk = 0 if and only if x = 0. Since kx − yk ≤ kx − zk + kz − yk for all x, y, z ∈ X, d(x, y) = kx − yk defines a metric in a normed space. In what follows normed paces will always be regarded as metric spaces with respect to the metric d. A normed space is called a Banach space if it is complete with respect to the metric d. Definition. Let X be a linear space over K (=R or C). The inner product (scalar product) is a function h·, ·i : X × X → K such that (1) hx, xi ≥ 0; (2) hx, xi = 0 if and only if x = 0; (3) hαx, yi = αhx, yi; (4) hx1 + x2, yi = hx1, yi + hx2, yi; (5) hx, yi = hy, xi, for all x, x1, x2, y ∈ X and all α ∈ K. As an obvious corollary we obtain hx, y1 + y2i = hx, y1i + hx, y2i, hx, αyi = αhx, yi , Date: February 12, 2009. 1 2 PIOTR HAJLASZ for all x, y1, y2 ∈ X and α ∈ K. For a space with an inner product we define kxk = phx, xi . Lemma 1.1 (Schwarz inequality).
    [Show full text]
  • 1 Approximating Integrals Using Taylor Polynomials 1 1.1 Definitions
    Seunghee Ye Ma 8: Week 7 Nov 10 Week 7 Summary This week, we will learn how we can approximate integrals using Taylor series and numerical methods. Topics Page 1 Approximating Integrals using Taylor Polynomials 1 1.1 Definitions . .1 1.2 Examples . .2 1.3 Approximating Integrals . .3 2 Numerical Integration 5 1 Approximating Integrals using Taylor Polynomials 1.1 Definitions When we first defined the derivative, recall that it was supposed to be the \instantaneous rate of change" of a function f(x) at a given point c. In other words, f 0 gives us a linear approximation of f(x) near c: for small values of " 2 R, we have f(c + ") ≈ f(c) + "f 0(c) But if f(x) has higher order derivatives, why stop with a linear approximation? Taylor series take this idea of linear approximation and extends it to higher order derivatives, giving us a better approximation of f(x) near c. Definition (Taylor Polynomial and Taylor Series) Let f(x) be a Cn function i.e. f is n-times continuously differentiable. Then, the n-th order Taylor polynomial of f(x) about c is: n X f (k)(c) T (f)(x) = (x − c)k n k! k=0 The n-th order remainder of f(x) is: Rn(f)(x) = f(x) − Tn(f)(x) If f(x) is C1, then the Taylor series of f(x) about c is: 1 X f (k)(c) T (f)(x) = (x − c)k 1 k! k=0 Note that the first order Taylor polynomial of f(x) is precisely the linear approximation we wrote down in the beginning.
    [Show full text]
  • A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- Sity, Pomona, CA 91768
    A Quotient Rule Integration by Parts Formula Jennifer Switkes ([email protected]), California State Polytechnic Univer- sity, Pomona, CA 91768 In a recent calculus course, I introduced the technique of Integration by Parts as an integration rule corresponding to the Product Rule for differentiation. I showed my students the standard derivation of the Integration by Parts formula as presented in [1]: By the Product Rule, if f (x) and g(x) are differentiable functions, then d f (x)g(x) = f (x)g(x) + g(x) f (x). dx Integrating on both sides of this equation, f (x)g(x) + g(x) f (x) dx = f (x)g(x), which may be rearranged to obtain f (x)g(x) dx = f (x)g(x) − g(x) f (x) dx. Letting U = f (x) and V = g(x) and observing that dU = f (x) dx and dV = g(x) dx, we obtain the familiar Integration by Parts formula UdV= UV − VdU. (1) My student Victor asked if we could do a similar thing with the Quotient Rule. While the other students thought this was a crazy idea, I was intrigued. Below, I derive a Quotient Rule Integration by Parts formula, apply the resulting integration formula to an example, and discuss reasons why this formula does not appear in calculus texts. By the Quotient Rule, if f (x) and g(x) are differentiable functions, then ( ) ( ) ( ) − ( ) ( ) d f x = g x f x f x g x . dx g(x) [g(x)]2 Integrating both sides of this equation, we get f (x) g(x) f (x) − f (x)g(x) = dx.
    [Show full text]
  • Operations on Power Series Related to Taylor Series
    Operations on Power Series Related to Taylor Series In this problem, we perform elementary operations on Taylor series – term by term differen­ tiation and integration – to obtain new examples of power series for which we know their sum. Suppose that a function f has a power series representation of the form: 1 2 X n f(x) = a0 + a1(x − c) + a2(x − c) + · · · = an(x − c) n=0 convergent on the interval (c − R; c + R) for some R. The results we use in this example are: • (Differentiation) Given f as above, f 0(x) has a power series expansion obtained by by differ­ entiating each term in the expansion of f(x): 1 0 X n−1 f (x) = a1 + a2(x − c) + 2a3(x − c) + · · · = nan(x − c) n=1 • (Integration) Given f as above, R f(x) dx has a power series expansion obtained by by inte­ grating each term in the expansion of f(x): 1 Z a1 a2 X an f(x) dx = C + a (x − c) + (x − c)2 + (x − c)3 + · · · = C + (x − c)n+1 0 2 3 n + 1 n=0 for some constant C depending on the choice of antiderivative of f. Questions: 1. Find a power series representation for the function f(x) = arctan(5x): (Note: arctan x is the inverse function to tan x.) 2. Use power series to approximate Z 1 2 sin(x ) dx 0 (Note: sin(x2) is a function whose antiderivative is not an elementary function.) Solution: 1 For question (1), we know that arctan x has a simple derivative: , which then has a power 1 + x2 1 2 series representation similar to that of , where we subsitute −x for x.
    [Show full text]
  • Introduction to the Modern Calculus of Variations
    MA4G6 Lecture Notes Introduction to the Modern Calculus of Variations Filip Rindler Spring Term 2015 Filip Rindler Mathematics Institute University of Warwick Coventry CV4 7AL United Kingdom [email protected] http://www.warwick.ac.uk/filiprindler Copyright ©2015 Filip Rindler. Version 1.1. Preface These lecture notes, written for the MA4G6 Calculus of Variations course at the University of Warwick, intend to give a modern introduction to the Calculus of Variations. I have tried to cover different aspects of the field and to explain how they fit into the “big picture”. This is not an encyclopedic work; many important results are omitted and sometimes I only present a special case of a more general theorem. I have, however, tried to strike a balance between a pure introduction and a text that can be used for later revision of forgotten material. The presentation is based around a few principles: • The presentation is quite “modern” in that I use several techniques which are perhaps not usually found in an introductory text or that have only recently been developed. • For most results, I try to use “reasonable” assumptions, not necessarily minimal ones. • When presented with a choice of how to prove a result, I have usually preferred the (in my opinion) most conceptually clear approach over more “elementary” ones. For example, I use Young measures in many instances, even though this comes at the expense of a higher initial burden of abstract theory. • Wherever possible, I first present an abstract result for general functionals defined on Banach spaces to illustrate the general structure of a certain result.
    [Show full text]
  • Derivative of Power Series and Complex Exponential
    LECTURE 4: DERIVATIVE OF POWER SERIES AND COMPLEX EXPONENTIAL The reason of dealing with power series is that they provide examples of analytic functions. P1 n Theorem 1. If n=0 anz has radius of convergence R > 0; then the function P1 n F (z) = n=0 anz is di®erentiable on S = fz 2 C : jzj < Rg; and the derivative is P1 n¡1 f(z) = n=0 nanz : Proof. (¤) We will show that j F (z+h)¡F (z) ¡ f(z)j ! 0 as h ! 0 (in C), whenever h ¡ ¢ n Pn n k n¡k jzj < R: Using the binomial theorem (z + h) = k=0 k h z we get F (z + h) ¡ F (z) X1 (z + h)n ¡ zn ¡ hnzn¡1 ¡ f(z) = a h n h n=0 µ ¶ X1 a Xn n = n ( hkzn¡k) h k n=0 k=2 µ ¶ X1 Xn n = a h( hk¡2zn¡k) n k n=0 k=2 µ ¶ X1 Xn¡2 n = a h( hjzn¡2¡j) (by putting j = k ¡ 2): n j + 2 n=0 j=0 ¡ n ¢ ¡n¡2¢ By using the easily veri¯able fact that j+2 · n(n ¡ 1) j ; we obtain µ ¶ F (z + h) ¡ F (z) X1 Xn¡2 n ¡ 2 j ¡ f(z)j · jhj n(n ¡ 1)ja j( jhjjjzjn¡2¡j) h n j n=0 j=0 X1 n¡2 = jhj n(n ¡ 1)janj(jzj + jhj) : n=0 P1 n¡2 We already know that the series n=0 n(n ¡ 1)janjjzj converges for jzj < R: Now, for jzj < R and h ! 0 we have jzj + jhj < R eventually.
    [Show full text]
  • Calculus Terminology
    AP Calculus BC Calculus Terminology Absolute Convergence Asymptote Continued Sum Absolute Maximum Average Rate of Change Continuous Function Absolute Minimum Average Value of a Function Continuously Differentiable Function Absolutely Convergent Axis of Rotation Converge Acceleration Boundary Value Problem Converge Absolutely Alternating Series Bounded Function Converge Conditionally Alternating Series Remainder Bounded Sequence Convergence Tests Alternating Series Test Bounds of Integration Convergent Sequence Analytic Methods Calculus Convergent Series Annulus Cartesian Form Critical Number Antiderivative of a Function Cavalieri’s Principle Critical Point Approximation by Differentials Center of Mass Formula Critical Value Arc Length of a Curve Centroid Curly d Area below a Curve Chain Rule Curve Area between Curves Comparison Test Curve Sketching Area of an Ellipse Concave Cusp Area of a Parabolic Segment Concave Down Cylindrical Shell Method Area under a Curve Concave Up Decreasing Function Area Using Parametric Equations Conditional Convergence Definite Integral Area Using Polar Coordinates Constant Term Definite Integral Rules Degenerate Divergent Series Function Operations Del Operator e Fundamental Theorem of Calculus Deleted Neighborhood Ellipsoid GLB Derivative End Behavior Global Maximum Derivative of a Power Series Essential Discontinuity Global Minimum Derivative Rules Explicit Differentiation Golden Spiral Difference Quotient Explicit Function Graphic Methods Differentiable Exponential Decay Greatest Lower Bound Differential
    [Show full text]
  • 20-Finding Taylor Coefficients
    Taylor Coefficients Learning goal: Let’s generalize the process of finding the coefficients of a Taylor polynomial. Let’s find a general formula for the coefficients of a Taylor polynomial. (Students complete worksheet series03 (Taylor coefficients).) 2 What we have discovered is that if we have a polynomial C0 + C1x + C2x + ! that we are trying to match values and derivatives to a function, then when we differentiate it a bunch of times, we 0 get Cnn!x + Cn+1(n+1)!x + !. Plugging in x = 0 and everything except Cnn! goes away. So to (n) match the nth derivative, we must have Cn = f (0)/n!. If we want to match the values and derivatives at some point other than zero, we will use the 2 polynomial C0 + C1(x – a) + C2(x – a) + !. Then when we differentiate, we get Cnn! + Cn+1(n+1)!(x – a) + !. Now, plugging in x = a makes everything go away, and we get (n) Cn = f (a)/n!. So we can generate a polynomial of any degree (well, as long as the function has enough derivatives!) that will match the function and its derivatives to that degree at a particular point a. We can also extend to create a power series called the Taylor series (or Maclaurin series if a is specifically 0). We have natural questions: • For what x does the series converge? • What does the series converge to? Is it, hopefully, f(x)? • If the series converges to the function, we can use parts of the series—Taylor polynomials—to approximate the function, at least nearby a.
    [Show full text]
  • Linear Spaces
    Chapter 2 Linear Spaces Contents FieldofScalars ........................................ ............. 2.2 VectorSpaces ........................................ .............. 2.3 Subspaces .......................................... .............. 2.5 Sumofsubsets........................................ .............. 2.5 Linearcombinations..................................... .............. 2.6 Linearindependence................................... ................ 2.7 BasisandDimension ..................................... ............. 2.7 Convexity ............................................ ............ 2.8 Normedlinearspaces ................................... ............... 2.9 The `p and Lp spaces ............................................. 2.10 Topologicalconcepts ................................... ............... 2.12 Opensets ............................................ ............ 2.13 Closedsets........................................... ............. 2.14 Boundedsets......................................... .............. 2.15 Convergence of sequences . ................... 2.16 Series .............................................. ............ 2.17 Cauchysequences .................................... ................ 2.18 Banachspaces....................................... ............... 2.19 Completesubsets ....................................... ............. 2.19 Transformations ...................................... ............... 2.21 Lineartransformations.................................. ................ 2.21
    [Show full text]
  • General Inner Product & Fourier Series
    General Inner Products 1 General Inner Product & Fourier Series Advanced Topics in Linear Algebra, Spring 2014 Cameron Braithwaite 1 General Inner Product The inner product is an algebraic operation that takes two vectors of equal length and com- putes a single number, a scalar. It introduces a geometric intuition for length and angles of vectors. The inner product is a generalization of the dot product which is the more familiar operation that's specific to the field of real numbers only. Euclidean space which is limited to 2 and 3 dimensions uses the dot product. The inner product is a structure that generalizes to vector spaces of any dimension. The utility of the inner product is very significant when proving theorems and runing computations. An important aspect that can be derived is the notion of convergence. Building on convergence we can move to represenations of functions, specifally periodic functions which show up frequently. The Fourier series is an expansion of periodic functions specific to sines and cosines. By the end of this paper we will be able to use a Fourier series to represent a wave function. To build toward this result many theorems are stated and only a few have proofs while some proofs are trivial and left for the reader to save time and space. Definition 1.11 Real Inner Product Let V be a real vector space and a; b 2 V . An inner product on V is a function h ; i : V x V ! R satisfying the following conditions: (a) hαa + α0b; ci = αha; ci + α0hb; ci (b) hc; αa + α0bi = αhc; ai + α0hc; bi (c) ha; bi = hb; ai (d) ha; aiis a positive real number for any a 6= 0 Definition 1.12 Complex Inner Product Let V be a vector space.
    [Show full text]
  • Function Space Tensor Decomposition and Its Application in Sports Analytics
    East Tennessee State University Digital Commons @ East Tennessee State University Electronic Theses and Dissertations Student Works 12-2019 Function Space Tensor Decomposition and its Application in Sports Analytics Justin Reising East Tennessee State University Follow this and additional works at: https://dc.etsu.edu/etd Part of the Applied Statistics Commons, Multivariate Analysis Commons, and the Other Applied Mathematics Commons Recommended Citation Reising, Justin, "Function Space Tensor Decomposition and its Application in Sports Analytics" (2019). Electronic Theses and Dissertations. Paper 3676. https://dc.etsu.edu/etd/3676 This Thesis - Open Access is brought to you for free and open access by the Student Works at Digital Commons @ East Tennessee State University. It has been accepted for inclusion in Electronic Theses and Dissertations by an authorized administrator of Digital Commons @ East Tennessee State University. For more information, please contact [email protected]. Function Space Tensor Decomposition and its Application in Sports Analytics A thesis presented to the faculty of the Department of Mathematics East Tennessee State University In partial fulfillment of the requirements for the degree Master of Science in Mathematical Sciences by Justin Reising December 2019 Jeff Knisley, Ph.D. Nicole Lewis, Ph.D. Michele Joyner, Ph.D. Keywords: Sports Analytics, PCA, Tensor Decomposition, Functional Analysis. ABSTRACT Function Space Tensor Decomposition and its Application in Sports Analytics by Justin Reising Recent advancements in sports information and technology systems have ushered in a new age of applications of both supervised and unsupervised analytical techniques in the sports domain. These automated systems capture large volumes of data points about competitors during live competition.
    [Show full text]