A New Analysis on Variability and Predictability of Seasonal Rainfall

Total Page:16

File Type:pdf, Size:1020Kb

A New Analysis on Variability and Predictability of Seasonal Rainfall JP1.8 WAVELET ANALYSIS ON VARIABILITY, TELECONNECTIVITY AND PREDICTABILITY OF EAST AFRICA RAINFALL Davison Mwalea, Thian Yew Gana* and Samuel Shenb a Department of Civil and Environment Engineering, University of Alberta, Alberta, Canada b Department of Mathematics, University of Alberta, Alberta, Canada 1. INTRODUCTION In investigating EA rainfall variability and its relationship to other climatic Problems of quantifying rainfall elements it has become imperative to variability in space with respect to climatic analyze the irregularly distributed events in elements such as SST, winds, air time that depict non-stationary power at temperature and specific humidity have many different frequencies. been a subject of numerous studies for To meet this requirement, this study Eastern Africa, (EA) (e.g., Mutai et al. 1998; used wavelet analysis and wavelet based 2002; Ntale et al. 2003), and other studies empirical orthogonal function analysis as summarized in Beltrando (1990). (WEOF), also known as wavelet principal Following some of these studies, EA has component analysis (WLPCA) and wavelet been divided into between 6 and 20 zones independent component analysis (WLICA) of homogenous rainfall variability, found to to study the variability, teleconnectivity and teleconnect to SST variability in the Atlantic, predictability of EA rainfall. In this study, two Indian and Pacific oceans, (Mutai et al. sets of information from the wavelet spectra 1998; Ntale et al. 2003) and atmospheric are used. These are (1) energy coefficients signals, Philippon et al. (2002). Many of (also known as power) of individual scales of these studies applied a combination of time the wavelet spectra, and (2) the average domain multivariate statistics such as wavelet spectra energy over some harmonic analysis, Fourier analysis, linear significant scales (called scale averaged regressions and cross correlations to climate wavelet power, SAWP). Combinations of data to determine relationships between and empirical orthogonal function analysis and among the climate elements. A common and wavelets analysis have previously been well-recognized shortcoming of such applied to multivariate statistical process approaches has been the assumption that monitoring (e.g., Bakshi 1998). climate data is stationary and linear, a criterion few data sets from natural 2. DATA AND METHODS. phenomena satisfy (Huang et al., 1998). With unstable relationships between EA has two rainy seasons, the rainfall in EA and SST in the oceans and September-November (SON) and March- atmospheric circulation fields highlighted by May (MAM). Monthly rainfall data (1950- erratic rainfall between 1965 and 1997, one 1995) from 21 grid locations at a resolution of the major challenges for meteorologists of 2.5°×3.75° latitude and longitude was has been to predict the nature of this extracted for EA (4ºS-12ºS, 28ºE-44ºE), see variability in the rainfall of EA over different Fig 1.The rainfall data was provided by the spatial and temporal scales. Potts (1971), UK meteorological office. and Rodhe and Virji (1996) showed that The monthly rainfall data was oscillatory peaks of 2-2.5, 3.5 and 5.6 years transformed into seasonal data by summing in addition to other peaks associated with EL the monthly values for each grid location. Nino (e.g., Ropelewski and Halpert 1987) Monthly SST anomaly grid data (1950-1995) exist in EA rainfall. However, it has never at 5o x 5o latitude and longitude resolution been clear as to when these oscillations was extracted from the Indian (20ºN-40ºS, occurred in the rainfall. 40ºE-105ºE) and Atlantic (10ºN-30ºS, 50ºW- 10ºE) Ocean. This data was transformed *Corresponding Author address: Thian Yew Gan, Univ. into seasonal and annual data by computing of Alberta, Dept of Civil and Environmental Engineering, 3-month averages for JFM, AMJ, JAS and Edmonton, Alberta AB T6G 2G7, Canada; Email: [email protected] SON and annual averages, respectively. The SST dataset is part of MOHSST6 and 1 was provided by the UK meteorological continuous or a discrete values. The office. quantity a −1/ 2 in Equation (2) is an energy normalization term, which ensures that energy of the mother, and daughter wavelets remain the same over all scales and making it possible to directly compare wavelet transforms of one time series with another (Torrence and Compo, 1998). The wavelet transform of a real signal X (t) with respect to the mother wavelet is a convolution integral given as 1 T t − b W ()b,a = X ()t Ψ * dt (3) a ∫0 a where ψ* is the complex conjugate of ψ. In Fig. 1. (a) Study location (b) detailed this equation, W (b,a) is a wavelet location and description of EA (Uganda, spectrum, a matrix of energy coefficients of Kenya and Tanzania) the decomposed time series X (t). A faster and much more efficient way to compute the wavelet transform is done in the Fourier 2.2 Wavelet Analysis space using the Fourier transform of a discrete time series, X (t), as Wavelets are a set of limited duration waves, also called daughter T ∩ ∩ * wavelets, because they are formed by iω nδt k (4) dilations and translations of a single Wt ()a = ∑ X Ψ (sω k )e prototype wavelet function ψ(t), where t is k =0 real valued, called the basic or mother where the caret symbolizes Fourier wavelet (Castleman, 1996). The mother Transform, k is the frequency index (0,…T) wavelet designed to oscillate like a wave, is and Ψ (sω ) is the Fourier transform of the required to span an area that sums to zero, k wavelet function. The wavelet spectrum was and die out rapidly to zero as t tends to computed using a discrete set of 20 scales infinity to satisfy the so called “admissibility” starting at 2 years in fractional power of two condition. using Ψ()t dt = 0 (1) jδj ∫ s j = so 2 (5) A set of wavelets can be generated by where s0 is twice the sampling rate,j translating and scaling the basic wavelet as =0,1,….20, and δj = 0.25, thus giving follows scales(periods) ranging from 2 to 64 years. The wavelet transform of a time series 1 t − b contains a wealth of information, which can Ψa,b ()t = Ψ (2) a a be condensed over a range of scales to construct the scale averaged wavelet power (SAWP) and used in multivariate analysis. where the scale (width) of the wavelet and 2 translated position along the t-axis are a and j2 W a 2 δ jδ t t ( j ) (6) b respectively. By continuously varying a Wt = ∑ long b, a picture is constructed depicting Cδ j= j1 a j how the energy over various frequencies where Cδ is the reconstruction factor that varies with time. The parameters a and b takes on values depending on the mother are real and a, always positive, may take 2 wavelet used, δj is a factor for scale makes the extracted components as averaging and δt is the sampling period. The independent as possible. To obtain the initial global wavelet spectrum shows dominant most important independent components, oscillations present in a time series. The pre-whitening, a process of projecting the local wavelet power shows how the original data into a subspace spanned by dominant oscillations vary with time. To the first few PCs is employed. Using these examine whether two remotely located projections optimization algorithms are used multivariate time series are related to each to estimate the independent components. other or if one modulates the other over a ICA is considered an extension to PCA and number of scales, the individual scale power used in this study as a complementary tool or the SAWP can be constructed by varying to EOF analysis, to allow the underlying the scale, a, and computing the weighted structure of the data to be more readily sum of the wavelet power over those scales. observed. Using the results of global spectra computed for some selected SST fields located in the 2.5 Artificial Neural Network driven Indian and Atlantic Oceans and rainfall fields by Genetic Algorithm (ANN-GA) located in EA, we extracted energy of some for rainfall prediction individual scale and SAWP from the 2-8 year range. To compute the wavelet power The Genetic algorithm calibrated for this study, the Morlet wavelet (k = 6), neural network used in this study (Fig not was used because its structure resembles shown) consists of a population of feed that of a rainfall time series. forward neural networks embedded in a Genetic algorithm routine. The ANN Ψ (t) = π-1/4ei6te-t2/2 (7) parameters are iteratively improved via genetic evolution (selection, crossover and 2.3 Wavelet Empirical Orthogonal mutation) to more accurately model the joint Function Analysis (WEOF). SST-rainfall variability. The objective function used is a combination of the Empirical Orthogonal Functions Pearson correlation and the root mean analysis (EOF) has been widely used (e.g. square error (RMSE). For each neural Kutzbach, 1967) for analyzing spatial and network of the population, the predictand, y, temporal variability of physical fields to is obtained as a nonlinear translation of the objectively identify the spatially uncorrelated weighted average of the PCs of raw data, x modes of variability of a given field. In this study EOF is used on individual scale power 2 1 1 2 y = f w f w x + b + b (8) or SAWP. Since the WLPCs are obtained 2 ∑ ( 1 ∑ ( ) ) from SAWP, they are interpreted as ‘frequency compacted’ energy variability. To where x is the standardized anomaly and w1 identify and delineate temporally and and w2 are weights for each solution to the spatially uncorrelated patterns at regional hidden and the output layers respectively, scale, we applied the WEOF analysis on and b1 and b2 are the bias vectors SAWP of the SST of the South Atlantic and associated with the hidden and output Indian Ocean and rainfall time series of EA.
Recommended publications
  • Lecture 19: Wavelet Compression of Time Series and Images
    Lecture 19: Wavelet compression of time series and images c Christopher S. Bretherton Winter 2014 Ref: Matlab Wavelet Toolbox help. 19.1 Wavelet compression of a time series The last section of wavelet leleccum notoolbox.m demonstrates the use of wavelet compression on a time series. The idea is to keep the wavelet coefficients of largest amplitude and zero out the small ones. 19.2 Wavelet analysis/compression of an image Wavelet analysis is easily extended to two-dimensional images or datasets (data matrices), by first doing a wavelet transform of each column of the matrix, then transforming each row of the result (see wavelet image). The wavelet coeffi- cient matrix has the highest level (largest-scale) averages in the first rows/columns, then successively smaller detail scales further down the rows/columns. The ex- ample also shows fine results with 50-fold data compression. 19.3 Continuous Wavelet Transform (CWT) Given a continuous signal u(t) and an analyzing wavelet (x), the CWT has the form Z 1 s − t W (λ, t) = λ−1=2 ( )u(s)ds (19.3.1) −∞ λ Here λ, the scale, is a continuous variable. We insist that have mean zero and that its square integrates to 1. The continuous Haar wavelet is defined: 8 < 1 0 < t < 1=2 (t) = −1 1=2 < t < 1 (19.3.2) : 0 otherwise W (λ, t) is proportional to the difference of running means of u over successive intervals of length λ/2. 1 Amath 482/582 Lecture 19 Bretherton - Winter 2014 2 In practice, for a discrete time series, the integral is evaluated as a Riemann sum using the Matlab wavelet toolbox function cwt.
    [Show full text]
  • Wavelet Entropy Energy Measure (WEEM): a Multiscale Measure to Grade a Geophysical System's Predictability
    EGU21-703, updated on 28 Sep 2021 https://doi.org/10.5194/egusphere-egu21-703 EGU General Assembly 2021 © Author(s) 2021. This work is distributed under the Creative Commons Attribution 4.0 License. Wavelet Entropy Energy Measure (WEEM): A multiscale measure to grade a geophysical system's predictability Ravi Kumar Guntu and Ankit Agarwal Indian Institute of Technology Roorkee, Hydrology, Roorkee, India ([email protected]) Model-free gradation of predictability of a geophysical system is essential to quantify how much inherent information is contained within the system and evaluate different forecasting methods' performance to get the best possible prediction. We conjecture that Multiscale Information enclosed in a given geophysical time series is the only input source for any forecast model. In the literature, established entropic measures dealing with grading the predictability of a time series at multiple time scales are limited. Therefore, we need an additional measure to quantify the information at multiple time scales, thereby grading the predictability level. This study introduces a novel measure, Wavelet Entropy Energy Measure (WEEM), based on Wavelet entropy to investigate a time series's energy distribution. From the WEEM analysis, predictability can be graded low to high. The difference between the entropy of a wavelet energy distribution of a time series and entropy of wavelet energy of white noise is the basis for gradation. The metric quantifies the proportion of the deterministic component of a time series in terms of energy concentration, and its range varies from zero to one. One corresponds to high predictable due to its high energy concentration and zero representing a process similar to the white noise process having scattered energy distribution.
    [Show full text]
  • Adaptive Wavelet Clustering for Highly Noisy Data
    Adaptive Wavelet Clustering for Highly Noisy Data Zengjian Chen Jiayi Liu Yihe Deng Department of Computer Science Department of Computer Science Department of Mathematics Huazhong University of University of Massachusetts Amherst University of California, Los Angeles Science and Technology Massachusetts, USA California, USA Wuhan, China [email protected] [email protected] [email protected] Kun He* John E. Hopcroft Department of Computer Science Department of Computer Science Huazhong University of Science and Technology Cornell University Wuhan, China Ithaca, NY, USA [email protected] [email protected] Abstract—In this paper we make progress on the unsupervised Based on the pioneering work of Sheikholeslami that applies task of mining arbitrarily shaped clusters in highly noisy datasets, wavelet transform, originally used for signal processing, on which is a task present in many real-world applications. Based spatial data clustering [12], we propose a new wavelet based on the fundamental work that first applies a wavelet transform to data clustering, we propose an adaptive clustering algorithm, algorithm called AdaWave that can adaptively and effectively denoted as AdaWave, which exhibits favorable characteristics for uncover clusters in highly noisy data. To tackle general appli- clustering. By a self-adaptive thresholding technique, AdaWave cations, we assume that the clusters in a dataset do not follow is parameter free and can handle data in various situations. any specific distribution and can be arbitrarily shaped. It is deterministic, fast in linear time, order-insensitive, shape- To show the hardness of the clustering task, we first design insensitive, robust to highly noisy data, and requires no pre- knowledge on data models.
    [Show full text]
  • A Fourier-Wavelet Monte Carlo Method for Fractal Random Fields
    JOURNAL OF COMPUTATIONAL PHYSICS 132, 384±408 (1997) ARTICLE NO. CP965647 A Fourier±Wavelet Monte Carlo Method for Fractal Random Fields Frank W. Elliott Jr., David J. Horntrop, and Andrew J. Majda Courant Institute of Mathematical Sciences, 251 Mercer Street, New York, New York 10012 Received August 2, 1996; revised December 23, 1996 2 2H k[v(x) 2 v(y)] l 5 CHux 2 yu , (1.1) A new hierarchical method for the Monte Carlo simulation of random ®elds called the Fourier±wavelet method is developed and where 0 , H , 1 is the Hurst exponent and k?l denotes applied to isotropic Gaussian random ®elds with power law spectral the expected value. density functions. This technique is based upon the orthogonal Here we develop a new Monte Carlo method based upon decomposition of the Fourier stochastic integral representation of the ®eld using wavelets. The Meyer wavelet is used here because a wavelet expansion of the Fourier space representation of its rapid decay properties allow for a very compact representation the fractal random ®elds in (1.1). This method is capable of the ®eld. The Fourier±wavelet method is shown to be straightfor- of generating a velocity ®eld with the Kolmogoroff spec- ward to implement, given the nature of the necessary precomputa- trum (H 5 Ad in (1.1)) over many (10 to 15) decades of tions and the run-time calculations, and yields comparable results scaling behavior comparable to the physical space multi- with scaling behavior over as many decades as the physical space multiwavelet methods developed recently by two of the authors.
    [Show full text]
  • Biostatistics for Oral Healthcare
    Biostatistics for Oral Healthcare Jay S. Kim, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Ronald J. Dailey, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Biostatistics for Oral Healthcare Biostatistics for Oral Healthcare Jay S. Kim, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Ronald J. Dailey, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 JayS.Kim, PhD, is Professor of Biostatistics at Loma Linda University, CA. A specialist in this area, he has been teaching biostatistics since 1997 to students in public health, medical school, and dental school. Currently his primary responsibility is teaching biostatistics courses to hygiene students, predoctoral dental students, and dental residents. He also collaborates with the faculty members on a variety of research projects. Ronald J. Dailey is the Associate Dean for Academic Affairs at Loma Linda and an active member of American Dental Educational Association. C 2008 by Blackwell Munksgaard, a Blackwell Publishing Company Editorial Offices: Blackwell Publishing Professional, 2121 State Avenue, Ames, Iowa 50014-8300, USA Tel: +1 515 292 0140 9600 Garsington Road, Oxford OX4 2DQ Tel: 01865 776868 Blackwell Publishing Asia Pty Ltd, 550 Swanston Street, Carlton South, Victoria 3053, Australia Tel: +61 (0)3 9347 0300 Blackwell Wissenschafts Verlag, Kurf¨urstendamm57, 10707 Berlin, Germany Tel: +49 (0)30 32 79 060 The right of the Author to be identified as the Author of this Work has been asserted in accordance with the Copyright, Designs and Patents Act 1988. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by the UK Copyright, Designs and Patents Act 1988, without the prior permission of the publisher.
    [Show full text]
  • A Practical Guide to Wavelet Analysis
    A Practical Guide to Wavelet Analysis Christopher Torrence and Gilbert P. Compo Program in Atmospheric and Oceanic Sciences, University of Colorado, Boulder, Colorado ABSTRACT A practical step-by-step guide to wavelet analysis is given, with examples taken from time series of the El Niño– Southern Oscillation (ENSO). The guide includes a comparison to the windowed Fourier transform, the choice of an appropriate wavelet basis function, edge effects due to finite-length time series, and the relationship between wavelet scale and Fourier frequency. New statistical significance tests for wavelet power spectra are developed by deriving theo- retical wavelet spectra for white and red noise processes and using these to establish significance levels and confidence intervals. It is shown that smoothing in time or scale can be used to increase the confidence of the wavelet spectrum. Empirical formulas are given for the effect of smoothing on significance levels and confidence intervals. Extensions to wavelet analysis such as filtering, the power Hovmöller, cross-wavelet spectra, and coherence are described. The statistical significance tests are used to give a quantitative measure of changes in ENSO variance on interdecadal timescales. Using new datasets that extend back to 1871, the Niño3 sea surface temperature and the Southern Oscilla- tion index show significantly higher power during 1880–1920 and 1960–90, and lower power during 1920–60, as well as a possible 15-yr modulation of variance. The power Hovmöller of sea level pressure shows significant variations in 2–8-yr wavelet power in both longitude and time. 1. Introduction complete description of geophysical applications can be found in Foufoula-Georgiou and Kumar (1995), Wavelet analysis is becoming a common tool for while a theoretical treatment of wavelet analysis is analyzing localized variations of power within a time given in Daubechies (1992).
    [Show full text]
  • An Overview of Wavelet Transform Concepts and Applications
    An overview of wavelet transform concepts and applications Christopher Liner, University of Houston February 26, 2010 Abstract The continuous wavelet transform utilizing a complex Morlet analyzing wavelet has a close connection to the Fourier transform and is a powerful analysis tool for decomposing broadband wavefield data. A wide range of seismic wavelet applications have been reported over the last three decades, and the free Seismic Unix processing system now contains a code (succwt) based on the work reported here. Introduction The continuous wavelet transform (CWT) is one method of investigating the time-frequency details of data whose spectral content varies with time (non-stationary time series). Moti- vation for the CWT can be found in Goupillaud et al. [12], along with a discussion of its relationship to the Fourier and Gabor transforms. As a brief overview, we note that French geophysicist J. Morlet worked with non- stationary time series in the late 1970's to find an alternative to the short-time Fourier transform (STFT). The STFT was known to have poor localization in both time and fre- quency, although it was a first step beyond the standard Fourier transform in the analysis of such data. Morlet's original wavelet transform idea was developed in collaboration with the- oretical physicist A. Grossmann, whose contributions included an exact inversion formula. A series of fundamental papers flowed from this collaboration [16, 12, 13], and connections were soon recognized between Morlet's wavelet transform and earlier methods, including harmonic analysis, scale-space representations, and conjugated quadrature filters. For fur- ther details, the interested reader is referred to Daubechies' [7] account of the early history of the wavelet transform.
    [Show full text]
  • Big Data for Reliability Engineering: Threat and Opportunity
    Reliability, February 2016 Big Data for Reliability Engineering: Threat and Opportunity Vitali Volovoi Independent Consultant [email protected] more recently, analytics). It shares with the rest of the fields Abstract - The confluence of several technologies promises under this umbrella the need to abstract away most stormy waters ahead for reliability engineering. News reports domain-specific information, and to use tools that are mainly are full of buzzwords relevant to the future of the field—Big domain-independent1. As a result, it increasingly shares the Data, the Internet of Things, predictive and prescriptive lingua franca of modern systems engineering—probability and analytics—the sexier sisters of reliability engineering, both statistics that are required to balance the otherwise orderly and exciting and threatening. Can we reliability engineers join the deterministic engineering world. party and suddenly become popular (and better paid), or are And yet, reliability engineering does not wear the fancy we at risk of being superseded and driven into obsolescence? clothes of its sisters. There is nothing privileged about it. It is This article argues that“big-picture” thinking, which is at the rarely studied in engineering schools, and it is definitely not core of the concept of the System of Systems, is key for a studied in business schools! Instead, it is perceived as a bright future for reliability engineering. necessary evil (especially if the reliability issues in question are safety-related). The community of reliability engineers Keywords - System of Systems, complex systems, Big Data, consists of engineers from other fields who were mainly Internet of Things, industrial internet, predictive analytics, trained on the job (instead of receiving formal degrees in the prescriptive analytics field).
    [Show full text]
  • Volatility Estimation of Stock Prices Using Garch Method
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by International Institute for Science, Technology and Education (IISTE): E-Journals European Journal of Business and Management www.iiste.org ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online) Vol.7, No.19, 2015 Volatility Estimation of Stock Prices using Garch Method Koima, J.K, Mwita, P.N Nassiuma, D.K Kabarak University Abstract Economic decisions are modeled based on perceived distribution of the random variables in the future, assessment and measurement of the variance which has a significant impact on the future profit or losses of particular portfolio. The ability to accurately measure and predict the stock market volatility has a wide spread implications. Volatility plays a very significant role in many financial decisions. The main purpose of this study is to examine the nature and the characteristics of stock market volatility of Kenyan stock markets and its stylized facts using GARCH models. Symmetric volatility model namly GARCH model was used to estimate volatility of stock returns. GARCH (1, 1) explains volatility of Kenyan stock markets and its stylized facts including volatility clustering, fat tails and mean reverting more satisfactorily.The results indicates the evidence of time varying stock return volatility over the sampled period of time. In conclusion it follows that in a financial crisis; the negative returns shocks have higher volatility than positive returns shocks. Keywords: GARCH, Stylized facts, Volatility clustering INTRODUCTION Volatility forecasting in financial market is very significant particularly in Investment, financial risk management and monetory policy making Poon and Granger (2003).Because of the link between volatility and risk,volatility can form a basis for efficient price discovery.Volatility implying predictability is very important phenomenon for traders and medium term - investors.
    [Show full text]
  • Fourier, Wavelet and Monte Carlo Methods in Computational Finance
    Fourier, Wavelet and Monte Carlo Methods in Computational Finance Kees Oosterlee1;2 1CWI, Amsterdam 2Delft University of Technology, the Netherlands AANMPDE-9-16, 7/7/2016 Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Joint work with Fang Fang, Marjon Ruijter, Luis Ortiz, Shashi Jain, Alvaro Leitao, Fei Cong, Qian Feng Agenda Derivatives pricing, Feynman-Kac Theorem Fourier methods Basics of COS method; Basics of SWIFT method; Options with early-exercise features COS method for Bermudan options Monte Carlo method BSDEs, BCOS method (very briefly) Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Agenda Derivatives pricing, Feynman-Kac Theorem Fourier methods Basics of COS method; Basics of SWIFT method; Options with early-exercise features COS method for Bermudan options Monte Carlo method BSDEs, BCOS method (very briefly) Joint work with Fang Fang, Marjon Ruijter, Luis Ortiz, Shashi Jain, Alvaro Leitao, Fei Cong, Qian Feng Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Feynman-Kac theorem: Z T v(t; x) = E g(s; Xs )ds + h(XT ) ; t where Xs is the solution to the FSDE dXs = µ(Xs )ds + σ(Xs )d!s ; Xt = x: Feynman-Kac Theorem The linear partial differential equation: @v(t; x) + Lv(t; x) + g(t; x) = 0; v(T ; x) = h(x); @t with operator 1 Lv(t; x) = µ(x)Dv(t; x) + σ2(x)D2v(t; x): 2 Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 2 / 51 Feynman-Kac Theorem The linear partial differential equation: @v(t; x) + Lv(t; x) + g(t; x) = 0; v(T ; x) = h(x); @t with operator 1 Lv(t; x) = µ(x)Dv(t; x) + σ2(x)D2v(t; x): 2 Feynman-Kac theorem: Z T v(t; x) = E g(s; Xs )ds + h(XT ) ; t where Xs is the solution to the FSDE dXs = µ(Xs )ds + σ(Xs )d!s ; Xt = x: Kees Oosterlee (CWI, TU Delft) Comp.
    [Show full text]
  • A Python Library for Neuroimaging Based Machine Learning
    Brain Predictability toolbox: a Python library for neuroimaging based machine learning 1 1 2 1 1 1 Hahn, S. ,​ Yuan, D.K. ,​ Thompson, W.K. ,​ Owens, M. ,​ Allgaier, N .​ and Garavan, H ​ ​ ​ ​ ​ ​ 1. Departments of Psychiatry and Complex Systems, University of Vermont, Burlington, VT 05401, USA 2. Division of Biostatistics, Department of Family Medicine and Public Health, University of California, San Diego, La Jolla, CA 92093, USA Abstract Summary Brain Predictability toolbox (BPt) represents a unified framework of machine learning (ML) tools designed to work with both tabulated data (in particular brain, psychiatric, behavioral, and physiological variables) and neuroimaging specific derived data (e.g., brain volumes and surfaces). This package is suitable for investigating a wide range of different neuroimaging based ML questions, in particular, those queried from large human datasets. Availability and Implementation BPt has been developed as an open-source Python 3.6+ package hosted at https://github.com/sahahn/BPt under MIT License, with documentation provided at ​ https://bpt.readthedocs.io/en/latest/, and continues to be actively developed. The project can be ​ downloaded through the github link provided. A web GUI interface based on the same code is currently under development and can be set up through docker with instructions at https://github.com/sahahn/BPt_app. ​ Contact Please contact Sage Hahn at [email protected] ​ Main Text 1 Introduction Large datasets in all domains are becoming increasingly prevalent as data from smaller existing studies are pooled and larger studies are funded. This increase in available data offers an unprecedented opportunity for researchers interested in applying machine learning (ML) based methodologies, especially those working in domains such as neuroimaging where data collection is quite expensive.
    [Show full text]
  • BIOSTATS Documentation
    BIOSTATS Documentation BIOSTATS is a collection of R functions written to aid in the statistical analysis of ecological data sets using both univariate and multivariate procedures. All of these functions make use of existing R functions and many are simply convenience wrappers for functions contained in other popular libraries, such as VEGAN and LABDSV for multivariate statistics, however others are custom functions written using functions in the base or stats libraries. Author: Kevin McGarigal, Professor Department of Environmental Conservation University of Massachusetts, Amherst Date Last Updated: 9 November 2016 Functions: all.subsets.gam.. 3 all.subsets.glm. 6 box.plots. 12 by.names. 14 ci.lines. 16 class.monte. 17 clus.composite.. 19 clus.stats. 21 cohen.kappa. 23 col.summary. 25 commission.mc. 27 concordance. 29 contrast.matrix. 33 cov.test. 35 data.dist. 37 data.stand. 41 data.trans. 44 dist.plots. 48 distributions. 50 drop.var. 53 edf.plots.. 55 ecdf.plots. 57 foa.plots.. 59 hclus.cophenetic. 62 hclus.scree. 64 hclus.table. 66 hist.plots. 68 intrasetcor. 70 Biostats Library 2 kappa.mc. 72 lda.structure.. 74 mantel2. 76 mantel.part. 78 mrpp2. 82 mv.outliers.. 84 nhclus.scree. 87 nmds.monte. 89 nmds.scree.. 91 norm.test. 93 ordi.monte.. 95 ordi.overlay. 97 ordi.part.. 99 ordi.scree. 105 pca.communality. 107 pca.eigenval. 109 pca.eigenvec. 111 pca.structure. 113 plot.anosim. 115 plot.mantel. 117 plot.mrpp.. 119 plot.ordi.part. 121 qqnorm.plots.. 123 ran.split. ..
    [Show full text]