Geostatistical Model, Covariance Structure and Cokriging
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Stationary Processes and Their Statistical Properties
Stationary Processes and Their Statistical Properties Brian Borchers March 29, 2001 1 Stationary processes A discrete time stochastic process is a sequence of random variables Z1, Z2, :::. In practice we will typically analyze a single realization z1, z2, :::, zn of the stochastic process and attempt to esimate the statistical properties of the stochastic process from the realization. We will also consider the problem of predicting zn+1 from the previous elements of the sequence. We will begin by focusing on the very important class of stationary stochas- tic processes. A stochastic process is strictly stationary if its statistical prop- erties are unaffected by shifting the stochastic process in time. In particular, this means that if we take a subsequence Zk+1, :::, Zk+m, then the joint distribution of the m random variables will be the same no matter what k is. Stationarity requires that the mean of the stochastic process be a constant. E[Zk] = µ. and that the variance is constant 2 V ar[Zk] = σZ : Also, stationarity requires that the covariance of two elements separated by a distance m is constant. That is, Cov(Zk;Zk+m) is constant. This covariance is called the autocovariance at lag m, and we will use the notation γm. Since Cov(Zk;Zk+m) = Cov(Zk+m;Zk), we need only find γm for m 0. The ≥ correlation of Zk and Zk+m is the autocorrelation at lag m. We will use the notation ρm for the autocorrelation. It is easy to show that γk ρk = : γ0 1 2 The autocovariance and autocorrelation ma- trices The covariance matrix for the random variables Z1, :::, Zn is called an auto- covariance matrix. -
Getting Started with Geostatistics
FedGIS Conference February 24–25, 2016 | Washington, DC Getting Started with Geostatistics Brett Rose Objectives • Using geography as an integrating factor. • Introduce a variety of core tools available in GIS. • Demonstrate the utility of spatial & Geo statistics for a range of applications. • Modeling relationships and driving decisions Outline • Spatial Statistics vs Geostatistics • Geostatistical Workflow • Exploratory Spatial Data Analysis • Interpolation methods • Deterministic vs Geostatistical •Geostatistical Analyst in ArcGIS • Demos • Class Exercise In ArcGIS Geostatistical Analyst • interactive ESDA • interactive modeling including variography • many Kriging models (6 + cokriging) • pre-processing of data - Decluster, detrend, transformation • model diagnostics and comparison Spatial Analyst • rich set of tools to perform cell-based (raster) analysis • kriging (2 models), IDW, nearest neighbor … Spatial Statistics • ESDA GP tools • analyzing the distribution of geographic features • identifying spatial patterns What are spatial statistics in a GIS environment? • Software-based tools, methods, and techniques developed specifically for use with geographic data. • Spatial statistics: – Describe and spatial distributions, spatial patterns, spatial processes model , and spatial relationships. – Incorporate space (area, length, proximity, orientation, and/or spatial relationships) directly into their mathematics. In many ways spatial statistics extenD what the eyes anD minD Do intuitively to assess spatial patterns, trenDs anD relationships. -
Stochastic Process - Introduction
Stochastic Process - Introduction • Stochastic processes are processes that proceed randomly in time. • Rather than consider fixed random variables X, Y, etc. or even sequences of i.i.d random variables, we consider sequences X0, X1, X2, …. Where Xt represent some random quantity at time t. • In general, the value Xt might depend on the quantity Xt-1 at time t-1, or even the value Xs for other times s < t. • Example: simple random walk . week 2 1 Stochastic Process - Definition • A stochastic process is a family of time indexed random variables Xt where t belongs to an index set. Formal notation, { t : ∈ ItX } where I is an index set that is a subset of R. • Examples of index sets: 1) I = (-∞, ∞) or I = [0, ∞]. In this case Xt is a continuous time stochastic process. 2) I = {0, ±1, ±2, ….} or I = {0, 1, 2, …}. In this case Xt is a discrete time stochastic process. • We use uppercase letter {Xt } to describe the process. A time series, {xt } is a realization or sample function from a certain process. • We use information from a time series to estimate parameters and properties of process {Xt }. week 2 2 Probability Distribution of a Process • For any stochastic process with index set I, its probability distribution function is uniquely determined by its finite dimensional distributions. •The k dimensional distribution function of a process is defined by FXX x,..., ( 1 x,..., k ) = P( Xt ≤ 1 ,..., xt≤ X k) x t1 tk 1 k for anyt 1 ,..., t k ∈ I and any real numbers x1, …, xk . -
Autocovariance Function Estimation Via Penalized Regression
Autocovariance Function Estimation via Penalized Regression Lina Liao, Cheolwoo Park Department of Statistics, University of Georgia, Athens, GA 30602, USA Jan Hannig Department of Statistics and Operations Research, University of North Carolina, Chapel Hill, NC 27599, USA Kee-Hoon Kang Department of Statistics, Hankuk University of Foreign Studies, Yongin, 449-791, Korea Abstract The work revisits the autocovariance function estimation, a fundamental problem in statistical inference for time series. We convert the function estimation problem into constrained penalized regression with a generalized penalty that provides us with flexible and accurate estimation, and study the asymptotic properties of the proposed estimator. In case of a nonzero mean time series, we apply a penalized regression technique to a differenced time series, which does not require a separate detrending procedure. In penalized regression, selection of tuning parameters is critical and we propose four different data-driven criteria to determine them. A simulation study shows effectiveness of the tuning parameter selection and that the proposed approach is superior to three existing methods. We also briefly discuss the extension of the proposed approach to interval-valued time series. Some key words: Autocovariance function; Differenced time series; Regularization; Time series; Tuning parameter selection. 1 1 Introduction Let fYt; t 2 T g be a stochastic process such that V ar(Yt) < 1, for all t 2 T . The autoco- variance function of fYtg is given as γ(s; t) = Cov(Ys;Yt) for all s; t 2 T . In this work, we consider a regularly sampled time series f(i; Yi); i = 1; ··· ; ng. Its model can be written as Yi = g(i) + i; i = 1; : : : ; n; (1.1) where g is a smooth deterministic function and the error is assumed to be a weakly stationary 2 process with E(i) = 0, V ar(i) = σ and Cov(i; j) = γ(ji − jj) for all i; j = 1; : : : ; n: The estimation of the autocovariance function γ (or autocorrelation function) is crucial to determine the degree of serial correlation in a time series. -
An Introduction to Spatial Autocorrelation and Kriging
An Introduction to Spatial Autocorrelation and Kriging Matt Robinson and Sebastian Dietrich RenR 690 – Spring 2016 Tobler and Spatial Relationships • Tobler’s 1st Law of Geography: “Everything is related to everything else, but near things are more related than distant things.”1 • Simple but powerful concept • Patterns exist across space • Forms basic foundation for concepts related to spatial dependency Waldo R. Tobler (1) Tobler W., (1970) "A computer movie simulating urban growth in the Detroit region". Economic Geography, 46(2): 234-240. Spatial Autocorrelation (SAC) – What is it? • Autocorrelation: A variable is correlated with itself (literally!) • Spatial Autocorrelation: Values of a random variable, at paired points, are more or less similar as a function of the distance between them 2 Closer Points more similar = Positive Autocorrelation Closer Points less similar = Negative Autocorrelation (2) Legendre P. Spatial Autocorrelation: Trouble or New Paradigm? Ecology. 1993 Sep;74(6):1659–1673. What causes Spatial Autocorrelation? (1) Artifact of Experimental Design (sample sites not random) Parent Plant (2) Interaction of variables across space (see below) Univariate case – response variable is correlated with itself Eg. Plant abundance higher (clustered) close to other plants (seeds fall and germinate close to parent). Multivariate case – interactions of response and predictor variables due to inherent properties of the variables Eg. Location of seed germination function of wind and preferred soil conditions Mechanisms underlying patterns will depend on study system!! Why is it important? Presence of SAC can be good or bad (depends on your objectives) Good: If SAC exists, it may allow reliable estimation at nearby, non-sampled sites (interpolation). -
Master of Science in Geospatial Technologies Geostatistics
Instituto Superior de Estatística e Gestão de Informação Universidade Nova de Lisboa MasterMaster ofof ScienceScience inin GeospatialGeospatial TechnologiesTechnologies GeostatisticsGeostatistics PredictionsPredictions withwith GeostatisticsGeostatistics CarlosCarlos AlbertoAlberto FelgueirasFelgueiras [email protected]@isegi.unl.pt 1 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics Contents Stochastic Predictions Introduction Deterministic x Stochastic Methods Regionalized Variables Stationary Hypothesis Kriging Prediction – Simple, Ordinary and with Trend Kriging Prediction - Example Cokriging CrossValidation and Validation Problems with Geostatistic Estimators Advantages on using geostatistics 2 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics o sample locations Deterministic x Stochastic Methods z*(u) = K + estimation locations • Deterministic •The z *(u) value is estimated as a Deterministic Variable. An unique value is associated to its spatial location. • No uncertainties are associated to the z*(u)? estimations + • Stochastic •The z *(u) value is considered a Random Variable that has a probability distribution function associated to its possible values • Uncertainties can be associated to the estimations • Important: Samples are realizations of cdf Random Variables 3 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics • Random Variables and Random -
Conditional Simulation
Basics in Geostatistics 3 Geostatistical Monte-Carlo methods: Conditional simulation Hans Wackernagel MINES ParisTech NERSC • April 2013 http://hans.wackernagel.free.fr Basic concepts Geostatistics Hans Wackernagel (MINES ParisTech) Basics in Geostatistics 3 NERSC • April 2013 2 / 34 Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x). Kriging: Best Linear Unbiased Estimation of point values (or spatial averages) at any location of a region. Conditional simulation: generate an ensemble of realizations of the random function, conditional upon data. Statistics not linearly related to data can be computed from this ensemble. Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x). Kriging: Best Linear Unbiased Estimation of point values (or spatial averages) at any location of a region. Conditional simulation: generate an ensemble of realizations of the random function, conditional upon data. Statistics not linearly related to data can be computed from this ensemble. Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x). -
An Introduction to Model-Based Geostatistics
An Introduction to Model-based Geostatistics Peter J Diggle School of Health and Medicine, Lancaster University and Department of Biostatistics, Johns Hopkins University September 2009 Outline • What is geostatistics? • What is model-based geostatistics? • Two examples – constructing an elevation surface from sparse data – tropical disease prevalence mapping Example: data surface elevation Z 6 65 70 75 80 85 90 95100 5 4 Y 3 2 1 0 1 2 3 X 4 5 6 Geostatistics • traditionally, a self-contained methodology for spatial prediction, developed at Ecole´ des Mines, Fontainebleau, France • nowadays, that part of spatial statistics which is concerned with data obtained by spatially discrete sampling of a spatially continuous process Kriging: find the linear combination of the data that best predicts the value of the surface at an arbitrary location x Model-based Geostatistics • the application of general principles of statistical modelling and inference to geostatistical problems – formulate a statistical model for the data – fit the model using likelihood-based methods – use the fitted model to make predictions Kriging: minimum mean square error prediction under Gaussian modelling assumptions Gaussian geostatistics (simplest case) Model 2 • Stationary Gaussian process S(x): x ∈ IR E[S(x)] = µ Cov{S(x), S(x′)} = σ2ρ(kx − x′k) 2 • Mutually independent Yi|S(·) ∼ N(S(x),τ ) Point predictor: Sˆ(x) = E[S(x)|Y ] • linear in Y = (Y1, ..., Yn); • interpolates Y if τ 2 = 0 • called simple kriging in classical geostatistics Predictive distribution • choose the -
Efficient Geostatistical Simulation for Spatial Uncertainty Propagation
Efficient geostatistical simulation for spatial uncertainty propagation Stelios Liodakis Phaedon Kyriakidis Petros Gaganis University of the Aegean Cyprus University of Technology University of the Aegean University Hill 2-8 Saripolou Str., 3036 University Hill, 81100 Mytilene, Greece Lemesos, Cyprus Mytilene, Greece [email protected] [email protected] [email protected] Abstract Spatial uncertainty and error propagation have received significant attention in GIScience over the last two decades. Uncertainty propagation via Monte Carlo simulation using simple random sampling constitutes the method of choice in GIScience and environmental modeling in general. In this spatial context, geostatistical simulation is often employed for generating simulated realizations of attribute values in space, which are then used as model inputs for uncertainty propagation purposes. In the case of complex models with spatially distributed parameters, however, Monte Carlo simulation could become computationally expensive due to the need for repeated model evaluations; e.g., models involving detailed analytical solutions over large (often three dimensional) discretization grids. A novel simulation method is proposed for generating representative spatial attribute realizations from a geostatistical model, in that these realizations span in a better way (being more dissimilar than what is dictated by change) the range of possible values corresponding to that geostatistical model. It is demonstrated via a synthetic case study, that the simulations produced by the proposed method exhibit much smaller sampling variability, hence better reproduce the statistics of the geostatistical model. In terms of uncertainty propagation, such a closer reproduction of target statistics is shown to yield a better reproduction of model output statistics with respect to simple random sampling using the same number of realizations. -
ON the USE of NON-EUCLIDEAN ISOTROPY in GEOSTATISTICS Frank C
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Collection Of Biostatistics Research Archive Johns Hopkins University, Dept. of Biostatistics Working Papers 12-5-2005 ON THE USE OF NON-EUCLIDEAN ISOTROPY IN GEOSTATISTICS Frank C. Curriero Department of Environmental Health Sciences and Biostatistics, Johns Hopkins Bloomberg School of Public Health, [email protected] Suggested Citation Curriero, Frank C., "ON THE USE OF NON-EUCLIDEAN ISOTROPY IN GEOSTATISTICS" (December 2005). Johns Hopkins University, Dept. of Biostatistics Working Papers. Working Paper 94. http://biostats.bepress.com/jhubiostat/paper94 This working paper is hosted by The Berkeley Electronic Press (bepress) and may not be commercially reproduced without the permission of the copyright holder. Copyright © 2011 by the authors On the Use of Non-Euclidean Isotropy in Geostatistics Frank C. Curriero Department of Environmental Health Sciences and Department of Biostatistics The Johns Hopkins University Bloomberg School of Public Health 615 N. Wolfe Street Baltimore, MD 21205 [email protected] Summary. This paper investigates the use of non-Euclidean distances to characterize isotropic spatial dependence for geostatistical related applications. A simple example is provided to demonstrate there are no guarantees that existing covariogram and variogram functions remain valid (i.e. positive de¯nite or conditionally negative de¯nite) when used with a non-Euclidean distance measure. Furthermore, satisfying the conditions of a metric is not su±cient to ensure the distance measure can be used with existing functions. Current literature is not clear on these topics. There are certain distance measures that when used with existing covariogram and variogram functions remain valid, an issue that is explored. -
Covariances of ARMA Models
Statistics 910, #9 1 Covariances of ARMA Processes Overview 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation function 5. Discussion Review of ARMA processes ARMA process A stationary solution fXtg (or if its mean is not zero, fXt − µg) of the linear difference equation Xt − φ1Xt−1 − · · · − φpXt−p = wt + θ1wt−1 + ··· + θqwt−q φ(B)Xt = θ(B)wt (1) 2 where wt denotes white noise, wt ∼ WN(0; σ ). Definition 3.5 adds the • identifiability condition that the polynomials φ(z) and θ(z) have no zeros in common and the • normalization condition that φ(0) = θ(0) = 1. Causal process A stationary process fXtg is said to be causal if there ex- ists a summable sequence (some require square summable (`2), others want more and require absolute summability (`1)) sequence f jg such that fXtg has the one-sided moving average representation 1 X Xt = jwt−j = (B)wt : (2) j=0 Proposition 3.1 states that a stationary ARMA process fXtg is causal if and only if (iff) the zeros of the autoregressive polynomial Statistics 910, #9 2 φ(z) lie outside the unit circle (i.e., φ(z) 6= 0 for jzj ≤ 1). Since φ(0) = 1, φ(z) > 0 for jzj ≤ 1. (The unit circle in the complex plane consists of those x 2 C for which jzj = 1; the closed unit disc includes the interior of the unit circle as well as the circle; the open unit disc consists of jzj < 1.) If the zeros of φ(z) lie outside the unit circle, then we can invert each Qp of the factors (1 − B=zj) that make up φ(B) = j=1(1 − B=zj) one at a time (as when back-substituting in the derivation of the AR(1) representation). -
Covariance Functions
C. E. Rasmussen & C. K. I. Williams, Gaussian Processes for Machine Learning, the MIT Press, 2006, ISBN 026218253X. c 2006 Massachusetts Institute of Technology. www.GaussianProcess.org/gpml Chapter 4 Covariance Functions We have seen that a covariance function is the crucial ingredient in a Gaussian process predictor, as it encodes our assumptions about the function which we wish to learn. From a slightly different viewpoint it is clear that in supervised learning the notion of similarity between data points is crucial; it is a basic similarity assumption that points with inputs x which are close are likely to have similar target values y, and thus training points that are near to a test point should be informative about the prediction at that point. Under the Gaussian process view it is the covariance function that defines nearness or similarity. An arbitrary function of input pairs x and x0 will not, in general, be a valid valid covariance covariance function.1 The purpose of this chapter is to give examples of some functions commonly-used covariance functions and to examine their properties. Section 4.1 defines a number of basic terms relating to covariance functions. Section 4.2 gives examples of stationary, dot-product, and other non-stationary covariance functions, and also gives some ways to make new ones from old. Section 4.3 introduces the important topic of eigenfunction analysis of covariance functions, and states Mercer’s theorem which allows us to express the covariance function (under certain conditions) in terms of its eigenfunctions and eigenvalues. The covariance functions given in section 4.2 are valid when the input domain X is a subset of RD.