Geostatistical Model, Covariance Structure and Cokriging

Total Page:16

File Type:pdf, Size:1020Kb

Geostatistical Model, Covariance Structure and Cokriging Introduction Geostatistical Model Covariance structure Cokriging Geostatistical Model, Covariance structure and Cokriging Hans Wackernagel¹ ¹Equipe de Géostatistique — MINES ParisTech www.cg.ensmp.fr Statistics and Machine Learning Interface Meeting Manchester, July 2009 Statistics & Machine Learning • Manchester, July 2009 Introduction Geostatistical Model Covariance structure Cokriging Introduction Statistics & Machine Learning • Manchester, July 2009 Geostatistics and Gaussian processes Geostatistics is not limited to Gaussian processes, it usually refers to the concept of random functions, it may also build on concepts from random sets theory. Geostatistics Geostatistics: is mostly known for the kriging techniques, nowadays deals much with geostatistical simulation. Bayesian inference of geostatistical parameters has also become a topic of research. Sequential data assimilation is an extension of geostatistics using a mechanistic model to describe the time dynamics. In this talk: we will stay with linear (Gaussian) geostatistics, concentrate on kriging in a multi-scale and multi-variate context. A typical application may be: the response surface estimation problem eventually with several correlated response variables. Statistical inference of parameters will not be discussed. Statistics vs Machine Learning ? Necessity of an interface meeting Differences (subjective): geostatistics favours interpretability of the statistical model, machine learning stresses prediction performance and computational perfomance of algorithms. Ideally both should be achieved. Geostatistics: definition Geostatistics is an application of the theory of regionalized variables to the problem of predicting spatial phenomena. (G. MATHERON, 1970) Usually we consider the regionalized variable z(x) to be a realization of a random function Z(x). Stationarity For the top series: we think of a (2nd order) stationary model For the bottom series: a mean and a finite variance do not make sense, rather the realization of a non-stationary process without drift. Second-order stationary model Mean and covariance are translation invariant The mean of the random function does not depend on x: h i E Z(x) = m The covariance depends on length and orientation of the vector h linking two points x and x0 = x + h: h i cov(Z(x);Z(x0)) = C(h) = E Z(x) − m · Z(x+h) − m Non-stationary model (without drift) Variance of increments is translation invariant The mean of increments does not depend on x and is zero: h i E Z(x+h) − Z(x) = m(h) = 0 The variance of increments depends only on h: h i var Z (x+h)−Z (x) = 2g(h) This is called intrinsic stationarity. Intrinsic stationarity does not imply 2nd order stationarity. 2nd order stationarity implies stationary increments. The variogram With intrinsic stationarity: 1 h 2 i g(h) = E Z (x+h) − Z(x) 2 Properties - zero at the origin g(0) = 0 - positive values g(h) ≥ 0 - even function g(h) = g(−h) The covariance function is bounded by the variance: C(0) = s 2 ≥ jC(h)j The variogram is not bounded. A variogram can always be constructed from a given covariance function: g(h) = C(0) − C(h) The converse is not true. What is a variogram ? A covariance function is a positive definite function. What is a variogram? A variogram is a conditionnally negative definite function. In particular: any variogram matrix Γ = [g(xa −xb )] is conditionally negative semi-definite, > > [wa ] g(xa −xb ) [wa ] = w Γw ≤ 0 for any set of weights with n ∑ wa = 0: a=0 Ordinary kriging n n ? Estimator: Z (x0) = ∑ wa Z(xa ) with ∑ wa = 1 a=1 a=1 Solving: " n # ? argmin var(Z (x0) − Z(x0)) − 2m( ∑ wa − 1) w1;:::;wn;m a=1 yields the system: 8 n > w g(x −x ) + m = g(x −x ) 8a > ∑ b a b a 0 < b =1 n > w = 1 :> ∑ b b =1 n 2 and the kriging variance: sK = m + ∑ wa g(xa −x0) a=1 Kriging the mean Stationary model: Z (x) = Y (x) + m n n ? Estimator: M = ∑ wa Z(xa ) with ∑ wa = 1 a=1 a=1 Solving: " n # ? argmin var(M − m) − 2m( ∑ wa − 1) w1;:::;wn;m a=1 yields the system: 8 n > w C(x −x ) − m = 0 8a > ∑ b a b < b =1 n > w = 1 :> ∑ b b =1 2 and the kriging variance: sK = m Kriging a component S Stationary model: Z(x) = ∑ Yu(x) + m (Yu ? Yv foru 6= v) u=0 n n Estimator: Y ? (x ) = w Z(x ) with w = 0 u0 0 ∑ a a ∑ a a=1 a=1 Solving: " # n argmin var Y ? (x ) − Y (x ) − 2m w u0 0 u0 0 ∑ a w1;:::;wn;m a=1 yields the system: 8 n > w C(x −x ) − m = C (x −x ) 8a > ∑ b a b u0 a 0 < b =1 n > w = 0 :> ∑ b b =1 Mobile phone exposure of children by Liudmila KUDRYAVTSEVA http://perso.rd.francetelecom.fr/joe.wiart/ Child heads at different ages Phone position and child head Head of 12 year old child SAR exposure (simulated) Max SAR for different positions of phone The phone positions are characterized by two angles The SAR values are normalized with respect to 1 W. Variogram Anisotropic linear variogram model Max SAR kriged map Kriging standard deviations Kriging standard deviations Different sample design Introduction Geostatistical Model Covariance structure Cokriging Geostatistical filtering: Skagerrak SST Statistics & Machine Learning • Manchester, July 2009 NAR16 images on 26-27 april 2005 Sea-surface temperature (SST) NAR SST − 200504261000S NAR SST − 200504261200S 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 59˚ 59˚59˚ 59˚ 58˚ 58˚58˚ 58˚ 57˚ 57˚57˚ 57˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6 7 8 9 6 7 8 9 NAR SST − 200504262000S NAR SST − 200504270200S 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 59˚ 59˚59˚ 59˚ 58˚ 58˚58˚ 58˚ 57˚ 57˚57˚ 57˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6 7 8 9 6 7 8 9 Nested variogram model Nested scales modeled by sum of different variograms: micro-scale nugget-effect of .005 small scale spherical model (range .4 deg longitude, sill .06) large scale linear model Geostatistical filtering Small-scale variability of NAR16 images NAR SST − SHORT26apr10h NAR SST − SHORT26apr12h 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 59˚ 59˚59˚ 59˚ 58˚ 58˚58˚ 58˚ 57˚ 57˚57˚ 57˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ −1 0 1 −1 0 1 NAR SST − SHORT26apr20h NAR SST − SHORT27apr02h 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 59˚ 59˚59˚ 59˚ 58˚ 58˚58˚ 58˚ 57˚ 57˚57˚ 57˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ 6˚ 7˚ 8˚ 9˚ 10˚ 11˚ −1 0 1 −1 0 1 Introduction Geostatistical Model Covariance structure Cokriging Geostatistical Model Statistics & Machine Learning • Manchester, July 2009 Introduction Geostatistical Model Covariance structure Cokriging Linear model of coregionalization The linear model of coregionalization (LMC) combines: a linear model for different scales of the spatial variation, a linear model for components of the multivariate variation. Statistics & Machine Learning • Manchester, July 2009 Two linear models S Linear Model of Regionalization: Z(x) = ∑ Yu(x) u=0 h i E Yu(x+h) − Yu(x) = 0 h i E Yu(x+h) − Yu(x) · Yv (x+h) − Yv (x) = gu(h)duv N Linear Model of PCA: Zi = ∑ aip Yp p=1 h i E Yp = 0 cov Yp;Yq = 0 for p 6= q Linear Model of Coregionalization Spatial and multivariate representation of Zi (x) using p u uncorrelated factors Yu (x) with coefficients aip: S N u p Zi (x) = ∑ ∑ aip Yu (x) u=0 p=1 p Given u, all factors Yu (x) have the same variogram gu(h). This implies a multivariate nested variogram: S Γ(h) = ∑ Bu gu(h) u=0 Coregionalization matrices The coregionalization matrices Bu characterize the correlation between the variables Zi at different spatial scales. In practice: 1 A multivariate nested variogram model is fitted. 2 Each matrix is then decomposed using a PCA: N h u i h u u i Bu = bij = ∑ aip ajp p=1 yielding the coefficients of the LMC. LMC: intrinsic correlation When all coregionalization matrices are proportional to a matrix B: Bu = au B we have an intrinsically correlated LMC: S Γ(h) = B ∑ au gu(h) = B g(h) u=0 In practice, with intrinsic correlation, the eigenanalysis of the different Bu will yield: different sets of eigenvalues, but identical sets of eigenvectors. Regionalized Multivariate Data Analysis With intrinsic correlation: The factors are autokrigeable, i.e. the factors can be computed from a classical MDA on the variance-covariance matrix V =∼ B and are kriged subsequently. With spatial-scale dependent correlation: The factors are defined on the basis of the coregionalization matrices Bu and are cokriged subsequently. Need for a regionalized multivariate data analysis! Regionalized PCA ? Variables Zi (x) # no Intrinsic Correlation ? −! gij (h) = ∑ Bu gu(h) u # yes # PCA on B PCA on Bu ## Transform into Y Cokrige Y ? (x) u0p0 ## Krige Y ? (x) −! Map of PC p0 Introduction Geostatistical Model Covariance structure Cokriging Geostatistical filtering: Golfe du Lion SST Statistics & Machine Learning • Manchester, July 2009 Modeling of spatial variability as the sum of a small-scale and a large-scale process Variogram of SST SST on 7 june 2005 The variogram of the Nar16 im- age is fitted with a short- and a long-range structure (with geo- metrical anisotropy). The small-scale components of the NAR16 image and of corresponding MARS ocean-model output are extracted by geostatistical filtering. Geostatistical filtering Small scale (top) and large scale (bottom) features 3˚ 4˚ 5˚ 6˚ 7˚ 3˚ 4˚ 5˚ 6˚ 7˚ 43˚ 43˚ 43˚ 43˚ 3˚ 4˚ 5˚ 6˚ 7˚ 3˚ 4˚ 5˚ 6˚ 7˚ −0.5 0.0 0.5 −0.5 0.0 0.5 NAR16 image MARS model output 3˚ 4˚ 5˚ 6˚ 7˚ 3˚ 4˚ 5˚ 6˚ 7˚ 43˚ 43˚ 43˚ 43˚ 3˚ 4˚ 5˚ 6˚ 7˚ 3˚ 4˚ 5˚ 6˚ 7˚ 15 20 15 20 Zoom into NE corner 3˚ 4˚ 5˚ 6˚ 7˚ 8001000 900 700 100 600 600 200500 400 200500 43˚ 700 1100 43˚ 300 9001000 14001200 400 100 8001600 15001300300 1700 2400 200018002300 2100 1900 2200 300200 800 900 500 1000 1200 400 600700 1300 Bathymetry 1100 1500 1400 1700 1600 1800 1900 2000 3˚ 4˚ 5˚ 6˚ 7˚ 0 500
Recommended publications
  • Stationary Processes and Their Statistical Properties
    Stationary Processes and Their Statistical Properties Brian Borchers March 29, 2001 1 Stationary processes A discrete time stochastic process is a sequence of random variables Z1, Z2, :::. In practice we will typically analyze a single realization z1, z2, :::, zn of the stochastic process and attempt to esimate the statistical properties of the stochastic process from the realization. We will also consider the problem of predicting zn+1 from the previous elements of the sequence. We will begin by focusing on the very important class of stationary stochas- tic processes. A stochastic process is strictly stationary if its statistical prop- erties are unaffected by shifting the stochastic process in time. In particular, this means that if we take a subsequence Zk+1, :::, Zk+m, then the joint distribution of the m random variables will be the same no matter what k is. Stationarity requires that the mean of the stochastic process be a constant. E[Zk] = µ. and that the variance is constant 2 V ar[Zk] = σZ : Also, stationarity requires that the covariance of two elements separated by a distance m is constant. That is, Cov(Zk;Zk+m) is constant. This covariance is called the autocovariance at lag m, and we will use the notation γm. Since Cov(Zk;Zk+m) = Cov(Zk+m;Zk), we need only find γm for m 0. The ≥ correlation of Zk and Zk+m is the autocorrelation at lag m. We will use the notation ρm for the autocorrelation. It is easy to show that γk ρk = : γ0 1 2 The autocovariance and autocorrelation ma- trices The covariance matrix for the random variables Z1, :::, Zn is called an auto- covariance matrix.
    [Show full text]
  • Getting Started with Geostatistics
    FedGIS Conference February 24–25, 2016 | Washington, DC Getting Started with Geostatistics Brett Rose Objectives • Using geography as an integrating factor. • Introduce a variety of core tools available in GIS. • Demonstrate the utility of spatial & Geo statistics for a range of applications. • Modeling relationships and driving decisions Outline • Spatial Statistics vs Geostatistics • Geostatistical Workflow • Exploratory Spatial Data Analysis • Interpolation methods • Deterministic vs Geostatistical •Geostatistical Analyst in ArcGIS • Demos • Class Exercise In ArcGIS Geostatistical Analyst • interactive ESDA • interactive modeling including variography • many Kriging models (6 + cokriging) • pre-processing of data - Decluster, detrend, transformation • model diagnostics and comparison Spatial Analyst • rich set of tools to perform cell-based (raster) analysis • kriging (2 models), IDW, nearest neighbor … Spatial Statistics • ESDA GP tools • analyzing the distribution of geographic features • identifying spatial patterns What are spatial statistics in a GIS environment? • Software-based tools, methods, and techniques developed specifically for use with geographic data. • Spatial statistics: – Describe and spatial distributions, spatial patterns, spatial processes model , and spatial relationships. – Incorporate space (area, length, proximity, orientation, and/or spatial relationships) directly into their mathematics. In many ways spatial statistics extenD what the eyes anD minD Do intuitively to assess spatial patterns, trenDs anD relationships.
    [Show full text]
  • Stochastic Process - Introduction
    Stochastic Process - Introduction • Stochastic processes are processes that proceed randomly in time. • Rather than consider fixed random variables X, Y, etc. or even sequences of i.i.d random variables, we consider sequences X0, X1, X2, …. Where Xt represent some random quantity at time t. • In general, the value Xt might depend on the quantity Xt-1 at time t-1, or even the value Xs for other times s < t. • Example: simple random walk . week 2 1 Stochastic Process - Definition • A stochastic process is a family of time indexed random variables Xt where t belongs to an index set. Formal notation, { t : ∈ ItX } where I is an index set that is a subset of R. • Examples of index sets: 1) I = (-∞, ∞) or I = [0, ∞]. In this case Xt is a continuous time stochastic process. 2) I = {0, ±1, ±2, ….} or I = {0, 1, 2, …}. In this case Xt is a discrete time stochastic process. • We use uppercase letter {Xt } to describe the process. A time series, {xt } is a realization or sample function from a certain process. • We use information from a time series to estimate parameters and properties of process {Xt }. week 2 2 Probability Distribution of a Process • For any stochastic process with index set I, its probability distribution function is uniquely determined by its finite dimensional distributions. •The k dimensional distribution function of a process is defined by FXX x,..., ( 1 x,..., k ) = P( Xt ≤ 1 ,..., xt≤ X k) x t1 tk 1 k for anyt 1 ,..., t k ∈ I and any real numbers x1, …, xk .
    [Show full text]
  • Autocovariance Function Estimation Via Penalized Regression
    Autocovariance Function Estimation via Penalized Regression Lina Liao, Cheolwoo Park Department of Statistics, University of Georgia, Athens, GA 30602, USA Jan Hannig Department of Statistics and Operations Research, University of North Carolina, Chapel Hill, NC 27599, USA Kee-Hoon Kang Department of Statistics, Hankuk University of Foreign Studies, Yongin, 449-791, Korea Abstract The work revisits the autocovariance function estimation, a fundamental problem in statistical inference for time series. We convert the function estimation problem into constrained penalized regression with a generalized penalty that provides us with flexible and accurate estimation, and study the asymptotic properties of the proposed estimator. In case of a nonzero mean time series, we apply a penalized regression technique to a differenced time series, which does not require a separate detrending procedure. In penalized regression, selection of tuning parameters is critical and we propose four different data-driven criteria to determine them. A simulation study shows effectiveness of the tuning parameter selection and that the proposed approach is superior to three existing methods. We also briefly discuss the extension of the proposed approach to interval-valued time series. Some key words: Autocovariance function; Differenced time series; Regularization; Time series; Tuning parameter selection. 1 1 Introduction Let fYt; t 2 T g be a stochastic process such that V ar(Yt) < 1, for all t 2 T . The autoco- variance function of fYtg is given as γ(s; t) = Cov(Ys;Yt) for all s; t 2 T . In this work, we consider a regularly sampled time series f(i; Yi); i = 1; ··· ; ng. Its model can be written as Yi = g(i) + i; i = 1; : : : ; n; (1.1) where g is a smooth deterministic function and the error is assumed to be a weakly stationary 2 process with E(i) = 0, V ar(i) = σ and Cov(i; j) = γ(ji − jj) for all i; j = 1; : : : ; n: The estimation of the autocovariance function γ (or autocorrelation function) is crucial to determine the degree of serial correlation in a time series.
    [Show full text]
  • An Introduction to Spatial Autocorrelation and Kriging
    An Introduction to Spatial Autocorrelation and Kriging Matt Robinson and Sebastian Dietrich RenR 690 – Spring 2016 Tobler and Spatial Relationships • Tobler’s 1st Law of Geography: “Everything is related to everything else, but near things are more related than distant things.”1 • Simple but powerful concept • Patterns exist across space • Forms basic foundation for concepts related to spatial dependency Waldo R. Tobler (1) Tobler W., (1970) "A computer movie simulating urban growth in the Detroit region". Economic Geography, 46(2): 234-240. Spatial Autocorrelation (SAC) – What is it? • Autocorrelation: A variable is correlated with itself (literally!) • Spatial Autocorrelation: Values of a random variable, at paired points, are more or less similar as a function of the distance between them 2 Closer Points more similar = Positive Autocorrelation Closer Points less similar = Negative Autocorrelation (2) Legendre P. Spatial Autocorrelation: Trouble or New Paradigm? Ecology. 1993 Sep;74(6):1659–1673. What causes Spatial Autocorrelation? (1) Artifact of Experimental Design (sample sites not random) Parent Plant (2) Interaction of variables across space (see below) Univariate case – response variable is correlated with itself Eg. Plant abundance higher (clustered) close to other plants (seeds fall and germinate close to parent). Multivariate case – interactions of response and predictor variables due to inherent properties of the variables Eg. Location of seed germination function of wind and preferred soil conditions Mechanisms underlying patterns will depend on study system!! Why is it important? Presence of SAC can be good or bad (depends on your objectives) Good: If SAC exists, it may allow reliable estimation at nearby, non-sampled sites (interpolation).
    [Show full text]
  • Master of Science in Geospatial Technologies Geostatistics
    Instituto Superior de Estatística e Gestão de Informação Universidade Nova de Lisboa MasterMaster ofof ScienceScience inin GeospatialGeospatial TechnologiesTechnologies GeostatisticsGeostatistics PredictionsPredictions withwith GeostatisticsGeostatistics CarlosCarlos AlbertoAlberto FelgueirasFelgueiras [email protected]@isegi.unl.pt 1 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics Contents Stochastic Predictions Introduction Deterministic x Stochastic Methods Regionalized Variables Stationary Hypothesis Kriging Prediction – Simple, Ordinary and with Trend Kriging Prediction - Example Cokriging CrossValidation and Validation Problems with Geostatistic Estimators Advantages on using geostatistics 2 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics o sample locations Deterministic x Stochastic Methods z*(u) = K + estimation locations • Deterministic •The z *(u) value is estimated as a Deterministic Variable. An unique value is associated to its spatial location. • No uncertainties are associated to the z*(u)? estimations + • Stochastic •The z *(u) value is considered a Random Variable that has a probability distribution function associated to its possible values • Uncertainties can be associated to the estimations • Important: Samples are realizations of cdf Random Variables 3 Master of Science in Geoespatial Technologies PredictionsPredictions withwith GeostatisticsGeostatistics • Random Variables and Random
    [Show full text]
  • Conditional Simulation
    Basics in Geostatistics 3 Geostatistical Monte-Carlo methods: Conditional simulation Hans Wackernagel MINES ParisTech NERSC • April 2013 http://hans.wackernagel.free.fr Basic concepts Geostatistics Hans Wackernagel (MINES ParisTech) Basics in Geostatistics 3 NERSC • April 2013 2 / 34 Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x). Kriging: Best Linear Unbiased Estimation of point values (or spatial averages) at any location of a region. Conditional simulation: generate an ensemble of realizations of the random function, conditional upon data. Statistics not linearly related to data can be computed from this ensemble. Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x). Kriging: Best Linear Unbiased Estimation of point values (or spatial averages) at any location of a region. Conditional simulation: generate an ensemble of realizations of the random function, conditional upon data. Statistics not linearly related to data can be computed from this ensemble. Concepts Geostatistical model The experimental variogram serves to analyze the spatial structure of a regionalized variable z(x). It is fitted with a variogram model which is the structure function of a random function. The regionalized variable (reality) is viewed as one realization of the random function Z(x).
    [Show full text]
  • An Introduction to Model-Based Geostatistics
    An Introduction to Model-based Geostatistics Peter J Diggle School of Health and Medicine, Lancaster University and Department of Biostatistics, Johns Hopkins University September 2009 Outline • What is geostatistics? • What is model-based geostatistics? • Two examples – constructing an elevation surface from sparse data – tropical disease prevalence mapping Example: data surface elevation Z 6 65 70 75 80 85 90 95100 5 4 Y 3 2 1 0 1 2 3 X 4 5 6 Geostatistics • traditionally, a self-contained methodology for spatial prediction, developed at Ecole´ des Mines, Fontainebleau, France • nowadays, that part of spatial statistics which is concerned with data obtained by spatially discrete sampling of a spatially continuous process Kriging: find the linear combination of the data that best predicts the value of the surface at an arbitrary location x Model-based Geostatistics • the application of general principles of statistical modelling and inference to geostatistical problems – formulate a statistical model for the data – fit the model using likelihood-based methods – use the fitted model to make predictions Kriging: minimum mean square error prediction under Gaussian modelling assumptions Gaussian geostatistics (simplest case) Model 2 • Stationary Gaussian process S(x): x ∈ IR E[S(x)] = µ Cov{S(x), S(x′)} = σ2ρ(kx − x′k) 2 • Mutually independent Yi|S(·) ∼ N(S(x),τ ) Point predictor: Sˆ(x) = E[S(x)|Y ] • linear in Y = (Y1, ..., Yn); • interpolates Y if τ 2 = 0 • called simple kriging in classical geostatistics Predictive distribution • choose the
    [Show full text]
  • Efficient Geostatistical Simulation for Spatial Uncertainty Propagation
    Efficient geostatistical simulation for spatial uncertainty propagation Stelios Liodakis Phaedon Kyriakidis Petros Gaganis University of the Aegean Cyprus University of Technology University of the Aegean University Hill 2-8 Saripolou Str., 3036 University Hill, 81100 Mytilene, Greece Lemesos, Cyprus Mytilene, Greece [email protected] [email protected] [email protected] Abstract Spatial uncertainty and error propagation have received significant attention in GIScience over the last two decades. Uncertainty propagation via Monte Carlo simulation using simple random sampling constitutes the method of choice in GIScience and environmental modeling in general. In this spatial context, geostatistical simulation is often employed for generating simulated realizations of attribute values in space, which are then used as model inputs for uncertainty propagation purposes. In the case of complex models with spatially distributed parameters, however, Monte Carlo simulation could become computationally expensive due to the need for repeated model evaluations; e.g., models involving detailed analytical solutions over large (often three dimensional) discretization grids. A novel simulation method is proposed for generating representative spatial attribute realizations from a geostatistical model, in that these realizations span in a better way (being more dissimilar than what is dictated by change) the range of possible values corresponding to that geostatistical model. It is demonstrated via a synthetic case study, that the simulations produced by the proposed method exhibit much smaller sampling variability, hence better reproduce the statistics of the geostatistical model. In terms of uncertainty propagation, such a closer reproduction of target statistics is shown to yield a better reproduction of model output statistics with respect to simple random sampling using the same number of realizations.
    [Show full text]
  • ON the USE of NON-EUCLIDEAN ISOTROPY in GEOSTATISTICS Frank C
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Collection Of Biostatistics Research Archive Johns Hopkins University, Dept. of Biostatistics Working Papers 12-5-2005 ON THE USE OF NON-EUCLIDEAN ISOTROPY IN GEOSTATISTICS Frank C. Curriero Department of Environmental Health Sciences and Biostatistics, Johns Hopkins Bloomberg School of Public Health, [email protected] Suggested Citation Curriero, Frank C., "ON THE USE OF NON-EUCLIDEAN ISOTROPY IN GEOSTATISTICS" (December 2005). Johns Hopkins University, Dept. of Biostatistics Working Papers. Working Paper 94. http://biostats.bepress.com/jhubiostat/paper94 This working paper is hosted by The Berkeley Electronic Press (bepress) and may not be commercially reproduced without the permission of the copyright holder. Copyright © 2011 by the authors On the Use of Non-Euclidean Isotropy in Geostatistics Frank C. Curriero Department of Environmental Health Sciences and Department of Biostatistics The Johns Hopkins University Bloomberg School of Public Health 615 N. Wolfe Street Baltimore, MD 21205 [email protected] Summary. This paper investigates the use of non-Euclidean distances to characterize isotropic spatial dependence for geostatistical related applications. A simple example is provided to demonstrate there are no guarantees that existing covariogram and variogram functions remain valid (i.e. positive de¯nite or conditionally negative de¯nite) when used with a non-Euclidean distance measure. Furthermore, satisfying the conditions of a metric is not su±cient to ensure the distance measure can be used with existing functions. Current literature is not clear on these topics. There are certain distance measures that when used with existing covariogram and variogram functions remain valid, an issue that is explored.
    [Show full text]
  • Covariances of ARMA Models
    Statistics 910, #9 1 Covariances of ARMA Processes Overview 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation function 5. Discussion Review of ARMA processes ARMA process A stationary solution fXtg (or if its mean is not zero, fXt − µg) of the linear difference equation Xt − φ1Xt−1 − · · · − φpXt−p = wt + θ1wt−1 + ··· + θqwt−q φ(B)Xt = θ(B)wt (1) 2 where wt denotes white noise, wt ∼ WN(0; σ ). Definition 3.5 adds the • identifiability condition that the polynomials φ(z) and θ(z) have no zeros in common and the • normalization condition that φ(0) = θ(0) = 1. Causal process A stationary process fXtg is said to be causal if there ex- ists a summable sequence (some require square summable (`2), others want more and require absolute summability (`1)) sequence f jg such that fXtg has the one-sided moving average representation 1 X Xt = jwt−j = (B)wt : (2) j=0 Proposition 3.1 states that a stationary ARMA process fXtg is causal if and only if (iff) the zeros of the autoregressive polynomial Statistics 910, #9 2 φ(z) lie outside the unit circle (i.e., φ(z) 6= 0 for jzj ≤ 1). Since φ(0) = 1, φ(z) > 0 for jzj ≤ 1. (The unit circle in the complex plane consists of those x 2 C for which jzj = 1; the closed unit disc includes the interior of the unit circle as well as the circle; the open unit disc consists of jzj < 1.) If the zeros of φ(z) lie outside the unit circle, then we can invert each Qp of the factors (1 − B=zj) that make up φ(B) = j=1(1 − B=zj) one at a time (as when back-substituting in the derivation of the AR(1) representation).
    [Show full text]
  • Covariance Functions
    C. E. Rasmussen & C. K. I. Williams, Gaussian Processes for Machine Learning, the MIT Press, 2006, ISBN 026218253X. c 2006 Massachusetts Institute of Technology. www.GaussianProcess.org/gpml Chapter 4 Covariance Functions We have seen that a covariance function is the crucial ingredient in a Gaussian process predictor, as it encodes our assumptions about the function which we wish to learn. From a slightly different viewpoint it is clear that in supervised learning the notion of similarity between data points is crucial; it is a basic similarity assumption that points with inputs x which are close are likely to have similar target values y, and thus training points that are near to a test point should be informative about the prediction at that point. Under the Gaussian process view it is the covariance function that defines nearness or similarity. An arbitrary function of input pairs x and x0 will not, in general, be a valid valid covariance covariance function.1 The purpose of this chapter is to give examples of some functions commonly-used covariance functions and to examine their properties. Section 4.1 defines a number of basic terms relating to covariance functions. Section 4.2 gives examples of stationary, dot-product, and other non-stationary covariance functions, and also gives some ways to make new ones from old. Section 4.3 introduces the important topic of eigenfunction analysis of covariance functions, and states Mercer’s theorem which allows us to express the covariance function (under certain conditions) in terms of its eigenfunctions and eigenvalues. The covariance functions given in section 4.2 are valid when the input domain X is a subset of RD.
    [Show full text]