G/Technology Gaussian Gaussian Process Models in Spatial Data Mining

Total Page:16

File Type:pdf, Size:1020Kb

G/Technology Gaussian Gaussian Process Models in Spatial Data Mining G G/Technology to estimate variances at unsampled locations, aiding in the design of targeted sampling strategies. Intergraph: Real Time Operational Geospatial Applications Historical Background The underlying ideas behind GPs can be traced back to the geostatistics technique called kriging [4], named after Gaussian the South African miner Danie Krige. Kriging in this liter- ature was used to model response variables (e. g., ozone Hurricane Wind Fields, Multivariate Modeling concentrations) over 2D spatial fields as realizations of a stochastic process. Sacks et al. [12] described the use of kriging to model (deterministic) computer experiments. Gaussian Process Models It took more than a decade from this point for the larger in Spatial Data Mining computer science community to investigate GPs for pat- tern analysis purposes. Thus, in the recent past, GPs have NAREN RAMAKRISHNAN1,CHRIS BAILEY-KELLOGG2 witnessed a revival primarily due to work in the statisti- 1 Department of Computer Science, cal pattern recognition community [5] and graphical mod- Virginia Tech, Blacksburg, VA, USA els literature [3]. Neal established the connection between 2 Department of Computer Science, Gaussian processes and neural networks with an infinite Dartmouth College, Hanover, NH, USA number of hidden units [8]. Such relationships allow us to take traditional learning techniques and re-express them Synonyms as imposing a particular covariance structure on the joint distribution of inputs. For instance, we can take a trained Active data mining neural network and mine the covariance structure implied by the weights (given mild assumptions such as a Gaus- Definition sian prior over the weight space). Williams motivates the usefulness of such studies and describes common covari- Gaussian processes (GPs) are local approximation tech- ance functions [14]. Williams and Barber [15] describe niques that model spatial data by placing (and updating) how the Gaussian process framework can be extended to priors on the covariance structures underlying the data. classification in which the modeled variable is categorical. Originally developed for geo-spatial contexts, they are also Since these publications were introduced, interest in GPs applicable in general contexts that involve computing and has exploded with rapid publications in conferences such modeling with multi-level spatial aggregates, e. g., mod- as ICML, NIPS; see also the recently published book by eling a configuration space for crystallographic design, Rasmussen and Williams [11]. casting folding energies as a function of a protein’s con- tact map, and formulation of vaccination policies taking into account social dynamics of individuals. Typically, we Scientific Fundamentals assume a parametrized covariance structure underlying the A GP can be formally defined as a collection of random data to be modeled. We estimate the covariance parameters variables, any finite subset of which have a (multivariate) conditional on the locations for which we have observed normal distribution. For simplicity, we assume 2D spatial- data, and use the inferred structure to make predictions at ly distributed (scalar) response variables ti, one for each new locations. GPs have a probabilistic basis that allow us location xi = [xi1, xi2] where we have collected a data sam- 326 Gaussian Process Models in Spatial Data Mining ple. Observe that, in the limiting case, each random vari- where the last step follows by conditional independence of able has a Gaussian distribution (but it is not true that any {t1, t2,...,tn} w.r.t. xn+1 and the part of Covn+1 not con- collection of Gaussian random variables will induce a GP). tained in Covn. The denominator in the above expression Given a dataset D = {xi, ti}, i = 1... n, and a new data is another Gaussian random variable given by: point xn+1, a GP can be used to model the posterior P(t | D, x ) (which would also be a Gaussian). This P(t ,t ,...,t | x , x ,...,x , Cov ) n+1 n+1 1 2 n 1 2 n n is essentially what many Bayesian modeling techniques do = 1 − 1 −1 T (e. g., least squares approximation with normally distribut- exp [t1,t2,...,tn]Covn [t1,t2,...,tn] . λ2 2 ed noise), however, it is the specifics of how the posterior is modeled that make GPs distinct as a class of modeling Putting it all together, we get: techniques. To make a prediction of tn+1 at a point xn+1, GPs place P(t + |t ,t ,...,t , x , x ,...,x , Cov + ) n 1 1 2 n 1 2 n+1 n 1 greater reliance on ti’s from nearby points. This reliance is λ 2 1 −1 T specified in the form of a covariance prior for the process. = exp − [t ,t ,...,t + ]Cov [t ,t ,...,t + ] λ 2 1 2 n 1 n+1 1 2 n 1 One example of a covariance prior is: 1 1 −1 T ! " − [t ,t ,...,tn]Cov [t ,t ,...,tn] . 2 2 1 2 n 1 2 1 2 Cov(ti,tj) = α exp − a (x − x ) . (1) 2 k ik jk k=1 Computing the mean and variance of this Gaussian distri- Intuitively, this function captures the notion that response bution, we get an estimate of tn+1 as: variables at nearby points must have high correlation. In ˆ T −1 tn+1 = k Cov [t1,t2,...,tn] , (2) Eq. 1, α is an overall scaling term, whereas a1, a2 define n the length scales for the two dimensions. However, this pri- or (or even its posterior) does not directly allow us to deter- and our uncertainty in this estimates as: mine t from t since the structure only captures the covari- j i 2 T −1 σˆ = k − k Cov k , (3) ance; predictions of a response variable for new sample tn+1 n locations are thus conditionally dependent on the measured response variables and their sample locations. Hence, we where kT represents the n-vector of covariances with the must first estimate the covariance parameters (a1, a2,and new data point: α) from D, and then use these parameters along with D to T predict tn+1 at xn+1. k = [Cov(x1, xn+1) Cov(x2, xn+1) ... Cov(xn, xn+1)] , Before covering the learning procedure for the covariance α parameters (a1, a2,and ), it is helpful to develop expres- and k is the (n +1, n + 1) entry of Covn+1. Equations 2 sions for the posterior of the response variable in terms of and 3, together, give us both an approximation at any giv- these parameters. Since the jpdf of the response variables en point and an uncertainty in this approximation; they P(t1, t2,...,tn+1) is modeled Gaussian (we will assume will serve as the basic building blocks for closing-the-loop a mean of zero), we can write: between data modeling and higher level mining function- ality. 1 P(t ,t ,...,t + | x , x ,...,x , + ) = 1 2 n 1 1 2 n+1 Covn 1 λ The above expressions can be alternatively derived by 1 positing a linear probabilistic model and optimizing for 1 · − −1 T the MSE (mean squared error) between observed and pre- exp [t1,t2,...,tn+1]Covn+1 [t1,t2,...,tn+1] 2 dicted response values (e. g., see [12]). In this sense, the Gaussian process model considered here is also known as where we ignore λ1 as it is simply a normalizing factor. the BLUE (best linear unbiased estimator), but GPs are not Here, Covn+1 is the covariance matrix formed from the restricted to linear combinations of basis functions. (n + 1) data values (x1, x2,...,xn+1). A distribution for the To apply GP modeling to a given dataset, one must first unknown variable tn+1 can then be obtained as: ensure that the chosen covariance structure matches the P(tn+1|t1,t2,...,tn, x1, x2,...,xn+1, Covn+1) data characteristics. The above example used a station- ( | ) ary structure which applies when the covariance is trans- = P t1,t2,...,tn+1 x1, x2,...,xn+1, Covn+1 P(t1,t2,...,tn | x1, x2,...,xn+1, Covn+1) lation invariant. Various other functions have been studied in the literature (e. g., see [7,9,12]), all of which satisfy the P(t1,t2,...,tn+1 | x1, x2,...,xn+1, Covn+1) = , required property of positive definiteness of a covariance P(t1,t2,...,tn | x1, x2,...,xn, Covn) Gaussian Process Models in Spatial Data Mining 327 G Gaussian Process Models in Spatial Data Mining, Figure 1 Active mining with Gaussian processes. An initial sample of data points (a;shownasred circles) gives a preliminary approximation to the target function (b). Active sampling suggests new locations (c; blue diamonds) that improve the quality of approximation (d) matrix. The simplest covariance function yields a diago- including a constant term (gives another parameter to be nal matrix, but this means that no data sample can have an estimated) in the covariance formulation. influence on other locations, and the GP approach offers Learning the GP parameters θ = (a1, a2, α) can be under- no particular advantages. In general, by placing a pri- taken in the maximum likelihood (ML) and maximum or directly on the function space, GPs are appropriate a posteriori (MAP) frameworks, or in the true Bayesian for modeling ‘smooth’ functions. The terms a1, a2 cap- setting where we obtain a distribution over values. The log- ture how quickly the influence of a data sample decays likelihood for the parameters is given by: in each direction and, thus, the length scales for smooth- ness. L = log P(t1,t2,...,tn|x1, x2,...,xn, θ) An important point to note is that even though the GP real- n 1 ization is one of a random process, we can nevertheless = c + log P(θ) − log(2π) − log | Covn | 2 2 build a GP model for deterministic functions by choos- 1 −1 T ing a covariance structure that ensures the diagonal cor- − [t1,t2,...,tn]Cov [t1,t2,...,tn] .
Recommended publications
  • Scalable Nonparametric Bayesian Inference on Point Processes with Gaussian Processes
    Scalable Nonparametric Bayesian Inference on Point Processes with Gaussian Processes Yves-Laurent Kom Samo [email protected] Stephen Roberts [email protected] Deparment of Engineering Science and Oxford-Man Institute, University of Oxford Abstract 2. Related Work In this paper we propose an efficient, scalable Non-parametric inference on point processes has been ex- non-parametric Gaussian process model for in- tensively studied in the literature. Rathbum & Cressie ference on Poisson point processes. Our model (1994) and Moeller et al. (1998) used a finite-dimensional does not resort to gridding the domain or to intro- piecewise constant log-Gaussian for the intensity function. ducing latent thinning points. Unlike competing Such approximations are limited in that the choice of the 3 models that scale as O(n ) over n data points, grid on which to represent the intensity function is arbitrary 2 our model has a complexity O(nk ) where k and one has to trade-off precision with computational com- n. We propose a MCMC sampler and show that plexity and numerical accuracy, with the complexity being the model obtained is faster, more accurate and cubic in the precision and exponential in the dimension of generates less correlated samples than competing the input space. Kottas (2006) and Kottas & Sanso (2007) approaches on both synthetic and real-life data. used a Dirichlet process mixture of Beta distributions as Finally, we show that our model easily handles prior for the normalised intensity function of a Poisson data sizes not considered thus far by alternate ap- process. Cunningham et al.
    [Show full text]
  • Deep Neural Networks As Gaussian Processes
    Published as a conference paper at ICLR 2018 DEEP NEURAL NETWORKS AS GAUSSIAN PROCESSES Jaehoon Lee∗y, Yasaman Bahri∗y, Roman Novak , Samuel S. Schoenholz, Jeffrey Pennington, Jascha Sohl-Dickstein Google Brain fjaehlee, yasamanb, romann, schsam, jpennin, [email protected] ABSTRACT It has long been known that a single-layer fully-connected neural network with an i.i.d. prior over its parameters is equivalent to a Gaussian process (GP), in the limit of infinite network width. This correspondence enables exact Bayesian inference for infinite width neural networks on regression tasks by means of evaluating the corresponding GP. Recently, kernel functions which mimic multi-layer random neural networks have been developed, but only outside of a Bayesian framework. As such, previous work has not identified that these kernels can be used as co- variance functions for GPs and allow fully Bayesian prediction with a deep neural network. In this work, we derive the exact equivalence between infinitely wide deep net- works and GPs. We further develop a computationally efficient pipeline to com- pute the covariance function for these GPs. We then use the resulting GPs to per- form Bayesian inference for wide deep neural networks on MNIST and CIFAR- 10. We observe that trained neural network accuracy approaches that of the corre- sponding GP with increasing layer width, and that the GP uncertainty is strongly correlated with trained network prediction error. We further find that test perfor- mance increases as finite-width trained networks are made wider and more similar to a GP, and thus that GP predictions typically outperform those of finite-width networks.
    [Show full text]
  • Crime Mapping and Analysis by Community Organizations
    U.S. Department of Justice Office of Justice Programs National Institute of Justice National Institute of Justice R e s e a r c h i n B r i e f March 2001 Issues and Findings Crime Mapping and Analysis Discussed in this Brief: An as- sessment of how community orga- by Community Organizations nizations in Hartford, Connecticut, used the Neighborhood Problem Solving (NPS) system, a computer- in Hartford, Connecticut based mapping and crime analysis technology. The NPS system en- By Thomas Rich abled users to create a variety of maps and other reports depicting Crime mapping has become increasingly community policing field offices. Devel- crime conditions by accessing a popular among law enforcement agencies oped with NIJ funding, the system enables database containing the most re- and has enjoyed high visibility at the Fed- users to create a variety of maps and other cent 2 years of incident-level police information. eral level, in the media, and among the reports depicting crime conditions by us- largest police departments in the Nation, ing a database containing the most recent Key issues: The objective in Hart- most notably those in Chicago and New 2 years of incident-level police informa- ford was to extend basic mapping York City. However, there have been few tion, including citizen-initiated calls for and crime analysis technologies efforts to make crime mapping capabilities service, reported crimes, and arrests. Al- beyond the law enforcement com- available to community residents and or- though some police departments routinely munity by putting them into the ganizations, although a National Institute publish aggregate crime statistics and a hands of neighborhood-based organizations so they could analyze of Justice (NIJ)-funded project in Chicago few publish incident-level crime informa- incident-level data and produce in the late 1980s aimed at introducing tion on the Internet, the objective in Hart- their own maps and reports.
    [Show full text]
  • BOSTON POLICE DEPARTMENT | BOSTON REGIONAL INTELLIGENCE CENTER Alexa Lambros, Class of 2018
    BOSTON POLICE DEPARTMENT | BOSTON REGIONAL INTELLIGENCE CENTER Alexa Lambros, Class of 2018 The BRIC interns with the BPD Chief of Police! MY ROLE: ASSISTANT CRIME ANALYST/RESEARCH ASSISTANT SPECIAL PROJECT • Mission/Goal Community policing; “working in partnership with the community to fight crime, STANDARD RESPONSIBILITIES reduce fear and improve the quality of life in our • Took raw data from UASI neighborhoods” (BPDnews.com) DAILY INFORMATION INFORMATION crime mapping database of all property stolen from • Size 20th largest LEA in the country; 3rd largest in SUMMARY REQUESTS • Fulfilled information retrieval 9/2015 – 2/2016 New England (Wikipedia) • Assembled packets of recent gang • Utilized Excel skills to • Organization 8 Bureaus and 12 districts incidents and summarized reports requests for Boston/UASI police, correctional, and probation officers organize/categorize data • Focus/Success Utilizes unique combination of • Researched arrests from Urban • Evaulated relationships community relations and statistical analysis to Area Security Initiative (UASI) between numerous elements of reduce crime throughout city; over the last four districts and summarized reports BULLETINS collected data and drew decades, Boston has achieved a large decrease in • Compiled incident summaries on • Created ID Wanted, Wanted, relevant conclusions its overall crime rate (BPDnews.com) violent and property crime within Missing Person, and BOLO bulletins • IMPORTANCE: Created Boston and UASI districts into BRIC's at request of Boston/UASI officers infographic
    [Show full text]
  • Gaussian Process Dynamical Models for Human Motion
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 2, FEBRUARY 2008 283 Gaussian Process Dynamical Models for Human Motion Jack M. Wang, David J. Fleet, Senior Member, IEEE, and Aaron Hertzmann, Member, IEEE Abstract—We introduce Gaussian process dynamical models (GPDMs) for nonlinear time series analysis, with applications to learning models of human pose and motion from high-dimensional motion capture data. A GPDM is a latent variable model. It comprises a low- dimensional latent space with associated dynamics, as well as a map from the latent space to an observation space. We marginalize out the model parameters in closed form by using Gaussian process priors for both the dynamical and the observation mappings. This results in a nonparametric model for dynamical systems that accounts for uncertainty in the model. We demonstrate the approach and compare four learning algorithms on human motion capture data, in which each pose is 50-dimensional. Despite the use of small data sets, the GPDM learns an effective representation of the nonlinear dynamics in these spaces. Index Terms—Machine learning, motion, tracking, animation, stochastic processes, time series analysis. Ç 1INTRODUCTION OOD statistical models for human motion are important models such as hidden Markov model (HMM) and linear Gfor many applications in vision and graphics, notably dynamical systems (LDS) are efficient and easily learned visual tracking, activity recognition, and computer anima- but limited in their expressiveness for complex motions. tion. It is well known in computer vision that the estimation More expressive models such as switching linear dynamical of 3D human motion from a monocular video sequence is systems (SLDS) and nonlinear dynamical systems (NLDS), highly ambiguous.
    [Show full text]
  • Motor Vehicle Theft: Crime and Spatial Analysis in a Non-Urban Region
    The author(s) shown below used Federal funds provided by the U.S. Department of Justice and prepared the following final report: Document Title: Motor Vehicle Theft: Crime and Spatial Analysis in a Non-Urban Region Author(s): Deborah Lamm Weisel ; William R. Smith ; G. David Garson ; Alexi Pavlichev ; Julie Wartell Document No.: 215179 Date Received: August 2006 Award Number: 2003-IJ-CX-0162 This report has not been published by the U.S. Department of Justice. To provide better customer service, NCJRS has made this Federally- funded grant final report available electronically in addition to traditional paper copies. Opinions or points of view expressed are those of the author(s) and do not necessarily reflect the official position or policies of the U.S. Department of Justice. This document is a research report submitted to the U.S. Department of Justice. This report has not been published by the Department. Opinions or points of view expressed are those of the author(s) and do not necessarily reflect the official position or policies of the U.S. Department of Justice. EXECUTIVE SUMMARY Crime and Spatial Analysis of Vehicle Theft in a Non-Urban Region Motor vehicle theft in non-urban areas does not reveal the well-recognized hot spots often associated with crime in urban areas. These findings resulted from a study of 2003 vehicle thefts in a four-county region of western North Carolina comprised primarily of small towns and unincorporated areas. While the study suggested that point maps have limited value for areas with low volume and geographically-dispersed crime, the steps necessary to create regional maps – including collecting and validating crime locations with Global Positioning System (GPS) coordinates – created a reliable dataset that permitted more in-depth analysis.
    [Show full text]
  • Modelling Multi-Object Activity by Gaussian Processes 1
    LOY et al.: MODELLING MULTI-OBJECT ACTIVITY BY GAUSSIAN PROCESSES 1 Modelling Multi-object Activity by Gaussian Processes Chen Change Loy School of EECS [email protected] Queen Mary University of London Tao Xiang E1 4NS London, UK [email protected] Shaogang Gong [email protected] Abstract We present a new approach for activity modelling and anomaly detection based on non-parametric Gaussian Process (GP) models. Specifically, GP regression models are formulated to learn non-linear relationships between multi-object activity patterns ob- served from semantically decomposed regions in complex scenes. Predictive distribu- tions are inferred from the regression models to compare with the actual observations for real-time anomaly detection. The use of a flexible, non-parametric model alleviates the difficult problem of selecting appropriate model complexity encountered in parametric models such as Dynamic Bayesian Networks (DBNs). Crucially, our GP models need fewer parameters; they are thus less likely to overfit given sparse data. In addition, our approach is robust to the inevitable noise in activity representation as noise is modelled explicitly in the GP models. Experimental results on a public traffic scene show that our models outperform DBNs in terms of anomaly sensitivity, noise robustness, and flexibil- ity in modelling complex activity. 1 Introduction Activity modelling and automatic anomaly detection in video have received increasing at- tention due to the recent large-scale deployments of surveillance cameras. These tasks are non-trivial because complex activity patterns in a busy public space involve multiple objects interacting with each other over space and time, whilst anomalies are often rare, ambigu- ous and can be easily confused with noise caused by low image quality, unstable lighting condition and occlusion.
    [Show full text]
  • Financial Time Series Volatility Analysis Using Gaussian Process State-Space Models
    Financial Time Series Volatility Analysis Using Gaussian Process State-Space Models by Jianan Han Bachelor of Engineering, Hebei Normal University, China, 2010 A thesis presented to Ryerson University in partial fulfillment of the requirements for the degree of Master of Applied Science in the Program of Electrical and Computer Engineering Toronto, Ontario, Canada, 2015 c Jianan Han 2015 AUTHOR'S DECLARATION FOR ELECTRONIC SUBMISSION OF A THESIS I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I authorize Ryerson University to lend this thesis to other institutions or individuals for the purpose of scholarly research. I further authorize Ryerson University to reproduce this thesis by photocopying or by other means, in total or in part, at the request of other institutions or individuals for the purpose of scholarly research. I understand that my dissertation may be made electronically available to the public. ii Financial Time Series Volatility Analysis Using Gaussian Process State-Space Models Master of Applied Science 2015 Jianan Han Electrical and Computer Engineering Ryerson University Abstract In this thesis, we propose a novel nonparametric modeling framework for financial time series data analysis, and we apply the framework to the problem of time varying volatility modeling. Existing parametric models have a rigid transition function form and they often have over-fitting problems when model parameters are estimated using maximum likelihood methods. These drawbacks effect the models' forecast performance. To solve this problem, we take Bayesian nonparametric modeling approach.
    [Show full text]
  • Gaussian-Random-Process.Pdf
    The Gaussian Random Process Perhaps the most important continuous state-space random process in communications systems in the Gaussian random process, which, we shall see is very similar to, and shares many properties with the jointly Gaussian random variable that we studied previously (see lecture notes and chapter-4). X(t); t 2 T is a Gaussian r.p., if, for any positive integer n, any choice of coefficients ak; 1 k n; and any choice ≤ ≤ of sample time tk ; 1 k n; the random variable given by the following weighted sum of random variables is Gaussian:2 T ≤ ≤ X(t) = a1X(t1) + a2X(t2) + ::: + anX(tn) using vector notation we can express this as follows: X(t) = [X(t1);X(t2);:::;X(tn)] which, according to our study of jointly Gaussian r.v.s is an n-dimensional Gaussian r.v.. Hence, its pdf is known since we know the pdf of a jointly Gaussian random variable. For the random process, however, there is also the nasty little parameter t to worry about!! The best way to see the connection to the Gaussian random variable and understand the pdf of a random process is by example: Example: Determining the Distribution of a Gaussian Process Consider a continuous time random variable X(t); t with continuous state-space (in this case amplitude) defined by the following expression: 2 R X(t) = Y1 + tY2 where, Y1 and Y2 are independent Gaussian distributed random variables with zero mean and variance 2 2 σ : Y1;Y2 N(0; σ ). The problem is to find the one and two dimensional probability density functions of the random! process X(t): The one-dimensional distribution of a random process, also known as the univariate distribution is denoted by the notation FX;1(u; t) and defined as: P r X(t) u .
    [Show full text]
  • Crime Mapping: Spatial and Temporal Challenges
    CHAPTER 2 Crime Mapping: Spatial and Temporal Challenges JERRY RATCLIFFE INTRODUCTION Crime opportunities are neither uniformly nor randomly organized in space and time. As a result, crime mappers can unlock these spatial patterns and strive for a better theoretical understanding of the role of geography and opportunity, as well as enabling practical crime prevention solutions that are tailored to specific places. The evolution of crime mapping has heralded a new era in spatial criminology, and a re-emergence of the importance of place as one of the cornerstones essential to an understanding of crime and criminality. While early criminological inquiry in France and Britain had a spatial component, much of mainstream criminology for the last century has labored to explain criminality from a dispositional per- spective, trying to explain why a particular offender or group has a propensity to commit crime. This traditional perspective resulted in criminologists focusing on individuals or on communities where the community extended from the neighborhood to larger aggregations (Weisburd et al. 2004). Even when the results lacked ambiguity, the findings often lacked policy relevance. However, crime mapping has revived interest and reshaped many criminol- ogists appreciation for the importance of local geography as a determinant of crime that may be as important as criminal motivation. Between the individual and large urban areas (such as cities and regions) lies a spatial scale where crime varies considerably and does so at a frame of reference that is often amenable to localized crime prevention techniques. For example, without the opportunity afforded by enabling environmental weaknesses, such as poorly lit streets, lack of protective surveillance, or obvious victims (such as overtly wealthy tourists or unsecured vehicles), many offenders would not be as encouraged to commit crime.
    [Show full text]
  • Gaussian Markov Processes
    C. E. Rasmussen & C. K. I. Williams, Gaussian Processes for Machine Learning, the MIT Press, 2006, ISBN 026218253X. c 2006 Massachusetts Institute of Technology. www.GaussianProcess.org/gpml Appendix B Gaussian Markov Processes Particularly when the index set for a stochastic process is one-dimensional such as the real line or its discretization onto the integer lattice, it is very interesting to investigate the properties of Gaussian Markov processes (GMPs). In this Appendix we use X(t) to define a stochastic process with continuous time pa- rameter t. In the discrete time case the process is denoted ...,X−1,X0,X1,... etc. We assume that the process has zero mean and is, unless otherwise stated, stationary. A discrete-time autoregressive (AR) process of order p can be written as AR process p X Xt = akXt−k + b0Zt, (B.1) k=1 where Zt ∼ N (0, 1) and all Zt’s are i.i.d. Notice the order-p Markov property that given the history Xt−1,Xt−2,..., Xt depends only on the previous p X’s. This relationship can be conveniently expressed as a graphical model; part of an AR(2) process is illustrated in Figure B.1. The name autoregressive stems from the fact that Xt is predicted from the p previous X’s through a regression equation. If one stores the current X and the p − 1 previous values as a state vector, then the AR(p) scalar process can be written equivalently as a vector AR(1) process. Figure B.1: Graphical model illustrating an AR(2) process.
    [Show full text]
  • FULLY BAYESIAN FIELD SLAM USING GAUSSIAN MARKOV RANDOM FIELDS Huan N
    Asian Journal of Control, Vol. 18, No. 5, pp. 1–14, September 2016 Published online in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/asjc.1237 FULLY BAYESIAN FIELD SLAM USING GAUSSIAN MARKOV RANDOM FIELDS Huan N. Do, Mahdi Jadaliha, Mehmet Temel, and Jongeun Choi ABSTRACT This paper presents a fully Bayesian way to solve the simultaneous localization and spatial prediction problem using a Gaussian Markov random field (GMRF) model. The objective is to simultaneously localize robotic sensors and predict a spatial field of interest using sequentially collected noisy observations by robotic sensors. The set of observations consists of the observed noisy positions of robotic sensing vehicles and noisy measurements of a spatial field. To be flexible, the spatial field of interest is modeled by a GMRF with uncertain hyperparameters. We derive an approximate Bayesian solution to the problem of computing the predictive inferences of the GMRF and the localization, taking into account observations, uncertain hyperparameters, measurement noise, kinematics of robotic sensors, and uncertain localization. The effectiveness of the proposed algorithm is illustrated by simulation results as well as by experiment results. The experiment results successfully show the flexibility and adaptability of our fully Bayesian approach ina data-driven fashion. Key Words: Vision-based localization, spatial modeling, simultaneous localization and mapping (SLAM), Gaussian process regression, Gaussian Markov random field. I. INTRODUCTION However, there are a number of inapplicable situa- tions. For example, underwater autonomous gliders for Simultaneous localization and mapping (SLAM) ocean sampling cannot find usual geometrical models addresses the problem of a robot exploring an unknown from measurements of environmental variables such as environment under localization uncertainty [1].
    [Show full text]