Methods of Monte Carlo Simulation II

Total Page:16

File Type:pdf, Size:1020Kb

Methods of Monte Carlo Simulation II Methods of Monte Carlo Simulation II Ulm University Institute of Stochastics Lecture Notes Dr. Tim Brereton Summer Term 2014 Ulm, 2014 2 Contents 1 SomeSimpleStochasticProcesses 7 1.1 StochasticProcesses . 7 1.2 RandomWalks .......................... 7 1.2.1 BernoulliProcesses . 7 1.2.2 RandomWalks ...................... 10 1.2.3 ProbabilitiesofRandomWalks . 13 1.2.4 Distribution of Xn .................... 13 1.2.5 FirstPassageTime . 14 2 Estimators 17 2.1 Bias, Variance, the Central Limit Theorem and Mean Square Error................................ 19 2.2 Non-AsymptoticErrorBounds. 22 2.3 Big O and Little o Notation ................... 23 3 Markov Chains 25 3.1 SimulatingMarkovChains . 28 3.1.1 Drawing from a Discrete Uniform Distribution . 28 3.1.2 Drawing From A Discrete Distribution on a Small State Space ........................... 28 3.1.3 SimulatingaMarkovChain . 28 3.2 Communication .......................... 29 3.3 TheStrongMarkovProperty . 30 3.4 RecurrenceandTransience . 31 3.4.1 RecurrenceofRandomWalks . 33 3.5 InvariantDistributions . 34 3.6 LimitingDistribution. 36 3.7 Reversibility............................ 37 4 The Poisson Process 39 4.1 Point Processes on [0, )..................... 39 ∞ 3 4 CONTENTS 4.2 PoissonProcess .......................... 41 4.2.1 Order Statistics and the Distribution of Arrival Times 44 4.2.2 DistributionofArrivalTimes . 45 4.3 SimulatingPoissonProcesses. 46 4.3.1 Using the Infinitesimal Definition to Simulate Approx- imately .......................... 46 4.3.2 SimulatingtheArrivalTimes . 47 4.3.3 SimulatingtheInter-ArrivalTimes . 48 4.4 InhomogenousPoissonProcesses. 48 4.5 Simulating an Inhomogenous Poisson Process . 49 4.5.1 Acceptance-Rejection. 49 4.5.2 Infinitesimal Approach (Approximate) . 50 4.6 CompoundPoissonProcesses . 51 5 ContinuousTimeMarkovChains 53 5.1 TransitionFunction. 53 5.2 InfinitesimalGenerator . 54 5.3 ContinuousTimeMarkovChains . 54 5.4 TheJumpChainandHoldingTimes . 55 5.5 Examples of Continuous Time Markov Chains . 56 5.5.1 PoissonProcess . 56 5.5.2 Birth-DeathProcess . 56 5.6 Simulating Continuous Time Markov Chains . 56 5.7 The Relationship Between P and Q intheFiniteCase . 58 5.8 Irreducibility, Recurrence and Positive Recurrence . ..... 59 5.9 Invariant Measures and Stationary Distribution . .. 61 5.9.1 ReversibilityandDetailedBalance . 62 6 Gaussian Processes 63 6.1 TheMultivariateNormalDistribution. 63 6.1.1 Symmetric Positive Definite and Semi-Positive Definite Matrices.......................... 65 6.1.2 DensitiesofMultivariateNormals . 66 6.1.3 SimulatingMultivariateNormals . 67 6.2 SimulatingaGaussianProcessesVersion1 . 68 6.3 Stationary and Weak Stationary Gaussian Processes . .. 71 6.4 FiniteDimensionalDistributions . 71 6.5 Marginal and Conditional Multivariate Normal Distributions . 72 6.6 InterpolatingGaussianProcesses . 73 6.7 MarkovianGaussianProcesses . 75 6.8 BrownianMotion ......................... 77 CONTENTS 5 7 Random Fields 81 7.1 GaussianRandomFields . 81 7.2 MarkovRandomFields. 85 7.3 GaussianRandomMarkovFields . 86 6 CONTENTS Chapter 1 Some Simple Stochastic Processes 1.1 Stochastic Processes To begin, we need to define the basic objects we will be learning about. Definition 1.1.1 (Stochastic Process). A stochastic process is a set of ran- dom variables X , taking values in a state space , with index sex I R. { i}i∈I X ⊂ In general, i represents a point in time. However, it could represent a point in 1D space as well. We will say a process is discrete time if I is discrete. For example, I could be N or 1, 2, 3, 10, 20 . Normally, when we talk about discrete time { } stochastic processes, we will use the index n (e.g., X N). { n}n∈ We will say a process is continuous time if I is an interval. For example [0, ) or [1, 2]. Normally, when we talk about continuous time stochastic processes,∞ we will use the index t (e.g. X ). { t}t∈[0,T ] 1.2 Random Walks 1.2.1 Bernoulli Processes One of the simplest stochastic processes is a random walk. However, even though the random walk is very simple, it has a number of properties that will be important when we think about more complicated processes. To define a random walk, we begin with an even simpler process called a Bernoulli process. A Bernoulli process is a discrete time process taking values in the state space 0, 1 . { } 7 8 CHAPTER 1. SIMPLE STOCHASTIC PROCESSES ETC. Definition 1.2.1 (Bernoulli Process). A Bernoulli process with parameter p [0, 1] is a sequence of independent and identically distributed (i.i.d.) ran- dom∈ variables, Y , such that { n}n≥1 P(Y =1)=1 P(Y =0)= p. (1.1) 1 − 1 A variable with the probability mass function (pmf) described by (1.1) is called a Bernoulli random variable with distribution Ber(p). In the case where p = 1/2, you can think of a Bernoulli process as a sequence of fair coin tosses. In the case where p = 1/2, you can think of a Bernoulli process as a sequence of tosses of an unfair6 coin. Note that if U U(0, 1), then P(U p) = p for p [0, 1]. This means if we define the random∼ variable Y = I(≤U p), where∈I( ) is the indicator function and U U(0, 1), we have ≤ · ∼ P(Y =1)=1 P(Y =0)= p. − In Matlab, we can make these variables as follows. Listing 1.1: Generating a Bernoulli Random Variable 1 Y = (rand <= p) If we want to make a realization of the first n steps of a Bernoulli process, we simply make n such variables. One way to do this is the following. Listing 1.2: Generating a Bernoulli Process 1 1 n = 20; p = 0.6; Y = zeros(N,1); 2 3 for i = 1:n 4 Y(i) = (rand <= p); 5 end We can do this more quickly in Matlab though. Listing 1.3: Generating a Bernoulli Process 2 1 n = 20; p = 0.6; 2 3 Y = (rand(n,1) <= p); It is important to be able to visualize stochastic objects. One way to represent / visualize a Bernoulli process is to put n =1, 2,... on the x-axis and the values of Yn on the y-axis. I draw a line (almost of length 1) at each value of Yn, as this is easier to see than dots. Figure 1.2.1 shows a realization of the first 20 steps of a Bernoulli process. 1.2. RANDOM WALKS 9 Listing 1.4: Generating and Visualizing a Bernoulli Process 1 n = 20; p = 0.6; 2 3 Y = (rand(n,1) <= p); 4 5 clf; 6 axis([1 n+1 -0.5 1.5]); 7 for i = 1:n 8 line([i, i+.9],[Y(i) Y(i)]); 9 end 1 n Y 0 2 4 6 8 10 n 12 14 16 18 20 Figure 1.2.1: A realization of the first 20 steps of a Bernoulli process. Now, it is easy to write out the probability of seeing a particular realiza- tion of n steps of a Bernoulli process. This is given by P(Y = y ,Y = y ,...,Y = y )= pNU (1 p)n−NU . n n n−1 n−1 1 1 − where NU is the number of times the Bernoulli process takes the value 1. More technically, we define NU as N =# 0 i n : y =1 , U { ≤ ≤ i } where # denotes the cardinality of the set. 10 CHAPTER 1. SIMPLE STOCHASTIC PROCESSES ETC. 1.2.2 Random Walks Now that we can generate Bernoulli processes, we are ready to consider our main object of interest in this chapter: the random walk. The random walk is a discrete time random process taking values in the state space Z. Definition 1.2.2 (Random Walk). Given a Bernoulli process Yn n≥1 with parameter p, we define a random walk X with parameter{p and} initial { n}n≥0 condition X0 = x0 by X = X + (2Y 1). n+1 n n+1 − Note that (2Yn 1) is 1 if Yn = 1 and 1 if Yn = 0. So, essentially, at each step of the process,− it goes up or down− by 1. There are lots of ways to think about random walks. One way is in terms of gambling (which is a classical setting for probability). Think of a game which is free to enter. The person running the game flips a coin. With probability p, the coin shows heads and you win AC1. With probability 1 p, − you lose AC1. If you start with ACx0, and assuming you can go into debt, Xn is the random variable describing how much money you have after n games. Given that we know how to generate a Bernoulli process, it is straight- forward to generate a random walk. Listing 1.5: Generating a Random Walk 1 1 n = 100; X = zeros(n,1); 2 p = 0.5; X_0 = 0; 3 4 Y = (rand(n,1) <= p); 5 X(1) = X_0 + 2*Y(1) - 1; 6 for i = 2:n 7 X(i) = X(i-1) + 2*Y(i) - 1; 8 end A more compact way is as follows. Listing 1.6: Generating a Random Walk 2 1 n = 100; X = zeros(n,1); 2 p = 0.5; X_0 = 0; 3 4 Y = (rand(n,1) <= p); 5 X = X_0 + cumsum((2*Y - 1)); It is good to be able to plot these. Here, we have at least two options. One is to draw a line for each value of Xn, as we did for the Bernoulli process. 1.2. RANDOM WALKS 11 Because the value of Xn n≥0 changes at every value of n, we can draw the lines to be length 1. This{ approach} is nice, because it reinforces the idea that Xn n≥0 jumps at each value of n (this is made obvious by the gaps between the{ lines).} Listing 1.7: Plotting a Random Walk 1 1 clf 2 axis([1 n+1 min(X)-1 max(X)+1]); 3 for i = 1:n 4 line([i,i+1],[X(i) X(i)],’LineWidth’,3); 5 end For larger values of n, it is help to draw vertical lines as well as horizontal lines (otherwise, things get a bit hard to see).
Recommended publications
  • Stationary Processes and Their Statistical Properties
    Stationary Processes and Their Statistical Properties Brian Borchers March 29, 2001 1 Stationary processes A discrete time stochastic process is a sequence of random variables Z1, Z2, :::. In practice we will typically analyze a single realization z1, z2, :::, zn of the stochastic process and attempt to esimate the statistical properties of the stochastic process from the realization. We will also consider the problem of predicting zn+1 from the previous elements of the sequence. We will begin by focusing on the very important class of stationary stochas- tic processes. A stochastic process is strictly stationary if its statistical prop- erties are unaffected by shifting the stochastic process in time. In particular, this means that if we take a subsequence Zk+1, :::, Zk+m, then the joint distribution of the m random variables will be the same no matter what k is. Stationarity requires that the mean of the stochastic process be a constant. E[Zk] = µ. and that the variance is constant 2 V ar[Zk] = σZ : Also, stationarity requires that the covariance of two elements separated by a distance m is constant. That is, Cov(Zk;Zk+m) is constant. This covariance is called the autocovariance at lag m, and we will use the notation γm. Since Cov(Zk;Zk+m) = Cov(Zk+m;Zk), we need only find γm for m 0. The ≥ correlation of Zk and Zk+m is the autocorrelation at lag m. We will use the notation ρm for the autocorrelation. It is easy to show that γk ρk = : γ0 1 2 The autocovariance and autocorrelation ma- trices The covariance matrix for the random variables Z1, :::, Zn is called an auto- covariance matrix.
    [Show full text]
  • 12 : Conditional Random Fields 1 Hidden Markov Model
    10-708: Probabilistic Graphical Models 10-708, Spring 2014 12 : Conditional Random Fields Lecturer: Eric P. Xing Scribes: Qin Gao, Siheng Chen 1 Hidden Markov Model 1.1 General parametric form In hidden Markov model (HMM), we have three sets of parameters, j i transition probability matrix A : p(yt = 1jyt−1 = 1) = ai;j; initialprobabilities : p(y1) ∼ Multinomial(π1; π2; :::; πM ); i emission probabilities : p(xtjyt) ∼ Multinomial(bi;1; bi;2; :::; bi;K ): 1.2 Inference k k The inference can be done with forward algorithm which computes αt ≡ µt−1!t(k) = P (x1; :::; xt−1; xt; yt = 1) recursively by k k X i αt = p(xtjyt = 1) αt−1ai;k; (1) i k k and the backward algorithm which computes βt ≡ µt t+1(k) = P (xt+1; :::; xT jyt = 1) recursively by k X i i βt = ak;ip(xt+1jyt+1 = 1)βt+1: (2) i Another key quantity is the conditional probability of any hidden state given the entire sequence, which can be computed by the dot product of forward message and backward message by, i i i i X i;j γt = p(yt = 1jx1:T ) / αtβt = ξt ; (3) j where we define, i;j i j ξt = p(yt = 1; yt−1 = 1; x1:T ); i j / µt−1!t(yt = 1)µt t+1(yt+1 = 1)p(xt+1jyt+1)p(yt+1jyt); i j i = αtβt+1ai;jp(xt+1jyt+1 = 1): The implementation in Matlab can be vectorized by using, i Bt(i) = p(xtjyt = 1); j i A(i; j) = p(yt+1 = 1jyt = 1): 1 2 12 : Conditional Random Fields The relation of those quantities can be simply written in pseudocode as, T αt = (A αt−1): ∗ Bt; βt = A(βt+1: ∗ Bt+1); T ξt = (αt(βt+1: ∗ Bt+1) ): ∗ A; γt = αt: ∗ βt: 1.3 Learning 1.3.1 Supervised Learning The supervised learning is trivial if only we know the true state path.
    [Show full text]
  • Examples of Stationary Time Series
    Statistics 910, #2 1 Examples of Stationary Time Series Overview 1. Stationarity 2. Linear processes 3. Cyclic models 4. Nonlinear models Stationarity Strict stationarity (Defn 1.6) Probability distribution of the stochastic process fXtgis invariant under a shift in time, P (Xt1 ≤ x1;Xt2 ≤ x2;:::;Xtk ≤ xk) = F (xt1 ; xt2 ; : : : ; xtk ) = F (xh+t1 ; xh+t2 ; : : : ; xh+tk ) = P (Xh+t1 ≤ x1;Xh+t2 ≤ x2;:::;Xh+tk ≤ xk) for any time shift h and xj. Weak stationarity (Defn 1.7) (aka, second-order stationarity) The mean and autocovariance of the stochastic process are finite and invariant under a shift in time, E Xt = µt = µ Cov(Xt;Xs) = E (Xt−µt)(Xs−µs) = γ(t; s) = γ(t−s) The separation rather than location in time matters. Equivalence If the process is Gaussian with finite second moments, then weak stationarity is equivalent to strong stationarity. Strict stationar- ity implies weak stationarity only if the necessary moments exist. Relevance Stationarity matters because it provides a framework in which averaging makes sense. Unless properties like the mean and covariance are either fixed or \evolve" in a known manner, we cannot average the observed data. What operations produce a stationary process? Can we recognize/identify these in data? Statistics 910, #2 2 Moving Average White noise Sequence of uncorrelated random variables with finite vari- ance, ( 2 often often σw = 1 if t = s; E Wt = µ = 0 Cov(Wt;Ws) = 0 otherwise The input component (fXtg in what follows) is often modeled as white noise. Strict white noise replaces uncorrelated by independent. Moving average A stochastic process formed by taking a weighted average of another time series, often formed from white noise.
    [Show full text]
  • Stationary Processes
    Stationary processes Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania [email protected] http://www.seas.upenn.edu/users/~aribeiro/ November 25, 2019 Stoch. Systems Analysis Stationary processes 1 Stationary stochastic processes Stationary stochastic processes Autocorrelation function and wide sense stationary processes Fourier transforms Linear time invariant systems Power spectral density and linear filtering of stochastic processes Stoch. Systems Analysis Stationary processes 2 Stationary stochastic processes I All probabilities are invariant to time shits, i.e., for any s P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2;:::; X (tK + s) ≥ xK ] = P[X (t1) ≥ x1; X (t2) ≥ x2;:::; X (tK ) ≥ xK ] I If above relation is true process is called strictly stationary (SS) I First order stationary ) probs. of single variables are shift invariant P[X (t + s) ≥ x] = P [X (t) ≥ x] I Second order stationary ) joint probs. of pairs are shift invariant P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2] = P [X (t1) ≥ x1; X (t2) ≥ x2] Stoch. Systems Analysis Stationary processes 3 Pdfs and moments of stationary process I For SS process joint cdfs are shift invariant. Whereby, pdfs also are fX (t+s)(x) = fX (t)(x) = fX (0)(x) := fX (x) I As a consequence, the mean of a SS process is constant Z 1 Z 1 µ(t) := E [X (t)] = xfX (t)(x) = xfX (x) = µ −∞ −∞ I The variance of a SS process is also constant Z 1 Z 1 2 2 2 var [X (t)] := (x − µ) fX (t)(x) = (x − µ) fX (x) = σ −∞ −∞ I The power of a SS process (second moment) is also constant Z 1 Z 1 2 2 2 2 2 E X (t) := x fX (t)(x) = x fX (x) = σ + µ −∞ −∞ Stoch.
    [Show full text]
  • Solutions to Exercises in Stationary Stochastic Processes for Scientists and Engineers by Lindgren, Rootzén and Sandsten Chapman & Hall/CRC, 2013
    Solutions to exercises in Stationary stochastic processes for scientists and engineers by Lindgren, Rootzén and Sandsten Chapman & Hall/CRC, 2013 Georg Lindgren, Johan Sandberg, Maria Sandsten 2017 CENTRUM SCIENTIARUM MATHEMATICARUM Faculty of Engineering Centre for Mathematical Sciences Mathematical Statistics 1 Solutions to exercises in Stationary stochastic processes for scientists and engineers Mathematical Statistics Centre for Mathematical Sciences Lund University Box 118 SE-221 00 Lund, Sweden http://www.maths.lu.se c Georg Lindgren, Johan Sandberg, Maria Sandsten, 2017 Contents Preface v 2 Stationary processes 1 3 The Poisson process and its relatives 5 4 Spectral representations 9 5 Gaussian processes 13 6 Linear filters – general theory 17 7 AR, MA, and ARMA-models 21 8 Linear filters – applications 25 9 Frequency analysis and spectral estimation 29 iii iv CONTENTS Preface This booklet contains hints and solutions to exercises in Stationary stochastic processes for scientists and engineers by Georg Lindgren, Holger Rootzén, and Maria Sandsten, Chapman & Hall/CRC, 2013. The solutions have been adapted from course material used at Lund University on first courses in stationary processes for students in engineering programs as well as in mathematics, statistics, and science programs. The web page for the course during the fall semester 2013 gives an example of a schedule for a seven week period: http://www.maths.lu.se/matstat/kurser/fms045mas210/ Note that the chapter references in the material from the Lund University course do not exactly agree with those in the printed volume. v vi CONTENTS Chapter 2 Stationary processes 2:1. (a) 1, (b) a + b, (c) 13, (d) a2 + b2, (e) a2 + b2, (f) 1.
    [Show full text]
  • MCMC Learning (Slides)
    Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions MCMC Learning Varun Kanade Elchanan Mossel UC Berkeley UC Berkeley August 30, 2013 Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Outline Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Uniform Distribution Learning • Unknown target function f : {−1; 1gn ! {−1; 1g from some class C • Uniform distribution over {−1; 1gn • Random Examples: Monotone Decision Trees [OS06] • Random Walk: DNF expressions [BMOS03] • Membership Query: DNF, TOP [J95] • Main Tool: Discrete Fourier Analysis X Y f (x) = f^(S)χS (x); χS (x) = xi S⊆[n] i2S • Can utilize sophisticated results: hypercontractivity, invariance, etc. • Connections to cryptography, hardness, de-randomization etc. • Unfortunately, too much of an idealization. In practice, variables are correlated. Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Markov Random Fields • Graph G = ([n]; E). Each node takes some value in finite set A. n • Distribution over A : (for φC non-negative, Z normalization constant) 1 Y Pr((σ ) ) = φ ((σ ) ) v v2[n] Z C v v2C clique C Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Markov Random Fields • MRFs widely used in vision, computational biology, biostatistics etc. • Extensive Algorithmic Theory for sampling from MRFs, recovering parameters and structures • Learning Question: Given f : An ! {−1; 1g. (How) Can we learn with respect to MRF distribution? • Can we utilize the structure of the MRF to aid in learning? Uniform Distribution Learning Markov Random Fields Harmonic Analysis Experiments and Questions Learning Model • Let M be a MRF with distribution π and f : An ! {−1; 1g the target function • Learning algorithm gets i.i.d.
    [Show full text]
  • Random Processes
    Chapter 6 Random Processes Random Process • A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). • For a fixed (sample path): a random process is a time varying function, e.g., a signal. – For fixed t: a random process is a random variable. • If one scans all possible outcomes of the underlying random experiment, we shall get an ensemble of signals. • Random Process can be continuous or discrete • Real random process also called stochastic process – Example: Noise source (Noise can often be modeled as a Gaussian random process. An Ensemble of Signals Remember: RV maps Events à Constants RP maps Events à f(t) RP: Discrete and Continuous The set of all possible sample functions {v(t, E i)} is called the ensemble and defines the random process v(t) that describes the noise source. Sample functions of a binary random process. RP Characterization • Random variables x 1 , x 2 , . , x n represent amplitudes of sample functions at t 5 t 1 , t 2 , . , t n . – A random process can, therefore, be viewed as a collection of an infinite number of random variables: RP Characterization – First Order • CDF • PDF • Mean • Mean-Square Statistics of a Random Process RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function of timeà We use second order estimation RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function
    [Show full text]
  • Markov Random Fields: – Geman, Geman, “Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images”, IEEE PAMI 6, No
    MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Conglin Lu Medical Image Display and Analysis Group University of North Carolina, Chapel Hill MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Outline: Background; Definition and properties of MRF; Computation; MRF m-reps models. MarkovMarkov RandomRandom FieldsFields Model a large collection of random variables with complex dependency relationships among them. MarkovMarkov RandomRandom FieldsFields • A model based approach; • Has been applied to a variety of problems: - Speech recognition - Natural language processing - Coding - Image analysis - Neural networks - Artificial intelligence • Usually used within the Bayesian framework. TheThe BayesianBayesian ParadigmParadigm X = space of the unknown variables, e.g. labels; Y = space of data (observations), e.g. intensity values; Given an observation y∈Y, want to make inference about x∈X. TheThe BayesianBayesian ParadigmParadigm Prior PX : probability distribution on X; Likelihood PY|X : conditional distribution of Y given X; Statistical inference is based on the posterior distribution PX|Y ∝ PX •PY|X . TheThe PriorPrior DistributionDistribution • Describes our assumption or knowledge about the model; • X is usually a high dimensional space. PX describes the joint distribution of a large number of random variables; • How do we define PX? MarkovMarkov RandomRandom FieldsFields withwith ApplicationsApplications toto MM--repsreps ModelsModels Outline: 9 Background; Definition and properties of MRF; Computation; MRF m-reps models. AssumptionsAssumptions •X = {Xs}s∈S, where each Xs is a random variable; S is an index set and is finite; • There is a common state space R:Xs∈R for all s ∈ S; | R | is finite; • Let Ω = {ω=(x , ..., x ): x ∈ , 1≤i≤N} be s1 sN si R the set of all possible configurations.
    [Show full text]
  • Anomaly Detection Based on Wavelet Domain GARCH Random Field Modeling Amir Noiboar and Israel Cohen, Senior Member, IEEE
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 45, NO. 5, MAY 2007 1361 Anomaly Detection Based on Wavelet Domain GARCH Random Field Modeling Amir Noiboar and Israel Cohen, Senior Member, IEEE Abstract—One-dimensional Generalized Autoregressive Con- Markov noise. It is also claimed that objects in imagery create a ditional Heteroscedasticity (GARCH) model is widely used for response over several scales in a multiresolution representation modeling financial time series. Extending the GARCH model to of an image, and therefore, the wavelet transform can serve as multiple dimensions yields a novel clutter model which is capable of taking into account important characteristics of a wavelet-based a means for computing a feature set for input to a detector. In multiscale feature space, namely heavy-tailed distributions and [17], a multiscale wavelet representation is utilized to capture innovations clustering as well as spatial and scale correlations. We periodical patterns of various period lengths, which often ap- show that the multidimensional GARCH model generalizes the pear in natural clutter images. In [12], the orientation and scale casual Gauss Markov random field (GMRF) model, and we de- selectivity of the wavelet transform are related to the biological velop a multiscale matched subspace detector (MSD) for detecting anomalies in GARCH clutter. Experimental results demonstrate mechanisms of the human visual system and are utilized to that by using a multiscale MSD under GARCH clutter modeling, enhance mammographic features. rather than GMRF clutter modeling, a reduced false-alarm rate Statistical models for clutter and anomalies are usually can be achieved without compromising the detection rate. related to the Gaussian distribution due to its mathematical Index Terms—Anomaly detection, Gaussian Markov tractability.
    [Show full text]
  • 4.2 Autoregressive (AR) Moving Average Models Are Causal Linear Processes by Definition. There Is Another Class of Models, Based
    4.2 Autoregressive (AR) Moving average models are causal linear processes by definition. There is another class of models, based on a recursive formulation similar to the exponentially weighted moving average. Definition 4.11 (Autoregressive AR(p)). Suppose φ1,...,φp ∈ R are constants 2 2 and (Wi) ∼ WN(σ ). The AR(p) process with parameters σ , φ1,...,φp is defined through p Xi = Wi + φjXi−j, (3) Xj=1 whenever such stationary process (Xi) exists. Remark 4.12. The process in Definition 4.11 is sometimes called a stationary AR(p) process. It is possible to consider a ‘non-stationary AR(p) process’ for any φ1,...,φp satisfying (3) for i ≥ 0 by letting for example Xi = 0 for i ∈ [−p+1, 0]. Example 4.13 (Variance and autocorrelation of AR(1) process). For the AR(1) process, whenever it exits, we must have 2 2 γ0 = Var(Xi) = Var(φ1Xi−1 + Wi)= φ1γ0 + σ , which implies that we must have |φ1| < 1, and σ2 γ0 = 2 . 1 − φ1 We may also calculate for j ≥ 1 j γj = E[XiXi−j]= E[(φ1Xi−1 + Wi)Xi−j]= φ1E[Xi−1Xi−j]= φ1γ0, j which gives that ρj = φ1. Example 4.14. Simulation of an AR(1) process. phi_1 <- 0.7 x <- arima.sim(model=list(ar=phi_1), 140) # This is the explicit simulation: gamma_0 <- 1/(1-phi_1^2) x_0 <- rnorm(1)*sqrt(gamma_0) x <- filter(rnorm(140), phi_1, method = "r", init = x_0) Example 4.15. Consider a stationary AR(1) process. We may write n−1 n j Xi = φ1Xi−1 + Wi = ··· = φ1 Xi−n + φ1Wi−j.
    [Show full text]
  • A Novel Approach for Markov Random Field with Intractable Normalising Constant on Large Lattices
    A novel approach for Markov Random Field with intractable normalising constant on large lattices W. Zhu ∗ and Y. Fany February 19, 2018 Abstract The pseudo likelihood method of Besag (1974), has remained a popular method for estimating Markov random field on a very large lattice, despite various documented deficiencies. This is partly because it remains the only computationally tractable method for large lattices. We introduce a novel method to estimate Markov random fields defined on a regular lattice. The method takes advantage of conditional independence structures and recur- sively decomposes a large lattice into smaller sublattices. An approximation is made at each decomposition. Doing so completely avoids the need to com- pute the troublesome normalising constant. The computational complexity arXiv:1601.02410v1 [stat.ME] 11 Jan 2016 is O(N), where N is the the number of pixels in lattice, making it computa- tionally attractive for very large lattices. We show through simulation, that the proposed method performs well, even when compared to the methods using exact likelihoods. Keywords: Markov random field, normalizing constant, conditional indepen- dence, decomposition, Potts model. ∗School of Mathematics and Statistics, University of New South Wales, Sydney 2052 Australia. Communicating Author Wanchuang Zhu: Email [email protected]. ySchool of Mathematics and Statistics, University of New South Wales, Sydney 2052 Australia. Email [email protected]. 1 1 Introduction Markov random field (MRF) models have an important role in modelling spa- tially correlated datasets. They have been used extensively in image and texture analyses ( Nott and Ryden´ 1999, Hurn et al. 2003), image segmentation (Pal and Pal 1993, Van Leemput et al.
    [Show full text]
  • Interacting Particle Systems MA4H3
    Interacting particle systems MA4H3 Stefan Grosskinsky Warwick, 2009 These notes and other information about the course are available on http://www.warwick.ac.uk/˜masgav/teaching/ma4h3.html Contents Introduction 2 1 Basic theory3 1.1 Continuous time Markov chains and graphical representations..........3 1.2 Two basic IPS....................................6 1.3 Semigroups and generators.............................9 1.4 Stationary measures and reversibility........................ 13 2 The asymmetric simple exclusion process 18 2.1 Stationary measures and conserved quantities................... 18 2.2 Currents and conservation laws........................... 23 2.3 Hydrodynamics and the dynamic phase transition................. 26 2.4 Open boundaries and matrix product ansatz.................... 30 3 Zero-range processes 34 3.1 From ASEP to ZRPs................................ 34 3.2 Stationary measures................................. 36 3.3 Equivalence of ensembles and relative entropy................... 39 3.4 Phase separation and condensation......................... 43 4 The contact process 46 4.1 Mean field rate equations.............................. 46 4.2 Stochastic monotonicity and coupling....................... 48 4.3 Invariant measures and critical values....................... 51 d 4.4 Results for Λ = Z ................................. 54 1 Introduction Interacting particle systems (IPS) are models for complex phenomena involving a large number of interrelated components. Examples exist within all areas of natural and social sciences, such as traffic flow on highways, pedestrians or constituents of a cell, opinion dynamics, spread of epi- demics or fires, reaction diffusion systems, crystal surface growth, chemotaxis, financial markets... Mathematically the components are modeled as particles confined to a lattice or some discrete geometry. Their motion and interaction is governed by local rules. Often microscopic influences are not accesible in full detail and are modeled as effective noise with some postulated distribution.
    [Show full text]