Chapter 9 Random Processes

Total Page:16

File Type:pdf, Size:1020Kb

Chapter 9 Random Processes Henry Stark and John W. Woods, Probability, Statistics, and Random Variables for Engineers, 4th ed., Pearson Education Inc., 2012. ISBN: 978-0-13-231123-6 Chapter 9 Random Processes Sections 9.1 Basic Definitions 544 9.2 Some Important Random Processes 548 9.3 Continuous-Time Linear Systems with Random Inputs 572 White Noise 577 9.4 Some Useful Classifications of Random Processes 578 Stationarity 579 9.5 Wide-Sense Stationary Processes and LSI Systems 581 Wide-Sense Stationary Case 582 Power Spectral Density 584 An Interpretation of the Power Spectral Density 586 More on White Noise 590 Stationary Processes and Differential Equations 596 9.6 Periodic and Cyclostationary Processes 600 9.7 Vector Processes and State Equations 606 State Equations 608 Summary 611 Problems 611 References 633 Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 1 of 78 ECE 3800 9.1 Basic Concepts A random process is a collection of time functions and an associated probability description. When a continuous or discrete or mixed process in time/space can be describe mathematically as a function containing one or more random variables. A sinusoidal waveform with a random amplitude. A sinusoidal waveform with a random phase. A sequence of digital symbols, each taking on a random value for a defined time period (e.g. amplitude, phase, frequency). A random walk (2-D or 3-D movement of a particle) The entire collection of possible time functions is an ensemble, designated as xt , where one particular member of the ensemble, designated as xt, is a sample function of the ensemble. In general only one sample function of a random process can be observed! Think of: X t Asinwt , 0 2 where A and w are known constants. Note that once a sample has been observed … x1 t1 Asinwt1 the function is known for all time, t. Note that, xt2 is a second time sample of the same random process and does not provide any “new information” about the value of the random variable. x1 t2 Asinwt2 There are many similar ensembles in engineering, where the sample function, once known, provides a continuing solution. In many cases, an entire system design approach is based on either assuming that randomness remains or is removed once actual measurements are taken! For example, in communications there is a significant difference between coherent (phase and frequency) demodulation and non-coherent (i.e. unknown starting phase) demodulation. On the other hand, another measurement in a different environment might measure x2 t1 A2 sinwt1 2 In this “space” the random variables could take on other values within the defined ranges. Thus an entire “ensemble” of possibilities may exist based on the random variables defined in the random process. Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 2 of 78 ECE 3800 For example, assume that there is a known AM signal transmitted: st 1 b Atsinwt at an undetermined distance the signal is received as yt 1 b At sinwt , 0 2 The received signal is mixed and low pass filtered … xt h t y t cos wt ht 1 b Atsinwt coswt,0 2 xt h t y t cos wt ht 1 b At 0.5sin2 wt sin ,0 2 If the filter removes the 2wt term, we have 1 b At xt h t y t cos w t sin , 0 2 2 Notice that based on the value of the random variable, the output can change significantly! From producing no output signal, ( 0, ), to having the output be positive or negative ( 0to or to 2 ). P.S. This is not how you perform non-coherent AM demodulation. To perform coherent AM demodulation, all I need to do is measured the value of the random variable and use it to insure that the output is a maximum (i.e. mix with coswt m , where. m t1 Note: the phase is a function of frequency, time, and distance from the transmitter. Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 3 of 78 ECE 3800 From our textbook Random Stochastic Sequence Definition 8.1-1. Let ,,P be a probability space. Let . Let Xn,be a mapping of the sample space into a space of complex-valued sequences on some index set Z. If, for each fixed integer n Z , Xn, is a random variable, then Xn, is a ransom (stochastic) sequence. The index set Z is all integers, n , padded with zeros if necessary, Definition 9.1-1. Let ,,P be a probability space. then define a mapping of X from the sample space to a space of continuous time functions. The elements in this space will be called sample functions. This mapping is called a random process if at each fixed time the mapping is a random variable, that is, Xt, for each fixed t on the real line t . Example sets of random sequence. Figure 8.1-1 Illustration of the concept of random sequence X(n,ζ), where the ζ domain (i.e., the sample space Ω) consists of just ten values. (Samples connected only for plot.) Example sets of random process. Figure 9.1-1 A random process for a continuous sample space Ω = [0,10]. Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 4 of 78 ECE 3800 Example 9.1-2 Separable random process may be constructed by combining a deterministic sequence with one or more random variables. The classic example already shown is a sinusoid with random amplitude and phase: X t, A sin2 f 0 t Where the amplitude and phase are R.V. defined based on the probability space selected. Example 9.1‐3 A random process used to model a continuous sequence of random communication symbols. X t An pt Tn n In a communication class, Dr. Bazuin would typically use the following X t An p t n T , for pt non zero for 0 t k T n Here An is the amplitude and phase of a complex communication symbol and p(t) is the deterministic time function, the simplest of which is a rectangular pulse in time. This can be used to describe a wide range of digital communication systems, including; Phase- Shift Keyed (PSK) or Quadrature Amplitude Modulation (QAM) communication signals. Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 5 of 78 ECE 3800 The application of the Expected Value Operator Moments play an important role and, for Ergodic Processes, they can be estimated from a single process in time of the infinite number that may be possible. Therefore, X t EX t and the correlation functions (auto- and cross-correlation) * RXX t1,t2 EX t1 X t2 * RXY t1,t2 EX t1 Yt2 and the covariance functions (auto- and cross-correlation) * K XX t1,t2 EX t1 X t1 X t2 X t2 * K XY t1,t2 EX t1 X t1 Yt2 Y t2 with * K XX t1,t2 RXX t1,t2 X t1 X t2 Note that the variance can be computed from the auto-covariance as * 2 K XX t,t EX t X t X t X t X t and the “power” function can be computed from the auto-correlation * 2 RXX t,t EX t X t EX t 2 2 2 For real X(t) RXX t,t EX t X t X t Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W. Woods, Pearson Education, Inc., 2012. B.J. Bazuin, Fall 2016 6 of 78 ECE 3800 Example 9.1‐5 Auto‐correlation of a sinusoid with random phase Think of: X t Asinwt , where A and w are known constants. And theta is a uniform pdf covering the unit circle. The mean is computed as X t EX t EAsinwt X t EX t A Esinwt 1 t E X t A sin wt d X 2 A t EX t coswt X 2 A A t EX t coswt cos wt 0 0 X 2 2 ( What would happen if 0 instead? ) The auto-correlation is computed as * * RXX t1,t2 EX t1 X t2 EAsinwt1 Asinwt2 R t ,t E X t X t * A2 E 1 cos w t t 1 cos w t t 2 XX 1 2 1 2 2 1 2 2 1 2 A2 A2 R t ,t cosw t t Ecosw t t 2 XX 1 2 2 1 2 2 1 2 A2 A2 R t ,t cosw t t 0 cosw t t XX 1 2 2 1 2 2 1 2 ( This works if 0 instead. ) Note that if A was a random variable (independent of phase) we would have … EA2 EA2 R t ,t cosw t t R cosw XX 1 2 2 1 2 XX 2 and we would still have EA t EX t 0 0 X 2 Note: this Random Process is Wide-Sense stationary (mean and variance not a function of time) Notes and figures are based on or taken from materials in the course textbook: Probability, Statistics and Random Processes for Engineers, 4th ed., Henry Stark and John W.
Recommended publications
  • Stationary Processes and Their Statistical Properties
    Stationary Processes and Their Statistical Properties Brian Borchers March 29, 2001 1 Stationary processes A discrete time stochastic process is a sequence of random variables Z1, Z2, :::. In practice we will typically analyze a single realization z1, z2, :::, zn of the stochastic process and attempt to esimate the statistical properties of the stochastic process from the realization. We will also consider the problem of predicting zn+1 from the previous elements of the sequence. We will begin by focusing on the very important class of stationary stochas- tic processes. A stochastic process is strictly stationary if its statistical prop- erties are unaffected by shifting the stochastic process in time. In particular, this means that if we take a subsequence Zk+1, :::, Zk+m, then the joint distribution of the m random variables will be the same no matter what k is. Stationarity requires that the mean of the stochastic process be a constant. E[Zk] = µ. and that the variance is constant 2 V ar[Zk] = σZ : Also, stationarity requires that the covariance of two elements separated by a distance m is constant. That is, Cov(Zk;Zk+m) is constant. This covariance is called the autocovariance at lag m, and we will use the notation γm. Since Cov(Zk;Zk+m) = Cov(Zk+m;Zk), we need only find γm for m 0. The ≥ correlation of Zk and Zk+m is the autocorrelation at lag m. We will use the notation ρm for the autocorrelation. It is easy to show that γk ρk = : γ0 1 2 The autocovariance and autocorrelation ma- trices The covariance matrix for the random variables Z1, :::, Zn is called an auto- covariance matrix.
    [Show full text]
  • Methods of Monte Carlo Simulation II
    Methods of Monte Carlo Simulation II Ulm University Institute of Stochastics Lecture Notes Dr. Tim Brereton Summer Term 2014 Ulm, 2014 2 Contents 1 SomeSimpleStochasticProcesses 7 1.1 StochasticProcesses . 7 1.2 RandomWalks .......................... 7 1.2.1 BernoulliProcesses . 7 1.2.2 RandomWalks ...................... 10 1.2.3 ProbabilitiesofRandomWalks . 13 1.2.4 Distribution of Xn .................... 13 1.2.5 FirstPassageTime . 14 2 Estimators 17 2.1 Bias, Variance, the Central Limit Theorem and Mean Square Error................................ 19 2.2 Non-AsymptoticErrorBounds. 22 2.3 Big O and Little o Notation ................... 23 3 Markov Chains 25 3.1 SimulatingMarkovChains . 28 3.1.1 Drawing from a Discrete Uniform Distribution . 28 3.1.2 Drawing From A Discrete Distribution on a Small State Space ........................... 28 3.1.3 SimulatingaMarkovChain . 28 3.2 Communication .......................... 29 3.3 TheStrongMarkovProperty . 30 3.4 RecurrenceandTransience . 31 3.4.1 RecurrenceofRandomWalks . 33 3.5 InvariantDistributions . 34 3.6 LimitingDistribution. 36 3.7 Reversibility............................ 37 4 The Poisson Process 39 4.1 Point Processes on [0, )..................... 39 ∞ 3 4 CONTENTS 4.2 PoissonProcess .......................... 41 4.2.1 Order Statistics and the Distribution of Arrival Times 44 4.2.2 DistributionofArrivalTimes . 45 4.3 SimulatingPoissonProcesses. 46 4.3.1 Using the Infinitesimal Definition to Simulate Approx- imately .......................... 46 4.3.2 SimulatingtheArrivalTimes . 47 4.3.3 SimulatingtheInter-ArrivalTimes . 48 4.4 InhomogenousPoissonProcesses. 48 4.5 Simulating an Inhomogenous Poisson Process . 49 4.5.1 Acceptance-Rejection. 49 4.5.2 Infinitesimal Approach (Approximate) . 50 4.6 CompoundPoissonProcesses . 51 5 ContinuousTimeMarkovChains 53 5.1 TransitionFunction. 53 5.2 InfinitesimalGenerator . 54 5.3 ContinuousTimeMarkovChains .
    [Show full text]
  • Power Comparisons of Five Most Commonly Used Autocorrelation Tests
    Pak.j.stat.oper.res. Vol.16 No. 1 2020 pp119-130 DOI: http://dx.doi.org/10.18187/pjsor.v16i1.2691 Pakistan Journal of Statistics and Operation Research Power Comparisons of Five Most Commonly Used Autocorrelation Tests Stanislaus S. Uyanto School of Economics and Business, Atma Jaya Catholic University of Indonesia, [email protected] Abstract In regression analysis, autocorrelation of the error terms violates the ordinary least squares assumption that the error terms are uncorrelated. The consequence is that the estimates of coefficients and their standard errors will be wrong if the autocorrelation is ignored. There are many tests for autocorrelation, we want to know which test is more powerful. We use Monte Carlo methods to compare the power of five most commonly used tests for autocorrelation, namely Durbin-Watson, Breusch-Godfrey, Box–Pierce, Ljung Box, and Runs tests in two different linear regression models. The results indicate the Durbin-Watson test performs better in the regression model without lagged dependent variable, although the advantage over the other tests reduce with increasing autocorrelation and sample sizes. For the model with lagged dependent variable, the Breusch-Godfrey test is generally superior to the other tests. Key Words: Correlated error terms; Ordinary least squares assumption; Residuals; Regression diagnostic; Lagged dependent variable. Mathematical Subject Classification: 62J05; 62J20. 1. Introduction An important assumption of the classical linear regression model states that there is no autocorrelation in the error terms. Autocorrelation of the error terms violates the ordinary least squares assumption that the error terms are uncorrelated, meaning that the Gauss Markov theorem (Plackett, 1949, 1950; Greene, 2018) does not apply, and that ordinary least squares estimators are no longer the best linear unbiased estimators.
    [Show full text]
  • LECTURES 2 - 3 : Stochastic Processes, Autocorrelation Function
    LECTURES 2 - 3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series fXtg is a series of observations taken sequentially over time: xt is an observation recorded at a specific time t. Characteristics of times series data: observations are dependent, become available at equally spaced time points and are time-ordered. This is a discrete time series. The purposes of time series analysis are to model and to predict or forecast future values of a series based on the history of that series. 2.2 Some descriptive techniques. (Based on [BD] x1.3 and x1.4) ......................................................................................... Take a step backwards: how do we describe a r.v. or a random vector? ² for a r.v. X: 2 d.f. FX (x) := P (X · x), mean ¹ = EX and variance σ = V ar(X). ² for a r.vector (X1;X2): joint d.f. FX1;X2 (x1; x2) := P (X1 · x1;X2 · x2), marginal d.f.FX1 (x1) := P (X1 · x1) ´ FX1;X2 (x1; 1) 2 2 mean vector (¹1; ¹2) = (EX1; EX2), variances σ1 = V ar(X1); σ2 = V ar(X2), and covariance Cov(X1;X2) = E(X1 ¡ ¹1)(X2 ¡ ¹2) ´ E(X1X2) ¡ ¹1¹2. Often we use correlation = normalized covariance: Cor(X1;X2) = Cov(X1;X2)=fσ1σ2g ......................................................................................... To describe a process X1;X2;::: we define (i) Def. Distribution function: (fi-di) d.f. Ft1:::tn (x1; : : : ; xn) = P (Xt1 · x1;:::;Xtn · xn); i.e. this is the joint d.f. for the vector (Xt1 ;:::;Xtn ). (ii) First- and Second-order moments. ² Mean: ¹X (t) = EXt 2 2 2 2 ² Variance: σX (t) = E(Xt ¡ ¹X (t)) ´ EXt ¡ ¹X (t) 1 ² Autocovariance function: γX (t; s) = Cov(Xt;Xs) = E[(Xt ¡ ¹X (t))(Xs ¡ ¹X (s))] ´ E(XtXs) ¡ ¹X (t)¹X (s) (Note: this is an infinite matrix).
    [Show full text]
  • Examples of Stationary Time Series
    Statistics 910, #2 1 Examples of Stationary Time Series Overview 1. Stationarity 2. Linear processes 3. Cyclic models 4. Nonlinear models Stationarity Strict stationarity (Defn 1.6) Probability distribution of the stochastic process fXtgis invariant under a shift in time, P (Xt1 ≤ x1;Xt2 ≤ x2;:::;Xtk ≤ xk) = F (xt1 ; xt2 ; : : : ; xtk ) = F (xh+t1 ; xh+t2 ; : : : ; xh+tk ) = P (Xh+t1 ≤ x1;Xh+t2 ≤ x2;:::;Xh+tk ≤ xk) for any time shift h and xj. Weak stationarity (Defn 1.7) (aka, second-order stationarity) The mean and autocovariance of the stochastic process are finite and invariant under a shift in time, E Xt = µt = µ Cov(Xt;Xs) = E (Xt−µt)(Xs−µs) = γ(t; s) = γ(t−s) The separation rather than location in time matters. Equivalence If the process is Gaussian with finite second moments, then weak stationarity is equivalent to strong stationarity. Strict stationar- ity implies weak stationarity only if the necessary moments exist. Relevance Stationarity matters because it provides a framework in which averaging makes sense. Unless properties like the mean and covariance are either fixed or \evolve" in a known manner, we cannot average the observed data. What operations produce a stationary process? Can we recognize/identify these in data? Statistics 910, #2 2 Moving Average White noise Sequence of uncorrelated random variables with finite vari- ance, ( 2 often often σw = 1 if t = s; E Wt = µ = 0 Cov(Wt;Ws) = 0 otherwise The input component (fXtg in what follows) is often modeled as white noise. Strict white noise replaces uncorrelated by independent. Moving average A stochastic process formed by taking a weighted average of another time series, often formed from white noise.
    [Show full text]
  • 5. the Student T Distribution
    Virtual Laboratories > 4. Special Distributions > 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 5. The Student t Distribution In this section we will study a distribution that has special importance in statistics. In particular, this distribution will arise in the study of a standardized version of the sample mean when the underlying distribution is normal. The Probability Density Function Suppose that Z has the standard normal distribution, V has the chi-squared distribution with n degrees of freedom, and that Z and V are independent. Let Z T= √V/n In the following exercise, you will show that T has probability density function given by −(n +1) /2 Γ((n + 1) / 2) t2 f(t)= 1 + , t∈ℝ ( n ) √n π Γ(n / 2) 1. Show that T has the given probability density function by using the following steps. n a. Show first that the conditional distribution of T given V=v is normal with mean 0 a nd variance v . b. Use (a) to find the joint probability density function of (T,V). c. Integrate the joint probability density function in (b) with respect to v to find the probability density function of T. The distribution of T is known as the Student t distribution with n degree of freedom. The distribution is well defined for any n > 0, but in practice, only positive integer values of n are of interest. This distribution was first studied by William Gosset, who published under the pseudonym Student. In addition to supplying the proof, Exercise 1 provides a good way of thinking of the t distribution: the t distribution arises when the variance of a mean 0 normal distribution is randomized in a certain way.
    [Show full text]
  • Alternative Tests for Time Series Dependence Based on Autocorrelation Coefficients
    Alternative Tests for Time Series Dependence Based on Autocorrelation Coefficients Richard M. Levich and Rosario C. Rizzo * Current Draft: December 1998 Abstract: When autocorrelation is small, existing statistical techniques may not be powerful enough to reject the hypothesis that a series is free of autocorrelation. We propose two new and simple statistical tests (RHO and PHI) based on the unweighted sum of autocorrelation and partial autocorrelation coefficients. We analyze a set of simulated data to show the higher power of RHO and PHI in comparison to conventional tests for autocorrelation, especially in the presence of small but persistent autocorrelation. We show an application of our tests to data on currency futures to demonstrate their practical use. Finally, we indicate how our methodology could be used for a new class of time series models (the Generalized Autoregressive, or GAR models) that take into account the presence of small but persistent autocorrelation. _______________________________________________________________ An earlier version of this paper was presented at the Symposium on Global Integration and Competition, sponsored by the Center for Japan-U.S. Business and Economic Studies, Stern School of Business, New York University, March 27-28, 1997. We thank Paul Samuelson and the other participants at that conference for useful comments. * Stern School of Business, New York University and Research Department, Bank of Italy, respectively. 1 1. Introduction Economic time series are often characterized by positive
    [Show full text]
  • 1 One Parameter Exponential Families
    1 One parameter exponential families The world of exponential families bridges the gap between the Gaussian family and general dis- tributions. Many properties of Gaussians carry through to exponential families in a fairly precise sense. • In the Gaussian world, there exact small sample distributional results (i.e. t, F , χ2). • In the exponential family world, there are approximate distributional results (i.e. deviance tests). • In the general setting, we can only appeal to asymptotics. A one-parameter exponential family, F is a one-parameter family of distributions of the form Pη(dx) = exp (η · t(x) − Λ(η)) P0(dx) for some probability measure P0. The parameter η is called the natural or canonical parameter and the function Λ is called the cumulant generating function, and is simply the normalization needed to make dPη fη(x) = (x) = exp (η · t(x) − Λ(η)) dP0 a proper probability density. The random variable t(X) is the sufficient statistic of the exponential family. Note that P0 does not have to be a distribution on R, but these are of course the simplest examples. 1.0.1 A first example: Gaussian with linear sufficient statistic Consider the standard normal distribution Z e−z2=2 P0(A) = p dz A 2π and let t(x) = x. Then, the exponential family is eη·x−x2=2 Pη(dx) / p 2π and we see that Λ(η) = η2=2: eta= np.linspace(-2,2,101) CGF= eta**2/2. plt.plot(eta, CGF) A= plt.gca() A.set_xlabel(r'$\eta$', size=20) A.set_ylabel(r'$\Lambda(\eta)$', size=20) f= plt.gcf() 1 Thus, the exponential family in this setting is the collection F = fN(η; 1) : η 2 Rg : d 1.0.2 Normal with quadratic sufficient statistic on R d As a second example, take P0 = N(0;Id×d), i.e.
    [Show full text]
  • 3 Autocorrelation
    3 Autocorrelation Autocorrelation refers to the correlation of a time series with its own past and future values. Autocorrelation is also sometimes called “lagged correlation” or “serial correlation”, which refers to the correlation between members of a series of numbers arranged in time. Positive autocorrelation might be considered a specific form of “persistence”, a tendency for a system to remain in the same state from one observation to the next. For example, the likelihood of tomorrow being rainy is greater if today is rainy than if today is dry. Geophysical time series are frequently autocorrelated because of inertia or carryover processes in the physical system. For example, the slowly evolving and moving low pressure systems in the atmosphere might impart persistence to daily rainfall. Or the slow drainage of groundwater reserves might impart correlation to successive annual flows of a river. Or stored photosynthates might impart correlation to successive annual values of tree-ring indices. Autocorrelation complicates the application of statistical tests by reducing the effective sample size. Autocorrelation can also complicate the identification of significant covariance or correlation between time series (e.g., precipitation with a tree-ring series). Autocorrelation implies that a time series is predictable, probabilistically, as future values are correlated with current and past values. Three tools for assessing the autocorrelation of a time series are (1) the time series plot, (2) the lagged scatterplot, and (3) the autocorrelation function. 3.1 Time series plot Positively autocorrelated series are sometimes referred to as persistent because positive departures from the mean tend to be followed by positive depatures from the mean, and negative departures from the mean tend to be followed by negative departures (Figure 3.1).
    [Show full text]
  • Random Variables and Probability Distributions 1.1
    RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS 1. DISCRETE RANDOM VARIABLES 1.1. Definition of a Discrete Random Variable. A random variable X is said to be discrete if it can assume only a finite or countable infinite number of distinct values. A discrete random variable can be defined on both a countable or uncountable sample space. 1.2. Probability for a discrete random variable. The probability that X takes on the value x, P(X=x), is defined as the sum of the probabilities of all sample points in Ω that are assigned the value x. We may denote P(X=x) by p(x) or pX (x). The expression pX (x) is a function that assigns probabilities to each possible value x; thus it is often called the probability function for the random variable X. 1.3. Probability distribution for a discrete random variable. The probability distribution for a discrete random variable X can be represented by a formula, a table, or a graph, which provides pX (x) = P(X=x) for all x. The probability distribution for a discrete random variable assigns nonzero probabilities to only a countable number of distinct x values. Any value x not explicitly assigned a positive probability is understood to be such that P(X=x) = 0. The function pX (x)= P(X=x) for each x within the range of X is called the probability distribution of X. It is often called the probability mass function for the discrete random variable X. 1.4. Properties of the probability distribution for a discrete random variable.
    [Show full text]
  • Spatial Autocorrelation: Covariance and Semivariance Semivariance
    Spatial Autocorrelation: Covariance and Semivariancence Lily Housese ­P eters GEOG 593 November 10, 2009 Quantitative Terrain Descriptorsrs Covariance and Semivariogram areare numericc methods used to describe the character of the terrain (ex. Terrain roughness, irregularity) Terrain character has important implications for: 1. the type of sampling strategy chosen 2. estimating DTM accuracy (after sampling and reconstruction) Spatial Autocorrelationon The First Law of Geography ““ Everything is related to everything else, but near things are moo re related than distant things.” (Waldo Tobler) The value of a variable at one point in space is related to the value of that same variable in a nearby location Ex. Moranan ’s I, Gearyary ’s C, LISA Positive Spatial Autocorrelation (Neighbors are similar) Negative Spatial Autocorrelation (Neighbors are dissimilar) R(d) = correlation coefficient of all the points with horizontal interval (d) Covariance The degree of similarity between pairs of surface points The value of similarity is an indicator of the complexity of the terrain surface Smaller the similarity = more complex the terrain surface V = Variance calculated from all N points Cov (d) = Covariance of all points with horizontal interval d Z i = Height of point i M = average height of all points Z i+d = elevation of the point with an interval of d from i Semivariancee Expresses the degree of relationship between points on a surface Equal to half the variance of the differences between all possible points spaced a constant distance apart
    [Show full text]
  • Stationary Processes
    Stationary processes Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania [email protected] http://www.seas.upenn.edu/users/~aribeiro/ November 25, 2019 Stoch. Systems Analysis Stationary processes 1 Stationary stochastic processes Stationary stochastic processes Autocorrelation function and wide sense stationary processes Fourier transforms Linear time invariant systems Power spectral density and linear filtering of stochastic processes Stoch. Systems Analysis Stationary processes 2 Stationary stochastic processes I All probabilities are invariant to time shits, i.e., for any s P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2;:::; X (tK + s) ≥ xK ] = P[X (t1) ≥ x1; X (t2) ≥ x2;:::; X (tK ) ≥ xK ] I If above relation is true process is called strictly stationary (SS) I First order stationary ) probs. of single variables are shift invariant P[X (t + s) ≥ x] = P [X (t) ≥ x] I Second order stationary ) joint probs. of pairs are shift invariant P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2] = P [X (t1) ≥ x1; X (t2) ≥ x2] Stoch. Systems Analysis Stationary processes 3 Pdfs and moments of stationary process I For SS process joint cdfs are shift invariant. Whereby, pdfs also are fX (t+s)(x) = fX (t)(x) = fX (0)(x) := fX (x) I As a consequence, the mean of a SS process is constant Z 1 Z 1 µ(t) := E [X (t)] = xfX (t)(x) = xfX (x) = µ −∞ −∞ I The variance of a SS process is also constant Z 1 Z 1 2 2 2 var [X (t)] := (x − µ) fX (t)(x) = (x − µ) fX (x) = σ −∞ −∞ I The power of a SS process (second moment) is also constant Z 1 Z 1 2 2 2 2 2 E X (t) := x fX (t)(x) = x fX (x) = σ + µ −∞ −∞ Stoch.
    [Show full text]