The Consistency of the Subjective Concept of Randomness

Total Page:16

File Type:pdf, Size:1020Kb

The Consistency of the Subjective Concept of Randomness Quarterly Journal of Experimental Psychology The consistency of the subjective concept of randomness Journal: Quarterly Journal of Experimental Psychology Manuscript ID Draft Manuscript Type: Standard Article Date Submitted by the Author: n/a Complete List of Authors: Yu, Ru Qi; University of British Columbia, Department of Psychology Gunn, Jordan; Princeton University Osherson, Dan; Princeton University, Zhao, Jiaying; University of British Columbia, Department of Psychology Keywords: randomness, modalities, alternation bias, anchoring effect, features URL: http:/mc.manuscriptcentral.com/pqje Page 1 of 18 Quarterly Journal of Experimental Psychology RUNNING HEAD: CONSISTENCY OF RANDOMNESS 1 1 2 3 4 5 6 7 8 9 10 11 12 The consistency of the subjective concept of randomness 13 14 15 Ru Qi Yu 1, Jordan Gunn 2, Daniel Osherson 2, and Jiaying Zhao 1,3 16 17 18 1 19 Department of Psychology, University of British Columbia 20 2Department of Psychology, Princeton University 21 3Institute for Resources, Environment and Sustainability, University of British Columbia 22 23 24 25 26 27 28 Manuscript type: Standard article 29 Main text word count: 4439 30 31 Abstract word count: 197/200 32 33 34 35 36 37 38 Please address correspondence to: 39 40 Jiaying Zhao 41 Department of Psychology 42 Institute for Resources, Environment and Sustainability 43 44 University of British Columbia 45 Vancouver, B.C., Canada, V6T 1Z4 46 Telephone: (1)-604-827-2203 47 Fax: (1)-604-827-2203 48 Email: [email protected] 49 50 51 52 53 54 55 56 57 58 59 60 URL: http:/mc.manuscriptcentral.com/pqje Quarterly Journal of Experimental Psychology Page 2 of 18 2 1 2 CONSISTENCY OF RANDOMNESS 3 Abstract 4 5 6 A pervasive bias in the subjective concept of randomness is that people often expect random 7 8 sequences to exhibit more alternations than produced by genuine random processes. Although 9 10 11 this over-alternation bias has received much attention, what is less known is the stability of the 12 13 bias. Here we examine two important aspects of the over-alternation bias: first, whether this bias 14 15 is present in stimuli that vary across feature dimensions, sensory modalities, presentation modes, 16 17 18 and probing methods; and second, how consistent the bias is across these stimulus variations. In 19 20 Experiment 1, participants adjusted sequences until they looked maximally random. The 21 22 sequences were presented as temporal streams of colors, shapes, auditory tones, or tiled as spatial 23 24 25 matrices. In Experiment 2, participants produced random matrices by adjusting the color of each 26 27 cell. We found that participants judged and produced over-alternating stimuli as the most 28 29 random. Importantly, this bias was consistent across presentation modes (temporal vs. spatial), 30 31 32 feature dimensions (color vs. shape), sensory modalities (visual vs. auditory), speed (fast vs. 33 34 slow), stimulus size (small vs. large matrices), and probing methods (adjusting the generating 35 36 37 process vs. individual bits). Overall, the results suggest that the subjective concept of 38 39 randomness is highly stable across stimulus variations. 40 41 42 43 44 Keywords : randomness, modalities, alternation bias, anchoring effect, features 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 URL: http:/mc.manuscriptcentral.com/pqje Page 3 of 18 Quarterly Journal of Experimental Psychology 3 1 2 CONSISTENCY OF RANDOMNESS 3 Introduction 4 5 6 In common parlance the term “random” is applied to sequences of events that appear 7 8 sufficiently disorderly or unstructured. For example, the string hthhtthtttht of heads and tails 9 10 11 from coin tosses might qualify as random, whereas hhhhhhtttttt would not. A contrasting usage, 12 13 adopted here, applies the term to certain mechanisms for generating events, namely, whose 14 15 successive outputs are independent and unbiased. A standard example is a device D that tosses a 16 17 18 fair coin repeatedly (ignoring issues about predictability assuming all forces were known). Any 19 20 sequence of heads and tails produced by such a device is qualified as “randomly generated (by 21 22 D)” regardless of its pattern. Thus, the terminology in this paper follows a “process” rather than 23 24 25 “product” conception of randomness (see Eagle, 2014; Earman, 1986 for an extended discussion 26 27 between the two approaches). Specifically, in our usage a “random” stimulus (or pattern) is an 28 29 object that has been produced by a random process. Non-random stimuli are defined as 30 31 32 productions from a distorted random source. 33 34 The randomness of the coin flipper D can be compromised in various ways, for example, 35 36 37 by making one of its two outcomes more likely than the other. Instead of introducing a bias to 38 39 the probabilities of the two outcomes, here we consider deviations from stochastic independence 40 41 by allowing previous flips to influence the next one, while maintaining the equal frequency of 42 43 44 the two outcomes. This allows us to identify the lay conception of randomness along a 45 46 continuum. Specifically, for each number x in the unit interval (from 0 to 1), let D(x) generate a 47 48 sequence of bits consisting of zeros and ones as follows: 49 50 51 52 53 54 55 56 57 58 59 60 URL: http:/mc.manuscriptcentral.com/pqje Quarterly Journal of Experimental Psychology Page 4 of 18 4 1 2 CONSISTENCY OF RANDOMNESS 3 (1) Sequence generation using the device D(x): An unbiased coin toss determines the first bit. 4 5 th 6 Suppose that the n bit has been constructed (for n ≥ 1). Then with probability x the n + 7 8 1st bit is set equal to the opposite of the n th bit; with probability 1 − x the n + 1 st bit is set 9 10 th 11 equal to the n bit. Repeat this process to generate a sequence of any length. 12 13 14 15 16 This procedure was introduced by Zhao, Hahn, and Osherson (2014). It can be seen that 17 18 D(.5) is a genuinely random device. For x < .5, D(x) tends to repeat itself, resulting in long runs 19 20 whereas for x > .5, D(x) tends to alternate. In particular, D(0) is uniform, either 0000… or 21 22 23 1111…, while D(1) consists of perfectly alternating bits, either 0101… or 1010…. The expected 24 25 proportion of each bit is 50%, for all x ∈ (0, 1). For any sequence produced by D(x), the 26 27 28 expected proportion of alternation — called the “switch rate” — is x. The switch rate of any 29 30 sequence is calculated by the number of switches between two successive bits divided by the 31 32 total number of bits in the sequence minus one. For example, the switch rate of the sequence of 33 34 35 11111 or 00000 is 0, and the switch rate of 01010 is 1. 36 37 A large body of research on randomness perception has revealed a pervasive bias in 38 39 40 people’s concept of “random.” That is, people often expect random sequences to exhibit shorter 41 42 runs (greater alternation) than typically produced by random devices, or in our terms, x > .5 43 44 (Kahneman & Tversky, 1972, Falk & Konold, 1997; Lopes & Oden, 1987; Nickerson & Butler, 45 46 47 2009; Wagenaar, 1972; for reviews on theoretical analysis of major experimental findings see 48 49 Bar-Hillel & Wagenaar, 1991; Oskarsson et al., 2009). 50 51 The over-alternation bias has been typically examined using judgment or production 52 53 54 paradigms (e.g., Kahneman & Tversky, 1972; Wagenaar, 1972), where participants either judged 55 56 how random a sequence appeared, or produced a sequence as if it were generated by a random 57 58 59 60 URL: http:/mc.manuscriptcentral.com/pqje Page 5 of 18 Quarterly Journal of Experimental Psychology 5 1 2 CONSISTENCY OF RANDOMNESS 3 process. The over-alternation bias explains the gambler’s fallacy (Kahneman & Tversky, 1972; 4 5 6 Reuter et al., 2005; Wagenaar, 1988), and can be driven by limitations in working memory 7 8 (Baddeley, 1966; Hahn & Warren, 2009; Kareev, 1992). Despite the overwhelming evidence for 9 10 11 the over-alternation bias, what has been given little attention is whether this bias is consistent 12 13 across sensory modalities, stimulus feature dimensions, presentation modes, and tasks. 14 15 The primary goal of the current study is thus to examine the concept of randomness 16 17 18 across multiple domains. This will offer insights on the stability or consistency of the over- 19 20 alternation bias. In Experiment 1, participants observed a sequence of bits produced by D(x) for 21 22 as long as they wanted. They were asked to adjust the sequence by increasing the alternations or 23 24 25 the repetitions of the bits, in order to make the sequence look maximally random. In other words, 26 27 they manipulated the switch rate x until they were satisfied that the sequence was being 28 29 generated “randomly.” We then measured how close x was to .5 (genuine randomness). Of 30 31 32 principal interest was the stability of the switch rate across variations in the stimuli that represent 33 34 the bits generated by D(x). In Experiment 2, we examined the stability of the switch rate in a 35 36 37 randomness production task. The two experiments investigated the consistency of people’s 38 39 randomness concept across stimulus features, modalities, presentations, and tasks. A summary of 40 41 the stimulus conditions was shown in Table 1 below.
Recommended publications
  • Wavelet Entropy Energy Measure (WEEM): a Multiscale Measure to Grade a Geophysical System's Predictability
    EGU21-703, updated on 28 Sep 2021 https://doi.org/10.5194/egusphere-egu21-703 EGU General Assembly 2021 © Author(s) 2021. This work is distributed under the Creative Commons Attribution 4.0 License. Wavelet Entropy Energy Measure (WEEM): A multiscale measure to grade a geophysical system's predictability Ravi Kumar Guntu and Ankit Agarwal Indian Institute of Technology Roorkee, Hydrology, Roorkee, India ([email protected]) Model-free gradation of predictability of a geophysical system is essential to quantify how much inherent information is contained within the system and evaluate different forecasting methods' performance to get the best possible prediction. We conjecture that Multiscale Information enclosed in a given geophysical time series is the only input source for any forecast model. In the literature, established entropic measures dealing with grading the predictability of a time series at multiple time scales are limited. Therefore, we need an additional measure to quantify the information at multiple time scales, thereby grading the predictability level. This study introduces a novel measure, Wavelet Entropy Energy Measure (WEEM), based on Wavelet entropy to investigate a time series's energy distribution. From the WEEM analysis, predictability can be graded low to high. The difference between the entropy of a wavelet energy distribution of a time series and entropy of wavelet energy of white noise is the basis for gradation. The metric quantifies the proportion of the deterministic component of a time series in terms of energy concentration, and its range varies from zero to one. One corresponds to high predictable due to its high energy concentration and zero representing a process similar to the white noise process having scattered energy distribution.
    [Show full text]
  • 3 Autocorrelation
    3 Autocorrelation Autocorrelation refers to the correlation of a time series with its own past and future values. Autocorrelation is also sometimes called “lagged correlation” or “serial correlation”, which refers to the correlation between members of a series of numbers arranged in time. Positive autocorrelation might be considered a specific form of “persistence”, a tendency for a system to remain in the same state from one observation to the next. For example, the likelihood of tomorrow being rainy is greater if today is rainy than if today is dry. Geophysical time series are frequently autocorrelated because of inertia or carryover processes in the physical system. For example, the slowly evolving and moving low pressure systems in the atmosphere might impart persistence to daily rainfall. Or the slow drainage of groundwater reserves might impart correlation to successive annual flows of a river. Or stored photosynthates might impart correlation to successive annual values of tree-ring indices. Autocorrelation complicates the application of statistical tests by reducing the effective sample size. Autocorrelation can also complicate the identification of significant covariance or correlation between time series (e.g., precipitation with a tree-ring series). Autocorrelation implies that a time series is predictable, probabilistically, as future values are correlated with current and past values. Three tools for assessing the autocorrelation of a time series are (1) the time series plot, (2) the lagged scatterplot, and (3) the autocorrelation function. 3.1 Time series plot Positively autocorrelated series are sometimes referred to as persistent because positive departures from the mean tend to be followed by positive depatures from the mean, and negative departures from the mean tend to be followed by negative departures (Figure 3.1).
    [Show full text]
  • Biostatistics for Oral Healthcare
    Biostatistics for Oral Healthcare Jay S. Kim, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Ronald J. Dailey, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Biostatistics for Oral Healthcare Biostatistics for Oral Healthcare Jay S. Kim, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 Ronald J. Dailey, Ph.D. Loma Linda University School of Dentistry Loma Linda, California 92350 JayS.Kim, PhD, is Professor of Biostatistics at Loma Linda University, CA. A specialist in this area, he has been teaching biostatistics since 1997 to students in public health, medical school, and dental school. Currently his primary responsibility is teaching biostatistics courses to hygiene students, predoctoral dental students, and dental residents. He also collaborates with the faculty members on a variety of research projects. Ronald J. Dailey is the Associate Dean for Academic Affairs at Loma Linda and an active member of American Dental Educational Association. C 2008 by Blackwell Munksgaard, a Blackwell Publishing Company Editorial Offices: Blackwell Publishing Professional, 2121 State Avenue, Ames, Iowa 50014-8300, USA Tel: +1 515 292 0140 9600 Garsington Road, Oxford OX4 2DQ Tel: 01865 776868 Blackwell Publishing Asia Pty Ltd, 550 Swanston Street, Carlton South, Victoria 3053, Australia Tel: +61 (0)3 9347 0300 Blackwell Wissenschafts Verlag, Kurf¨urstendamm57, 10707 Berlin, Germany Tel: +49 (0)30 32 79 060 The right of the Author to be identified as the Author of this Work has been asserted in accordance with the Copyright, Designs and Patents Act 1988. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by the UK Copyright, Designs and Patents Act 1988, without the prior permission of the publisher.
    [Show full text]
  • The Bayesian Approach to Statistics
    THE BAYESIAN APPROACH TO STATISTICS ANTHONY O’HAGAN INTRODUCTION the true nature of scientific reasoning. The fi- nal section addresses various features of modern By far the most widely taught and used statisti- Bayesian methods that provide some explanation for the rapid increase in their adoption since the cal methods in practice are those of the frequen- 1980s. tist school. The ideas of frequentist inference, as set out in Chapter 5 of this book, rest on the frequency definition of probability (Chapter 2), BAYESIAN INFERENCE and were developed in the first half of the 20th century. This chapter concerns a radically differ- We first present the basic procedures of Bayesian ent approach to statistics, the Bayesian approach, inference. which depends instead on the subjective defini- tion of probability (Chapter 3). In some respects, Bayesian methods are older than frequentist ones, Bayes’s Theorem and the Nature of Learning having been the basis of very early statistical rea- Bayesian inference is a process of learning soning as far back as the 18th century. Bayesian from data. To give substance to this statement, statistics as it is now understood, however, dates we need to identify who is doing the learning and back to the 1950s, with subsequent development what they are learning about. in the second half of the 20th century. Over that time, the Bayesian approach has steadily gained Terms and Notation ground, and is now recognized as a legitimate al- ternative to the frequentist approach. The person doing the learning is an individual This chapter is organized into three sections.
    [Show full text]
  • Big Data for Reliability Engineering: Threat and Opportunity
    Reliability, February 2016 Big Data for Reliability Engineering: Threat and Opportunity Vitali Volovoi Independent Consultant [email protected] more recently, analytics). It shares with the rest of the fields Abstract - The confluence of several technologies promises under this umbrella the need to abstract away most stormy waters ahead for reliability engineering. News reports domain-specific information, and to use tools that are mainly are full of buzzwords relevant to the future of the field—Big domain-independent1. As a result, it increasingly shares the Data, the Internet of Things, predictive and prescriptive lingua franca of modern systems engineering—probability and analytics—the sexier sisters of reliability engineering, both statistics that are required to balance the otherwise orderly and exciting and threatening. Can we reliability engineers join the deterministic engineering world. party and suddenly become popular (and better paid), or are And yet, reliability engineering does not wear the fancy we at risk of being superseded and driven into obsolescence? clothes of its sisters. There is nothing privileged about it. It is This article argues that“big-picture” thinking, which is at the rarely studied in engineering schools, and it is definitely not core of the concept of the System of Systems, is key for a studied in business schools! Instead, it is perceived as a bright future for reliability engineering. necessary evil (especially if the reliability issues in question are safety-related). The community of reliability engineers Keywords - System of Systems, complex systems, Big Data, consists of engineers from other fields who were mainly Internet of Things, industrial internet, predictive analytics, trained on the job (instead of receiving formal degrees in the prescriptive analytics field).
    [Show full text]
  • Volatility Estimation of Stock Prices Using Garch Method
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by International Institute for Science, Technology and Education (IISTE): E-Journals European Journal of Business and Management www.iiste.org ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online) Vol.7, No.19, 2015 Volatility Estimation of Stock Prices using Garch Method Koima, J.K, Mwita, P.N Nassiuma, D.K Kabarak University Abstract Economic decisions are modeled based on perceived distribution of the random variables in the future, assessment and measurement of the variance which has a significant impact on the future profit or losses of particular portfolio. The ability to accurately measure and predict the stock market volatility has a wide spread implications. Volatility plays a very significant role in many financial decisions. The main purpose of this study is to examine the nature and the characteristics of stock market volatility of Kenyan stock markets and its stylized facts using GARCH models. Symmetric volatility model namly GARCH model was used to estimate volatility of stock returns. GARCH (1, 1) explains volatility of Kenyan stock markets and its stylized facts including volatility clustering, fat tails and mean reverting more satisfactorily.The results indicates the evidence of time varying stock return volatility over the sampled period of time. In conclusion it follows that in a financial crisis; the negative returns shocks have higher volatility than positive returns shocks. Keywords: GARCH, Stylized facts, Volatility clustering INTRODUCTION Volatility forecasting in financial market is very significant particularly in Investment, financial risk management and monetory policy making Poon and Granger (2003).Because of the link between volatility and risk,volatility can form a basis for efficient price discovery.Volatility implying predictability is very important phenomenon for traders and medium term - investors.
    [Show full text]
  • A Python Library for Neuroimaging Based Machine Learning
    Brain Predictability toolbox: a Python library for neuroimaging based machine learning 1 1 2 1 1 1 Hahn, S. ,​ Yuan, D.K. ,​ Thompson, W.K. ,​ Owens, M. ,​ Allgaier, N .​ and Garavan, H ​ ​ ​ ​ ​ ​ 1. Departments of Psychiatry and Complex Systems, University of Vermont, Burlington, VT 05401, USA 2. Division of Biostatistics, Department of Family Medicine and Public Health, University of California, San Diego, La Jolla, CA 92093, USA Abstract Summary Brain Predictability toolbox (BPt) represents a unified framework of machine learning (ML) tools designed to work with both tabulated data (in particular brain, psychiatric, behavioral, and physiological variables) and neuroimaging specific derived data (e.g., brain volumes and surfaces). This package is suitable for investigating a wide range of different neuroimaging based ML questions, in particular, those queried from large human datasets. Availability and Implementation BPt has been developed as an open-source Python 3.6+ package hosted at https://github.com/sahahn/BPt under MIT License, with documentation provided at ​ https://bpt.readthedocs.io/en/latest/, and continues to be actively developed. The project can be ​ downloaded through the github link provided. A web GUI interface based on the same code is currently under development and can be set up through docker with instructions at https://github.com/sahahn/BPt_app. ​ Contact Please contact Sage Hahn at [email protected] ​ Main Text 1 Introduction Large datasets in all domains are becoming increasingly prevalent as data from smaller existing studies are pooled and larger studies are funded. This increase in available data offers an unprecedented opportunity for researchers interested in applying machine learning (ML) based methodologies, especially those working in domains such as neuroimaging where data collection is quite expensive.
    [Show full text]
  • BIOSTATS Documentation
    BIOSTATS Documentation BIOSTATS is a collection of R functions written to aid in the statistical analysis of ecological data sets using both univariate and multivariate procedures. All of these functions make use of existing R functions and many are simply convenience wrappers for functions contained in other popular libraries, such as VEGAN and LABDSV for multivariate statistics, however others are custom functions written using functions in the base or stats libraries. Author: Kevin McGarigal, Professor Department of Environmental Conservation University of Massachusetts, Amherst Date Last Updated: 9 November 2016 Functions: all.subsets.gam.. 3 all.subsets.glm. 6 box.plots. 12 by.names. 14 ci.lines. 16 class.monte. 17 clus.composite.. 19 clus.stats. 21 cohen.kappa. 23 col.summary. 25 commission.mc. 27 concordance. 29 contrast.matrix. 33 cov.test. 35 data.dist. 37 data.stand. 41 data.trans. 44 dist.plots. 48 distributions. 50 drop.var. 53 edf.plots.. 55 ecdf.plots. 57 foa.plots.. 59 hclus.cophenetic. 62 hclus.scree. 64 hclus.table. 66 hist.plots. 68 intrasetcor. 70 Biostats Library 2 kappa.mc. 72 lda.structure.. 74 mantel2. 76 mantel.part. 78 mrpp2. 82 mv.outliers.. 84 nhclus.scree. 87 nmds.monte. 89 nmds.scree.. 91 norm.test. 93 ordi.monte.. 95 ordi.overlay. 97 ordi.part.. 99 ordi.scree. 105 pca.communality. 107 pca.eigenval. 109 pca.eigenvec. 111 pca.structure. 113 plot.anosim. 115 plot.mantel. 117 plot.mrpp.. 119 plot.ordi.part. 121 qqnorm.plots.. 123 ran.split. ..
    [Show full text]
  • Public Health & Intelligence
    Public Health & Intelligence PHI Trend Analysis Guidance Document Control Version Version 1.5 Date Issued March 2017 Author Róisín Farrell, David Walker / HI Team Comments to [email protected] Version Date Comment Author Version 1.0 July 2016 1st version of paper David Walker, Róisín Farrell Version 1.1 August Amending format; adding Róisín Farrell 2016 sections Version 1.2 November Adding content to sections Róisín Farrell, Lee 2016 1 and 2 Barnsdale Version 1.3 November Amending as per Róisín Farrell 2016 suggestions from SAG Version 1.4 December Structural changes Róisín Farrell, Lee 2016 Barnsdale Version 1.5 March 2017 Amending as per Róisín Farrell, Lee suggestions from SAG Barnsdale Contents 1. Introduction ........................................................................................................................................... 1 2. Understanding trend data for descriptive analysis ................................................................................. 1 3. Descriptive analysis of time trends ......................................................................................................... 2 3.1 Smoothing ................................................................................................................................................ 2 3.1.1 Simple smoothing ............................................................................................................................ 2 3.1.2 Median smoothing ..........................................................................................................................
    [Show full text]
  • A Review of Graph and Network Complexity from an Algorithmic Information Perspective
    entropy Review A Review of Graph and Network Complexity from an Algorithmic Information Perspective Hector Zenil 1,2,3,4,5,*, Narsis A. Kiani 1,2,3,4 and Jesper Tegnér 2,3,4,5 1 Algorithmic Dynamics Lab, Centre for Molecular Medicine, Karolinska Institute, 171 77 Stockholm, Sweden; [email protected] 2 Unit of Computational Medicine, Department of Medicine, Karolinska Institute, 171 77 Stockholm, Sweden; [email protected] 3 Science for Life Laboratory (SciLifeLab), 171 77 Stockholm, Sweden 4 Algorithmic Nature Group, Laboratoire de Recherche Scientifique (LABORES) for the Natural and Digital Sciences, 75005 Paris, France 5 Biological and Environmental Sciences and Engineering Division (BESE), King Abdullah University of Science and Technology (KAUST), Thuwal 23955, Saudi Arabia * Correspondence: [email protected] or [email protected] Received: 21 June 2018; Accepted: 20 July 2018; Published: 25 July 2018 Abstract: Information-theoretic-based measures have been useful in quantifying network complexity. Here we briefly survey and contrast (algorithmic) information-theoretic methods which have been used to characterize graphs and networks. We illustrate the strengths and limitations of Shannon’s entropy, lossless compressibility and algorithmic complexity when used to identify aspects and properties of complex networks. We review the fragility of computable measures on the one hand and the invariant properties of algorithmic measures on the other demonstrating how current approaches to algorithmic complexity are misguided and suffer of similar limitations than traditional statistical approaches such as Shannon entropy. Finally, we review some current definitions of algorithmic complexity which are used in analyzing labelled and unlabelled graphs. This analysis opens up several new opportunities to advance beyond traditional measures.
    [Show full text]
  • Introduction to Bayesian Inference and Modeling Edps 590BAY
    Introduction to Bayesian Inference and Modeling Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2019 Introduction What Why Probability Steps Example History Practice Overview ◮ What is Bayes theorem ◮ Why Bayesian analysis ◮ What is probability? ◮ Basic Steps ◮ An little example ◮ History (not all of the 705+ people that influenced development of Bayesian approach) ◮ In class work with probabilities Depending on the book that you select for this course, read either Gelman et al. Chapter 1 or Kruschke Chapters 1 & 2. C.J. Anderson (Illinois) Introduction Fall 2019 2.2/ 29 Introduction What Why Probability Steps Example History Practice Main References for Course Throughout the coures, I will take material from ◮ Gelman, A., Carlin, J.B., Stern, H.S., Dunson, D.B., Vehtari, A., & Rubin, D.B. (20114). Bayesian Data Analysis, 3rd Edition. Boco Raton, FL, CRC/Taylor & Francis.** ◮ Hoff, P.D., (2009). A First Course in Bayesian Statistical Methods. NY: Sringer.** ◮ McElreath, R.M. (2016). Statistical Rethinking: A Bayesian Course with Examples in R and Stan. Boco Raton, FL, CRC/Taylor & Francis. ◮ Kruschke, J.K. (2015). Doing Bayesian Data Analysis: A Tutorial with JAGS and Stan. NY: Academic Press.** ** There are e-versions these of from the UofI library. There is a verson of McElreath, but I couldn’t get if from UofI e-collection. C.J. Anderson (Illinois) Introduction Fall 2019 3.3/ 29 Introduction What Why Probability Steps Example History Practice Bayes Theorem A whole semester on this? p(y|θ)p(θ) p(θ|y)= p(y) where ◮ y is data, sample from some population.
    [Show full text]
  • Autocorrelation
    Autocorrelation David Gerbing School of Business Administration Portland State University January 30, 2016 Autocorrelation The difference between an actual data value and the forecasted data value from a model is the residual for that forecasted value. Residual: ei = Yi − Y^i One of the assumptions of the least squares estimation procedure for the coefficients of a regression model is that the residuals are purely random. One consequence of randomness is that the residuals would not correlate with anything else, including with each other at different time points. A value above the mean, that is, a value with a positive residual, would contain no information as to whether the next value in time would have a positive residual, or negative residual, with a data value below the mean. For example, flipping a fair coin yields random flips, with half of the flips resulting in a Head and the other half a Tail. If a Head is scored as a 1 and a Tail as a 0, and the probability of both Heads and Tails is 0.5, then calculate the value of the population mean as: Population Mean: µ = (0:5)(1) + (0:5)(0) = :5 The forecast of the outcome of the next flip of a fair coin is the mean of the process, 0.5, which is stable over time. What are the corresponding residuals? A residual value is the difference of the corresponding data value minus the mean. With this scoring system, a Head generates a positive residual from the mean, µ, Head: ei = 1 − µ = 1 − 0:5 = 0:5 A Tail generates a negative residual from the mean, Tail: ei = 0 − µ = 0 − 0:5 = −0:5 The error terms of the coin flips are independent of each other, so if the current flip is a Head, or if the last 5 flips are Heads, the forecast for the next flip is still µ = :5.
    [Show full text]