1 | Page JASP – Bayesian Inference. Dr Mark Goss-Sampson
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Anomalous Perception in a Ganzfeld Condition - a Meta-Analysis of More Than 40 Years Investigation [Version 1; Peer Review: Awaiting Peer Review]
F1000Research 2021, 10:234 Last updated: 08 SEP 2021 RESEARCH ARTICLE Stage 2 Registered Report: Anomalous perception in a Ganzfeld condition - A meta-analysis of more than 40 years investigation [version 1; peer review: awaiting peer review] Patrizio E. Tressoldi 1, Lance Storm2 1Studium Patavinum, University of Padua, Padova, Italy 2School of Psychology, University of Adelaide, Adelaide, Australia v1 First published: 24 Mar 2021, 10:234 Open Peer Review https://doi.org/10.12688/f1000research.51746.1 Latest published: 24 Mar 2021, 10:234 https://doi.org/10.12688/f1000research.51746.1 Reviewer Status AWAITING PEER REVIEW Any reports and responses or comments on the Abstract article can be found at the end of the article. This meta-analysis is an investigation into anomalous perception (i.e., conscious identification of information without any conventional sensorial means). The technique used for eliciting an effect is the ganzfeld condition (a form of sensory homogenization that eliminates distracting peripheral noise). The database consists of studies published between January 1974 and December 2020 inclusive. The overall effect size estimated both with a frequentist and a Bayesian random-effect model, were in close agreement yielding an effect size of .088 (.04-.13). This result passed four publication bias tests and seems not contaminated by questionable research practices. Trend analysis carried out with a cumulative meta-analysis and a meta-regression model with Year of publication as covariate, did not indicate sign of decline of this effect size. The moderators analyses show that selected participants outcomes were almost three-times those obtained by non-selected participants and that tasks that simulate telepathic communication show a two- fold effect size with respect to tasks requiring the participants to guess a target. -
Dentist's Knowledge of Evidence-Based Dentistry and Digital Applications Resources in Saudi Arabia
PTB Reports Research Article Dentist’s Knowledge of Evidence-based Dentistry and Digital Applications Resources in Saudi Arabia Yousef Ahmed Alomi*, BSc. Pharm, MSc. Clin Pharm, BCPS, BCNSP, DiBA, CDE ABSTRACT Critical Care Clinical Pharmacists, TPN Objectives: Drug information resources provide clinicians with safer use of medications and play a Clinical Pharmacist, Freelancer Business vital role in improving drug safety. Evidence-based medicine (EBM) has become essential to medical Planner, Content Editor, and Data Analyst, Riyadh, SAUDI ARABIA. practice; however, EBM is still an emerging dentistry concept. Therefore, in this study, we aimed to Anwar Mouslim Alshammari, B.D.S explore dentists’ knowledge about evidence-based dentistry resources in Saudi Arabia. Methods: College of Destiney, Hail University, SAUDI This is a 4-month cross-sectional study conducted to analyze dentists’ knowledge about evidence- ARABIA. based dentistry resources in Saudi Arabia. We included dentists from interns to consultants and Hanin Sumaydan Saleam Aljohani, those across all dentistry specialties and located in Saudi Arabia. The survey collected demographic Ministry of Health, Riyadh, SAUDI ARABIA. information and knowledge of resources on dental drugs. The knowledge of evidence-based dental care and knowledge of dental drug information applications. The survey was validated through the Correspondence: revision of expert reviewers and pilot testing. Moreover, various reliability tests had been done with Dr. Yousef Ahmed Alomi, BSc. Pharm, the study. The data were collected through the Survey Monkey system and analyzed using Statistical MSc. Clin Pharm, BCPS, BCNSP, DiBA, CDE Critical Care Clinical Pharmacists, TPN Package of Social Sciences (SPSS) and Jeffery’s Amazing Statistics Program (JASP). -
11. Correlation and Linear Regression
11. Correlation and linear regression The goal in this chapter is to introduce correlation and linear regression. These are the standard tools that statisticians rely on when analysing the relationship between continuous predictors and continuous outcomes. 11.1 Correlations In this section we’ll talk about how to describe the relationships between variables in the data. To do that, we want to talk mostly about the correlation between variables. But first, we need some data. 11.1.1 The data Table 11.1: Descriptive statistics for the parenthood data. variable min max mean median std. dev IQR Dan’s grumpiness 41 91 63.71 62 10.05 14 Dan’s hours slept 4.84 9.00 6.97 7.03 1.02 1.45 Dan’s son’s hours slept 3.25 12.07 8.05 7.95 2.07 3.21 ............................................................................................ Let’s turn to a topic close to every parent’s heart: sleep. The data set we’ll use is fictitious, but based on real events. Suppose I’m curious to find out how much my infant son’s sleeping habits affect my mood. Let’s say that I can rate my grumpiness very precisely, on a scale from 0 (not at all grumpy) to 100 (grumpy as a very, very grumpy old man or woman). And lets also assume that I’ve been measuring my grumpiness, my sleeping patterns and my son’s sleeping patterns for - 251 - quite some time now. Let’s say, for 100 days. And, being a nerd, I’ve saved the data as a file called parenthood.csv. -
Arxiv:1803.00360V2 [Stat.CO] 2 Mar 2018 Prone to Misinterpretation [2, 3]
Computing Bayes factors to measure evidence from experiments: An extension of the BIC approximation Thomas J. Faulkenberry∗ Tarleton State University Bayesian inference affords scientists with powerful tools for testing hypotheses. One of these tools is the Bayes factor, which indexes the extent to which support for one hypothesis over another is updated after seeing the data. Part of the hesitance to adopt this approach may stem from an unfamiliarity with the computational tools necessary for computing Bayes factors. Previous work has shown that closed form approximations of Bayes factors are relatively easy to obtain for between-groups methods, such as an analysis of variance or t-test. In this paper, I extend this approximation to develop a formula for the Bayes factor that directly uses infor- mation that is typically reported for ANOVAs (e.g., the F ratio and degrees of freedom). After giving two examples of its use, I report the results of simulations which show that even with minimal input, this approximate Bayes factor produces similar results to existing software solutions. Note: to appear in Biometrical Letters. I. INTRODUCTION A. The Bayes factor Bayesian inference is a method of measurement that is Hypothesis testing is the primary tool for statistical based on the computation of P (H j D), which is called inference across much of the biological and behavioral the posterior probability of a hypothesis H, given data D. sciences. As such, most scientists are trained in classical Bayes' theorem casts this probability as null hypothesis significance testing (NHST). The scenario for testing a hypothesis is likely familiar to most readers of this journal. -
1 Estimation and Beyond in the Bayes Universe
ISyE8843A, Brani Vidakovic Handout 7 1 Estimation and Beyond in the Bayes Universe. 1.1 Estimation No Bayes estimate can be unbiased but Bayesians are not upset! No Bayes estimate with respect to the squared error loss can be unbiased, except in a trivial case when its Bayes’ risk is 0. Suppose that for a proper prior ¼ the Bayes estimator ±¼(X) is unbiased, Xjθ (8θ)E ±¼(X) = θ: This implies that the Bayes risk is 0. The Bayes risk of ±¼(X) can be calculated as repeated expectation in two ways, θ Xjθ 2 X θjX 2 r(¼; ±¼) = E E (θ ¡ ±¼(X)) = E E (θ ¡ ±¼(X)) : Thus, conveniently choosing either EθEXjθ or EX EθjX and using the properties of conditional expectation we have, θ Xjθ 2 θ Xjθ X θjX X θjX 2 r(¼; ±¼) = E E θ ¡ E E θ±¼(X) ¡ E E θ±¼(X) + E E ±¼(X) θ Xjθ 2 θ Xjθ X θjX X θjX 2 = E E θ ¡ E θ[E ±¼(X)] ¡ E ±¼(X)E θ + E E ±¼(X) θ Xjθ 2 θ X X θjX 2 = E E θ ¡ E θ ¢ θ ¡ E ±¼(X)±¼(X) + E E ±¼(X) = 0: Bayesians are not upset. To check for its unbiasedness, the Bayes estimator is averaged with respect to the model measure (Xjθ), and one of the Bayesian commandments is: Thou shall not average with respect to sample space, unless you have Bayesian design in mind. Even frequentist agree that insisting on unbiasedness can lead to bad estimators, and that in their quest to minimize the risk by trading off between variance and bias-squared a small dosage of bias can help. -
Statistical Analysis in JASP
Copyright © 2018 by Mark A Goss-Sampson. All rights reserved. This book or any portion thereof may not be reproduced or used in any manner whatsoever without the express written permission of the author except for the purposes of research, education or private study. CONTENTS PREFACE .................................................................................................................................................. 1 USING THE JASP INTERFACE .................................................................................................................... 2 DESCRIPTIVE STATISTICS ......................................................................................................................... 8 EXPLORING DATA INTEGRITY ................................................................................................................ 15 ONE SAMPLE T-TEST ............................................................................................................................. 22 BINOMIAL TEST ..................................................................................................................................... 25 MULTINOMIAL TEST .............................................................................................................................. 28 CHI-SQUARE ‘GOODNESS-OF-FIT’ TEST............................................................................................. 30 MULTINOMIAL AND Χ2 ‘GOODNESS-OF-FIT’ TEST. .......................................................................... -
Hierarchical Models & Bayesian Model Selection
Hierarchical Models & Bayesian Model Selection Geoffrey Roeder Departments of Computer Science and Statistics University of British Columbia Jan. 20, 2016 Geoffrey Roeder Hierarchical Models & Bayesian Model Selection Jan. 20, 2016 Contact information Please report any typos or errors to geoff[email protected] Geoffrey Roeder Hierarchical Models & Bayesian Model Selection Jan. 20, 2016 Outline 1 Hierarchical Bayesian Modelling Coin toss redux: point estimates for θ Hierarchical models Application to clinical study 2 Bayesian Model Selection Introduction Bayes Factors Shortcut for Marginal Likelihood in Conjugate Case Geoffrey Roeder Hierarchical Models & Bayesian Model Selection Jan. 20, 2016 Outline 1 Hierarchical Bayesian Modelling Coin toss redux: point estimates for θ Hierarchical models Application to clinical study 2 Bayesian Model Selection Introduction Bayes Factors Shortcut for Marginal Likelihood in Conjugate Case Geoffrey Roeder Hierarchical Models & Bayesian Model Selection Jan. 20, 2016 Let Y be a random variable denoting number of observed heads in n coin tosses. Then, we can model Y ∼ Bin(n; θ), with probability mass function n p(Y = y j θ) = θy (1 − θ)n−y (1) y We want to estimate the parameter θ Coin toss: point estimates for θ Probability model Consider the experiment of tossing a coin n times. Each toss results in heads with probability θ and tails with probability 1 − θ Geoffrey Roeder Hierarchical Models & Bayesian Model Selection Jan. 20, 2016 We want to estimate the parameter θ Coin toss: point estimates for θ Probability model Consider the experiment of tossing a coin n times. Each toss results in heads with probability θ and tails with probability 1 − θ Let Y be a random variable denoting number of observed heads in n coin tosses. -
Resources for Mini Meta-Analysis and Related Things Updated on June 25, 2018 Here Are Some Resources That I and Others Find
Resources for Mini Meta-Analysis and Related Things Updated on June 25, 2018 Here are some resources that I and others find useful. I know my mini meta SPPC article and Excel were not exhaustive and I didn’t cover many things, so I hope this resource page will give you some answers. I’ll update this page whenever I find something new or receive suggestions. Meta-Analysis Guides & Workshops • David Wilson, who wrote Practical Meta-Analysis, has a website containing PowerPoints and macros (for SPSS, Stata, and SAS) as well as some useful links on meta-analysis o I created an Excel template based on his slides (read the Excel page to see how this template differs from my previous SPPC template) • McShane and Böckenholt (2017) wrote an article in the Journal of Consumer Research similar to my mini meta paper. It comes with a very cool website that calculates meta- analysis for you: https://blakemcshane.shinyapps.io/spmetacase1/ • Courtney Soderberg from the Center for Open Science gave a mini meta-analysis workshop at SIPS 2017. All materials are publicly available: https://osf.io/nmdtq/ • Adam Pegler wrote a tutorial blog on using R to calculate mini meta based on my article • The founders of MetaLab at Stanford made video tutorials on meta-analysis in general • Geoff Cumming has a video tutorial on meta-analysis and other “new statistics” or read his Psychological Science article on the “new statistics” • Patrick Forscher’s meta-analysis syllabus and class materials are available online • A blog post containing 5 tips to understand and carefully -
Bayes Factors in Practice Author(S): Robert E
Bayes Factors in Practice Author(s): Robert E. Kass Source: Journal of the Royal Statistical Society. Series D (The Statistician), Vol. 42, No. 5, Special Issue: Conference on Practical Bayesian Statistics, 1992 (2) (1993), pp. 551-560 Published by: Blackwell Publishing for the Royal Statistical Society Stable URL: http://www.jstor.org/stable/2348679 Accessed: 10/08/2010 18:09 Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at http://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in the JSTOR archive only for your personal, non-commercial use. Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained at http://www.jstor.org/action/showPublisher?publisherCode=black. Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission. JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Blackwell Publishing and Royal Statistical Society are collaborating with JSTOR to digitize, preserve and extend access to Journal of the Royal Statistical Society. -
Bayes Factor Consistency
Bayes Factor Consistency Siddhartha Chib John M. Olin School of Business, Washington University in St. Louis and Todd A. Kuffner Department of Mathematics, Washington University in St. Louis July 1, 2016 Abstract Good large sample performance is typically a minimum requirement of any model selection criterion. This article focuses on the consistency property of the Bayes fac- tor, a commonly used model comparison tool, which has experienced a recent surge of attention in the literature. We thoroughly review existing results. As there exists such a wide variety of settings to be considered, e.g. parametric vs. nonparametric, nested vs. non-nested, etc., we adopt the view that a unified framework has didactic value. Using the basic marginal likelihood identity of Chib (1995), we study Bayes factor asymptotics by decomposing the natural logarithm of the ratio of marginal likelihoods into three components. These are, respectively, log ratios of likelihoods, prior densities, and posterior densities. This yields an interpretation of the log ra- tio of posteriors as a penalty term, and emphasizes that to understand Bayes factor consistency, the prior support conditions driving posterior consistency in each respec- tive model under comparison should be contrasted in terms of the rates of posterior contraction they imply. Keywords: Bayes factor; consistency; marginal likelihood; asymptotics; model selection; nonparametric Bayes; semiparametric regression. 1 1 Introduction Bayes factors have long held a special place in the Bayesian inferential paradigm, being the criterion of choice in model comparison problems for such Bayesian stalwarts as Jeffreys, Good, Jaynes, and others. An excellent introduction to Bayes factors is given by Kass & Raftery (1995). -
Sequence Ambiguity Determined from Routine Pol Sequencing Is a Reliable Tool for Real-Time Surveillance of HIV Incidence Trends
Infection, Genetics and Evolution 69 (2019) 146–152 Contents lists available at ScienceDirect Infection, Genetics and Evolution journal homepage: www.elsevier.com/locate/meegid Research paper Sequence ambiguity determined from routine pol sequencing is a reliable T tool for real-time surveillance of HIV incidence trends ⁎ Maja M. Lunara, Snježana Židovec Lepejb, Mario Poljaka, a Institute of Microbiology and Immunology, Faculty of Medicine, University of Ljubljana, Zaloška 4, Ljubljana 1105, Slovenia b Dr. Fran Mihaljevič University Hospital for Infectious Diseases, Mirogojska 8, Zagreb 10000, Croatia ARTICLE INFO ABSTRACT Keywords: Identifying individuals recently infected with HIV has been of great significance for close monitoring of HIV HIV-1 epidemic dynamics. Low HIV sequence ambiguity (SA) has been described as a promising marker of recent Incidence infection in previous studies. This study explores the utility of SA defined as a proportion of ambiguous nu- Sequence ambiguity cleotides detected in baseline pol sequences as a tool for routine real-time surveillance of HIV incidence trends at Mixed base calls a national level. Surveillance A total of 353 partial HIV-1 pol sequences obtained from persons diagnosed with HIV infection in Slovenia from 2000 to 2012 were studied, and SA was reported as a percentage of ambiguous base calls. Patients were characterized as recently infected by examining anti-HIV serological patterns and/or using commercial HIV-1 incidence assays (BED and/or LAg-Avidity assay). A mean SA of 0.29%, 0.14%, and 0.19% was observed for infections classified as recent by BED, LAg, or anti-HIV serological results, respectively. Welch's t-test showed a significant difference in the SA of recent versus long-standing infections (p < 0.001). -
9. Categorical Data Analysis
9. Categorical data analysis Now that we’ve covered the basic theory behind hypothesis testing it’s time to start looking at specific tests that are commonly used in psychology. So where should we start? Not every textbook agrees on where to start, but I’m going to start with “χ2 tests” (this chapter, pronounced “chi-square”1)and “t-tests” (Chapter 10). Both of these tools are very frequently used in scientific practice, and whilst they’re not as powerful as “regression” (Chapter 11) and “analysis of variance” (Chapter 12)they’re much easier to understand. The term “categorical data” is just another name for “nominal scale data”. It’s nothing that we haven’t already discussed, it’s just that in the context of data analysis people tend to use the term “categorical data” rather than “nominal scale data”. I don’t know why. In any case, categorical data analysis refers to a collection of tools that you can use when your data are nominal scale. However, there are a lot of different tools that can be used for categorical data analysis, and this chapter covers only a few of the more common ones. 9.1 The χ2 (chi-square) goodness-of-fit test The χ2 goodness-of-fit test is one of the oldest hypothesis tests around. It was invented by Karl Pearson around the turn of the century (Pearson 1900), with some corrections made later by Sir Ronald Fisher (Fisher 1922a). It tests whether an observed frequency distribution of a nominal variable matches an expected frequency distribution.