Bayes, Jeffreys, Prior Distributions and the Philosophy of Statistics

Total Page:16

File Type:pdf, Size:1020Kb

Bayes, Jeffreys, Prior Distributions and the Philosophy of Statistics Statistical Science 2009, Vol. 24, No. 2, 176–178 DOI: 10.1214/09-STS284D Main article DOI: 10.1214/09-STS284 c Institute of Mathematical Statistics, 2009 Bayes, Jeffreys, Prior Distributions and the Philosophy of Statistics1 Andrew Gelman I actually own a copy of Harold Jeffreys’s The- tion of Jeffreys’s ideas is to explicitly include model ory of Probability but have only read small bits of checking in the process of data analysis. it, most recently over a decade ago to confirm that, indeed, Jeffreys was not too proud to use a classi- THE ROLE OF THE PRIOR DISTRIBUTION cal chi-squared p-value when he wanted to check the IN BAYESIAN DATA ANALYSIS misfit of a model to data (Gelman, Meng and Stern, At least in the field of statistics, Jeffreys is best 2006). I do, however, feel that it is important to un- known for his eponymous prior distribution and, derstand where our probability models come from, more generally, for the principle of constructing non- and I welcome the opportunity to use the present ar- informative, or minimally informative, or objective, ticle by Robert, Chopin and Rousseau as a platform 2 or reference prior distributions from the likelihood for further discussion of foundational issues. (see, for example, Kass and Wasserman, 1996). But In this brief discussion I will argue the follow- it can notoriously difficult to choose among nonin- ing: (1) in thinking about prior distributions, we formative priors; and, even more importantly, seem- should go beyond Jeffreys’s principles and move to- ingly noninformative distributions can sometimes have ward weakly informative priors; (2) it is natural for strong and undesirable implications, as I have found those of us who work in social and computational in my own experience (Gelman, 1996, 2006). As a re- sciences to favor complex models, contra Jeffreys’s sult I have become a convert to the cause of weakly preference for simplicity; and (3) a key generaliza- informative priors, which attempt to let the data speak while being strong enough to exclude vari- Andrew Gelman is Professor, Department of Statistics ous “unphysical” possibilities which, if not blocked, and Department of Political Science, Columbia can take over a posterior distribution in settings University e-mail: [email protected]; URL: with sparse data—a situation which is increasingly http://www.stat.columbia.edu/˜gelman. present as we continue to develop the techniques of 1Discussion of “Harold Jeffreys’s Theory of Probability re- visited,” by Christian Robert, Nicolas Chopin, and Judith working with complex hierarchical and nonparamet- Rousseau, for Statistical Science. ric models. arXiv:1001.2968v1 [stat.ME] 18 Jan 2010 This is an electronic reprint of the original article HOW THE SOCIAL AND COMPUTATIONAL published by the Institute of Mathematical Statistics in SCIENCES DIFFER FROM PHYSICS Statistical Science, 2009, Vol. 24, No. 2, 176–178. This reprint differs from the original in pagination and Robert, Chopin and Rousseau trace the applica- typographic detail. tion of Ockham’s razor (the preference for simpler 2On the topic of other books on the foundations of Bayesian models) from Jeffreys’s discussion of the law of grav- statistics, I confess to having found Savage (1954) to be nearly ity through later work of a mathematical statisti- unreadable, a book too much of a product of its time in its cian (Jim Berger), an astronomer (Bill Jefferys) and enthusiasm for game theory as a solution to all problems, an attitude which I find charming in the classic work of Luce and a physicist (David MacKay). From their perspec- Raiffa (1957) but more of annoyance in a book of statistical tive, Ockham’s razor seems unquestionably reason- methods. When it comes to Cold War-era foundational work able, with the only point of debate being the extent on Bayesian statistics, I much prefer the work of Lindley, in to which Bayesian inference automatically encom- his 1965 book and elsewhere. passes it. Also, I would be disloyal to my coauthors if I did not report My own perspective as a social scientist is com- that, despite what is said in the second footnote in the arti- cle under discussion, there is at least one other foundational pletely different. I’ve just about never heard some- Bayesian text of 1990s vintage that continues to receive more one in social science object to the inclusion of a citations than Jeffreys. variable or an interaction in a model; rather, the 1 2 A. GELMAN most serious criticisms of a model involve worries BAYESIAN INFERENCE VS. BAYESIAN DATA that certain potentially important factors have not ANALYSIS been included. In the social science problems I’ve One of my own epiphanies—actually stimulated seen, Ockham’s razor is at best an irrelevance and by the writings of E. T. Jaynes, yet another Bayesian at worse can lead to acceptance of models that are physicist—and incorporated into the title of my own missing key features that the data could actually book on Bayesian statistics, is that sometimes the provide information on. As such, I am no fan of most important thing to come out of an inference is methods such as BIC that attempt to justify the the rejection of the model on which it is based. Data use of simple models that do not fit observed data. analysis includes model building and criticism, not Don’t get me wrong—all the time I use simple mod- merely inference. Only through careful model build- els that don’t fit the data—but no amount of BIC ing is such definitive rejection possible. This idea— 3 will make me feel good about it! the comparison of predictive inferences to data— I much prefer Radford Neal’s line from his Ph.D. was forcefully put into Bayesian terms nearly thirty thesis: years ago by Box (1980) and Rubin (1984) but is even now still only gradually becoming standard in Sometimes a simple model will outperform Bayesian practice. a more complex model. Nevertheless, I A famous empiricist once said, “With great power [Neal] believe that deliberately limiting the comes great responsibility.” In Bayesian terms, the complexity of the model is not fruitful when stronger we make our model—following the excellent the problem is evidently complex. Instead, precepts of Jeffreys and Jaynes—the more able we if a simple model is found that outper- will be to find the model’s flaws and thus perform forms some particular complex model, the scientific learning. appropriate response is to define a differ- To roughly translate into philosophy-of-science jar- ent complex model that captures what- gon: Bayesian inference within a model is “normal ever aspect of the problem led to the sim- science,” and “scientific revolution” is the process of ple model performing well. checking a model, seeing its mismatches with reality, and coming up with a replacement. The revolution This is not really a Bayesian or a non-Bayesian is the glamour boy in this scenario, but, as Kuhn issue: complicated models with virtually unlimited (1962) emphasized, it is only the careful work of nor- nonlinearity and interactions are being developed mal science that makes the revolution possible: the using Bayesian principles. See, for example, Dun- better we can work out the implications of a theory, son (2006) and Chipman, George and McCulloch the more effectively we can find its flaws and thus (2008). To put it another way, you can be a prac- learn about nature.4 In this chicken-and-egg process, ticing Bayesian and prefer simpler models, or be a both normal science (Bayesian inference) and revo- practicing Bayesian and prefer complicated models. lution (Bayesian model revision) are useful, and they Or you can follow similar inclinations toward sim- feed upon each other. It is in this sense that graph- plicity or complexity from various non-Bayesian per- ical methods and exploratory data analysis can be spectives. viewed as explicitly Bayesian, as tools for comparing My point here is only that the Ockhamite tenden- posterior predictions to data (Gelman, 2003). cies of Jeffreys and his followers up to and includ- To get back to the Robert, Chopin, and Rousseau ing MacKay may derive, to some extent, from the article: I am suggesting that their identification (and simplicity of the best models of physics, the sense Jeffreys’s) of Bayesian data analysis with Bayesian that good science moves from the particular to the inference is limiting and, in practice, puts an unre- general—an attitude that does not fit in so well with alistic burden on any model. modern social and computational science. 4As Kuhn may very well have written had he lived long enough, scientific progress is fractal, with episodes of nor- 3See Gelman and Rubin (1995) for a fuller expression of this mal science and mini-revolutions happening over the period position, and Raftery (1995) for a defense of BIC in general of minutes, hours, days, and years, as well as the more familiar and in the context of two applications in sociology. examples of paradigms lasting over decades or centuries. COMMENT 3 CONCLUSION Gelman, A. (2003). A Bayesian formulation of exploratory data analysis and goodness-of-fit testing. Internat. Statist. If you wanted to do foundational research in statis- Rev. 71 369–382. tics in the mid-twentieth century, you had to be bit Gelman, A. (2006). Prior distributions for variance param- of a mathematician, whether you wanted to or not. eters in hierarchical models. Bayesian Anal. 1 515–533. As Robert, Chopin, and Rousseau’s own work re- MR2221284 veals, if you want to do statistical research at the Gelman, A., Carlin, J.
Recommended publications
  • The Interpretation of Probability: Still an Open Issue? 1
    philosophies Article The Interpretation of Probability: Still an Open Issue? 1 Maria Carla Galavotti Department of Philosophy and Communication, University of Bologna, Via Zamboni 38, 40126 Bologna, Italy; [email protected] Received: 19 July 2017; Accepted: 19 August 2017; Published: 29 August 2017 Abstract: Probability as understood today, namely as a quantitative notion expressible by means of a function ranging in the interval between 0–1, took shape in the mid-17th century, and presents both a mathematical and a philosophical aspect. Of these two sides, the second is by far the most controversial, and fuels a heated debate, still ongoing. After a short historical sketch of the birth and developments of probability, its major interpretations are outlined, by referring to the work of their most prominent representatives. The final section addresses the question of whether any of such interpretations can presently be considered predominant, which is answered in the negative. Keywords: probability; classical theory; frequentism; logicism; subjectivism; propensity 1. A Long Story Made Short Probability, taken as a quantitative notion whose value ranges in the interval between 0 and 1, emerged around the middle of the 17th century thanks to the work of two leading French mathematicians: Blaise Pascal and Pierre Fermat. According to a well-known anecdote: “a problem about games of chance proposed to an austere Jansenist by a man of the world was the origin of the calculus of probabilities”2. The ‘man of the world’ was the French gentleman Chevalier de Méré, a conspicuous figure at the court of Louis XIV, who asked Pascal—the ‘austere Jansenist’—the solution to some questions regarding gambling, such as how many dice tosses are needed to have a fair chance to obtain a double-six, or how the players should divide the stakes if a game is interrupted.
    [Show full text]
  • Statistical Science Meets Philosophy of Science Part 2: Shallow Versus Deep Explorations
    RMM Vol. 3, 2012, 71–107 Special Topic: Statistical Science and Philosophy of Science Edited by Deborah G. Mayo, Aris Spanos and Kent W. Staley http://www.rmm-journal.de/ Deborah Mayo Statistical Science Meets Philosophy of Science Part 2: Shallow versus Deep Explorations Abstract: Inability to clearly defend against the criticisms of frequentist methods has turned many a frequentist away from venturing into foundational battlegrounds. Conceding the dis- torted perspectives drawn from overly literal and radical expositions of what Fisher, Ney- man, and Pearson ‘really thought’, some deny they matter to current practice. The goal of this paper is not merely to call attention to the howlers that pass as legitimate crit- icisms of frequentist error statistics, but also to sketch the main lines of an alternative statistical philosophy within which to better articulate the roles and value of frequentist tools. 1. Comedy Hour at the Bayesian Retreat Overheard at the comedy hour at the Bayesian retreat: Did you hear the one about the frequentist. “who defended the reliability of his radiation reading, despite using a broken radiometer, on the grounds that most of the time he uses one that works, so on average he’s pretty reliable?” or “who claimed that observing ‘heads’ on a biased coin that lands heads with probability .05 is evidence of a statistically significant improve- ment over the standard treatment of diabetes, on the grounds that such an event occurs with low probability (.05)?” Such jests may work for an after-dinner laugh, but if it turns out that, despite being retreads of ‘straw-men’ fallacies, they form the basis of why some statis- ticians and philosophers reject frequentist methods, then they are not such a laughing matter.
    [Show full text]
  • PHIL 8670 (Fall, 2015): Philosophy of Statistics
    PHIL 8670 (Fall, 2015): Philosophy of Statistics Samuel C. Fletcher September 12, 2015 Primary Email: scfl[email protected] Office Phone: 612-625-4325 Office Hours: 754 Heller Hall, Tuesdays 3 pm { 4 pm, and by appointment Class Period: 731 Heller Hall, Tuesdays 4 pm { 6 pm Course Description Statistical methods form a significant part of the methodology of every scientific discipline. This seminar will survey a number of central philosophical and conceptual issues in statistics. After a brief review of elementary probability theory and its interpretation, we will consider at length both the classical and Bayesian schools of statistics: their formulations, and arguments pro and con. We will then consider a third school, likelihoodism, that is sometimes considered a compromise between classical and Bayesian methodologies and has gained favor among some philosophers of science. This leads to the likelihood principle, a natural focal point in the debate between these schools on the nature of evidence. The remainder of the seminar will be split between topics on statistical modeling (model verification and model selection), connections between the philosophy of statistics and more traditional topics in the philosophy of science (simplicity, induction and confirmation) and the relation between statistics and causation (randomness, randomization and causal inference). Evaluation Each week, there will be a semi-formal presentation of that week's reading assignment aimed at stimulating a critical discussion. For the first meeting (or two) I will be responsible for presenting, and for each of the remaining meetings, one or two students will be responsible. Students not presenting during a given week are responsible instead for submitting a one- paragraph question about a philosophical/conceptual issue arising in that week's readings to those presenting (and me).
    [Show full text]
  • Fusion Confusion? Comments on Nancy Reid: “BFF Four–Are We Converging?”
    Fusion Confusion? Comments on Nancy Reid: “BFF Four–Are we Converging?” Deborah G. Mayo The Fourth Bayesian, Fiducial and Frequentist Workshop (BFF4): Harvard University May 2, 2017 <1> I’m delighted to Be part of a workshop linking statistics and philosophy of statistics! I thank the organizers for inviting me. Nancy Reid’s “BFF Four–Are we Converging?” gives numerous avenues for discussion She zeroes in on obstacles to fusion: Confusion or disagreement on the nature of probability and its use in statistical inference <2> From Nancy Reid: Nature of probability probability to describe physical haphazard variability • probabilities represent features of the “real” world in idealized form • suBject to empirical test and improvement • conclusions of statistical analysis expressed in terms of interpretaBle parameters • enhanced understanding of the data generating process probability to describe the uncertainty of knowledge • measures rational, supposedly impersonal, degree of Belief given relevant information (Jeffreys) • measures a particular person’s degree of Belief, suBject typically to some constraints of self-consistency … • often linked with personal decision-making <3> • As is common, she laBels the second “epistemological” But a key question for me is: what’s relevant for a normative epistemology, for an account of what’s warranted/unwarranted to infer <4> Reid quite rightly asks: • in what sense are confidence distribution functions, significance functions, structural or fiducial probaBilities to be interpreted? • empirically? degree of Belief? • literature is not very clear <5> Reid: We may avoid the need for a different version of proBaBility by appeal to a notion of calibration (Cox 2006, Reid & Cox 2015) • This is my central focus I approach this indirectly, with analogy between philosophy of statistics and statistics <6> Carnap: Bayesians as Popper: Frequentists (N-P/Fisher) Can’t solve induction But can Build logics of induction or confirmation theories (e.g., Carnap 1962).
    [Show full text]
  • Teaching Statistics in a Critical Way: Historical, Philosophical and Political Aspects of Statistics
    ICOTS-7, 2006: François and Bracke (Refereed) TEACHING STATISTICS IN A CRITICAL WAY: HISTORICAL, PHILOSOPHICAL AND POLITICAL ASPECTS OF STATISTICS Karen François Free University Brussels, Belgium Nele Bracke Ghent University, Belgium [email protected] In this paper we sketch the history and the philosophy of statistics and probability theory and the connections to its political aspects. Knowledge of the cultural embeddedness of statistics and probability theory is an added value in the teaching thereof. The use of statistics and probability is a phenomenon with which everyone is confronted on a daily basis. Beside literacy, numeracy is an important challenge for education. In order to succeed in this task, the traditional curriculum (technique-oriented and individual, competition-oriented) will need to be sacrificed for a curriculum in which there is room for the cultural aspects of statistics and probability theory. For this purpose, cooperative learning is a didactic entry which is suitable for interaction and critical input of pupils. INTRODUCTION Looking at the history of statistics, one can show the growing importance of statistics in social sciences and in society in general. Understanding statistical representations, being able to interpret graphs or to give meaning to information in figures today are part of the skills that are generally assumed. Education, and particularly education in probability and statistics, face the challenge of helping to develop this competence. If education wishes to take up part of the responsibility to bring about a democratic state in which all have access to basic information like the media, then the development of a working knowledge of statistics and probability should form part of the curriculum.
    [Show full text]
  • Is the Philosophy of Probabilism an Obstacle to Statistical Fraud Busting? Deborah G
    Fraudbusting Feb 20, 2014 Mayo / 1 Is the Philosophy of Probabilism an Obstacle to Statistical Fraud Busting? Deborah G. Mayo 2013 was the “year of celebrating statistics”, but it might well have been dubbed the “year of frustration with statistics” Well-worn criticisms of how easy it is to lie with statistics have their own slogans: o Association is not causation. o Statistical significance is not substantive significance. o No evidence of risk is not evidence of no risk. o If you torture the data enough, they will confess. Fraudbusting Feb 20, 2014 Mayo / 2 Professional manuals and treatises written for popular consumption are rife with exposés of fallacies and foibles. • Task forces are organized, cottage industries regularly arise to expose the lack of replication and all manner of selection and publication biases. • The rise of “big data” might have made cookbook statistics easier but these fallacies were addressed by the founders of the tools. Fraudbusting Feb 20, 2014 Mayo / 3 R. A. Fisher (birthday: Feb 17) observed that: [t]he political principle that anything can be proved by statistics arises from the practice of presenting only a selected subset of the data available” (Fisher 1955, 75). • However, nowadays it’s the tools that are blamed instead of the abuser. • We don’t need “scaling up” so much as scaling back many misdiagnoses of what is going wrong. • In one sense, it’s blatantly obvious: sufficient finagling may practically predetermine that a researcher’s favorite hypothesis gets support, even if it’s unwarranted by evidence. Fraudbusting Feb 20, 2014 Mayo / 4 2.
    [Show full text]
  • Induction and Deduction in Bayesian Data Analysis*
    RMM Vol. 2, 2011, 67–78 Special Topic: Statistical Science and Philosophy of Science Edited by Deborah G. Mayo, Aris Spanos and Kent W. Staley http://www.rmm-journal.de/ Andrew Gelman Induction and Deduction in Bayesian Data Analysis* Abstract: The classical or frequentist approach to statistics (in which inference is centered on sig- nificance testing), is associated with a philosophy in which science is deductive and fol- lows Popper’s doctrine of falsification. In contrast, Bayesian inference is commonly asso- ciated with inductive reasoning and the idea that a model can be dethroned by a compet- ing model but can never be directly falsified by a significance test. The purpose of this article is to break these associations, which I think are incorrect and have been detrimen- tal to statistical practice, in that they have steered falsificationists away from the very useful tools of Bayesian inference and have discouraged Bayesians from checking the fit of their models. From my experience using and developing Bayesian methods in social and environmental science, I have found model checking and falsification to be central in the modeling process. 1. The Standard View of the Philosophy of Statistics, and Its Malign Influence on Statistical Practice Statisticians can be roughly divided into two camps, each with a clear configu- ration of practice and philosophy. I will divide some of the relevant adjectives into two columns: Frequentist Bayesian Objective Subjective Procedures Models P-values Bayes factors Deduction Induction Falsification Pr (model is true) This division is not absolute—in particular, it does not capture my own mix of philosophy and statistical practice—but it reflects a common alignment of be- liefs.
    [Show full text]
  • Frequentist Statistics As a Theory of Inductive Inference
    IMS Lecture Notes–Monograph Series 2nd Lehmann Symposium – Optimality Vol. 49 (2006) 77–97 c Institute of Mathematical Statistics, 2006 DOI: 10.1214/074921706000000400 Frequentist statistics as a theory of inductive inference Deborah G. Mayo1 and D. R. Cox2 Viriginia Tech and Nuffield College, Oxford Abstract: After some general remarks about the interrelation between philo- sophical and statistical thinking, the discussion centres largely on significance tests. These are defined as the calculation of p-values rather than as formal procedures for “acceptance” and “rejection.” A number of types of null hypoth- esis are described and a principle for evidential interpretation set out governing the implications of p-values in the specific circumstances of each application, as contrasted with a long-run interpretation. A variety of more complicated situations are discussed in which modification of the simple p-value may be essential. 1. Statistics and inductive philosophy 1.1. What is the Philosophy of Statistics? The philosophical foundations of statistics may be regarded as the study of the epistemological, conceptual and logical problems revolving around the use and in- terpretation of statistical methods, broadly conceived. As with other domains of philosophy of science, work in statistical science progresses largely without wor- rying about “philosophical foundations”. Nevertheless, even in statistical practice, debates about the different approaches to statistical analysis may influence and be influenced by general issues of the nature of inductive-statistical inference, and thus are concerned with foundational or philosophical matters. Even those who are largely concerned with applications are often interested in identifying general prin- ciples that underlie and justify the procedures they have come to value on relatively pragmatic grounds.
    [Show full text]
  • The Philosophy of Statistics Author(S): Dennis V
    The Philosophy of Statistics Author(s): Dennis V. Lindley Reviewed work(s): Source: Journal of the Royal Statistical Society. Series D (The Statistician), Vol. 49, No. 3 (2000), pp. 293-337 Published by: Wiley-Blackwell for the Royal Statistical Society Stable URL: http://www.jstor.org/stable/2681060 . Accessed: 05/09/2012 14:41 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Wiley-Blackwell and Royal Statistical Society are collaborating with JSTOR to digitize, preserve and extend access to Journal of the Royal Statistical Society. Series D (The Statistician). http://www.jstor.org The Statistician(2000) 49, Part3, pp. 293-337 The philosophy of statistics Dennis V. Lindley Minehead,UK [Received June 1999] Summary. This paper puts forwardan overallview of statistics.It is argued that statisticsis the studyof uncertainty.The many demonstrationsthat uncertaintiescan only combine accordingto the rules of the probabilitycalculus are summarized.The conclusion is thatstatistical inference is firmlybased on probabilityalone. Progress is thereforedependent on the constructionof a probabilitymodel; methods for doing this are considered. It is argued that the probabilitiesare personal. The roles of likelihoodand exchangeabilityare explained. Inferenceis onlyof value ifit can be used, so the extensionto decision analysis, incorporatingutility, is relatedto riskand to the use of statisticsin science and law.
    [Show full text]
  • Statistical Inference: the Big Picture1 Robert E
    Statistical Science 2011, Vol. 26, No. 1, 1–9 DOI: 10.1214/10-STS337 © Institute of Mathematical Statistics, 2011 Statistical Inference: The Big Picture1 Robert E. Kass Abstract. Statistics has moved beyond the frequentist-Bayesian controver- sies of the past. Where does this leave our ability to interpret results? I suggest that a philosophy compatible with statistical practice, labeled here statisti- cal pragmatism, serves as a foundation for inference. Statistical pragmatism is inclusive and emphasizes the assumptions that connect statistical models with observed data. I argue that introductory courses often mischaracterize the process of statistical inference and I propose an alternative “big picture” depiction. Key words and phrases: Bayesian, confidence, frequentist, statistical edu- cation, statistical pragmatism, statistical significance. 1. INTRODUCTION ple, these caricatures forfeit an opportunity to articulate a fundamental attitude of statistical practice. The protracted battle for the foundations of statis- Most modern practitioners have, I think, an open- tics, joined vociferously by Fisher, Jeffreys, Neyman, minded view about alternative modes of inference, but Savage and many disciples, has been deeply illuminat- are acutely aware of theoretical assumptions and the ing, but it has left statistics without a philosophy that matches contemporary attitudes. Because each camp many ways they may be mistaken. I would suggest that took as its goal exclusive ownership of inference, each it makes more sense to place in the center of our log- was doomed to failure. We have all, or nearly all, ical framework the match or mismatch of theoretical moved past these old debates, yet our textbook expla- assumptions with the real world of data.
    [Show full text]
  • 1 Excerpts from Stanford's Encyclopedia of Philosophy Article
    Excerpts from Stanford’s Encyclopedia of Philosophy article on the Philosophy of Statistics article at http://plato.stanford.edu/entries/statistics/ Now what does the probability function mean? The mathematical notion of probability does not provide an answer. The function P may be interpreted as physical, namely the frequency or propensity of the occurrence of a state of affairs, often referred to as the chance, or else as epistemic, namely the degree of belief in the occurrence of the state of affairs, the willingness to act on its assumption, a degree of support or confirmation, or similar. Physical probability and classical statistics In the sciences, the idea that probabilities express physical states of affairs, often called chances or stochastic processes, is most prominent. They are relative frequencies in series of events or, alternatively, they are tendencies or propensities in the systems that realize those events. More precisely, the probability attached to the property of an event type can be understood as the frequency or tendency with which that property manifests in a series of events of that type. The notion of physical probability is connected to one of the major theories of statistical method, which has come to be called classical statistics. It was developed roughly in the first half of the 20th century, mostly by mathematicians and working scientists like Fisher (1925, 1935, 1956), Wald (1939, 1950), Neyman and Pearson (1928, 1933, 1967)… Physical probability cannot meaningfully be attributed to statistical hypotheses, since hypotheses do not have tendencies or frequencies with which they come about: they are categorically true or false, once and for all.
    [Show full text]
  • Bayesian Philosophy of Science
    Jan Sprenger Stephan Hartmann Bayesian Philosophy of Science Variations on a Theme by the Reverend Thomas Bayes p(EjH) p(HjE) = p(H) p(E) Ich bin ein Esel, und will getreu, wie meine Väter, die Alten, an der alten, lieben Eselei, am Eseltume halten. Und weil ich ein Esel, so rat ich euch, den Esel zum König zu wählen; wir stiften das große Eselreich, wo nur die Esel befehlen. Wir sind alle Esel! I-A! I-A! Wir sind keine Pferdeknechte. Fort mit den Rossen! Es lebe—Hurra!— Der König vom Eselsgeschlechte! Heinrich Heine Contents Theme: Bayesian Philosophy of Science 1 Variation 1: Learning Conditional Information 31 1.1 The Kullback-Leibler Divergence and Probabilistic Updating 33 1.2 Three Challenges for Minimizing Divergence . 38 1.3 Meeting the Challenges . 40 1.4 Discussion . 48 1.5 Proofs of the Theorems . 51 Variation 2: Confirmation 57 2.1 Motivating Bayesian Confirmation Theory . 58 2.2 Confirmation as (Increase in) Firmness . 59 2.3 The Plurality of Bayesian Confirmation Measures . 69 2.4 Discussion . 73 Variation 3: The Problem of Old Evidence 77 3.1 The Dynamic POE: The GJN Approach . 79 3.2 Solving the Dynamic POE: Alternative Explanations . 83 3.3 The Static POE: A Counterfactual Perspective . 86 3.4 Solving the Hybrid POE: Learning Explanatory Relationships 88 3.5 Discussion . 92 3.6 Proofs of the Theorems . 96 Variation 4: The No Alternatives Argument 101 4.1 Modeling the No Alternatives Argument . 102 4.2 Results and Discussion . 107 4.3 Proof of the Theorems . 111 v vi Contents Variation 5: Scientific Realism and the No Miracles Argument 115 5.1 The Probabilistic No Miracles Argument .
    [Show full text]