Statistical Science Meets Philosophy of Science Part 2: Shallow Versus Deep Explorations

Total Page:16

File Type:pdf, Size:1020Kb

Statistical Science Meets Philosophy of Science Part 2: Shallow Versus Deep Explorations RMM Vol. 3, 2012, 71–107 Special Topic: Statistical Science and Philosophy of Science Edited by Deborah G. Mayo, Aris Spanos and Kent W. Staley http://www.rmm-journal.de/ Deborah Mayo Statistical Science Meets Philosophy of Science Part 2: Shallow versus Deep Explorations Abstract: Inability to clearly defend against the criticisms of frequentist methods has turned many a frequentist away from venturing into foundational battlegrounds. Conceding the dis- torted perspectives drawn from overly literal and radical expositions of what Fisher, Ney- man, and Pearson ‘really thought’, some deny they matter to current practice. The goal of this paper is not merely to call attention to the howlers that pass as legitimate crit- icisms of frequentist error statistics, but also to sketch the main lines of an alternative statistical philosophy within which to better articulate the roles and value of frequentist tools. 1. Comedy Hour at the Bayesian Retreat Overheard at the comedy hour at the Bayesian retreat: Did you hear the one about the frequentist. “who defended the reliability of his radiation reading, despite using a broken radiometer, on the grounds that most of the time he uses one that works, so on average he’s pretty reliable?” or “who claimed that observing ‘heads’ on a biased coin that lands heads with probability .05 is evidence of a statistically significant improve- ment over the standard treatment of diabetes, on the grounds that such an event occurs with low probability (.05)?” Such jests may work for an after-dinner laugh, but if it turns out that, despite being retreads of ‘straw-men’ fallacies, they form the basis of why some statis- ticians and philosophers reject frequentist methods, then they are not such a laughing matter. But surely the drubbing of frequentist methods could not be based on a collection of howlers, could it? I invite the reader to stay and find out. If we are to take the criticisms seriously, and put to one side the possibility that they are deliberate distortions of frequentist statistical methods, we need 72 Deborah Mayo to identify their sources. To this end I consider two interrelated areas around which to organize foundational issues in statistics: (1) the roles of probability in induction and inference, and (2) the nature and goals of statistical inference in science or learning. Frequentist sampling statistics, which I prefer to call ‘error statistics’, continues to be raked over the coals in the foundational literature, but with little scrutiny of the presuppositions about goals and methods, without which the criticisms lose all force. First, there is the supposition that an adequate account must assign degrees of probability to hypotheses, an assumption often called probabilism. Second, there is the assumption that the main, if not the only, goal of error-statistical methods is to evaluate long-run error rates. Given the wide latitude with which some critics define ‘controlling long-run error’, it is not surprising to find them arguing that (i) error statisticians approve of silly methods, and/or (ii) rival (e.g., Bayesian) accounts also satisfy error statistical demands. Absent this sleight of hand, Bayesian celebrants would have to go straight to the finale of their entertainment hour: a rousing rendition of ‘There’s No Theorem Like Bayes’s Theorem’. Never mind that frequentists have responded to these criticisms, they keep popping up (verbatim) in every Bayesian and some non-Bayesian textbooks and articles on philosophical foundations. No wonder that statistician Stephen Senn is inclined to “describe a Bayesian as one who has a reverential awe for all opin- ions except those of a frequentist statistician” (Senn 2011, 59, this special topic of RMM). Never mind that a correct understanding of the error-statistical de- mands belies the assumption that any method (with good performance proper- ties in the asymptotic long-run) succeeds in satisfying error-statistical demands. The difficulty of articulating a statistical philosophy that fully explains the basis for both (i) insisting on error-statistical guarantees, while (ii) avoiding pathological examples in practice, has turned many a frequentist away from venturing into foundational battlegrounds. Some even concede the distorted perspectives drawn from overly literal and radical expositions of what Fisher, Neyman, and Pearson ‘really thought’. I regard this as a shallow way to do foundations. Here is where I view my contribution—as a philosopher of science—to the long-standing debate: not merely to call attention to the howlers that pass as legitimate criticisms of frequentist error statistics, but also to sketch the main lines of an alternative statistical philosophy within which to better articulate the roles and value of frequentist tools. Let me be clear that I do not consider this the only philosophical framework for frequentist statistics—different terminology could do as well. I will consider myself successful if I can provide one way of building, or one standpoint from which to build, a frequentist, error-statistical philosophy. Here I mostly sketch key ingredients and report on updates in a larger, ongoing project. Statistical Science Meets Philosophy of Science Part 2 73 2. Popperians Are to Frequentists as Carnapians Are to Bayesians Statisticians do, from time to time, allude to better-known philosophers of sci- ence (e.g., Popper). The familiar philosophy/statistics analogy—that Popper is to frequentists as Carnap is to Bayesians—is worth exploring more deeply, most notably the contrast between the popular conception of Popperian falsification and inductive probabilism. Popper himself remarked: “In opposition to [the] inductivist attitude, I assert that C(H,x) must not be interpreted as the degree of corroboration of H by x, unless x reports the results of our sincere efforts to overthrow H. The re- quirement of sincerity cannot be formalized—no more than the in- ductivist requirement that x must represent our total observational knowledge.” (Popper 1959, 418, I replace ‘e’ with ‘x’) In contrast with the more familiar reference to Popperian falsification, and its apparent similarity to statistical significance testing, here we see Popper allud- ing to failing to reject, or what he called the “corroboration” of hypothesis H. Popper chides the inductivist for making it too easy for agreements between data x and H to count as giving H a degree of confirmation. “Observations or experiments can be accepted as supporting a theory (or a hypothesis, or a scientific assertion) only if these observations or experiments are severe tests of the theory—or in other words, only if they result from serious attempts to refute the theory.” (Popper 1994, 89) (Note the similarity to Peirce in Mayo 2011, 87, this special topic of RMM.) 2.1 Severe Tests Popper did not mean to cash out ‘sincerity’ psychologically of course, but in some objective manner. Further, high corroboration must be ascertainable: ‘sincerely trying’ to find flaws will not suffice. Although Popper never adequately cashed out his intuition, there is clearly something right in this requirement. It is the gist of an experimental principle presumably accepted by Bayesians and fre- quentists alike, thereby supplying a minimal basis to philosophically scrutinize different methods. (Mayo 2011, section 2.5, this special topic of RMM) Error-statistical tests lend themselves to the philosophical standpoint re- flected in the severity demand. Pretty clearly, evidence is not being taken se- riously in appraising hypothesis H if it is predetermined that, even if H is false, a way would be found to either obtain, or interpret, data as agreeing with (or ‘passing’) hypothesis H. Here is one of many ways to state this: Severity Requirement (weakest): An agreement between data x and H fails to count as evidence for a hypothesis or claim H if the test 74 Deborah Mayo would yield (with high probability) so good an agreement even if H is false. Because such a test procedure had little or no ability to find flaws in H, finding none would scarcely count in H’s favor. 2.1.1 Example: Negative Pressure Tests on the Deep Horizon Rig Did the negative pressure readings provide ample evidence that: H0: leaking gases, if any, were within the bounds of safety (e.g., less than θ0)? Not if the rig workers kept decreasing the pressure until H passed, rather than performing a more stringent test (e.g., a so-called ‘cement bond log’ using acous- tics). Such a lowering of the hurdle for passing H0 made it too easy to pass H0 even if it was false, i.e., even if in fact: H1: the pressure build-up was in excess of θ0. That ‘the negative pressure readings were misinterpreted’, meant that it was incorrect to construe them as indicating H0. If such negative readings would be expected, say, 80 percent of the time, even if H1 is true, then H0 might be said to have passed a test with only .2 severity. Using Popper’s nifty abbreviation, it could be said to have low corroboration, .2. So the error probability associ- ated with the inference to H1 would be .8—clearly high. This is not a posterior probability, but it does just what we want it to do. 2.2 Another Egregious Violation of the Severity Requirement Too readily interpreting data as agreeing with or fitting hypothesis H is not the only way to violate the severity requirement. Using utterly irrelevant evi- dence, such as the result of a coin flip to appraise a diabetes treatment, would be another way. In order for data x to succeed in corroborating H with severity, two things are required: (i) x must fit H, for an adequate notion of fit, and (ii) the test must have a reasonable probability of finding worse agreement with H, were H false.
Recommended publications
  • The Interpretation of Probability: Still an Open Issue? 1
    philosophies Article The Interpretation of Probability: Still an Open Issue? 1 Maria Carla Galavotti Department of Philosophy and Communication, University of Bologna, Via Zamboni 38, 40126 Bologna, Italy; [email protected] Received: 19 July 2017; Accepted: 19 August 2017; Published: 29 August 2017 Abstract: Probability as understood today, namely as a quantitative notion expressible by means of a function ranging in the interval between 0–1, took shape in the mid-17th century, and presents both a mathematical and a philosophical aspect. Of these two sides, the second is by far the most controversial, and fuels a heated debate, still ongoing. After a short historical sketch of the birth and developments of probability, its major interpretations are outlined, by referring to the work of their most prominent representatives. The final section addresses the question of whether any of such interpretations can presently be considered predominant, which is answered in the negative. Keywords: probability; classical theory; frequentism; logicism; subjectivism; propensity 1. A Long Story Made Short Probability, taken as a quantitative notion whose value ranges in the interval between 0 and 1, emerged around the middle of the 17th century thanks to the work of two leading French mathematicians: Blaise Pascal and Pierre Fermat. According to a well-known anecdote: “a problem about games of chance proposed to an austere Jansenist by a man of the world was the origin of the calculus of probabilities”2. The ‘man of the world’ was the French gentleman Chevalier de Méré, a conspicuous figure at the court of Louis XIV, who asked Pascal—the ‘austere Jansenist’—the solution to some questions regarding gambling, such as how many dice tosses are needed to have a fair chance to obtain a double-six, or how the players should divide the stakes if a game is interrupted.
    [Show full text]
  • PHIL 8670 (Fall, 2015): Philosophy of Statistics
    PHIL 8670 (Fall, 2015): Philosophy of Statistics Samuel C. Fletcher September 12, 2015 Primary Email: scfl[email protected] Office Phone: 612-625-4325 Office Hours: 754 Heller Hall, Tuesdays 3 pm { 4 pm, and by appointment Class Period: 731 Heller Hall, Tuesdays 4 pm { 6 pm Course Description Statistical methods form a significant part of the methodology of every scientific discipline. This seminar will survey a number of central philosophical and conceptual issues in statistics. After a brief review of elementary probability theory and its interpretation, we will consider at length both the classical and Bayesian schools of statistics: their formulations, and arguments pro and con. We will then consider a third school, likelihoodism, that is sometimes considered a compromise between classical and Bayesian methodologies and has gained favor among some philosophers of science. This leads to the likelihood principle, a natural focal point in the debate between these schools on the nature of evidence. The remainder of the seminar will be split between topics on statistical modeling (model verification and model selection), connections between the philosophy of statistics and more traditional topics in the philosophy of science (simplicity, induction and confirmation) and the relation between statistics and causation (randomness, randomization and causal inference). Evaluation Each week, there will be a semi-formal presentation of that week's reading assignment aimed at stimulating a critical discussion. For the first meeting (or two) I will be responsible for presenting, and for each of the remaining meetings, one or two students will be responsible. Students not presenting during a given week are responsible instead for submitting a one- paragraph question about a philosophical/conceptual issue arising in that week's readings to those presenting (and me).
    [Show full text]
  • Fusion Confusion? Comments on Nancy Reid: “BFF Four–Are We Converging?”
    Fusion Confusion? Comments on Nancy Reid: “BFF Four–Are we Converging?” Deborah G. Mayo The Fourth Bayesian, Fiducial and Frequentist Workshop (BFF4): Harvard University May 2, 2017 <1> I’m delighted to Be part of a workshop linking statistics and philosophy of statistics! I thank the organizers for inviting me. Nancy Reid’s “BFF Four–Are we Converging?” gives numerous avenues for discussion She zeroes in on obstacles to fusion: Confusion or disagreement on the nature of probability and its use in statistical inference <2> From Nancy Reid: Nature of probability probability to describe physical haphazard variability • probabilities represent features of the “real” world in idealized form • suBject to empirical test and improvement • conclusions of statistical analysis expressed in terms of interpretaBle parameters • enhanced understanding of the data generating process probability to describe the uncertainty of knowledge • measures rational, supposedly impersonal, degree of Belief given relevant information (Jeffreys) • measures a particular person’s degree of Belief, suBject typically to some constraints of self-consistency … • often linked with personal decision-making <3> • As is common, she laBels the second “epistemological” But a key question for me is: what’s relevant for a normative epistemology, for an account of what’s warranted/unwarranted to infer <4> Reid quite rightly asks: • in what sense are confidence distribution functions, significance functions, structural or fiducial probaBilities to be interpreted? • empirically? degree of Belief? • literature is not very clear <5> Reid: We may avoid the need for a different version of proBaBility by appeal to a notion of calibration (Cox 2006, Reid & Cox 2015) • This is my central focus I approach this indirectly, with analogy between philosophy of statistics and statistics <6> Carnap: Bayesians as Popper: Frequentists (N-P/Fisher) Can’t solve induction But can Build logics of induction or confirmation theories (e.g., Carnap 1962).
    [Show full text]
  • Teaching Statistics in a Critical Way: Historical, Philosophical and Political Aspects of Statistics
    ICOTS-7, 2006: François and Bracke (Refereed) TEACHING STATISTICS IN A CRITICAL WAY: HISTORICAL, PHILOSOPHICAL AND POLITICAL ASPECTS OF STATISTICS Karen François Free University Brussels, Belgium Nele Bracke Ghent University, Belgium [email protected] In this paper we sketch the history and the philosophy of statistics and probability theory and the connections to its political aspects. Knowledge of the cultural embeddedness of statistics and probability theory is an added value in the teaching thereof. The use of statistics and probability is a phenomenon with which everyone is confronted on a daily basis. Beside literacy, numeracy is an important challenge for education. In order to succeed in this task, the traditional curriculum (technique-oriented and individual, competition-oriented) will need to be sacrificed for a curriculum in which there is room for the cultural aspects of statistics and probability theory. For this purpose, cooperative learning is a didactic entry which is suitable for interaction and critical input of pupils. INTRODUCTION Looking at the history of statistics, one can show the growing importance of statistics in social sciences and in society in general. Understanding statistical representations, being able to interpret graphs or to give meaning to information in figures today are part of the skills that are generally assumed. Education, and particularly education in probability and statistics, face the challenge of helping to develop this competence. If education wishes to take up part of the responsibility to bring about a democratic state in which all have access to basic information like the media, then the development of a working knowledge of statistics and probability should form part of the curriculum.
    [Show full text]
  • Is the Philosophy of Probabilism an Obstacle to Statistical Fraud Busting? Deborah G
    Fraudbusting Feb 20, 2014 Mayo / 1 Is the Philosophy of Probabilism an Obstacle to Statistical Fraud Busting? Deborah G. Mayo 2013 was the “year of celebrating statistics”, but it might well have been dubbed the “year of frustration with statistics” Well-worn criticisms of how easy it is to lie with statistics have their own slogans: o Association is not causation. o Statistical significance is not substantive significance. o No evidence of risk is not evidence of no risk. o If you torture the data enough, they will confess. Fraudbusting Feb 20, 2014 Mayo / 2 Professional manuals and treatises written for popular consumption are rife with exposés of fallacies and foibles. • Task forces are organized, cottage industries regularly arise to expose the lack of replication and all manner of selection and publication biases. • The rise of “big data” might have made cookbook statistics easier but these fallacies were addressed by the founders of the tools. Fraudbusting Feb 20, 2014 Mayo / 3 R. A. Fisher (birthday: Feb 17) observed that: [t]he political principle that anything can be proved by statistics arises from the practice of presenting only a selected subset of the data available” (Fisher 1955, 75). • However, nowadays it’s the tools that are blamed instead of the abuser. • We don’t need “scaling up” so much as scaling back many misdiagnoses of what is going wrong. • In one sense, it’s blatantly obvious: sufficient finagling may practically predetermine that a researcher’s favorite hypothesis gets support, even if it’s unwarranted by evidence. Fraudbusting Feb 20, 2014 Mayo / 4 2.
    [Show full text]
  • Induction and Deduction in Bayesian Data Analysis*
    RMM Vol. 2, 2011, 67–78 Special Topic: Statistical Science and Philosophy of Science Edited by Deborah G. Mayo, Aris Spanos and Kent W. Staley http://www.rmm-journal.de/ Andrew Gelman Induction and Deduction in Bayesian Data Analysis* Abstract: The classical or frequentist approach to statistics (in which inference is centered on sig- nificance testing), is associated with a philosophy in which science is deductive and fol- lows Popper’s doctrine of falsification. In contrast, Bayesian inference is commonly asso- ciated with inductive reasoning and the idea that a model can be dethroned by a compet- ing model but can never be directly falsified by a significance test. The purpose of this article is to break these associations, which I think are incorrect and have been detrimen- tal to statistical practice, in that they have steered falsificationists away from the very useful tools of Bayesian inference and have discouraged Bayesians from checking the fit of their models. From my experience using and developing Bayesian methods in social and environmental science, I have found model checking and falsification to be central in the modeling process. 1. The Standard View of the Philosophy of Statistics, and Its Malign Influence on Statistical Practice Statisticians can be roughly divided into two camps, each with a clear configu- ration of practice and philosophy. I will divide some of the relevant adjectives into two columns: Frequentist Bayesian Objective Subjective Procedures Models P-values Bayes factors Deduction Induction Falsification Pr (model is true) This division is not absolute—in particular, it does not capture my own mix of philosophy and statistical practice—but it reflects a common alignment of be- liefs.
    [Show full text]
  • Bayes, Jeffreys, Prior Distributions and the Philosophy of Statistics
    Statistical Science 2009, Vol. 24, No. 2, 176–178 DOI: 10.1214/09-STS284D Main article DOI: 10.1214/09-STS284 c Institute of Mathematical Statistics, 2009 Bayes, Jeffreys, Prior Distributions and the Philosophy of Statistics1 Andrew Gelman I actually own a copy of Harold Jeffreys’s The- tion of Jeffreys’s ideas is to explicitly include model ory of Probability but have only read small bits of checking in the process of data analysis. it, most recently over a decade ago to confirm that, indeed, Jeffreys was not too proud to use a classi- THE ROLE OF THE PRIOR DISTRIBUTION cal chi-squared p-value when he wanted to check the IN BAYESIAN DATA ANALYSIS misfit of a model to data (Gelman, Meng and Stern, At least in the field of statistics, Jeffreys is best 2006). I do, however, feel that it is important to un- known for his eponymous prior distribution and, derstand where our probability models come from, more generally, for the principle of constructing non- and I welcome the opportunity to use the present ar- informative, or minimally informative, or objective, ticle by Robert, Chopin and Rousseau as a platform 2 or reference prior distributions from the likelihood for further discussion of foundational issues. (see, for example, Kass and Wasserman, 1996). But In this brief discussion I will argue the follow- it can notoriously difficult to choose among nonin- ing: (1) in thinking about prior distributions, we formative priors; and, even more importantly, seem- should go beyond Jeffreys’s principles and move to- ingly noninformative distributions can sometimes have ward weakly informative priors; (2) it is natural for strong and undesirable implications, as I have found those of us who work in social and computational in my own experience (Gelman, 1996, 2006).
    [Show full text]
  • Frequentist Statistics As a Theory of Inductive Inference
    IMS Lecture Notes–Monograph Series 2nd Lehmann Symposium – Optimality Vol. 49 (2006) 77–97 c Institute of Mathematical Statistics, 2006 DOI: 10.1214/074921706000000400 Frequentist statistics as a theory of inductive inference Deborah G. Mayo1 and D. R. Cox2 Viriginia Tech and Nuffield College, Oxford Abstract: After some general remarks about the interrelation between philo- sophical and statistical thinking, the discussion centres largely on significance tests. These are defined as the calculation of p-values rather than as formal procedures for “acceptance” and “rejection.” A number of types of null hypoth- esis are described and a principle for evidential interpretation set out governing the implications of p-values in the specific circumstances of each application, as contrasted with a long-run interpretation. A variety of more complicated situations are discussed in which modification of the simple p-value may be essential. 1. Statistics and inductive philosophy 1.1. What is the Philosophy of Statistics? The philosophical foundations of statistics may be regarded as the study of the epistemological, conceptual and logical problems revolving around the use and in- terpretation of statistical methods, broadly conceived. As with other domains of philosophy of science, work in statistical science progresses largely without wor- rying about “philosophical foundations”. Nevertheless, even in statistical practice, debates about the different approaches to statistical analysis may influence and be influenced by general issues of the nature of inductive-statistical inference, and thus are concerned with foundational or philosophical matters. Even those who are largely concerned with applications are often interested in identifying general prin- ciples that underlie and justify the procedures they have come to value on relatively pragmatic grounds.
    [Show full text]
  • The Philosophy of Statistics Author(S): Dennis V
    The Philosophy of Statistics Author(s): Dennis V. Lindley Reviewed work(s): Source: Journal of the Royal Statistical Society. Series D (The Statistician), Vol. 49, No. 3 (2000), pp. 293-337 Published by: Wiley-Blackwell for the Royal Statistical Society Stable URL: http://www.jstor.org/stable/2681060 . Accessed: 05/09/2012 14:41 Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at . http://www.jstor.org/page/info/about/policies/terms.jsp . JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Wiley-Blackwell and Royal Statistical Society are collaborating with JSTOR to digitize, preserve and extend access to Journal of the Royal Statistical Society. Series D (The Statistician). http://www.jstor.org The Statistician(2000) 49, Part3, pp. 293-337 The philosophy of statistics Dennis V. Lindley Minehead,UK [Received June 1999] Summary. This paper puts forwardan overallview of statistics.It is argued that statisticsis the studyof uncertainty.The many demonstrationsthat uncertaintiescan only combine accordingto the rules of the probabilitycalculus are summarized.The conclusion is thatstatistical inference is firmlybased on probabilityalone. Progress is thereforedependent on the constructionof a probabilitymodel; methods for doing this are considered. It is argued that the probabilitiesare personal. The roles of likelihoodand exchangeabilityare explained. Inferenceis onlyof value ifit can be used, so the extensionto decision analysis, incorporatingutility, is relatedto riskand to the use of statisticsin science and law.
    [Show full text]
  • Statistical Inference: the Big Picture1 Robert E
    Statistical Science 2011, Vol. 26, No. 1, 1–9 DOI: 10.1214/10-STS337 © Institute of Mathematical Statistics, 2011 Statistical Inference: The Big Picture1 Robert E. Kass Abstract. Statistics has moved beyond the frequentist-Bayesian controver- sies of the past. Where does this leave our ability to interpret results? I suggest that a philosophy compatible with statistical practice, labeled here statisti- cal pragmatism, serves as a foundation for inference. Statistical pragmatism is inclusive and emphasizes the assumptions that connect statistical models with observed data. I argue that introductory courses often mischaracterize the process of statistical inference and I propose an alternative “big picture” depiction. Key words and phrases: Bayesian, confidence, frequentist, statistical edu- cation, statistical pragmatism, statistical significance. 1. INTRODUCTION ple, these caricatures forfeit an opportunity to articulate a fundamental attitude of statistical practice. The protracted battle for the foundations of statis- Most modern practitioners have, I think, an open- tics, joined vociferously by Fisher, Jeffreys, Neyman, minded view about alternative modes of inference, but Savage and many disciples, has been deeply illuminat- are acutely aware of theoretical assumptions and the ing, but it has left statistics without a philosophy that matches contemporary attitudes. Because each camp many ways they may be mistaken. I would suggest that took as its goal exclusive ownership of inference, each it makes more sense to place in the center of our log- was doomed to failure. We have all, or nearly all, ical framework the match or mismatch of theoretical moved past these old debates, yet our textbook expla- assumptions with the real world of data.
    [Show full text]
  • 1 Excerpts from Stanford's Encyclopedia of Philosophy Article
    Excerpts from Stanford’s Encyclopedia of Philosophy article on the Philosophy of Statistics article at http://plato.stanford.edu/entries/statistics/ Now what does the probability function mean? The mathematical notion of probability does not provide an answer. The function P may be interpreted as physical, namely the frequency or propensity of the occurrence of a state of affairs, often referred to as the chance, or else as epistemic, namely the degree of belief in the occurrence of the state of affairs, the willingness to act on its assumption, a degree of support or confirmation, or similar. Physical probability and classical statistics In the sciences, the idea that probabilities express physical states of affairs, often called chances or stochastic processes, is most prominent. They are relative frequencies in series of events or, alternatively, they are tendencies or propensities in the systems that realize those events. More precisely, the probability attached to the property of an event type can be understood as the frequency or tendency with which that property manifests in a series of events of that type. The notion of physical probability is connected to one of the major theories of statistical method, which has come to be called classical statistics. It was developed roughly in the first half of the 20th century, mostly by mathematicians and working scientists like Fisher (1925, 1935, 1956), Wald (1939, 1950), Neyman and Pearson (1928, 1933, 1967)… Physical probability cannot meaningfully be attributed to statistical hypotheses, since hypotheses do not have tendencies or frequencies with which they come about: they are categorically true or false, once and for all.
    [Show full text]
  • Bayesian Philosophy of Science
    Jan Sprenger Stephan Hartmann Bayesian Philosophy of Science Variations on a Theme by the Reverend Thomas Bayes p(EjH) p(HjE) = p(H) p(E) Ich bin ein Esel, und will getreu, wie meine Väter, die Alten, an der alten, lieben Eselei, am Eseltume halten. Und weil ich ein Esel, so rat ich euch, den Esel zum König zu wählen; wir stiften das große Eselreich, wo nur die Esel befehlen. Wir sind alle Esel! I-A! I-A! Wir sind keine Pferdeknechte. Fort mit den Rossen! Es lebe—Hurra!— Der König vom Eselsgeschlechte! Heinrich Heine Contents Theme: Bayesian Philosophy of Science 1 Variation 1: Learning Conditional Information 31 1.1 The Kullback-Leibler Divergence and Probabilistic Updating 33 1.2 Three Challenges for Minimizing Divergence . 38 1.3 Meeting the Challenges . 40 1.4 Discussion . 48 1.5 Proofs of the Theorems . 51 Variation 2: Confirmation 57 2.1 Motivating Bayesian Confirmation Theory . 58 2.2 Confirmation as (Increase in) Firmness . 59 2.3 The Plurality of Bayesian Confirmation Measures . 69 2.4 Discussion . 73 Variation 3: The Problem of Old Evidence 77 3.1 The Dynamic POE: The GJN Approach . 79 3.2 Solving the Dynamic POE: Alternative Explanations . 83 3.3 The Static POE: A Counterfactual Perspective . 86 3.4 Solving the Hybrid POE: Learning Explanatory Relationships 88 3.5 Discussion . 92 3.6 Proofs of the Theorems . 96 Variation 4: The No Alternatives Argument 101 4.1 Modeling the No Alternatives Argument . 102 4.2 Results and Discussion . 107 4.3 Proof of the Theorems . 111 v vi Contents Variation 5: Scientific Realism and the No Miracles Argument 115 5.1 The Probabilistic No Miracles Argument .
    [Show full text]