Enhanced Step-By-Step Approach for the Use of Big Data in Modelling for Official Statistics

Total Page:16

File Type:pdf, Size:1020Kb

Enhanced Step-By-Step Approach for the Use of Big Data in Modelling for Official Statistics Enhanced step-by-step approach for the use of big data in modelling for official statistics Dario Buono1, George Kapetanios2, Massimiliano Marcellino3, Gian Luigi Mazzi4, Fotis Papailias5 Paper prepared for the 16th Conference of IAOS OECD Headquarters, Paris, France, 19-21 September 2018 Session 1.A., Day 1, 19/09, 11:00: Use of Big Data for compiling statistics 1 Eurostat, European Commission. Email: [email protected] 2 King’s College London. Email: [email protected] 3 Bocconi University. Email: [email protected] 4 GOPA. Email: [email protected] 5 King’s College London. Email: [email protected] Dario Buono [email protected], European Commission George Kapetanios [email protected] King’s College London Massimiliano Marcellino [email protected] Bocconi University Gian Luigi Mazzi [email protected] GOPA Fotis Papailias [email protected] King’s College London Enhanced step-by-step approach for the use of big data in modelling for official statistics DRAFT VERSION DD/MM/2018 PLEASE DO NOT CITE (optional) Prepared for the 16th Conference of the International Association of Official Statisticians (IAOS) OECD Headquarters, Paris, France, 19-21 September 2018 Note (optional): This Working Paper should not be reported as representing the views of the Author organisation/s. The views expressed are those of the author(s). ABSTRACT In this paper we describe a step-by-step approach for the use of big data in nowcasting exercises and, possibly, in the construction of high frequency and new indicators. We also provide a set of recommendations associated to each step which are based on the outcome of the other tasks of the project as well as on the outcome of the previous project on big data and macroeconomic nowcasting. Specifically, we consider issues related to data sources, access, preparation, cleaning/filtering, modelling, and evaluation. Keywords: Big Data, Macroeconomic Nowcasting, Econometric Models, construction of new indicators; Outliers; Seasonality; Unstructured data. 3 1 INTRODUCTION Recent advances in IT and the social use of Internet related applications make nowadays a vast amount of information available in real time. The associated Big Data are potentially very useful for a variety of applications, ranging from marketing to tackling tax evasion. From the point of view of statistical institutes, the main question is whether and to what extent Big Data are valuable to expand, check and improve the data production process. In this paper we focus on three particular cases of using Big Data in official statistics. The first one is macroeconomic nowcasting, thus potentially enhancing the timely availability and precision of early estimates of key macroeconomic variables, and possibly providing new Big Data based coincident and leading indicators of economic activity. The second is represented by the possibility of using big data in the construction of new high frequency versions of already existing macroeconomic indicators. This could, for example, concern the estimation of daily or weekly proxies of GDP or of inflation. An example can be found in Aprigliano et al. (2016). The last one concerns the possibility of using big data to construct new qualitative or quantitative indicators measuring phenomena for which official statistics had problems to find appropriate information to satisfactorily measuring or describing them. This could be the case of pressure indicators (like for prices or for consumers) and of wellbeing ones. While clearly potentially of important value in the contexts mentioned above, Big Data still raises a number of old and new issues, related to data availability and accessibility, data preparation, cleaning and filtering, data modelling, validation and evaluation. All these aspects have a great relevance in the context of official statistics. Furthermore, the use of Big Data also requires proper statistical and econometric modelling as well as evaluation techniques. A wide range of methods are available, but they should be used in a careful way as they have been developed either (in the econometric literature) to handle large but not really big datasets of time series, or (in the statistical and machine learning literatures) to model really big datasets of iid observations, while we would like to work with big sets of correlated time series. Broadly speaking, the available methods are based either on summarizing the many variables, or on selecting them, or on combining many small models, after a proper data pre-treatment. Obviously the above considerations apply to the case of numerical big data previously converted into a set of time-series. This approach is especially relevant for nowcasting exercises or even to construct high frequency indicators. Nevertheless when constructing new indicators, especially qualitative ones, it is possible that non-numerical big data (such as textual or image ones) have to be considered. In such situation, since the conversion into time series or even panel data is not feasible, we are obliged to work in unstructured data and the problems are even more difficult to solve. In this case, we need to base our process of construction on typically multivariate tools of data exploration such as data mining or data analytics (or their extensions such as text mining). The characteristics of these methods call for an even higher carefulness in considering the outcome of such investigations. Given this complex background, in this document we develop a step-by-step approach, ac- companied by specific recommendations for the use of big data for macroeconomic 4 nowcasting, construction of high frequency indicators and of new indicators. The approach proposed here in- tends to generalise and extend the one already proposed by Mazzi (2016) and Kapetanios et al. (2017). Besides the generalisation to different kind of usages of big data, the extensions proposed here mainly concern: • the analysis of the usefulness of various types of big data when constructing different indicators, • the methods for converting unstructured numerical big data into time series structured ones, • the seasonal and calendar adjustment, • the identification of advantages and drawbacks of statistical methods to be used when dealing with non-numerical big data. Keeping unchanged the philosophy used in the previous papers, the step-by-step approach pro- posed below is intended to guide researchers and statisticians from the identification and the choice of Big Data to its pre-treatment and econometric/statistical modelling until the construction of indicators and the comparative evaluation of results. With this step-by-step approach we aim to provide researchers and statisticians with a very useful, flexible and complete tool for decision making about: if, when, how and for what using big data in official statistics. More specifically: • Step 1 considers the clear specification of the statistical problem (i.e. nowcasting versus high frequency indicator versus new indicator); • Step 2 refers to the definition of an appropriate IT environment for collecting, storing, treating and analyzing big data; • Step 3 considers Big Data potential usefulness within the statistical exercises as defined in step 1; • Step 4 discusses issues related to the design of the exercise: possible big data sources, definition of the type of variables/information to be extracted from big-data, etc.; • Step 5 assesses big-data accessibility and the quality of envisaged big-data sources; • Step 6 focuses on big data specific features and on possible ways to move from unstructured to structured data; • Step 7 deals with the approaches to replace missing observations as well as to remove and correct outliers; • Step 8 deals with the ways for filtering out seasonal and other very short-term components from big data; • Step 9 concerns the presence of bias in big data and the ways to correct it; 5 • Step 10 analyzes the design of the econometric modelling strategy in case of the availability of numerical big data structured in time series form; • Step 11 considers the statistical exploratory techniques to be used in case of non- numerical or numerical but not structured big data; • Step 12 deals with the evaluation of the results mainly obtained in steps 10 and 11 (choice of benchmark model or nowcasts, evaluation metrics, criteria for assessing the significance of the accuracy and/or timeliness gains, the statistical and economic validity of the results, etc.); • Step 13 considers the overall implementation of big data based results within the regular production and dissemination process; All these steps are discussed in the various subsections of Section 2. Each subsection is structured into two parts: the first one providing a description of the problem and related issues and the second one providing specific recommendations. Section 3 proposes a more compact presentation of the step-by-step approach and briefly concludes. 2 A STEP-BY-STEP APPROACH In this section the 13 steps listed above are presented in details, focusing on the main theoretical and empirical aspects characterizing each step and deriving recommendations reflecting the experience of the authors in dealing with big data as well as some relevant literature contributions in this field. 2.1 Step 1: Identification of the Statistical Problem 2.1.1 Description Before starting the search phase for identifying possibly useful big data, it is important to clarify in the most accurate way the problem we want to address. In particular, in case of a nowcasting exercise, it is important to specify the expected gain in terms of timeliness and the frequency for nowcasting updating (especially if the estimates already start during the reference period). In case of the estimation of high frequency indicators it is important to define the frequency at which the indicator is expected to be estimated as well as some feasible targets in terms of timeliness. For new indicators it is essential to clearly define what they are supposed to measure, if they should be qualitative or quantitative as well as the expected frequency of computation and, if relevant, the timelines of production.
Recommended publications
  • Computational Economics and Econometrics
    Computational Economics and Econometrics A. Ronald Gallant Penn State University c 2015 A. Ronald Gallant Course Website http://www.aronaldg.org/courses/compecon Go to website and discuss • Preassignment • Course plan (briefly now, in more detail a few slides later) • Source code • libscl • Lectures • Homework • Projects Course Objective – Intro Introduce modern methods of computation and numerical anal- ysis to enable students to solve computationally intensive prob- lems in economics, econometrics, and finance. The key concepts to be mastered are the following: • The object oriented programming style. • The use of standard data structures. • Implementation and use of a matrix class. • Implementation and use of numerical algorithms. • Practical applications. • Parallel processing. Details follow Object Oriented Programming Object oriented programming is a style of programming devel- oped to support modern computing projects. Much of the devel- opment was in the commercial sector. The features of interest to us are the following: • The computer code closely resembles the way we think about a problem. • The computer code is compartmentalized into objects that perform clearly specified tasks. Importantly, this allows one to work on one part of the code without having to remember how the other parts work: se- lective ignorance • One can use inheritance and virtual functions both to describe the project design as interfaces to its objects and to permit polymorphism. Inter- faces cause the compiler to enforce our design, relieving us of the chore. Polymorphism allows us to easily swap in and out objects so we can try different models, different algorithms, etc. • The structures used to implement objects are much more flexible than the minimalist types of non-object oriented language structures such as subroutines, functions, and static common storage.
    [Show full text]
  • Computational Topics in Economics and Econometrics Spring Semester 2003 Tuesday 12-2 Room 315 Christopher Flinn Room 302 Christopher.fl[email protected]
    Computational Topics in Economics and Econometrics Spring Semester 2003 Tuesday 12-2 Room 315 Christopher Flinn Room 302 christopher.fl[email protected] Introduction: The objective of this course is to develop computational modeling skills that will increase the stu- dent’s capacity for performing quantiative research in economics. The basic premise of the course is the following. To adequately explain social phenomena of interest to economists requires the use of relatively sophisticated, albeit stylized, models. These models, whether involving only the solution of individual decision rules in a dynamic context or the determination of equilibria in a general equilibrium model, can only be solved numerically. We will review techniques that allow the solution of a large class of fixed point problems, which virtually all economic problems can be cast as, and will illustrate and compare techniques in the context of diverse examples. As an integral part of the course, students will be expected to contribute problems of their own, hopefully taken from current research questions they are investigating. We will also study the use of computational-intensive techniques in econometrics. Over the past several decades there have been a lot of exciting developments in this area due to enormous increases in processor speed and affordability. Topics in this area that we will treat include general computational issues in the estimation of nonlinear models, simulation estimation methods (maximum likelihood and method of moments), bootstrap methods, and the E-M algorithm. Examples from my own and other applied economists’ research will be used to illustrate the methods where appropriate. Because applied numerical analysis is at least one-half art form, it will be extremely useful for all of us to hear from individuals who have successfully completed large research projects that required them to solve daunting computational problems in the process.
    [Show full text]
  • Complexity Economics Theory and Computational Methods
    Complexity Economics Theory and Computational Methods Dr. Claudius Gräbner Dr. Torsten Heinrich Institute for the Comprehensive Analysis of the Economy (ICAE) Institute for New Economic Thinking (INET) & Oxford Martin School Johannes Kepler University Linz University of Oxford & & Institute for Socio-economics University of Bremen University of Duisburg-Essen & www.torsten-heinrich.com ZOE. Institute for Future-Fit Economies @TorstenHeinriX www.claudius-graebner.com @ClaudiusGraebner Summer Academy for Pluralism in Economics Exploring Economics 10.08.2018 Plan for first lecture • Getting to know ourselves • Clarify organizational issues and resolve open issues • Introduce and explain background of course outline • Start with meta-theoretical framework 2 Claudius Gräbner & Torsten Heinrich: Complexity Economics. Theory and Comp Methods @ Exploring Economics Summer Academy, Aug 9-16 2019 Intro Claudius Torsten • Studied economics in Dresden and Madrid • Studied „Staatswissenschaften“, i.e. economics, law, and social sciences in Erfurt • PhD and Habilitation in economics in Bremen, Germany • PhD in economics in Bremen, Germany • Now: Post Doc at Oxford University • Several stays in Santa Fe, USA • Institute for New Economic Thinking (INET), • Now: Post Doc at the Johannes Kepler University in Oxford Martin School (OMS) & Wolfson College Linz, Austria & University of Duisburg-Essen, Germany • Main interests: • Main interests: • Complex systems science and agent-based • Economic development, technological change computational economics and
    [Show full text]
  • Simulating Inflation 1
    Running head: SIMULATING INFLATION 1 Simulating Inflation: Is the Economy a Big Balloon? Ryan Heathcote A Senior Thesis submitted in partial fulfillment of the requirements for graduation in the Honors Program Liberty University Spring 2013 SIMULATING INFLATION 2 Acceptance of Senior Honors Thesis This Senior Honors Thesis is accepted in partial fulfillment of the requirements for graduation from the Honors Program of Liberty University. ______________________________ Mark Shaneck, Ph.D. Thesis Chair ______________________________ Terry Metzgar, Ph.D. Committee Member ______________________________ Monty Kester, Ph.D. Committee Member ______________________________ James H. Nutter, D.A. Honors Director ______________________________ Date SIMULATING INFLATION 3 Abstract Inflation is a common problem in modern economics, and it seems to persist, requiring government financial policy intervention on a regular basis in order to properly manage. The mechanics of inflation are difficult to understand, since the best metric modern society has for inflation involves taking samples of prices across the country. A simulation can give deeper insight into more than the mere fact that prices rise over time: a simulation can help to answer the “why” question. The problem with this statement is that developing a simulation is not as simple as writing some code. A simulation is developed according to a paradigm, and paradigms are not created equal. Research reveals that traditional paradigms that impose order from the outside do not mimic reality and thus do not give a good degree of accuracy. By contrast, Agent Based Modelling models reality very closely and is consequently able to simulate economic systems much more accurately. SIMULATING INFLATION 4 Simulating Inflation: Is the Economy a Big Balloon? It is often said that life’s two inevitabilities are death and taxes.
    [Show full text]
  • Dr Martin Kliem
    Martin Kliem Deutsche Bundesbank Office Phone: +49-(0)69-9566-4759 Research Centre Mail: [email protected] Wilhelm-Epstein-Straße 14 60431 Frankfurt am Main Germany EDUCATION 05/2005 - 07/2009 Doctorate in Economics at the Humboldt-Universität zu Berlin (Ph.D. equivalent) Advisor: Prof. Harald Uhlig, Ph.D. Thesis: "Essays on Asset Pricing and the Macroeconomy" 10/1998 - 08/2004 Diploma in Business Administration at Humboldt-Universität zu Berlin (M.A. equivalent) 10/1998 - 04/2004 Diploma in Economics at Humboldt-Universität zu Berlin (M.A. equivalent) WORK EXPERIENCE 04/2009 - present Economist, Deutsche Bundesbank, Research Centre 09/2014 - 11/2014 European University Institute, Florence Pierre Werner Fellow at the Robert Schuman Centre of Advanced Studies 05/2005 - 03/2009 Humboldt-Universität zu Berlin Research associate at the Collaborative Research Center 649 "Eco- nomic Risk" 09/2004 - 04/2005 Consultant, J.O.R. FinConsult GmbH, Berlin RESEARCH INTERESTS Macroeconomics, Applied Macroeconometrics, Asset Pricing, Fiscal and Monetary Policy 1 PUBLICATIONS Monetary-fiscal policy interaction and fiscal inflation: a tale of three countries, with Alexan- der Kriwoluzky and Samad Sarferaz, European Economic Review, Vol. 88, 158-184, 2016. On the low-frequency relationship between public deficits and inflation, with Alexander Kri- woluzky and Samad Sarferaz, Journal of Applied Econometrics, Vol. 31(3), 566-583, 2016. Bayesian estimation of a dynamic stochastic general equlibrium model with asset prices, with Harald Uhlig, Quantitative Economcis, Vol. 7(1), 257-287, 2016. Toward a Taylor rule for fiscal policy, with A. Kriwoluzky, Review of Economic Dynamics, Vol. 17(2), 294-302, 2014. Assessing macro-financial linkages: A model comparison exercise, with R.
    [Show full text]
  • Agent-Based Computational Economics Documentation Release 0.9.3B0
    Agent-Based Computational Economics Documentation Release 0.9.3b0 Davoud Taghawi-Nejad Jul 24, 2021 Contents 1 Introduction 3 1.1 Design..................................................3 1.2 Download and Installation........................................7 1.3 Interactive jupyter notebook Tutorial..................................9 1.4 Walk through...............................................9 1.5 Tutorial for Plant Modeling....................................... 16 1.6 Examples................................................. 20 1.7 unit testing................................................ 24 2 Simulation Programming 25 2.1 The simulation in start.py........................................ 25 2.2 Agents.................................................. 28 2.3 Groups.................................................. 29 2.4 Physical goods and services....................................... 31 2.5 Trade................................................... 32 2.6 Messaging................................................ 32 2.7 Firm and production........................................... 32 2.8 Household and consumption....................................... 35 2.9 Observing agents and logging...................................... 36 2.10 Retrieval of the simulation results.................................... 37 2.11 NotEnoughGoods Exception....................................... 38 3 Advanced 39 3.1 Quote................................................... 39 3.2 Spatial and Netlogo like Models..................................... 39
    [Show full text]
  • Economy-Wide Modeling: Evaluating the Economic Impacts of Air Regulations
    Economy-Wide Modeling: Evaluating the Economic Impacts of Air Regulations DRAFT: June 14, 2016 Prepared for the U.S. EPA Science Advisory Board Panel (SAB) on Economy-Wide Modeling of the Benefits and Costs of Environmental Regulation This paper has been developed to inform the deliberations of the SAB Panel on the technical merits and challenges of economy-wide modeling for an air regulation. It is not an official U.S. Environmental Protection Agency (EPA) report nor does it necessarily represent official policies or views of the U.S. EPA. 1 Table of Contents 1. Introduction .......................................................................................................................................... 4 2. Measuring Economic Impacts and the Role of CGE Modeling at EPA .................................................. 5 3. Quantitative evaluation of economic impacts by EPA ........................................................................ 10 3.1 Regional or State-Implemented Emission Targets ............................................................................ 11 3.2 Other Types of Air Regulations ......................................................................................................... 13 3.2.1 Single Sector Emission Rate Limits ............................................................................................. 14 3.2.2 Multi-Sector Boiler or Engine-Level Emission Limits ................................................................. 15 3.3.3 Federal Product Standards ........................................................................................................
    [Show full text]
  • Lawrence J. Christiano
    1 August, 2016 LAWRENCE J. CHRISTIANO VITA Personal Data Home Address Office Address 2517 Thornwood Department of Economics Wilmette, Illinois 60091 Northwestern University (847) 251-5710 Evanston, Illinois 60208 Citizenship: U.S.A. (847) 491-8231 Education Degree Field Institution Year Ph.D. Economics Columbia University 1982 M.Sc. Econometrics/Mathematical London School of Economics 1977 Economics M.A. Economics University of Minnesota 1975 B.A. History/Economics University of Minnesota 1973 Experience September 1 2002 – present: Alfred W. Chase Chair in Business Institutions, Northwestern University. Summers 1999, 2000, 2001: Visiting Scholar, International Monetary Fund. Summers, 2001 and 2002: Visiting Scholar, European Central Bank. Fall 1999 – present: Short courses on ‘Monetary Economics’ at Stockholm School of Economics, International Monetary Fund, Banca D’Italia, Central Bank of Portugal, Central Bank of Chile, European Central Bank, University of Bonn, New University of Lisbon, Humboldt University, Bocconi University, Massachusetts Institute of Technology, Study Center Gerzensee, Central Bank of Israel, Central Bank of Turkey, European Business Cycle Network, Brazilian Central Bank, Reserve Bank of New Zealand, Swiss National Bank, University of Amsterdam, University of Toronto, American Economic Association, Central Bank of Peru, Central Bank of Colombia, Central 2 Bank of Hungary, Central Bank of Czech Republic, Kiel Institute, Central Bank of Korea, Hanqing Advanced Institution of Economics and Finance, Renmin University of China, Shanghai Advanced Institute of Finance. Summer 1996 – 2007, consultant, Federal Reserve Bank of Cleveland. Fall 2008 - present, consultant, Federal Reserve Bank of Atlanta. July, 1995: consultant, Board of Governors, Federal Reserve System. August 1993, May-June 1998, Summers 1999, 2000, 2001: Visiting Scholar, International Monetary Fund.
    [Show full text]
  • Preface to the Special Issue on Computational Economics
    OPERATIONS RESEARCH informs Vol. 58, No. 4, Part 2 of 2, July–August 2010, pp. 1035–1036 ® issn 0030-364X eissn 1526-5463 10 5804 1035 doi 10.1287/opre.1100.0842 © 2010 INFORMS Preface to the Special Issue on Computational Economics Kenneth Judd Revolution and Peace, Hoover Institution on War, Stanford, California 94305, [email protected] Garrett van Ryzin Graduate School of Business, Columbia University, New York, New York 10027, [email protected] Economics is thestudy of how scarceresourcesareallo- Costs,” by Schulz and Uhan, looks at thecomplexityof cated. Operations research studies how to accomplish goals computing theleastcorevaluein cooperativegameswith in the least costly manner. These fields have much to offer supermodular cost structures. Such games arise in a variety each other in terms of challenging problems that need to of operations research problems—scheduling, for example. be solved and the techniques to solve them. This was the Theauthors show that such problemsarestrongly NP hard, caseafterWorld War II, partly becausetheindividuals who but for a special case of single-machine scheduling the went on to be the leading scholars in economics and oper- Shapleyvalueis in theleastcore. ations research worked together during WWII. In fact, the “Multigrid Techniques in Economics,” by Speight, is a two fields share many early luminaries, including Arrow, good example of how advances in computational physics, Dantzig, Holt, Kantorovich, Koopmans, Modigliani, Scarf, in this casecomputational fluid dynamics, can beapplied and von Neumann. Unfortunately, as new generations of to economic problems. Multigrid methods are a relatively scholars took charge, those ties weakened, resulting in lit- new approach to solving partial differential equations, and tle interaction between operations research scholars and this paper explains how they can be applied to dynamic economists in the past three decades.
    [Show full text]
  • Applied Computational Economics Spring 2019 Tuesday, 5:30 – 8:15 Pm Center City 906 ______Contact Information Instructor: Dr
    ECON 6257-U90 : Applied Computational Economics Spring 2019 Tuesday, 5:30 – 8:15 pm Center City 906 ______________________________________________________________________________ Contact Information Instructor: Dr. Hwan C. Lin Office: 217B Friday Building Office Hours: T, R, 1:00 – 2:30 pm (main campus) Office Phone: (704) 687-7646 E-mail: [email protected] Web: http://belkcollege.uncc.edu/hwlin/ Course Description This course introduces computational approaches for solving economic models. Topics include: interpolation and approximation techniques, numerical optimization, numerical solutions to systems of nonlinear equations, quadrature formulas for numerical integration, Monte Carlo simulation, and basic solution algorithms for economic dynamics. Graduate Credit (3). On Demand. Prerequisites: ECON 6201 & ECON 6202. Course Objectives The primary objective of the course is to teach how to compute equilibriums in economic models that cannot be solved analytically. It requires students to learn to formulate economic problems in computationally tractable forms, and then use numerical analysis techniques to solve them. After this class, students should be able to formulate economic models, design solution algorithms, calibrate models, and perform simulations and policy experiments. Instructional Method The course is offered with classroom lectures. A wide range of numerical methods and their applications to economic models will be discussed. Students are required to learn computer programming to implement numerical methods to solve economic problems. Computer Programming Language Students need to understand and use some programming languages. My suggestion is for students to learn Python, which is an open-source general-purpose programming language. With clear and simple syntax, Python is powerful for scientific computation. For computational purposes, you will need to download the following four core components: Python (http://python.org/), Numpy, Scipy (http://www.scipy.org/Download) Matplotlib (http://matplotlib.sourceforge.net/).
    [Show full text]
  • 1 6 Measuring and Interpreting the Economic Impact of Tourism: 20/20
    6 Measuring and Interpreting the Economic Impact of Tourism: 20/20 Hindsight and Foresight Douglas Frechtling and Egon Smeral Published in Douglas G. Pearce and Richard W. Butler (2010), TOURISM RESEARCH: A 20-20 VISION, Oxford, England: Goodfellow Publishing Measuring the economic consequences of tourism activities on national and sub-national (regional) areas provides valuable input for critically addressing issues of importance to at least four key stakeholder groups in the geographic area of interest. These groups are public officials, business owners and managers, employees of tourism establishments and other residents of the host communities. Before we address our topic, it is essential to define it. We had thought that standard English usage over the last three decades was pretty well established. As Reece (2010) recently affirmed, the economic impact of tourism is comprised of “changes in regional employment, incomes, tax payments, and other measures of economic activity . that result from a region’s tourism development” (p. 50), except that we understand this to apply to nations as well as regions. Reece follows a long train of authors who have defined “tourism economic impact” broadly (e.g., Goeldner and Ritchie 2003, Mason 2003, Smith 2000, Frechtling 1994). Some now claim that “tourism economic impact” measures only result from complex economic equilibrium models (Dwyer, et al., 2008), a clear break with tradition and the ways the term is commonly used in tourism research today. According to the tradition of tourism economic
    [Show full text]
  • Computational Economics and Economic Theory: Substitutes Or Complements? Kenneth L
    Computational Economics and Economic Theory: Substitutes or Complements? Kenneth L. Judd January 11, 1996∗ Abstract. This essay examines the idea and potential of a “compu- tational approach to theory,” discusses methodological issues raised by such computational methods, and outlines the problems associated with the dissem- ination of computational methods and the exposition of computational results. We argue that the study of a theory need not be conÞned to proving theo- rems, that current and future computer technologies create new possibilities for theoretical analysis, and that by resolving these issues we will create an intellectual atmosphere in which computational methods can make substantial contributions to economic analysis. Contents 1 ComputationandScience........................... 4 2 WhatCanEconomistsCompute?....................... 6 3 TheAdvantagesofDeductiveTheory..................... 8 4 TheoreticalAnalysisversusProvingTheorems................ 9 5 The Excessive Simplicity and Quantitative Irrelevance of Tractable Theo- reticalModels.................................. 11 6 The“BlackBox”Criticism........................... 12 7 The Real Issue: Where to Approximate, not Whether ............ 12 8 PartnersinAnalysis:StrongComplementarities............... 13 9 SettingPriorities:QuantitativeTestsforImportance............ 14 10TheoryasExploration............................. 15 11PerturbationMethods:CommonGround................... 18 12ComputationalAnalysesintheAbsenceofTheorems............ 20 13 Computational Considerations
    [Show full text]