Bayesian Methods for Solving Estimation and Forecasting Problems in the High-Frequency Trading Environment

Total Page:16

File Type:pdf, Size:1020Kb

Bayesian Methods for Solving Estimation and Forecasting Problems in the High-Frequency Trading Environment Bayesian methods for solving estimation and forecasting problems in the high-frequency trading environment Paul Alexander Bilokon Christ Church University of Oxford A thesis submitted in partial fulfillment of the MSc in Mathematical Finance 16 December, 2016 To God. And to my parents, Alex and Nataliya Bilokon. With love and gratitude. To the memory of Rudolf Emil Kálmán (1930–2016), who laid down the intellectual foundations on which this work is built. i Abstract We examine modern stochastic filtering and Markov chain Monte Carlo (MCMC) methods and consider their applications in finance, especially electronic trading. Stochastic filtering methods have found many applications, from Space Shuttles to self-driving cars. We review some classical and modern algorithms and show how they can be used to estimate and forecast econometric models, stochastic volatility and term structure of risky bonds. We discuss the practicalities, such as outlier filtering, parameter estimation, and diagnostics. We focus on one particular application, stochastic volatility with leverage, and show how recent advances in filtering methods can help in this application: kernel density estimation can be used to estimate the predicted observation, filter out outliers, detect structural change, and improve the root mean square error while preventing discontinuities due to the resampling step. We then take a closer look at the discretisation of the continuous-time stochastic volatility models and show how an alternative discretisation, based on what we call a filtering Euler–Maruyama scheme, together with our generalisation of Gaussian assumed density filters to arbitrary (not nec- essarily additive) correlated process and observation noises, gives rise to a new, very fast approxi- mate filter for stochastic volatility with leverage. Its accuracy falls short of particle filters but beats the unscented Kálmán filter. Due to its speed and reliance exclusively on scalar computations this filter will be particularly useful in a high-frequency trading environment. In the final chapter we examine the leverage effect in high-frequency trade data, using last data point interpolation, tick and wall-clock time and generalise the models to take into account the time intervals between the ticks. We use a combination of MCMC methods and particle filtering methods. The robustness of the latter helps estimate parameters and compute Bayes factors. The speed and precision of modern filtering algorithms enables real-time filtering and prediction of the state. Acknowledgements First and foremost, I would like to thank my supervisor, Dan Jones, who not only taught and instructed me academically, but also provided advice, guidance, and support in very challenging circumstances. Thank you very much, Dan, for putting up with me during this project. I would like to thank my mentors and former colleagues, Dan Crisan (Imperial College), Rob Smith (Citi- group), William Osborn (formerly Citigroup), and Martin Zinkin (formerly Deutsche Bank), who have taught me much of what I know about stochastic filtering. Rob not only taught me a lot, but was very helpful and understanding when I was going through a difficult time. Martin showed me a level of scientific and software development skill which I didn’t know existed. I would also like to thank my former manager at Deutsche Bank, Jason Batt, for his help and encouragement. I am very grateful to my PhD supervisor, Abbas Edalat, and scientific mentors, Claudio Albanese, Iain Clark, Achim Jung, Moe Hafiz, and Berc Rustem, for teaching me to think scientifically. I would also like to mention my quant and technology colleagues at Deutsche Bank Markets Electronic Trading who all, one way or another, taught me various aspects of the science, technology, and business of electronic trading: Tahsin Alam, Zar Amrolia, Benedict Carter, Teddy Cho, Luis Cota, Matt Daly, Madhucchand Darbha, Alex Doran, Marina Duma, Gareth Evans, Brian Gabillet, Yelena Gilbert, Thomas Giverin, Hakan Guney, James Gwinnutt, William Hall, Matt Harvey, Mridula Iyer, Chris-M Jones, Sergey Korovin, Egor Kraev, Leonid Kraginskiy, Peter Lee, Ewan Lowe, Phil Morris, Ian Muetzelfeldt, Roel Oomen, Helen O’Regan, Kir- ill Petunin, Stanislav Pustygin, Claude Rosenstrauch, Lubomir Schmidt, Torsten Schöneborn, Sumit Sengupta, Alexander Stepanov, Viktoriya Stepanova, Ian Thomson, Krzysztof Zylak. It has been an honour and privilege to work with you all while I was writing this dissertation. My friends have provided the much-needed encouragement and support: Sofiko Abanoidze, John and Anastasia Aston, Ester Goldberg, James Gwinnutt, Moe Hafiz, Lyudmila Meshchersky, Laly Nickatsadze, Adel Sabirova, Stephen Scott, Andrei Serjantov, Irakli Shanidze, Ed Silantyev, Nitza and Robin Spiro, Natasha Svetlova, Alexey Taycher, Nataliya Tkachuk, Jay See Tow, Douglas Vieira. Finally, I would like to acknowledge Saeed Amen, Matthew Dixon, Attila Agod, and Jan Novotny at Thale- sians, who have not only been great colleagues, but also great friends. I am especially grateful to Saeed for his patience, support, and friendship over the years. iii Nomenclature the Halmos symbol, which stands for quod erat demonstrandum A := f1, 2, 3g is equal by definition to (left to right) f1, 2, 3g =: A is equal by definition to (right to left) N natural numbers N0 ditto, explicitly incl. 0 N∗ nonzero natural numbers R reals, (−¥, +¥) R+ positive reals, (0, +¥) + R0 nonnegative reals, [0, +¥) + R0 extended nonnegative reals, [0, +¥] F, G,... name of a collection A, B,... name of a set fA, B,...g definition of a collection fO 2 T j O clopeng —"— f1, 2, 3g definition of a set fx 2 N j x eveng —"— (ak) name of element of a sequence (ak)k2N —"— ¥ (ak)k=1 —"— (2, 4, 6, . .) definition of a sequence (2, 4, 6) definition of a tuple x 2 N j x even such that iv Z ⊆ R subset R ⊇ Z superset Z ( R proper subset R ) Z proper superset A × B the Cartesian product of the sets A and B + C ! R0 function type + f : C ! R0 —"— x 7! x2 function definition f : x 7! x2 —"— f (A) the image of the set A under the function f f −1(A) the inverse image of the set A under the function f kxkp p-norm; in particular, when p = 2, Euclidean norm v name of a vector A name of a matrix v| transpose vec A vectorisation of a matrix s(A) the s-algebra generated by the family of sets A p(· j ·) a generic conditional probability density or mass function, with the arguments making it clear which conditional distribution it relates to T the time set X in the context of filtering, the state process S the state space B(S) the Borel s-algebra of S fX gt2T the usual augmentation of the filtration generated by X U (a, b) continuous uniform distribution with support [a, b] N m, s2 univariate normal (Gaussian) distribution with mean m and variance s2 N (m, S) multivariate normal (Gaussian) distribution with mean m and covariance matrix S v j x; m, s2, j m, s2 the probability density function (pdf), with argument x, of a univariate normal (Gaus- sian) distribution with mean m and variance s2 j (x; m, S), j (m, S) the probability density function (pdf), with argument x, of a multivariate normal (Gaussian) distribution with mean m and covariance matrix S AR autoregressive (process) ARIMA autoregressive integrated moving average (process) ARMA autoregressive moving average (process) ARCH autoregressive conditional heteroscedasticity BM geometric Brownian motion BUGS Bayesian inference Using Gibbs Sampling — a modelling language CUSUM Cumulative Sum of Recursive Residual DAG Directed Acyclic Graph EBFS empirical Bratley–Fox–Schrage EGARCH exponential generalised autoregressive conditional heteroscedasticity EKF extended Kálmán filter ELK empirical Law–Kelton EM Euler–Maruyama (scheme) GARCH generalised autoregressive conditional heteroscedasticity GASF Gaussian approximation recursive filter GBM geometric Brownian motion GDP gross domestic product HMM hidden Markov model i.i.d. independent and identically distributed (random variables) KDE kernel density estimator (estimate, estimation) KF Kálmán filter LTRF long-term relative frequency MA moving average (process) MCMC Markov chain Monte Carlo vi MISE mean integrated square error MLE maximum likelihood estimator (estimate, estimation) MMSE minimum mean square error estimator (estimate, estimation) MSE mean square error NLN normal–lognormal mixture (distribution) OU Ornstein-Uhlenbeck (process) post-RPF post-regularised particle filter QML quasi-maximum likelihood RMSE root mean square error SDE stochastic differential equation SIR sequential importance resampling SIS sequential importance sampling SISR sequential importance sampling with resampling SMC sequential Monte Carlo SV stochastic volatilty SVL stochastic volatility with leverage SVL2 stochastic volatility with the second correlation structure SVLJ stochastic volatility with leverage and jumps UKF unscented Kálmán filter wcSVL wall-clock stochastic volatility with leverage Z-spread zero volatility spread vii Contents Nomenclature iv Contents vii List of tables xi List of figures xii List of algorithms xiv List of listings xv Preface xvii 1 Volatility 1 1.1 Introduction . .1 1.2 The early days of volatility modelling: constant volatility models . .1 1.3 Evidence against the constant volatility assumption . .2 1.4 The two families of models . .3 1.4.1 The GARCH family . .3 1.4.2 The stochastic volatility family . .4 1.5 The statistical leverage effect . .6 1.6 A stochastic volatility model with leverage and jumps . .6 1.7 Conclusions and further reading . .9 viii 2 The filtering framework 11 2.1 Introduction . 11 2.2 Special case: general state-space models . 13 2.3 Particle filtering methods . 13 2.4 Applying the particle filter to the stochastic volatility model with leverage and jumps . 15 2.5 The Kálmán filter . 18 2.6 Some examples of linear-Gaussian state space models . 19 2.6.1 A non-financial example: the Newtonian system . 20 2.6.2 Autoregressive moving average models . 20 2.6.3 Continuous-time stochastic processes: the Wiener process, geometric Brownian motion, and the Ornstein–Uhlenbeck process .
Recommended publications
  • Two Principles of Evidence and Their Implications for the Philosophy of Scientific Method
    TWO PRINCIPLES OF EVIDENCE AND THEIR IMPLICATIONS FOR THE PHILOSOPHY OF SCIENTIFIC METHOD by Gregory Stephen Gandenberger BA, Philosophy, Washington University in St. Louis, 2009 MA, Statistics, University of Pittsburgh, 2014 Submitted to the Graduate Faculty of the Kenneth P. Dietrich School of Arts and Sciences in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Pittsburgh 2015 UNIVERSITY OF PITTSBURGH KENNETH P. DIETRICH SCHOOL OF ARTS AND SCIENCES This dissertation was presented by Gregory Stephen Gandenberger It was defended on April 14, 2015 and approved by Edouard Machery, Pittsburgh, Dietrich School of Arts and Sciences Satish Iyengar, Pittsburgh, Dietrich School of Arts and Sciences John Norton, Pittsburgh, Dietrich School of Arts and Sciences Teddy Seidenfeld, Carnegie Mellon University, Dietrich College of Humanities & Social Sciences James Woodward, Pittsburgh, Dietrich School of Arts and Sciences Dissertation Director: Edouard Machery, Pittsburgh, Dietrich School of Arts and Sciences ii Copyright © by Gregory Stephen Gandenberger 2015 iii TWO PRINCIPLES OF EVIDENCE AND THEIR IMPLICATIONS FOR THE PHILOSOPHY OF SCIENTIFIC METHOD Gregory Stephen Gandenberger, PhD University of Pittsburgh, 2015 The notion of evidence is of great importance, but there are substantial disagreements about how it should be understood. One major locus of disagreement is the Likelihood Principle, which says roughly that an observation supports a hypothesis to the extent that the hy- pothesis predicts it. The Likelihood Principle is supported by axiomatic arguments, but the frequentist methods that are most commonly used in science violate it. This dissertation advances debates about the Likelihood Principle, its near-corollary the Law of Likelihood, and related questions about statistical practice.
    [Show full text]
  • Testing for Cointegration When Some of The
    EconometricTheory, 11, 1995, 984-1014. Printed in the United States of America. TESTINGFOR COINTEGRATION WHEN SOME OF THE COINTEGRATINGVECTORS ARE PRESPECIFIED MICHAELT.K. HORVATH Stanford University MARKW. WATSON Princeton University Manyeconomic models imply that ratios, simpledifferences, or "spreads"of variablesare I(O).In these models, cointegratingvectors are composedof l's, O's,and - l's and containno unknownparameters. In this paper,we develop tests for cointegrationthat can be appliedwhen some of the cointegratingvec- tors are prespecifiedunder the null or underthe alternativehypotheses. These tests are constructedin a vectorerror correction model and are motivatedas Waldtests from a Gaussianversion of the model. Whenall of the cointegrat- ing vectorsare prespecifiedunder the alternative,the tests correspondto the standardWald tests for the inclusionof errorcorrection terms in the VAR. Modificationsof this basictest are developedwhen a subsetof the cointegrat- ing vectorscontain unknown parameters. The asymptoticnull distributionsof the statisticsare derived,critical values are determined,and the local power propertiesof the test are studied.Finally, the test is appliedto data on for- eign exchangefuture and spot pricesto test the stabilityof the forward-spot premium. 1. INTRODUCTION Economic models often imply that variables are cointegrated with simple and known cointegrating vectors. Examples include the neoclassical growth model, which implies that income, consumption, investment, and the capi- tal stock will grow in a balanced way, so that any stochastic growth in one of the series must be matched by corresponding growth in the others. Asset This paper has benefited from helpful comments by Neil Ericsson, Gordon Kemp, Andrew Levin, Soren Johansen, John McDermott, Pierre Perron, Peter Phillips, James Stock, and three referees.
    [Show full text]
  • Conditional Heteroscedastic Cointegration Analysis with Structural Breaks a Study on the Chinese Stock Markets
    Conditional Heteroscedastic Cointegration Analysis with Structural Breaks A study on the Chinese stock markets Authors: Supervisor: Andrea P.G. Kratz Frederik Lundtofte Heli M.K. Raulamo VT 2011 Abstract A large number of studies have shown that macroeconomic variables can explain co- movements in stock market returns in developed markets. The purpose of this paper is to investigate whether this relation also holds in China’s two stock markets. By doing a heteroscedastic cointegration analysis, the long run relation is investigated. The results show that it is difficult to determine if a cointegrating relationship exists. This could be caused by conditional heteroscedasticity and possible structural break(s) apparent in the sample. Keywords: cointegration, conditional heteroscedasticity, structural break, China, global financial crisis Table of contents 1. Introduction ............................................................................................................................ 3 2. The Chinese stock market ...................................................................................................... 5 3. Previous research .................................................................................................................... 7 3.1. Stock market and macroeconomic variables ................................................................ 7 3.2. The Chinese market ..................................................................................................... 9 4. Theory .................................................................................................................................
    [Show full text]
  • Lecture 18 Cointegration
    RS – EC2 - Lecture 18 Lecture 18 Cointegration 1 Spurious Regression • Suppose yt and xt are I(1). We regress yt against xt. What happens? • The usual t-tests on regression coefficients can show statistically significant coefficients, even if in reality it is not so. • This the spurious regression problem (Granger and Newbold (1974)). • In a Spurious Regression the errors would be correlated and the standard t-statistic will be wrongly calculated because the variance of the errors is not consistently estimated. Note: This problem can also appear with I(0) series –see, Granger, Hyung and Jeon (1998). 1 RS – EC2 - Lecture 18 Spurious Regression - Examples Examples: (1) Egyptian infant mortality rate (Y), 1971-1990, annual data, on Gross aggregate income of American farmers (I) and Total Honduran money supply (M) ŷ = 179.9 - .2952 I - .0439 M, R2 = .918, DW = .4752, F = 95.17 (16.63) (-2.32) (-4.26) Corr = .8858, -.9113, -.9445 (2). US Export Index (Y), 1960-1990, annual data, on Australian males’ life expectancy (X) ŷ = -2943. + 45.7974 X, R2 = .916, DW = .3599, F = 315.2 (-16.70) (17.76) Corr = .9570 (3) Total Crime Rates in the US (Y), 1971-1991, annual data, on Life expectancy of South Africa (X) ŷ = -24569 + 628.9 X, R2 = .811, DW = .5061, F = 81.72 (-6.03) (9.04) Corr = .9008 Spurious Regression - Statistical Implications • Suppose yt and xt are unrelated I(1) variables. We run the regression: y t x t t • True value of β=0. The above is a spurious regression and et ∼ I(1).
    [Show full text]
  • Testing Linear Restrictions on Cointegration Vectors: Sizes and Powers of Wald Tests in Finite Samples
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Haug, Alfred A. Working Paper Testing linear restrictions on cointegration vectors: Sizes and powers of Wald tests in finite samples Technical Report, No. 1999,04 Provided in Cooperation with: Collaborative Research Center 'Reduction of Complexity in Multivariate Data Structures' (SFB 475), University of Dortmund Suggested Citation: Haug, Alfred A. (1999) : Testing linear restrictions on cointegration vectors: Sizes and powers of Wald tests in finite samples, Technical Report, No. 1999,04, Universität Dortmund, Sonderforschungsbereich 475 - Komplexitätsreduktion in Multivariaten Datenstrukturen, Dortmund This Version is available at: http://hdl.handle.net/10419/77134 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung
    [Show full text]
  • SUPPLEMENTARY APPENDIX a Time Series Model of Interest Rates
    SUPPLEMENTARY APPENDIX A Time Series Model of Interest Rates With the Effective Lower Bound⇤ Benjamin K. Johannsen† Elmar Mertens Federal Reserve Board Bank for International Settlements April 16, 2018 Abstract This appendix contains supplementary results as well as further descriptions of computa- tional procedures for our paper. Section I, describes the MCMC sampler used in estimating our model. Section II describes the computation of predictive densities. Section III reports ad- ditional estimates of trends and stochastic volatilities well as posterior moments of parameter estimates from our baseline model. Section IV reports estimates from an alternative version of our model, where the CBO unemployment rate gap is used as business cycle measure in- stead of the CBO output gap. Section V reports trend estimates derived from different variable orderings in the gap VAR of our model. Section VI compares the forecasting performance of our model to the performance of the no-change forecast from the random-walk model over a period that begins in 1985 and ends in 2017:Q2. Sections VII and VIII describe the particle filtering methods used for the computation of marginal data densities as well as the impulse responses. ⇤The views in this paper do not necessarily represent the views of the Bank for International Settlements, the Federal Reserve Board, any other person in the Federal Reserve System or the Federal Open Market Committee. Any errors or omissions should be regarded as solely those of the authors. †For correspondence: Benjamin K. Johannsen, Board of Governors of the Federal Reserve System, Washington D.C. 20551. email [email protected].
    [Show full text]
  • Sequential Monte Carlo Methods for Inference and Prediction of Latent Time-Series
    SSStttooonnnyyy BBBrrrooooookkk UUUnnniiivvveeerrrsssiiitttyyy The official electronic file of this thesis or dissertation is maintained by the University Libraries on behalf of The Graduate School at Stony Brook University. ©©© AAAllllll RRRiiiggghhhtttsss RRReeessseeerrrvvveeeddd bbbyyy AAAuuuttthhhooorrr... Sequential Monte Carlo Methods for Inference and Prediction of Latent Time-series A Dissertation presented by Iñigo Urteaga to The Graduate School in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in Electrical Engineering Stony Brook University August 2016 Stony Brook University The Graduate School Iñigo Urteaga We, the dissertation committee for the above candidate for the Doctor of Philosophy degree, hereby recommend acceptance of this dissertation. Petar M. Djuri´c,Dissertation Advisor Professor, Department of Electrical and Computer Engineering Mónica F. Bugallo, Chairperson of Defense Associate Professor, Department of Electrical and Computer Engineering Yue Zhao Assistant Professor, Department of Electrical and Computer Engineering Svetlozar Rachev Professor, Department of Applied Math and Statistics This dissertation is accepted by the Graduate School. Nancy Goroff Interim Dean of the Graduate School ii Abstract of the Dissertation Sequential Monte Carlo Methods for Inference and Prediction of Latent Time-series by Iñigo Urteaga Doctor of Philosophy in Electrical Engineering Stony Brook University 2016 In the era of information-sensing mobile devices, the Internet- of-Things and Big Data, research on advanced methods for extracting information from data has become extremely critical. One important task in this area of work is the analysis of time-varying phenomena, observed sequentially in time. This endeavor is relevant in many applications, where the goal is to infer the dynamics of events of interest described by the data, as soon as new data-samples are acquired.
    [Show full text]
  • Identification of Spikes in Time Series Arxiv:1801.08061V1 [Stat.AP]
    Identification of Spikes in Time Series Dana E. Goin1 and Jennifer Ahern1 1 Division of Epidemiology, School of Public Health, University of California, Berkeley, California January 25, 2018 arXiv:1801.08061v1 [stat.AP] 24 Jan 2018 1 Abstract Identification of unexpectedly high values in a time series is useful for epi- demiologists, economists, and other social scientists interested in the effect of an exposure spike on an outcome variable. However, the best method to identify spikes in time series is not known. This paper aims to fill this gap by testing the performance of several spike detection methods in a sim- ulation setting. We created simulations parameterized by monthly violence rates in nine California cities that represented different series features, and randomly inserted spikes into the series. We then compared the ability to detect spikes of the following methods: ARIMA modeling, Kalman filtering and smoothing, wavelet modeling with soft thresholding, and an iterative outlier detection method. We varied the magnitude of spikes from 10-50% of the mean rate over the study period and varied the number of spikes inserted from 1 to 10. We assessed performance of each method using sensitivity and specificity. The Kalman filtering and smoothing procedure had the best over- all performance. We applied Kalman filtering and smoothing to the monthly violence rates in nine California cities and identified spikes in the rate over the 2005-2012 period. 2 List of Tables 1 ARIMA models and parameters by city . 21 2 Average sensitivity of spike identification methods for spikes of magnitudes ranging from 10-50% increase over series mean .
    [Show full text]
  • The Role of Models and Probabilities in the Monetary Policy Process
    1017-01 BPEA/Sims 12/30/02 14:48 Page 1 CHRISTOPHER A. SIMS Princeton University The Role of Models and Probabilities in the Monetary Policy Process This is a paper on the way data relate to decisionmaking in central banks. One component of the paper is based on a series of interviews with staff members and a few policy committee members of four central banks: the Swedish Riksbank, the European Central Bank (ECB), the Bank of England, and the U.S. Federal Reserve. These interviews focused on the policy process and sought to determine how forecasts were made, how uncertainty was characterized and handled, and what role formal economic models played in the process at each central bank. In each of the four central banks, “subjective” forecasting, based on data analysis by sectoral “experts,” plays an important role. At the Federal Reserve, a seventeen-year record of model-based forecasts can be com- pared with a longer record of subjective forecasts, and a second compo- nent of this paper is an analysis of these records. Two of the central banks—the Riksbank and the Bank of England— have explicit inflation-targeting policies that require them to set quantita- tive targets for inflation and to publish, several times a year, their forecasts of inflation. A third component of the paper discusses the effects of such a policy regime on the policy process and on the role of models within it. The large models in use in central banks today grew out of a first generation of large models that were thought to be founded on the statisti- cal theory of simultaneous-equations models.
    [Show full text]
  • Lecture: Introduction to Cointegration Applied Econometrics
    Lecture: Introduction to Cointegration Applied Econometrics Jozef Barunik IES, FSV, UK Summer Semester 2010/2011 Jozef Barunik (IES, FSV, UK) Lecture: Introduction to Cointegration Summer Semester 2010/2011 1 / 18 Introduction Readings Readings 1 The Royal Swedish Academy of Sciences (2003): Time Series Econometrics: Cointegration and Autoregressive Conditional Heteroscedasticity, downloadable from: http://www-stat.wharton.upenn.edu/∼steele/HoldingPen/NobelPrizeInfo.pdf 2 Granger,C.W.J. (2003): Time Series, Cointegration and Applications, Nobel lecture, December 8, 2003 3 Harris Using Cointegration Analysis in Econometric Modelling, 1995 (Useful applied econometrics textbook focused solely on cointegration) 4 Almost all textbooks cover the introduction to cointegration Engle-Granger procedure (single equation procedure), Johansen multivariate framework (covered in the following lecture) Jozef Barunik (IES, FSV, UK) Lecture: Introduction to Cointegration Summer Semester 2010/2011 2 / 18 Introduction Outline Outline of the today's talk What is cointegration? Deriving Error-Correction Model (ECM) Engle-Granger procedure Jozef Barunik (IES, FSV, UK) Lecture: Introduction to Cointegration Summer Semester 2010/2011 3 / 18 Introduction Outline Outline of the today's talk What is cointegration? Deriving Error-Correction Model (ECM) Engle-Granger procedure Jozef Barunik (IES, FSV, UK) Lecture: Introduction to Cointegration Summer Semester 2010/2011 3 / 18 Introduction Outline Outline of the today's talk What is cointegration? Deriving Error-Correction Model (ECM) Engle-Granger procedure Jozef Barunik (IES, FSV, UK) Lecture: Introduction to Cointegration Summer Semester 2010/2011 3 / 18 Introduction Outline Robert F. Engle and Clive W.J. Granger Robert F. Engle shared the Nobel prize (2003) \for methods of analyzing economic time series with time-varying volatility (ARCH) with Clive W.
    [Show full text]
  • On the Power and Size Properties of Cointegration Tests in the Light of High-Frequency Stylized Facts
    Journal of Risk and Financial Management Article On the Power and Size Properties of Cointegration Tests in the Light of High-Frequency Stylized Facts Christopher Krauss *,† and Klaus Herrmann † University of Erlangen-Nürnberg, Lange Gasse 20, 90403 Nürnberg, Germany; [email protected] * Correspondence: [email protected]; Tel.: +49-0911-5302-278 † The views expressed here are those of the authors and not necessarily those of affiliated institutions. Academic Editor: Teodosio Perez-Amaral Received: 10 October 2016; Accepted: 31 January 2017; Published: 7 February 2017 Abstract: This paper establishes a selection of stylized facts for high-frequency cointegrated processes, based on one-minute-binned transaction data. A methodology is introduced to simulate cointegrated stock pairs, following none, some or all of these stylized facts. AR(1)-GARCH(1,1) and MR(3)-STAR(1)-GARCH(1,1) processes contaminated with reversible and non-reversible jumps are used to model the cointegration relationship. In a Monte Carlo simulation, the power and size properties of ten cointegration tests are assessed. We find that in high-frequency settings typical for stock price data, power is still acceptable, with the exception of strong or very frequent non-reversible jumps. Phillips–Perron and PGFF tests perform best. Keywords: cointegration testing; high-frequency; stylized facts; conditional heteroskedasticity; smooth transition autoregressive models 1. Introduction The concept of cointegration has been empirically applied to a wide range of financial and macroeconomic data. In recent years, interest has surged in identifying cointegration relationships in high-frequency financial market data (see, among others, [1–5]). However, it remains unclear whether standard cointegration tests are truly robust against the specifics of high-frequency settings.
    [Show full text]
  • Making Wald Tests Work for Cointegrated VAR Systems *
    ECONOMETRIC REVIEWS, 15(4), 369-386 (1996) Making Wald Tests Work for Cointegrated VAR Systems * Juan J. Dolado and Helmut Lütkepohl CEMF1 Humboldt-Universitiit Casado del Alisal, 5 Spandauer Strasse 1 28014 Madrid, Spain 10178 Berlin, Germany Abstract Wald tests of restrictions on the coefficients of vector autoregressive (VAR) processes are known to have nonstandard asymptotic properties for 1(1) and cointegrated sys- tems of variables. A simple device is proposed which guarantees that Wald tests have asymptotic X2-distributions under general conditions. If the true generation process is a VAR(p) it is proposed to fit a VAR(p+1) to the data and perform a Wald test on the coefficients of the first p lags only. The power properties of the modified tests are studied both analytically and numerically by means of simple illustrative examples. 1 Introduction Wald tests are standard tools for testing restrictions on the coefficients of vector au- toregressive (VAR) processes. Their conceptual simplicity and easy applicability make them attractive for applied work to carry out statistical inference on hypotheses of interest. For instance, a typical example is the test of Granger-causality in the VAR *The research for this paper was supported by funds frorn the Deutsche Forschungsgerneinschaft, Sonderforschungsbereich 373. We are indebted to R. Mestre for invaluable research assistance. We have benefited frorn cornrnents by Peter C.B. Phillips, Enrique Sentana, Jürgen Wolters, three anony- rnous referees and participants at the European Meeting of the Econornetric Society, Maastricht 1994. 369 1 Copyright © 1996 by Mareel Dekker, Ine. 370 DOLADO AND LÜTKEPOHL framework where the null hypothesis is formulated as zero restrictions on the coeffi- cients of the lags of a subset of the variables.
    [Show full text]