ESAIM: PROCEEDINGS AND SURVEYS, October 2015, Vol. 51, p. 320-336 A. Garivier et al, Editors
AN INTRODUCTION TO ECONOPHYSICS AND QUANTITATIVE FINANCE
Remy´ Chicheportiche1, Thibault Jaisson2, Iacopo Mastromatteo3 and Vincent Vargas4
Abstract. This paper gives an account of the talks given by the authors at the 2014 MAS conference in Toulouse. These talks present recent research in the field of econophysics and quantitative finance.
R´esum´e. Cet article rend compte des expos´esdonn´espar les auteurs durant les journ´eesMAS 2014 `aToulouse. Ces expos´espr´esentent des recherches r´ecentes dans le domaine de l’´econophysique et de la finance quantitative.
Key words or phrases: Econophysics, multifractals, Hawkes processes, market impact. AMS 2000 subject classifications: 28A80, 60G15, 60G25, 60G57, 60F05, 60F17, 60G55, 62P05 .
1. Introduction In this paper, we will try to present recent trends of research related to the field of econophysics and quantitative finance. Therefore, it could be useful to first explain a bit how the field of econophysics differs from classical mathematical finance. Modern mathematical finance is really born with the founding paper by Black-Scholes on option pricing [10]. The purpose of the pioneering work [10] is to price standard vanilla options, i.e. a contract which gives its holder the right to buy a stock at a future time T at some fixed price K. The main assumption in [10] is to suppose that the stock price process (St)t≥0 is a geometric Brownian motion (GBM) with drift
σ2 (µ− )t+σBt St = e 2 where µ is a drift (supposed to be constant) and σ the average volatility. The main novelty of this work and Merton’s work [39] is to combine portfolio theory with option pricing; more precisely, the seller of an option can cover his risk by constituting a dynamic portfolio where he invests his money on a bank account but also on the stock St. Though the idea of using a dynamic portfolio in the theory of option pricing is very powerful and general, a few amazing features arise in the specific context of the GBM: • There is no arbitrage, i.e. the possibility to make riskless money starting with nothing; • The option price does not depend on the drift µ; • With an appropriate portfolio, one can perfectly hedge the risk of an option: the market is said to be complete.
1 Chaire de Finance Quantitative, Ecole´ Centrale Paris, 92290 Chˆatenay-Malabry. 2 Centre de Math´ematiquesAppliqu´ees,CNRS, Ecole´ Polytechnique, UMR 7641, 91128 Palaiseau. 3 Centre de Math´ematiquesAppliqu´ees,CNRS, Ecole´ Polytechnique, UMR 7641, 91128 Palaiseau. 4 ENS Ulm, DMA, 45 rue d’Ulm, 75005 Paris. c EDP Sciences, SMAI 2015
Article published online by EDP Sciences and available at http://www.esaim-proc.org or http://dx.doi.org/10.1051/proc/201551017 ESAIM: PROCEEDINGS AND SURVEYS 321
These features are very desirable for an option seller and make markets look perfectly safe. However, all market participants know that a GBM is not a good model for financial stocks. Indeed, a Brownian motion is Gaussian hence excludes large stock movements in a way which is incompatible with what is observed during a financial crisis for instance. Also, Brownian motion has independent increments whereas it is not hard to convince oneself that stocks have temporal correlations: for instance, it is well known that volatility on markets is highly correlated in time. In particular, the pool of all option prices observed on markets cannot be reproduced at some fixed instant using a GBM model. This led to the introduction of the local volatility model by Dupire in [21]. Though the model looks very counter-intuitive, it can fit any pool of option prices at some fixed instant and it shares exactly the same specific features as the GBM: no arbitrage, completeness, etc. . . Hence, Dupire’s work came as a relief for the community of mathematical finance! So, case closed? Well not really because, although the model fits any pool of option prices at some fixed initial time, the model and the pool of option prices observed on the market diverge dangerously under many realistic scenarios when time passes. This clearly indicates that one should work with realistic models of stock prices: this brings us precisely to the field of econophysics. In some sense, econophysics can be seen as the endless quest of understanding and modeling stock prices. The pillars of the field are the so-called stylized facts which are the universal statistical features observed on all financial assets. With respect to this problem, one should distinguish two time scales of observation corresponding to low-frequency and high-frequency finance. More precisely, there is some time threshold τc around 10 minutes such that prices on time lags below τc behave differently than on time lags above τc. Therefore, there exist low-frequency and high-frequency stylized facts. When one deals with high frequency finance, one typically studies quantities more general than the price process (which is in fact ill-defined due to the discretization effects at small time scales) like the order book dynamics: in that case, one is interested in the stylized facts of the order book, i.e. its universal statistical features. Now, typical basic questions in econophysics are • Can one construct a model which satisfies these stylized facts at all time scales? • Can one construct a (agent-based) model which explains the high-frequency stylized facts (order book dynamics, price formation process, etc...)? The next sections will tackle these questions. Each section corresponds to a talk given by one of the authors at the 2014 MAS conference. In particular, the first two sections will study two low-frequency models for the price process: the multifractal random walk (talk by V. Vargas) and the self-reflective dynamics of the stock return fluctuations (talk by R. Chicheportiche). The third section will then present the nearly unstable Hawkes model (talk by T. Jaisson), which reproduces properties of the order flow at both high and low frequencies. The fourth section will finally focus on the problem of market impact (talk by I. Mastromatteo), by illustrating the predictions of a latent liquidity model.
2. Forecasting volatility with the multifractal random walk (MRW) V. Vargas 2.1. Definition of the MRW and stylized facts The multifractal random walk (MRW) can be seen as a stochastic volatility model, i.e. the price process (St)t≥0 solves the following SDE
dSt/St = σtdWt (1) where (Wt)t≥0 is a Brownian motion and the volatility process is independent from W and has the following form λωT −λ2E[(ωT )2] σt = σe t t 322 ESAIM: PROCEEDINGS AND SURVEYS where ωT is a centered Gaussian “process” (independent of W ) with covariance
T E[ωT ωT ] = ln+ , (2) s t |t − s|
+ T with ln (x) = max(ln x, 0). The problem is that (ωt )t∈R lives in the space of distributions (in the sense of Schwartz) but not in the space of functions hence σt is ill-defined. To make sense of this expression, first note that if (σt)t∈R is a nice process, Brownian scaling gives the following expression
R 1 R 2 σsdWs− σ ds 1 2 s BM[0,t]− M[0,t] St = S0e [0,t] [0,t] = S0e 2 (3) where B is a Brownian motion and M is the random measure given by the following expression:
Z t 2 M[0, t] = σs ds, t ≥ 0. 0 Notice that (3) makes sense in great generality for any (random) measure M (whereas definition (1) makes sense when σt is a function, i.e. M is absolutely continuous with respect to the Lebesgue measure). Now, it happens that Kahane’s theory of Gaussian multiplicative chaos [30] enables to make sense of the following formal measure Z t T 2 2λωT −2λ2E[(ωT )2] M [0, t] = σ e s s ds, t ≥ 0, 0 T T T T where ω is the Gaussian process with covariance (2). M is defined by a limit procedure M = lim Mτ where τ→0
t Z τ,T 2 τ,T 2 T 2 2λωs −2λ E[(ωs ) ] Mτ [0, t] = σ e ds, t ≥ 0, 0 and ωτ,T is the smooth Gaussian process:
T E[ωτ,T ωτ,T ] = ln+ . s t |t − s| + τ
The MRW model is then (3) with M T as choice of M and was first introduced and studied by Bacry, Delour and Muzy in [5] (the idea of looking at a time changed Brownian motion as a model for a stock price goes back to the work of Mandelbrot and Taylor [35]). One of the interesting features of the MRW model is that it exhibits many of the low-frequency stylized facts observed on financial assets; in particular: • the volatility fluctuates and is approximately lognormal • the volatility is highly correlated and exhibits long range correlations. (Law) Note however that the (recentered) model is symmetric since (BM T [0,t])t = −(BM T [0,t])t and therefore it does not reproduce the celebrated leverage effect (see [15]). Roughly, the leverage effect is a quantitative measure of the fact that large drawdowns of a stock increases its volatility more than large upward movements. This effect is essentially observed on indexes and large market capitalization stocks. Therefore, the MRW model is essentially used to model currencies where there is usually little leverage effect. 2.2. Forecasting the log-volatility Since the MRW is a good model for financial assets, it is natural to try to establish prediction formulas for the volatility of the model, namely M T . In order to use these prediction formulas, one must also calibrate the model on financial data. The value of λ2 is rather stable in time on an asset and its value (which depends on ESAIM: PROCEEDINGS AND SURVEYS 323 the asset) is found to be roughly in the interval [0.01; 0.08]. However, the calibration of σ and T to the market is much more delicate. Indeed, we can make the following observation: Lemma 2.1 (Invariance by integral scale change). For all T < T 0:
0 (Law) 2 T 0 T 2λΩT 0/T −2λ ln T (M [0, t])t∈[0,T ] = e T (M [0, t])t∈[0,T ]
T 0 T where ΩT 0/T is a centered Gaussian variable of variance ln T that is independent of (M ([0, t]))t∈[0,T ]. As a consequence of the above lemma, if the observation window of M T is of length less than or equal to T , it is impossible to determine σ and T (ill-posed problem). Typically, when you try to measure the integral scale T on a financial asset, you find it comparable to the scale of your observation window, hence the measure is unreliable. However, one idea to overcome the problem is to let T → ∞ in order to get rid of the determination of σ and T at the same time! On the level of the log volatility, recall that ωT is a Gaussian measure on the space of tempered distributions S0(R) (in the sense of Schwartz) such that:
R T R T 2 i φ(t)ωt dt −1/2E(( φ(t)ωt dt) ) ∀φ ∈ S(R),E(e R ) = e R RR + −1/2 2 φ(t)φ(s) ln (T/|t−s|)dsdt = e R .
Now, we want to let T → ∞ in the above formula. In order to make sense of the limit, it is rather straightforward to see that φ must be such that R φ(t)dt = 0 in which case we get the following convergence R
RR + RR −1/2 2 φ(t)φ(s) ln (T/|t−s|)dsdt −1/2 2 φ(t)φ(s) ln(1/|t−s|)dsdt e R → e R . T →∞
One then naturally introduces ω, a Gaussian measure on the quotient space S0(R)/R. Now, if one considers R S0( ) = {φ ∈ S( ); φ(t)dt = 0} then: R R R
R R 2 i φ(t)ωtdt −1/2E(( φ(t)ωtdt) ) ∀φ ∈ S0(R),E(e R ) = e R RR −1/2 2 φ(t)φ(s) ln(1/|t−s|)dsdt = e R
R |φˆ(ξ)|2 −1/4 |ξ| dξ = e R .
Therefore, ω can be seen as a rigorous definition of the celebrated 1/f noise studied in physics. We consider the reproducing kernel Hilbert space of ω: Z 1/2 0 2 H (R) = {f ∈ S (R)/R ; |ξ||fˆ(ξ)| dξ < ∞}. R Now, one can prove the following prediction formula: Theorem 2.1 ([36]). For all f ∈ H1/2(R), √ 1 Z 0 t ∀t > 0 E[ωt|(ωs)s<0 = f] = √ f(s)ds. π −∞ (t − s) −s
In fact, one can use the above theorem to get exact prediction formulas for ωT . In order to state the result, we introduce the standard Sobolev space: Z 1/2 0 2 1/2 2 H (R) = {f ∈ S (R); (1 + |ξ| ) |fˆ(ξ)| dξ < ∞}. R 324 ESAIM: PROCEEDINGS AND SURVEYS
Let L be some observation window and T the integral scale. There is an explicit kernel KL,T (t, s) (see [36] for the exact expression) such that:
Corollary 2.1 ([36]). For all f ∈ H1/2(R),
Z 0 T T ∀t ∈]0,T − 2L[,E[ωt |(ωs )−2L
Finally, one can discretize the above formulas to get approximate formulas for the discrete model. In that case, one can also transfer the prediction formulas to the volatility itself (in which case, the prediction formulas will also depend on λ2): see [36] for more details.
3. Modeling stock returns fluctuations R. Chicheportiche One of the most striking universal stylized facts of financial returns is the volatility clustering effect, which was first reported by Mandelbrot as early as 1963 [34]. He noted that large changes in the asset price St tend to be followed by large changes, of either sign, and small changes tend to be followed by small changes. The first quantitative description of this effect was the ARCH model proposed by Engle in 1982 [22]. It formalizes Mandelbrot’s hunch in the simplest possible way, by postulating that log-returns rt ≡ ln St − ln St−1 can be modeled as conditionally Gaussian random variables rt = σtξt, where the residuals ξ’s are IID random variables, of zero mean and variance equal to unity (while many papers take these ξ’s to be Gaussian, it is preferable to be agnostic about their univariate distribution. In fact a consensus emerges on that the ξ’s themselves have fat-tails: asset returns are not conditionally Gaussian and “true jumps” do occur). The volatility σt is time dependent and evolves according to: 2 2 2 σt = s + grt−1. (4) In words, this equation means that the (squared) volatility today is equal to a baseline level s2, plus a self- 2 exciting term that describes the influence of yesterday’s perceived volatility rt−1 on today’s activity, through a feedback parameter g. It soon became apparent that the above model is far too simple to account for empirical data. For one thing, it is unable to account for the long memory nature of volatility fluctuations. It is also arbitrary in at least two ways: 2 • First, there is no reason to limit the feedback effect to the previous day only. One can replace rt−1 by a weighted moving average of past squared returns, leading to a large family of models such as:
∞ 2 2 X 2 σt = s + k(τ)rt−τ . (5) τ=1
A slowly (power-law) decaying kernel k(τ) is able to account for the long memory of volatility. • Second, there is no reason to single out the day as the only time scale to define the returns: in principle, returns over different time scales could also feedback on the volatility today [11,12,33,41]. For extended time scales longer than the day, this leads to another natural extension of the model as:
∞ 2 2 X X (`)2 σt = s + g`(τ)Rt−τ , (6) ` τ=1
(`) P` where Rt = τ=1 rt−τ is the cumulative, ` day return between t − ` and t. ESAIM: PROCEEDINGS AND SURVEYS 325
The common point to the zoo of generalizations of the initial ARCH model is that the current level of 2 volatility σt is expressed as a quadratic form of past realized returns. The most general model of this kind, called QARCH (for Quadratic ARCH), is due to Sentana [43], and reads:
∞ ∞ 2 2 X X 0 σt = s + L(τ) rt−τ + K(τ, τ ) rt−τ rt−τ 0 . (7) τ=1 τ,τ 0=1
2 The QARCH can be seen as a general discrete-time model for the dependence of σt on all past returns {rt0 }t0 Tr K < 1. (8) In this case, the volatility is a stationary process such that hσ2i ≡ E[σ2] = s2/(1 − Tr K): the feedback-induced increase of the volatility only involves the diagonal elements of K. Note also that the leverage kernel L(τ) does not appear in this equation. Empirical study and dataset We have calibrated the above QARCH model using an appropriate methodology presented in detail in [17]. In particular, we have estimated the matrix elements K(τ, τ 0) both under Maximum Likelihood (ML) and Method of Moments (GMM). We have also implemented a constrained model imposing a simple Long-Term structure on K (ML+LT). We worked with daily stock prices (Open, Close, High, Low) for N = 280 names present in the S&P-500 index from Jan. 2000 to Dec. 2009 (T = 2515 days), without interruption. The reference price for day t is defined to be the close of that day Ct. The true volatility is of course unobservable; we have replaced it by the Rogers-Satchell estimator [42]. 3.1. Large-dimensional effects The ubiquitous empirical long-memory correlations of the volatility,