Markov Determinantal Point Processes

Total Page:16

File Type:pdf, Size:1020Kb

Markov Determinantal Point Processes Markov Determinantal Point Processes Raja Hafiz Affandi Alex Kulesza Emily B. Fox Department of Statistics Dept. of Computer and Information Science Department of Statistics The Wharton School University of Pennsylvania The Wharton School University of Pennsylvania [email protected] University of Pennsylvania [email protected] [email protected] Abstract locations are likewise over-dispersed relative to uniform placement (Bernstein and Gobbel, 1979). Likewise, many A determinantal point process (DPP) is a random practical tasks can be posed in terms of diverse subset selec- process useful for modeling the combinatorial tion. For example, one might want to select a set of frames problem of subset selection. In particular, DPPs from a movie that are representative of its content. Clearly, encourage a random subset Y to contain a diverse diversity is preferable to avoid redundancy; likewise, each set of items selected from a base set Y. For ex- frame should be of high quality. A motivating example we ample, we might use a DPP to display a set of consider throughout the paper is the task of selecting a di- news headlines that are relevant to a user’s inter- verse yet relevant set of news headlines to display to a user. ests while covering a variety of topics. Suppose, One could imagine employing binary Markov random fields however, that we are asked to sequentially se- with negative correlations, but such models often involve lect multiple diverse sets of items, for example, notoriously intractable inference problems. displaying new headlines day-by-day. We might want these sets to be diverse not just individually Determinantal point processes (DPPs), which arise in ran- but also through time, offering headlines today dom matrix theory (Mehta and Gaudin, 1960; Ginibre, 1965) repul- that are unlike the ones shown yesterday. In this and quantum physics (Macchi, 1975), are a class of sive processes paper, we construct a Markov DPP (M-DPP) that and are natural models for subset selection problems where diversity is preferred. DPPs define the models a sequence of random sets fY tg. The proposed M-DPP defines a stationary process that probability of a subset in terms of the determinant of a maintains DPP margins. Crucially, the induced kernel submatrix, and with an appropriate definition of the kernel matrix they can be interpreted as inherently balanc- union process Zt ≡ Y t [Y t−1 is also marginally DPP-distributed. Jointly, these properties imply ing quality and diversity. DPPs are appealing in practice that the sequence of random sets are encouraged since they offer interpretability and tractable algorithms for to be diverse both at a given time step as well as exact inference. For example, one can compute marginal across time steps. We describe an exact, efficient and conditional probabilities and perform exact sampling. sampling procedure, and a method for incremen- DPPs have recently been employed for human pose estima- tally learning a quality measure over items in the tion, search diversification, and document summarization base set Y based on external preferences. We (Kulesza and Taskar, 2010, 2011a,b). apply the M-DPP to the task of sequentially dis- In this paper, our focus is instead on modeling diverse se- playing diverse and relevant news articles to a quences of subsets. For example, in displaying news head- user with topic preferences. lines from day to day, one aims to select articles that are relevant and diverse on any given day. Additionally, it is desirable to select articles that are diverse relative to those 1 INTRODUCTION previously shown. We construct a Markov DPP (M-DPP) for a sequence of random sets fY g. The proposed M-DPP Consider the combinatorial problem of subset selection. Bi- t defines a stationary process that maintains DPP margins, nary Markov random fields are commonly applied in this implying that Y is encouraged to be diverse at time t. Cru- setting, and in the case of positive correlations, yield subsets t cially, the induced union process Z ≡ Y [ Y is also that favor similar items. However, in many applications t t t−1 marginally DPP-distributed. Since this property implies the there is naturally a sense of repulsion. For example, re- diversity of Z , in addition to the individual diversity of pulsive processes arise in nature—trees tend to grow in t Y and Y , we conclude that Y is diverse from Y . the least occupied space (Neeff et al., 2005), and ant hill t t−1 t t−1 DPP M-DPP random set drawn according to P, then for every A ⊆ Y: P(Y ⊇ A) = det(KA) : (1) Here, KA ≡ [KA]i;j2A denotes the submatrix of K in- dexed by elements in A, and we adopt the convention that det(K;) = 1. We will refer to K as the marginal kernel. If we think of Kij as measuring the similarity between items i and j, then Time Time P(Y ⊇ fi; jg) = K K − K2 (2) Figure 1: A set of points on a line (y axis) drawn from ii jj ij a DPP independently over time (left) and from a M-DPP implies that Y is unlikely to contain both i and j when they (right). While DPP points are diverse only within time steps are very similar; that is, a DPP can be seen as modeling a (columns), M-DPP points are also diverse across time steps. collection of diverse items from the base set Y. For an illustration of the improved overall diversity when DPPs can alternatively be constructed via L-ensembles sampling from a M-DPP rather than independent sequential (Borodin and Rains, 2005). An L-ensemble is a probability Y sampling from a DPP, see Fig. 1. measure on 2 defined via a positive semidefinite matrix L indexed by elements of Y: Our specific construction of the M-DPP yields an exact sampling procedure that can be performed in polynomial det(LA) PL(Y = A) = ; (3) time. Additionally, we explore a method for incrementally det(L + I) learning the quality of each item in the base set Y based where I is the N ×N identity matrix. It can be shown that an on externally provided preferences. In particular, a decom- L-ensemble is a DPP with marginal kernel K = L(I+L)−1. position of the DPP kernel matrix has an interpretation as Conversely, a DPP with marginal kernel K has L-ensemble defining the quality of each item and pairwise similarities kernel L = K(I − K)−1 (when the inverse exists). between items. Our incremental learning procedure assumes a well-defined similarity metric and aims to learn features An intuitive way to think of the L-ensemble kernel L is as a of items that a user deems as preferable. These features are Gram matrix (Kulesza and Taskar, 2010): used to define the quality scores for each item. The M-DPP > Lij = qiφi φjqj ; (4) aids in the exploration of items of interest to the user by + providing sequentially diverse results. interpreting qi 2 R as representing the intrinsic quality n of an item i, and φi; φj 2 R as unit length feature vec- We study the empirical behavior of the M-DPP on a news tors representing the similarity between items i and j with task where the goal is to display diverse and high quality > φi φj 2 [−1; 1]. Under this framework, we can model qual- articles. Compared to choosing articles based on quality ity and similarity separately to encourage the DPP to choose alone or to sampling from an independent DPP at each time high quality items that are dissimilar to each other. This is step, we show that the M-DPP produces articles that are very useful in many applications. For example, in response significantly more diverse across time steps without large to a search query we can provide a very relevant (i.e. high sacrifices in quality. Furthermore, within a time step the quality) but diverse (i.e. dissimilar) list of results. M-DPP chooses articles with diversity comparable to that of the independent DPP; this is a direct result of the fact Conditional DPPs For any A; B ⊆ Y with A \ B = ;, that the M-DPP maintains DPP margins. We also consider it is straightforward to show that learning the quality function over time from the feedback of a user with topic preferences. In this setting, the M-DPP det(LA[B) PL(Y = A [ BjY ⊇ A) = ; (5) returns high quality results that are preferred by the user det(L + IYnA) while simultaneously exploring the topic space more quickly where I is a matrix with ones on the diagonal entries than baseline methods, leading to improved coverage. YnA indexed by the elements of Y n A and zeros elsewhere. 2 DETERMINANTAL POINT This conditional distribution is itself a DPP over the ele- PROCESSES ments of Y n A (Borodin and Rains, 2005). In particular, suppose Y is DPP-distributed with L-ensemble kernel L, A random point process P on a discrete base set Y = and condition on the fact that Y ⊇ A. Then the set Y n A f1;:::;Ng is a probability measure on the set 2Y of all is DPP-distributed with marginal and L-ensemble kernels subsets of Y. Let K be a semidefinite matrix with rows KA = I − (L + I )−1 (6) and columns indexed by the elements of Y. P is called a YnA YnA determinantal point process (DPP) if there exists K I −1 LA = (L + I )−1 − I: (7) (all eigenvalues less than or equal to 1) such that if Y is a YnA YnA Here, [·]YnA denotes the submatrix of the argument indexed immediately the joint probability by elements in Y n A.
Recommended publications
  • POISSON PROCESSES 1.1. the Rutherford-Chadwick-Ellis
    POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS 1.1. The Rutherford-Chadwick-Ellis Experiment. About 90 years ago Ernest Rutherford and his collaborators at the Cavendish Laboratory in Cambridge conducted a series of pathbreaking experiments on radioactive decay. In one of these, a radioactive substance was observed in N = 2608 time intervals of 7.5 seconds each, and the number of decay particles reaching a counter during each period was recorded. The table below shows the number Nk of these time periods in which exactly k decays were observed for k = 0,1,2,...,9. Also shown is N pk where k pk = (3.87) exp 3.87 =k! {− g The parameter value 3.87 was chosen because it is the mean number of decays/period for Rutherford’s data. k Nk N pk k Nk N pk 0 57 54.4 6 273 253.8 1 203 210.5 7 139 140.3 2 383 407.4 8 45 67.9 3 525 525.5 9 27 29.2 4 532 508.4 10 16 17.1 5 408 393.5 ≥ This is typical of what happens in many situations where counts of occurences of some sort are recorded: the Poisson distribution often provides an accurate – sometimes remarkably ac- curate – fit. Why? 1.2. Poisson Approximation to the Binomial Distribution. The ubiquity of the Poisson distri- bution in nature stems in large part from its connection to the Binomial and Hypergeometric distributions. The Binomial-(N ,p) distribution is the distribution of the number of successes in N independent Bernoulli trials, each with success probability p.
    [Show full text]
  • Rescaling Marked Point Processes
    1 Rescaling Marked Point Processes David Vere-Jones1, Victoria University of Wellington Frederic Paik Schoenberg2, University of California, Los Angeles Abstract In 1971, Meyer showed how one could use the compensator to rescale a multivari- ate point process, forming independent Poisson processes with intensity one. Meyer’s result has been generalized to multi-dimensional point processes. Here, we explore gen- eralization of Meyer’s theorem to the case of marked point processes, where the mark space may be quite general. Assuming simplicity and the existence of a conditional intensity, we show that a marked point process can be transformed into a compound Poisson process with unit total rate and a fixed mark distribution. ** Key words: residual analysis, random time change, Meyer’s theorem, model evaluation, intensity, compensator, Poisson process. ** 1 Department of Mathematics, Victoria University of Wellington, P.O. Box 196, Welling- ton, New Zealand. [email protected] 2 Department of Statistics, 8142 Math-Science Building, University of California, Los Angeles, 90095-1554. [email protected] Vere-Jones and Schoenberg. Rescaling Marked Point Processes 2 1 Introduction. Before other matters, both authors would like express their appreciation to Daryl for his stimulating and forgiving company, and to wish him a long and fruitful continuation of his mathematical, musical, woodworking, and many other activities. In particular, it is a real pleasure for the first author to acknowledge his gratitude to Daryl for his hard work, good humour, generosity and continuing friendship throughout the development of (innumerable draft versions and now even two editions of) their joint text on point processes.
    [Show full text]
  • Lecture 26 : Poisson Point Processes
    Lecture 26 : Poisson Point Processes STAT205 Lecturer: Jim Pitman Scribe: Ben Hough <[email protected]> In this lecture, we consider a measure space (S; S; µ), where µ is a σ-finite measure (i.e. S may be written as a disjoint union of sets of finite µ-measure). 26.1 The Poisson Point Process Definition 26.1 A Poisson Point Process (P.P.P.) with intensity µ is a collection of random variables N(A; !), A 2 S, ! 2 Ω defined on a probability space (Ω; F; P) such that: 1. N(·; !) is a counting measure on (S; S) for each ! 2 Ω. 2. N(A; ·) is Poisson with mean µ(A): − P e µ(A)(µ(A))k (N(A) = k) = k! all A 2 S. 3. If A1, A2, ... are disjoint sets then N(A1; ·), N(A2; ·), ... are independent random variables. Theorem 26.2 P.P.P.'s exist. Proof Sketch: It's not enough to quote Kolmogorov's Extension Theorem. The only convincing argument is to give an explicit constuction from sequences of independent random variables. We begin by considering the case µ(S) < 1. P µ(A) 1. Take X1, X2, ... to be i.i.d. random variables so that (Xi 2 A) = µ(S) . 2. Take N(S) to be a Poisson random variable with mean µ(S), independent of the Xi's. Assume all random varibles are defined on the same probability space (Ω; F; P). N(S) 3. Define N(A) = i=1 1(Xi2A), for all A 2 S. P 1 Now verify that this N(A; !) is a P.P.P.
    [Show full text]
  • A Course in Interacting Particle Systems
    A Course in Interacting Particle Systems J.M. Swart January 14, 2020 arXiv:1703.10007v2 [math.PR] 13 Jan 2020 2 Contents 1 Introduction 7 1.1 General set-up . .7 1.2 The voter model . .9 1.3 The contact process . 11 1.4 Ising and Potts models . 14 1.5 Phase transitions . 17 1.6 Variations on the voter model . 20 1.7 Further models . 22 2 Continuous-time Markov chains 27 2.1 Poisson point sets . 27 2.2 Transition probabilities and generators . 30 2.3 Poisson construction of Markov processes . 31 2.4 Examples of Poisson representations . 33 3 The mean-field limit 35 3.1 Processes on the complete graph . 35 3.2 The mean-field limit of the Ising model . 36 3.3 Analysis of the mean-field model . 38 3.4 Functions of Markov processes . 42 3.5 The mean-field contact process . 47 3.6 The mean-field voter model . 49 3.7 Exercises . 51 4 Construction and ergodicity 53 4.1 Introduction . 53 4.2 Feller processes . 54 4.3 Poisson construction . 63 4.4 Generator construction . 72 4.5 Ergodicity . 79 4.6 Application to the Ising model . 81 4.7 Further results . 85 5 Monotonicity 89 5.1 The stochastic order . 89 5.2 The upper and lower invariant laws . 94 5.3 The contact process . 97 5.4 Other examples . 100 3 4 CONTENTS 5.5 Exercises . 101 6 Duality 105 6.1 Introduction . 105 6.2 Additive systems duality . 106 6.3 Cancellative systems duality . 113 6.4 Other dualities .
    [Show full text]
  • Point Processes, Temporal
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by CiteSeerX Point processes, temporal David R. Brillinger, Peter M. Guttorp & Frederic Paik Schoenberg Volume 3, pp 1577–1581 in Encyclopedia of Environmetrics (ISBN 0471 899976) Edited by Abdel H. El-Shaarawi and Walter W. Piegorsch John Wiley & Sons, Ltd, Chichester, 2002 is an encyclopedic account of the theory of the Point processes, temporal subject. A temporal point process is a random process whose Examples realizations consist of the times fjg, j 2 , j D 0, š1, š2,... of isolated events scattered in time. A Examples of point processes abound in the point process is also known as a counting process or environment; we have already mentioned times a random scatter. The times may correspond to events of floods. There are also times of earthquakes, of several types. fires, deaths, accidents, hurricanes, storms (hale, Figure 1 presents an example of temporal point ice, thunder), volcanic eruptions, lightning strikes, process data. The figure actually provides three dif- tornadoes, power outages, chemical spills (see ferent ways of representing the timing of floods Meteorological extremes; Natural disasters). on the Amazon River near Manaus, Brazil, dur- ing the period 1892–1992 (see Hydrological ex- tremes)[7]. Questions The formal use of the concept of point process has a long history going back at least to the life tables The questions that scientists ask involving point of Graunt [14]. Physicists contributed many ideas in process data include the following. Is a point pro- the first half of the twentieth century; see, for ex- cess associated with another process? Is the associ- ample, [23].
    [Show full text]
  • Martingale Representation in the Enlargement of the Filtration Generated by a Point Process Paolo Di Tella, Monique Jeanblanc
    Martingale Representation in the Enlargement of the Filtration Generated by a Point Process Paolo Di Tella, Monique Jeanblanc To cite this version: Paolo Di Tella, Monique Jeanblanc. Martingale Representation in the Enlargement of the Filtration Generated by a Point Process. 2019. hal-02146840 HAL Id: hal-02146840 https://hal.archives-ouvertes.fr/hal-02146840 Preprint submitted on 4 Jun 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Martingale Representation in the Enlargement of the Filtration Generated by a Point Process Paolo Di Tella1;2 and Monique Jeanblanc3 Abstract Let X be a point process and let X denote the filtration generated by X. In this paper we study martingale representation theorems in the filtration G obtained as an initial and progressive enlargement of the filtration X. In particular, the progressive enlargement is done by means of a whole point process H. We work here in full generality, without requiring any further assumption on the process H and we recover the special case in which X is enlarged progressively by a random time t. Keywords: Point processes, martingale representation, progressive enlargement, initial enlargement, random measures.
    [Show full text]
  • Notes on the Poisson Point Process
    Notes on the Poisson point process Paul Keeler January 5, 2016 This work is licensed under a “CC BY-SA 3.0” license. Abstract The Poisson point process is a type of random object known as a point pro- cess that has been the focus of much study and application. This survey aims to give an accessible but detailed account of the Poisson point process by covering its history, mathematical definitions in a number of settings, and key properties as well detailing various terminology and applications of the process, with rec- ommendations for further reading. 1 Introduction In probability, statistics and related fields, a Poisson point process or a Poisson process or a Poisson point field is a type of random object known as a point pro- cess or point field that consists of randomly positioned points located on some underlying mathematical space [68]. The process has convenient mathematical properties [39], which has led to it being frequently defined in Euclidean space and used as a mathematical model for seemingly random processes in numer- ous disciplines such as astronomy [5], biology [53], ecology [70], geology [16], physics [61], image processing [12], and telecommunications [7][28]. The Poisson point process is often defined on the real line playing an impor- tant role in the field of queueing theory [40] where it is used to model certain random events happening in time such as the arrival of customers at a store or phone calls at an exchange. In the plane, the point process, also known as a spa- tial Poisson process [9], may represent scattered objects such as users in a wireless network [2, 6, 7, 27], particles colliding into a detector, or trees in a forest [68].
    [Show full text]
  • Stochastic Differential Equations with Jumps
    Probability Surveys Vol. 1 (2004) 1–19 ISSN: 1549-5787 DOI: 10.1214/154957804100000015 Stochastic differential equations with jumps∗,† Richard F. Bass Department of Mathematics, University of Connecticut Storrs, CT 06269-3009 e-mail: [email protected] Abstract: This paper is a survey of uniqueness results for stochastic dif- ferential equations with jumps and regularity results for the corresponding harmonic functions. Keywords and phrases: stochastic differential equations, jumps, mar- tingale problems, pathwise uniqueness, Harnack inequality, harmonic func- tions, Dirichlet forms. AMS 2000 subject classifications: Primary 60H10; secondary 60H30, 60J75. Received September 2003. 1. Introduction Researchers have increasingly been studying models from economics and from the natural sciences where the underlying randomness contains jumps. To give an example from financial mathematics, the classical model for a stock price is that of a geometric Brownian motion. However, wars, decisions of the Federal Reserve and other central banks, and other news can cause the stock price to make a sudden shift. To model this, one would like to represent the stock price by a process that has jumps. This paper is a survey of some aspects of stochastic differential equations (SDEs) with jumps. As will quickly become apparent, the theory of SDEs with jumps is nowhere near as well developed as the theory of continuous SDEs, and in fact, is still in a relatively primitive stage. In my opinion, the field is both fertile and important. To encourage readers to undertake research in this area, I have mentioned some open problems. arXiv:math/0309246v2 [math.PR] 9 Nov 2005 Section 2 is a description of stochastic integration when there are jumps.
    [Show full text]
  • Stochastic Point Processes
    Stochastic Point Processes Cengiz Öztireli, 2015 Distributions in Nature [Filckr user Hans Dekker] Sampling • Conversion from continuous to discrete • Integration in rendering [Filckr user Josh Pesavento ] Point Distributions • All kinds of patterns Studying Point Distributions • How can we analyze patterns? Periodogram Density [Heck et al. 2013] [Faal 2011] Point Processes • Formal characterization of point patterns Point Processes • Definition Point Process Point Processes • Infinite point processes Observation window Point Process Statistics • Correlations as probabilities %(n)(x , , x )dV dV = p(x , , x N) 1 ··· n 1 ··· n 1 ··· n| Product density Small volumes Points in space Point process xi Point Process Statistics • Correlations as probabilities – First order product density %(1)(x)=λ(x) Expected number of points around x x Measures local density Point Process Statistics • Correlations as probabilities – First order product density %(1)(x)=λ(x) Point Process Statistics • Correlations as probabilities – First order product density %(1)(x)=λ(x) Constant Point Process Statistics • Correlations as probabilities – Second order product density %(2)(x, y)=%(x, y) Expected number of points around x & y y x Measures the joint probability p(x, y) Point Process Statistics • Correlations as probabilities – Higher order? %(3)(x, y, z) z Expected number of points around x y z y The “second order dogma” in physics x Point Process Statistics • Summary: – First and second order product densities %(1)(x)=λ(x) %(2)(x, y)=%(x, y) y x x Point Process Statistics • Example: homogenous Poisson process – a.k.a random sampling p(x)=p p(x, y)=p(x)p(y) 2 λ(x)dV = p p(x, y)=%(x, y)dVxdVy = p(x)p(y)=λ(x)dVxλ(y)dVy = λ 2 λ(xp)=(x, yλ)=%(x, y)dVxdVy = p(x)p(y)=λ(x)dVxλ(y)dVy = λ 2 p(x, y)=%(x, y)dVxdVy = p(x)p(y)=λ(x)dVxλ(y)dVy = λ %(x, y)=λ(x)λ(y)=λ2 Stationary Processes Stationary Isotropic (translation invariant) (translation & rotation invariant) [Zhou et al.
    [Show full text]
  • Point Processes on Directed Linear Network Arxiv:1812.09071V2 [Math
    Point processes on directed linear network Jakob G. Rasmussen and Heidi S. Christensen Department of Mathematical Sciences, Aalborg University Abstract In this paper we consider point processes specified on directed linear net- works, i.e. linear networks with associated directions. We adapt the so-called conditional intensity function used for specifying point processes on the time line to the setting of directed linear networks. For models specified by such a conditional intensity function, we derive an explicit expression for the like- lihood function, specify two simulation algorithms (the inverse method and Ogata’s modified thinning algorithm), and consider methods for model check- ing through the use of residuals. We also extend the results and methods to the case of a marked point process on a directed linear network. Further- more, we consider specific classes of point process models on directed linear networks (Poisson processes, Hawkes processes, non-linear Hawkes processes, self-correcting processes, and marked Hawkes processes), all adapted from well- known models in the temporal setting. Finally, we apply the results and meth- ods to analyse simulated and neurological data. Keywords: Conditional intensity, dendrite network, Hawkes process, self-correcting process 1 Introduction Point processes on linear networks are important for modelling events or objects on a real network, such as a road or river network. In recent years there have been a fair amount of papers on functional summary statistics and models for point processes specified on linear networks (see Okabe and Yamada, 2001; Ang et al., arXiv:1812.09071v2 [math.ST] 2 Jan 2019 2012; Baddeley et al., 2014; McSwiggan et al., 2016; Baddeley et al., 2017; Rakshit et al., 2017).
    [Show full text]
  • ARCH/GARCH Models in Applied Financial Econometrics
    JWPR026-Fabozzi c114-NP November 27, 2007 10:17 1 CHAPTER NP 2 3 4 5 6 ARCH/GARCH Models in Applied 7 8 Financial Econometrics 9 10 11 ROBERT F. ENGLE, PhD 12 Michael Armellino Professorship in the Management of Financial Services, Leonard N. Stern School 13 of Business, New York University 14 15 SERGIO M. FOCARDI 16 Partner, The Intertek Group 17 18 19 FRANK J. FABOZZI, PhD, CFA, CPA 20 Professor in the Practice of Finance, School of Management, Yale University 21 22 23 24 25 Review of Linear Regression and Autoregressive Integration of First, Second, and Higher Moments 7 26 Models 2 Generalizations to High-Frequency Data 7 27 ARCH/GARCH Models 3 Multivariate Extensions 9 28 Application to Value at Risk 5 Summary 10 29 Why ARCH/GARCH? 5 References 10 30 Generalizations of the ARCH/GARCH Models 7 31 32 33 34 Abstract: Volatility is a key parameter used in many financial applications, from deriva- 35 tives valuation to asset management and risk management. Volatility measures the size 36 of the errors made in modeling returns and other financial variables. It was discov- 37 ered that, for vast classes of models, the average size of volatility is not constant but 38 changes with time and is predictable. Autoregressive conditional heteroskedasticity 39 (ARCH)/generalized autoregressive conditional heteroskedasticity (GARCH) models 40 and stochastic volatility models are the main tools used to model and forecast volatil- 41 ity. Moving from single assets to portfolios made of multiple assets, we find that not 42 only idiosyncratic volatilities but also correlations and covariances between assets are 43 time varying and predictable.
    [Show full text]
  • On Directed Random Graphs and Greedy Walks on Point Processes
    UPPSALA DISSERTATIONS IN MATHEMATICS 97 On Directed Random Graphs and Greedy Walks on Point Processes Katja Gabrysch Department of Mathematics Uppsala University UPPSALA 2016 Dissertation presented at Uppsala University to be publicly examined in Polhemsalen, Ångströmlaboratoriet, Lägerhyddsvägen 1, Uppsala, Friday, 9 December 2016 at 13:15 for the degree of Doctor of Philosophy. The examination will be conducted in English. Faculty examiner: Professor Thomas Mountford (EPFL, Switzerland). Abstract Gabrysch, K. 2016. On Directed Random Graphs and Greedy Walks on Point Processes. Uppsala Dissertations in Mathematics 97. 28 pp. Uppsala: Department of Mathematics. ISBN 978-91-506-2608-7. This thesis consists of an introduction and five papers, of which two contribute to the theory of directed random graphs and three to the theory of greedy walks on point processes. We consider a directed random graph on a partially ordered vertex set, with an edge between any two com- parable vertices present with probability p, independently of all other edges, and each edge is directed from the vertex with smaller label to the vertex with larger label. In Paper I we consider a directed random graph on 2 with the vertices ordered according to the product order and we show that the limiting distribution of the centered and rescaled length of the longest path from (0,0) to (n, na ), a<3/14, is ℤthe Tracy-Widom distribution. In Paper II we show that, under a suitable rescaling, the closure of vertex 0 of a directed random graph on with edge probability n−1 converges⌊ in⌋ distribution to the Poisson-weighted infinite tree.
    [Show full text]