Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing Haim H

Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing Haim H

3248 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 6, JUNE 2011 Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing Haim H. Permuter, Member, IEEE, Young-Han Kim, Member, IEEE, and Tsachy Weissman, Senior Member, IEEE Abstract—We investigate the role of directed information in plays a central role in Shannon’s random coding ar- portfolio theory, data compression, and statistics with causality gument, because the probability that independently drawn se- constraints. In particular, we show that directed information is quences and an upper bound on the increment in growth rates of optimal portfolios in a stock market due to causal side information. This “look” as if they were drawn jointly decays exponentially in upper bound is tight for gambling in a horse race, which is an with as the first order in the exponent. Shannon extreme case of stock markets. Directed information also char- also proved a dual result [2] that the rate distortion function acterizes the value of causal side information in instantaneous , the minimum compression rate to describe a source compression and quantifies the benefit of causal inference in joint by its reconstruction within average distortion ,isgiven compression of two stochastic processes. In hypothesis testing, directed information evaluates the best error exponent for testing by . In another duality result (the whether a random process causally influences another process Lagrange duality this time) to (1), Gallager [3] proved the min- or not. These results lead to a natural interpretation of directed imax redundancy theorem, connecting the redundancy of the information s@ n 3 nA as the amount of information that n universal lossless source code to the maximum mutual infor- a random sequence a@IYPY FFFYnA causally provides n mation (capacity) of the channel with conditional distribution about another random sequence a@IYPY FFFYnA.A new measure, directed lautum information, is also introduced and consisting of the set of possible source distributions (cf. [4]). interpreted in portfolio theory, data compression, and hypothesis It has been shown that mutual information plays important testing. roles in problems beyond those related to describing sources Index Terms—Causal conditioning, causal side information, di- or transferring information through channels. Among the most rected information, hypothesis testing, instantaneous compression, celebrated of such examples is the use of mutual information in Kelly gambling, lautum information, portfolio theory. gambling. In 1956, Kelly [5] showed that if a horse race outcome can be represented as an independent and identically distributed (i.i.d.) random variable , and the gambler has some side infor- I. INTRODUCTION mation relevant to the outcome of the race, then the mutual in- formation captures the difference between growth rates UTUAL information between two random of the optimal gambler’s wealth with and without side informa- variables and arises as the canonical answer to a M tion . Thus, Kelly’s result provides an interpretation of mutual variety of questions in science and engineering. Most notably, information as the financial value of side information Shannon [1] showed that the capacity , the maximal data rate for gambling in the horse race . for reliable communication, of a discrete memoryless channel In order to tackle problems arising in information systems with input and output is given by with causally dependent components, Massey [6], inspired by Marko’s work [7] on bidirectional communication, coined the (1) notion of “directed information” from to , defined as Shannon’s channel coding theorem leads naturally to the opera- (2) tional interpretation of mutual information as the amount of uncertainty about that can be re- and showed that the normalized maximum directed information duced by observing , or equivalently, the amount of informa- upper bounds the capacity of channels with feedback. Subse- tion that can provide about . Indeed, mutual information quently, it was shown that directed information, as defined by Massey, indeed characterizes the capacity of channels with feed- Manuscript received December 24, 2009; revised November 24, 2010; ac- back [8]–[16] and the rate distortion function with feedforward cepted November 27, 2010. Date of current version May 25, 2011. This work was supported in part by NSF Grant CCF-0729195 and in part by BSF Grant [17]. Note that directed information (2) can also be rewritten as 2008402. H. H. Permuter was supported in part by the Marie Curie Reintegra- tion fellowship. (3) H. H. Permuter is with the Department of Electrical and Computer Engi- neering, Ben-Gurion University of the Negev, Beer-Sheva 84105, Israel (e-mail: [email protected]). each term of which corresponds to the achievable rate at time Y.-H. Kim is with the Department of Electrical and Computer Engineering, University of California San Diego, La Jolla, CA 92093-0407 USA (e-mail: given side information (cf. [11] for the details). [email protected]). Recently, directed information has also been used in models of T. Weissman is with the Information Systems Lab, Stanford University, Stan- computational biology [18]–[21] and in the context linear pre- ford, CA 94305 USA (e-mail: [email protected]). Communicated by I. Kontoyiannis, Associate Editor for Shannon Theory. diction representation for the rate distortion function of a sta- Digital Object Identifier 10.1109/TIT.2011.2136270 tionary Gaussian source [22]. 0018-9448/$26.00 © 2011 IEEE PERMUTER et al.: INTERPRETATIONS OF DIRECTED INFORMATION IN PORTFOLIO THEORY, DATA COMPRESSION, AND HYPOTHESIS TESTING 3249 The main contribution of this paper is showing that directed Under this notation, directed information defined in (2) can be information has natural interpretations in portfolio theory, com- rewritten as pression, and statistics when causality constraints exist. In stock market investment (Section III), directed information between (9) the stock price and side information is an upper bound which hints, in a rough analogy to mutual informa- on the increase in growth rates due to causal side informa- tion, a possible interpretation of directed information tion. This upper bound is tight when specialized to gambling as the amount in horse races. In data compression (Section IV), we show that of information that causally available side information can directed information characterizes the value of causal side infor- provide about . mation in instantaneous compression and it also quantifies the Note that the channel capacity results involve the term role of causal inference in joint compression of two stochastic , which measures the amount of information processes. In hypothesis testing (Section V), we show that di- transfer over the forward link from to . In gambling, rected information is the exponent of the minimum type II error however, the increase in growth rate is due to the side infor- probability when one is to decide whether has causal influ- mation (the backward link), and, therefore, the expression ence on or not. Finally, we introduce the notion of directed appears. Throughout the paper we also use the lautum1 information (Section VI), which extends the notion of notation which denotes the directed infor- lautum information by Palomar and Verdú [23] to accommodate mation between the -tuple , i.e., the null symbol causality. We briefly discuss its role in horse race gambling, data followed by , and , that is, compression, and hypothesis testing. II. PRELIMINARIES:DIRECTED INFORMATION AND (10) CAUSAL CONDITIONING Throughout this paper, we use the causal conditioning nota- Using the causal conditioning notation, given in (8), the directed tion developed by Kramer [8]. We denote by information can be written as the probability mass function of causally (11) conditioned on for some integer , which is defined as Directed information and mutual information obey the conser- vation law (4) (12) By convention, if , then is set to null, i.e., if which was established by Massey and Massey [24]. The conser- then is just . In particular, we use vation law is a direct consequence of the chain rule (7), and we extensively the cases show later in Section IV-B that it has a natural interpretation as a conservation of a mismatch cost in data compression. (5) The causally conditional entropy rate of a random process given another random process and the directed information rate from to are defined respectively as (6) (13) Using the chain rule, it is easily verified that (14) (7) when these limits exist. In particular, when is stationary The causally conditional entropy and ergodic, both quantities are well-defined, namely, the limits in are defined respectively as (13) and (14) exist [8, Properties 3.5 and 3.6]. III. PORTFOLIO THEORY Here we show that directed information is an upper bound on the increment in growth rates of optimal portfolios in a stock market due to causal side information. We start by considering a special case of gambling in a horse race market and show that the upper bound is tight. Then we consider the general case of stock market investment. (8) A. Horse Race Gambling With Causal Side Information 1Lautum (“elegant” in Latin) is the reverse spelling of “mutual” as aptly Assume that there are racing horses and let denote the coined in [23]. winning horse at time , i.e., . At time 3250 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 6, JUNE 2011 , the gambler has some side information relevant to the perfor- Furthermore, note that the best strategy is greedy, namely, at any mance of the horses, which we denote as . For instance, a pos- time the best strategy is to maximize regardless of sible side information may be the health condition of the horses the horizon . In other words, at any time the gambler should and the jockeys on the day of the race. We assume that the gam- maximize the expected growth rate, i.e., . bler invests all his/her capital in the horse race as a function of Proof: Consider the previous horse race outcomes and side information up to time .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    12 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us