International Statistical Review (2010), 00, 0, 1–30 doi:10.1111/j.1751-5823.2010.00105.x
Information Measures in Perspective
Nader Ebrahimi1, Ehsan S. Soofi2 and Refik Soyer3
1Division of Statistics, Northern Illinois University, DeKalb, IL 60155, USA E-mail: [email protected] 2Sheldon B. Lubar School of Business, University of Wisconsin-Milwaukee, P.O. Box 742, Milwaukee, WI 53201, USA E-mail: [email protected] 3Department of Decision Sciences, George Washington University, Washington, DC 20052, USA E-mail: [email protected]
Summary Information-theoretic methodologies are increasingly being used in various disciplines. Frequently an information measure is adapted for a problem, yet the perspective of information as the unifying notion is overlooked. We set forth this perspective through presenting information-theoretic methodologies for a set of problems in probability and statistics. Our focal measures are Shannon entropy and Kullback–Leibler information. The background topics for these measures include notions of uncertainty and information, their axiomatic foundation, interpretations, properties, and generalizations. Topics with broad methodological applications include discrepancy between distributions, derivation of probability models, dependence between variables, and Bayesian analysis. More specific methodological topics include model selection, limiting distributions, optimal prior distribution and design of experiment, modeling duration variables, order statistics, data disclosure, and relative importance of predictors. Illustrations range from very basic to highly technical ones that draw attention to subtle points.
Key words: Bayesian information; dynamic information; entropy; Kullback–Leibler information; mutual information.
1 Introduction The information theory offers measures that have axiomatic foundation and are capable of handling diverse problems in a unified manner. However, the information methodologies are often developed in isolation, where a particular measure is used without consideration of the larger picture. This paper takes an integrative approach and draws the common properties of various information measures. This approach enables us to relate solutions to diverse problems. We present information-theoretic solutions to several statistical problems, ranging from very basic to highly technical. The paper is organized into the following parts:
(a) Basic notions and measures of uncertainty and information (Sections 2–4). (b) Information methodologies for model derivation and selection, measuring dependence, and Bayesian analysis (Sections 5–7). (c) Specific areas of applications (Section 8).