Hartley's Information Measure • Messages Are Strings of Characters

Hartley's Information Measure • Messages Are Strings of Characters

Hartley’s Information Measure • Messages are strings of characters from a fixed alphabet. • The amount of information contained in a message should be a function of the total number of possible mes- sages. • If you have an alphabet with s sym- bols, then there are s` messages of length, `. • The amount of information contained in two messages should be the sum of the information contained in the individual messages. Hartley’s Information Measure (contd.) • The amount of information in ` mes- sages of length one should equal the amount of information in one mes- sage of length `. It is clear that the only function which satisfies these requirements is the log function: `log(s)= log(s`). If the base of the logarithm is two, then the unit of information is the bit. Shannon’s Information Measure Let X be a discrete r.v. with n outcomes, {x1,...,xn}. The probability that the out- come will be xi is pX(xi). The informa- tion contained in a message about the outcome of X is: −log pX(xi). The avg. information or entropy of a message about the outcome of X is: n HX = − ∑ pX(xi)log pX(xi). i=1 Example Let X be a discrete r.v. with two out- comes, {x1,x2}. The probability that the outcome will be x1 is θ and the prob- ability that the outcome will be x2 is 1 − θ. The avg. information contained in a message about the outcome of X is: HX = −θlog(θ) − (1 − θ)log(1 − θ). We observe that the avg. information θ θ 1 is maximized when = 1 − = 2, in which case HX = 1 bit. Joint Information Let X be a discrete r.v. with outcomes, {x1,...,xn} and let Y be a discrete r.v. with outcomes, {y1,...,ym}. The proba- bility that the outcome of X is xi and the outcome of Y is y j is pXY (xi,y j). The amount of information contained in a message about the outcome of X and Y is: −log pXY (xi,y j). The avg. information or entropy of a message about the outcome of X and Y is: n m HXY = − ∑ ∑ pXY (xi,y j)log pXY (xi,y j). i=1 j=1 1 line 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 10 20 30 40 50 60 70 80 90 100 Figure 1: HX = −θlog(θ) − (1 − θ)log(1 − θ). Properties of Shannon’s Measure • HX is continuous in the pX(xi). • HX is symmetric. That is, HX = HY when pY (x1)= pX(x2) and pY (x2)= pX(x1). More generally, HX is in- variant under permutation of the dis- tribution function, pX . • HX is additive. That is, when X and Y are independent r.v.’s, then HXY = HX + HY . • HX is maximum when all of the pX(xi)’s are equal. • HX is minimum when one of the pX(xi)’s equals one. Additivity Example Let X and Y be fair dice. The avg. amount of information contained in a message about the outcome of X and Y is: 6 6 1 1 HXY = − ∑ ∑ log ≈ 5.16 bits. i=1 j=1 36 36 The avg. amount of information con- tained in a message about the outcome of X is: 6 1 1 HX = − ∑ log ≈ 2.58 bits. i=1 6 6 Since HX = HY , it follows that HX + HY ≈ 5.16 bits. Symmetry Example • Let X be a discrete r.v. with out- comes, {A,G,C,T}, which occur with 1 1 1 1 probabilities, {4, 8, 8, 2}. • Let Y be a discrete r.v. with out- comes, {♣,♠,♦,♥}, which occur with 1 1 1 1 probabilities, {4, 8, 2, 8}. • The avg. amount of information con- tained in a message about the out- come of X is: 1 1 1 1 1 1 1 1 H = − log − log − log − log X 4 4 8 8 8 8 2 2 = 1.75 bits Symmetry Example (contd.) • The avg. amount of information con- tained in a message about the out- come of Y is: 1 1 1 1 1 1 1 1 H = − log − log − log − log Y 4 4 8 8 2 2 8 8 = 1.75 bits Theorem 1.1 Let X be a discrete r.v. with n outcomes, {x1,...,xn}. The probability that the out- come will be xi is pX(xi). Then • HX ≤ logn with HX = logn if and only if for all i it is true that pX(xi)= 1/n. • HX ≥ 0 with HX = 0 if and only if there exists a k such that pX(xk)= 1. Theorem 1.1 (contd.) Proof: HX − logn = n − ∑ pX(xi)log pX(xi) − logn = i=1 n n − ∑ pX(xi)log pX(xi)−∑ pX(xi)logn = i=1 i=1 n − ∑ pX(xi)(log pX(xi)+ logn)= i=1 n 1 ∑ p (x )log . X i np (x ) i=1 X i Theorem 1.1 (contd.) From the inequality lna ≤ a−1 and the fact that loga = lna/ln2: lna ≤ (a − 1) lna/ln2 ≤ (a − 1)/ln2 loga ≤ (a − 1)/ln2 loga ≤ (a − 1)lne/ln2 loga ≤ (a − 1)loge 10 line 1 line 2 8 6 4 2 0 -2 -4 0 1 2 3 4 5 6 7 8 9 10 Figure 2: lna ≤ a − 1. Theorem 1.1 (contd.) Using this result in the expression for HX − logn yields: n 1 H − logn = ∑ p (x )log X X i np (x ) i=1 X i n 1 ≤ ∑ p (x ) − 1 loge X i np (x ) i=1 X i n 1 n ≤ ∑ − ∑ pX(xi) loge i=1 n i=1 ! 1 ≤ n − 1 loge n ≤ 0 Theorem 1.1 (contd.) This proves that HX ≤ logn. To prove that HX ≥ 0, we observe that: •∀i pX(xi) ≥ 0 •∀i − log pX(xi) ≥ 0 It follows that: n − ∑ pX(xi)log pX(xi) ≥ 0. i=1 Maximum Entropy Let X be a r.v. with outcomes, {x1,...,xn}. These outcomes occur with probability, pX(xi) = 1/n for all i. The avg. in- formation contained in a message about the outcome of X is: n HX = − ∑ pX(xi)log pX(xi) i=1 n 1 1 = − ∑ log i=1 n n 1 1 n = − log ∑ 1 n n i=1 1 1 1 = − log n = −log n n n = logn Maximum Entropy Example Let X be a discrete r.v. with outcomes, {A,G,C,T}. These outcomes occur with 1 1 1 1 probabilities, {4, 4, 4, 4}. The avg. amount of information contained in a message about the outcome of X is: 1 1 1 1 1 1 1 1 H = − log − log − log − log X 4 4 4 4 4 4 4 4 = 2 bits The genome of the bacterium, E. coli, is a DNA molecule consisting of 4 × 106 base pairs. The maximum amount of information stored in the E. coli genome is therefore 8 × 106 bits. Minimum Entropy Example Let X be a discrete r.v. with outcomes, {A,G,C,T}. These outcomes occur with probabilities, {0,1,0,0}. The avg. amount of information contained in a message about the outcome of X is: HX = −0log0 − 1log1 − 0log0 − 0log0 = 0 bits..

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    19 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us