On a Variational Definition for the Jensen-Shannon Symmetrization Of

On a Variational Definition for the Jensen-Shannon Symmetrization Of

entropy Article On a Variational Definition for the Jensen-Shannon Symmetrization of Distances Based on the Information Radius Frank Nielsen Sony Computer Science Laboratories, Tokyo 141-0022, Japan; [email protected] Abstract: We generalize the Jensen-Shannon divergence and the Jensen-Shannon diversity index by considering a variational definition with respect to a generic mean, thereby extending the notion of Sibson’s information radius. The variational definition applies to any arbitrary distance and yields a new way to define a Jensen-Shannon symmetrization of distances. When the variational optimization is further constrained to belong to prescribed families of probability measures, we get relative Jensen-Shannon divergences and their equivalent Jensen-Shannon symmetrizations of distances that generalize the concept of information projections. Finally, we touch upon applications of these variational Jensen-Shannon divergences and diversity indices to clustering and quantization tasks of probability measures, including statistical mixtures. Keywords: Jensen-Shannon divergence; diversity index; Rényi entropy; information radius; information projection; exponential family; Bregman divergence; Fenchel–Young divergence; Bregman information; q-exponential family; q-divergence; Bhattacharyya distance; centroid; clustering 1. Introduction: Background and Motivations Citation: Nielsen, F. On a Variational Definition for the Jensen-Shannon The goal of the author is to methodologically contribute to an extension of the Sib- Symmetrization of Distances Based son’s information radius [1] and also concentrate on analysis of the specified families of on the Information Radius. Entropy distributions called exponential families [2]. 2021, 23, 464. https://doi.org/ Let (X , F) denote a measurable space [3] with sample space X and s-algebra F on 10.3390/e23040464 the set X . The Jensen-Shannon divergence [4] (JSD) between two probability measures P and Q (or probability distributions) on (X , F) is defined by: Academic Editor: Yong Deng 1 P + Q P + Q DJS[P, Q] := DKL P : + DKL Q : , (1) Received: 12 March 2021 2 2 2 Accepted: 9 April 2021 Published: 14 April 2021 where DKL denotes the Kullback–Leibler divergence [5,6] (KLD): ( R dP(x) Publisher’s Note: MDPI stays neutral X log dQ(x) dP, P Q DKL[P : Q] := (2) with regard to jurisdictional claims in +¥, P 6 Q published maps and institutional affil- iations. dP where P Q means that P is absolutely continuous with respect to Q [3], and dQ is the Radon–Nikodym derivative of P with respect to Q. Equation (2) can be rewritten using the chain rule as: ( R dP(x) dP(x) X dQ(x) log dQ(x) dQ, P Q DKL[P : Q] := (3) Copyright: © 2021 by the author. +¥, P 6 Q Licensee MDPI, Basel, Switzerland. dP This article is an open access article Consider a measure m for which both the Radon–Nikodym derivatives p := dm distributed under the terms and dP P+Q and q := dm exist (e.g., m = 2 ). Subsequently the Kullback–Leibler divergence can conditions of the Creative Commons be rewritten as (see Equation (2.5) page 5 of [5] and page 251 of the Cover & Thomas’ Attribution (CC BY) license (https:// textbook [6]): creativecommons.org/licenses/by/ 4.0/). Entropy 2021, 23, 464. https://doi.org/10.3390/e23040464 https://www.mdpi.com/journal/entropy Entropy 2021, 23, 464 2 of 28 Z p(x) DKL[p : q] := p(x) log dm(x). (4) X q(x) Denote by D = D(X ) the set of all densities with full support X (Radon–Nikodym derivatives of probability measures with respect to m): Z D(X ) := p : X! R : p(x) > 0 m-almost everywhere, p(x)dm(x) = 1 . X Subsequently, the Jensen-Shannon divergence [4] between two densities p and q of D is defined by: 1 p + q p + q D [p, q] := D p : + D q : . (5) JS 2 KL 2 KL 2 d d d Often, one considers the Lebesgue measure [3] m = mL on (R , B(R )), where B(R ) is X the Borel s-algebra, or the counting measure [3] m = m# on (X , 2 ) where X is a countable set, for defining the measure space (X , F, m). The JSD belongs to the class of f -divergences [7–9] which are known as the invariant decomposable divergences of information geometry (see [10], pp. 52–57). Although the KLD is asymmetric (i.e., DKL[p : q] 6= DKL[q : p]), the JSD is symmetric (i.e., DJS[p, q] = DJS[q, p]). The notation ‘:’ is used as a parameter separator to indicate that the parameters are not permutation invariant, and that the order of parameters is important. In this work, a distance D(O1 : O2) is a measure of dissimilarity between two objects O1 and O2, which do not need to be symmetric or satisfy the triangle inequality of metric distances. A distance only satisfies the identity of indiscernibles: D(O1 : O2) = 0 if and only if O1 = O2. When the objects O1 and O2 are probability densities with respect to m, we call this distance a statistical distance, use the brackets to enclose the arguments of the statistical distance (i.e., D[O1 : O2]), and we have D[O1 : O2] = 0 if and only if O1(x) = O2(x) m-almost everywhere. The 2-point JSD of Equation (4) can be extended to a weighted set of n densities P := n f(w1, p1), ... , (wn, pn)g (with positive wi’s normalized to sum up to unity, i.e., ∑i=1 wi = 1) thus providing a diversity index, i.e., a n-point JSD for P: n DJS(P) := ∑ wiDKL[pi : p¯], (6) i=1 n where p¯ := ∑i=1 wi pi denotes the statistical mixture [11] of the densities of P. We have 1 1 DJS[p : q] = DJS(f( 2 , p), ( 2 , q)g). We call DJS(P) the Jensen-Shannon diversity index. The KLD is also called the relative entropy since it can be expressed as the difference between the cross entropy h[p : q] and the entropy h[p]: Z p(x) DKL[p : q] := p(x) log dm(x) (7) X q(x) Z Z = p(x) log p(x)dm(x) − p(x) log q(x)dm(x), (8) X X = h[p : q] − h[p], (9) with the cross-entropy and entropy defined, respectively, by Z h[p : q] := − p(x) log q(x)dm(x), (10) X Z h[p] := − p(x) log p(x)dm(x). (11) X Because h[p] = h[p : p], we may say that the entropy is the self-cross-entropy. Entropy 2021, 23, 464 3 of 28 When m is the Lebesgue measure, the Shannon entropy is also called the differential entropy [6]. Although the discrete entropy H[p] = − ∑i pi log pi (i.e., entropy with respect to the counting measure) is always positive and bounded by log jX j, the differential entropy may be negative (e.g., entropy of a Gaussian distribution with small variance). The Jensen-Shannon divergence of Equation (6) can be rewritten as: n DJS[p, q] = h[p¯] − ∑ wih[pi] := J−h[p, q]. (12) i=1 The JSD representation of Equation (12) is a Jensen divergence [12] for the strictly convex negentropy F(p) = −h[p], since the entropy function h[.] is strictly concave. Therefore, it is appropriate to call this divergence the Jensen-Shannon divergence. ( ) ( ) Because pi x ≤ pi x = 1 , it can be shown that the Jensen-Shannon diversity p¯(x) wi pi(x) wi n index is upper bounded by H(w) := − ∑i=1 wi log wi, the discrete Shannon entropy. Thus, the Jensen-Shannon diversity index is bounded by log n, and the 2-point JSD is bounded by log 2, although the KLD is unbounded and it may even be equal to +¥ when the definite integral diverges (e.g., KLD between the standard Cauchy distribution and the standard Gaussian distribution). Another nice property of the JSD is that its square root yields a metric distance [13,14]. This property further holds for the quantum JSD [15]. The JSD has gained interest in machine learning. See, for example, the Generative Adversarial Networks [16] (GANs) in deep learning [17], where it was proven that minimizing the GAN objective function by adversarial training is equivalent to minimizing a JSD. 1 To delineate the different roles that are played by the factor 2 in the ordinary Jensen- Shannon divergence (i.e., in weighting the two KLDs and in weighting the two densities), let us introduce two scalars a, b 2 (0, 1), and define a generic (a, b)-skewed Jensen-Shannon divergence, as follows: DJS,a,b[p : q] := (1 − b)DKL[p : ma] + bDKL[q : ma], (13) = (1 − b)h[p : ma] + bh[q : ma] − (1 − b)h[p] − bh[q], (14) = h[mb : ma] − ((1 − b)h[p] + bh[q]), (15) where ma := (1 − a)p + aq and mb := (1 − b)p + bq. This identity holds, because DJS,a,b 1 1 is bounded by (1 − b) log 1−a + b log a , see [18]. Thus, when b = a, we have DJS,a[p, q] = DJS,a,a[p, q] = h[ma] − ((1 − a)h[p] + ah[q]), since the self-cross entropy corresponds to the entropy: h[ma : ma] = h[ma]. A f -divergence [9,19,20] is defined for a convex generator f , which is strictly convex at 1 (to satisfy the identity of the indiscernibles) and that satisfies f (1) = 0, by Z q(x) I [p : q] := p(x) f dm(x) ≥ f (1) = 0, (16) f p(x) where the right-hand-side follows from Jensen’s inequality [20]. For example, the total 1 R variation distance DTV[p : q] = 2 X jp(x) − q(x)jdm(x) is a f -divergence for the generator ( ) = j − j [ ] = [ ] ( ) fTV u u 1 : DTV p : q IfTV p : q .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    28 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us