Web-Books
im Austria-Forum
Austria-Forum
Web-Books
Naturwissenschaften
Physik
Differential Geometrical Theory of Statistics
Seite - 296 -
  • Benutzer
  • Version
    • Vollversion
    • Textversion
  • Sprache
    • Deutsch
    • English - Englisch

Seite - 296 - in Differential Geometrical Theory of Statistics

Bild der Seite - 296 -

Bild der Seite - 296 - in Differential Geometrical Theory of Statistics

Text der Seite - 296 -

Entropy2016,18, 442 Without loss of generality, consider GMMs in the formm(x) = ∑ki=1wip(x;μi,Σi) (Σi=σ 2 i for univariate Gaussians). The mean μ¯ of the mixture is μ¯=∑ki=1wiμi and the variance is σ¯2=E[m2]−E[m]2. SinceE[m2]=∑ki=1wi ∫ x2p(x;μi,Σi)dx=∑ki=1wi ( μ2i+σ 2 i ) ,wededuce that σ¯2= k ∑ i=1 wi(μ2i+σ 2 i )− ( k ∑ i=1 wiμi )2 = k ∑ i=1 wi [ (μi− μ¯)2+σ2i ] . The entropyof a randomvariablewith aprescribedvariance σ¯2 ismaximal for theGaussian distribution with the same variance σ¯2, see [4]. Since the differential entropy of a Gaussian is log(σ¯ √ 2πe),wededuce that theentropyof theGMMisupperboundedby H(m)≤ 1 2 log(2πe)+ 1 2 log k ∑ i=1 wi [ (μi− μ¯)2+σ2i ] . Thisupperboundcanbeeasilygeneralizedtoarbitrarydimensionality.Weget thefollowinglemma: Lemma 2. The entropy of a d-variate GMM m(x) = ∑ki=1wip(x;μi,Σi) is upper bounded by d 2 log(2πe)+ 1 2 logdetΣ,whereΣ=∑ k i=1wi(μiμ i +Σi)− ( ∑ki=1wiμi )( ∑ki=1wiμ i ) . Ingeneral, exponential familieshavefinitemomentsof anyorder [17]: Inparticular,wehave E[t(X)]=∇F(θ)andV[t(X)]=∇2F(θ). ForGaussiandistribution,wehavethesufficientstatistics t(x)= (x,x2) so thatE[t(X)] =∇F(θ)yields themeanandvariance fromthe log-normalizer. It is easy togeneralizeLemma2tomixturesofexponential familydistributions. Note that thisbound(calledtheMaximumEntropyUpperBoundin[13],MEUB) is tightwhen theGMMapproximatesasingleGaussian. It is fast tocomputecomparedto theboundreported in [9] thatusesTaylor’ sexpansionof the log-sumof themixturedensity. Asimilar argument cannotbeapplied for a lowerboundsinceaGMMwithagivenvariance mayhaveentropytendingto−∞. Forexample,assumethe2-componentmixture’smean iszero,and that thevarianceapproximates1by takingm(x)= 12G(x;−1, )+ 12G(x;1, )whereGdenotes the Gaussiandensity. Letting →0,weget theentropytendingto−∞. Weremarkthatour log-sum-expinequality techniqueyieldsa log2additiveapproximationrange in thecaseofaGaussianmixturewith twocomponents. It thusgeneralizes theboundsreported in [7] toGMMswitharbitraryvariances thatarenotnecessarilyequal. Tosee theboundgap,wehave −∑ r ∫ Ir m(x) ( logk+ logmax i wipi(x) ) dx≤H(m) ≤−∑ r ∫ Ir m(x)max { logmax i wipi(x), logk+ logmin i wipi(x) } dx. (27) Therefore thegapisatmost Δ=min { ∑ r ∫ Ir m(x) log maxiwipi(x) miniwipi(x) dx, logk } =min { ∑ s ∑ r ∫ Ir wsps(x) log maxiwipi(x) miniwipi(x) dx, logk } . (28) Thustocompute thegaperrorboundof thedifferentialentropy,weneedto integrate termsin the form ∫ wapa(x) log wbpb(x) wcpc(x) dx. 296
zurück zum  Buch Differential Geometrical Theory of Statistics"
Differential Geometrical Theory of Statistics
Titel
Differential Geometrical Theory of Statistics
Autoren
Frédéric Barbaresco
Frank Nielsen
Herausgeber
MDPI
Ort
Basel
Datum
2017
Sprache
englisch
Lizenz
CC BY-NC-ND 4.0
ISBN
978-3-03842-425-3
Abmessungen
17.0 x 24.4 cm
Seiten
476
Schlagwörter
Entropy, Coding Theory, Maximum entropy, Information geometry, Computational Information Geometry, Hessian Geometry, Divergence Geometry, Information topology, Cohomology, Shape Space, Statistical physics, Thermodynamics
Kategorien
Naturwissenschaften Physik
Web-Books
Bibliothek
Datenschutz
Impressum
Austria-Forum
Austria-Forum
Web-Books
Differential Geometrical Theory of Statistics