Seite - 301 - in Differential Geometrical Theory of Statistics
Bild der Seite - 301 -
Text der Seite - 301 -
Entropy2016,18, 442
LetuspartitionthesupportX=unionmulti r=1Irarbitrarilyintoelementaryranges,whichdonotnecessarily
correspond to the envelopes. Denote by MI the probabilitymass of amixturem(x) in the range I:
MI= â«
Im(x)dx. Then
Df(m :mâČ)= â
r=1 MIr â«
Ir m(x)
MIr f ( mâČ(x)
m(x) )
dx.
Note that in range Ir, m(x)
MIr is a unit weight distribution. Thus by Jensenâs inequality
f(E[X])â€E[f(X)],weget
Df(m :mâČ)â„ â
r=1 MIr f (â«
Ir m(x)
MIr mâČ(x)
m(x) dx )
= â
r=1 MIr f ( MâČIr
MIr )
. (52)
Notice that the RHS of Equation (52) is the f-divergence between (MI1, · · · ,MI ) and
(MâČI1, · · · ,MâČI ), denoted byDIf (m : mâČ). In the special case that = 1 and I1 = X , the above
Equation(52) turnsout tobe theusualGibbsâ inequality:Df(m :mâČ)â„ f(1), andCsiszĂĄrgenerator
is chosenso that f(1)= 0. In conclusion, foraïŹxed (coarse-grained) countablepartitionofX ,we
recover thewell-knowinformationmonotonicity [46]of the f-divergences:
Df(m :mâČ)â„DIf (m :mâČ)â„0.
Inpractice,wegetclosed-formlowerboundswhenMI= â« b
a m(x)dx=Ί(b)âΊ(a) isavailable
inclosed-form,whereΊ(·)denote theCDF. Inparticular, ifm(x) isamixturemodel, then itsCDFcan
becomputedbylinearlycombiningtheCDFsof its components.
Towrapup,wehaveproved that coarse-grainingbymakingaïŹnitepartitionof the support
X yieldsa lowerboundon the f-divergencebyvirtueof the informationmonotonicity. Therefore,
insteadofdoingMonteCarlostochastic integration:
DËnf(m :m âČ)= 1
n n
â
i=1 f ( mâČ(xi)
m(xi) )
,
with x1, . . . ,xn âŒi.i.d. m(x), it could be better to sort those n samples and consider the
coarse-grainedpartition:
I=(ââ,x(1)]âȘ (
unionmultinâ1i=1 (x(i),x(i+1)] )
âȘ(x(n),+â)
togetaguaranteed lowerboundonthe f-divergence.Wewillcall thisboundCGQLBforCoarseGraining
QuantizationLowerBound.
Givenabudgetofn splittingpointson the rangeX , itwouldbe interesting toïŹnd thebestn
points thatmaximize the lowerboundDIf (m :m âČ). This isongoingresearch.
6. Experiments
Weperformanempiricalstudytoverifyourtheoreticalbounds.Wesimulatefourpairsofmixture
models{(EMM1,EMM2),(RMM1,RMM2),(GMM1,GMM2),(GaMM1,GaMM2)}as the test subjects. Thecomponent
type is impliedbythemodelname,whereGaMMstands forGammamixtures. Thecomponentsofeach
mixturemodelaregivenas follows.
1. EMM1âs components, in the form (λi,wi), are given by (0.1,1/3), (0.5,1/3), (1,1/3); EMM2âs
componentsare (2,0.2), (10,0.4), (20,0.4).
2. RMM1âs components, in the form (Ïi,wi), aregivenby (0.5,1/3), (2,1/3), (10,1/3);RMM2 consists
of (5,0.25), (60,0.25), (100,0.5).
301
Differential Geometrical Theory of Statistics
- Titel
- Differential Geometrical Theory of Statistics
- Autoren
- Frédéric Barbaresco
- Frank Nielsen
- Herausgeber
- MDPI
- Ort
- Basel
- Datum
- 2017
- Sprache
- englisch
- Lizenz
- CC BY-NC-ND 4.0
- ISBN
- 978-3-03842-425-3
- Abmessungen
- 17.0 x 24.4 cm
- Seiten
- 476
- Schlagwörter
- Entropy, Coding Theory, Maximum entropy, Information geometry, Computational Information Geometry, Hessian Geometry, Divergence Geometry, Information topology, Cohomology, Shape Space, Statistical physics, Thermodynamics
- Kategorien
- Naturwissenschaften Physik