Page - 449 - in Differential Geometrical Theory of Statistics
Image of the Page - 449 -
Text of the Page - 449 -
Entropy2016,18, 110
3.2. EntropyandStatistics of theGilbertâVarshamovLine
TheGilbertâVarshamovlineR=1âHq(δ)canbecharacterizedstatisticallly. Suchastatistical
descriptioncanbeobtainedbyconsideringtheShannonRandomCodeEnsemble (SRCE).Theseare
randomcodesobtainedbychoosingcodewordsas independentrandomvariableswithrespect toa
uniformBernoullimeasure, so thatacode isdescribedbyarandomlychosensetofdifferentwordsof
lengthnoccurringwithprobabilityqân, see [26,27]. There isnoapriori reasonwhythe typeofcodes
weconsiderhere,withcodewords formedusingthesyntacticparametersofnatural languages,would
be linear. Thus,weconsider thegeneral settingofunstructuredcodes,as inSectionVof [27].
TheHammingvolumeVolq(n,d= nδ), that is, thenumberofwordsof lengthnatHamming
distanceatmostd fromagivenone, canbeestimated in termsof theq-aryShannonentropy
Hq(δ)= δ logq(qâ1)âδ logqδâ(1âδ) logq(1âδ)
in the form
q(Hq(δ)âo(1))nâ¤Volq(n,d=nδ)= d
â
j=0 (
n
j )
(qâ1)j⤠qHq(δ)n.
Theexpectationvalue for the randomvariable counting thenumberofunorderedpairsofdistinct
codewordswithHammingdistanceatmostd is thenestimatedas
Eâź ( qk
2 )
Volq(n,d)qânâź qn(Hq(δ)â1+2R)+o(n).
Thisestimate is thenused(see [26,27]) toshowthat theprobability tohavecodes in theSRCEwith
Hq(δ)âĽmax{1â2R,0}+ isboundedbyqâ n(1+o(1)). Byasimilarargument (seeSectionVof [27]
and Proposition 2.2 of [16]) it is shown that the probability that Hq(δ) ⼠1âR+ is bounded
byqân (1+o(1)).
While, by this typeof argument, one can see theGilbertâVarshamov line as representing the
typicalbehaviorof sufďŹcientlyrandomcodes, theasymptoticbounddoesnothaveasimilarstatistical
interpretation. It does have, however, a relation toKolmogorov complexity, which is relevant to
thepoint ofviewdiscussed in thepresentpaper. The relationbetweenasymptotic boundof error
correcting codesandKolmogorovcomplexitywasdescribed in [16]. We recall it in the rest of this
section,alongwith its implications for the linguisticapplicationsweareconsidering.
3.3. KolmogorovComplexity
Werefer thereaderto[30] foranextensivetreatmentofKolmogorovcomplexityanditsproperties.
Werecallheresomebasic factsweneedin the following.
LetTU be auniversal Turingmachine, that is, a Turingmachine that can simulate any other
arbitraryTuringmachine,byreadingontapeboth the inputandthedescriptionof theTuringmachine
it shouldsimulate.ApreďŹxTuringmachine isaTuringmachinewithunidirectional inputandoutput
tapesandbidirectionalwork tapes. ThesetofprogramsPonwhichapreďŹxTuringmachinehalts
formsapreďŹxcode.
Givenastringw inanalphabetA, thepreďŹxKolmogorovcomplexity isgivenbyminimal length
ofaprogramforwhichtheuniversalpreďŹxTuringmachineTU outputsw,
KTU(w)= minP:TU(P)=w (P).
There isauniversalityproperty.Namely,givenanyotherpreďŹxTuringmachineT, onehas
KT(w)â¤KTU(w)+cT,
449
Differential Geometrical Theory of Statistics
- Title
- Differential Geometrical Theory of Statistics
- Authors
- FrĂŠdĂŠric Barbaresco
- Frank Nielsen
- Editor
- MDPI
- Location
- Basel
- Date
- 2017
- Language
- English
- License
- CC BY-NC-ND 4.0
- ISBN
- 978-3-03842-425-3
- Size
- 17.0 x 24.4 cm
- Pages
- 476
- Keywords
- Entropy, Coding Theory, Maximum entropy, Information geometry, Computational Information Geometry, Hessian Geometry, Divergence Geometry, Information topology, Cohomology, Shape Space, Statistical physics, Thermodynamics
- Categories
- Naturwissenschaften Physik