Web-Books
im Austria-Forum
Austria-Forum
Web-Books
Naturwissenschaften
Physik
Differential Geometrical Theory of Statistics
Seite - 449 -
  • Benutzer
  • Version
    • Vollversion
    • Textversion
  • Sprache
    • Deutsch
    • English - Englisch

Seite - 449 - in Differential Geometrical Theory of Statistics

Bild der Seite - 449 -

Bild der Seite - 449 - in Differential Geometrical Theory of Statistics

Text der Seite - 449 -

Entropy2016,18, 110 3.2. EntropyandStatistics of theGilbert–VarshamovLine TheGilbert–VarshamovlineR=1−Hq(δ)canbecharacterizedstatisticallly. Suchastatistical descriptioncanbeobtainedbyconsideringtheShannonRandomCodeEnsemble (SRCE).Theseare randomcodesobtainedbychoosingcodewordsas independentrandomvariableswithrespect toa uniformBernoullimeasure, so thatacode isdescribedbyarandomlychosensetofdifferentwordsof lengthnoccurringwithprobabilityq−n, see [26,27]. There isnoapriori reasonwhythe typeofcodes weconsiderhere,withcodewords formedusingthesyntacticparametersofnatural languages,would be linear. Thus,weconsider thegeneral settingofunstructuredcodes,as inSectionVof [27]. TheHammingvolumeVolq(n,d= nδ), that is, thenumberofwordsof lengthnatHamming distanceatmostd fromagivenone, canbeestimated in termsof theq-aryShannonentropy Hq(δ)= δ logq(q−1)−δ logqδ−(1−δ) logq(1−δ) in the form q(Hq(δ)−o(1))n≤Volq(n,d=nδ)= d ∑ j=0 ( n j ) (q−1)j≤ qHq(δ)n. Theexpectationvalue for the randomvariable counting thenumberofunorderedpairsofdistinct codewordswithHammingdistanceatmostd is thenestimatedas E∼ ( qk 2 ) Volq(n,d)q−n∼ qn(Hq(δ)−1+2R)+o(n). Thisestimate is thenused(see [26,27]) toshowthat theprobability tohavecodes in theSRCEwith Hq(δ)≥max{1−2R,0}+ isboundedbyq− n(1+o(1)). Byasimilarargument (seeSectionVof [27] and Proposition 2.2 of [16]) it is shown that the probability that Hq(δ) ≥ 1−R+ is bounded byq−n (1+o(1)). While, by this typeof argument, one can see theGilbert–Varshamov line as representing the typicalbehaviorof sufficientlyrandomcodes, theasymptoticbounddoesnothaveasimilarstatistical interpretation. It does have, however, a relation toKolmogorov complexity, which is relevant to thepoint ofviewdiscussed in thepresentpaper. The relationbetweenasymptotic boundof error correcting codesandKolmogorovcomplexitywasdescribed in [16]. We recall it in the rest of this section,alongwith its implications for the linguisticapplicationsweareconsidering. 3.3. KolmogorovComplexity Werefer thereaderto[30] foranextensivetreatmentofKolmogorovcomplexityanditsproperties. Werecallheresomebasic factsweneedin the following. LetTU be auniversal Turingmachine, that is, a Turingmachine that can simulate any other arbitraryTuringmachine,byreadingontapeboth the inputandthedescriptionof theTuringmachine it shouldsimulate.AprefixTuringmachine isaTuringmachinewithunidirectional inputandoutput tapesandbidirectionalwork tapes. ThesetofprogramsPonwhichaprefixTuringmachinehalts formsaprefixcode. Givenastringw inanalphabetA, theprefixKolmogorovcomplexity isgivenbyminimal length ofaprogramforwhichtheuniversalprefixTuringmachineTU outputsw, KTU(w)= minP:TU(P)=w (P). There isauniversalityproperty.Namely,givenanyotherprefixTuringmachineT, onehas KT(w)≤KTU(w)+cT, 449
zurück zum  Buch Differential Geometrical Theory of Statistics"
Differential Geometrical Theory of Statistics
Titel
Differential Geometrical Theory of Statistics
Autoren
Frédéric Barbaresco
Frank Nielsen
Herausgeber
MDPI
Ort
Basel
Datum
2017
Sprache
englisch
Lizenz
CC BY-NC-ND 4.0
ISBN
978-3-03842-425-3
Abmessungen
17.0 x 24.4 cm
Seiten
476
Schlagwörter
Entropy, Coding Theory, Maximum entropy, Information geometry, Computational Information Geometry, Hessian Geometry, Divergence Geometry, Information topology, Cohomology, Shape Space, Statistical physics, Thermodynamics
Kategorien
Naturwissenschaften Physik
Web-Books
Bibliothek
Datenschutz
Impressum
Austria-Forum
Austria-Forum
Web-Books
Differential Geometrical Theory of Statistics