Web-Books
in the Austria-Forum
Austria-Forum
Web-Books
Naturwissenschaften
Physik
Differential Geometrical Theory of Statistics
Page - 375 -
  • User
  • Version
    • full version
    • text only version
  • Language
    • Deutsch - German
    • English

Page - 375 - in Differential Geometrical Theory of Statistics

Image of the Page - 375 -

Image of the Page - 375 - in Differential Geometrical Theory of Statistics

Text of the Page - 375 -

Entropy2016,18, 98 Givenasetofobservations{Y1, · · · ,YN}arisingfromEquation(27)whereM isunknown, the BICconsistsofchoosingtheparameter: M¯=argmaxMBIC(M) where: BIC(M)=LL− 1 2 ×DF× log(N) (31) Here,LL is the log-likelihoodgivenby: LL= N ∑ n=1 log ( M ∑ k=1 ˆkp(Yn|Yˆk, σˆk) ) (32) andDF is thenumberofdegreesof freedomof thestatisticalmodel: DF=M×m(m+1) 2 +M+M−1 (33) InFormula (32), (ˆk,Yˆk, σˆk)1≤k≤M areobtained fromanEMalgorithmas stated inSection4.1 assumingtheexactdimension isM. Finally,note that inFormula (33),M×m(m+1)2 (respectivelyM andM−1)corresponds to thenumberofdegreesof freedomassociatedwith (Yˆk)1≤k≤M (respectively (σˆk)1≤k≤M and (ˆk)1≤k≤M). 5.ApplicationtoClassificationofDataonPm Recently,severalapproacheshaveusedtheRiemanniandistanceingeneralasthemaininnovation in imageorsignalclassificationproblems[2,15,34]. It turnsout that theuseof thisdistance leads to moreaccurate results (incomparison, forexample,with theEuclideandistance). This sectionproposes anapplicationthat followsasimilarapproach,but inadditionto theRiemanniandistance, italsorelies onastatisticalapproach. It considers theapplicationof theRiemannianLaplacedistribution(RLD) to theclassificationofdata inPm andgivesanoriginalLaplaceclassificationrule,whichcanbeusedto carryout thetaskofclassification,eveninthepresenceofoutliers. Italsoapplies thisclassificationrule to theproblemof textureclassification incomputervision, showingthat it leads to improvedresults in comparisonwithrecent literature. Section5.1considers, fromthepointofviewofstatistical learning, theclassificationofdatawith values inPm. GivendatapointsY1, · · · ,YN∈Pm, thisproceeds in twosteps, calledthe learningphase andtheclassificationphase, respectively. The learningphaseuncovers theclassstructureof thedata, byestimatingamixturemodelusing theEMalgorithmdeveloped inSection4.1. Once training is accomplished,datapointsaresubdividedintodisjointclasses.Classificationconsistsofassociating eachnewdatapoint to themostsuitableclass. For this,anewclassificationrulewillbeestablished andshowntobeoptimal. Section5.2 is the implementationof theLaplaceclassificationrule togetherwith theBICcriterion to textureclassification incomputervision. Ithighlights theadvantageof theLaplacedistribution in thepresenceofoutliersandshowsitsbetterperformancecomparedtorecentapproaches. 5.1. ClassificationUsingMixturesofLaplaceDistributions Assumetobegivenasetof trainingdataY1, · · · ,YN. Thesearenowmodeledasarealizationofa mixtureofLaplacedistributions: p(Y)= M ∑ μ=1 μ×p(Y|Y¯μ,σμ) (34) 375
back to the  book Differential Geometrical Theory of Statistics"
Differential Geometrical Theory of Statistics
Title
Differential Geometrical Theory of Statistics
Authors
Frédéric Barbaresco
Frank Nielsen
Editor
MDPI
Location
Basel
Date
2017
Language
English
License
CC BY-NC-ND 4.0
ISBN
978-3-03842-425-3
Size
17.0 x 24.4 cm
Pages
476
Keywords
Entropy, Coding Theory, Maximum entropy, Information geometry, Computational Information Geometry, Hessian Geometry, Divergence Geometry, Information topology, Cohomology, Shape Space, Statistical physics, Thermodynamics
Categories
Naturwissenschaften Physik
Web-Books
Library
Privacy
Imprint
Austria-Forum
Austria-Forum
Web-Books
Differential Geometrical Theory of Statistics