Web-Books
in the Austria-Forum
Austria-Forum
Web-Books
Informatik
Short-Term Load Forecasting by Artificial Intelligent Technologies
Page - 163 -
  • User
  • Version
    • full version
    • text only version
  • Language
    • Deutsch - German
    • English

Page - 163 - in Short-Term Load Forecasting by Artificial Intelligent Technologies

Image of the Page - 163 -

Image of the Page - 163 - in Short-Term Load Forecasting by Artificial Intelligent Technologies

Text of the Page - 163 -

Energies2018,11, 2038 whereT is thenumberof leaves in the treewith leafweightsw = (w1,w2, . . . ,wT).Using thesecond orderTaylorexpansion, (3) canbesimplifiedto: L˜(t) = n ∑ i= 1 [ gi ft(xi)+ 1 2 hi f2t (xi) ] +γT+ 1 2 λ T ∑ j= 1 w2j (4) wheregi = ∂yˆ(t−1) l (yi, yˆ (t−1))andhi = ∂2yˆ(t−1) l (yi, yˆ (t−1)). Denotingby Ij = {i|q(xi) = j} the instancesetof leaf j,wecanrewrite (4), as follows: L˜(t) = T ∑ j= 1 ⎡⎣⎛⎝∑ i∈Ij gi ⎞⎠wj+ 12 ⎛⎝∑ i∈Ij hi+λ ⎞⎠w2j ⎤⎦+γT (5) Therefore, theoptimalweight isgivenby: w∗j = − ∑i∈Ij gi ∑i∈Ij hi+λ (6) andthecorrespondingoptimalobjectiveby: L˜(t)(q) = −1 2 T ∑ j= 1 ( ∑i∈Ij gi )2 ∑i∈Ij hi+λ + γT (7) whereq represents theoptimal treestructurewithT leavesandleafweightsw∗ = ( w∗1,w ∗ 2, . . . ,w ∗ T ) . Duetothe impossibilityofenumeratingall thepossible treestructuresq, agreedyalgorithmisused (itstartswithasingle leafandaddsbranches iteratively).Denotingby ILand IR the instancesetsof left andrightnodesafter thesplit, I = IL∪ IR, thereductionintheobjectiveafter thesplit isgivenby: Lsplit = 12 ⎡⎢⎣ ( ∑i∈IL gi )2 ∑i∈IL hi+λ + ( ∑i∈IR gi )2 ∑i∈IR hi+λ − (∑i∈I gi) 2 ∑i∈I hi+λ ⎤⎥⎦−γ (8) Thetaskofsearchingthebestsplithasbeendevelopedintwoscenarios: anexactgreedyalgorithm (itenumeratesall thepossiblesplitsonall the features,which iscomputationaldemanding)andan approximategreedyalgorithmforbigdatasets, see [37] formoredetails. Themain difference between random forest and boosting is that the former builds the base learners independently throughbootstrapsamplingonthetrainingdataset,while the latterobtains themsequentially focusingontheerrorsof theprevious iterationandusinggradientdescentmethods. Somestrengthsof theXGBoost implementationcomparingtoothermethodsare: • Anexactgreedyalgorithmisavailable. • Approximateglobalandapproximate localalgorithmsareavailable forbigdatasets. • Itperformsparallel learning. Besides, aneffectivecache-awareblockstructure isavailable for out-of-core tree learning. • It isefficient incaseofsparse inputdata (includingthepresenceofmissingvalues). The extremegradient boostingmethod (XGBoost) has been implemented bymeans of theR package“xgboost”, see [38]. Apart fromitshighlycomputationalefficiency, theXGBoostoffersagreatflexibility,but it requires settingupmore thanthe tenparameters thatcouldnotbe learnedfromthedata. Taking intoaccount thatRpackage“xgboost”doesnothaveanyhyperparameter tuning, theparameter tuningcanbe 163
back to the  book Short-Term Load Forecasting by Artificial Intelligent Technologies"
Short-Term Load Forecasting by Artificial Intelligent Technologies
Title
Short-Term Load Forecasting by Artificial Intelligent Technologies
Authors
Wei-Chiang Hong
Ming-Wei Li
Guo-Feng Fan
Editor
MDPI
Location
Basel
Date
2019
Language
English
License
CC BY 4.0
ISBN
978-3-03897-583-0
Size
17.0 x 24.4 cm
Pages
448
Keywords
Scheduling Problems in Logistics, Transport, Timetabling, Sports, Healthcare, Engineering, Energy Management
Category
Informatik
Web-Books
Library
Privacy
Imprint
Austria-Forum
Austria-Forum
Web-Books
Short-Term Load Forecasting by Artificial Intelligent Technologies