Difference between revisions of "Gradient Boosted Decision Tree"
From Cohen Courses
Jump to navigationJump to search (Created page with 'GBDT is an additive regression algorithm consisting of an ensemble of trees, fitted to current residuals, gradients of the loss function, in a forward step-wise manner. It iterat…') |
|||
Line 13: | Line 13: | ||
The optimal weights of trees <math>\beta_{t}</math> are determined by | The optimal weights of trees <math>\beta_{t}</math> are determined by | ||
− | <math> | + | <math>\beta_{t}=\operatorname{argmin}_{\beta}\overset{N}{\underset{i}{\sum}}L(y_{i},f_{t-1}(x_{i})+\beta T(x_{i},\theta))</math> |
(Source: [Dong et al WWW 2010]) | (Source: [Dong et al WWW 2010]) |
Revision as of 12:00, 29 March 2011
GBDT is an additive regression algorithm consisting of an ensemble of trees, fitted to current residuals, gradients of the loss function, in a forward step-wise manner. It iteratively fits an additive model as
such that a certain loss function is minimized, where is a tree at iteration , weighted by parameter , with a finite number of parameters, and is the learning rate. At iteration , tree is induced to fit the negative gradient by least squares. That is
where is the gradient over current prediction function
The optimal weights of trees are determined by
(Source: [Dong et al WWW 2010])