Difference between revisions of "Gradient Boosted Decision Tree"

From Cohen Courses
Jump to navigationJump to search
(Created page with 'GBDT is an additive regression algorithm consisting of an ensemble of trees, fitted to current residuals, gradients of the loss function, in a forward step-wise manner. It iterat…')
 
Line 13: Line 13:
 
The optimal weights of trees <math>\beta_{t}</math> are determined by
 
The optimal weights of trees <math>\beta_{t}</math> are determined by
  
<math><\beta_{t}=\operatorname{argmin}_{\beta}\overset{N}{\underset{i}{\sum}}L(y_{i},f_{t-1}(x_{i})+\beta T(x_{i},\theta))/math>
+
<math>\beta_{t}=\operatorname{argmin}_{\beta}\overset{N}{\underset{i}{\sum}}L(y_{i},f_{t-1}(x_{i})+\beta T(x_{i},\theta))</math>
  
 
(Source: [Dong et al WWW 2010])
 
(Source: [Dong et al WWW 2010])

Revision as of 12:00, 29 March 2011

GBDT is an additive regression algorithm consisting of an ensemble of trees, fitted to current residuals, gradients of the loss function, in a forward step-wise manner. It iteratively fits an additive model as

such that a certain loss function is minimized, where is a tree at iteration , weighted by parameter , with a finite number of parameters, and is the learning rate. At iteration , tree is induced to fit the negative gradient by least squares. That is

where is the gradient over current prediction function

The optimal weights of trees are determined by

(Source: [Dong et al WWW 2010])