LogitBoost

LogitBoost is a boosting algorithm formulated by Jerome Friedman, Trevor Hastie, and Robert Tibshirani. The original paper[1] casts the AdaBoost algorithm into a statistical framework. Specifically, if one considers AdaBoost as a generalized additive model and then applies the cost functional of logistic regression, one can derive the LogitBoost algorithm.

LogitBoost is a boosting algorithm formulated by Jerome FriedmanTrevor Hastie, and Robert Tibshirani. The original paper[1] casts the AdaBoost algorithm into a statistical framework. Specifically, if one considers AdaBoost as a generalized additive model and then applies the cost functional of logistic regression, one can derive the LogitBoost algorithm.

Minimizing the LogitBoost cost functional[edit]

LogitBoost can be seen as a convex optimization. Specifically, given that we seek an additive model of the form

f = \sum_t \alpha_t h_t

the LogitBoost algorithm minimizes the logistic loss:

\sum_i \log\left( 1 + e^{-y_i f(x_i)}\right)

References[edit]

  1. Jump up^ Jerome Friedman, Trevor Hastie and Robert Tibshirani. Additive logistic regression: a statistical view of boosting. Annals of Statistics 28(2), 2000. 337–407. http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.51.9525

See also[edit]

RELATED ARTICLESExplain
Machine Learning Methods & Algorithms
Supervised learning
Ensemble learning
Boosting (meta-algorithm)
LogitBoost
Randomized weighted majority algorithm
Weighted Majority Algorithm (WMA)
Graph of this discussion
Enter the title of your article


Enter a short (max 500 characters) summation of your article
Enter the main body of your article
Lock
+Comments (0)
+Citations (0)
+About
Enter comment

Select article text to quote
welcome text

First name   Last name 

Email

Skip