Misplaced Pages

LogitBoost

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Boosting algorithm

In machine learning and computational learning theory, LogitBoost is a boosting algorithm formulated by Jerome Friedman, Trevor Hastie, and Robert Tibshirani.

The original paper casts the AdaBoost algorithm into a statistical framework. Specifically, if one considers AdaBoost as a generalized additive model and then applies the cost function of logistic regression, one can derive the LogitBoost algorithm.

Minimizing the LogitBoost cost function

LogitBoost can be seen as a convex optimization. Specifically, given that we seek an additive model of the form

f = t α t h t {\displaystyle f=\sum _{t}\alpha _{t}h_{t}}

the LogitBoost algorithm minimizes the logistic loss:

i log ( 1 + e y i f ( x i ) ) {\displaystyle \sum _{i}\log \left(1+e^{-y_{i}f(x_{i})}\right)}

See also

References

  1. Friedman, Jerome; Hastie, Trevor; Tibshirani, Robert (2000). "Additive logistic regression: a statistical view of boosting". Annals of Statistics. 28 (2): 337–407. CiteSeerX 10.1.1.51.9525. doi:10.1214/aos/1016218223.
  2. "Machine Learning Algorithms for Beginners". 22 September 2023. Retrieved 2023-10-01.
Stub icon

This artificial intelligence-related article is a stub. You can help Misplaced Pages by expanding it.

Categories: