Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revisionNext revisionBoth sides next revision | ||
data_mining:xgboost [2019/05/03 00:52] – [XGBoost] phreazer | data_mining:xgboost [2019/05/03 01:10] – phreazer | ||
---|---|---|---|
Line 19: | Line 19: | ||
\hat{y}_i = \sum^K_{k=1} f_k(x_i), f_k \in F | \hat{y}_i = \sum^K_{k=1} f_k(x_i), f_k \in F | ||
$$ | $$ | ||
+ | |||
+ | ===== Gradient boosting ===== | ||
$F$ is space of functions containing all regression trees | $F$ is space of functions containing all regression trees | ||
Line 57: | Line 59: | ||
* Logistic loss $l(y_i, | * Logistic loss $l(y_i, | ||
- | Stochastic Gradient | + | Stochastic Gradient |
Solution is **additive training**: Start with constant prediction, add a new function each time. | Solution is **additive training**: Start with constant prediction, add a new function each time. | ||
Line 80: | Line 82: | ||
- | Taylor expansion | + | ==== Taylor expansion |
Use taylor expansion to approximate a function through a power series (polynom). | Use taylor expansion to approximate a function through a power series (polynom). |