Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
data_mining:xgboost [2019/05/02 23:10] – phreazer | data_mining:xgboost [2020/08/02 14:12] (current) – phreazer | ||
---|---|---|---|
Line 22: | Line 22: | ||
===== Gradient boosting ===== | ===== Gradient boosting ===== | ||
- | $F$ is space of functions containing all regression trees | + | * $F$ is space of functions containing all regression trees |
- | $K$ is number of trees | + | |
- | $f_k(x_i)$ is regression tree that maps a attribute to a score | + | |
Learn functions (trees) instead of weights in $R^d$. | Learn functions (trees) instead of weights in $R^d$. | ||
Line 35: | Line 35: | ||
Learning objective: | Learning objective: | ||
- | * Training loss: Fit of the functions to the points | + | |
- | * Regularization: | + | |
Objective: | Objective: | ||
Line 95: | Line 95: | ||
$$\sum^n_{i=1} [l(y_i, | $$\sum^n_{i=1} [l(y_i, | ||
- | With removed constants | + | With removed constants |
$$\sum^n_{i=1} [g_if_t(x_i) + \frac{1}{2}h_if_t^2(x_i)] + \Omega(f_t)$$ | $$\sum^n_{i=1} [g_if_t(x_i) + \frac{1}{2}h_if_t^2(x_i)] + \Omega(f_t)$$ | ||
So that learning function only influences $g_i$ and $h_i$ while rest stays the same. | So that learning function only influences $g_i$ and $h_i$ while rest stays the same. |