Boosting
Posted: 2017-02-08 , Modified: 2017-02-08
Tags: Boosting
Posted: 2017-02-08 , Modified: 2017-02-08
Tags: Boosting
Ref:
Generalization of linear predictors to address 2 issues
AdaBoost relies on the family of hypothesis classes obtained by composing a linear predictor on top of simple classes.
\(C\) is strongly PAC-learnable by \(A\) if:
\(C\) is weakly PAC-learnable by \(A\) if:
[Q: what is an explicit example of a provable weak learner?]
Q: Is weak learning equivalent to strong learning?
A: Yes if you increase the hypothesis class.
Given \((x_i,y_i)\) with \(y_i\in \{-1,+1\}\) and access to a weak learner \(A\), find \(H\) such \(\Pj(err_D(H)\le \ep)\ge 1-\de\) (learn strongly).
Idea: produce different distributions \(D\) from \(\mathcal D\). Pick distributions at each round that provide info about points that are hard to learn.
Q: Can we unify this with multiplicative weights? This seems like some kind of dual. (Check multiplicative weights paper?)