Boosting

Posted: 2017-02-08 , Modified: 2017-02-08

Tags: Boosting

Ref:

Generalization of linear predictors to address 2 issues

AdaBoost relies on the family of hypothesis classes obtained by composing a linear predictor on top of simple classes.

Weak learnability

\(C\) is strongly PAC-learnable by \(A\) if:

\(C\) is weakly PAC-learnable by \(A\) if:

[Q: what is an explicit example of a provable weak learner?]

Q: Is weak learning equivalent to strong learning?

A: Yes if you increase the hypothesis class.

Boosting problem

Given \((x_i,y_i)\) with \(y_i\in \{-1,+1\}\) and access to a weak learner \(A\), find \(H\) such \(\Pj(err_D(H)\le \ep)\ge 1-\de\) (learn strongly).

AdaBoost

Idea: produce different distributions \(D\) from \(\mathcal D\). Pick distributions at each round that provide info about points that are hard to learn.

Q: Can we unify this with multiplicative weights? This seems like some kind of dual. (Check multiplicative weights paper?)