For a given learning scenario, are there any thumb rules to figure out whether to use bagging or boosting? I understand that one reduces variance and the other bias, which might be a preliminary criteria to guide the choice of the particular ensemble method to be used. However, I was wondering whether people have come across other empirical factors which might lead to favoring one over the other !!

asked Sep 13 '10 at 14:58

IdleBrain's gravatar image

IdleBrain
91337


4 Answers:

Usually what you want is boosting, I think, that generally improves the performance of your classifier (by adding one layer of depth to the model). Bagging seems to be more useful when you want to derive confidence intervals and/or probability estimates from classifiers that don't naturally give you these informations, not necessarily to improve performance as such.

answered Sep 13 '10 at 15:42

Alexandre%20Passos's gravatar image

Alexandre Passos ♦
2554154278421

In general, boosting tends to learn faster than bagging as it concentrates more in the misclassified cases. Another difference is that the weak learners used in boosting can be weak classifiers (e.g., small decision trees) but for bagging the weak learners actually need to be reasonably strong (e.g., large decision trees). Boosting however is a bit more sensitive to errors in training data (classification noise). For some early experimental comparison, see this paper. But then there exists more robust variants of boosting that are less sensitive to classification noise.

answered Sep 13 '10 at 16:11

spinxl39's gravatar image

spinxl39
3698114869

You provided excellent links. Thank you.

(Sep 15 '10 at 17:01) Lucian Sasu

I lean more towards bagging than boosting myself. A key issue for me is how noisy is the data I'm working with. Usually, the noise is non-trivial and bagging is a safer choice than vanilla boosting. (I don't personally have experience w/ the robust boosting variants.)

My advice is to try both if it is not too hard. Just be sure to a) use enough base models to reach the asymptotic performance, and b) use early stopping on validation data to prevent boosting from overfitting.

Also, I know the conventional wisdom is to use weak models with boosting, but I've seen empirical evidence suggesting that you can do better with more powerful models (e.g., big decision trees).

answered Sep 15 '10 at 13:15

Art%20Munson's gravatar image

Art Munson
64611316

1

On a more technical level, both bagging and boosting can reduce both bias and variance. Performance improvements from bagging are mostly from variance reduction, whereas improvements from boosting are more from bias reduction than variance reduction. I recommend Bauer & Kohavi's paper "An empirical comparison of voting classification algorithms: bagging, boosting, and variants" for those interested.

(Sep 15 '10 at 13:24) Art Munson
-1

When in doubt, measure.

If you don't have a theoretical basis for choosing one other the other, compare them using some test data and use the one that scored better.

answered Sep 16 '10 at 00:04

Robert%20Layton's gravatar image

Robert Layton
1625122637

Your answer
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.