Sciweavers

ALT
2010
Springer

Approximation Stability and Boosting

14 years 2 months ago
Approximation Stability and Boosting
Stability has been explored to study the performance of learning algorithms in recent years and it has been shown that stability is sufficient for generalization and is sufficient and necessary for consistency of ERM in the general learning setting. Previous studies showed that AdaBoost has almost-everywhere uniform stability if the base learner has L1 stability. The L1 stability, however, is too restrictive and we show that AdaBoost becomes constant learner if the base learner is not real-valued learner. Considering that AdaBoost is mostly successful as a classification algorithm, stability analysis for AdaBoost when the base learner is not real-valued learner is an important yet unsolved problem. In this paper, we introduce the approximation stability and prove that approximation stability is sufficient for generalization, and sufficient and necessary for learnability of AERM in the general learning setting. We prove that AdaBoost has approximation stability and thus has good general...
Wei Gao, Zhi-Hua Zhou
Added 26 Oct 2010
Updated 26 Oct 2010
Type Conference
Year 2010
Where ALT
Authors Wei Gao, Zhi-Hua Zhou
Comments (0)