Sciweavers

141 search results - page 8 / 29
» Learning r-of-k Functions by Boosting
Sort
View
EUROCOLT
1995
Springer
13 years 10 months ago
A decision-theoretic generalization of on-line learning and an application to boosting
k. The model we study can be interpreted as a broad, abstract extension of the well-studied on-line prediction model to a general decision-theoretic setting. We show that the multi...
Yoav Freund, Robert E. Schapire
COLT
2008
Springer
13 years 8 months ago
On the Equivalence of Weak Learnability and Linear Separability: New Relaxations and Efficient Boosting Algorithms
Boosting algorithms build highly accurate prediction mechanisms from a collection of lowaccuracy predictors. To do so, they employ the notion of weak-learnability. The starting po...
Shai Shalev-Shwartz, Yoram Singer
ICML
2008
IEEE
14 years 7 months ago
Random classification noise defeats all convex potential boosters
A broad class of boosting algorithms can be interpreted as performing coordinate-wise gradient descent to minimize some potential function of the margins of a data set. This class...
Philip M. Long, Rocco A. Servedio
ICML
2007
IEEE
14 years 7 months ago
On learning with dissimilarity functions
We study the problem of learning a classification task in which only a dissimilarity function of the objects is accessible. That is, data are not represented by feature vectors bu...
Liwei Wang, Cheng Yang, Jufu Feng
ICML
2004
IEEE
14 years 7 months ago
Surrogate maximization/minimization algorithms for AdaBoost and the logistic regression model
Surrogate maximization (or minimization) (SM) algorithms are a family of algorithms that can be regarded as a generalization of expectation-maximization (EM) algorithms. There are...
Zhihua Zhang, James T. Kwok, Dit-Yan Yeung