Sciweavers

COLT
2005
Springer

Leaving the Span

14 years 5 months ago
Leaving the Span
We discuss a simple sparse linear problem that is hard to learn with any algorithm that uses a linear combination of the training instances as its weight vector. The hardness holds even if we allow the learner to embed the instances into any higher dimensional feature space (and use a kernel function to define the dot product between the embedded instances). These algorithms are inherently limited by the fact that after seeing k instances only a weight space of dimension k can be spanned. Our hardness result is surprising because the same problem can be efficiently learned using the exponentiated gradient (EG) algorithm: Now the component-wise logarithms of the weights are essentially a linear combination of the training instances and after seeing k instances the space of possible weight vectors can contain up to 2k unit vectors. The EG algorithm enforces additional constraints on the weights (all must be non-negative and sum to one) and in some cases these constraints alone allow th...
Manfred K. Warmuth, S. V. N. Vishwanathan
Added 26 Jun 2010
Updated 26 Jun 2010
Type Conference
Year 2005
Where COLT
Authors Manfred K. Warmuth, S. V. N. Vishwanathan
Comments (0)