Sciweavers

1106 search results - page 168 / 222
» On regularization algorithms in learning theory
Sort
View
ICML
2006
IEEE
16 years 5 months ago
How boosting the margin can also boost classifier complexity
Boosting methods are known not to usually overfit training data even as the size of the generated classifiers becomes large. Schapire et al. attempted to explain this phenomenon i...
Lev Reyzin, Robert E. Schapire
ATAL
2010
Springer
15 years 5 months ago
Cultivating desired behaviour: policy teaching via environment-dynamics tweaks
In this paper we study, for the first time explicitly, the implications of endowing an interested party (i.e. a teacher) with the ability to modify the underlying dynamics of the ...
Zinovi Rabinovich, Lachlan Dufton, Kate Larson, Ni...
JAIR
2006
110views more  JAIR 2006»
15 years 4 months ago
Domain Adaptation for Statistical Classifiers
The most basic assumption used in statistical learning theory is that training data and test data are drawn from the same underlying distribution. Unfortunately, in many applicati...
Hal Daumé III, Daniel Marcu
ALT
2008
Springer
16 years 1 months ago
On-Line Probability, Complexity and Randomness
Abstract. Classical probability theory considers probability distributions that assign probabilities to all events (at least in the finite case). However, there are natural situat...
Alexey V. Chernov, Alexander Shen, Nikolai K. Vere...
COLT
1999
Springer
15 years 8 months ago
Regret Bounds for Prediction Problems
We present a unified framework for reasoning about worst-case regret bounds for learning algorithms. This framework is based on the theory of duality of convex functions. It brin...
Geoffrey J. Gordon