Sciweavers

180 search results - page 9 / 36
» Improved bounds on the sample complexity of learning
Sort
View
DCG
2008
104views more  DCG 2008»
13 years 7 months ago
Finding the Homology of Submanifolds with High Confidence from Random Samples
Recently there has been a lot of interest in geometrically motivated approaches to data analysis in high dimensional spaces. We consider the case where data is drawn from sampling...
Partha Niyogi, Stephen Smale, Shmuel Weinberger
ICML
2007
IEEE
14 years 8 months ago
Asymptotic Bayesian generalization error when training and test distributions are different
In supervised learning, we commonly assume that training and test data are sampled from the same distribution. However, this assumption can be violated in practice and then standa...
Keisuke Yamazaki, Klaus-Robert Müller, Masash...
ML
2002
ACM
145views Machine Learning» more  ML 2002»
13 years 7 months ago
Boosting Methods for Regression
In this paper we examine ensemble methods for regression that leverage or "boost" base regressors by iteratively calling them on modified samples. The most successful lev...
Nigel Duffy, David P. Helmbold
TNN
2010
143views Management» more  TNN 2010»
13 years 2 months ago
Using unsupervised analysis to constrain generalization bounds for support vector classifiers
Abstract--A crucial issue in designing learning machines is to select the correct model parameters. When the number of available samples is small, theoretical sample-based generali...
Sergio Decherchi, Sandro Ridella, Rodolfo Zunino, ...
NIPS
2004
13 years 9 months ago
Efficient Kernel Machines Using the Improved Fast Gauss Transform
The computation and memory required for kernel machines with N training samples is at least O(N2 ). Such a complexity is significant even for moderate size problems and is prohibi...
Changjiang Yang, Ramani Duraiswami, Larry S. Davis