This contribution develops a theoretical framework that takes into account the effect of approximate optimization on learning algorithms. The analysis shows distinct tradeoffs for...
We combine three threads of research on approximate dynamic programming: sparse random sampling of states, value function and policy approximation using local models, and using lo...
Semi-supervised methods use unlabeled data in addition to labeled data to construct predictors. While existing semi-supervised methods have shown some promising empirical performa...
We present a new class of models for high-dimensional nonparametric regression and classification called sparse additive models (SpAM). Our methods combine ideas from sparse line...
Pradeep D. Ravikumar, Han Liu, John D. Lafferty, L...
To accelerate the training of kernel machines, we propose to map the input data to a randomized low-dimensional feature space and then apply existing fast linear methods. The feat...
Assessing similarity between features is a key step in object recognition and scene categorization tasks. We argue that knowledge on the distribution of distances generated by sim...
Gertjan J. Burghouts, Arnold W. M. Smeulders, Jan-...
Reward-modulated spike-timing-dependent plasticity (STDP) has recently emerged as a candidate for a learning rule that could explain how local learning rules at single synapses su...
Robert A. Legenstein, Dejan Pecevski, Wolfgang Maa...
In recent years, the language model Latent Dirichlet Allocation (LDA), which clusters co-occurring words into topics, has been widely applied in the computer vision field. Howeve...
Rosetta is one of the leading algorithms for protein structure prediction today. It is a Monte Carlo energy minimization method requiring many random restarts to find structures ...
Ben Blum, Michael I. Jordan, David Kim, Rhiju Das,...