Sciweavers

1062 search results - page 63 / 213
» Sublinear Optimization for Machine Learning
Sort
View
ICML
2009
IEEE
14 years 9 months ago
Learning when to stop thinking and do something!
An anytime algorithm is capable of returning a response to the given task at essentially any time; typically the quality of the response improves as the time increases. Here, we c...
Barnabás Póczos, Csaba Szepesv&aacut...
ICML
2009
IEEE
14 years 9 months ago
Robust feature extraction via information theoretic learning
In this paper, we present a robust feature extraction framework based on informationtheoretic learning. Its formulated objective aims at simultaneously maximizing the Renyi's...
Xiaotong Yuan, Bao-Gang Hu
ECML
2005
Springer
14 years 2 months ago
Active Learning for Probability Estimation Using Jensen-Shannon Divergence
Active selection of good training examples is an important approach to reducing data-collection costs in machine learning; however, most existing methods focus on maximizing classi...
Prem Melville, Stewart M. Yang, Maytal Saar-Tsecha...
ICML
1994
IEEE
14 years 16 days ago
Markov Games as a Framework for Multi-Agent Reinforcement Learning
In the Markov decision process (MDP) formalization of reinforcement learning, a single adaptive agent interacts with an environment defined by a probabilistic transition function....
Michael L. Littman
NIPS
2008
13 years 10 months ago
Optimization on a Budget: A Reinforcement Learning Approach
Many popular optimization algorithms, like the Levenberg-Marquardt algorithm (LMA), use heuristic-based "controllers" that modulate the behavior of the optimizer during ...
Paul Ruvolo, Ian R. Fasel, Javier R. Movellan