Sciweavers

101 search results - page 3 / 21
» Convergence of Gradient Dynamics with a Variable Learning Ra...
Sort
View
NIPS
1993
13 years 8 months ago
Optimal Stochastic Search and Adaptive Momentum
Stochastic optimization algorithms typically use learning rate schedules that behave asymptotically as (t) = 0=t. The ensemble dynamics (Leen and Moody, 1993) for such algorithms ...
Todd K. Leen, Genevieve B. Orr
AI
2002
Springer
13 years 6 months ago
Multiagent learning using a variable learning rate
Learning to act in a multiagent environment is a difficult problem since the normal definition of an optimal policy no longer applies. The optimal policy at any moment depends on ...
Michael H. Bowling, Manuela M. Veloso
COLT
2000
Springer
13 years 11 months ago
Estimation and Approximation Bounds for Gradient-Based Reinforcement Learning
We model reinforcement learning as the problem of learning to control a Partially Observable Markov Decision Process (  ¢¡¤£¦¥§  ), and focus on gradient ascent approache...
Peter L. Bartlett, Jonathan Baxter
MOC
2002
77views more  MOC 2002»
13 years 6 months ago
Directional Newton methods in n variables
Directional Newton methods for functions f of n variables are shown to converge, under standard assumptions, to a solution of f(x) = 0. The rate of convergence is quadratic, for ne...
Yuri Levin, Adi Ben-Israel
JMLR
2006
97views more  JMLR 2006»
13 years 7 months ago
Learning Coordinate Covariances via Gradients
We introduce an algorithm that learns gradients from samples in the supervised learning framework. An error analysis is given for the convergence of the gradient estimated by the ...
Sayan Mukherjee, Ding-Xuan Zhou