Sciweavers

1340 search results - page 6 / 268
» Kalman Temporal Differences
Sort
View
ICML
2006
IEEE
14 years 8 months ago
Relational temporal difference learning
We introduce relational temporal difference learning as an effective approach to solving multi-agent Markov decision problems with large state spaces. Our algorithm uses temporal ...
Nima Asgharbeygi, David J. Stracuzzi, Pat Langley
CEC
2010
IEEE
13 years 7 months ago
Coevolutionary Temporal Difference Learning for small-board Go
—In this paper we apply Coevolutionary Temporal Difference Learning (CTDL), a hybrid of coevolutionary search and reinforcement learning proposed in our former study, to evolve s...
Krzysztof Krawiec, Marcin Szubert
GECCO
2009
Springer
142views Optimization» more  GECCO 2009»
14 years 1 months ago
A stopping criterion based on Kalman estimation techniques with several progress indicators
The need for a stopping criterion in MOEA’s is a repeatedly mentioned matter in the domain of MOOP’s, even though it is usually left aside as secondary, while stopping criteri...
José Luis Guerrero, Jesús Garc&iacut...
NIPS
2008
13 years 8 months ago
Temporal Difference Based Actor Critic Learning - Convergence and Neural Implementation
Actor-critic algorithms for reinforcement learning are achieving renewed popularity due to their good convergence properties in situations where other approaches often fail (e.g.,...
Dotan Di Castro, Dmitry Volkinshtein, Ron Meir
ICML
2009
IEEE
14 years 8 months ago
Regularization and feature selection in least-squares temporal difference learning
We consider the task of reinforcement learning with linear value function approximation. Temporal difference algorithms, and in particular the Least-Squares Temporal Difference (L...
J. Zico Kolter, Andrew Y. Ng