Sciweavers

495 search results - page 16 / 99
» Constructing States for Reinforcement Learning
Sort
View
ICML
2008
IEEE
14 years 8 months ago
Automatic discovery and transfer of MAXQ hierarchies
We present an algorithm, HI-MAT (Hierarchy Induction via Models And Trajectories), that discovers MAXQ task hierarchies by applying dynamic Bayesian network models to a successful...
Neville Mehta, Soumya Ray, Prasad Tadepalli, Thoma...
ROBOCUP
2007
Springer
167views Robotics» more  ROBOCUP 2007»
14 years 1 months ago
Cooperative/Competitive Behavior Acquisition Based on State Value Estimation of Others
The existing reinforcement learning approaches have been suffering from the curse of dimension problem when they are applied to multiagent dynamic environments. One of the typical...
Kentarou Noma, Yasutake Takahashi, Minoru Asada
ATAL
2004
Springer
14 years 1 months ago
From Global Selective Perception to Local Selective Perception
This paper presents a reinforcement learning algorithm used to allocate tasks to agents in an uncertain real-time environment. In such environment, tasks have to be analyzed and a...
Sébastien Paquet, Nicolas Bernier, Brahim C...
ECML
2003
Springer
14 years 29 days ago
Could Active Perception Aid Navigation of Partially Observable Grid Worlds?
Due to the unavoidable fact that a robot’s sensors will be limited in some manner, it is entirely possible that it can find itself unable to distinguish between differing state...
Paul A. Crook, Gillian Hayes
IWANN
1999
Springer
13 years 12 months ago
Using Temporal Neighborhoods to Adapt Function Approximators in Reinforcement Learning
To avoid the curse of dimensionality, function approximators are used in reinforcement learning to learn value functions for individual states. In order to make better use of comp...
R. Matthew Kretchmar, Charles W. Anderson