Sciweavers

22 search results - page 4 / 5
» Tighter Value Function Bounds for Bayesian Reinforcement Lea...
Sort
View
ICML
2001
IEEE
14 years 8 months ago
Off-Policy Temporal Difference Learning with Function Approximation
We introduce the first algorithm for off-policy temporal-difference learning that is stable with linear function approximation. Off-policy learning is of interest because it forms...
Doina Precup, Richard S. Sutton, Sanjoy Dasgupta
IJCAI
2007
13 years 9 months ago
Using Linear Programming for Bayesian Exploration in Markov Decision Processes
A key problem in reinforcement learning is finding a good balance between the need to explore the environment and the need to gain rewards by exploiting existing knowledge. Much ...
Pablo Samuel Castro, Doina Precup

Publication
222views
14 years 4 months ago
Algorithms and Bounds for Rollout Sampling Approximate Policy Iteration
Abstract: Several approximate policy iteration schemes without value functions, which focus on policy representation using classifiers and address policy learning as a supervis...
Christos Dimitrakakis, Michail G. Lagoudakis
ICML
2002
IEEE
14 years 8 months ago
Learning from Scarce Experience
Searching the space of policies directly for the optimal policy has been one popular method for solving partially observable reinforcement learning problems. Typically, with each ...
Leonid Peshkin, Christian R. Shelton
CORR
2000
Springer
92views Education» more  CORR 2000»
13 years 7 months ago
Predicting the expected behavior of agents that learn about agents: the CLRI framework
We describe a framework and equations used to model and predict the behavior of multi-agent systems (MASs) with learning agents. A difference equation is used for calculating the ...
José M. Vidal, Edmund H. Durfee