Sciweavers

683 search results - page 121 / 137
» Coarticulation in Markov Decision Processes
Sort
View
FLAIRS
2004
13 years 9 months ago
State Space Reduction For Hierarchical Reinforcement Learning
er provides new techniques for abstracting the state space of a Markov Decision Process (MDP). These techniques extend one of the recent minimization models, known as -reduction, ...
Mehran Asadi, Manfred Huber
NIPS
2001
13 years 9 months ago
Variance Reduction Techniques for Gradient Estimates in Reinforcement Learning
Policy gradient methods for reinforcement learning avoid some of the undesirable properties of the value function approaches, such as policy degradation (Baxter and Bartlett, 2001...
Evan Greensmith, Peter L. Bartlett, Jonathan Baxte...
NIPS
2001
13 years 9 months ago
Multiagent Planning with Factored MDPs
We present a principled and efficient planning algorithm for cooperative multiagent dynamic systems. A striking feature of our method is that the coordination and communication be...
Carlos Guestrin, Daphne Koller, Ronald Parr
IJCAI
2003
13 years 9 months ago
Approximating Optimal Policies for Agents with Limited Execution Resources
An agent with limited consumable execution resources needs policies that attempt to achieve good performance while respecting these limitations. Otherwise, an agent (such as a pla...
Dmitri A. Dolgov, Edmund H. Durfee
IJCAI
2003
13 years 9 months ago
Generalizing Plans to New Environments in Relational MDPs
A longstanding goal in planning research is the ability to generalize plans developed for some set of environments to a new but similar environment, with minimal or no replanning....
Carlos Guestrin, Daphne Koller, Chris Gearhart, Ne...