Sciweavers

1138 search results - page 170 / 228
» Feature Markov Decision Processes
Sort
View
NIPS
2007
15 years 5 months ago
What makes some POMDP problems easy to approximate?
Point-based algorithms have been surprisingly successful in computing approximately optimal solutions for partially observable Markov decision processes (POMDPs) in high dimension...
David Hsu, Wee Sun Lee, Nan Rong
128
Voted
AAAI
2006
15 years 4 months ago
Incremental Least Squares Policy Iteration for POMDPs
We present a new algorithm, called incremental least squares policy iteration (ILSPI), for finding the infinite-horizon stationary policy for partially observable Markov decision ...
Hui Li, Xuejun Liao, Lawrence Carin
137
Voted
FLAIRS
2004
15 years 4 months ago
State Space Reduction For Hierarchical Reinforcement Learning
er provides new techniques for abstracting the state space of a Markov Decision Process (MDP). These techniques extend one of the recent minimization models, known as -reduction, ...
Mehran Asadi, Manfred Huber
124
Voted
NIPS
2001
15 years 4 months ago
Variance Reduction Techniques for Gradient Estimates in Reinforcement Learning
Policy gradient methods for reinforcement learning avoid some of the undesirable properties of the value function approaches, such as policy degradation (Baxter and Bartlett, 2001...
Evan Greensmith, Peter L. Bartlett, Jonathan Baxte...
129
Voted
IJCAI
2003
15 years 4 months ago
Approximating Optimal Policies for Agents with Limited Execution Resources
An agent with limited consumable execution resources needs policies that attempt to achieve good performance while respecting these limitations. Otherwise, an agent (such as a pla...
Dmitri A. Dolgov, Edmund H. Durfee