Sciweavers

802 search results - page 67 / 161
» Experts in a Markov Decision Process
Sort
View
ICML
2008
IEEE
14 years 8 months ago
Reinforcement learning with limited reinforcement: using Bayes risk for active learning in POMDPs
Partially Observable Markov Decision Processes (POMDPs) have succeeded in planning domains that require balancing actions that increase an agent's knowledge and actions that ...
Finale Doshi, Joelle Pineau, Nicholas Roy
ICALP
2009
Springer
14 years 8 months ago
Reachability in Stochastic Timed Games
We define stochastic timed games, which extend two-player timed games with probabilities (following a recent approach by Baier et al), and which extend in a natural way continuous-...
Patricia Bouyer, Vojtech Forejt
ECML
2007
Springer
14 years 2 months ago
Safe Q-Learning on Complete History Spaces
In this article, we present an idea for solving deterministic partially observable markov decision processes (POMDPs) based on a history space containing sequences of past observat...
Stephan Timmer, Martin Riedmiller
GECCO
2004
Springer
142views Optimization» more  GECCO 2004»
14 years 1 months ago
Improving MACS Thanks to a Comparison with 2TBNs
Abstract. Factored Markov Decision Processes is the theoretical framework underlying multi-step Learning Classifier Systems research. This framework is mostly used in the context ...
Olivier Sigaud, Thierry Gourdin, Pierre-Henri Wuil...
ECSQARU
2001
Springer
14 years 9 days ago
Space-Progressive Value Iteration: An Anytime Algorithm for a Class of POMDPs
Abstract. Finding optimal policies for general partially observable Markov decision processes (POMDPs) is computationally difficult primarily due to the need to perform dynamic-pr...
Nevin Lianwen Zhang, Weihong Zhang