Sciweavers

201 search results - page 8 / 41
» Solving Concurrent Markov Decision Processes
Sort
View
ICRA
2007
IEEE
154views Robotics» more  ICRA 2007»
14 years 1 months ago
Oracular Partially Observable Markov Decision Processes: A Very Special Case
— We introduce the Oracular Partially Observable Markov Decision Process (OPOMDP), a type of POMDP in which the world produces no observations; instead there is an “oracle,” ...
Nicholas Armstrong-Crews, Manuela M. Veloso
IJCAI
2007
13 years 8 months ago
Using Linear Programming for Bayesian Exploration in Markov Decision Processes
A key problem in reinforcement learning is finding a good balance between the need to explore the environment and the need to gain rewards by exploiting existing knowledge. Much ...
Pablo Samuel Castro, Doina Precup
UAI
1998
13 years 8 months ago
Structured Reachability Analysis for Markov Decision Processes
Recent research in decision theoretic planning has focussedon making the solution of Markov decision processes (MDPs) more feasible. We develop a family of algorithms for structur...
Craig Boutilier, Ronen I. Brafman, Christopher W. ...
ICTAI
2000
IEEE
13 years 11 months ago
Building efficient partial plans using Markov decision processes
Markov Decision Processes (MDP) have been widely used as a framework for planning under uncertainty. They allow to compute optimal sequences of actions in order to achieve a given...
Pierre Laroche
CDC
2009
IEEE
169views Control Systems» more  CDC 2009»
14 years 4 days ago
Parametric regret in uncertain Markov decision processes
— We consider decision making in a Markovian setup where the reward parameters are not known in advance. Our performance criterion is the gap between the performance of the best ...
Huan Xu, Shie Mannor