Sciweavers

499 search results - page 11 / 100
» Model Minimization in Markov Decision Processes
Sort
View
ATAL
2007
Springer
13 years 11 months ago
Modeling plan coordination in multiagent decision processes
In multiagent planning, it is often convenient to view a problem as two subproblems: agent local planning and coordination. Thus, we can classify agent activities into two categor...
Ping Xuan
ICMLA
2008
13 years 9 months ago
Prediction-Directed Compression of POMDPs
High dimensionality of belief space in Partially Observable Markov Decision Processes (POMDPs) is one of the major causes that severely restricts the applicability of this model. ...
Abdeslam Boularias, Masoumeh T. Izadi, Brahim Chai...
CORR
2008
Springer
91views Education» more  CORR 2008»
13 years 7 months ago
Significant Diagnostic Counterexamples in Probabilistic Model Checking
Abstract. This paper presents a novel technique for counterexample generation in probabilistic model checking of Markov chains and Markov Decision Processes. (Finite) paths in coun...
Miguel E. Andrés, Pedro R. D'Argenio, Peter...
NIPS
2004
13 years 8 months ago
A Cost-Shaping LP for Bellman Error Minimization with Performance Guarantees
We introduce a new algorithm based on linear programming that approximates the differential value function of an average-cost Markov decision process via a linear combination of p...
Daniela Pucci de Farias, Benjamin Van Roy
IROS
2009
IEEE
206views Robotics» more  IROS 2009»
14 years 2 months ago
Bayesian reinforcement learning in continuous POMDPs with gaussian processes
— Partially Observable Markov Decision Processes (POMDPs) provide a rich mathematical model to handle realworld sequential decision processes but require a known model to be solv...
Patrick Dallaire, Camille Besse, Stéphane R...