We study observation-based strategies for partially-observable Markov decision processes (POMDPs) with parity objectives. An observationbased strategy relies on partial information...
Krishnendu Chatterjee, Laurent Doyen, Thomas A. He...
Many stochastic planning problems can be represented using Markov Decision Processes (MDPs). A difficulty with using these MDP representations is that the common algorithms for so...
We study the computational complexity of some central analysis problems for One-Counter Markov Decision Processes (OC-MDPs), a class of finitely-presented, countable-state MDPs. O...
Tomas Brazdil, Vaclav Brozek, Kousha Etessami, Ant...
We study and provide efficient algorithms for multi-objective model checking problems for Markov Decision Processes (MDPs). Given an MDP, M, and given multiple linear-time (ω-regu...
Kousha Etessami, Marta Z. Kwiatkowska, Moshe Y. Va...
The majority of the work in the area of Markov decision processes has focused on expected values of rewards in the objective function and expected costs in the constraints. Althou...