Probabilistic planning problems are typically modeled as a Markov Decision Process (MDP). MDPs, while an otherwise expressive model, allow only for sequential, non-durative action...
In this work we present an approach to solving time-critical decision-making problems by taking advantage of domain structure to expand the amountof time available for processing ...
This paper investigates the impact of symbolic search for solving domain-independent action planning problems with binary decision diagrams (BDDs). Polynomial upper and exponential...
Markov Decision Processes (MDP) have been widely used as a framework for planning under uncertainty. They allow to compute optimal sequences of actions in order to achieve a given...
Partially Observable Markov Decision Processes (POMDPs) provide a general framework for AI planning, but they lack the structure for representing real world planning problems in a...