Sciweavers

423 search results - page 43 / 85
» Multi-objective Model Checking of Markov Decision Processes
Sort
View
AIPS
2008
13 years 11 months ago
Multiagent Planning Under Uncertainty with Stochastic Communication Delays
We consider the problem of cooperative multiagent planning under uncertainty, formalized as a decentralized partially observable Markov decision process (Dec-POMDP). Unfortunately...
Matthijs T. J. Spaan, Frans A. Oliehoek, Nikos A. ...
JAIR
2008
107views more  JAIR 2008»
13 years 8 months ago
Planning with Durative Actions in Stochastic Domains
Probabilistic planning problems are typically modeled as a Markov Decision Process (MDP). MDPs, while an otherwise expressive model, allow only for sequential, non-durative action...
Mausam, Daniel S. Weld
ECAI
1998
Springer
14 years 27 days ago
Optimal Scheduling of Dynamic Progressive Processing
Progressive processing allows a system to satisfy a set of requests under time pressure by limiting the amount of processing allocated to each task based on a predefined hierarchic...
Abdel-Illah Mouaddib, Shlomo Zilberstein
STACS
2012
Springer
12 years 4 months ago
Stabilization of Branching Queueing Networks
Queueing networks are gaining attraction for the performance analysis of parallel computer systems. A Jackson network is a set of interconnected servers, where the completion of a...
Tomás Brázdil, Stefan Kiefer
IAT
2005
IEEE
14 years 2 months ago
Decomposing Large-Scale POMDP Via Belief State Analysis
Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing ...
Xin Li, William K. Cheung, Jiming Liu