Sciweavers

423 search results - page 43 / 85
» Multi-objective Model Checking of Markov Decision Processes
Sort
View
123
Voted
AIPS
2008
15 years 5 months ago
Multiagent Planning Under Uncertainty with Stochastic Communication Delays
We consider the problem of cooperative multiagent planning under uncertainty, formalized as a decentralized partially observable Markov decision process (Dec-POMDP). Unfortunately...
Matthijs T. J. Spaan, Frans A. Oliehoek, Nikos A. ...
JAIR
2008
107views more  JAIR 2008»
15 years 3 months ago
Planning with Durative Actions in Stochastic Domains
Probabilistic planning problems are typically modeled as a Markov Decision Process (MDP). MDPs, while an otherwise expressive model, allow only for sequential, non-durative action...
Mausam, Daniel S. Weld
136
Voted
ECAI
1998
Springer
15 years 7 months ago
Optimal Scheduling of Dynamic Progressive Processing
Progressive processing allows a system to satisfy a set of requests under time pressure by limiting the amount of processing allocated to each task based on a predefined hierarchic...
Abdel-Illah Mouaddib, Shlomo Zilberstein
STACS
2012
Springer
13 years 11 months ago
Stabilization of Branching Queueing Networks
Queueing networks are gaining attraction for the performance analysis of parallel computer systems. A Jackson network is a set of interconnected servers, where the completion of a...
Tomás Brázdil, Stefan Kiefer
IAT
2005
IEEE
15 years 9 months ago
Decomposing Large-Scale POMDP Via Belief State Analysis
Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing ...
Xin Li, William K. Cheung, Jiming Liu