Sciweavers

43 search results - page 5 / 9
» Multiagent Planning with Factored MDPs
Sort
View
AAAI
2012
11 years 10 months ago
Planning in Factored Action Spaces with Symbolic Dynamic Programming
We consider symbolic dynamic programming (SDP) for solving Markov Decision Processes (MDP) with factored state and action spaces, where both states and actions are described by se...
Aswin Raghavan, Saket Joshi, Alan Fern, Prasad Tad...
ATAL
2003
Springer
14 years 27 days ago
Minimizing communication cost in a distributed Bayesian network using a decentralized MDP
In complex distributed applications, a problem is often decomposed into a set of subproblems that are distributed to multiple agents. We formulate this class of problems with a tw...
Jiaying Shen, Victor R. Lesser, Norman Carver
UAI
2000
13 years 9 months ago
Fast Planning in Stochastic Games
Stochastic games generalize Markov decision processes MDPs to a multiagent setting by allowing the state transitions to depend jointly on all player actions, and having rewards de...
Michael J. Kearns, Yishay Mansour, Satinder P. Sin...
ATAL
2009
Springer
14 years 2 months ago
Constraint-based dynamic programming for decentralized POMDPs with structured interactions
Decentralized partially observable MDPs (DEC-POMDPs) provide a rich framework for modeling decision making by a team of agents. Despite rapid progress in this area, the limited sc...
Akshat Kumar, Shlomo Zilberstein
AAAI
1997
13 years 9 months ago
Model Minimization in Markov Decision Processes
Many stochastic planning problems can be represented using Markov Decision Processes (MDPs). A difficulty with using these MDP representations is that the common algorithms for so...
Thomas Dean, Robert Givan