Sciweavers

32 search results - page 5 / 7
» Optimal and Approximate Q-value Functions for Decentralized ...
Sort
View
IAT
2005
IEEE
14 years 1 months ago
Decomposing Large-Scale POMDP Via Belief State Analysis
Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing ...
Xin Li, William K. Cheung, Jiming Liu
HICSS
2003
IEEE
207views Biometrics» more  HICSS 2003»
14 years 28 days ago
Formalizing Multi-Agent POMDP's in the context of network routing
This paper uses partially observable Markov decision processes (POMDP’s) as a basic framework for MultiAgent planning. We distinguish three perspectives: first one is that of a...
Bharaneedharan Rathnasabapathy, Piotr J. Gmytrasie...
ATAL
2003
Springer
14 years 27 days ago
Transition-independent decentralized markov decision processes
There has been substantial progress with formal models for sequential decision making by individual agents using the Markov decision process (MDP). However, similar treatment of m...
Raphen Becker, Shlomo Zilberstein, Victor R. Lesse...
CORR
2006
Springer
130views Education» more  CORR 2006»
13 years 7 months ago
On optimal quantization rules for some sequential decision problems
We consider the problem of sequential decentralized detection, a problem that entails several interdependent choices: the choice of a stopping rule (specifying the sample size), a...
XuanLong Nguyen, Martin J. Wainwright, Michael I. ...
AAAI
2010
13 years 9 months ago
Robust Policy Computation in Reward-Uncertain MDPs Using Nondominated Policies
The precise specification of reward functions for Markov decision processes (MDPs) is often extremely difficult, motivating research into both reward elicitation and the robust so...
Kevin Regan, Craig Boutilier