Sciweavers

85 search results - page 7 / 17
» Solving Stochastic Planning Problems with Large State and Ac...
Sort
View
IAT
2005
IEEE
14 years 1 months ago
Decomposing Large-Scale POMDP Via Belief State Analysis
Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing ...
Xin Li, William K. Cheung, Jiming Liu
VALUETOOLS
2006
ACM
176views Hardware» more  VALUETOOLS 2006»
14 years 1 months ago
How to solve large scale deterministic games with mean payoff by policy iteration
Min-max functions are dynamic programming operators of zero-sum deterministic games with finite state and action spaces. The problem of computing the linear growth rate of the or...
Vishesh Dhingra, Stephane Gaubert
IJCAI
2001
13 years 8 months ago
Symbolic Dynamic Programming for First-Order MDPs
We present a dynamic programming approach for the solution of first-order Markov decisions processes. This technique uses an MDP whose dynamics is represented in a variant of the ...
Craig Boutilier, Raymond Reiter, Bob Price
AAAI
2011
12 years 7 months ago
Linear Dynamic Programs for Resource Management
Sustainable resource management in many domains presents large continuous stochastic optimization problems, which can often be modeled as Markov decision processes (MDPs). To solv...
Marek Petrik, Shlomo Zilberstein
AI
2000
Springer
13 years 7 months ago
Stochastic dynamic programming with factored representations
Markov decisionprocesses(MDPs) haveproven to be popular models for decision-theoretic planning, but standard dynamic programming algorithms for solving MDPs rely on explicit, stat...
Craig Boutilier, Richard Dearden, Moisés Go...