Sciweavers

107 search results - page 8 / 22
» Approximate Linear Programming for Constrained Partially Obs...
Sort
View
ICML
2006
IEEE
14 years 1 months ago
Automatic basis function construction for approximate dynamic programming and reinforcement learning
We address the problem of automatically constructing basis functions for linear approximation of the value function of a Markov Decision Process (MDP). Our work builds on results ...
Philipp W. Keller, Shie Mannor, Doina Precup
IJCAI
2007
13 years 9 months ago
Learning from Partial Observations
We present a general machine learning framework for modelling the phenomenon of missing information in data. We propose a masking process model to capture the stochastic nature of...
Loizos Michael
UAI
2008
13 years 9 months ago
Partitioned Linear Programming Approximations for MDPs
Approximate linear programming (ALP) is an efficient approach to solving large factored Markov decision processes (MDPs). The main idea of the method is to approximate the optimal...
Branislav Kveton, Milos Hauskrecht
IJRR
2010
162views more  IJRR 2010»
13 years 6 months ago
Planning under Uncertainty for Robotic Tasks with Mixed Observability
Partially observable Markov decision processes (POMDPs) provide a principled, general framework for robot motion planning in uncertain and dynamic environments. They have been app...
Sylvie C. W. Ong, Shao Wei Png, David Hsu, Wee Sun...
IJCAI
2003
13 years 8 months ago
Taming Decentralized POMDPs: Towards Efficient Policy Computation for Multiagent Settings
The problem of deriving joint policies for a group of agents that maximize some joint reward function can be modeled as a decentralized partially observable Markov decision proces...
Ranjit Nair, Milind Tambe, Makoto Yokoo, David V. ...