Sciweavers

129 search results - page 17 / 26
» Automatic Recovery Using Bounded Partially Observable Markov...
Sort
View
ATAL
2007
Springer
14 years 1 months ago
Graphical models for online solutions to interactive POMDPs
We develop a new graphical representation for interactive partially observable Markov decision processes (I-POMDPs) that is significantly more transparent and semantically clear t...
Prashant Doshi, Yifeng Zeng, Qiongyu Chen
ATAL
2005
Springer
14 years 1 months ago
Exploiting belief bounds: practical POMDPs for personal assistant agents
Agents or agent teams deployed to assist humans often face the challenges of monitoring the state of key processes in their environment (including the state of their human users t...
Pradeep Varakantham, Rajiv T. Maheswaran, Milind T...
NIPS
2001
13 years 9 months ago
Variance Reduction Techniques for Gradient Estimates in Reinforcement Learning
Policy gradient methods for reinforcement learning avoid some of the undesirable properties of the value function approaches, such as policy degradation (Baxter and Bartlett, 2001...
Evan Greensmith, Peter L. Bartlett, Jonathan Baxte...
JAIR
2006
160views more  JAIR 2006»
13 years 7 months ago
Anytime Point-Based Approximations for Large POMDPs
The Partially Observable Markov Decision Process has long been recognized as a rich framework for real-world planning and control problems, especially in robotics. However exact s...
Joelle Pineau, Geoffrey J. Gordon, Sebastian Thrun
MOR
2008
87views more  MOR 2008»
13 years 7 months ago
On Near Optimality of the Set of Finite-State Controllers for Average Cost POMDP
We consider the average cost problem for partially observable Markov decision processes (POMDP) with finite state, observation, and control spaces. We prove that there exists an -...
Huizhen Yu, Dimitri P. Bertsekas