Sciweavers

98 search results - page 6 / 20
» Using Rewards for Belief State Updates in Partially Observab...
Sort
View
ECSQARU
2001
Springer
14 years 2 days ago
Space-Progressive Value Iteration: An Anytime Algorithm for a Class of POMDPs
Abstract. Finding optimal policies for general partially observable Markov decision processes (POMDPs) is computationally difficult primarily due to the need to perform dynamic-pr...
Nevin Lianwen Zhang, Weihong Zhang
IJRR
2010
162views more  IJRR 2010»
13 years 6 months ago
Planning under Uncertainty for Robotic Tasks with Mixed Observability
Partially observable Markov decision processes (POMDPs) provide a principled, general framework for robot motion planning in uncertain and dynamic environments. They have been app...
Sylvie C. W. Ong, Shao Wei Png, David Hsu, Wee Sun...
ICML
1995
IEEE
14 years 8 months ago
Learning Policies for Partially Observable Environments: Scaling Up
Partially observable Markov decision processes (pomdp's) model decision problems in which an agent tries to maximize its reward in the face of limited and/or noisy sensor fee...
Michael L. Littman, Anthony R. Cassandra, Leslie P...
GLOBECOM
2010
IEEE
13 years 5 months ago
Cooperative Relay Scheduling under Partial State Information in Energy Harvesting Sensor Networks
Abstract--Sensors equipped with energy harvesting and cooperative communication capabilities are a viable solution to the power limitations of Wireless Sensor Networks (WSNs) assoc...
Huijiang Li, Neeraj Jaggi, Biplab Sikdar
ICML
2008
IEEE
14 years 8 months ago
Reinforcement learning with limited reinforcement: using Bayes risk for active learning in POMDPs
Partially Observable Markov Decision Processes (POMDPs) have succeeded in planning domains that require balancing actions that increase an agent's knowledge and actions that ...
Finale Doshi, Joelle Pineau, Nicholas Roy