Sciweavers

71 search results - page 7 / 15
» A Behavior Adaptation Algorithm based on Hierarchical Partia...
Sort
View
CORR
2010
Springer
105views Education» more  CORR 2010»
13 years 6 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
AAAI
2004
13 years 8 months ago
Stochastic Local Search for POMDP Controllers
The search for finite-state controllers for partially observable Markov decision processes (POMDPs) is often based on approaches like gradient ascent, attractive because of their ...
Darius Braziunas, Craig Boutilier
ICTAI
2005
IEEE
14 years 1 months ago
Planning with POMDPs Using a Compact, Logic-Based Representation
Partially Observable Markov Decision Processes (POMDPs) provide a general framework for AI planning, but they lack the structure for representing real world planning problems in a...
Chenggang Wang, James G. Schmolze
AIPS
2008
13 years 9 months ago
HiPPo: Hierarchical POMDPs for Planning Information Processing and Sensing Actions on a Robot
Flexible general purpose robots need to tailor their visual processing to their task, on the fly. We propose a new approach to this within a planning framework, where the goal is ...
Mohan Sridharan, Jeremy L. Wyatt, Richard Dearden
AAAI
2007
13 years 9 months ago
Scaling Up: Solving POMDPs through Value Based Clustering
Partially Observable Markov Decision Processes (POMDPs) provide an appropriately rich model for agents operating under partial knowledge of the environment. Since finding an opti...
Yan Virin, Guy Shani, Solomon Eyal Shimony, Ronen ...