Sciweavers

499 search results - page 72 / 100
» Model Minimization in Markov Decision Processes
Sort
View
ICRA
2003
IEEE
167views Robotics» more  ICRA 2003»
14 years 1 months ago
Local exploration: online algorithms and a probabilistic framework
— Mapping an environment with an imaging sensor becomes very challenging if the environment to be mapped is unknown and has to be explored. Exploration involves the planning of v...
Volkan Isler, Sampath Kannan, Kostas Daniilidis
AAAI
2007
13 years 11 months ago
Authorial Idioms for Target Distributions in TTD-MDPs
In designing Markov Decision Processes (MDP), one must define the world, its dynamics, a set of actions, and a reward function. MDPs are often applied in situations where there i...
David L. Roberts, Sooraj Bhat, Kenneth St. Clair, ...
ICML
2006
IEEE
14 years 9 months ago
An analytic solution to discrete Bayesian reinforcement learning
Reinforcement learning (RL) was originally proposed as a framework to allow agents to learn in an online fashion as they interact with their environment. Existing RL algorithms co...
Pascal Poupart, Nikos A. Vlassis, Jesse Hoey, Kevi...
AAAI
2007
13 years 11 months ago
Continuous State POMDPs for Object Manipulation Tasks
My research focus is on using continuous state partially observable Markov decision processes (POMDPs) to perform object manipulation tasks using a robotic arm. During object mani...
Emma Brunskill
ICPR
2008
IEEE
14 years 9 months ago
HOPS: Efficient region labeling using Higher Order Proxy Neighborhoods
We present the Higher Order Proxy Neighborhoods (HOPS) approach to modeling higher order neighborhoods in Markov Random Fields (MRFs). HOPS incorporates more context information i...
Albert Y. C. Chen, Jason J. Corso, Le Wang