Sciweavers

129 search results - page 12 / 26
» Automatic Recovery Using Bounded Partially Observable Markov...
Sort
View
ICMLA
2009
13 years 5 months ago
Sensitivity Analysis of POMDP Value Functions
In sequential decision making under uncertainty, as in many other modeling endeavors, researchers observe a dynamical system and collect data measuring its behavior over time. The...
Stéphane Ross, Masoumeh T. Izadi, Mark Merc...
AROBOTS
2008
166views more  AROBOTS 2008»
13 years 6 months ago
User-adapted plan recognition and user-adapted shared control: A Bayesian approach to semi-autonomous wheelchair driving
Abstract Many elderly and physically impaired people experience difficulties when maneuvering a powered wheelchair. In order to provide improved maneuvering, powered wheelchairs ha...
Eric Demeester, Alexander Hüntemann, Dirk Van...
CORR
2010
Springer
105views Education» more  CORR 2010»
13 years 6 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
ICTAI
2005
IEEE
14 years 1 months ago
Planning with POMDPs Using a Compact, Logic-Based Representation
Partially Observable Markov Decision Processes (POMDPs) provide a general framework for AI planning, but they lack the structure for representing real world planning problems in a...
Chenggang Wang, James G. Schmolze
ATAL
2007
Springer
13 years 11 months ago
Interactive dynamic influence diagrams
This paper extends the framework of dynamic influence diagrams (DIDs) to the multi-agent setting. DIDs are computational representations of the Partially Observable Markov Decisio...
Kyle Polich, Piotr J. Gmytrasiewicz