Sciweavers

1138 search results - page 10 / 228
» Feature Markov Decision Processes
Sort
View
AAAI
2004
13 years 8 months ago
Solving Concurrent Markov Decision Processes
Typically, Markov decision problems (MDPs) assume a single action is executed per decision epoch, but in the real world one may frequently execute certain actions in parallel. Thi...
Mausam, Daniel S. Weld
AAAI
2007
13 years 9 months ago
Purely Epistemic Markov Decision Processes
Planning under uncertainty involves two distinct sources of uncertainty: uncertainty about the effects of actions and uncertainty about the current state of the world. The most wi...
Régis Sabbadin, Jérôme Lang, N...
ECAI
2010
Springer
13 years 8 months ago
On Finding Compromise Solutions in Multiobjective Markov Decision Processes
A Markov Decision Process (MDP) is a general model for solving planning problems under uncertainty. It has been extended to multiobjective MDP to address multicriteria or multiagen...
Patrice Perny, Paul Weng
CVIU
2010
163views more  CVIU 2010»
13 years 7 months ago
Automated handwashing assistance for persons with dementia using video and a partially observable Markov decision process
This paper presents a real-time vision-based system to assist a person with dementia wash their hands. The system uses only video inputs, and assistance is given as either verbal ...
Jesse Hoey, Pascal Poupart, Axel von Bertoldi, Tam...
ECML
2005
Springer
14 years 28 days ago
Using Rewards for Belief State Updates in Partially Observable Markov Decision Processes
Partially Observable Markov Decision Processes (POMDP) provide a standard framework for sequential decision making in stochastic environments. In this setting, an agent takes actio...
Masoumeh T. Izadi, Doina Precup