Sciweavers

332 search results - page 10 / 67
» Ranking policies in discrete Markov decision processes
Sort
View
NIPS
2004
13 years 9 months ago
Experts in a Markov Decision Process
We consider an MDP setting in which the reward function is allowed to change during each time step of play (possibly in an adversarial manner), yet the dynamics remain fixed. Simi...
Eyal Even-Dar, Sham M. Kakade, Yishay Mansour
TSMC
2011
258views more  TSMC 2011»
13 years 2 months ago
Cross-Entropy Optimization of Control Policies With Adaptive Basis Functions
—This paper introduces an algorithm for direct search of control policies in continuous-state discrete-action Markov decision processes. The algorithm looks for the best closed-l...
Lucian Busoniu, Damien Ernst, Bart De Schutter, Ro...
HCI
2009
13 years 5 months ago
Partially Observable Markov Decision Process (POMDP) Technologies for Sign Language Based Human-Computer Interaction
Sign language (SL) recognition modules in human-computer interaction systems need to be both fast and reliable. In cases where multiple sets of features are extracted from the SL d...
Sylvie C. W. Ong, David Hsu, Wee Sun Lee, Hanna Ku...
LICS
2007
IEEE
14 years 2 months ago
Limits of Multi-Discounted Markov Decision Processes
Markov decision processes (MDPs) are controllable discrete event systems with stochastic transitions. The payoff received by the controller can be evaluated in different ways, dep...
Hugo Gimbert, Wieslaw Zielonka
STACS
2007
Springer
14 years 2 months ago
Pure Stationary Optimal Strategies in Markov Decision Processes
Markov decision processes (MDPs) are controllable discrete event systems with stochastic transitions. Performances of an MDP are evaluated by a payoff function. The controller of ...
Hugo Gimbert