Sciweavers

102 search results - page 10 / 21
» MDPs with Non-Deterministic Policies
Sort
View
ACL
2000
13 years 8 months ago
Spoken Dialogue Management Using Probabilistic Reasoning
Spoken dialogue managers have benefited from using stochastic planners such as Markov Decision Processes (MDPs). However, so far, MDPs do not handle well noisy and ambiguous speec...
Nicholas Roy, Joelle Pineau, Sebastian Thrun
AAAI
1997
13 years 8 months ago
Structured Solution Methods for Non-Markovian Decision Processes
Markov Decision Processes (MDPs), currently a popular method for modeling and solving decision theoretic planning problems, are limited by the Markovian assumption: rewards and dy...
Fahiem Bacchus, Craig Boutilier, Adam J. Grove
ICML
2008
IEEE
14 years 8 months ago
Learning all optimal policies with multiple criteria
We describe an algorithm for learning in the presence of multiple criteria. Our technique generalizes previous approaches in that it can learn optimal policies for all linear pref...
Leon Barrett, Srini Narayanan
AAAI
2006
13 years 9 months ago
Factored MDP Elicitation and Plan Display
The software suite we will demonstrate at AAAI '06 was designed around planning with factored Markov decision processes (MDPs). It is a user-friendly suite that facilitates d...
Krol Kevin Mathias, Casey Lengacher, Derek William...
ML
2002
ACM
154views Machine Learning» more  ML 2002»
13 years 7 months ago
Technical Update: Least-Squares Temporal Difference Learning
TD() is a popular family of algorithms for approximate policy evaluation in large MDPs. TD() works by incrementally updating the value function after each observed transition. It h...
Justin A. Boyan