Sciweavers

499 search results - page 64 / 100
» Model Minimization in Markov Decision Processes
Sort
View
IJCAI
2007
13 years 10 months ago
The Value of Observation for Monitoring Dynamic Systems
We consider the fundamental problem of monitoring (i.e. tracking) the belief state in a dynamic system, when the model is only approximately correct and when the initial belief st...
Eyal Even-Dar, Sham M. Kakade, Yishay Mansour
CVPR
2008
IEEE
14 years 3 months ago
The statistical modelling of fingerprint minutiae distribution with implications for fingerprint individuality studies
The spatial distribution of fingerprint minutiae is a core problem in the fingerprint individuality study, the cornerstone of the fingerprint authentication technology. Previously...
Jiansheng Chen, Yiu Sang Moon
ICASSP
2010
IEEE
13 years 8 months ago
Automatic state discovery for unstructured audio scene classification
In this paper we present a novel scheme for unstructured audio scene classification that possesses three highly desirable and powerful features: autonomy, scalability, and robust...
Julian Ramos, Sajid M. Siddiqi, Artur Dubrawski, G...
PKDD
2010
Springer
129views Data Mining» more  PKDD 2010»
13 years 7 months ago
Smarter Sampling in Model-Based Bayesian Reinforcement Learning
Abstract. Bayesian reinforcement learning (RL) is aimed at making more efficient use of data samples, but typically uses significantly more computation. For discrete Markov Decis...
Pablo Samuel Castro, Doina Precup
ICML
2006
IEEE
14 years 9 months ago
Using inaccurate models in reinforcement learning
In the model-based policy search approach to reinforcement learning (RL), policies are found using a model (or "simulator") of the Markov decision process. However, for ...
Pieter Abbeel, Morgan Quigley, Andrew Y. Ng