Sciweavers

683 search results - page 38 / 137
» Coarticulation in Markov Decision Processes
Sort
View
ICML
2004
IEEE
14 years 9 months ago
Utile distinction hidden Markov models
This paper addresses the problem of constructing good action selection policies for agents acting in partially observable environments, a class of problems generally known as Part...
Daan Wierstra, Marco Wiering
SIGMETRICS
2000
ACM
105views Hardware» more  SIGMETRICS 2000»
14 years 1 months ago
Using the exact state space of a Markov model to compute approximate stationary measures
We present a new approximation algorithm based on an exact representation of the state space S, using decision diagrams, and of the transition rate matrix R, using Kronecker algeb...
Andrew S. Miner, Gianfranco Ciardo, Susanna Donate...
UAI
2000
13 years 10 months ago
PEGASUS: A policy search method for large MDPs and POMDPs
We propose a new approach to the problem of searching a space of policies for a Markov decision process (MDP) or a partially observable Markov decision process (POMDP), given a mo...
Andrew Y. Ng, Michael I. Jordan
IJFCS
2008
130views more  IJFCS 2008»
13 years 9 months ago
Equivalence of Labeled Markov Chains
We consider the equivalence problem for labeled Markov chains (LMCs), where each state is labeled with an observation. Two LMCs are equivalent if every finite sequence of observat...
Laurent Doyen, Thomas A. Henzinger, Jean-Fran&cced...
CVPR
2012
IEEE
11 years 11 months ago
RALF: A reinforced active learning formulation for object class recognition
Active learning aims to reduce the amount of labels required for classification. The main difficulty is to find a good trade-off between exploration and exploitation of the lab...
Sandra Ebert, Mario Fritz, Bernt Schiele