Sciweavers

499 search results - page 43 / 100
» Model Minimization in Markov Decision Processes
Sort
View
CORR
2006
Springer
113views Education» more  CORR 2006»
13 years 8 months ago
A Unified View of TD Algorithms; Introducing Full-Gradient TD and Equi-Gradient Descent TD
This paper addresses the issue of policy evaluation in Markov Decision Processes, using linear function approximation. It provides a unified view of algorithms such as TD(), LSTD()...
Manuel Loth, Philippe Preux
WEBI
2010
Springer
13 years 6 months ago
Impacts of Analysts' Cognitive Styles on the Analytic Process
A user's cognitive style has been found to affect how they search for information, how they analyze the information, and how they make decisions in an analytical process. In ...
Eugene Santos Jr., Hien Nguyen, Fei Yu, Deqing Li,...
ICCV
2011
IEEE
12 years 8 months ago
Perturb-and-MAP Random Fields: Using Discrete Optimization\\to Learn and Sample from Energy Models
We propose a novel way to induce a random field from an energy function on discrete labels. It amounts to locally injecting noise to the energy potentials, followed by finding t...
George Papandreou, Alan L. Yuille
ICIP
2003
IEEE
14 years 10 months ago
A probabilistic framework for image segmentation
A new probabilistic image segmentation model based on hypothesis testing and Gibbs Random Fields is introduced. First, a probabilistic difference measure derived from a set of hyp...
Slawo Wesolkowski, Paul W. Fieguth
PRICAI
2000
Springer
14 years 9 days ago
Generating Hierarchical Structure in Reinforcement Learning from State Variables
This paper presents the CQ algorithm which decomposes and solves a Markov Decision Process (MDP) by automatically generating a hierarchy of smaller MDPs using state variables. The ...
Bernhard Hengst