Sciweavers

771 search results - page 95 / 155
» Markov Decision Processes with Arbitrary Reward Processes
Sort
View
AIPS
1998
15 years 5 months ago
Solving Stochastic Planning Problems with Large State and Action Spaces
Planning methods for deterministic planning problems traditionally exploit factored representations to encode the dynamics of problems in terms of a set of parameters, e.g., the l...
Thomas Dean, Robert Givan, Kee-Eung Kim
SBMF
2009
Springer
126views Formal Methods» more  SBMF 2009»
15 years 8 months ago
Undecidability Results for Distributed Probabilistic Systems
Abstract. In the verification of concurrent systems involving probabilities, the aim is to find out the maximum/minimum probability that a given event occurs (examples of such ev...
Sergio Giro
TSP
2010
14 years 10 months ago
Universal randomized switching
Abstract--In this paper, we consider a competitive approach to sequential decision problems, suitable for a variety of signal processing applications where at each of a succession ...
Suleyman Serdar Kozat, Andrew C. Singer
UAI
2000
15 years 5 months ago
Value-Directed Belief State Approximation for POMDPs
We consider the problem belief-state monitoring for the purposes of implementing a policy for a partially-observable Markov decision process (POMDP), specifically how one might ap...
Pascal Poupart, Craig Boutilier
JMLR
2010
125views more  JMLR 2010»
14 years 10 months ago
Variational methods for Reinforcement Learning
We consider reinforcement learning as solving a Markov decision process with unknown transition distribution. Based on interaction with the environment, an estimate of the transit...
Thomas Furmston, David Barber