Sciweavers

682 search results - page 54 / 137
» One-Counter Markov Decision Processes
Sort
View
127
Voted
ICML
2006
IEEE
16 years 4 months ago
PAC model-free reinforcement learning
For a Markov Decision Process with finite state (size S) and action spaces (size A per state), we propose a new algorithm--Delayed Q-Learning. We prove it is PAC, achieving near o...
Alexander L. Strehl, Lihong Li, Eric Wiewiora, Joh...
119
Voted
ICML
2006
IEEE
16 years 4 months ago
An intrinsic reward mechanism for efficient exploration
How should a reinforcement learning agent act if its sole purpose is to efficiently learn an optimal policy for later use? In other words, how should it explore, to be able to exp...
Özgür Simsek, Andrew G. Barto
120
Voted
ICRA
2007
IEEE
126views Robotics» more  ICRA 2007»
15 years 10 months ago
A formal framework for robot learning and control under model uncertainty
— While the Partially Observable Markov Decision Process (POMDP) provides a formal framework for the problem of robot control under uncertainty, it typically assumes a known and ...
Robin Jaulmes, Joelle Pineau, Doina Precup
ICML
2006
IEEE
15 years 9 months ago
Automatic basis function construction for approximate dynamic programming and reinforcement learning
We address the problem of automatically constructing basis functions for linear approximation of the value function of a Markov Decision Process (MDP). Our work builds on results ...
Philipp W. Keller, Shie Mannor, Doina Precup
ICVS
2001
Springer
15 years 8 months ago
Adapting Object Recognition across Domains: A Demonstration
High-level vision systems use object, scene or domain specific knowledge to interpret images. Unfortunately, this knowledge has to be acquired for every domain. This makes it diffi...
Bruce A. Draper, Ulrike Ahlrichs, Dietrich Paulus