Sciweavers

166 search results - page 22 / 34
» Safe exploration for reinforcement learning
Sort
View
ICML
2009
IEEE
14 years 9 months ago
Near-Bayesian exploration in polynomial time
We consider the exploration/exploitation problem in reinforcement learning (RL). The Bayesian approach to model-based RL offers an elegant solution to this problem, by considering...
J. Zico Kolter, Andrew Y. Ng
IHI
2010
197views Healthcare» more  IHI 2010»
13 years 3 months ago
Beyond safe harbor: automatic discovery of health information de-identification policy alternatives
Regulations in various countries permit the reuse of health information without patient authorization provided the data is "de-identified". In the United States, for ins...
Kathleen Benitez, Grigorios Loukides, Bradley Mali...
NIPS
2003
13 years 10 months ago
Approximate Planning in POMDPs with Macro-Actions
Recent research has demonstrated that useful POMDP solutions do not require consideration of the entire belief space. We extend this idea with the notion of temporal abstraction. ...
Georgios Theocharous, Leslie Pack Kaelbling
CORR
2012
Springer
216views Education» more  CORR 2012»
12 years 4 months ago
Fractional Moments on Bandit Problems
Reinforcement learning addresses the dilemma between exploration to find profitable actions and exploitation to act according to the best observations already made. Bandit proble...
Ananda Narayanan B., Balaraman Ravindran
PKDD
2010
Springer
122views Data Mining» more  PKDD 2010»
13 years 7 months ago
Exploration in Relational Worlds
Abstract. One of the key problems in model-based reinforcement learning is balancing exploration and exploitation. Another is learning and acting in large relational domains, in wh...
Tobias Lang, Marc Toussaint, Kristian Kersting