Sciweavers

495 search results - page 58 / 99
» Constructing States for Reinforcement Learning
Sort
View
NIPS
1993
15 years 5 months ago
Using Local Trajectory Optimizers to Speed Up Global Optimization in Dynamic Programming
Dynamic programming provides a methodology to develop planners and controllers for nonlinear systems. However, general dynamic programming is computationally intractable. We have ...
Christopher G. Atkeson
128
Voted
ICML
2010
IEEE
15 years 5 months ago
Finite-Sample Analysis of LSTD
In this paper we consider the problem of policy evaluation in reinforcement learning, i.e., learning the value function of a fixed policy, using the least-squares temporal-differe...
Alessandro Lazaric, Mohammad Ghavamzadeh, Ré...
AVSS
2007
IEEE
15 years 10 months ago
Vehicular traffic density estimation via statistical methods with automated state learning
This paper proposes a novel approach of combining an unsupervised clustering scheme called AutoClass with Hidden Markov Models (HMMs) to determine the traffic density state in a R...
Evan Tan, Jing Chen
ICRA
2009
IEEE
121views Robotics» more  ICRA 2009»
15 years 10 months ago
Learning sequential visual attention control through dynamic state space discretization
² Similar to humans and primates, artificial creatures like robots are limited in terms of allocation of their resources to huge sensory and perceptual information. Serial process...
Ali Borji, Majid Nili Ahmadabadi, Babak Nadjar Ara...
ICRA
2010
IEEE
143views Robotics» more  ICRA 2010»
15 years 2 months ago
Apprenticeship learning via soft local homomorphisms
Abstract— We consider the problem of apprenticeship learning when the expert’s demonstration covers only a small part of a large state space. Inverse Reinforcement Learning (IR...
Abdeslam Boularias, Brahim Chaib-draa