Sciweavers

91 search results - page 2 / 19
» Event-based control using quadratic approximate value functi...
Sort
View
NN
2010
Springer
187views Neural Networks» more  NN 2010»
13 years 2 months ago
Efficient exploration through active learning for value function approximation in reinforcement learning
Appropriately designing sampling policies is highly important for obtaining better control policies in reinforcement learning. In this paper, we first show that the least-squares ...
Takayuki Akiyama, Hirotaka Hachiya, Masashi Sugiya...
CDC
2010
IEEE
13 years 2 months ago
Stochastic approximation for consensus with general time-varying weight matrices
This paper considers consensus problems with delayed noisy measurements, and stochastic approximation is used to achieve mean square consensus. For stochastic approximation based c...
Minyi Huang
AAAI
2008
13 years 9 months ago
Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation
Off-policy reinforcement learning is aimed at efficiently reusing data samples gathered in the past, which is an essential problem for physically grounded AI as experiments are us...
Hirotaka Hachiya, Takayuki Akiyama, Masashi Sugiya...
SIAMCO
2002
121views more  SIAMCO 2002»
13 years 7 months ago
Consistent Approximations and Approximate Functions and Gradients in Optimal Control
As shown in [7], optimal control problems with either ODE or PDE dynamics can be solved efficiently using a setting of consistent approximations obtained by numerical discretizati...
Olivier Pironneau, Elijah Polak
JMLR
2006
143views more  JMLR 2006»
13 years 7 months ago
Geometric Variance Reduction in Markov Chains: Application to Value Function and Gradient Estimation
We study a sequential variance reduction technique for Monte Carlo estimation of functionals in Markov Chains. The method is based on designing sequential control variates using s...
Rémi Munos