Sciweavers

NIPS
1996

Reinforcement Learning for Mixed Open-loop and Closed-loop Control

14 years 25 days ago
Reinforcement Learning for Mixed Open-loop and Closed-loop Control
Closed-loop control relies on sensory feedback that is usually assumed to be free. But if sensing incurs a cost, it may be coste ective to take sequences of actions in open-loop mode. We describe a reinforcement learning algorithm that learns to combine open-loop and closed-loop control when sensing incurs a cost. Although we assume reliable sensors, use of open-loop control means that actions must sometimes be taken when the current state of the controlled system is uncertain. This is a special case of the hidden-state problem in reinforcement learning, and to cope, our algorithm relies on short-term memory. The main result of the paper is a rule that signi cantly limits exploration of possible memory states by pruning memory states for which the estimated value of information is greater than its cost. We prove that this rule allows convergence to an optimal policy.
Eric A. Hansen, Andrew G. Barto, Shlomo Zilberstei
Added 02 Nov 2010
Updated 02 Nov 2010
Type Conference
Year 1996
Where NIPS
Authors Eric A. Hansen, Andrew G. Barto, Shlomo Zilberstein
Comments (0)