Sciweavers

GECCO
2005
Springer

Co-evolving recurrent neurons learn deep memory POMDPs

14 years 5 months ago
Co-evolving recurrent neurons learn deep memory POMDPs
Recurrent neural networks are theoretically capable of learning complex temporal sequences, but training them through gradient-descent is too slow and unstable for practical use in reinforcement learning environments. Neuroevolution, the evolution of artificial neural networks using genetic algorithms, can potentially solve real-world reinforcement learning tasks that require deep use of memory, i.e. memory spanning hundreds or thousands of inputs, by searching the space of recurrent neural networks directly. In this paper, we introduce a new neuroevolution algorithm called Hierarchical Enforced SubPopulations that simultaneously evolves networks at two levels of granularity: full networks and network components or neurons. We demonstrate the method in two POMDP tasks that involve temporal dependencies of up to thousands of time-steps, and show that it is faster and simpler than the current best conventional reinforcement learning system on these tasks. Categories and Subject Descrip...
Faustino J. Gomez, Jürgen Schmidhuber
Added 27 Jun 2010
Updated 27 Jun 2010
Type Conference
Year 2005
Where GECCO
Authors Faustino J. Gomez, Jürgen Schmidhuber
Comments (0)