Normal fuzzy CMAC neural network performs well because of its fast learning speed and local generalization capability for approximating nonlinear functions. However, it requires hu...
Floriberto Ortiz Rodriguez, Wen Yu, Marco A. Moren...
Abstract. We apply Long Short-Term Memory (LSTM) recurrent neural networks to a large corpus of unprompted speech- the German part of the VERBMOBIL corpus. Training first on a fra...
Nicole Beringer, Alex Graves, Florian Schiel, J&uu...
Training recurrent neural networks is hard. Recently it has however been discovered that it is possible to just construct a random recurrent topology, and only train a single linea...
Benjamin Schrauwen, David Verstraeten, Jan M. Van ...
—Recurrent neural networks processing symbolic strings can be regarded as adaptive neural parsers. Given a set of positive and negative examples, picked up from a given language,...
Marco Gori, Marco Maggini, Enrico Martinelli, Giov...
Paper [1] aimed at providing a unified presentation of neural network architectures. We show in the present comment (i) that the canonical form of recurrent neural networks presen...