Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error
Glüge, Stefan; ; ; (2014). Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error. Neurocomputing, 141. 54 - 64. Peer reviewed.
In general, recurrent neural networks have difficulties in learning long-term dependencies. The segmented-memory recurrent neural network (SMRNN) architecture together with the extended realtime recurrent learning (eRTRL) algorithm was proposed to circumvent this problem. Due to its computational complexity eRTRL becomes impractical with increasing network size. Therefore, we introduce the less complex extended backpropagation through time (eBPTT) for SMRNN together with a layer-local unsupervised pre-training procedure. A comparison on the information latching problem showed that eRTRL is better able to handle the latching of information over longer periods of time, even
though eBPTT guaranteed a better generalisation when training was successful. Further, pre-training significantly improved the ability to learn long-term dependencies with eBPTT. Therefore, the proposed eBPTT algorithm is suited for tasks that require big networks where eRTRL is impractical. The pretraining
procedure itself is independent of the supervised learning algorithm and can improve learning in SMRNN in general.