arXiv Analytics

Sign in

arXiv:2105.00357 [cs.LG]AbstractReferencesReviewsResources

RotLSTM: Rotating Memories in Recurrent Neural Networks

Vlad Velici, Adam Prügel-Bennett

Published 2021-05-01Version 1

Long Short-Term Memory (LSTM) units have the ability to memorise and use long-term dependencies between inputs to generate predictions on time series data. We introduce the concept of modifying the cell state (memory) of LSTMs using rotation matrices parametrised by a new set of trainable weights. This addition shows significant increases of performance on some of the tasks from the bAbI dataset.

Related articles: Most relevant | Search more
arXiv:2103.15589 [cs.LG] (Published 2021-03-26)
Backpropagation Through Time For Networks With Long-Term Dependencies
arXiv:1809.05896 [cs.LG] (Published 2018-09-16)
Classifying Process Instances Using Recurrent Neural Networks
arXiv:1704.05119 [cs.LG] (Published 2017-04-17)
Exploring Sparsity in Recurrent Neural Networks