arXiv Analytics

Sign in

arXiv:1802.10410 [cs.LG]AbstractReferencesReviewsResources

Tensor Decomposition for Compressing Recurrent Neural Network

Andros Tjandra, Sakriani Sakti, Satoshi Nakamura

Published 2018-02-28Version 1

In the machine learning fields, Recurrent Neural Network (RNN) has become a popular algorithm for sequential data modeling. However, behind the impressive performance, RNNs require a large number of parameters for both training and inference. In this paper, we are trying to reduce the number of parameters and maintain the expressive power from RNN simultaneously. We utilize several tensor decompositions method including CANDECOMP/PARAFAC (CP), Tucker decomposition and Tensor Train(TT) to re-parameterize the Gated Recurrent Unit (GRU) RNN. We evaluate all tensor-based RNNs performance on sequence modeling tasks with a various number of parameters. Based on our experiment results, TT-GRU achieved the best results in a various number of parameters compared to other decomposition methods.

Related articles: Most relevant | Search more
arXiv:1705.08052 [cs.LG] (Published 2017-05-23)
Compressing Recurrent Neural Network with Tensor Train
arXiv:2306.13264 [cs.LG] (Published 2023-06-23)
FedSelect: Customized Selection of Parameters for Fine-Tuning during Personalized Federated Learning
arXiv:1806.01248 [cs.LG] (Published 2018-06-04)
Dynamically Hierarchy Revolution: DirNet for Compressing Recurrent Neural Network on Mobile Devices