arXiv Analytics

Sign in

arXiv:2308.06780 [cs.LG]AbstractReferencesReviewsResources

Neural Networks at a Fraction with Pruned Quaternions

Sahel Mohammad Iqbal, Subhankar Mishra

Published 2023-08-13Version 1

Contemporary state-of-the-art neural networks have increasingly large numbers of parameters, which prevents their deployment on devices with limited computational power. Pruning is one technique to remove unnecessary weights and reduce resource requirements for training and inference. In addition, for ML tasks where the input data is multi-dimensional, using higher-dimensional data embeddings such as complex numbers or quaternions has been shown to reduce the parameter count while maintaining accuracy. In this work, we conduct pruning on real and quaternion-valued implementations of different architectures on classification tasks. We find that for some architectures, at very high sparsity levels, quaternion models provide higher accuracies than their real counterparts. For example, at the task of image classification on CIFAR-10 using Conv-4, at $3\%$ of the number of parameters as the original model, the pruned quaternion version outperforms the pruned real by more than $10\%$. Experiments on various network architectures and datasets show that for deployment in extremely resource-constrained environments, a sparse quaternion network might be a better candidate than a real sparse model of similar architecture.

Journal: In Proceedings of the 6th Joint International Conference on Data Science & Management of Data (CODS-COMAD '23). ACM. 19-27 (2023)
Categories: cs.LG, cs.CV
Related articles: Most relevant | Search more
arXiv:1511.05497 [cs.LG] (Published 2015-11-17)
Learning the Architecture of Deep Neural Networks
arXiv:2205.10662 [cs.LG] (Published 2022-05-21)
Equivariant Mesh Attention Networks
arXiv:2210.09459 [cs.LG] (Published 2022-10-17)
Extensible Proxy for Efficient NAS