arXiv Analytics

Sign in

arXiv:2307.09994 [cs.LG]AbstractReferencesReviewsResources

Impact of Disentanglement on Pruning Neural Networks

Carl Shneider, Peyman Rostami, Anis Kacem, Nilotpal Sinha, Abd El Rahman Shabayek, Djamila Aouada

Published 2023-07-19Version 1

Deploying deep learning neural networks on edge devices, to accomplish task specific objectives in the real-world, requires a reduction in their memory footprint, power consumption, and latency. This can be realized via efficient model compression. Disentangled latent representations produced by variational autoencoder (VAE) networks are a promising approach for achieving model compression because they mainly retain task-specific information, discarding useless information for the task at hand. We make use of the Beta-VAE framework combined with a standard criterion for pruning to investigate the impact of forcing the network to learn disentangled representations on the pruning process for the task of classification. In particular, we perform experiments on MNIST and CIFAR10 datasets, examine disentanglement challenges, and propose a path forward for future works.

Related articles: Most relevant | Search more
arXiv:2009.08576 [cs.LG] (Published 2020-09-18)
Pruning Neural Networks at Initialization: Why are We Missing the Mark?
arXiv:2012.08749 [cs.LG] (Published 2020-12-16)
Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks
arXiv:2411.00273 [cs.LG] (Published 2024-11-01)
Efficient Model Compression for Bayesian Neural Networks