arXiv Analytics

Sign in

arXiv:1905.04967 [cs.LG]AbstractReferencesReviewsResources

Implicit Filter Sparsification In Convolutional Neural Networks

Dushyant Mehta, Kwang In Kim, Christian Theobalt

Published 2019-05-13Version 1

We show implicit filter level sparsity manifests in convolutional neural networks (CNNs) which employ Batch Normalization and ReLU activation, and are trained with adaptive gradient descent techniques and L2 regularization or weight decay. Through an extensive empirical study (Mehta et al., 2019) we hypothesize the mechanism behind the sparsification process, and find surprising links to certain filter sparsification heuristics proposed in literature. Emergence of, and the subsequent pruning of selective features is observed to be one of the contributing mechanisms, leading to feature sparsity at par or better than certain explicit sparsification / pruning approaches. In this workshop article we summarize our findings, and point out corollaries of selective-featurepenalization which could also be employed as heuristics for filter pruning

Comments: ODML-CDNNR 2019 (ICML'19 workshop) extended abstract of the CVPR 2019 paper "On Implicit Filter Level Sparsity in Convolutional Neural Networks, Mehta et al." (arXiv:1811.12495)
Categories: cs.LG, cs.CV, stat.ML
Related articles: Most relevant | Search more
arXiv:1809.10463 [cs.LG] (Published 2018-09-27)
Learning to Train a Binary Neural Network
arXiv:1511.06067 [cs.LG] (Published 2015-11-19)
Convolutional neural networks with low-rank regularization
arXiv:1707.09641 [cs.LG] (Published 2017-07-30)
Visual Explanations for Convolutional Neural Networks via Input Resampling